JoPmt commited on
Commit
6eba67d
·
verified ·
1 Parent(s): cd896eb

Upload model

Browse files
sam/adapter_config.json CHANGED
@@ -22,198 +22,198 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
 
 
25
  "up_blocks.1.attentions.2.proj_in",
26
- "down_blocks.2.attentions.0.proj_out",
27
- "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v",
28
- "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0",
29
- "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k",
30
- "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k",
31
- "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0",
32
- "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.2",
33
- "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k",
34
- "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k",
35
- "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k",
36
- "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k",
37
- "up_blocks.3.attentions.2.proj_out",
38
- "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj",
39
- "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q",
40
- "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0",
41
- "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q",
42
- "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj",
43
- "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0",
44
- "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k",
45
- "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v",
46
- "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj",
47
  "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v",
48
- "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.2",
49
- "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0",
50
- "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0",
51
  "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k",
52
- "down_blocks.0.attentions.1.proj_in",
53
- "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v",
54
- "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k",
55
- "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v",
56
  "up_blocks.3.attentions.1.proj_in",
57
- "up_blocks.1.attentions.1.proj_out",
 
 
 
 
 
 
 
 
 
 
 
58
  "mid_block.attentions.0.transformer_blocks.0.attn1.to_q",
59
- "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q",
60
- "mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0",
61
- "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k",
62
  "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.2",
63
- "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k",
64
- "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k",
65
- "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q",
66
- "up_blocks.1.attentions.0.proj_out",
67
  "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
  "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k",
69
- "mid_block.attentions.0.proj_in",
70
- "up_blocks.2.attentions.2.proj_out",
71
- "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0",
72
- "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k",
 
 
 
 
 
 
 
 
 
 
 
 
73
  "mid_block.attentions.0.proj_out",
74
- "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.2",
 
 
 
 
 
 
75
  "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0",
76
- "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj",
77
- "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v",
78
- "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2",
79
- "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k",
80
- "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.2",
81
- "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0",
82
- "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2",
 
 
 
83
  "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k",
 
84
  "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q",
85
- "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.2",
86
- "down_blocks.0.attentions.1.proj_out",
87
- "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
88
  "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.2",
89
- "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v",
90
- "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0",
91
- "mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj",
92
- "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k",
93
- "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0",
94
- "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q",
95
- "down_blocks.1.attentions.0.proj_in",
96
- "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q",
97
  "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v",
98
- "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q",
99
- "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2",
100
- "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q",
101
- "down_blocks.1.attentions.1.proj_out",
102
- "up_blocks.2.attentions.0.proj_in",
103
  "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q",
104
- "down_blocks.0.attentions.0.proj_in",
105
- "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k",
106
  "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q",
107
- "up_blocks.3.attentions.0.proj_out",
108
- "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v",
109
- "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k",
110
- "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k",
111
- "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v",
112
- "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k",
113
- "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj",
114
- "down_blocks.2.attentions.1.proj_in",
115
- "up_blocks.3.attentions.1.proj_out",
116
- "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.2",
117
- "up_blocks.1.attentions.2.proj_out",
118
- "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k",
119
- "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k",
120
- "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q",
121
  "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v",
122
- "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2",
 
 
 
123
  "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0",
124
- "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q",
125
- "mid_block.attentions.0.transformer_blocks.0.attn2.to_v",
126
- "up_blocks.3.attentions.0.proj_in",
127
- "up_blocks.1.attentions.0.proj_in",
128
- "mid_block.attentions.0.transformer_blocks.0.ff.net.2",
129
- "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q",
130
- "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v",
131
- "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q",
132
- "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2",
133
- "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0",
134
- "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.0.proj",
135
- "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v",
136
- "up_blocks.2.attentions.1.proj_out",
137
- "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0",
138
- "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.0.proj",
139
- "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.0.proj",
140
  "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k",
141
- "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0",
142
- "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v",
143
- "mid_block.attentions.0.transformer_blocks.0.attn1.to_v",
144
  "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v",
 
 
 
 
 
 
 
145
  "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.2",
146
- "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v",
147
- "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.0.proj",
148
- "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj",
149
- "mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0",
150
- "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0",
151
  "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q",
152
- "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v",
153
- "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q",
154
- "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj",
155
- "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v",
156
- "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q",
157
- "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q",
158
- "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v",
159
- "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0",
160
- "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v",
161
  "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0",
162
- "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v",
163
- "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0",
164
- "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0",
165
- "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q",
166
- "up_blocks.2.attentions.0.proj_out",
167
- "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k",
168
- "up_blocks.3.attentions.2.proj_in",
169
- "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q",
170
- "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q",
171
- "down_blocks.2.attentions.1.proj_out",
172
  "down_blocks.1.attentions.0.proj_out",
173
- "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj",
174
- "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v",
175
- "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0",
176
- "mid_block.attentions.0.transformer_blocks.0.attn2.to_k",
177
- "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0",
178
- "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v",
179
- "up_blocks.2.attentions.1.proj_in",
180
- "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q",
181
- "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q",
182
- "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q",
183
- "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0",
184
- "down_blocks.0.attentions.0.proj_out",
185
- "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0",
186
- "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.2",
187
- "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v",
188
- "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0",
189
- "down_blocks.1.attentions.1.proj_in",
190
  "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q",
191
- "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q",
192
- "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v",
193
- "up_blocks.1.attentions.1.proj_in",
194
- "mid_block.attentions.0.transformer_blocks.0.attn2.to_q",
195
  "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj",
196
- "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k",
197
- "mid_block.attentions.0.transformer_blocks.0.attn1.to_k",
198
- "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj",
199
- "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q",
200
- "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0",
201
- "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v",
202
- "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v",
203
- "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q",
204
- "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k",
205
- "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v",
206
- "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q",
207
- "up_blocks.2.attentions.2.proj_in",
208
- "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0",
209
- "down_blocks.2.attentions.0.proj_in",
210
- "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k",
211
  "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0",
212
- "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0",
213
- "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0",
214
- "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v",
215
- "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k",
216
- "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj"
217
  ],
218
  "task_type": null,
219
  "use_dora": false,
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v",
26
+ "mid_block.attentions.0.transformer_blocks.0.attn2.to_k",
27
  "up_blocks.1.attentions.2.proj_in",
28
+ "up_blocks.1.attentions.2.proj_out",
29
+ "mid_block.attentions.0.transformer_blocks.0.attn2.to_v",
30
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q",
31
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k",
32
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q",
33
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v",
35
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k",
 
 
36
  "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k",
37
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k",
38
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k",
39
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0",
40
+ "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0",
41
  "up_blocks.3.attentions.1.proj_in",
42
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v",
43
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v",
44
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v",
45
+ "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj",
46
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0",
47
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k",
48
+ "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k",
49
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q",
50
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q",
51
+ "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.0.proj",
52
+ "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj",
53
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0",
54
  "mid_block.attentions.0.transformer_blocks.0.attn1.to_q",
 
 
 
55
  "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.2",
 
 
 
 
56
  "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v",
57
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0",
58
+ "down_blocks.1.attentions.1.proj_out",
59
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v",
60
+ "down_blocks.2.attentions.0.proj_out",
61
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q",
62
+ "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k",
63
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k",
64
+ "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj",
65
+ "up_blocks.3.attentions.0.proj_in",
66
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0",
67
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v",
68
+ "up_blocks.2.attentions.0.proj_in",
69
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q",
70
+ "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2",
71
+ "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2",
72
+ "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj",
73
+ "mid_block.attentions.0.transformer_blocks.0.attn2.to_q",
74
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v",
75
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q",
76
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q",
77
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0",
78
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k",
79
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0",
80
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k",
81
+ "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0",
82
+ "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2",
83
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v",
84
  "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k",
85
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0",
86
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k",
87
+ "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0",
88
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0",
89
+ "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj",
90
+ "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.2",
91
+ "mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj",
92
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0",
93
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k",
94
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v",
95
+ "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v",
96
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0",
97
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0",
98
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k",
99
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0",
100
+ "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0",
101
  "mid_block.attentions.0.proj_out",
102
+ "down_blocks.0.attentions.0.proj_in",
103
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v",
104
+ "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.0.proj",
105
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q",
106
+ "down_blocks.1.attentions.0.proj_in",
107
+ "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.0.proj",
108
+ "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.2",
109
  "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0",
110
+ "down_blocks.2.attentions.1.proj_out",
111
+ "up_blocks.1.attentions.0.proj_in",
112
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0",
113
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k",
114
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v",
115
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k",
116
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k",
117
+ "up_blocks.2.attentions.1.proj_in",
118
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k",
119
+ "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q",
120
  "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k",
121
+ "up_blocks.2.attentions.1.proj_out",
122
  "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q",
123
+ "up_blocks.1.attentions.1.proj_out",
124
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q",
125
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0",
126
+ "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.2",
127
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q",
128
+ "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k",
129
+ "up_blocks.2.attentions.2.proj_out",
130
+ "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v",
131
+ "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q",
132
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0",
133
+ "down_blocks.0.attentions.0.proj_out",
134
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v",
135
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v",
136
+ "up_blocks.2.attentions.0.proj_out",
137
+ "up_blocks.3.attentions.1.proj_out",
138
+ "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v",
139
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v",
140
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k",
141
+ "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v",
142
+ "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.0.proj",
143
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v",
144
+ "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj",
145
+ "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj",
146
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q",
147
+ "mid_block.attentions.0.transformer_blocks.0.ff.net.2",
148
  "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.2",
149
+ "up_blocks.3.attentions.2.proj_in",
150
+ "down_blocks.0.attentions.1.proj_in",
151
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k",
 
 
 
 
 
152
  "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v",
153
+ "up_blocks.3.attentions.0.proj_out",
 
 
 
 
154
  "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q",
155
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q",
 
156
  "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q",
157
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q",
158
+ "mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0",
159
+ "down_blocks.2.attentions.0.proj_in",
 
 
 
 
 
 
 
 
 
 
 
160
  "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v",
161
+ "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.2",
162
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q",
163
+ "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj",
164
+ "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.2",
165
  "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0",
166
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0",
167
+ "up_blocks.3.attentions.2.proj_out",
168
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v",
169
+ "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k",
 
 
 
 
 
 
 
 
 
 
 
 
170
  "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k",
171
+ "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k",
172
+ "mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0",
 
173
  "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v",
174
+ "up_blocks.1.attentions.0.proj_out",
175
+ "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q",
176
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v",
177
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0",
178
+ "down_blocks.1.attentions.1.proj_in",
179
+ "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.2",
180
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0",
181
  "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.2",
182
+ "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.2",
183
+ "up_blocks.2.attentions.2.proj_in",
184
+ "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2",
185
+ "up_blocks.1.attentions.1.proj_in",
 
186
  "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q",
187
+ "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q",
188
+ "down_blocks.2.attentions.1.proj_in",
189
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q",
190
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k",
191
+ "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2",
192
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k",
 
 
 
193
  "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0",
194
+ "mid_block.attentions.0.proj_in",
195
+ "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v",
196
+ "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q",
197
+ "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v",
198
+ "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0",
199
+ "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj",
200
+ "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q",
201
+ "down_blocks.0.attentions.1.proj_out",
 
 
202
  "down_blocks.1.attentions.0.proj_out",
203
+ "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0",
204
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q",
205
+ "mid_block.attentions.0.transformer_blocks.0.attn1.to_k",
206
+ "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q",
 
 
 
 
 
 
 
 
 
 
 
 
 
207
  "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q",
208
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k",
 
 
 
209
  "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj",
210
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211
  "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0",
212
+ "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v",
213
+ "mid_block.attentions.0.transformer_blocks.0.attn1.to_v",
214
+ "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj",
215
+ "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0",
216
+ "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q"
217
  ],
218
  "task_type": null,
219
  "use_dora": false,
sam/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:806debef51f61b2d89892ef7ee739c28a496007d6d93b74e9f31b87dade775be
3
  size 108588496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9ec0bac97c2c2916c0ed85507d83d1d4e78001895c2814539d51590d8855353
3
  size 108588496