Spaces:
Runtime error
Runtime error
Adjustments
Browse files
app.py
CHANGED
@@ -151,12 +151,7 @@ def optimize_network(
|
|
151 |
# np.random.seed(seed)
|
152 |
# torch.manual_seed(seed)
|
153 |
# random.seed(seed)
|
154 |
-
|
155 |
-
layer = int(layer)
|
156 |
-
cutn = int(cutn)
|
157 |
-
num_iterations = int(num_iterations)
|
158 |
-
neuron = int(neuron)
|
159 |
-
display_rate = int(display_rate)
|
160 |
|
161 |
make_cutouts = MakeCutouts(clip_models[clip_model].visual.input_resolution, cutn)
|
162 |
loss_fn = CLIPActivationLoss(clip_models[clip_model].visual.transformer.resblocks[layer],
|
@@ -240,7 +235,12 @@ def inference(
|
|
240 |
maximize,
|
241 |
display_rate = 20
|
242 |
):
|
243 |
-
|
|
|
|
|
|
|
|
|
|
|
244 |
opt_type = 'MADGRAD'
|
245 |
seed = 20
|
246 |
save_progress_video = True
|
|
|
151 |
# np.random.seed(seed)
|
152 |
# torch.manual_seed(seed)
|
153 |
# random.seed(seed)
|
154 |
+
save_progress_video = True
|
|
|
|
|
|
|
|
|
|
|
155 |
|
156 |
make_cutouts = MakeCutouts(clip_models[clip_model].visual.input_resolution, cutn)
|
157 |
loss_fn = CLIPActivationLoss(clip_models[clip_model].visual.transformer.resblocks[layer],
|
|
|
235 |
maximize,
|
236 |
display_rate = 20
|
237 |
):
|
238 |
+
layer = int(layer)
|
239 |
+
cutn = int(cutn)
|
240 |
+
num_iterations = int(num_iterations)
|
241 |
+
neuron = int(neuron)
|
242 |
+
display_rate = int(display_rate)
|
243 |
+
|
244 |
opt_type = 'MADGRAD'
|
245 |
seed = 20
|
246 |
save_progress_video = True
|