Spaces:
Runtime error
Runtime error
Commit
·
2821541
1
Parent(s):
14fbc60
app.py
CHANGED
|
@@ -87,7 +87,7 @@ def write_video(file_name, images, fps=15):
|
|
| 87 |
|
| 88 |
|
| 89 |
image_encoder = CLIPVisionModelWithProjection.from_pretrained("h94/IP-Adapter", subfolder="sdxl_models/image_encoder", torch_dtype=dtype,
|
| 90 |
-
device_map='
|
| 91 |
#vae = AutoencoderTiny.from_pretrained("madebyollin/taesd", torch_dtype=dtype)
|
| 92 |
|
| 93 |
# vae = ConsistencyDecoderVAE.from_pretrained("openai/consistency-decoder", torch_dtype=dtype)
|
|
@@ -209,7 +209,7 @@ def pluck_img(user_id, user_emb):
|
|
| 209 |
not_rated_rows = prevs_df[[i[1]['user:rating'].get(user_id, 'gone') == 'gone' for i in prevs_df.iterrows()]]
|
| 210 |
while len(not_rated_rows) == 0:
|
| 211 |
not_rated_rows = prevs_df[[i[1]['user:rating'].get(user_id, 'gone') == 'gone' for i in prevs_df.iterrows()]]
|
| 212 |
-
time.sleep(.
|
| 213 |
print(f'user {user_id} has rated all rows')
|
| 214 |
# TODO optimize this lol
|
| 215 |
best_sim = -100000
|
|
|
|
| 87 |
|
| 88 |
|
| 89 |
image_encoder = CLIPVisionModelWithProjection.from_pretrained("h94/IP-Adapter", subfolder="sdxl_models/image_encoder", torch_dtype=dtype,
|
| 90 |
+
device_map='cuda')
|
| 91 |
#vae = AutoencoderTiny.from_pretrained("madebyollin/taesd", torch_dtype=dtype)
|
| 92 |
|
| 93 |
# vae = ConsistencyDecoderVAE.from_pretrained("openai/consistency-decoder", torch_dtype=dtype)
|
|
|
|
| 209 |
not_rated_rows = prevs_df[[i[1]['user:rating'].get(user_id, 'gone') == 'gone' for i in prevs_df.iterrows()]]
|
| 210 |
while len(not_rated_rows) == 0:
|
| 211 |
not_rated_rows = prevs_df[[i[1]['user:rating'].get(user_id, 'gone') == 'gone' for i in prevs_df.iterrows()]]
|
| 212 |
+
time.sleep(.1)
|
| 213 |
print(f'user {user_id} has rated all rows')
|
| 214 |
# TODO optimize this lol
|
| 215 |
best_sim = -100000
|