multimodalart HF Staff commited on
Commit
076e3f4
·
verified ·
1 Parent(s): f293335

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -46,13 +46,13 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
46
  model_checkpoint = "gokaygokay/Flux-Prompt-Enhance"
47
  tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
48
  model = AutoModelForSeq2SeqLM.from_pretrained(model_checkpoint)
49
- enhancer = pipeline('text2text-generation',
50
- model=model,
51
- tokenizer=tokenizer,
52
- repetition_penalty= 1.2,
53
- device=device
54
- )
55
- enhancer.to(device)
56
  max_target_length = 256
57
 
58
  @spaces.GPU
 
46
  model_checkpoint = "gokaygokay/Flux-Prompt-Enhance"
47
  tokenizer = AutoTokenizer.from_pretrained(model_checkpoint)
48
  model = AutoModelForSeq2SeqLM.from_pretrained(model_checkpoint)
49
+ enhancer = pipeline(
50
+ 'text2text-generation',
51
+ model=model,
52
+ tokenizer=tokenizer,
53
+ repetition_penalty= 1.2,
54
+ device=device
55
+ )
56
  max_target_length = 256
57
 
58
  @spaces.GPU