watsonpro commited on
Commit
5f8b425
·
verified ·
1 Parent(s): bb0ecfa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -38
app.py CHANGED
@@ -1,43 +1,15 @@
1
  import gradio as gr
2
- from transformers import AutoTokenizer, AutoModelForCausalLM
3
- import torch
4
 
5
- model = AutoModelForCausalLM.from_pretrained(
6
- "tiiuae/falcon-7b-instruct",
7
- torch_dtype=torch.bfloat16,
8
- trust_remote_code=True,
9
- device_map="auto",
10
- low_cpu_mem_usage=True,
11
- )
12
 
13
- tokenizer = AutoTokenizer.from_pretrained("tiiuae/falcon-7b-instruct")
 
 
14
 
15
- def generate_text(input_text):
16
- input_ids = tokenizer.encode(input_text, return_tensors="pt")
17
- attention_mask = torch.ones(input_ids.shape)
18
-
19
- output = model.generate(
20
- input_ids,
21
- attention_mask=attention_mask,
22
- max_length=200,
23
- do_sample=True,
24
- top_k=10,
25
- num_return_sequences=1,
26
- eos_token_id=tokenizer.eos_token_id,
27
- )
28
-
29
- output_text = tokenizer.decode(output[0], skip_special_tokens=True)
30
- print(output_text)
31
-
32
- # Remove Prompt Echo from Generated Text
33
- cleaned_output_text = output_text.replace(input_text, "")
34
- return cleaned_output_text
35
-
36
- text_generation_interface = gr.Interface(
37
- fn=generate_text,
38
- inputs=[
39
- gr.inputs.Textbox(label="Input Text"),
40
- ],
41
- outputs=gr.inputs.Textbox(label="Generated Text"),
42
- title="Falcon-7B Instruct",
43
  ).launch()
 
1
  import gradio as gr
2
+ from transformers import pipeline
 
3
 
4
+ pipeline = pipeline(task="image-classification", model="julien-c/hotdog-not-hotdog")
 
 
 
 
 
 
5
 
6
+ def predict(image):
7
+ predictions = pipeline(image)
8
+ return {p["label"]: p["score"] for p in predictions}
9
 
10
+ gr.Interface(
11
+ predict,
12
+ inputs=gr.Image(label="Upload hot dog candidate", type="filepath"),
13
+ outputs=gr.Label(num_top_classes=2),
14
+ title="Hot Dog? Or Not?",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ).launch()