Spaces:
Sleeping
Sleeping
Upload folder using huggingface_hub
Browse files- .gitattributes +2 -6
- README.md +4 -3
- app.py +48 -0
- dragQueen1.jpg +0 -0
- dragQueen2.jpg +0 -0
- female.jpg +0 -0
- femaleAngry1.jpg +0 -0
- femaleAngry2.jpg +0 -0
- femaleAsian.jpg +0 -0
- femaleDefault.jpg +0 -0
- femaleMuscle1.jpg +0 -0
- femaleMuscle2.jpg +0 -0
- gradio_cached_examples/14/log.csv +15 -0
- gradio_cached_examples/14/output/tmp1pi9dt6x.json +1 -0
- gradio_cached_examples/14/output/tmp49qkmuyr.json +1 -0
- gradio_cached_examples/14/output/tmp4joshavx.json +1 -0
- gradio_cached_examples/14/output/tmp5cete_c0.json +1 -0
- gradio_cached_examples/14/output/tmpde335yc3.json +1 -0
- gradio_cached_examples/14/output/tmpdq9wuff4.json +1 -0
- gradio_cached_examples/14/output/tmpemjlfl_y.json +1 -0
- gradio_cached_examples/14/output/tmpi8pyk2w1.json +1 -0
- gradio_cached_examples/14/output/tmpi_2oem5f.json +1 -0
- gradio_cached_examples/14/output/tmpmk4dpz_q.json +1 -0
- gradio_cached_examples/14/output/tmpp15raev4.json +1 -0
- gradio_cached_examples/14/output/tmpr_f180a9.json +1 -0
- gradio_cached_examples/14/output/tmpsn63qolo.json +1 -0
- gradio_cached_examples/14/output/tmpxp18339a.json +1 -0
- male.jpg +0 -0
- maleAsian.jpg +0 -0
- maleCrying2.jpg +0 -0
- maleCrying2No.jpg +0 -0
- maleDefault.jpg +0 -0
- maleEurope.jpg +0 -0
- model.pkl +3 -0
- requirements.txt +3 -0
.gitattributes
CHANGED
@@ -2,13 +2,11 @@
|
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
-
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
-
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
*.npy filter=lfs diff=lfs merge=lfs -text
|
@@ -16,20 +14,18 @@
|
|
16 |
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
*.pkl filter=lfs diff=lfs merge=lfs -text
|
|
|
22 |
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
-
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
-
*.
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
|
|
5 |
*.ftz filter=lfs diff=lfs merge=lfs -text
|
6 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
7 |
*.h5 filter=lfs diff=lfs merge=lfs -text
|
8 |
*.joblib filter=lfs diff=lfs merge=lfs -text
|
9 |
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
|
|
10 |
*.model filter=lfs diff=lfs merge=lfs -text
|
11 |
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
12 |
*.npy filter=lfs diff=lfs merge=lfs -text
|
|
|
14 |
*.onnx filter=lfs diff=lfs merge=lfs -text
|
15 |
*.ot filter=lfs diff=lfs merge=lfs -text
|
16 |
*.parquet filter=lfs diff=lfs merge=lfs -text
|
|
|
17 |
*.pickle filter=lfs diff=lfs merge=lfs -text
|
18 |
*.pkl filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
*.pt filter=lfs diff=lfs merge=lfs -text
|
21 |
*.pth filter=lfs diff=lfs merge=lfs -text
|
22 |
*.rar filter=lfs diff=lfs merge=lfs -text
|
|
|
23 |
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
24 |
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
|
|
25 |
*.tflite filter=lfs diff=lfs merge=lfs -text
|
26 |
*.tgz filter=lfs diff=lfs merge=lfs -text
|
27 |
*.wasm filter=lfs diff=lfs merge=lfs -text
|
28 |
*.xz filter=lfs diff=lfs merge=lfs -text
|
29 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
31 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,12 +1,13 @@
|
|
1 |
---
|
2 |
title: GenderClassifier
|
3 |
-
emoji:
|
4 |
-
colorFrom:
|
5 |
-
colorTo:
|
6 |
sdk: gradio
|
7 |
sdk_version: 3.47.1
|
8 |
app_file: app.py
|
9 |
pinned: false
|
|
|
10 |
---
|
11 |
|
12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
1 |
---
|
2 |
title: GenderClassifier
|
3 |
+
emoji: 🦀
|
4 |
+
colorFrom: green
|
5 |
+
colorTo: indigo
|
6 |
sdk: gradio
|
7 |
sdk_version: 3.47.1
|
8 |
app_file: app.py
|
9 |
pinned: false
|
10 |
+
license: apache-2.0
|
11 |
---
|
12 |
|
13 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
app.py
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
## import gradio as gr
|
2 |
+
#
|
3 |
+
## def greet(name):
|
4 |
+
## return "Hello " + name + "!!"
|
5 |
+
#
|
6 |
+
## iface = gr.Interface(fn=greet, inputs="text", outputs="text")
|
7 |
+
## iface.launch()
|
8 |
+
|
9 |
+
import gradio as gr
|
10 |
+
from fastai.vision.all import *
|
11 |
+
import skimage
|
12 |
+
import pathlib
|
13 |
+
temp = pathlib.PosixPath
|
14 |
+
pathlib.PosixPath = pathlib.WindowsPath
|
15 |
+
pathlib.PosixPath = temp
|
16 |
+
|
17 |
+
learn = load_learner('model.pkl')
|
18 |
+
labels = learn.dls.vocab
|
19 |
+
def predict(img):
|
20 |
+
img = PILImage.create(img)
|
21 |
+
pred,pred_idx,probs = learn.predict(img)
|
22 |
+
return {labels[i]: float(probs[i]) for i in range(len(labels))}
|
23 |
+
|
24 |
+
title = "Female/Male Classifier"
|
25 |
+
description = "A Female/Male classifier trained on the duckduckgo search result with fastai. Created as a demo for Gradio and HuggingFace Spaces."
|
26 |
+
## article="<p style='text-align: center'><a href='https://tmabraham.github.io/blog/gradio_hf_spaces_tutorial' target='_blank'>Blog post</a></p>"
|
27 |
+
examples = ['femaleDefault.jpg', 'maleDefault.jpg',
|
28 |
+
'dragQueen1.jpg', 'dragQueen2.jpg',
|
29 |
+
'femaleAngry1.jpg', 'femaleAngry2.jpg',
|
30 |
+
'femaleMuscle1.jpg', 'femaleMuscle2.jpg',
|
31 |
+
'maleAsian.jpg', 'maleEurope.jpg',
|
32 |
+
'femaleAsian.jpg', 'femaleDefault.jpg',
|
33 |
+
'maleCrying2.jpg', 'maleCrying2No.jpg']
|
34 |
+
#interpretation='default'
|
35 |
+
enable_queue=True
|
36 |
+
#
|
37 |
+
## gr.Interface(fn=predict,inputs=gr.inputs.Image(shape=(512, 512)),outputs=gr.outputs.Label(),title=title,description=description,article=article,examples=examples,interpretation=interpretation,enable_queue=enable_queue).launch()
|
38 |
+
#
|
39 |
+
gr.Interface(
|
40 |
+
fn=predict,
|
41 |
+
inputs=gr.inputs.Image(shape=(512, 512)),
|
42 |
+
outputs=gr.outputs.Label(),
|
43 |
+
title=title,
|
44 |
+
description=description,
|
45 |
+
examples=examples,
|
46 |
+
cache_examples=True,
|
47 |
+
examples_per_page=2,
|
48 |
+
enable_queue=enable_queue).launch(share=True)
|
dragQueen1.jpg
ADDED
![]() |
dragQueen2.jpg
ADDED
![]() |
female.jpg
ADDED
![]() |
femaleAngry1.jpg
ADDED
![]() |
femaleAngry2.jpg
ADDED
![]() |
femaleAsian.jpg
ADDED
![]() |
femaleDefault.jpg
ADDED
![]() |
femaleMuscle1.jpg
ADDED
![]() |
femaleMuscle2.jpg
ADDED
![]() |
gradio_cached_examples/14/log.csv
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
output,flag,username,timestamp
|
2 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpemjlfl_y.json,,,2023-10-11 15:31:19.087730
|
3 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpr_f180a9.json,,,2023-10-11 15:31:19.130380
|
4 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmp1pi9dt6x.json,,,2023-10-11 15:31:19.167941
|
5 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpde335yc3.json,,,2023-10-11 15:31:19.208476
|
6 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmp5cete_c0.json,,,2023-10-11 15:31:19.250513
|
7 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpi8pyk2w1.json,,,2023-10-11 15:31:19.288551
|
8 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmp4joshavx.json,,,2023-10-11 15:31:19.331076
|
9 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpmk4dpz_q.json,,,2023-10-11 15:31:19.375151
|
10 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpi_2oem5f.json,,,2023-10-11 15:31:19.420194
|
11 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpp15raev4.json,,,2023-10-11 15:31:19.479757
|
12 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpxp18339a.json,,,2023-10-11 15:31:19.524283
|
13 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmp49qkmuyr.json,,,2023-10-11 15:31:19.567493
|
14 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpdq9wuff4.json,,,2023-10-11 15:31:19.605011
|
15 |
+
C:\Users\Be\Desktop\temp\Gender_Classifier\gradio_cached_examples\14\output\tmpsn63qolo.json,,,2023-10-11 15:31:19.665570
|
gradio_cached_examples/14/output/tmp1pi9dt6x.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.9019985198974609}, {"label": "male", "confidence": 0.09800148755311966}]}
|
gradio_cached_examples/14/output/tmp49qkmuyr.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.9623780846595764}, {"label": "male", "confidence": 0.037621963769197464}]}
|
gradio_cached_examples/14/output/tmp4joshavx.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.7219564914703369}, {"label": "male", "confidence": 0.2780435383319855}]}
|
gradio_cached_examples/14/output/tmp5cete_c0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "male", "confidences": [{"label": "male", "confidence": 0.661678671836853}, {"label": "female", "confidence": 0.338321328163147}]}
|
gradio_cached_examples/14/output/tmpde335yc3.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.558527946472168}, {"label": "male", "confidence": 0.4414720833301544}]}
|
gradio_cached_examples/14/output/tmpdq9wuff4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "male", "confidences": [{"label": "male", "confidence": 0.6026121377944946}, {"label": "female", "confidence": 0.397387832403183}]}
|
gradio_cached_examples/14/output/tmpemjlfl_y.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.9623780846595764}, {"label": "male", "confidence": 0.037621963769197464}]}
|
gradio_cached_examples/14/output/tmpi8pyk2w1.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.9751496911048889}, {"label": "male", "confidence": 0.024850372225046158}]}
|
gradio_cached_examples/14/output/tmpi_2oem5f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "male", "confidences": [{"label": "male", "confidence": 0.8869091868400574}, {"label": "female", "confidence": 0.11309077590703964}]}
|
gradio_cached_examples/14/output/tmpmk4dpz_q.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "male", "confidences": [{"label": "male", "confidence": 0.5785160064697266}, {"label": "female", "confidence": 0.4214840531349182}]}
|
gradio_cached_examples/14/output/tmpp15raev4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "male", "confidences": [{"label": "male", "confidence": 0.9677399396896362}, {"label": "female", "confidence": 0.032260045409202576}]}
|
gradio_cached_examples/14/output/tmpr_f180a9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "male", "confidences": [{"label": "male", "confidence": 0.9945085644721985}, {"label": "female", "confidence": 0.005491500720381737}]}
|
gradio_cached_examples/14/output/tmpsn63qolo.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.5097442865371704}, {"label": "male", "confidence": 0.4902556538581848}]}
|
gradio_cached_examples/14/output/tmpxp18339a.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"label": "female", "confidences": [{"label": "female", "confidence": 0.9533503651618958}, {"label": "male", "confidence": 0.046649616211652756}]}
|
male.jpg
ADDED
![]() |
maleAsian.jpg
ADDED
![]() |
maleCrying2.jpg
ADDED
![]() |
maleCrying2No.jpg
ADDED
![]() |
maleDefault.jpg
ADDED
![]() |
maleEurope.jpg
ADDED
![]() |
model.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:157754422adf7e02d69a3717f46ccd195cb28e21b46f70539d05b10397882bdd
|
3 |
+
size 46957103
|
requirements.txt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
gradio
|
2 |
+
fastai
|
3 |
+
scikit-image
|