|
import gradio as gr |
|
from transformers import pipeline, ViTForImageClassification, ViTImageProcessor |
|
import numpy as np |
|
from PIL import Image |
|
import cv2 as cv |
|
import dlib |
|
import warnings |
|
import logging |
|
from typing import Optional |
|
from pytorch_grad_cam import run_dff_on_image, GradCAM |
|
from pytorch_grad_cam.utils.model_targets import ClassifierOutputTarget |
|
from pytorch_grad_cam.utils.image import show_cam_on_image |
|
import torch |
|
from face_grab import FaceGrabber |
|
from gradcam import GradCam |
|
from torchvision import transforms |
|
|
|
|
|
logging.basicConfig(level=logging.INFO) |
|
|
|
def grab_faces(img: np.ndarray) -> Optional[np.ndarray]: |
|
cascades = [ |
|
"haarcascade_frontalface_default.xml", |
|
"haarcascade_frontalface_alt.xml", |
|
"haarcascade_frontalface_alt2.xml", |
|
"haarcascade_frontalface_alt_tree.xml" |
|
] |
|
|
|
detector = dlib.get_frontal_face_detector() |
|
predictor = dlib.shape_predictor("shape_predictor_68_face_landmarks_GTX.dat") |
|
mmod = dlib.cnn_face_detection_model_v1("mmod_human_face_detector.dat") |
|
|
|
paddingBy = 0.1 |
|
|
|
gray = cv.cvtColor(img, cv.COLOR_BGR2GRAY) |
|
|
|
detected = None |
|
|
|
if detected is None: |
|
faces = detector(gray) |
|
if len(faces) > 0: |
|
detected = faces[0] |
|
detected = (detected.left(), detected.top(), detected.width(), detected.height()) |
|
logging.info("Face detected by dlib") |
|
|
|
if detected is None: |
|
faces = mmod(img) |
|
if len(faces) > 0: |
|
detected = faces[0] |
|
detected = (detected.rect.left(), detected.rect.top(), detected.rect.width(), detected.rect.height()) |
|
logging.info("Face detected by mmod") |
|
|
|
for cascade in cascades: |
|
cascadeClassifier = cv.CascadeClassifier(cv.data.haarcascades + cascade) |
|
faces = cascadeClassifier.detectMultiScale(gray, scaleFactor=1.3, minNeighbors=5) |
|
if len(faces) > 0: |
|
detected = faces[0] |
|
logging.info(f"Face detected by {cascade}") |
|
break |
|
|
|
if detected is not None: |
|
x, y, w, h = detected |
|
padW = int(paddingBy * w) |
|
padH = int(paddingBy * h) |
|
imgH, imgW, _ = img.shape |
|
x = max(0, x - padW) |
|
y = max(0, y - padH) |
|
w = min(imgW - x, w + 2 * padW) |
|
h = min(imgH - y, h + 2 * padH) |
|
x = max(0, x - (w - detected[2]) // 2) |
|
y = max(0, y - (h - detected[3]) // 2) |
|
face = img[y:y+h, x:x+w] |
|
return face |
|
|
|
return None |
|
|
|
model = ViTForImageClassification.from_pretrained("ongkn/attraction-classifier") |
|
processor = ViTImageProcessor.from_pretrained("ongkn/attraction-classifier") |
|
|
|
pipe = pipeline("image-classification", model=model, feature_extractor=processor) |
|
|
|
faceGrabber = FaceGrabber() |
|
gradCam = GradCam() |
|
|
|
targetsForGradCam = [ClassifierOutputTarget(gradCam.category_name_to_index(model, "pos")), |
|
ClassifierOutputTarget(gradCam.category_name_to_index(model, "neg"))] |
|
targetLayerDff = model.vit.layernorm |
|
targetLayerGradCam = model.vit.encoder.layer[-2].output |
|
|
|
def classify_image(input): |
|
face = faceGrabber.grab_faces(np.array(input)) |
|
if face is None: |
|
return "No face detected", 0, input |
|
face = Image.fromarray(face) |
|
imgTensor = transforms.ToTensor()(face) |
|
tensor = transforms.ToTensor()(face) |
|
dffImage = run_dff_on_image(model=model, |
|
target_layer=targetLayerDff, |
|
classifier=model.classifier, |
|
img_pil=face, |
|
img_tensor=tensor, |
|
reshape_transform=gradCam.reshape_transform_vit_huggingface, |
|
n_components=5, |
|
top_k=10, |
|
threshold=0, |
|
) |
|
gradCamImage = gradCam.run_grad_cam_on_image(model=model, |
|
target_layer=targetLayerGradCam, |
|
classifier=model.classifier, |
|
img_pil=face, |
|
img_tensor=tensor, |
|
reshape_transform=gradCam.reshape_transform_vit_huggingface, |
|
n_components=5, |
|
top_k=10, |
|
threshold=0, |
|
) |
|
result = pipe(face) |
|
return result[0]["label"], result[0]["score"], face, dffImage, gradCamImage |
|
|
|
iface = gr.Interface( |
|
fn=classify_image, |
|
inputs="image", |
|
outputs=["text", "number", "image", "image", "image"], |
|
title="Attraction Classifier - subjective", |
|
description=f"Takes in a (224, 224) image and outputs an attraction class: {'pos', 'neg'}, along with a GradCam/DFF explanation. Face detection, cropping, and resizing are done internally. Uploaded images are not stored by us, but may be stored by HF. Refer to their privacy policy for details." |
|
) |
|
iface.launch() |