Spaces:
Sleeping
Sleeping
File size: 2,161 Bytes
516d9b1 bef99af 516d9b1 bef99af 516d9b1 08ba7c3 ad1fcc6 bef99af 516d9b1 190a01a 924cefe bfe786b b4aea34 8d7dd3a 924cefe 8d7dd3a ad1fcc6 8d7dd3a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 |
#!/usr/bin/env python
from __future__ import annotations
import pathlib
import cv2
import gradio as gr
import huggingface_hub
import insightface
import numpy as np
import onnxruntime as ort
TITLE = "insightface Person Detection"
DESCRIPTION = "https://github.com/deepinsight/insightface/tree/master/examples/person_detection"
def load_model():
path = huggingface_hub.hf_hub_download("public-data/insightface", "models/scrfd_person_2.5g.onnx")
options = ort.SessionOptions()
options.intra_op_num_threads = 8
options.inter_op_num_threads = 8
session = ort.InferenceSession(
path, sess_options=options, providers=["CPUExecutionProvider"]
)
model = insightface.model_zoo.retinaface.RetinaFace(model_file=path, session=session)
return model
def detect_person(
img: np.ndarray, detector: insightface.model_zoo.retinaface.RetinaFace
) -> np.ndarray:
bboxes, _ = detector.detect(img)
bboxes = np.round(bboxes[:, :4]).astype(int)
return bboxes
def extract_persons(image: np.ndarray, bboxes: np.ndarray) -> list[np.ndarray]:
person_images = []
for bbox in bboxes:
x1, y1, x2, y2 = bbox
person_image = image[y1:y2, x1:x2] # Crop the detected person
person_images.append(person_image)
return person_images
detector = load_model()
detector.prepare(-1, nms_thresh=0.5, input_size=(640, 640))
def detect(image: np.ndarray) -> list[np.ndarray]:
image = image[:, :, ::-1] # RGB -> BGR
bboxes = detect_person(image, detector)
person_images = extract_persons(image, bboxes) # Extract each person as a separate image
return [person_img[:, :, ::-1] for person_img in person_images] # BGR -> RGB
examples = sorted(pathlib.Path("images").glob("*.jpg"))
# Forcing PNG format
gr.processing_utils.ENCODING_FORMAT = "PNG"
demo = gr.Interface(
fn=detect,
inputs=gr.Image(label="Input", type="numpy"),
outputs=gr.Gallery(label="Detected Persons"), # Display multiple images in a gallery
examples=examples,
examples_per_page=30,
title=TITLE,
description=DESCRIPTION,
)
if __name__ == "__main__":
demo.queue(max_size=10).launch()
|