Spaces:
Running
Running
File size: 835 Bytes
4b51ee4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 |
from transformers import CLIPProcessor, CLIPModel
from PIL import Image
model = CLIPModel.from_pretrained("openai/clip-vit-base-patch32")
processor = CLIPProcessor.from_pretrained("openai/clip-vit-base-patch32")
def is_safe_image(
model,
processor,
image,
):
# Load image
# image = Image.open(
# r"F:\om\2025\fastsdcpumcp\fastsdcpu\results\829a2123-92c8-4957-ad2f-06365a19665a-1.png"
# )
categories = ["safe", "nsfw"]
inputs = processor(
text=categories,
images=image,
return_tensors="pt",
padding=True,
)
outputs = model(**inputs)
logits_per_image = outputs.logits_per_image
probs = logits_per_image.softmax(dim=1)
safe_prob = dict(zip(categories, probs[0].tolist()))
print(safe_prob)
return safe_prob["safe"] > safe_prob["nsfw"]
|