Spaces:
Runtime error
Runtime error
"""PyTorch Hub models | |
Usage: | |
import torch | |
model = torch.hub.load('repo', 'model') | |
""" | |
from pathlib import Path | |
import torch | |
from models.yolo import Model | |
from utils.general import check_requirements, set_logging | |
from utils.google_utils import attempt_download | |
from utils.torch_utils import select_device | |
dependencies = ['torch', 'yaml'] | |
check_requirements(Path(__file__).parent / 'requirements.txt', exclude=('pycocotools', 'thop')) | |
set_logging() | |
def create(name, pretrained, channels, classes, autoshape): | |
"""Creates a specified model | |
Arguments: | |
name (str): name of model, i.e. 'yolov7' | |
pretrained (bool): load pretrained weights into the model | |
channels (int): number of input channels | |
classes (int): number of model classes | |
Returns: | |
pytorch model | |
""" | |
try: | |
cfg = list((Path(__file__).parent / 'cfg').rglob(f'{name}.yaml'))[0] # model.yaml path | |
model = Model(cfg, channels, classes) | |
if pretrained: | |
fname = f'{name}.pt' # checkpoint filename | |
attempt_download(fname) # download if not found locally | |
ckpt = torch.load(fname, map_location=torch.device('cpu')) # load | |
msd = model.state_dict() # model state_dict | |
csd = ckpt['model'].float().state_dict() # checkpoint state_dict as FP32 | |
csd = {k: v for k, v in csd.items() if msd[k].shape == v.shape} # filter | |
model.load_state_dict(csd, strict=False) # load | |
if len(ckpt['model'].names) == classes: | |
model.names = ckpt['model'].names # set class names attribute | |
if autoshape: | |
model = model.autoshape() # for file/URI/PIL/cv2/np inputs and NMS | |
device = select_device('0' if torch.cuda.is_available() else 'cpu') # default to GPU if available | |
return model.to(device) | |
except Exception as e: | |
s = 'Cache maybe be out of date, try force_reload=True.' | |
raise Exception(s) from e | |
def custom(path_or_model='path/to/model.pt', autoshape=True): | |
"""custom mode | |
Arguments (3 options): | |
path_or_model (str): 'path/to/model.pt' | |
path_or_model (dict): torch.load('path/to/model.pt') | |
path_or_model (nn.Module): torch.load('path/to/model.pt')['model'] | |
Returns: | |
pytorch model | |
""" | |
model = torch.load(path_or_model) if isinstance(path_or_model, str) else path_or_model # load checkpoint | |
if isinstance(model, dict): | |
model = model['ema' if model.get('ema') else 'model'] # load model | |
hub_model = Model(model.yaml).to(next(model.parameters()).device) # create | |
hub_model.load_state_dict(model.float().state_dict()) # load state_dict | |
hub_model.names = model.names # class names | |
if autoshape: | |
hub_model = hub_model.autoshape() # for file/URI/PIL/cv2/np inputs and NMS | |
device = select_device('0' if torch.cuda.is_available() else 'cpu') # default to GPU if available | |
return hub_model.to(device) | |
def yolov7(pretrained=True, channels=3, classes=80, autoshape=True): | |
return create('yolov7', pretrained, channels, classes, autoshape) | |
if __name__ == '__main__': | |
model = custom(path_or_model='yolov7.pt') # custom example | |
# model = create(name='yolov7', pretrained=True, channels=3, classes=80, autoshape=True) # pretrained example | |
# Verify inference | |
import numpy as np | |
from PIL import Image | |
imgs = [np.zeros((640, 480, 3))] | |
results = model(imgs) # batched inference | |
results.print() | |
results.save() | |