benjolo commited on
Commit
480569a
·
verified ·
1 Parent(s): 0a2fac0

removing force download from model loading

Browse files
Files changed (1) hide show
  1. backend/main.py +3 -4
backend/main.py CHANGED
@@ -129,12 +129,11 @@ static_files = {
129
  },
130
  }
131
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
132
- processor = AutoProcessor.from_pretrained("facebook/seamless-m4t-v2-large", force_download=True)
133
- #cache_dir="/.cache"
134
 
135
  # PM - hardcoding temporarily as my GPU doesnt have enough vram
136
- # model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large").to("cpu")
137
- model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large", force_download=True).to(device)
138
 
139
 
140
  bytes_data = bytearray()
 
129
  },
130
  }
131
  device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
132
+ # processor = AutoProcessor.from_pretrained("facebook/seamless-m4t-v2-large", force_download=True)
133
+ processor = AutoProcessor.from_pretrained("facebook/seamless-m4t-v2-large")
134
 
135
  # PM - hardcoding temporarily as my GPU doesnt have enough vram
136
+ model = SeamlessM4Tv2Model.from_pretrained("facebook/seamless-m4t-v2-large").to(device)
 
137
 
138
 
139
  bytes_data = bytearray()