hanhainebula commited on
Commit
e575418
·
1 Parent(s): 51b49ff

update etag_timeout, TIMEOUT=300

Browse files
Files changed (3) hide show
  1. app.py +3 -2
  2. src/backend.py +8 -7
  3. src/envs.py +1 -0
app.py CHANGED
@@ -13,7 +13,8 @@ from src.envs import (
13
  UNZIP_TARGET_DIR,
14
  TIME_DURATION,
15
  EVAL_K_VALUES,
16
- SUBMIT_INFOS_TABLE_COLS
 
17
  )
18
  from src.css_html_js import custom_css
19
 
@@ -36,7 +37,7 @@ def load_submit_infos_df():
36
  repo_id=SUBMIT_INFOS_REPO,
37
  repo_type="dataset",
38
  local_dir=SUBMIT_INFOS_DIR,
39
- etag_timeout=30
40
  )
41
  submit_infos_save_path = os.path.join(SUBMIT_INFOS_DIR, SUBMIT_INFOS_FILE_NAME)
42
 
 
13
  UNZIP_TARGET_DIR,
14
  TIME_DURATION,
15
  EVAL_K_VALUES,
16
+ SUBMIT_INFOS_TABLE_COLS,
17
+ TIMEOUT,
18
  )
19
  from src.css_html_js import custom_css
20
 
 
37
  repo_id=SUBMIT_INFOS_REPO,
38
  repo_type="dataset",
39
  local_dir=SUBMIT_INFOS_DIR,
40
+ etag_timeout=TIMEOUT,
41
  )
42
  submit_infos_save_path = os.path.join(SUBMIT_INFOS_DIR, SUBMIT_INFOS_FILE_NAME)
43
 
src/backend.py CHANGED
@@ -15,7 +15,8 @@ from src.envs import (
15
  API,
16
  ZIP_CACHE_DIR,
17
  SEARCH_RESULTS_REPO, RESULTS_REPO, SUBMIT_INFOS_REPO,
18
- make_clickable_model
 
19
  )
20
 
21
  logger = logging.getLogger(__name__)
@@ -171,7 +172,7 @@ def pull_search_results(
171
  repo_id=SUBMIT_INFOS_REPO,
172
  repo_type="dataset",
173
  local_dir=submit_infos_dir,
174
- etag_timeout=30
175
  )
176
 
177
  logger.warning(f"Start from commit: {start_commit_id}")
@@ -181,7 +182,7 @@ def pull_search_results(
181
  repo_type="dataset",
182
  revision=start_commit_id,
183
  local_dir=hf_search_results_repo_dir,
184
- etag_timeout=30,
185
  allow_patterns=['*.json']
186
  )
187
  cur_file_paths = get_file_list(hf_search_results_repo_dir, allowed_suffixes=['.json'])
@@ -190,7 +191,7 @@ def pull_search_results(
190
  repo_id=SEARCH_RESULTS_REPO,
191
  repo_type="dataset",
192
  local_dir=hf_search_results_repo_dir,
193
- etag_timeout=30,
194
  allow_patterns=['*.json']
195
  )
196
  cur_file_paths = get_file_list(hf_search_results_repo_dir, allowed_suffixes=['.json'])
@@ -204,13 +205,13 @@ def pull_search_results(
204
  repo_id=RESULTS_REPO,
205
  repo_type="dataset",
206
  local_dir=hf_eval_results_repo_dir,
207
- etag_timeout=30
208
  )
209
  API.snapshot_download(
210
  repo_id=SEARCH_RESULTS_REPO,
211
  repo_type="dataset",
212
  local_dir=hf_search_results_repo_dir,
213
- etag_timeout=30,
214
  allow_patterns=['*.json']
215
  )
216
  except Exception as e:
@@ -247,7 +248,7 @@ def pull_search_results(
247
  repo_id=SEARCH_RESULTS_REPO,
248
  repo_type="dataset",
249
  local_dir=ZIP_CACHE_DIR,
250
- etag_timeout=30,
251
  allow_patterns=[f'*{zip_file_name}']
252
  )
253
  zip_file_path = get_zip_file_path(zip_file_name)
 
15
  API,
16
  ZIP_CACHE_DIR,
17
  SEARCH_RESULTS_REPO, RESULTS_REPO, SUBMIT_INFOS_REPO,
18
+ make_clickable_model,
19
+ TIMEOUT,
20
  )
21
 
22
  logger = logging.getLogger(__name__)
 
172
  repo_id=SUBMIT_INFOS_REPO,
173
  repo_type="dataset",
174
  local_dir=submit_infos_dir,
175
+ etag_timeout=TIMEOUT,
176
  )
177
 
178
  logger.warning(f"Start from commit: {start_commit_id}")
 
182
  repo_type="dataset",
183
  revision=start_commit_id,
184
  local_dir=hf_search_results_repo_dir,
185
+ etag_timeout=TIMEOUT,
186
  allow_patterns=['*.json']
187
  )
188
  cur_file_paths = get_file_list(hf_search_results_repo_dir, allowed_suffixes=['.json'])
 
191
  repo_id=SEARCH_RESULTS_REPO,
192
  repo_type="dataset",
193
  local_dir=hf_search_results_repo_dir,
194
+ etag_timeout=TIMEOUT,
195
  allow_patterns=['*.json']
196
  )
197
  cur_file_paths = get_file_list(hf_search_results_repo_dir, allowed_suffixes=['.json'])
 
205
  repo_id=RESULTS_REPO,
206
  repo_type="dataset",
207
  local_dir=hf_eval_results_repo_dir,
208
+ etag_timeout=TIMEOUT
209
  )
210
  API.snapshot_download(
211
  repo_id=SEARCH_RESULTS_REPO,
212
  repo_type="dataset",
213
  local_dir=hf_search_results_repo_dir,
214
+ etag_timeout=TIMEOUT,
215
  allow_patterns=['*.json']
216
  )
217
  except Exception as e:
 
248
  repo_id=SEARCH_RESULTS_REPO,
249
  repo_type="dataset",
250
  local_dir=ZIP_CACHE_DIR,
251
+ etag_timeout=TIMEOUT,
252
  allow_patterns=[f'*{zip_file_name}']
253
  )
254
  zip_file_path = get_zip_file_path(zip_file_name)
src/envs.py CHANGED
@@ -34,6 +34,7 @@ HF_EVAL_RESULTS_REPO_DIR = os.path.join(CACHE_PATH, "eval_results")
34
  UNZIP_TARGET_DIR = os.path.join(CACHE_PATH, "unzip_target_dir")
35
 
36
  TIME_DURATION = 600 # seconds
 
37
 
38
  EVAL_K_VALUES = [1, 3, 5, 10, 50, 100, 1000]
39
 
 
34
  UNZIP_TARGET_DIR = os.path.join(CACHE_PATH, "unzip_target_dir")
35
 
36
  TIME_DURATION = 600 # seconds
37
+ TIMEOUT=300 # seconds
38
 
39
  EVAL_K_VALUES = [1, 3, 5, 10, 50, 100, 1000]
40