David Pomerenke commited on
Commit
ddde856
·
1 Parent(s): c567aee

Use uv in GH action pipeline

Browse files
.github/workflows/nightly-evals.yml CHANGED
@@ -18,8 +18,8 @@ jobs:
18
 
19
  - name: Install dependencies
20
  run: |
21
- python -m pip install --upgrade pip
22
- pip install -r requirements_dev.txt
23
 
24
  - name: Run evaluations
25
  env:
@@ -27,9 +27,8 @@ jobs:
27
  ELEVENLABS_API_KEY: ${{ secrets.ELEVENLABS_API_KEY }}
28
  HUGGINGFACE_ACCESS_TOKEN: ${{ secrets.HUGGINGFACE_ACCESS_TOKEN }}
29
  run: |
30
- cd evals
31
- python download_data.py
32
- python main.py
33
 
34
  - name: Commit changes
35
  run: |
@@ -41,7 +40,7 @@ jobs:
41
 
42
  - name: Install Hugging Face dependencies
43
  run: |
44
- python -m pip install huggingface_hub
45
 
46
  - name: Upload to Hugging Face
47
  env:
 
18
 
19
  - name: Install dependencies
20
  run: |
21
+ curl -LsSf https://astral.sh/uv/install.sh | sh
22
+ uv sync --frozen
23
 
24
  - name: Run evaluations
25
  env:
 
27
  ELEVENLABS_API_KEY: ${{ secrets.ELEVENLABS_API_KEY }}
28
  HUGGINGFACE_ACCESS_TOKEN: ${{ secrets.HUGGINGFACE_ACCESS_TOKEN }}
29
  run: |
30
+ uv run download_data.py
31
+ uv run main.py
 
32
 
33
  - name: Commit changes
34
  run: |
 
40
 
41
  - name: Install Hugging Face dependencies
42
  run: |
43
+ uv add huggingface_hub
44
 
45
  - name: Upload to Hugging Face
46
  env:
requirements_dev.txt DELETED
@@ -1,297 +0,0 @@
1
- # This file was autogenerated by uv via the following command:
2
- # uv pip compile pyproject.toml -o requirements_dev.txt --extra dev
3
- aiohappyeyeballs==2.6.1
4
- # via aiohttp
5
- aiohttp==3.11.18
6
- # via
7
- # datasets
8
- # fsspec
9
- aiolimiter==1.2.1
10
- # via languagebench (pyproject.toml)
11
- aiosignal==1.3.2
12
- # via aiohttp
13
- annotated-types==0.7.0
14
- # via pydantic
15
- anyio==4.9.0
16
- # via
17
- # httpx
18
- # openai
19
- # starlette
20
- async-timeout==5.0.1
21
- # via aiohttp
22
- attrs==25.3.0
23
- # via aiohttp
24
- bert-score==0.3.13
25
- # via languagebench (pyproject.toml)
26
- certifi==2025.4.26
27
- # via
28
- # httpcore
29
- # httpx
30
- # requests
31
- charset-normalizer==3.4.2
32
- # via requests
33
- click==8.1.8
34
- # via
35
- # jiwer
36
- # uvicorn
37
- colorama==0.4.6
38
- # via sacrebleu
39
- contourpy==1.3.2
40
- # via matplotlib
41
- cycler==0.12.1
42
- # via matplotlib
43
- datasets==3.5.1
44
- # via evaluate
45
- dill==0.3.8
46
- # via
47
- # datasets
48
- # evaluate
49
- # multiprocess
50
- distro==1.9.0
51
- # via openai
52
- elevenlabs==1.57.0
53
- # via languagebench (pyproject.toml)
54
- evaluate==0.4.0
55
- # via languagebench (pyproject.toml)
56
- exceptiongroup==1.2.2
57
- # via anyio
58
- fastapi==0.115.12
59
- # via languagebench (pyproject.toml)
60
- filelock==3.18.0
61
- # via
62
- # datasets
63
- # huggingface-hub
64
- # torch
65
- # transformers
66
- fonttools==4.57.0
67
- # via matplotlib
68
- frozenlist==1.6.0
69
- # via
70
- # aiohttp
71
- # aiosignal
72
- fsspec==2025.3.0
73
- # via
74
- # datasets
75
- # evaluate
76
- # huggingface-hub
77
- # torch
78
- h11==0.16.0
79
- # via
80
- # httpcore
81
- # uvicorn
82
- httpcore==1.0.9
83
- # via httpx
84
- httpx==0.28.1
85
- # via
86
- # elevenlabs
87
- # openai
88
- huggingface-hub==0.30.2
89
- # via
90
- # languagebench (pyproject.toml)
91
- # datasets
92
- # evaluate
93
- # tokenizers
94
- # transformers
95
- idna==3.10
96
- # via
97
- # anyio
98
- # httpx
99
- # requests
100
- # yarl
101
- jinja2==3.1.6
102
- # via torch
103
- jiter==0.9.0
104
- # via openai
105
- jiwer==3.1.0
106
- # via languagebench (pyproject.toml)
107
- joblib==1.5.0
108
- # via languagebench (pyproject.toml)
109
- kiwisolver==1.4.8
110
- # via matplotlib
111
- langcodes==3.5.0
112
- # via languagebench (pyproject.toml)
113
- language-data==1.3.0
114
- # via
115
- # languagebench (pyproject.toml)
116
- # langcodes
117
- lxml==5.4.0
118
- # via sacrebleu
119
- marisa-trie==1.2.1
120
- # via language-data
121
- markdown-it-py==3.0.0
122
- # via rich
123
- markupsafe==3.0.2
124
- # via jinja2
125
- matplotlib==3.10.1
126
- # via bert-score
127
- mdurl==0.1.2
128
- # via markdown-it-py
129
- mpmath==1.3.0
130
- # via sympy
131
- multidict==6.4.3
132
- # via
133
- # aiohttp
134
- # yarl
135
- multiprocess==0.70.16
136
- # via
137
- # datasets
138
- # evaluate
139
- networkx==3.4.2
140
- # via torch
141
- numpy==2.2.5
142
- # via
143
- # languagebench (pyproject.toml)
144
- # bert-score
145
- # contourpy
146
- # datasets
147
- # evaluate
148
- # matplotlib
149
- # pandas
150
- # sacrebleu
151
- # transformers
152
- openai==1.77.0
153
- # via languagebench (pyproject.toml)
154
- packaging==25.0
155
- # via
156
- # bert-score
157
- # datasets
158
- # evaluate
159
- # huggingface-hub
160
- # matplotlib
161
- # transformers
162
- pandas==2.2.3
163
- # via
164
- # languagebench (pyproject.toml)
165
- # bert-score
166
- # datasets
167
- # evaluate
168
- pillow==11.2.1
169
- # via matplotlib
170
- portalocker==3.1.1
171
- # via sacrebleu
172
- propcache==0.3.1
173
- # via
174
- # aiohttp
175
- # yarl
176
- protobuf==6.30.2
177
- # via languagebench (pyproject.toml)
178
- pyarrow==20.0.0
179
- # via datasets
180
- pydantic==2.11.4
181
- # via
182
- # elevenlabs
183
- # fastapi
184
- # openai
185
- pydantic-core==2.33.2
186
- # via
187
- # elevenlabs
188
- # pydantic
189
- pygments==2.19.1
190
- # via rich
191
- pyparsing==3.2.3
192
- # via matplotlib
193
- python-dateutil==2.9.0.post0
194
- # via
195
- # matplotlib
196
- # pandas
197
- python-dotenv==1.1.0
198
- # via languagebench (pyproject.toml)
199
- pytz==2025.2
200
- # via pandas
201
- pyyaml==6.0.2
202
- # via
203
- # datasets
204
- # huggingface-hub
205
- # transformers
206
- rapidfuzz==3.13.0
207
- # via jiwer
208
- regex==2024.11.6
209
- # via
210
- # sacrebleu
211
- # tiktoken
212
- # transformers
213
- requests==2.32.3
214
- # via
215
- # bert-score
216
- # datasets
217
- # elevenlabs
218
- # evaluate
219
- # huggingface-hub
220
- # responses
221
- # tiktoken
222
- # transformers
223
- responses==0.18.0
224
- # via evaluate
225
- rich==14.0.0
226
- # via languagebench (pyproject.toml)
227
- sacrebleu==2.5.1
228
- # via languagebench (pyproject.toml)
229
- safetensors==0.5.3
230
- # via transformers
231
- sentencepiece==0.2.0
232
- # via languagebench (pyproject.toml)
233
- setuptools==80.3.1
234
- # via marisa-trie
235
- six==1.17.0
236
- # via python-dateutil
237
- sniffio==1.3.1
238
- # via
239
- # anyio
240
- # openai
241
- starlette==0.46.2
242
- # via fastapi
243
- sympy==1.14.0
244
- # via torch
245
- tabulate==0.9.0
246
- # via sacrebleu
247
- tiktoken==0.9.0
248
- # via languagebench (pyproject.toml)
249
- tokenizers==0.21.1
250
- # via transformers
251
- torch==2.7.0
252
- # via bert-score
253
- tqdm==4.67.1
254
- # via
255
- # languagebench (pyproject.toml)
256
- # bert-score
257
- # datasets
258
- # evaluate
259
- # huggingface-hub
260
- # openai
261
- # transformers
262
- transformers==4.51.3
263
- # via
264
- # languagebench (pyproject.toml)
265
- # bert-score
266
- typing-extensions==4.13.2
267
- # via
268
- # anyio
269
- # elevenlabs
270
- # fastapi
271
- # huggingface-hub
272
- # multidict
273
- # openai
274
- # pydantic
275
- # pydantic-core
276
- # rich
277
- # torch
278
- # typing-inspection
279
- # uvicorn
280
- typing-inspection==0.4.0
281
- # via pydantic
282
- tzdata==2025.2
283
- # via pandas
284
- urllib3==2.4.0
285
- # via
286
- # requests
287
- # responses
288
- uvicorn==0.34.2
289
- # via languagebench (pyproject.toml)
290
- websockets==15.0.1
291
- # via elevenlabs
292
- xxhash==3.5.0
293
- # via
294
- # datasets
295
- # evaluate
296
- yarl==1.20.0
297
- # via aiohttp