hysts HF Staff commited on
Commit
f8ca87f
·
1 Parent(s): 98d1bde
.pre-commit-config.yaml CHANGED
@@ -1,6 +1,6 @@
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
- rev: v4.6.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
@@ -18,13 +18,15 @@ repos:
18
  hooks:
19
  - id: docformatter
20
  args: ["--in-place"]
21
- - repo: https://github.com/pycqa/isort
22
- rev: 5.13.2
23
  hooks:
24
- - id: isort
25
- args: ["--profile", "black"]
 
 
26
  - repo: https://github.com/pre-commit/mirrors-mypy
27
- rev: v1.9.0
28
  hooks:
29
  - id: mypy
30
  args: ["--ignore-missing-imports"]
@@ -35,14 +37,8 @@ repos:
35
  "types-PyYAML",
36
  "types-pytz",
37
  ]
38
- - repo: https://github.com/psf/black
39
- rev: 24.4.0
40
- hooks:
41
- - id: black
42
- language_version: python3.10
43
- args: ["--line-length", "119"]
44
  - repo: https://github.com/kynan/nbstripout
45
- rev: 0.7.1
46
  hooks:
47
  - id: nbstripout
48
  args:
@@ -51,7 +47,7 @@ repos:
51
  "metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
52
  ]
53
  - repo: https://github.com/nbQA-dev/nbQA
54
- rev: 1.8.5
55
  hooks:
56
  - id: nbqa-black
57
  - id: nbqa-pyupgrade
 
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
+ rev: v5.0.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
 
18
  hooks:
19
  - id: docformatter
20
  args: ["--in-place"]
21
+ - repo: https://github.com/astral-sh/ruff-pre-commit
22
+ rev: v0.8.4
23
  hooks:
24
+ - id: ruff
25
+ args: ["--fix"]
26
+ - id: ruff-format
27
+ args: ["--line-length", "119"]
28
  - repo: https://github.com/pre-commit/mirrors-mypy
29
+ rev: v1.14.0
30
  hooks:
31
  - id: mypy
32
  args: ["--ignore-missing-imports"]
 
37
  "types-PyYAML",
38
  "types-pytz",
39
  ]
 
 
 
 
 
 
40
  - repo: https://github.com/kynan/nbstripout
41
+ rev: 0.8.1
42
  hooks:
43
  - id: nbstripout
44
  args:
 
47
  "metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
48
  ]
49
  - repo: https://github.com/nbQA-dev/nbQA
50
+ rev: 1.9.1
51
  hooks:
52
  - id: nbqa-black
53
  - id: nbqa-pyupgrade
.python-version ADDED
@@ -0,0 +1 @@
 
 
1
+ 3.10
.vscode/extensions.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "recommendations": [
3
+ "ms-python.python",
4
+ "charliermarsh.ruff",
5
+ "streetsidesoftware.code-spell-checker",
6
+ "tamasfe.even-better-toml"
7
+ ]
8
+ }
.vscode/settings.json CHANGED
@@ -2,29 +2,20 @@
2
  "editor.formatOnSave": true,
3
  "files.insertFinalNewline": false,
4
  "[python]": {
5
- "editor.defaultFormatter": "ms-python.black-formatter",
6
  "editor.formatOnType": true,
7
  "editor.codeActionsOnSave": {
 
8
  "source.organizeImports": "explicit"
9
  }
10
  },
11
  "[jupyter]": {
12
  "files.insertFinalNewline": false
13
  },
14
- "black-formatter.args": [
15
- "--line-length=119"
16
- ],
17
- "isort.args": ["--profile", "black"],
18
- "flake8.args": [
19
- "--max-line-length=119"
20
- ],
21
- "ruff.lint.args": [
22
- "--line-length=119"
23
- ],
24
  "notebook.output.scrolling": true,
25
  "notebook.formatOnCellExecution": true,
26
  "notebook.formatOnSave.enabled": true,
27
- "notebook.codeActionsOnSave": {
28
- "source.organizeImports": "explicit"
29
- }
30
  }
 
2
  "editor.formatOnSave": true,
3
  "files.insertFinalNewline": false,
4
  "[python]": {
5
+ "editor.defaultFormatter": "charliermarsh.ruff",
6
  "editor.formatOnType": true,
7
  "editor.codeActionsOnSave": {
8
+ "source.fixAll.ruff": "explicit",
9
  "source.organizeImports": "explicit"
10
  }
11
  },
12
  "[jupyter]": {
13
  "files.insertFinalNewline": false
14
  },
 
 
 
 
 
 
 
 
 
 
15
  "notebook.output.scrolling": true,
16
  "notebook.formatOnCellExecution": true,
17
  "notebook.formatOnSave.enabled": true,
18
+ "notebook.codeActionsOnSave": {
19
+ "source.organizeImports": "explicit"
20
+ }
21
  }
README.md CHANGED
@@ -4,7 +4,7 @@ emoji: 📊
4
  colorFrom: green
5
  colorTo: purple
6
  sdk: gradio
7
- sdk_version: 4.36.1
8
  app_file: app.py
9
  pinned: true
10
  header: mini
 
4
  colorFrom: green
5
  colorTo: purple
6
  sdk: gradio
7
+ sdk_version: 5.9.1
8
  app_file: app.py
9
  pinned: true
10
  header: mini
app.py CHANGED
@@ -48,8 +48,7 @@ DEFAULT_COLUMNS = [
48
  def update_num_papers(df: pd.DataFrame) -> str:
49
  if "claimed" in df.columns:
50
  return f"{len(df)} / {len(paper_list.df_raw)} ({len(df[df['claimed'].str.contains('✅')])} claimed)"
51
- else:
52
- return f"{len(df)} / {len(paper_list.df_raw)}"
53
 
54
 
55
  def update_df(
@@ -73,7 +72,7 @@ def update_df(
73
  )
74
 
75
 
76
- with gr.Blocks(css="style.css") as demo:
77
  gr.Markdown(DESCRIPTION)
78
  with gr.Accordion(label="Tutorial", open=True):
79
  gr.Markdown(TUTORIAL)
@@ -121,7 +120,7 @@ with gr.Blocks(css="style.css") as demo:
121
  type="pandas",
122
  row_count=(0, "dynamic"),
123
  interactive=False,
124
- height=1000,
125
  elem_id="table",
126
  wrap=True,
127
  )
 
48
  def update_num_papers(df: pd.DataFrame) -> str:
49
  if "claimed" in df.columns:
50
  return f"{len(df)} / {len(paper_list.df_raw)} ({len(df[df['claimed'].str.contains('✅')])} claimed)"
51
+ return f"{len(df)} / {len(paper_list.df_raw)}"
 
52
 
53
 
54
  def update_df(
 
72
  )
73
 
74
 
75
+ with gr.Blocks(css_paths="style.css") as demo:
76
  gr.Markdown(DESCRIPTION)
77
  with gr.Accordion(label="Tutorial", open=True):
78
  gr.Markdown(TUTORIAL)
 
120
  type="pandas",
121
  row_count=(0, "dynamic"),
122
  interactive=False,
123
+ max_height=1000,
124
  elem_id="table",
125
  wrap=True,
126
  )
papers.py CHANGED
@@ -19,7 +19,7 @@ ABSTRACT_RETRIEVER.search("LLM")
19
 
20
 
21
  class PaperList:
22
- COLUMN_INFO = [
23
  ["Title", "str"],
24
  ["Authors", "str"],
25
  ["Type", "str"],
@@ -33,16 +33,16 @@ class PaperList:
33
  ["Models", "markdown"],
34
  ["Datasets", "markdown"],
35
  ["claimed", "markdown"],
36
- ]
37
 
38
- def __init__(self):
39
  self.df_raw = self.get_df()
40
  self.df_prettified = self.prettify(self.df_raw)
41
 
42
  @staticmethod
43
  def get_df() -> pd.DataFrame:
44
- df = pd.merge(
45
- left=datasets.load_dataset("ICLR2024/ICLR2024-papers", split="train").to_pandas(),
46
  right=datasets.load_dataset("ICLR2024/ICLR2024-paper-stats", split="train").to_pandas(),
47
  on="id",
48
  how="left",
@@ -102,14 +102,14 @@ class PaperList:
102
  return pd.DataFrame(rows, columns=PaperList.get_column_names())
103
 
104
  @staticmethod
105
- def get_column_names():
106
  return list(map(operator.itemgetter(0), PaperList.COLUMN_INFO))
107
 
108
  def get_column_datatypes(self, column_names: list[str]) -> list[str]:
109
  mapping = dict(self.COLUMN_INFO)
110
  return [mapping[name] for name in column_names]
111
 
112
- def search(
113
  self,
114
  title_search_query: str,
115
  abstract_search_query: str,
 
19
 
20
 
21
  class PaperList:
22
+ COLUMN_INFO = (
23
  ["Title", "str"],
24
  ["Authors", "str"],
25
  ["Type", "str"],
 
33
  ["Models", "markdown"],
34
  ["Datasets", "markdown"],
35
  ["claimed", "markdown"],
36
+ )
37
 
38
+ def __init__(self) -> None:
39
  self.df_raw = self.get_df()
40
  self.df_prettified = self.prettify(self.df_raw)
41
 
42
  @staticmethod
43
  def get_df() -> pd.DataFrame:
44
+ df = datasets.load_dataset("ICLR2024/ICLR2024-papers", split="train").to_pandas()
45
+ df = df.merge(
46
  right=datasets.load_dataset("ICLR2024/ICLR2024-paper-stats", split="train").to_pandas(),
47
  on="id",
48
  how="left",
 
102
  return pd.DataFrame(rows, columns=PaperList.get_column_names())
103
 
104
  @staticmethod
105
+ def get_column_names() -> list[str]:
106
  return list(map(operator.itemgetter(0), PaperList.COLUMN_INFO))
107
 
108
  def get_column_datatypes(self, column_names: list[str]) -> list[str]:
109
  mapping = dict(self.COLUMN_INFO)
110
  return [mapping[name] for name in column_names]
111
 
112
+ def search( # noqa: C901
113
  self,
114
  title_search_query: str,
115
  abstract_search_query: str,
pyproject.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [project]
2
+ name = "iclr2024-papers"
3
+ version = "0.1.0"
4
+ description = ""
5
+ readme = "README.md"
6
+ requires-python = ">=3.10"
7
+ dependencies = [
8
+ "datasets>=3.2.0",
9
+ "gradio>=5.9.1",
10
+ "hf-transfer>=0.1.8",
11
+ "ragatouille>=0.0.8.post4",
12
+ "setuptools>=75.6.0",
13
+ ]
14
+
15
+ [tool.ruff]
16
+ line-length = 119
17
+
18
+ [tool.ruff.lint]
19
+ select = ["ALL"]
20
+ ignore = [
21
+ "COM812", # missing-trailing-comma
22
+ "D203", # one-blank-line-before-class
23
+ "D213", # multi-line-summary-second-line
24
+ "E501", # line-too-long
25
+ "SIM117", # multiple-with-statements
26
+ ]
27
+ extend-ignore = [
28
+ "D100", # undocumented-public-module
29
+ "D101", # undocumented-public-class
30
+ "D102", # undocumented-public-method
31
+ "D103", # undocumented-public-function
32
+ "D104", # undocumented-public-package
33
+ "D105", # undocumented-magic-method
34
+ "D107", # undocumented-public-init
35
+ "EM101", # raw-string-in-exception
36
+ "FBT001", # boolean-type-hint-positional-argument
37
+ "FBT002", # boolean-default-value-positional-argument
38
+ "PD901", # pandas-df-variable-name
39
+ "PGH003", # blanket-type-ignore
40
+ "PLR0913", # too-many-arguments
41
+ "PLR0915", # too-many-statements
42
+ "TRY003", # raise-vanilla-args
43
+ ]
44
+ unfixable = [
45
+ "F401", # unused-import
46
+ ]
47
+
48
+ [tool.ruff.format]
49
+ docstring-code-format = true
requirements.txt CHANGED
@@ -1,6 +1,534 @@
1
- datasets==2.19.1
2
- #gradio==4.36.1
3
- huggingface_hub==0.23.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  pandas==2.2.2
5
- ragatouille==0.0.8.post2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  tqdm==4.66.4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This file was autogenerated by uv via the following command:
2
+ # uv pip compile pyproject.toml -o requirements.txt
3
+ aiofiles==23.2.1
4
+ # via gradio
5
+ aiohappyeyeballs==2.4.4
6
+ # via aiohttp
7
+ aiohttp==3.11.11
8
+ # via
9
+ # datasets
10
+ # fsspec
11
+ # langchain
12
+ # llama-index-core
13
+ aiosignal==1.3.2
14
+ # via aiohttp
15
+ annotated-types==0.7.0
16
+ # via pydantic
17
+ anyio==4.7.0
18
+ # via
19
+ # gradio
20
+ # httpx
21
+ # openai
22
+ # starlette
23
+ async-timeout==4.0.3
24
+ # via
25
+ # aiohttp
26
+ # langchain
27
+ attrs==24.3.0
28
+ # via aiohttp
29
+ beautifulsoup4==4.12.3
30
+ # via llama-index-readers-file
31
+ bitarray==3.0.0
32
+ # via colbert-ai
33
+ blinker==1.9.0
34
+ # via flask
35
+ catalogue==2.0.10
36
+ # via srsly
37
+ certifi==2024.12.14
38
+ # via
39
+ # httpcore
40
+ # httpx
41
+ # llama-cloud
42
+ # requests
43
+ charset-normalizer==3.4.1
44
+ # via requests
45
+ click==8.1.8
46
+ # via
47
+ # flask
48
+ # llama-parse
49
+ # nltk
50
+ # typer
51
+ # uvicorn
52
+ colbert-ai==0.2.19
53
+ # via ragatouille
54
+ dataclasses-json==0.6.7
55
+ # via llama-index-core
56
+ datasets==3.2.0
57
+ # via
58
+ # iclr2024-papers (pyproject.toml)
59
+ # colbert-ai
60
+ deprecated==1.2.15
61
+ # via llama-index-core
62
+ dill==0.3.8
63
+ # via
64
+ # datasets
65
+ # multiprocess
66
+ dirtyjson==1.0.8
67
+ # via llama-index-core
68
+ distro==1.9.0
69
+ # via openai
70
+ exceptiongroup==1.2.2
71
+ # via anyio
72
+ faiss-cpu==1.9.0.post1
73
+ # via ragatouille
74
+ fast-pytorch-kmeans==0.2.0.1
75
+ # via ragatouille
76
+ fastapi==0.115.6
77
+ # via gradio
78
+ ffmpy==0.5.0
79
+ # via gradio
80
+ filelock==3.16.1
81
+ # via
82
+ # datasets
83
+ # huggingface-hub
84
+ # torch
85
+ # transformers
86
+ # triton
87
+ filetype==1.2.0
88
+ # via llama-index-core
89
+ flask==3.1.0
90
+ # via colbert-ai
91
+ frozenlist==1.5.0
92
+ # via
93
+ # aiohttp
94
+ # aiosignal
95
+ fsspec==2024.9.0
96
+ # via
97
+ # datasets
98
+ # gradio-client
99
+ # huggingface-hub
100
+ # llama-index-core
101
+ # torch
102
+ git-python==1.0.3
103
+ # via colbert-ai
104
+ gitdb==4.0.11
105
+ # via gitpython
106
+ gitpython==3.1.43
107
+ # via git-python
108
+ gradio==5.9.1
109
+ # via iclr2024-papers (pyproject.toml)
110
+ gradio-client==1.5.2
111
+ # via gradio
112
+ greenlet==3.1.1
113
+ # via sqlalchemy
114
+ h11==0.14.0
115
+ # via
116
+ # httpcore
117
+ # uvicorn
118
+ hf-transfer==0.1.8
119
+ # via iclr2024-papers (pyproject.toml)
120
+ httpcore==1.0.7
121
+ # via httpx
122
+ httpx==0.28.1
123
+ # via
124
+ # gradio
125
+ # gradio-client
126
+ # langsmith
127
+ # llama-cloud
128
+ # llama-index-core
129
+ # openai
130
+ # safehttpx
131
+ huggingface-hub==0.27.0
132
+ # via
133
+ # datasets
134
+ # gradio
135
+ # gradio-client
136
+ # sentence-transformers
137
+ # tokenizers
138
+ # transformers
139
+ idna==3.10
140
+ # via
141
+ # anyio
142
+ # httpx
143
+ # requests
144
+ # yarl
145
+ itsdangerous==2.2.0
146
+ # via flask
147
+ jinja2==3.1.5
148
+ # via
149
+ # flask
150
+ # gradio
151
+ # torch
152
+ jiter==0.8.2
153
+ # via openai
154
+ joblib==1.4.2
155
+ # via
156
+ # nltk
157
+ # scikit-learn
158
+ jsonpatch==1.33
159
+ # via langchain-core
160
+ jsonpointer==3.0.0
161
+ # via jsonpatch
162
+ langchain==0.3.13
163
+ # via ragatouille
164
+ langchain-core==0.3.28
165
+ # via
166
+ # langchain
167
+ # langchain-text-splitters
168
+ # ragatouille
169
+ langchain-text-splitters==0.3.4
170
+ # via langchain
171
+ langsmith==0.2.7
172
+ # via
173
+ # langchain
174
+ # langchain-core
175
+ llama-cloud==0.1.7
176
+ # via llama-index-indices-managed-llama-cloud
177
+ llama-index==0.12.9
178
+ # via ragatouille
179
+ llama-index-agent-openai==0.4.1
180
+ # via
181
+ # llama-index
182
+ # llama-index-program-openai
183
+ llama-index-cli==0.4.0
184
+ # via llama-index
185
+ llama-index-core==0.12.9
186
+ # via
187
+ # llama-index
188
+ # llama-index-agent-openai
189
+ # llama-index-cli
190
+ # llama-index-embeddings-openai
191
+ # llama-index-indices-managed-llama-cloud
192
+ # llama-index-llms-openai
193
+ # llama-index-multi-modal-llms-openai
194
+ # llama-index-program-openai
195
+ # llama-index-question-gen-openai
196
+ # llama-index-readers-file
197
+ # llama-index-readers-llama-parse
198
+ # llama-parse
199
+ llama-index-embeddings-openai==0.3.1
200
+ # via
201
+ # llama-index
202
+ # llama-index-cli
203
+ llama-index-indices-managed-llama-cloud==0.6.3
204
+ # via llama-index
205
+ llama-index-llms-openai==0.3.12
206
+ # via
207
+ # llama-index
208
+ # llama-index-agent-openai
209
+ # llama-index-cli
210
+ # llama-index-multi-modal-llms-openai
211
+ # llama-index-program-openai
212
+ # llama-index-question-gen-openai
213
+ llama-index-multi-modal-llms-openai==0.4.1
214
+ # via llama-index
215
+ llama-index-program-openai==0.3.1
216
+ # via
217
+ # llama-index
218
+ # llama-index-question-gen-openai
219
+ llama-index-question-gen-openai==0.3.0
220
+ # via llama-index
221
+ llama-index-readers-file==0.4.1
222
+ # via llama-index
223
+ llama-index-readers-llama-parse==0.4.0
224
+ # via llama-index
225
+ llama-parse==0.5.19
226
+ # via llama-index-readers-llama-parse
227
+ markdown-it-py==3.0.0
228
+ # via rich
229
+ markupsafe==2.1.5
230
+ # via
231
+ # gradio
232
+ # jinja2
233
+ # werkzeug
234
+ marshmallow==3.23.2
235
+ # via dataclasses-json
236
+ mdurl==0.1.2
237
+ # via markdown-it-py
238
+ mpmath==1.3.0
239
+ # via sympy
240
+ multidict==6.1.0
241
+ # via
242
+ # aiohttp
243
+ # yarl
244
+ multiprocess==0.70.16
245
+ # via datasets
246
+ mypy-extensions==1.0.0
247
+ # via typing-inspect
248
+ nest-asyncio==1.6.0
249
+ # via llama-index-core
250
+ networkx==3.4.2
251
+ # via
252
+ # llama-index-core
253
+ # torch
254
+ ninja==1.11.1.3
255
+ # via colbert-ai
256
+ nltk==3.9.1
257
+ # via
258
+ # llama-index
259
+ # llama-index-core
260
+ numpy==1.26.4
261
+ # via
262
+ # datasets
263
+ # faiss-cpu
264
+ # fast-pytorch-kmeans
265
+ # gradio
266
+ # langchain
267
+ # llama-index-core
268
+ # onnx
269
+ # pandas
270
+ # scikit-learn
271
+ # scipy
272
+ # sentence-transformers
273
+ # transformers
274
+ # voyager
275
+ nvidia-cublas-cu12==12.4.5.8
276
+ # via
277
+ # nvidia-cudnn-cu12
278
+ # nvidia-cusolver-cu12
279
+ # torch
280
+ nvidia-cuda-cupti-cu12==12.4.127
281
+ # via torch
282
+ nvidia-cuda-nvrtc-cu12==12.4.127
283
+ # via torch
284
+ nvidia-cuda-runtime-cu12==12.4.127
285
+ # via torch
286
+ nvidia-cudnn-cu12==9.1.0.70
287
+ # via torch
288
+ nvidia-cufft-cu12==11.2.1.3
289
+ # via torch
290
+ nvidia-curand-cu12==10.3.5.147
291
+ # via torch
292
+ nvidia-cusolver-cu12==11.6.1.9
293
+ # via torch
294
+ nvidia-cusparse-cu12==12.3.1.170
295
+ # via
296
+ # nvidia-cusolver-cu12
297
+ # torch
298
+ nvidia-ml-py==12.560.30
299
+ # via pynvml
300
+ nvidia-nccl-cu12==2.21.5
301
+ # via torch
302
+ nvidia-nvjitlink-cu12==12.4.127
303
+ # via
304
+ # nvidia-cusolver-cu12
305
+ # nvidia-cusparse-cu12
306
+ # torch
307
+ nvidia-nvtx-cu12==12.4.127
308
+ # via torch
309
+ onnx==1.17.0
310
+ # via ragatouille
311
+ openai==1.58.1
312
+ # via
313
+ # llama-index-agent-openai
314
+ # llama-index-embeddings-openai
315
+ # llama-index-llms-openai
316
+ orjson==3.10.13
317
+ # via
318
+ # gradio
319
+ # langsmith
320
+ packaging==24.2
321
+ # via
322
+ # datasets
323
+ # faiss-cpu
324
+ # gradio
325
+ # gradio-client
326
+ # huggingface-hub
327
+ # langchain-core
328
+ # marshmallow
329
+ # transformers
330
  pandas==2.2.2
331
+ # via
332
+ # datasets
333
+ # gradio
334
+ # llama-index-readers-file
335
+ pillow==11.0.0
336
+ # via
337
+ # gradio
338
+ # llama-index-core
339
+ # sentence-transformers
340
+ propcache==0.2.1
341
+ # via
342
+ # aiohttp
343
+ # yarl
344
+ protobuf==5.29.2
345
+ # via onnx
346
+ pyarrow==18.1.0
347
+ # via datasets
348
+ pydantic==2.10.4
349
+ # via
350
+ # fastapi
351
+ # gradio
352
+ # langchain
353
+ # langchain-core
354
+ # langsmith
355
+ # llama-cloud
356
+ # llama-index-core
357
+ # llama-parse
358
+ # openai
359
+ pydantic-core==2.27.2
360
+ # via pydantic
361
+ pydub==0.25.1
362
+ # via gradio
363
+ pygments==2.18.0
364
+ # via rich
365
+ pynvml==12.0.0
366
+ # via fast-pytorch-kmeans
367
+ pypdf==5.1.0
368
+ # via llama-index-readers-file
369
+ python-dateutil==2.9.0.post0
370
+ # via pandas
371
+ python-dotenv==1.0.1
372
+ # via colbert-ai
373
+ python-multipart==0.0.20
374
+ # via gradio
375
+ pytz==2024.2
376
+ # via pandas
377
+ pyyaml==6.0.2
378
+ # via
379
+ # datasets
380
+ # gradio
381
+ # huggingface-hub
382
+ # langchain
383
+ # langchain-core
384
+ # llama-index-core
385
+ # transformers
386
+ ragatouille==0.0.8.post4
387
+ # via iclr2024-papers (pyproject.toml)
388
+ regex==2024.11.6
389
+ # via
390
+ # nltk
391
+ # tiktoken
392
+ # transformers
393
+ requests==2.32.3
394
+ # via
395
+ # datasets
396
+ # huggingface-hub
397
+ # langchain
398
+ # langsmith
399
+ # llama-index-core
400
+ # requests-toolbelt
401
+ # tiktoken
402
+ # transformers
403
+ requests-toolbelt==1.0.0
404
+ # via langsmith
405
+ rich==13.9.4
406
+ # via typer
407
+ ruff==0.8.4
408
+ # via gradio
409
+ safehttpx==0.1.6
410
+ # via gradio
411
+ safetensors==0.4.5
412
+ # via transformers
413
+ scikit-learn==1.6.0
414
+ # via sentence-transformers
415
+ scipy==1.14.1
416
+ # via
417
+ # colbert-ai
418
+ # scikit-learn
419
+ # sentence-transformers
420
+ semantic-version==2.10.0
421
+ # via gradio
422
+ sentence-transformers==2.7.0
423
+ # via ragatouille
424
+ setuptools==75.6.0
425
+ # via iclr2024-papers (pyproject.toml)
426
+ shellingham==1.5.4
427
+ # via typer
428
+ six==1.17.0
429
+ # via python-dateutil
430
+ smmap==5.0.1
431
+ # via gitdb
432
+ sniffio==1.3.1
433
+ # via
434
+ # anyio
435
+ # openai
436
+ soupsieve==2.6
437
+ # via beautifulsoup4
438
+ sqlalchemy==2.0.36
439
+ # via
440
+ # langchain
441
+ # llama-index-core
442
+ srsly==2.4.8
443
+ # via ragatouille
444
+ starlette==0.41.3
445
+ # via
446
+ # fastapi
447
+ # gradio
448
+ striprtf==0.0.26
449
+ # via llama-index-readers-file
450
+ sympy==1.13.1
451
+ # via torch
452
+ tenacity==9.0.0
453
+ # via
454
+ # langchain
455
+ # langchain-core
456
+ # llama-index-core
457
+ threadpoolctl==3.5.0
458
+ # via scikit-learn
459
+ tiktoken==0.8.0
460
+ # via llama-index-core
461
+ tokenizers==0.21.0
462
+ # via transformers
463
+ tomlkit==0.13.2
464
+ # via gradio
465
+ torch==2.5.1
466
+ # via
467
+ # fast-pytorch-kmeans
468
+ # ragatouille
469
+ # sentence-transformers
470
  tqdm==4.66.4
471
+ # via
472
+ # colbert-ai
473
+ # datasets
474
+ # huggingface-hub
475
+ # llama-index-core
476
+ # nltk
477
+ # openai
478
+ # sentence-transformers
479
+ # transformers
480
+ transformers==4.47.1
481
+ # via
482
+ # colbert-ai
483
+ # ragatouille
484
+ # sentence-transformers
485
+ triton==3.1.0
486
+ # via torch
487
+ typer==0.15.1
488
+ # via gradio
489
+ typing-extensions==4.12.2
490
+ # via
491
+ # anyio
492
+ # fastapi
493
+ # gradio
494
+ # gradio-client
495
+ # huggingface-hub
496
+ # langchain-core
497
+ # llama-index-core
498
+ # multidict
499
+ # openai
500
+ # pydantic
501
+ # pydantic-core
502
+ # pypdf
503
+ # rich
504
+ # sqlalchemy
505
+ # torch
506
+ # typer
507
+ # typing-inspect
508
+ # uvicorn
509
+ typing-inspect==0.9.0
510
+ # via
511
+ # dataclasses-json
512
+ # llama-index-core
513
+ tzdata==2024.2
514
+ # via pandas
515
+ ujson==5.10.0
516
+ # via colbert-ai
517
+ urllib3==2.3.0
518
+ # via requests
519
+ uvicorn==0.34.0
520
+ # via gradio
521
+ voyager==2.1.0
522
+ # via ragatouille
523
+ websockets==14.1
524
+ # via gradio-client
525
+ werkzeug==3.1.3
526
+ # via flask
527
+ wrapt==1.17.0
528
+ # via
529
+ # deprecated
530
+ # llama-index-core
531
+ xxhash==3.5.0
532
+ # via datasets
533
+ yarl==1.18.3
534
+ # via aiohttp
uv.lock ADDED
The diff for this file is too large to render. See raw diff