.pre-commit-config.yaml CHANGED
@@ -1,6 +1,6 @@
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
- rev: v5.0.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
@@ -13,21 +13,48 @@ repos:
13
  args: ["--fix=lf"]
14
  - id: requirements-txt-fixer
15
  - id: trailing-whitespace
16
- - repo: https://github.com/astral-sh/ruff-pre-commit
17
- rev: v0.8.6
18
  hooks:
19
- - id: ruff
20
- args: ["--fix"]
21
- - id: ruff-format
 
 
 
 
22
  - repo: https://github.com/pre-commit/mirrors-mypy
23
- rev: v1.14.1
24
  hooks:
25
  - id: mypy
26
  args: ["--ignore-missing-imports"]
27
  additional_dependencies:
28
  [
29
  "types-python-slugify",
30
- "types-pytz",
31
- "types-PyYAML",
32
  "types-requests",
 
 
33
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
+ rev: v4.6.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
 
13
  args: ["--fix=lf"]
14
  - id: requirements-txt-fixer
15
  - id: trailing-whitespace
16
+ - repo: https://github.com/myint/docformatter
17
+ rev: v1.7.5
18
  hooks:
19
+ - id: docformatter
20
+ args: ["--in-place"]
21
+ - repo: https://github.com/pycqa/isort
22
+ rev: 5.13.2
23
+ hooks:
24
+ - id: isort
25
+ args: ["--profile", "black"]
26
  - repo: https://github.com/pre-commit/mirrors-mypy
27
+ rev: v1.9.0
28
  hooks:
29
  - id: mypy
30
  args: ["--ignore-missing-imports"]
31
  additional_dependencies:
32
  [
33
  "types-python-slugify",
 
 
34
  "types-requests",
35
+ "types-PyYAML",
36
+ "types-pytz",
37
  ]
38
+ - repo: https://github.com/psf/black
39
+ rev: 24.4.0
40
+ hooks:
41
+ - id: black
42
+ language_version: python3.10
43
+ args: ["--line-length", "119"]
44
+ - repo: https://github.com/kynan/nbstripout
45
+ rev: 0.7.1
46
+ hooks:
47
+ - id: nbstripout
48
+ args:
49
+ [
50
+ "--extra-keys",
51
+ "metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
52
+ ]
53
+ - repo: https://github.com/nbQA-dev/nbQA
54
+ rev: 1.8.5
55
+ hooks:
56
+ - id: nbqa-black
57
+ - id: nbqa-pyupgrade
58
+ args: ["--py37-plus"]
59
+ - id: nbqa-isort
60
+ args: ["--float-to-top"]
.python-version DELETED
@@ -1 +0,0 @@
1
- 3.10
 
 
.vscode/extensions.json DELETED
@@ -1,8 +0,0 @@
1
- {
2
- "recommendations": [
3
- "ms-python.python",
4
- "charliermarsh.ruff",
5
- "streetsidesoftware.code-spell-checker",
6
- "tamasfe.even-better-toml"
7
- ]
8
- }
 
 
 
 
 
 
 
 
 
.vscode/settings.json CHANGED
@@ -2,16 +2,29 @@
2
  "editor.formatOnSave": true,
3
  "files.insertFinalNewline": false,
4
  "[python]": {
5
- "editor.defaultFormatter": "charliermarsh.ruff",
6
  "editor.formatOnType": true,
7
  "editor.codeActionsOnSave": {
8
- "source.fixAll.ruff": "explicit",
9
  "source.organizeImports": "explicit"
10
  }
11
  },
12
  "[jupyter]": {
13
  "files.insertFinalNewline": false
14
  },
 
 
 
 
 
 
 
 
 
 
15
  "notebook.output.scrolling": true,
16
- "notebook.formatOnSave.enabled": true
 
 
 
 
17
  }
 
2
  "editor.formatOnSave": true,
3
  "files.insertFinalNewline": false,
4
  "[python]": {
5
+ "editor.defaultFormatter": "ms-python.black-formatter",
6
  "editor.formatOnType": true,
7
  "editor.codeActionsOnSave": {
 
8
  "source.organizeImports": "explicit"
9
  }
10
  },
11
  "[jupyter]": {
12
  "files.insertFinalNewline": false
13
  },
14
+ "black-formatter.args": [
15
+ "--line-length=119"
16
+ ],
17
+ "isort.args": ["--profile", "black"],
18
+ "flake8.args": [
19
+ "--max-line-length=119"
20
+ ],
21
+ "ruff.lint.args": [
22
+ "--line-length=119"
23
+ ],
24
  "notebook.output.scrolling": true,
25
+ "notebook.formatOnCellExecution": true,
26
+ "notebook.formatOnSave.enabled": true,
27
+ "notebook.codeActionsOnSave": {
28
+ "source.organizeImports": "explicit"
29
+ }
30
  }
README.md CHANGED
@@ -4,9 +4,9 @@ emoji: 📊
4
  colorFrom: green
5
  colorTo: purple
6
  sdk: gradio
7
- sdk_version: 5.12.0
8
  app_file: app.py
9
- pinned: true
10
  header: mini
11
  datasets:
12
  - ICLR2024/ICLR2024-papers
 
4
  colorFrom: green
5
  colorTo: purple
6
  sdk: gradio
7
+ sdk_version: 4.27.0
8
  app_file: app.py
9
+ pinned: false
10
  header: mini
11
  datasets:
12
  - ICLR2024/ICLR2024-papers
app.py CHANGED
@@ -8,6 +8,10 @@ from papers import PaperList
8
  DESCRIPTION = "# ICLR 2024 Papers"
9
 
10
  TUTORIAL = """\
 
 
 
 
11
  #### Tutorial for claiming the ICLR 2024 papers
12
 
13
  1. Find your paper in the table below.
@@ -22,10 +26,6 @@ The admin team will validate your request soon. Once confirmed, the Paper page w
22
  If you need further assistance, see the guide [here](https://huggingface.co/docs/hub/paper-pages#claiming-authorship-to-a-paper).
23
 
24
  If your paper is not yet indexed on Hugging Face, you can index it by following this [guide](https://huggingface.co/docs/hub/paper-pages#can-i-have-a-paper-page-even-if-i-have-no-modeldatasetspace) and open a PR with [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to add your Paper page to this Space.
25
-
26
- #### Tutorial for creating a PR
27
-
28
- To add data to the table below, please use [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to create a PR.
29
  """
30
 
31
 
@@ -35,10 +35,7 @@ DEFAULT_COLUMNS = [
35
  "Title",
36
  "Type",
37
  "Paper page",
38
- "👍",
39
- "💬",
40
  "OpenReview",
41
- "Project page",
42
  "GitHub",
43
  "Spaces",
44
  "Models",
@@ -48,7 +45,8 @@ DEFAULT_COLUMNS = [
48
  def update_num_papers(df: pd.DataFrame) -> str:
49
  if "claimed" in df.columns:
50
  return f"{len(df)} / {len(paper_list.df_raw)} ({len(df[df['claimed'].str.contains('✅')])} claimed)"
51
- return f"{len(df)} / {len(paper_list.df_raw)}"
 
52
 
53
 
54
  def update_df(
@@ -72,7 +70,7 @@ def update_df(
72
  )
73
 
74
 
75
- with gr.Blocks(css_paths="style.css") as demo:
76
  gr.Markdown(DESCRIPTION)
77
  with gr.Accordion(label="Tutorial", open=True):
78
  gr.Markdown(TUTORIAL)
@@ -120,7 +118,7 @@ with gr.Blocks(css_paths="style.css") as demo:
120
  type="pandas",
121
  row_count=(0, "dynamic"),
122
  interactive=False,
123
- max_height=1000,
124
  elem_id="table",
125
  wrap=True,
126
  )
 
8
  DESCRIPTION = "# ICLR 2024 Papers"
9
 
10
  TUTORIAL = """\
11
+ #### Tutorial for creating a PR
12
+
13
+ To add data to the table below, please use [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to create a PR.
14
+
15
  #### Tutorial for claiming the ICLR 2024 papers
16
 
17
  1. Find your paper in the table below.
 
26
  If you need further assistance, see the guide [here](https://huggingface.co/docs/hub/paper-pages#claiming-authorship-to-a-paper).
27
 
28
  If your paper is not yet indexed on Hugging Face, you can index it by following this [guide](https://huggingface.co/docs/hub/paper-pages#can-i-have-a-paper-page-even-if-i-have-no-modeldatasetspace) and open a PR with [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to add your Paper page to this Space.
 
 
 
 
29
  """
30
 
31
 
 
35
  "Title",
36
  "Type",
37
  "Paper page",
 
 
38
  "OpenReview",
 
39
  "GitHub",
40
  "Spaces",
41
  "Models",
 
45
  def update_num_papers(df: pd.DataFrame) -> str:
46
  if "claimed" in df.columns:
47
  return f"{len(df)} / {len(paper_list.df_raw)} ({len(df[df['claimed'].str.contains('✅')])} claimed)"
48
+ else:
49
+ return f"{len(df)} / {len(paper_list.df_raw)}"
50
 
51
 
52
  def update_df(
 
70
  )
71
 
72
 
73
+ with gr.Blocks(css="style.css") as demo:
74
  gr.Markdown(DESCRIPTION)
75
  with gr.Accordion(label="Tutorial", open=True):
76
  gr.Markdown(TUTORIAL)
 
118
  type="pandas",
119
  row_count=(0, "dynamic"),
120
  interactive=False,
121
+ height=1000,
122
  elem_id="table",
123
  wrap=True,
124
  )
papers.py CHANGED
@@ -19,36 +19,32 @@ ABSTRACT_RETRIEVER.search("LLM")
19
 
20
 
21
  class PaperList:
22
- COLUMN_INFO = (
23
  ["Title", "str"],
24
  ["Authors", "str"],
25
  ["Type", "str"],
26
  ["Paper page", "markdown"],
27
- ["👍", "number"],
28
- ["💬", "number"],
29
  ["OpenReview", "markdown"],
30
- ["Project page", "markdown"],
31
  ["GitHub", "markdown"],
32
  ["Spaces", "markdown"],
33
  ["Models", "markdown"],
34
  ["Datasets", "markdown"],
35
  ["claimed", "markdown"],
36
- )
37
 
38
- def __init__(self) -> None:
39
  self.df_raw = self.get_df()
40
  self.df_prettified = self.prettify(self.df_raw)
41
 
42
  @staticmethod
43
  def get_df() -> pd.DataFrame:
44
- df = datasets.load_dataset("ICLR2024/ICLR2024-papers", split="train").to_pandas()
45
- df = df.merge(
46
- right=datasets.load_dataset("ICLR2024/ICLR2024-paper-stats", split="train").to_pandas(),
47
  on="id",
48
  how="left",
49
  )
50
- keys = ["n_authors", "n_linked_authors", "upvotes", "num_comments"]
51
- df[keys] = df[keys].fillna(-1).astype(int)
52
  df["paper_page"] = df["arxiv_id"].apply(
53
  lambda arxiv_id: f"https://huggingface.co/papers/{arxiv_id}" if arxiv_id else ""
54
  )
@@ -66,19 +62,12 @@ class PaperList:
66
  n_linked_authors = "" if row.n_linked_authors == -1 else row.n_linked_authors
67
  n_authors = "" if row.n_authors == -1 else row.n_authors
68
  claimed_paper = "" if n_linked_authors == "" else f"{n_linked_authors}/{n_authors} {author_linked}"
69
- upvotes = "" if row.upvotes == -1 else row.upvotes
70
- num_comments = "" if row.num_comments == -1 else row.num_comments
71
 
72
  new_row = {
73
  "Title": row["title"],
74
  "Authors": ", ".join(row["authors"]),
75
  "Type": row["type"],
76
  "Paper page": PaperList.create_link(row["arxiv_id"], row["paper_page"]),
77
- "Project page": (
78
- PaperList.create_link("Project page", row["project_page"]) if row["project_page"] else ""
79
- ),
80
- "👍": upvotes,
81
- "💬": num_comments,
82
  "OpenReview": PaperList.create_link("OpenReview", row["OpenReview"]),
83
  "GitHub": "\n".join([PaperList.create_link("GitHub", url) for url in row["GitHub"]]),
84
  "Spaces": "\n".join(
@@ -102,14 +91,14 @@ class PaperList:
102
  return pd.DataFrame(rows, columns=PaperList.get_column_names())
103
 
104
  @staticmethod
105
- def get_column_names() -> list[str]:
106
  return list(map(operator.itemgetter(0), PaperList.COLUMN_INFO))
107
 
108
  def get_column_datatypes(self, column_names: list[str]) -> list[str]:
109
  mapping = dict(self.COLUMN_INFO)
110
  return [mapping[name] for name in column_names]
111
 
112
- def search( # noqa: C901
113
  self,
114
  title_search_query: str,
115
  abstract_search_query: str,
@@ -130,7 +119,7 @@ class PaperList:
130
  df = df[df["type"] == presentation_type]
131
 
132
  if "Paper page" in filter_names:
133
- df = df[df["paper_page"] != ""]
134
  if "GitHub" in filter_names:
135
  df = df[df["GitHub"].apply(len) > 0]
136
  if "Space" in filter_names:
 
19
 
20
 
21
  class PaperList:
22
+ COLUMN_INFO = [
23
  ["Title", "str"],
24
  ["Authors", "str"],
25
  ["Type", "str"],
26
  ["Paper page", "markdown"],
 
 
27
  ["OpenReview", "markdown"],
 
28
  ["GitHub", "markdown"],
29
  ["Spaces", "markdown"],
30
  ["Models", "markdown"],
31
  ["Datasets", "markdown"],
32
  ["claimed", "markdown"],
33
+ ]
34
 
35
+ def __init__(self):
36
  self.df_raw = self.get_df()
37
  self.df_prettified = self.prettify(self.df_raw)
38
 
39
  @staticmethod
40
  def get_df() -> pd.DataFrame:
41
+ df = pd.merge(
42
+ left=datasets.load_dataset("ICLR2024/ICLR2024-papers", split="train").to_pandas(),
43
+ right=datasets.load_dataset("ICLR2024/ICLR2024-num-claimed-papers", split="train").to_pandas(),
44
  on="id",
45
  how="left",
46
  )
47
+ df[["n_authors", "n_linked_authors"]] = df[["n_authors", "n_linked_authors"]].fillna(-1).astype(int)
 
48
  df["paper_page"] = df["arxiv_id"].apply(
49
  lambda arxiv_id: f"https://huggingface.co/papers/{arxiv_id}" if arxiv_id else ""
50
  )
 
62
  n_linked_authors = "" if row.n_linked_authors == -1 else row.n_linked_authors
63
  n_authors = "" if row.n_authors == -1 else row.n_authors
64
  claimed_paper = "" if n_linked_authors == "" else f"{n_linked_authors}/{n_authors} {author_linked}"
 
 
65
 
66
  new_row = {
67
  "Title": row["title"],
68
  "Authors": ", ".join(row["authors"]),
69
  "Type": row["type"],
70
  "Paper page": PaperList.create_link(row["arxiv_id"], row["paper_page"]),
 
 
 
 
 
71
  "OpenReview": PaperList.create_link("OpenReview", row["OpenReview"]),
72
  "GitHub": "\n".join([PaperList.create_link("GitHub", url) for url in row["GitHub"]]),
73
  "Spaces": "\n".join(
 
91
  return pd.DataFrame(rows, columns=PaperList.get_column_names())
92
 
93
  @staticmethod
94
+ def get_column_names():
95
  return list(map(operator.itemgetter(0), PaperList.COLUMN_INFO))
96
 
97
  def get_column_datatypes(self, column_names: list[str]) -> list[str]:
98
  mapping = dict(self.COLUMN_INFO)
99
  return [mapping[name] for name in column_names]
100
 
101
+ def search(
102
  self,
103
  title_search_query: str,
104
  abstract_search_query: str,
 
119
  df = df[df["type"] == presentation_type]
120
 
121
  if "Paper page" in filter_names:
122
+ df = df[df["paper_page"].notnull()]
123
  if "GitHub" in filter_names:
124
  df = df[df["GitHub"].apply(len) > 0]
125
  if "Space" in filter_names:
pyproject.toml DELETED
@@ -1,50 +0,0 @@
1
- [project]
2
- name = "iclr2024-papers"
3
- version = "0.1.0"
4
- description = ""
5
- readme = "README.md"
6
- requires-python = ">=3.10"
7
- dependencies = [
8
- "datasets>=3.2.0",
9
- "gradio>=5.12.0",
10
- "hf-transfer>=0.1.9",
11
- "ragatouille>=0.0.8.post4",
12
- "setuptools>=75.6.0",
13
- "spaces>=0.32.0",
14
- ]
15
-
16
- [tool.ruff]
17
- line-length = 119
18
-
19
- [tool.ruff.lint]
20
- select = ["ALL"]
21
- ignore = [
22
- "COM812", # missing-trailing-comma
23
- "D203", # one-blank-line-before-class
24
- "D213", # multi-line-summary-second-line
25
- "E501", # line-too-long
26
- "SIM117", # multiple-with-statements
27
- ]
28
- extend-ignore = [
29
- "D100", # undocumented-public-module
30
- "D101", # undocumented-public-class
31
- "D102", # undocumented-public-method
32
- "D103", # undocumented-public-function
33
- "D104", # undocumented-public-package
34
- "D105", # undocumented-magic-method
35
- "D107", # undocumented-public-init
36
- "EM101", # raw-string-in-exception
37
- "FBT001", # boolean-type-hint-positional-argument
38
- "FBT002", # boolean-default-value-positional-argument
39
- "PD901", # pandas-df-variable-name
40
- "PGH003", # blanket-type-ignore
41
- "PLR0913", # too-many-arguments
42
- "PLR0915", # too-many-statements
43
- "TRY003", # raise-vanilla-args
44
- ]
45
- unfixable = [
46
- "F401", # unused-import
47
- ]
48
-
49
- [tool.ruff.format]
50
- docstring-code-format = true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
requirements.txt CHANGED
@@ -1,545 +1,6 @@
1
- # This file was autogenerated by uv via the following command:
2
- # uv pip compile pyproject.toml -o requirements.txt
3
- aiofiles==23.2.1
4
- # via gradio
5
- aiohappyeyeballs==2.4.4
6
- # via aiohttp
7
- aiohttp==3.11.11
8
- # via
9
- # datasets
10
- # fsspec
11
- # langchain
12
- # llama-index-core
13
- aiosignal==1.3.2
14
- # via aiohttp
15
- annotated-types==0.7.0
16
- # via pydantic
17
- anyio==4.8.0
18
- # via
19
- # gradio
20
- # httpx
21
- # openai
22
- # starlette
23
- async-timeout==4.0.3
24
- # via
25
- # aiohttp
26
- # langchain
27
- attrs==24.3.0
28
- # via aiohttp
29
- beautifulsoup4==4.12.3
30
- # via llama-index-readers-file
31
- bitarray==3.0.0
32
- # via colbert-ai
33
- blinker==1.9.0
34
- # via flask
35
- catalogue==2.0.10
36
- # via srsly
37
- certifi==2024.12.14
38
- # via
39
- # httpcore
40
- # httpx
41
- # llama-cloud
42
- # requests
43
- charset-normalizer==3.4.1
44
- # via requests
45
- click==8.1.8
46
- # via
47
- # flask
48
- # llama-parse
49
- # nltk
50
- # typer
51
- # uvicorn
52
- colbert-ai==0.2.19
53
- # via ragatouille
54
- dataclasses-json==0.6.7
55
- # via llama-index-core
56
- datasets==3.2.0
57
- # via
58
- # iclr2024-papers (pyproject.toml)
59
- # colbert-ai
60
- deprecated==1.2.15
61
- # via llama-index-core
62
- dill==0.3.8
63
- # via
64
- # datasets
65
- # multiprocess
66
- dirtyjson==1.0.8
67
- # via llama-index-core
68
- distro==1.9.0
69
- # via openai
70
- exceptiongroup==1.2.2
71
- # via anyio
72
- faiss-cpu==1.9.0.post1
73
- # via ragatouille
74
- fast-pytorch-kmeans==0.2.0.1
75
- # via ragatouille
76
- fastapi==0.115.6
77
- # via gradio
78
- ffmpy==0.5.0
79
- # via gradio
80
- filelock==3.16.1
81
- # via
82
- # datasets
83
- # huggingface-hub
84
- # torch
85
- # transformers
86
- # triton
87
- filetype==1.2.0
88
- # via llama-index-core
89
- flask==3.1.0
90
- # via colbert-ai
91
- frozenlist==1.5.0
92
- # via
93
- # aiohttp
94
- # aiosignal
95
- fsspec==2024.9.0
96
- # via
97
- # datasets
98
- # gradio-client
99
- # huggingface-hub
100
- # llama-index-core
101
- # torch
102
- git-python==1.0.3
103
- # via colbert-ai
104
- gitdb==4.0.12
105
- # via gitpython
106
- gitpython==3.1.44
107
- # via git-python
108
- gradio==5.12.0
109
- # via
110
- # iclr2024-papers (pyproject.toml)
111
- # spaces
112
- gradio-client==1.5.4
113
- # via gradio
114
- greenlet==3.1.1
115
- # via sqlalchemy
116
- h11==0.14.0
117
- # via
118
- # httpcore
119
- # uvicorn
120
- hf-transfer==0.1.9
121
- # via iclr2024-papers (pyproject.toml)
122
- httpcore==1.0.7
123
- # via httpx
124
- httpx==0.28.1
125
- # via
126
- # gradio
127
- # gradio-client
128
- # langsmith
129
- # llama-cloud
130
- # llama-index-core
131
- # openai
132
- # safehttpx
133
- # spaces
134
- huggingface-hub==0.27.1
135
- # via
136
- # datasets
137
- # gradio
138
- # gradio-client
139
- # sentence-transformers
140
- # tokenizers
141
- # transformers
142
- idna==3.10
143
- # via
144
- # anyio
145
- # httpx
146
- # requests
147
- # yarl
148
- itsdangerous==2.2.0
149
- # via flask
150
- jinja2==3.1.5
151
- # via
152
- # flask
153
- # gradio
154
- # torch
155
- jiter==0.8.2
156
- # via openai
157
- joblib==1.4.2
158
- # via
159
- # nltk
160
- # scikit-learn
161
- jsonpatch==1.33
162
- # via langchain-core
163
- jsonpointer==3.0.0
164
- # via jsonpatch
165
- langchain==0.3.14
166
- # via ragatouille
167
- langchain-core==0.3.29
168
- # via
169
- # langchain
170
- # langchain-text-splitters
171
- # ragatouille
172
- langchain-text-splitters==0.3.5
173
- # via langchain
174
- langsmith==0.2.10
175
- # via
176
- # langchain
177
- # langchain-core
178
- llama-cloud==0.1.8
179
- # via llama-index-indices-managed-llama-cloud
180
- llama-index==0.12.10
181
- # via ragatouille
182
- llama-index-agent-openai==0.4.1
183
- # via
184
- # llama-index
185
- # llama-index-program-openai
186
- llama-index-cli==0.4.0
187
- # via llama-index
188
- llama-index-core==0.12.10.post1
189
- # via
190
- # llama-index
191
- # llama-index-agent-openai
192
- # llama-index-cli
193
- # llama-index-embeddings-openai
194
- # llama-index-indices-managed-llama-cloud
195
- # llama-index-llms-openai
196
- # llama-index-multi-modal-llms-openai
197
- # llama-index-program-openai
198
- # llama-index-question-gen-openai
199
- # llama-index-readers-file
200
- # llama-index-readers-llama-parse
201
- # llama-parse
202
- llama-index-embeddings-openai==0.3.1
203
- # via
204
- # llama-index
205
- # llama-index-cli
206
- llama-index-indices-managed-llama-cloud==0.6.3
207
- # via llama-index
208
- llama-index-llms-openai==0.3.13
209
- # via
210
- # llama-index
211
- # llama-index-agent-openai
212
- # llama-index-cli
213
- # llama-index-multi-modal-llms-openai
214
- # llama-index-program-openai
215
- # llama-index-question-gen-openai
216
- llama-index-multi-modal-llms-openai==0.4.2
217
- # via llama-index
218
- llama-index-program-openai==0.3.1
219
- # via
220
- # llama-index
221
- # llama-index-question-gen-openai
222
- llama-index-question-gen-openai==0.3.0
223
- # via llama-index
224
- llama-index-readers-file==0.4.3
225
- # via llama-index
226
- llama-index-readers-llama-parse==0.4.0
227
- # via llama-index
228
- llama-parse==0.5.19
229
- # via llama-index-readers-llama-parse
230
- markdown-it-py==3.0.0
231
- # via rich
232
- markupsafe==2.1.5
233
- # via
234
- # gradio
235
- # jinja2
236
- # werkzeug
237
- marshmallow==3.25.0
238
- # via dataclasses-json
239
- mdurl==0.1.2
240
- # via markdown-it-py
241
- mpmath==1.3.0
242
- # via sympy
243
- multidict==6.1.0
244
- # via
245
- # aiohttp
246
- # yarl
247
- multiprocess==0.70.16
248
- # via datasets
249
- mypy-extensions==1.0.0
250
- # via typing-inspect
251
- nest-asyncio==1.6.0
252
- # via llama-index-core
253
- networkx==3.4.2
254
- # via
255
- # llama-index-core
256
- # torch
257
- ninja==1.11.1.3
258
- # via colbert-ai
259
- nltk==3.9.1
260
- # via
261
- # llama-index
262
- # llama-index-core
263
- numpy==1.26.4
264
- # via
265
- # datasets
266
- # faiss-cpu
267
- # fast-pytorch-kmeans
268
- # gradio
269
- # langchain
270
- # llama-index-core
271
- # onnx
272
- # pandas
273
- # scikit-learn
274
- # scipy
275
- # sentence-transformers
276
- # transformers
277
- # voyager
278
- nvidia-cublas-cu12==12.4.5.8
279
- # via
280
- # nvidia-cudnn-cu12
281
- # nvidia-cusolver-cu12
282
- # torch
283
- nvidia-cuda-cupti-cu12==12.4.127
284
- # via torch
285
- nvidia-cuda-nvrtc-cu12==12.4.127
286
- # via torch
287
- nvidia-cuda-runtime-cu12==12.4.127
288
- # via torch
289
- nvidia-cudnn-cu12==9.1.0.70
290
- # via torch
291
- nvidia-cufft-cu12==11.2.1.3
292
- # via torch
293
- nvidia-curand-cu12==10.3.5.147
294
- # via torch
295
- nvidia-cusolver-cu12==11.6.1.9
296
- # via torch
297
- nvidia-cusparse-cu12==12.3.1.170
298
- # via
299
- # nvidia-cusolver-cu12
300
- # torch
301
- nvidia-ml-py==12.560.30
302
- # via pynvml
303
- nvidia-nccl-cu12==2.21.5
304
- # via torch
305
- nvidia-nvjitlink-cu12==12.4.127
306
- # via
307
- # nvidia-cusolver-cu12
308
- # nvidia-cusparse-cu12
309
- # torch
310
- nvidia-nvtx-cu12==12.4.127
311
- # via torch
312
- onnx==1.17.0
313
- # via ragatouille
314
- openai==1.59.6
315
- # via
316
- # llama-index-agent-openai
317
- # llama-index-embeddings-openai
318
- # llama-index-llms-openai
319
- orjson==3.10.14
320
- # via
321
- # gradio
322
- # langsmith
323
- packaging==24.2
324
- # via
325
- # datasets
326
- # faiss-cpu
327
- # gradio
328
- # gradio-client
329
- # huggingface-hub
330
- # langchain-core
331
- # marshmallow
332
- # spaces
333
- # transformers
334
- pandas==2.2.3
335
- # via
336
- # datasets
337
- # gradio
338
- # llama-index-readers-file
339
- pillow==11.1.0
340
- # via
341
- # gradio
342
- # llama-index-core
343
- # sentence-transformers
344
- propcache==0.2.1
345
- # via
346
- # aiohttp
347
- # yarl
348
- protobuf==5.29.3
349
- # via onnx
350
- psutil==5.9.8
351
- # via spaces
352
- pyarrow==18.1.0
353
- # via datasets
354
- pydantic==2.10.5
355
- # via
356
- # fastapi
357
- # gradio
358
- # langchain
359
- # langchain-core
360
- # langsmith
361
- # llama-cloud
362
- # llama-index-core
363
- # llama-parse
364
- # openai
365
- # spaces
366
- pydantic-core==2.27.2
367
- # via pydantic
368
- pydub==0.25.1
369
- # via gradio
370
- pygments==2.19.1
371
- # via rich
372
- pynvml==12.0.0
373
- # via fast-pytorch-kmeans
374
- pypdf==5.1.0
375
- # via llama-index-readers-file
376
- python-dateutil==2.9.0.post0
377
- # via pandas
378
- python-dotenv==1.0.1
379
- # via colbert-ai
380
- python-multipart==0.0.20
381
- # via gradio
382
- pytz==2024.2
383
- # via pandas
384
- pyyaml==6.0.2
385
- # via
386
- # datasets
387
- # gradio
388
- # huggingface-hub
389
- # langchain
390
- # langchain-core
391
- # llama-index-core
392
- # transformers
393
- ragatouille==0.0.8.post4
394
- # via iclr2024-papers (pyproject.toml)
395
- regex==2024.11.6
396
- # via
397
- # nltk
398
- # tiktoken
399
- # transformers
400
- requests==2.32.3
401
- # via
402
- # datasets
403
- # huggingface-hub
404
- # langchain
405
- # langsmith
406
- # llama-index-core
407
- # requests-toolbelt
408
- # spaces
409
- # tiktoken
410
- # transformers
411
- requests-toolbelt==1.0.0
412
- # via langsmith
413
- rich==13.9.4
414
- # via typer
415
- ruff==0.9.1
416
- # via gradio
417
- safehttpx==0.1.6
418
- # via gradio
419
- safetensors==0.5.2
420
- # via transformers
421
- scikit-learn==1.6.1
422
- # via sentence-transformers
423
- scipy==1.15.1
424
- # via
425
- # colbert-ai
426
- # scikit-learn
427
- # sentence-transformers
428
- semantic-version==2.10.0
429
- # via gradio
430
- sentence-transformers==2.7.0
431
- # via ragatouille
432
- setuptools==75.8.0
433
- # via iclr2024-papers (pyproject.toml)
434
- shellingham==1.5.4
435
- # via typer
436
- six==1.17.0
437
- # via python-dateutil
438
- smmap==5.0.2
439
- # via gitdb
440
- sniffio==1.3.1
441
- # via
442
- # anyio
443
- # openai
444
- soupsieve==2.6
445
- # via beautifulsoup4
446
- spaces==0.32.0
447
- # via iclr2024-papers (pyproject.toml)
448
- sqlalchemy==2.0.37
449
- # via
450
- # langchain
451
- # llama-index-core
452
- srsly==2.4.8
453
- # via ragatouille
454
- starlette==0.41.3
455
- # via
456
- # fastapi
457
- # gradio
458
- striprtf==0.0.26
459
- # via llama-index-readers-file
460
- sympy==1.13.1
461
- # via torch
462
- tenacity==9.0.0
463
- # via
464
- # langchain
465
- # langchain-core
466
- # llama-index-core
467
- threadpoolctl==3.5.0
468
- # via scikit-learn
469
- tiktoken==0.8.0
470
- # via llama-index-core
471
- tokenizers==0.21.0
472
- # via transformers
473
- tomlkit==0.13.2
474
- # via gradio
475
- torch==2.5.1
476
- # via
477
- # fast-pytorch-kmeans
478
- # ragatouille
479
- # sentence-transformers
480
- tqdm==4.67.1
481
- # via
482
- # colbert-ai
483
- # datasets
484
- # huggingface-hub
485
- # llama-index-core
486
- # nltk
487
- # openai
488
- # sentence-transformers
489
- # transformers
490
- transformers==4.48.0
491
- # via
492
- # colbert-ai
493
- # ragatouille
494
- # sentence-transformers
495
- triton==3.1.0
496
- # via torch
497
- typer==0.15.1
498
- # via gradio
499
- typing-extensions==4.12.2
500
- # via
501
- # anyio
502
- # fastapi
503
- # gradio
504
- # gradio-client
505
- # huggingface-hub
506
- # langchain-core
507
- # llama-index-core
508
- # multidict
509
- # openai
510
- # pydantic
511
- # pydantic-core
512
- # pypdf
513
- # rich
514
- # spaces
515
- # sqlalchemy
516
- # torch
517
- # typer
518
- # typing-inspect
519
- # uvicorn
520
- typing-inspect==0.9.0
521
- # via
522
- # dataclasses-json
523
- # llama-index-core
524
- tzdata==2024.2
525
- # via pandas
526
- ujson==5.10.0
527
- # via colbert-ai
528
- urllib3==2.3.0
529
- # via requests
530
- uvicorn==0.34.0
531
- # via gradio
532
- voyager==2.1.0
533
- # via ragatouille
534
- websockets==14.1
535
- # via gradio-client
536
- werkzeug==3.1.3
537
- # via flask
538
- wrapt==1.17.1
539
- # via
540
- # deprecated
541
- # llama-index-core
542
- xxhash==3.5.0
543
- # via datasets
544
- yarl==1.18.3
545
- # via aiohttp
 
1
+ datasets==2.18.0
2
+ #gradio==4.27.0
3
+ huggingface_hub==0.22.2
4
+ pandas==2.2.1
5
+ ragatouille==0.0.8.post2
6
+ tqdm==4.66.2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
uv.lock DELETED
The diff for this file is too large to render. See raw diff