Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Update 18449
#2
by
gsarti
- opened
- .pre-commit-config.yaml +36 -9
- .python-version +0 -1
- .vscode/extensions.json +0 -8
- .vscode/settings.json +16 -3
- README.md +2 -2
- app.py +8 -10
- papers.py +10 -21
- pyproject.toml +0 -50
- requirements.txt +6 -545
- uv.lock +0 -0
.pre-commit-config.yaml
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
repos:
|
2 |
- repo: https://github.com/pre-commit/pre-commit-hooks
|
3 |
-
rev:
|
4 |
hooks:
|
5 |
- id: check-executables-have-shebangs
|
6 |
- id: check-json
|
@@ -13,21 +13,48 @@ repos:
|
|
13 |
args: ["--fix=lf"]
|
14 |
- id: requirements-txt-fixer
|
15 |
- id: trailing-whitespace
|
16 |
-
- repo: https://github.com/
|
17 |
-
rev:
|
18 |
hooks:
|
19 |
-
- id:
|
20 |
-
args: ["--
|
21 |
-
|
|
|
|
|
|
|
|
|
22 |
- repo: https://github.com/pre-commit/mirrors-mypy
|
23 |
-
rev: v1.
|
24 |
hooks:
|
25 |
- id: mypy
|
26 |
args: ["--ignore-missing-imports"]
|
27 |
additional_dependencies:
|
28 |
[
|
29 |
"types-python-slugify",
|
30 |
-
"types-pytz",
|
31 |
-
"types-PyYAML",
|
32 |
"types-requests",
|
|
|
|
|
33 |
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
repos:
|
2 |
- repo: https://github.com/pre-commit/pre-commit-hooks
|
3 |
+
rev: v4.6.0
|
4 |
hooks:
|
5 |
- id: check-executables-have-shebangs
|
6 |
- id: check-json
|
|
|
13 |
args: ["--fix=lf"]
|
14 |
- id: requirements-txt-fixer
|
15 |
- id: trailing-whitespace
|
16 |
+
- repo: https://github.com/myint/docformatter
|
17 |
+
rev: v1.7.5
|
18 |
hooks:
|
19 |
+
- id: docformatter
|
20 |
+
args: ["--in-place"]
|
21 |
+
- repo: https://github.com/pycqa/isort
|
22 |
+
rev: 5.13.2
|
23 |
+
hooks:
|
24 |
+
- id: isort
|
25 |
+
args: ["--profile", "black"]
|
26 |
- repo: https://github.com/pre-commit/mirrors-mypy
|
27 |
+
rev: v1.9.0
|
28 |
hooks:
|
29 |
- id: mypy
|
30 |
args: ["--ignore-missing-imports"]
|
31 |
additional_dependencies:
|
32 |
[
|
33 |
"types-python-slugify",
|
|
|
|
|
34 |
"types-requests",
|
35 |
+
"types-PyYAML",
|
36 |
+
"types-pytz",
|
37 |
]
|
38 |
+
- repo: https://github.com/psf/black
|
39 |
+
rev: 24.4.0
|
40 |
+
hooks:
|
41 |
+
- id: black
|
42 |
+
language_version: python3.10
|
43 |
+
args: ["--line-length", "119"]
|
44 |
+
- repo: https://github.com/kynan/nbstripout
|
45 |
+
rev: 0.7.1
|
46 |
+
hooks:
|
47 |
+
- id: nbstripout
|
48 |
+
args:
|
49 |
+
[
|
50 |
+
"--extra-keys",
|
51 |
+
"metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
|
52 |
+
]
|
53 |
+
- repo: https://github.com/nbQA-dev/nbQA
|
54 |
+
rev: 1.8.5
|
55 |
+
hooks:
|
56 |
+
- id: nbqa-black
|
57 |
+
- id: nbqa-pyupgrade
|
58 |
+
args: ["--py37-plus"]
|
59 |
+
- id: nbqa-isort
|
60 |
+
args: ["--float-to-top"]
|
.python-version
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
3.10
|
|
|
|
.vscode/extensions.json
DELETED
@@ -1,8 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"recommendations": [
|
3 |
-
"ms-python.python",
|
4 |
-
"charliermarsh.ruff",
|
5 |
-
"streetsidesoftware.code-spell-checker",
|
6 |
-
"tamasfe.even-better-toml"
|
7 |
-
]
|
8 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.vscode/settings.json
CHANGED
@@ -2,16 +2,29 @@
|
|
2 |
"editor.formatOnSave": true,
|
3 |
"files.insertFinalNewline": false,
|
4 |
"[python]": {
|
5 |
-
"editor.defaultFormatter": "
|
6 |
"editor.formatOnType": true,
|
7 |
"editor.codeActionsOnSave": {
|
8 |
-
"source.fixAll.ruff": "explicit",
|
9 |
"source.organizeImports": "explicit"
|
10 |
}
|
11 |
},
|
12 |
"[jupyter]": {
|
13 |
"files.insertFinalNewline": false
|
14 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
"notebook.output.scrolling": true,
|
16 |
-
"notebook.
|
|
|
|
|
|
|
|
|
17 |
}
|
|
|
2 |
"editor.formatOnSave": true,
|
3 |
"files.insertFinalNewline": false,
|
4 |
"[python]": {
|
5 |
+
"editor.defaultFormatter": "ms-python.black-formatter",
|
6 |
"editor.formatOnType": true,
|
7 |
"editor.codeActionsOnSave": {
|
|
|
8 |
"source.organizeImports": "explicit"
|
9 |
}
|
10 |
},
|
11 |
"[jupyter]": {
|
12 |
"files.insertFinalNewline": false
|
13 |
},
|
14 |
+
"black-formatter.args": [
|
15 |
+
"--line-length=119"
|
16 |
+
],
|
17 |
+
"isort.args": ["--profile", "black"],
|
18 |
+
"flake8.args": [
|
19 |
+
"--max-line-length=119"
|
20 |
+
],
|
21 |
+
"ruff.lint.args": [
|
22 |
+
"--line-length=119"
|
23 |
+
],
|
24 |
"notebook.output.scrolling": true,
|
25 |
+
"notebook.formatOnCellExecution": true,
|
26 |
+
"notebook.formatOnSave.enabled": true,
|
27 |
+
"notebook.codeActionsOnSave": {
|
28 |
+
"source.organizeImports": "explicit"
|
29 |
+
}
|
30 |
}
|
README.md
CHANGED
@@ -4,9 +4,9 @@ emoji: 📊
|
|
4 |
colorFrom: green
|
5 |
colorTo: purple
|
6 |
sdk: gradio
|
7 |
-
sdk_version:
|
8 |
app_file: app.py
|
9 |
-
pinned:
|
10 |
header: mini
|
11 |
datasets:
|
12 |
- ICLR2024/ICLR2024-papers
|
|
|
4 |
colorFrom: green
|
5 |
colorTo: purple
|
6 |
sdk: gradio
|
7 |
+
sdk_version: 4.27.0
|
8 |
app_file: app.py
|
9 |
+
pinned: false
|
10 |
header: mini
|
11 |
datasets:
|
12 |
- ICLR2024/ICLR2024-papers
|
app.py
CHANGED
@@ -8,6 +8,10 @@ from papers import PaperList
|
|
8 |
DESCRIPTION = "# ICLR 2024 Papers"
|
9 |
|
10 |
TUTORIAL = """\
|
|
|
|
|
|
|
|
|
11 |
#### Tutorial for claiming the ICLR 2024 papers
|
12 |
|
13 |
1. Find your paper in the table below.
|
@@ -22,10 +26,6 @@ The admin team will validate your request soon. Once confirmed, the Paper page w
|
|
22 |
If you need further assistance, see the guide [here](https://huggingface.co/docs/hub/paper-pages#claiming-authorship-to-a-paper).
|
23 |
|
24 |
If your paper is not yet indexed on Hugging Face, you can index it by following this [guide](https://huggingface.co/docs/hub/paper-pages#can-i-have-a-paper-page-even-if-i-have-no-modeldatasetspace) and open a PR with [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to add your Paper page to this Space.
|
25 |
-
|
26 |
-
#### Tutorial for creating a PR
|
27 |
-
|
28 |
-
To add data to the table below, please use [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to create a PR.
|
29 |
"""
|
30 |
|
31 |
|
@@ -35,10 +35,7 @@ DEFAULT_COLUMNS = [
|
|
35 |
"Title",
|
36 |
"Type",
|
37 |
"Paper page",
|
38 |
-
"👍",
|
39 |
-
"💬",
|
40 |
"OpenReview",
|
41 |
-
"Project page",
|
42 |
"GitHub",
|
43 |
"Spaces",
|
44 |
"Models",
|
@@ -48,7 +45,8 @@ DEFAULT_COLUMNS = [
|
|
48 |
def update_num_papers(df: pd.DataFrame) -> str:
|
49 |
if "claimed" in df.columns:
|
50 |
return f"{len(df)} / {len(paper_list.df_raw)} ({len(df[df['claimed'].str.contains('✅')])} claimed)"
|
51 |
-
|
|
|
52 |
|
53 |
|
54 |
def update_df(
|
@@ -72,7 +70,7 @@ def update_df(
|
|
72 |
)
|
73 |
|
74 |
|
75 |
-
with gr.Blocks(
|
76 |
gr.Markdown(DESCRIPTION)
|
77 |
with gr.Accordion(label="Tutorial", open=True):
|
78 |
gr.Markdown(TUTORIAL)
|
@@ -120,7 +118,7 @@ with gr.Blocks(css_paths="style.css") as demo:
|
|
120 |
type="pandas",
|
121 |
row_count=(0, "dynamic"),
|
122 |
interactive=False,
|
123 |
-
|
124 |
elem_id="table",
|
125 |
wrap=True,
|
126 |
)
|
|
|
8 |
DESCRIPTION = "# ICLR 2024 Papers"
|
9 |
|
10 |
TUTORIAL = """\
|
11 |
+
#### Tutorial for creating a PR
|
12 |
+
|
13 |
+
To add data to the table below, please use [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to create a PR.
|
14 |
+
|
15 |
#### Tutorial for claiming the ICLR 2024 papers
|
16 |
|
17 |
1. Find your paper in the table below.
|
|
|
26 |
If you need further assistance, see the guide [here](https://huggingface.co/docs/hub/paper-pages#claiming-authorship-to-a-paper).
|
27 |
|
28 |
If your paper is not yet indexed on Hugging Face, you can index it by following this [guide](https://huggingface.co/docs/hub/paper-pages#can-i-have-a-paper-page-even-if-i-have-no-modeldatasetspace) and open a PR with [this Space](https://huggingface.co/spaces/ICLR2024/update-ICLR2024-papers) to add your Paper page to this Space.
|
|
|
|
|
|
|
|
|
29 |
"""
|
30 |
|
31 |
|
|
|
35 |
"Title",
|
36 |
"Type",
|
37 |
"Paper page",
|
|
|
|
|
38 |
"OpenReview",
|
|
|
39 |
"GitHub",
|
40 |
"Spaces",
|
41 |
"Models",
|
|
|
45 |
def update_num_papers(df: pd.DataFrame) -> str:
|
46 |
if "claimed" in df.columns:
|
47 |
return f"{len(df)} / {len(paper_list.df_raw)} ({len(df[df['claimed'].str.contains('✅')])} claimed)"
|
48 |
+
else:
|
49 |
+
return f"{len(df)} / {len(paper_list.df_raw)}"
|
50 |
|
51 |
|
52 |
def update_df(
|
|
|
70 |
)
|
71 |
|
72 |
|
73 |
+
with gr.Blocks(css="style.css") as demo:
|
74 |
gr.Markdown(DESCRIPTION)
|
75 |
with gr.Accordion(label="Tutorial", open=True):
|
76 |
gr.Markdown(TUTORIAL)
|
|
|
118 |
type="pandas",
|
119 |
row_count=(0, "dynamic"),
|
120 |
interactive=False,
|
121 |
+
height=1000,
|
122 |
elem_id="table",
|
123 |
wrap=True,
|
124 |
)
|
papers.py
CHANGED
@@ -19,36 +19,32 @@ ABSTRACT_RETRIEVER.search("LLM")
|
|
19 |
|
20 |
|
21 |
class PaperList:
|
22 |
-
COLUMN_INFO =
|
23 |
["Title", "str"],
|
24 |
["Authors", "str"],
|
25 |
["Type", "str"],
|
26 |
["Paper page", "markdown"],
|
27 |
-
["👍", "number"],
|
28 |
-
["💬", "number"],
|
29 |
["OpenReview", "markdown"],
|
30 |
-
["Project page", "markdown"],
|
31 |
["GitHub", "markdown"],
|
32 |
["Spaces", "markdown"],
|
33 |
["Models", "markdown"],
|
34 |
["Datasets", "markdown"],
|
35 |
["claimed", "markdown"],
|
36 |
-
|
37 |
|
38 |
-
def __init__(self)
|
39 |
self.df_raw = self.get_df()
|
40 |
self.df_prettified = self.prettify(self.df_raw)
|
41 |
|
42 |
@staticmethod
|
43 |
def get_df() -> pd.DataFrame:
|
44 |
-
df =
|
45 |
-
|
46 |
-
right=datasets.load_dataset("ICLR2024/ICLR2024-
|
47 |
on="id",
|
48 |
how="left",
|
49 |
)
|
50 |
-
|
51 |
-
df[keys] = df[keys].fillna(-1).astype(int)
|
52 |
df["paper_page"] = df["arxiv_id"].apply(
|
53 |
lambda arxiv_id: f"https://huggingface.co/papers/{arxiv_id}" if arxiv_id else ""
|
54 |
)
|
@@ -66,19 +62,12 @@ class PaperList:
|
|
66 |
n_linked_authors = "" if row.n_linked_authors == -1 else row.n_linked_authors
|
67 |
n_authors = "" if row.n_authors == -1 else row.n_authors
|
68 |
claimed_paper = "" if n_linked_authors == "" else f"{n_linked_authors}/{n_authors} {author_linked}"
|
69 |
-
upvotes = "" if row.upvotes == -1 else row.upvotes
|
70 |
-
num_comments = "" if row.num_comments == -1 else row.num_comments
|
71 |
|
72 |
new_row = {
|
73 |
"Title": row["title"],
|
74 |
"Authors": ", ".join(row["authors"]),
|
75 |
"Type": row["type"],
|
76 |
"Paper page": PaperList.create_link(row["arxiv_id"], row["paper_page"]),
|
77 |
-
"Project page": (
|
78 |
-
PaperList.create_link("Project page", row["project_page"]) if row["project_page"] else ""
|
79 |
-
),
|
80 |
-
"👍": upvotes,
|
81 |
-
"💬": num_comments,
|
82 |
"OpenReview": PaperList.create_link("OpenReview", row["OpenReview"]),
|
83 |
"GitHub": "\n".join([PaperList.create_link("GitHub", url) for url in row["GitHub"]]),
|
84 |
"Spaces": "\n".join(
|
@@ -102,14 +91,14 @@ class PaperList:
|
|
102 |
return pd.DataFrame(rows, columns=PaperList.get_column_names())
|
103 |
|
104 |
@staticmethod
|
105 |
-
def get_column_names()
|
106 |
return list(map(operator.itemgetter(0), PaperList.COLUMN_INFO))
|
107 |
|
108 |
def get_column_datatypes(self, column_names: list[str]) -> list[str]:
|
109 |
mapping = dict(self.COLUMN_INFO)
|
110 |
return [mapping[name] for name in column_names]
|
111 |
|
112 |
-
def search(
|
113 |
self,
|
114 |
title_search_query: str,
|
115 |
abstract_search_query: str,
|
@@ -130,7 +119,7 @@ class PaperList:
|
|
130 |
df = df[df["type"] == presentation_type]
|
131 |
|
132 |
if "Paper page" in filter_names:
|
133 |
-
df = df[df["paper_page"]
|
134 |
if "GitHub" in filter_names:
|
135 |
df = df[df["GitHub"].apply(len) > 0]
|
136 |
if "Space" in filter_names:
|
|
|
19 |
|
20 |
|
21 |
class PaperList:
|
22 |
+
COLUMN_INFO = [
|
23 |
["Title", "str"],
|
24 |
["Authors", "str"],
|
25 |
["Type", "str"],
|
26 |
["Paper page", "markdown"],
|
|
|
|
|
27 |
["OpenReview", "markdown"],
|
|
|
28 |
["GitHub", "markdown"],
|
29 |
["Spaces", "markdown"],
|
30 |
["Models", "markdown"],
|
31 |
["Datasets", "markdown"],
|
32 |
["claimed", "markdown"],
|
33 |
+
]
|
34 |
|
35 |
+
def __init__(self):
|
36 |
self.df_raw = self.get_df()
|
37 |
self.df_prettified = self.prettify(self.df_raw)
|
38 |
|
39 |
@staticmethod
|
40 |
def get_df() -> pd.DataFrame:
|
41 |
+
df = pd.merge(
|
42 |
+
left=datasets.load_dataset("ICLR2024/ICLR2024-papers", split="train").to_pandas(),
|
43 |
+
right=datasets.load_dataset("ICLR2024/ICLR2024-num-claimed-papers", split="train").to_pandas(),
|
44 |
on="id",
|
45 |
how="left",
|
46 |
)
|
47 |
+
df[["n_authors", "n_linked_authors"]] = df[["n_authors", "n_linked_authors"]].fillna(-1).astype(int)
|
|
|
48 |
df["paper_page"] = df["arxiv_id"].apply(
|
49 |
lambda arxiv_id: f"https://huggingface.co/papers/{arxiv_id}" if arxiv_id else ""
|
50 |
)
|
|
|
62 |
n_linked_authors = "" if row.n_linked_authors == -1 else row.n_linked_authors
|
63 |
n_authors = "" if row.n_authors == -1 else row.n_authors
|
64 |
claimed_paper = "" if n_linked_authors == "" else f"{n_linked_authors}/{n_authors} {author_linked}"
|
|
|
|
|
65 |
|
66 |
new_row = {
|
67 |
"Title": row["title"],
|
68 |
"Authors": ", ".join(row["authors"]),
|
69 |
"Type": row["type"],
|
70 |
"Paper page": PaperList.create_link(row["arxiv_id"], row["paper_page"]),
|
|
|
|
|
|
|
|
|
|
|
71 |
"OpenReview": PaperList.create_link("OpenReview", row["OpenReview"]),
|
72 |
"GitHub": "\n".join([PaperList.create_link("GitHub", url) for url in row["GitHub"]]),
|
73 |
"Spaces": "\n".join(
|
|
|
91 |
return pd.DataFrame(rows, columns=PaperList.get_column_names())
|
92 |
|
93 |
@staticmethod
|
94 |
+
def get_column_names():
|
95 |
return list(map(operator.itemgetter(0), PaperList.COLUMN_INFO))
|
96 |
|
97 |
def get_column_datatypes(self, column_names: list[str]) -> list[str]:
|
98 |
mapping = dict(self.COLUMN_INFO)
|
99 |
return [mapping[name] for name in column_names]
|
100 |
|
101 |
+
def search(
|
102 |
self,
|
103 |
title_search_query: str,
|
104 |
abstract_search_query: str,
|
|
|
119 |
df = df[df["type"] == presentation_type]
|
120 |
|
121 |
if "Paper page" in filter_names:
|
122 |
+
df = df[df["paper_page"].notnull()]
|
123 |
if "GitHub" in filter_names:
|
124 |
df = df[df["GitHub"].apply(len) > 0]
|
125 |
if "Space" in filter_names:
|
pyproject.toml
DELETED
@@ -1,50 +0,0 @@
|
|
1 |
-
[project]
|
2 |
-
name = "iclr2024-papers"
|
3 |
-
version = "0.1.0"
|
4 |
-
description = ""
|
5 |
-
readme = "README.md"
|
6 |
-
requires-python = ">=3.10"
|
7 |
-
dependencies = [
|
8 |
-
"datasets>=3.2.0",
|
9 |
-
"gradio>=5.12.0",
|
10 |
-
"hf-transfer>=0.1.9",
|
11 |
-
"ragatouille>=0.0.8.post4",
|
12 |
-
"setuptools>=75.6.0",
|
13 |
-
"spaces>=0.32.0",
|
14 |
-
]
|
15 |
-
|
16 |
-
[tool.ruff]
|
17 |
-
line-length = 119
|
18 |
-
|
19 |
-
[tool.ruff.lint]
|
20 |
-
select = ["ALL"]
|
21 |
-
ignore = [
|
22 |
-
"COM812", # missing-trailing-comma
|
23 |
-
"D203", # one-blank-line-before-class
|
24 |
-
"D213", # multi-line-summary-second-line
|
25 |
-
"E501", # line-too-long
|
26 |
-
"SIM117", # multiple-with-statements
|
27 |
-
]
|
28 |
-
extend-ignore = [
|
29 |
-
"D100", # undocumented-public-module
|
30 |
-
"D101", # undocumented-public-class
|
31 |
-
"D102", # undocumented-public-method
|
32 |
-
"D103", # undocumented-public-function
|
33 |
-
"D104", # undocumented-public-package
|
34 |
-
"D105", # undocumented-magic-method
|
35 |
-
"D107", # undocumented-public-init
|
36 |
-
"EM101", # raw-string-in-exception
|
37 |
-
"FBT001", # boolean-type-hint-positional-argument
|
38 |
-
"FBT002", # boolean-default-value-positional-argument
|
39 |
-
"PD901", # pandas-df-variable-name
|
40 |
-
"PGH003", # blanket-type-ignore
|
41 |
-
"PLR0913", # too-many-arguments
|
42 |
-
"PLR0915", # too-many-statements
|
43 |
-
"TRY003", # raise-vanilla-args
|
44 |
-
]
|
45 |
-
unfixable = [
|
46 |
-
"F401", # unused-import
|
47 |
-
]
|
48 |
-
|
49 |
-
[tool.ruff.format]
|
50 |
-
docstring-code-format = true
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
requirements.txt
CHANGED
@@ -1,545 +1,6 @@
|
|
1 |
-
|
2 |
-
#
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
aiohttp==3.11.11
|
8 |
-
# via
|
9 |
-
# datasets
|
10 |
-
# fsspec
|
11 |
-
# langchain
|
12 |
-
# llama-index-core
|
13 |
-
aiosignal==1.3.2
|
14 |
-
# via aiohttp
|
15 |
-
annotated-types==0.7.0
|
16 |
-
# via pydantic
|
17 |
-
anyio==4.8.0
|
18 |
-
# via
|
19 |
-
# gradio
|
20 |
-
# httpx
|
21 |
-
# openai
|
22 |
-
# starlette
|
23 |
-
async-timeout==4.0.3
|
24 |
-
# via
|
25 |
-
# aiohttp
|
26 |
-
# langchain
|
27 |
-
attrs==24.3.0
|
28 |
-
# via aiohttp
|
29 |
-
beautifulsoup4==4.12.3
|
30 |
-
# via llama-index-readers-file
|
31 |
-
bitarray==3.0.0
|
32 |
-
# via colbert-ai
|
33 |
-
blinker==1.9.0
|
34 |
-
# via flask
|
35 |
-
catalogue==2.0.10
|
36 |
-
# via srsly
|
37 |
-
certifi==2024.12.14
|
38 |
-
# via
|
39 |
-
# httpcore
|
40 |
-
# httpx
|
41 |
-
# llama-cloud
|
42 |
-
# requests
|
43 |
-
charset-normalizer==3.4.1
|
44 |
-
# via requests
|
45 |
-
click==8.1.8
|
46 |
-
# via
|
47 |
-
# flask
|
48 |
-
# llama-parse
|
49 |
-
# nltk
|
50 |
-
# typer
|
51 |
-
# uvicorn
|
52 |
-
colbert-ai==0.2.19
|
53 |
-
# via ragatouille
|
54 |
-
dataclasses-json==0.6.7
|
55 |
-
# via llama-index-core
|
56 |
-
datasets==3.2.0
|
57 |
-
# via
|
58 |
-
# iclr2024-papers (pyproject.toml)
|
59 |
-
# colbert-ai
|
60 |
-
deprecated==1.2.15
|
61 |
-
# via llama-index-core
|
62 |
-
dill==0.3.8
|
63 |
-
# via
|
64 |
-
# datasets
|
65 |
-
# multiprocess
|
66 |
-
dirtyjson==1.0.8
|
67 |
-
# via llama-index-core
|
68 |
-
distro==1.9.0
|
69 |
-
# via openai
|
70 |
-
exceptiongroup==1.2.2
|
71 |
-
# via anyio
|
72 |
-
faiss-cpu==1.9.0.post1
|
73 |
-
# via ragatouille
|
74 |
-
fast-pytorch-kmeans==0.2.0.1
|
75 |
-
# via ragatouille
|
76 |
-
fastapi==0.115.6
|
77 |
-
# via gradio
|
78 |
-
ffmpy==0.5.0
|
79 |
-
# via gradio
|
80 |
-
filelock==3.16.1
|
81 |
-
# via
|
82 |
-
# datasets
|
83 |
-
# huggingface-hub
|
84 |
-
# torch
|
85 |
-
# transformers
|
86 |
-
# triton
|
87 |
-
filetype==1.2.0
|
88 |
-
# via llama-index-core
|
89 |
-
flask==3.1.0
|
90 |
-
# via colbert-ai
|
91 |
-
frozenlist==1.5.0
|
92 |
-
# via
|
93 |
-
# aiohttp
|
94 |
-
# aiosignal
|
95 |
-
fsspec==2024.9.0
|
96 |
-
# via
|
97 |
-
# datasets
|
98 |
-
# gradio-client
|
99 |
-
# huggingface-hub
|
100 |
-
# llama-index-core
|
101 |
-
# torch
|
102 |
-
git-python==1.0.3
|
103 |
-
# via colbert-ai
|
104 |
-
gitdb==4.0.12
|
105 |
-
# via gitpython
|
106 |
-
gitpython==3.1.44
|
107 |
-
# via git-python
|
108 |
-
gradio==5.12.0
|
109 |
-
# via
|
110 |
-
# iclr2024-papers (pyproject.toml)
|
111 |
-
# spaces
|
112 |
-
gradio-client==1.5.4
|
113 |
-
# via gradio
|
114 |
-
greenlet==3.1.1
|
115 |
-
# via sqlalchemy
|
116 |
-
h11==0.14.0
|
117 |
-
# via
|
118 |
-
# httpcore
|
119 |
-
# uvicorn
|
120 |
-
hf-transfer==0.1.9
|
121 |
-
# via iclr2024-papers (pyproject.toml)
|
122 |
-
httpcore==1.0.7
|
123 |
-
# via httpx
|
124 |
-
httpx==0.28.1
|
125 |
-
# via
|
126 |
-
# gradio
|
127 |
-
# gradio-client
|
128 |
-
# langsmith
|
129 |
-
# llama-cloud
|
130 |
-
# llama-index-core
|
131 |
-
# openai
|
132 |
-
# safehttpx
|
133 |
-
# spaces
|
134 |
-
huggingface-hub==0.27.1
|
135 |
-
# via
|
136 |
-
# datasets
|
137 |
-
# gradio
|
138 |
-
# gradio-client
|
139 |
-
# sentence-transformers
|
140 |
-
# tokenizers
|
141 |
-
# transformers
|
142 |
-
idna==3.10
|
143 |
-
# via
|
144 |
-
# anyio
|
145 |
-
# httpx
|
146 |
-
# requests
|
147 |
-
# yarl
|
148 |
-
itsdangerous==2.2.0
|
149 |
-
# via flask
|
150 |
-
jinja2==3.1.5
|
151 |
-
# via
|
152 |
-
# flask
|
153 |
-
# gradio
|
154 |
-
# torch
|
155 |
-
jiter==0.8.2
|
156 |
-
# via openai
|
157 |
-
joblib==1.4.2
|
158 |
-
# via
|
159 |
-
# nltk
|
160 |
-
# scikit-learn
|
161 |
-
jsonpatch==1.33
|
162 |
-
# via langchain-core
|
163 |
-
jsonpointer==3.0.0
|
164 |
-
# via jsonpatch
|
165 |
-
langchain==0.3.14
|
166 |
-
# via ragatouille
|
167 |
-
langchain-core==0.3.29
|
168 |
-
# via
|
169 |
-
# langchain
|
170 |
-
# langchain-text-splitters
|
171 |
-
# ragatouille
|
172 |
-
langchain-text-splitters==0.3.5
|
173 |
-
# via langchain
|
174 |
-
langsmith==0.2.10
|
175 |
-
# via
|
176 |
-
# langchain
|
177 |
-
# langchain-core
|
178 |
-
llama-cloud==0.1.8
|
179 |
-
# via llama-index-indices-managed-llama-cloud
|
180 |
-
llama-index==0.12.10
|
181 |
-
# via ragatouille
|
182 |
-
llama-index-agent-openai==0.4.1
|
183 |
-
# via
|
184 |
-
# llama-index
|
185 |
-
# llama-index-program-openai
|
186 |
-
llama-index-cli==0.4.0
|
187 |
-
# via llama-index
|
188 |
-
llama-index-core==0.12.10.post1
|
189 |
-
# via
|
190 |
-
# llama-index
|
191 |
-
# llama-index-agent-openai
|
192 |
-
# llama-index-cli
|
193 |
-
# llama-index-embeddings-openai
|
194 |
-
# llama-index-indices-managed-llama-cloud
|
195 |
-
# llama-index-llms-openai
|
196 |
-
# llama-index-multi-modal-llms-openai
|
197 |
-
# llama-index-program-openai
|
198 |
-
# llama-index-question-gen-openai
|
199 |
-
# llama-index-readers-file
|
200 |
-
# llama-index-readers-llama-parse
|
201 |
-
# llama-parse
|
202 |
-
llama-index-embeddings-openai==0.3.1
|
203 |
-
# via
|
204 |
-
# llama-index
|
205 |
-
# llama-index-cli
|
206 |
-
llama-index-indices-managed-llama-cloud==0.6.3
|
207 |
-
# via llama-index
|
208 |
-
llama-index-llms-openai==0.3.13
|
209 |
-
# via
|
210 |
-
# llama-index
|
211 |
-
# llama-index-agent-openai
|
212 |
-
# llama-index-cli
|
213 |
-
# llama-index-multi-modal-llms-openai
|
214 |
-
# llama-index-program-openai
|
215 |
-
# llama-index-question-gen-openai
|
216 |
-
llama-index-multi-modal-llms-openai==0.4.2
|
217 |
-
# via llama-index
|
218 |
-
llama-index-program-openai==0.3.1
|
219 |
-
# via
|
220 |
-
# llama-index
|
221 |
-
# llama-index-question-gen-openai
|
222 |
-
llama-index-question-gen-openai==0.3.0
|
223 |
-
# via llama-index
|
224 |
-
llama-index-readers-file==0.4.3
|
225 |
-
# via llama-index
|
226 |
-
llama-index-readers-llama-parse==0.4.0
|
227 |
-
# via llama-index
|
228 |
-
llama-parse==0.5.19
|
229 |
-
# via llama-index-readers-llama-parse
|
230 |
-
markdown-it-py==3.0.0
|
231 |
-
# via rich
|
232 |
-
markupsafe==2.1.5
|
233 |
-
# via
|
234 |
-
# gradio
|
235 |
-
# jinja2
|
236 |
-
# werkzeug
|
237 |
-
marshmallow==3.25.0
|
238 |
-
# via dataclasses-json
|
239 |
-
mdurl==0.1.2
|
240 |
-
# via markdown-it-py
|
241 |
-
mpmath==1.3.0
|
242 |
-
# via sympy
|
243 |
-
multidict==6.1.0
|
244 |
-
# via
|
245 |
-
# aiohttp
|
246 |
-
# yarl
|
247 |
-
multiprocess==0.70.16
|
248 |
-
# via datasets
|
249 |
-
mypy-extensions==1.0.0
|
250 |
-
# via typing-inspect
|
251 |
-
nest-asyncio==1.6.0
|
252 |
-
# via llama-index-core
|
253 |
-
networkx==3.4.2
|
254 |
-
# via
|
255 |
-
# llama-index-core
|
256 |
-
# torch
|
257 |
-
ninja==1.11.1.3
|
258 |
-
# via colbert-ai
|
259 |
-
nltk==3.9.1
|
260 |
-
# via
|
261 |
-
# llama-index
|
262 |
-
# llama-index-core
|
263 |
-
numpy==1.26.4
|
264 |
-
# via
|
265 |
-
# datasets
|
266 |
-
# faiss-cpu
|
267 |
-
# fast-pytorch-kmeans
|
268 |
-
# gradio
|
269 |
-
# langchain
|
270 |
-
# llama-index-core
|
271 |
-
# onnx
|
272 |
-
# pandas
|
273 |
-
# scikit-learn
|
274 |
-
# scipy
|
275 |
-
# sentence-transformers
|
276 |
-
# transformers
|
277 |
-
# voyager
|
278 |
-
nvidia-cublas-cu12==12.4.5.8
|
279 |
-
# via
|
280 |
-
# nvidia-cudnn-cu12
|
281 |
-
# nvidia-cusolver-cu12
|
282 |
-
# torch
|
283 |
-
nvidia-cuda-cupti-cu12==12.4.127
|
284 |
-
# via torch
|
285 |
-
nvidia-cuda-nvrtc-cu12==12.4.127
|
286 |
-
# via torch
|
287 |
-
nvidia-cuda-runtime-cu12==12.4.127
|
288 |
-
# via torch
|
289 |
-
nvidia-cudnn-cu12==9.1.0.70
|
290 |
-
# via torch
|
291 |
-
nvidia-cufft-cu12==11.2.1.3
|
292 |
-
# via torch
|
293 |
-
nvidia-curand-cu12==10.3.5.147
|
294 |
-
# via torch
|
295 |
-
nvidia-cusolver-cu12==11.6.1.9
|
296 |
-
# via torch
|
297 |
-
nvidia-cusparse-cu12==12.3.1.170
|
298 |
-
# via
|
299 |
-
# nvidia-cusolver-cu12
|
300 |
-
# torch
|
301 |
-
nvidia-ml-py==12.560.30
|
302 |
-
# via pynvml
|
303 |
-
nvidia-nccl-cu12==2.21.5
|
304 |
-
# via torch
|
305 |
-
nvidia-nvjitlink-cu12==12.4.127
|
306 |
-
# via
|
307 |
-
# nvidia-cusolver-cu12
|
308 |
-
# nvidia-cusparse-cu12
|
309 |
-
# torch
|
310 |
-
nvidia-nvtx-cu12==12.4.127
|
311 |
-
# via torch
|
312 |
-
onnx==1.17.0
|
313 |
-
# via ragatouille
|
314 |
-
openai==1.59.6
|
315 |
-
# via
|
316 |
-
# llama-index-agent-openai
|
317 |
-
# llama-index-embeddings-openai
|
318 |
-
# llama-index-llms-openai
|
319 |
-
orjson==3.10.14
|
320 |
-
# via
|
321 |
-
# gradio
|
322 |
-
# langsmith
|
323 |
-
packaging==24.2
|
324 |
-
# via
|
325 |
-
# datasets
|
326 |
-
# faiss-cpu
|
327 |
-
# gradio
|
328 |
-
# gradio-client
|
329 |
-
# huggingface-hub
|
330 |
-
# langchain-core
|
331 |
-
# marshmallow
|
332 |
-
# spaces
|
333 |
-
# transformers
|
334 |
-
pandas==2.2.3
|
335 |
-
# via
|
336 |
-
# datasets
|
337 |
-
# gradio
|
338 |
-
# llama-index-readers-file
|
339 |
-
pillow==11.1.0
|
340 |
-
# via
|
341 |
-
# gradio
|
342 |
-
# llama-index-core
|
343 |
-
# sentence-transformers
|
344 |
-
propcache==0.2.1
|
345 |
-
# via
|
346 |
-
# aiohttp
|
347 |
-
# yarl
|
348 |
-
protobuf==5.29.3
|
349 |
-
# via onnx
|
350 |
-
psutil==5.9.8
|
351 |
-
# via spaces
|
352 |
-
pyarrow==18.1.0
|
353 |
-
# via datasets
|
354 |
-
pydantic==2.10.5
|
355 |
-
# via
|
356 |
-
# fastapi
|
357 |
-
# gradio
|
358 |
-
# langchain
|
359 |
-
# langchain-core
|
360 |
-
# langsmith
|
361 |
-
# llama-cloud
|
362 |
-
# llama-index-core
|
363 |
-
# llama-parse
|
364 |
-
# openai
|
365 |
-
# spaces
|
366 |
-
pydantic-core==2.27.2
|
367 |
-
# via pydantic
|
368 |
-
pydub==0.25.1
|
369 |
-
# via gradio
|
370 |
-
pygments==2.19.1
|
371 |
-
# via rich
|
372 |
-
pynvml==12.0.0
|
373 |
-
# via fast-pytorch-kmeans
|
374 |
-
pypdf==5.1.0
|
375 |
-
# via llama-index-readers-file
|
376 |
-
python-dateutil==2.9.0.post0
|
377 |
-
# via pandas
|
378 |
-
python-dotenv==1.0.1
|
379 |
-
# via colbert-ai
|
380 |
-
python-multipart==0.0.20
|
381 |
-
# via gradio
|
382 |
-
pytz==2024.2
|
383 |
-
# via pandas
|
384 |
-
pyyaml==6.0.2
|
385 |
-
# via
|
386 |
-
# datasets
|
387 |
-
# gradio
|
388 |
-
# huggingface-hub
|
389 |
-
# langchain
|
390 |
-
# langchain-core
|
391 |
-
# llama-index-core
|
392 |
-
# transformers
|
393 |
-
ragatouille==0.0.8.post4
|
394 |
-
# via iclr2024-papers (pyproject.toml)
|
395 |
-
regex==2024.11.6
|
396 |
-
# via
|
397 |
-
# nltk
|
398 |
-
# tiktoken
|
399 |
-
# transformers
|
400 |
-
requests==2.32.3
|
401 |
-
# via
|
402 |
-
# datasets
|
403 |
-
# huggingface-hub
|
404 |
-
# langchain
|
405 |
-
# langsmith
|
406 |
-
# llama-index-core
|
407 |
-
# requests-toolbelt
|
408 |
-
# spaces
|
409 |
-
# tiktoken
|
410 |
-
# transformers
|
411 |
-
requests-toolbelt==1.0.0
|
412 |
-
# via langsmith
|
413 |
-
rich==13.9.4
|
414 |
-
# via typer
|
415 |
-
ruff==0.9.1
|
416 |
-
# via gradio
|
417 |
-
safehttpx==0.1.6
|
418 |
-
# via gradio
|
419 |
-
safetensors==0.5.2
|
420 |
-
# via transformers
|
421 |
-
scikit-learn==1.6.1
|
422 |
-
# via sentence-transformers
|
423 |
-
scipy==1.15.1
|
424 |
-
# via
|
425 |
-
# colbert-ai
|
426 |
-
# scikit-learn
|
427 |
-
# sentence-transformers
|
428 |
-
semantic-version==2.10.0
|
429 |
-
# via gradio
|
430 |
-
sentence-transformers==2.7.0
|
431 |
-
# via ragatouille
|
432 |
-
setuptools==75.8.0
|
433 |
-
# via iclr2024-papers (pyproject.toml)
|
434 |
-
shellingham==1.5.4
|
435 |
-
# via typer
|
436 |
-
six==1.17.0
|
437 |
-
# via python-dateutil
|
438 |
-
smmap==5.0.2
|
439 |
-
# via gitdb
|
440 |
-
sniffio==1.3.1
|
441 |
-
# via
|
442 |
-
# anyio
|
443 |
-
# openai
|
444 |
-
soupsieve==2.6
|
445 |
-
# via beautifulsoup4
|
446 |
-
spaces==0.32.0
|
447 |
-
# via iclr2024-papers (pyproject.toml)
|
448 |
-
sqlalchemy==2.0.37
|
449 |
-
# via
|
450 |
-
# langchain
|
451 |
-
# llama-index-core
|
452 |
-
srsly==2.4.8
|
453 |
-
# via ragatouille
|
454 |
-
starlette==0.41.3
|
455 |
-
# via
|
456 |
-
# fastapi
|
457 |
-
# gradio
|
458 |
-
striprtf==0.0.26
|
459 |
-
# via llama-index-readers-file
|
460 |
-
sympy==1.13.1
|
461 |
-
# via torch
|
462 |
-
tenacity==9.0.0
|
463 |
-
# via
|
464 |
-
# langchain
|
465 |
-
# langchain-core
|
466 |
-
# llama-index-core
|
467 |
-
threadpoolctl==3.5.0
|
468 |
-
# via scikit-learn
|
469 |
-
tiktoken==0.8.0
|
470 |
-
# via llama-index-core
|
471 |
-
tokenizers==0.21.0
|
472 |
-
# via transformers
|
473 |
-
tomlkit==0.13.2
|
474 |
-
# via gradio
|
475 |
-
torch==2.5.1
|
476 |
-
# via
|
477 |
-
# fast-pytorch-kmeans
|
478 |
-
# ragatouille
|
479 |
-
# sentence-transformers
|
480 |
-
tqdm==4.67.1
|
481 |
-
# via
|
482 |
-
# colbert-ai
|
483 |
-
# datasets
|
484 |
-
# huggingface-hub
|
485 |
-
# llama-index-core
|
486 |
-
# nltk
|
487 |
-
# openai
|
488 |
-
# sentence-transformers
|
489 |
-
# transformers
|
490 |
-
transformers==4.48.0
|
491 |
-
# via
|
492 |
-
# colbert-ai
|
493 |
-
# ragatouille
|
494 |
-
# sentence-transformers
|
495 |
-
triton==3.1.0
|
496 |
-
# via torch
|
497 |
-
typer==0.15.1
|
498 |
-
# via gradio
|
499 |
-
typing-extensions==4.12.2
|
500 |
-
# via
|
501 |
-
# anyio
|
502 |
-
# fastapi
|
503 |
-
# gradio
|
504 |
-
# gradio-client
|
505 |
-
# huggingface-hub
|
506 |
-
# langchain-core
|
507 |
-
# llama-index-core
|
508 |
-
# multidict
|
509 |
-
# openai
|
510 |
-
# pydantic
|
511 |
-
# pydantic-core
|
512 |
-
# pypdf
|
513 |
-
# rich
|
514 |
-
# spaces
|
515 |
-
# sqlalchemy
|
516 |
-
# torch
|
517 |
-
# typer
|
518 |
-
# typing-inspect
|
519 |
-
# uvicorn
|
520 |
-
typing-inspect==0.9.0
|
521 |
-
# via
|
522 |
-
# dataclasses-json
|
523 |
-
# llama-index-core
|
524 |
-
tzdata==2024.2
|
525 |
-
# via pandas
|
526 |
-
ujson==5.10.0
|
527 |
-
# via colbert-ai
|
528 |
-
urllib3==2.3.0
|
529 |
-
# via requests
|
530 |
-
uvicorn==0.34.0
|
531 |
-
# via gradio
|
532 |
-
voyager==2.1.0
|
533 |
-
# via ragatouille
|
534 |
-
websockets==14.1
|
535 |
-
# via gradio-client
|
536 |
-
werkzeug==3.1.3
|
537 |
-
# via flask
|
538 |
-
wrapt==1.17.1
|
539 |
-
# via
|
540 |
-
# deprecated
|
541 |
-
# llama-index-core
|
542 |
-
xxhash==3.5.0
|
543 |
-
# via datasets
|
544 |
-
yarl==1.18.3
|
545 |
-
# via aiohttp
|
|
|
1 |
+
datasets==2.18.0
|
2 |
+
#gradio==4.27.0
|
3 |
+
huggingface_hub==0.22.2
|
4 |
+
pandas==2.2.1
|
5 |
+
ragatouille==0.0.8.post2
|
6 |
+
tqdm==4.66.2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
uv.lock
DELETED
The diff for this file is too large to render.
See raw diff
|
|