hayas commited on
Commit
61cf706
·
1 Parent(s): aad286b
Files changed (7) hide show
  1. .pre-commit-config.yaml +6 -24
  2. .python-version +1 -1
  3. README.md +2 -1
  4. app.py +4 -3
  5. pyproject.toml +29 -16
  6. requirements.txt +168 -103
  7. uv.lock +0 -0
.pre-commit-config.yaml CHANGED
@@ -1,6 +1,6 @@
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
- rev: v5.0.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
@@ -14,38 +14,20 @@ repos:
14
  - id: requirements-txt-fixer
15
  - id: trailing-whitespace
16
  - repo: https://github.com/astral-sh/ruff-pre-commit
17
- rev: v0.8.4
18
  hooks:
19
- - id: ruff
20
  args: ["--fix"]
21
  - id: ruff-format
22
- args: ["--line-length", "119"]
23
  - repo: https://github.com/pre-commit/mirrors-mypy
24
- rev: v1.14.0
25
  hooks:
26
  - id: mypy
27
  args: ["--ignore-missing-imports"]
28
  additional_dependencies:
29
  [
30
  "types-python-slugify",
31
- "types-requests",
32
- "types-PyYAML",
33
  "types-pytz",
 
 
34
  ]
35
- - repo: https://github.com/kynan/nbstripout
36
- rev: 0.8.1
37
- hooks:
38
- - id: nbstripout
39
- args:
40
- [
41
- "--extra-keys",
42
- "metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
43
- ]
44
- - repo: https://github.com/nbQA-dev/nbQA
45
- rev: 1.9.1
46
- hooks:
47
- - id: nbqa-black
48
- - id: nbqa-pyupgrade
49
- args: ["--py37-plus"]
50
- - id: nbqa-isort
51
- args: ["--float-to-top"]
 
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
+ rev: v6.0.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
 
14
  - id: requirements-txt-fixer
15
  - id: trailing-whitespace
16
  - repo: https://github.com/astral-sh/ruff-pre-commit
17
+ rev: v0.15.4
18
  hooks:
19
+ - id: ruff-check
20
  args: ["--fix"]
21
  - id: ruff-format
 
22
  - repo: https://github.com/pre-commit/mirrors-mypy
23
+ rev: v1.19.1
24
  hooks:
25
  - id: mypy
26
  args: ["--ignore-missing-imports"]
27
  additional_dependencies:
28
  [
29
  "types-python-slugify",
 
 
30
  "types-pytz",
31
+ "types-PyYAML",
32
+ "types-requests",
33
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.python-version CHANGED
@@ -1 +1 @@
1
- 3.10
 
1
+ 3.12
README.md CHANGED
@@ -4,7 +4,8 @@ emoji: 🐢
4
  colorFrom: purple
5
  colorTo: purple
6
  sdk: gradio
7
- sdk_version: 5.9.1
 
8
  app_file: app.py
9
  pinned: false
10
  ---
 
4
  colorFrom: purple
5
  colorTo: purple
6
  sdk: gradio
7
+ sdk_version: 6.8.0
8
+ python_version: 3.12.12
9
  app_file: app.py
10
  pinned: false
11
  ---
app.py CHANGED
@@ -106,7 +106,7 @@ def process_example(instruction: str, input_text: str) -> Iterator[str]:
106
  yield from run(instruction, input_text)
107
 
108
 
109
- with gr.Blocks(css_paths="style.css") as demo:
110
  gr.Markdown(DESCRIPTION)
111
 
112
  with gr.Row():
@@ -128,6 +128,7 @@ with gr.Blocks(css_paths="style.css") as demo:
128
  inputs=[instruction, input_text, max_new_tokens, temperature, top_p],
129
  outputs=output,
130
  api_name="run",
 
131
  )
132
 
133
  gr.Examples(
@@ -146,8 +147,8 @@ with gr.Blocks(css_paths="style.css") as demo:
146
  outputs=output,
147
  fn=process_example,
148
  cache_examples=os.getenv("CACHE_EXAMPLES") == "1",
149
- api_name=False,
150
  )
151
 
152
  if __name__ == "__main__":
153
- demo.launch()
 
106
  yield from run(instruction, input_text)
107
 
108
 
109
+ with gr.Blocks() as demo:
110
  gr.Markdown(DESCRIPTION)
111
 
112
  with gr.Row():
 
128
  inputs=[instruction, input_text, max_new_tokens, temperature, top_p],
129
  outputs=output,
130
  api_name="run",
131
+ api_visibility="public",
132
  )
133
 
134
  gr.Examples(
 
147
  outputs=output,
148
  fn=process_example,
149
  cache_examples=os.getenv("CACHE_EXAMPLES") == "1",
150
+ api_visibility="private",
151
  )
152
 
153
  if __name__ == "__main__":
154
+ demo.launch(css_paths="style.css")
pyproject.toml CHANGED
@@ -3,20 +3,19 @@ name = "swallow-13b-instruct"
3
  version = "0.1.0"
4
  description = ""
5
  readme = "README.md"
6
- requires-python = ">=3.10"
7
  dependencies = [
8
- "accelerate>=1.2.1",
9
- "bitsandbytes>=0.45.0",
10
- "blobfile>=3.0.0",
11
- "gradio>=5.9.1",
12
- "hf-transfer>=0.1.8",
13
- "protobuf>=5.29.2",
14
- "sentencepiece>=0.2.0",
15
- "setuptools>=75.6.0",
16
- "spaces>=0.31.1",
17
- "tiktoken>=0.8.0",
18
- "torch==2.4.0",
19
- "transformers>=4.47.1",
20
  ]
21
 
22
  [tool.ruff]
@@ -30,8 +29,7 @@ ignore = [
30
  "D213", # multi-line-summary-second-line
31
  "E501", # line-too-long
32
  "SIM117", # multiple-with-statements
33
- ]
34
- extend-ignore = [
35
  "D100", # undocumented-public-module
36
  "D101", # undocumented-public-class
37
  "D102", # undocumented-public-method
@@ -42,7 +40,7 @@ extend-ignore = [
42
  "EM101", # raw-string-in-exception
43
  "FBT001", # boolean-type-hint-positional-argument
44
  "FBT002", # boolean-default-value-positional-argument
45
- "PD901", # pandas-df-variable-name
46
  "PGH003", # blanket-type-ignore
47
  "PLR0913", # too-many-arguments
48
  "PLR0915", # too-many-statements
@@ -52,5 +50,20 @@ unfixable = [
52
  "F401", # unused-import
53
  ]
54
 
 
 
 
 
 
 
55
  [tool.ruff.format]
56
  docstring-code-format = true
 
 
 
 
 
 
 
 
 
 
3
  version = "0.1.0"
4
  description = ""
5
  readme = "README.md"
6
+ requires-python = ">=3.12"
7
  dependencies = [
8
+ "accelerate>=1.12.0",
9
+ "bitsandbytes>=0.49.2",
10
+ "blobfile>=3.2.0",
11
+ "gradio>=6.8.0",
12
+ "protobuf>=7.34.0",
13
+ "sentencepiece>=0.2.1",
14
+ "setuptools>=82.0.0",
15
+ "spaces>=0.47.0",
16
+ "tiktoken>=0.12.0",
17
+ "torch==2.9.1",
18
+ "transformers>=5.2.0",
 
19
  ]
20
 
21
  [tool.ruff]
 
29
  "D213", # multi-line-summary-second-line
30
  "E501", # line-too-long
31
  "SIM117", # multiple-with-statements
32
+ #
 
33
  "D100", # undocumented-public-module
34
  "D101", # undocumented-public-class
35
  "D102", # undocumented-public-method
 
40
  "EM101", # raw-string-in-exception
41
  "FBT001", # boolean-type-hint-positional-argument
42
  "FBT002", # boolean-default-value-positional-argument
43
+ "ISC001", # single-line-implicit-string-concatenation
44
  "PGH003", # blanket-type-ignore
45
  "PLR0913", # too-many-arguments
46
  "PLR0915", # too-many-statements
 
50
  "F401", # unused-import
51
  ]
52
 
53
+ [tool.ruff.lint.pydocstyle]
54
+ convention = "google"
55
+
56
+ [tool.ruff.lint.per-file-ignores]
57
+ "*.ipynb" = ["T201", "T203"]
58
+
59
  [tool.ruff.format]
60
  docstring-code-format = true
61
+
62
+ [dependency-groups]
63
+ dev = [
64
+ "pre-commit>=4.5.1",
65
+ "ruff>=0.15.4",
66
+ ]
67
+ hf-spaces = [
68
+ "datasets",
69
+ ]
requirements.txt CHANGED
@@ -1,90 +1,123 @@
1
  # This file was autogenerated by uv via the following command:
2
- # uv pip compile pyproject.toml -o requirements.txt
3
- accelerate==1.2.1
4
- # via swallow-13b-instruct (pyproject.toml)
5
- aiofiles==23.2.1
6
  # via gradio
 
 
 
 
 
 
 
 
 
 
7
  annotated-types==0.7.0
8
  # via pydantic
9
- anyio==4.7.0
10
  # via
11
  # gradio
12
  # httpx
13
  # starlette
14
- bitsandbytes==0.45.0
15
- # via swallow-13b-instruct (pyproject.toml)
16
- blobfile==3.0.0
17
- # via swallow-13b-instruct (pyproject.toml)
18
- certifi==2024.12.14
 
 
 
 
 
 
19
  # via
20
  # httpcore
21
  # httpx
22
  # requests
23
- charset-normalizer==3.4.1
24
  # via requests
25
- click==8.1.8
26
  # via
27
  # typer
28
  # uvicorn
29
- exceptiongroup==1.2.2
30
- # via anyio
31
- fastapi==0.115.6
 
 
 
 
 
 
 
32
  # via gradio
33
- ffmpy==0.5.0
34
  # via gradio
35
- filelock==3.16.1
36
  # via
37
  # blobfile
 
38
  # huggingface-hub
39
  # torch
40
- # transformers
41
- # triton
42
- fsspec==2024.12.0
43
  # via
 
 
 
 
 
44
  # gradio-client
45
  # huggingface-hub
46
  # torch
47
- gradio==5.9.1
48
  # via
49
- # swallow-13b-instruct (pyproject.toml)
50
  # spaces
51
- gradio-client==1.5.2
 
 
 
52
  # via gradio
53
- h11==0.14.0
54
  # via
55
  # httpcore
56
  # uvicorn
57
- hf-transfer==0.1.8
58
- # via swallow-13b-instruct (pyproject.toml)
59
- httpcore==1.0.7
60
  # via httpx
61
  httpx==0.28.1
62
  # via
 
63
  # gradio
64
  # gradio-client
 
65
  # safehttpx
66
  # spaces
67
- huggingface-hub==0.27.0
68
  # via
69
  # accelerate
 
70
  # gradio
71
  # gradio-client
72
  # tokenizers
73
  # transformers
74
- idna==3.10
75
  # via
76
  # anyio
77
  # httpx
78
  # requests
79
- jinja2==3.1.5
 
80
  # via
81
  # gradio
82
  # torch
83
- lxml==5.3.0
84
  # via blobfile
85
- markdown-it-py==3.0.0
86
  # via rich
87
- markupsafe==2.1.5
88
  # via
89
  # gradio
90
  # jinja2
@@ -92,174 +125,206 @@ mdurl==0.1.2
92
  # via markdown-it-py
93
  mpmath==1.3.0
94
  # via sympy
95
- networkx==3.4.2
 
 
 
 
 
 
96
  # via torch
97
- numpy==2.2.1
98
  # via
99
  # accelerate
100
  # bitsandbytes
 
101
  # gradio
102
  # pandas
103
  # transformers
104
- nvidia-cublas-cu12==12.1.3.1
105
  # via
106
  # nvidia-cudnn-cu12
107
  # nvidia-cusolver-cu12
108
  # torch
109
- nvidia-cuda-cupti-cu12==12.1.105
 
 
110
  # via torch
111
- nvidia-cuda-nvrtc-cu12==12.1.105
112
  # via torch
113
- nvidia-cuda-runtime-cu12==12.1.105
114
  # via torch
115
- nvidia-cudnn-cu12==9.1.0.70
116
  # via torch
117
- nvidia-cufft-cu12==11.0.2.54
118
  # via torch
119
- nvidia-curand-cu12==10.3.2.106
120
  # via torch
121
- nvidia-cusolver-cu12==11.4.5.107
122
  # via torch
123
- nvidia-cusparse-cu12==12.1.0.106
124
  # via
125
  # nvidia-cusolver-cu12
126
  # torch
127
- nvidia-nccl-cu12==2.20.5
128
  # via torch
129
- nvidia-nvjitlink-cu12==12.6.85
 
 
130
  # via
 
131
  # nvidia-cusolver-cu12
132
  # nvidia-cusparse-cu12
133
- nvidia-nvtx-cu12==12.1.105
 
134
  # via torch
135
- orjson==3.10.13
 
 
136
  # via gradio
137
- packaging==24.2
138
  # via
139
  # accelerate
 
 
140
  # gradio
141
  # gradio-client
142
  # huggingface-hub
143
  # spaces
144
  # transformers
145
- pandas==2.2.3
146
- # via gradio
147
- pillow==11.1.0
 
 
148
  # via gradio
149
- protobuf==5.29.2
150
- # via swallow-13b-instruct (pyproject.toml)
 
 
 
 
151
  psutil==5.9.8
152
  # via
153
  # accelerate
154
  # spaces
155
- pycryptodomex==3.21.0
 
 
156
  # via blobfile
157
- pydantic==2.10.4
158
  # via
159
  # fastapi
160
  # gradio
161
  # spaces
162
- pydantic-core==2.27.2
163
  # via pydantic
164
  pydub==0.25.1
165
  # via gradio
166
- pygments==2.18.0
167
  # via rich
168
  python-dateutil==2.9.0.post0
169
  # via pandas
170
- python-multipart==0.0.20
171
  # via gradio
172
- pytz==2024.2
173
- # via pandas
174
- pyyaml==6.0.2
175
  # via
176
  # accelerate
 
177
  # gradio
178
  # huggingface-hub
179
  # transformers
180
- regex==2024.11.6
181
  # via
182
  # tiktoken
183
  # transformers
184
- requests==2.32.3
185
  # via
186
- # huggingface-hub
187
  # spaces
188
  # tiktoken
189
- # transformers
190
- rich==13.9.4
191
  # via typer
192
- ruff==0.8.4
193
  # via gradio
194
- safehttpx==0.1.6
195
- # via gradio
196
- safetensors==0.4.5
197
  # via
198
  # accelerate
199
  # transformers
200
  semantic-version==2.10.0
201
  # via gradio
202
- sentencepiece==0.2.0
203
- # via swallow-13b-instruct (pyproject.toml)
204
- setuptools==75.6.0
205
- # via swallow-13b-instruct (pyproject.toml)
 
 
206
  shellingham==1.5.4
207
  # via typer
208
  six==1.17.0
209
  # via python-dateutil
210
- sniffio==1.3.1
211
- # via anyio
212
- spaces==0.31.1
213
- # via swallow-13b-instruct (pyproject.toml)
214
- starlette==0.41.3
215
  # via
216
  # fastapi
217
  # gradio
218
- sympy==1.13.3
219
  # via torch
220
- tiktoken==0.8.0
221
- # via swallow-13b-instruct (pyproject.toml)
222
- tokenizers==0.21.0
223
  # via transformers
224
- tomlkit==0.13.2
225
  # via gradio
226
- torch==2.4.0
227
  # via
228
- # swallow-13b-instruct (pyproject.toml)
229
  # accelerate
230
  # bitsandbytes
231
- tqdm==4.67.1
 
232
  # via
 
233
  # huggingface-hub
234
  # transformers
235
- transformers==4.47.1
236
- # via swallow-13b-instruct (pyproject.toml)
237
- triton==3.0.0
238
  # via torch
239
- typer==0.15.1
240
- # via gradio
241
- typing-extensions==4.12.2
242
  # via
 
 
 
 
 
 
243
  # anyio
244
- # bitsandbytes
245
  # fastapi
246
  # gradio
247
  # gradio-client
248
  # huggingface-hub
249
  # pydantic
250
  # pydantic-core
251
- # rich
252
  # spaces
 
253
  # torch
254
- # typer
255
- # uvicorn
256
- tzdata==2024.2
 
 
 
257
  # via pandas
258
- urllib3==2.3.0
259
  # via
260
  # blobfile
261
  # requests
262
- uvicorn==0.34.0
263
  # via gradio
264
- websockets==14.1
265
- # via gradio-client
 
 
 
1
  # This file was autogenerated by uv via the following command:
2
+ # uv export --no-hashes --no-dev --group hf-spaces --no-emit-package typer-slim -o requirements.txt
3
+ accelerate==1.12.0
4
+ # via swallow-13b-instruct
5
+ aiofiles==24.1.0
6
  # via gradio
7
+ aiohappyeyeballs==2.6.1
8
+ # via aiohttp
9
+ aiohttp==3.13.3
10
+ # via fsspec
11
+ aiosignal==1.4.0
12
+ # via aiohttp
13
+ annotated-doc==0.0.4
14
+ # via
15
+ # fastapi
16
+ # typer
17
  annotated-types==0.7.0
18
  # via pydantic
19
+ anyio==4.12.1
20
  # via
21
  # gradio
22
  # httpx
23
  # starlette
24
+ attrs==25.4.0
25
+ # via aiohttp
26
+ audioop-lts==0.2.2 ; python_full_version >= '3.13'
27
+ # via gradio
28
+ bitsandbytes==0.49.2
29
+ # via swallow-13b-instruct
30
+ blobfile==3.2.0
31
+ # via swallow-13b-instruct
32
+ brotli==1.2.0
33
+ # via gradio
34
+ certifi==2026.2.25
35
  # via
36
  # httpcore
37
  # httpx
38
  # requests
39
+ charset-normalizer==3.4.4
40
  # via requests
41
+ click==8.3.1
42
  # via
43
  # typer
44
  # uvicorn
45
+ colorama==0.4.6 ; sys_platform == 'win32'
46
+ # via
47
+ # click
48
+ # tqdm
49
+ datasets==4.6.1
50
+ dill==0.4.0
51
+ # via
52
+ # datasets
53
+ # multiprocess
54
+ fastapi==0.135.1
55
  # via gradio
56
+ ffmpy==1.0.0
57
  # via gradio
58
+ filelock==3.25.0
59
  # via
60
  # blobfile
61
+ # datasets
62
  # huggingface-hub
63
  # torch
64
+ frozenlist==1.8.0
 
 
65
  # via
66
+ # aiohttp
67
+ # aiosignal
68
+ fsspec==2026.2.0
69
+ # via
70
+ # datasets
71
  # gradio-client
72
  # huggingface-hub
73
  # torch
74
+ gradio==6.8.0
75
  # via
 
76
  # spaces
77
+ # swallow-13b-instruct
78
+ gradio-client==2.2.0
79
+ # via gradio
80
+ groovy==0.1.2
81
  # via gradio
82
+ h11==0.16.0
83
  # via
84
  # httpcore
85
  # uvicorn
86
+ hf-xet==1.3.2 ; platform_machine == 'AMD64' or platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64'
87
+ # via huggingface-hub
88
+ httpcore==1.0.9
89
  # via httpx
90
  httpx==0.28.1
91
  # via
92
+ # datasets
93
  # gradio
94
  # gradio-client
95
+ # huggingface-hub
96
  # safehttpx
97
  # spaces
98
+ huggingface-hub==1.5.0
99
  # via
100
  # accelerate
101
+ # datasets
102
  # gradio
103
  # gradio-client
104
  # tokenizers
105
  # transformers
106
+ idna==3.11
107
  # via
108
  # anyio
109
  # httpx
110
  # requests
111
+ # yarl
112
+ jinja2==3.1.6
113
  # via
114
  # gradio
115
  # torch
116
+ lxml==6.0.2
117
  # via blobfile
118
+ markdown-it-py==4.0.0
119
  # via rich
120
+ markupsafe==3.0.3
121
  # via
122
  # gradio
123
  # jinja2
 
125
  # via markdown-it-py
126
  mpmath==1.3.0
127
  # via sympy
128
+ multidict==6.7.1
129
+ # via
130
+ # aiohttp
131
+ # yarl
132
+ multiprocess==0.70.18
133
+ # via datasets
134
+ networkx==3.6.1
135
  # via torch
136
+ numpy==2.4.2
137
  # via
138
  # accelerate
139
  # bitsandbytes
140
+ # datasets
141
  # gradio
142
  # pandas
143
  # transformers
144
+ nvidia-cublas-cu12==12.8.4.1 ; platform_machine == 'x86_64' and sys_platform == 'linux'
145
  # via
146
  # nvidia-cudnn-cu12
147
  # nvidia-cusolver-cu12
148
  # torch
149
+ nvidia-cuda-cupti-cu12==12.8.90 ; platform_machine == 'x86_64' and sys_platform == 'linux'
150
+ # via torch
151
+ nvidia-cuda-nvrtc-cu12==12.8.93 ; platform_machine == 'x86_64' and sys_platform == 'linux'
152
  # via torch
153
+ nvidia-cuda-runtime-cu12==12.8.90 ; platform_machine == 'x86_64' and sys_platform == 'linux'
154
  # via torch
155
+ nvidia-cudnn-cu12==9.10.2.21 ; platform_machine == 'x86_64' and sys_platform == 'linux'
156
  # via torch
157
+ nvidia-cufft-cu12==11.3.3.83 ; platform_machine == 'x86_64' and sys_platform == 'linux'
158
  # via torch
159
+ nvidia-cufile-cu12==1.13.1.3 ; platform_machine == 'x86_64' and sys_platform == 'linux'
160
  # via torch
161
+ nvidia-curand-cu12==10.3.9.90 ; platform_machine == 'x86_64' and sys_platform == 'linux'
162
  # via torch
163
+ nvidia-cusolver-cu12==11.7.3.90 ; platform_machine == 'x86_64' and sys_platform == 'linux'
164
  # via torch
165
+ nvidia-cusparse-cu12==12.5.8.93 ; platform_machine == 'x86_64' and sys_platform == 'linux'
166
  # via
167
  # nvidia-cusolver-cu12
168
  # torch
169
+ nvidia-cusparselt-cu12==0.7.1 ; platform_machine == 'x86_64' and sys_platform == 'linux'
170
  # via torch
171
+ nvidia-nccl-cu12==2.27.5 ; platform_machine == 'x86_64' and sys_platform == 'linux'
172
+ # via torch
173
+ nvidia-nvjitlink-cu12==12.8.93 ; platform_machine == 'x86_64' and sys_platform == 'linux'
174
  # via
175
+ # nvidia-cufft-cu12
176
  # nvidia-cusolver-cu12
177
  # nvidia-cusparse-cu12
178
+ # torch
179
+ nvidia-nvshmem-cu12==3.3.20 ; platform_machine == 'x86_64' and sys_platform == 'linux'
180
  # via torch
181
+ nvidia-nvtx-cu12==12.8.90 ; platform_machine == 'x86_64' and sys_platform == 'linux'
182
+ # via torch
183
+ orjson==3.11.7
184
  # via gradio
185
+ packaging==26.0
186
  # via
187
  # accelerate
188
+ # bitsandbytes
189
+ # datasets
190
  # gradio
191
  # gradio-client
192
  # huggingface-hub
193
  # spaces
194
  # transformers
195
+ pandas==3.0.1
196
+ # via
197
+ # datasets
198
+ # gradio
199
+ pillow==12.1.1
200
  # via gradio
201
+ propcache==0.4.1
202
+ # via
203
+ # aiohttp
204
+ # yarl
205
+ protobuf==7.34.0
206
+ # via swallow-13b-instruct
207
  psutil==5.9.8
208
  # via
209
  # accelerate
210
  # spaces
211
+ pyarrow==23.0.1
212
+ # via datasets
213
+ pycryptodomex==3.23.0
214
  # via blobfile
215
+ pydantic==2.12.5
216
  # via
217
  # fastapi
218
  # gradio
219
  # spaces
220
+ pydantic-core==2.41.5
221
  # via pydantic
222
  pydub==0.25.1
223
  # via gradio
224
+ pygments==2.19.2
225
  # via rich
226
  python-dateutil==2.9.0.post0
227
  # via pandas
228
+ python-multipart==0.0.22
229
  # via gradio
230
+ pytz==2025.2
231
+ # via gradio
232
+ pyyaml==6.0.3
233
  # via
234
  # accelerate
235
+ # datasets
236
  # gradio
237
  # huggingface-hub
238
  # transformers
239
+ regex==2026.2.28
240
  # via
241
  # tiktoken
242
  # transformers
243
+ requests==2.32.5
244
  # via
245
+ # datasets
246
  # spaces
247
  # tiktoken
248
+ rich==14.3.3
 
249
  # via typer
250
+ safehttpx==0.1.7
251
  # via gradio
252
+ safetensors==0.7.0
 
 
253
  # via
254
  # accelerate
255
  # transformers
256
  semantic-version==2.10.0
257
  # via gradio
258
+ sentencepiece==0.2.1
259
+ # via swallow-13b-instruct
260
+ setuptools==82.0.0
261
+ # via
262
+ # swallow-13b-instruct
263
+ # torch
264
  shellingham==1.5.4
265
  # via typer
266
  six==1.17.0
267
  # via python-dateutil
268
+ spaces==0.47.0
269
+ # via swallow-13b-instruct
270
+ starlette==0.52.1
 
 
271
  # via
272
  # fastapi
273
  # gradio
274
+ sympy==1.14.0
275
  # via torch
276
+ tiktoken==0.12.0
277
+ # via swallow-13b-instruct
278
+ tokenizers==0.22.2
279
  # via transformers
280
+ tomlkit==0.13.3
281
  # via gradio
282
+ torch==2.9.1
283
  # via
 
284
  # accelerate
285
  # bitsandbytes
286
+ # swallow-13b-instruct
287
+ tqdm==4.67.3
288
  # via
289
+ # datasets
290
  # huggingface-hub
291
  # transformers
292
+ transformers==5.2.0
293
+ # via swallow-13b-instruct
294
+ triton==3.5.1 ; platform_machine == 'x86_64' and sys_platform == 'linux'
295
  # via torch
296
+ typer==0.24.1
 
 
297
  # via
298
+ # gradio
299
+ # huggingface-hub
300
+ # typer-slim
301
+ typing-extensions==4.15.0
302
+ # via
303
+ # aiosignal
304
  # anyio
 
305
  # fastapi
306
  # gradio
307
  # gradio-client
308
  # huggingface-hub
309
  # pydantic
310
  # pydantic-core
 
311
  # spaces
312
+ # starlette
313
  # torch
314
+ # typing-inspection
315
+ typing-inspection==0.4.2
316
+ # via
317
+ # fastapi
318
+ # pydantic
319
+ tzdata==2025.3 ; sys_platform == 'emscripten' or sys_platform == 'win32'
320
  # via pandas
321
+ urllib3==2.6.3
322
  # via
323
  # blobfile
324
  # requests
325
+ uvicorn==0.41.0
326
  # via gradio
327
+ xxhash==3.6.0
328
+ # via datasets
329
+ yarl==1.23.0
330
+ # via aiohttp
uv.lock CHANGED
The diff for this file is too large to render. See raw diff