hysts HF staff commited on
Commit
749d210
·
1 Parent(s): c9f3be5
.pre-commit-config.yaml CHANGED
@@ -1,6 +1,6 @@
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
- rev: v4.6.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
@@ -18,13 +18,15 @@ repos:
18
  hooks:
19
  - id: docformatter
20
  args: ["--in-place"]
21
- - repo: https://github.com/pycqa/isort
22
- rev: 5.13.2
23
  hooks:
24
- - id: isort
25
- args: ["--profile", "black"]
 
 
26
  - repo: https://github.com/pre-commit/mirrors-mypy
27
- rev: v1.9.0
28
  hooks:
29
  - id: mypy
30
  args: ["--ignore-missing-imports"]
@@ -35,14 +37,8 @@ repos:
35
  "types-PyYAML",
36
  "types-pytz",
37
  ]
38
- - repo: https://github.com/psf/black
39
- rev: 24.4.0
40
- hooks:
41
- - id: black
42
- language_version: python3.10
43
- args: ["--line-length", "119"]
44
  - repo: https://github.com/kynan/nbstripout
45
- rev: 0.7.1
46
  hooks:
47
  - id: nbstripout
48
  args:
@@ -51,7 +47,7 @@ repos:
51
  "metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
52
  ]
53
  - repo: https://github.com/nbQA-dev/nbQA
54
- rev: 1.8.5
55
  hooks:
56
  - id: nbqa-black
57
  - id: nbqa-pyupgrade
 
1
  repos:
2
  - repo: https://github.com/pre-commit/pre-commit-hooks
3
+ rev: v5.0.0
4
  hooks:
5
  - id: check-executables-have-shebangs
6
  - id: check-json
 
18
  hooks:
19
  - id: docformatter
20
  args: ["--in-place"]
21
+ - repo: https://github.com/astral-sh/ruff-pre-commit
22
+ rev: v0.8.4
23
  hooks:
24
+ - id: ruff
25
+ args: ["--fix"]
26
+ - id: ruff-format
27
+ args: ["--line-length", "119"]
28
  - repo: https://github.com/pre-commit/mirrors-mypy
29
+ rev: v1.14.0
30
  hooks:
31
  - id: mypy
32
  args: ["--ignore-missing-imports"]
 
37
  "types-PyYAML",
38
  "types-pytz",
39
  ]
 
 
 
 
 
 
40
  - repo: https://github.com/kynan/nbstripout
41
+ rev: 0.8.1
42
  hooks:
43
  - id: nbstripout
44
  args:
 
47
  "metadata.interpreter metadata.kernelspec cell.metadata.pycharm",
48
  ]
49
  - repo: https://github.com/nbQA-dev/nbQA
50
+ rev: 1.9.1
51
  hooks:
52
  - id: nbqa-black
53
  - id: nbqa-pyupgrade
.python-version ADDED
@@ -0,0 +1 @@
 
 
1
+ 3.10
.vscode/extensions.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "recommendations": [
3
+ "ms-python.python",
4
+ "charliermarsh.ruff",
5
+ "streetsidesoftware.code-spell-checker",
6
+ "tamasfe.even-better-toml"
7
+ ]
8
+ }
.vscode/settings.json CHANGED
@@ -2,29 +2,20 @@
2
  "editor.formatOnSave": true,
3
  "files.insertFinalNewline": false,
4
  "[python]": {
5
- "editor.defaultFormatter": "ms-python.black-formatter",
6
  "editor.formatOnType": true,
7
  "editor.codeActionsOnSave": {
 
8
  "source.organizeImports": "explicit"
9
  }
10
  },
11
  "[jupyter]": {
12
  "files.insertFinalNewline": false
13
  },
14
- "black-formatter.args": [
15
- "--line-length=119"
16
- ],
17
- "isort.args": ["--profile", "black"],
18
- "flake8.args": [
19
- "--max-line-length=119"
20
- ],
21
- "ruff.lint.args": [
22
- "--line-length=119"
23
- ],
24
  "notebook.output.scrolling": true,
25
  "notebook.formatOnCellExecution": true,
26
  "notebook.formatOnSave.enabled": true,
27
- "notebook.codeActionsOnSave": {
28
- "source.organizeImports": "explicit"
29
- }
30
  }
 
2
  "editor.formatOnSave": true,
3
  "files.insertFinalNewline": false,
4
  "[python]": {
5
+ "editor.defaultFormatter": "charliermarsh.ruff",
6
  "editor.formatOnType": true,
7
  "editor.codeActionsOnSave": {
8
+ "source.fixAll.ruff": "explicit",
9
  "source.organizeImports": "explicit"
10
  }
11
  },
12
  "[jupyter]": {
13
  "files.insertFinalNewline": false
14
  },
 
 
 
 
 
 
 
 
 
 
15
  "notebook.output.scrolling": true,
16
  "notebook.formatOnCellExecution": true,
17
  "notebook.formatOnSave.enabled": true,
18
+ "notebook.codeActionsOnSave": {
19
+ "source.organizeImports": "explicit"
20
+ }
21
  }
README.md CHANGED
@@ -4,7 +4,7 @@ emoji: 🐨
4
  colorFrom: pink
5
  colorTo: blue
6
  sdk: gradio
7
- sdk_version: 4.36.1
8
  app_file: app.py
9
  pinned: false
10
  license: mit
 
4
  colorFrom: pink
5
  colorTo: blue
6
  sdk: gradio
7
+ sdk_version: 5.9.1
8
  app_file: app.py
9
  pinned: false
10
  license: mit
app.py CHANGED
@@ -1,15 +1,15 @@
1
  #!/usr/bin/env python
2
 
3
  import os
 
4
  from threading import Thread
5
- from typing import Iterator
6
 
7
  import gradio as gr
8
  import spaces
9
  import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
11
 
12
- DESCRIPTION = "# Mistral-7B v0.2"
13
 
14
  if not torch.cuda.is_available():
15
  DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
@@ -19,7 +19,7 @@ DEFAULT_MAX_NEW_TOKENS = 1024
19
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
20
 
21
  if torch.cuda.is_available():
22
- model_id = "mistralai/Mistral-7B-Instruct-v0.2"
23
  model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16, device_map="auto")
24
  tokenizer = AutoTokenizer.from_pretrained(model_id)
25
 
@@ -27,17 +27,14 @@ if torch.cuda.is_available():
27
  @spaces.GPU
28
  def generate(
29
  message: str,
30
- chat_history: list[tuple[str, str]],
31
  max_new_tokens: int = 1024,
32
  temperature: float = 0.6,
33
  top_p: float = 0.9,
34
  top_k: int = 50,
35
  repetition_penalty: float = 1.2,
36
  ) -> Iterator[str]:
37
- conversation = []
38
- for user, assistant in chat_history:
39
- conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
40
- conversation.append({"role": "user", "content": message})
41
 
42
  input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt")
43
  if input_ids.shape[1] > MAX_INPUT_TOKEN_LENGTH:
@@ -66,7 +63,7 @@ def generate(
66
  yield "".join(outputs)
67
 
68
 
69
- chat_interface = gr.ChatInterface(
70
  fn=generate,
71
  additional_inputs=[
72
  gr.Slider(
@@ -113,16 +110,10 @@ chat_interface = gr.ChatInterface(
113
  ["How many hours does it take a man to eat a Helicopter?"],
114
  ["Write a 100-word article on 'Benefits of Open-Source in AI research'"],
115
  ],
 
 
 
116
  )
117
 
118
- with gr.Blocks(css="style.css") as demo:
119
- gr.Markdown(DESCRIPTION)
120
- gr.DuplicateButton(
121
- value="Duplicate Space for private use",
122
- elem_id="duplicate-button",
123
- visible=os.getenv("SHOW_DUPLICATE_BUTTON") == "1",
124
- )
125
- chat_interface.render()
126
-
127
  if __name__ == "__main__":
128
  demo.queue(max_size=20).launch()
 
1
  #!/usr/bin/env python
2
 
3
  import os
4
+ from collections.abc import Iterator
5
  from threading import Thread
 
6
 
7
  import gradio as gr
8
  import spaces
9
  import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
11
 
12
+ DESCRIPTION = "# Mistral-7B v0.3"
13
 
14
  if not torch.cuda.is_available():
15
  DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
 
19
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
20
 
21
  if torch.cuda.is_available():
22
+ model_id = "mistralai/Mistral-7B-Instruct-v0.3"
23
  model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16, device_map="auto")
24
  tokenizer = AutoTokenizer.from_pretrained(model_id)
25
 
 
27
  @spaces.GPU
28
  def generate(
29
  message: str,
30
+ chat_history: list[dict],
31
  max_new_tokens: int = 1024,
32
  temperature: float = 0.6,
33
  top_p: float = 0.9,
34
  top_k: int = 50,
35
  repetition_penalty: float = 1.2,
36
  ) -> Iterator[str]:
37
+ conversation = [*chat_history, {"role": "user", "content": message}]
 
 
 
38
 
39
  input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt")
40
  if input_ids.shape[1] > MAX_INPUT_TOKEN_LENGTH:
 
63
  yield "".join(outputs)
64
 
65
 
66
+ demo = gr.ChatInterface(
67
  fn=generate,
68
  additional_inputs=[
69
  gr.Slider(
 
110
  ["How many hours does it take a man to eat a Helicopter?"],
111
  ["Write a 100-word article on 'Benefits of Open-Source in AI research'"],
112
  ],
113
+ type="messages",
114
+ description=DESCRIPTION,
115
+ css_paths="style.css",
116
  )
117
 
 
 
 
 
 
 
 
 
 
118
  if __name__ == "__main__":
119
  demo.queue(max_size=20).launch()
pyproject.toml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [project]
2
+ name = "mistral-7b"
3
+ version = "0.1.0"
4
+ description = "Add your description here"
5
+ readme = "README.md"
6
+ requires-python = ">=3.10"
7
+ dependencies = [
8
+ "accelerate>=1.2.1",
9
+ "bitsandbytes>=0.45.0",
10
+ "gradio>=5.9.1",
11
+ "hf-transfer>=0.1.8",
12
+ "protobuf>=5.29.2",
13
+ "sentencepiece>=0.2.0",
14
+ "spaces>=0.31.1",
15
+ "torch==2.4.0",
16
+ "transformers>=4.47.1",
17
+ ]
18
+
19
+ [tool.ruff]
20
+ line-length = 119
21
+
22
+ [tool.ruff.lint]
23
+ select = ["ALL"]
24
+ ignore = [
25
+ "COM812", # missing-trailing-comma
26
+ "D203", # one-blank-line-before-class
27
+ "D213", # multi-line-summary-second-line
28
+ "E501", # line-too-long
29
+ "SIM117", # multiple-with-statements
30
+ ]
31
+ extend-ignore = [
32
+ "D100", # undocumented-public-module
33
+ "D101", # undocumented-public-class
34
+ "D102", # undocumented-public-method
35
+ "D103", # undocumented-public-function
36
+ "D104", # undocumented-public-package
37
+ "D105", # undocumented-magic-method
38
+ "D107", # undocumented-public-init
39
+ "EM101", # raw-string-in-exception
40
+ "FBT001", # boolean-type-hint-positional-argument
41
+ "FBT002", # boolean-default-value-positional-argument
42
+ "PD901", # pandas-df-variable-name
43
+ "PGH003", # blanket-type-ignore
44
+ "PLR0913", # too-many-arguments
45
+ "PLR0915", # too-many-statements
46
+ "TRY003", # raise-vanilla-args
47
+ ]
48
+ unfixable = [
49
+ "F401", # unused-import
50
+ ]
51
+
52
+ [tool.ruff.format]
53
+ docstring-code-format = true
requirements.txt CHANGED
@@ -1,8 +1,249 @@
1
- accelerate==0.31.0
2
- bitsandbytes==0.43.1
3
- gradio==4.36.1
4
- scipy==1.13.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  sentencepiece==0.2.0
6
- spaces==0.28.3
7
- torch==2.0.1
8
- transformers==4.41.2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This file was autogenerated by uv via the following command:
2
+ # uv pip compile pyproject.toml -o requirements.txt
3
+ accelerate==1.2.1
4
+ # via mistral-7b (pyproject.toml)
5
+ aiofiles==23.2.1
6
+ # via gradio
7
+ annotated-types==0.7.0
8
+ # via pydantic
9
+ anyio==4.7.0
10
+ # via
11
+ # gradio
12
+ # httpx
13
+ # starlette
14
+ bitsandbytes==0.45.0
15
+ # via mistral-7b (pyproject.toml)
16
+ certifi==2024.12.14
17
+ # via
18
+ # httpcore
19
+ # httpx
20
+ # requests
21
+ charset-normalizer==3.4.1
22
+ # via requests
23
+ click==8.1.8
24
+ # via
25
+ # typer
26
+ # uvicorn
27
+ exceptiongroup==1.2.2
28
+ # via anyio
29
+ fastapi==0.115.6
30
+ # via gradio
31
+ ffmpy==0.5.0
32
+ # via gradio
33
+ filelock==3.16.1
34
+ # via
35
+ # huggingface-hub
36
+ # torch
37
+ # transformers
38
+ # triton
39
+ fsspec==2024.12.0
40
+ # via
41
+ # gradio-client
42
+ # huggingface-hub
43
+ # torch
44
+ gradio==5.9.1
45
+ # via
46
+ # mistral-7b (pyproject.toml)
47
+ # spaces
48
+ gradio-client==1.5.2
49
+ # via gradio
50
+ h11==0.14.0
51
+ # via
52
+ # httpcore
53
+ # uvicorn
54
+ hf-transfer==0.1.8
55
+ # via mistral-7b (pyproject.toml)
56
+ httpcore==1.0.7
57
+ # via httpx
58
+ httpx==0.28.1
59
+ # via
60
+ # gradio
61
+ # gradio-client
62
+ # safehttpx
63
+ # spaces
64
+ huggingface-hub==0.27.0
65
+ # via
66
+ # accelerate
67
+ # gradio
68
+ # gradio-client
69
+ # tokenizers
70
+ # transformers
71
+ idna==3.10
72
+ # via
73
+ # anyio
74
+ # httpx
75
+ # requests
76
+ jinja2==3.1.5
77
+ # via
78
+ # gradio
79
+ # torch
80
+ markdown-it-py==3.0.0
81
+ # via rich
82
+ markupsafe==2.1.5
83
+ # via
84
+ # gradio
85
+ # jinja2
86
+ mdurl==0.1.2
87
+ # via markdown-it-py
88
+ mpmath==1.3.0
89
+ # via sympy
90
+ networkx==3.4.2
91
+ # via torch
92
+ numpy==2.2.1
93
+ # via
94
+ # accelerate
95
+ # bitsandbytes
96
+ # gradio
97
+ # pandas
98
+ # transformers
99
+ nvidia-cublas-cu12==12.1.3.1
100
+ # via
101
+ # nvidia-cudnn-cu12
102
+ # nvidia-cusolver-cu12
103
+ # torch
104
+ nvidia-cuda-cupti-cu12==12.1.105
105
+ # via torch
106
+ nvidia-cuda-nvrtc-cu12==12.1.105
107
+ # via torch
108
+ nvidia-cuda-runtime-cu12==12.1.105
109
+ # via torch
110
+ nvidia-cudnn-cu12==9.1.0.70
111
+ # via torch
112
+ nvidia-cufft-cu12==11.0.2.54
113
+ # via torch
114
+ nvidia-curand-cu12==10.3.2.106
115
+ # via torch
116
+ nvidia-cusolver-cu12==11.4.5.107
117
+ # via torch
118
+ nvidia-cusparse-cu12==12.1.0.106
119
+ # via
120
+ # nvidia-cusolver-cu12
121
+ # torch
122
+ nvidia-nccl-cu12==2.20.5
123
+ # via torch
124
+ nvidia-nvjitlink-cu12==12.6.85
125
+ # via
126
+ # nvidia-cusolver-cu12
127
+ # nvidia-cusparse-cu12
128
+ nvidia-nvtx-cu12==12.1.105
129
+ # via torch
130
+ orjson==3.10.13
131
+ # via gradio
132
+ packaging==24.2
133
+ # via
134
+ # accelerate
135
+ # gradio
136
+ # gradio-client
137
+ # huggingface-hub
138
+ # spaces
139
+ # transformers
140
+ pandas==2.2.3
141
+ # via gradio
142
+ pillow==11.0.0
143
+ # via gradio
144
+ protobuf==5.29.2
145
+ # via mistral-7b (pyproject.toml)
146
+ psutil==5.9.8
147
+ # via
148
+ # accelerate
149
+ # spaces
150
+ pydantic==2.10.4
151
+ # via
152
+ # fastapi
153
+ # gradio
154
+ # spaces
155
+ pydantic-core==2.27.2
156
+ # via pydantic
157
+ pydub==0.25.1
158
+ # via gradio
159
+ pygments==2.18.0
160
+ # via rich
161
+ python-dateutil==2.9.0.post0
162
+ # via pandas
163
+ python-multipart==0.0.20
164
+ # via gradio
165
+ pytz==2024.2
166
+ # via pandas
167
+ pyyaml==6.0.2
168
+ # via
169
+ # accelerate
170
+ # gradio
171
+ # huggingface-hub
172
+ # transformers
173
+ regex==2024.11.6
174
+ # via transformers
175
+ requests==2.32.3
176
+ # via
177
+ # huggingface-hub
178
+ # spaces
179
+ # transformers
180
+ rich==13.9.4
181
+ # via typer
182
+ ruff==0.8.4
183
+ # via gradio
184
+ safehttpx==0.1.6
185
+ # via gradio
186
+ safetensors==0.4.5
187
+ # via
188
+ # accelerate
189
+ # transformers
190
+ semantic-version==2.10.0
191
+ # via gradio
192
  sentencepiece==0.2.0
193
+ # via mistral-7b (pyproject.toml)
194
+ shellingham==1.5.4
195
+ # via typer
196
+ six==1.17.0
197
+ # via python-dateutil
198
+ sniffio==1.3.1
199
+ # via anyio
200
+ spaces==0.31.1
201
+ # via mistral-7b (pyproject.toml)
202
+ starlette==0.41.3
203
+ # via
204
+ # fastapi
205
+ # gradio
206
+ sympy==1.13.3
207
+ # via torch
208
+ tokenizers==0.21.0
209
+ # via transformers
210
+ tomlkit==0.13.2
211
+ # via gradio
212
+ torch==2.4.0
213
+ # via
214
+ # mistral-7b (pyproject.toml)
215
+ # accelerate
216
+ # bitsandbytes
217
+ tqdm==4.67.1
218
+ # via
219
+ # huggingface-hub
220
+ # transformers
221
+ transformers==4.47.1
222
+ # via mistral-7b (pyproject.toml)
223
+ triton==3.0.0
224
+ # via torch
225
+ typer==0.15.1
226
+ # via gradio
227
+ typing-extensions==4.12.2
228
+ # via
229
+ # anyio
230
+ # bitsandbytes
231
+ # fastapi
232
+ # gradio
233
+ # gradio-client
234
+ # huggingface-hub
235
+ # pydantic
236
+ # pydantic-core
237
+ # rich
238
+ # spaces
239
+ # torch
240
+ # typer
241
+ # uvicorn
242
+ tzdata==2024.2
243
+ # via pandas
244
+ urllib3==2.3.0
245
+ # via requests
246
+ uvicorn==0.34.0
247
+ # via gradio
248
+ websockets==14.1
249
+ # via gradio-client
style.css CHANGED
@@ -9,9 +9,3 @@ h1 {
9
  background: #1565c0;
10
  border-radius: 100vh;
11
  }
12
-
13
- .contain {
14
- max-width: 900px;
15
- margin: auto;
16
- padding-top: 1.5rem;
17
- }
 
9
  background: #1565c0;
10
  border-radius: 100vh;
11
  }
 
 
 
 
 
 
uv.lock ADDED
The diff for this file is too large to render. See raw diff