Update README.md
Browse files
README.md
CHANGED
|
@@ -210,7 +210,6 @@ from vllm.sampling_params import SamplingParams
|
|
| 210 |
model_name = "mistralai/Pixtral-12B-2409"
|
| 211 |
max_img_per_msg = 5
|
| 212 |
|
| 213 |
-
|
| 214 |
llm = LLM(model=model_name, tokenizer_mode="mistral", limit_mm_per_prompt={"image": max_img_per_msg}, max_model_len=32768)
|
| 215 |
|
| 216 |
def load_system_prompt(repo_id: str, filename: str) -> str:
|
|
@@ -221,7 +220,6 @@ def load_system_prompt(repo_id: str, filename: str) -> str:
|
|
| 221 |
|
| 222 |
SYSTEM_PROMPT = load_system_prompt(model_name, "vision_system_prompt.txt")
|
| 223 |
|
| 224 |
-
|
| 225 |
url = "https://huggingface.co/datasets/patrickvonplaten/random_img/resolve/main/yosemite.png"
|
| 226 |
|
| 227 |
messages = [
|
|
|
|
| 210 |
model_name = "mistralai/Pixtral-12B-2409"
|
| 211 |
max_img_per_msg = 5
|
| 212 |
|
|
|
|
| 213 |
llm = LLM(model=model_name, tokenizer_mode="mistral", limit_mm_per_prompt={"image": max_img_per_msg}, max_model_len=32768)
|
| 214 |
|
| 215 |
def load_system_prompt(repo_id: str, filename: str) -> str:
|
|
|
|
| 220 |
|
| 221 |
SYSTEM_PROMPT = load_system_prompt(model_name, "vision_system_prompt.txt")
|
| 222 |
|
|
|
|
| 223 |
url = "https://huggingface.co/datasets/patrickvonplaten/random_img/resolve/main/yosemite.png"
|
| 224 |
|
| 225 |
messages = [
|