[Model][Bugfix]: correct Aria model output (#12309)

Signed-off-by: xffxff <1247714429@qq.com>
This commit is contained in:
zhou fan
2025-01-22 19:39:19 +08:00
committed by GitHub
parent cd7b6f0857
commit 528dbcac7d
2 changed files with 54 additions and 3 deletions

View File

@ -28,9 +28,10 @@ def run_aria(question: str, modality: str):
llm = LLM(model=model_name,
max_model_len=4096,
max_num_seqs=2,
dtype="bfloat16",
disable_mm_preprocessor_cache=args.disable_mm_preprocessor_cache)
prompt = (f"<|im_start|>user\n<fim_prefix><|img|><fim_suffix>\n{question}"
prompt = (f"<|im_start|>user\n<fim_prefix><|img|><fim_suffix>{question}"
"<|im_end|>\n<|im_start|>assistant\n")
stop_token_ids = [93532, 93653, 944, 93421, 1019, 93653, 93519]