[Bugfix] Fix guidance backend for Qwen models (#16210)
Signed-off-by: Benjamin Chislett <benjamin.chislett@centml.ai>
This commit is contained in:
parent
ed636d99ca
commit
3147586ebd
@ -46,7 +46,8 @@ class GuidanceBackend(StructuredOutputBackend):
|
|||||||
in vllm_config.decoding_config.guided_decoding_backend)
|
in vllm_config.decoding_config.guided_decoding_backend)
|
||||||
|
|
||||||
tokenizer = tokenizer_group.get_lora_tokenizer(None)
|
tokenizer = tokenizer_group.get_lora_tokenizer(None)
|
||||||
self.ll_tokenizer = llguidance_hf.from_tokenizer(tokenizer, None)
|
self.ll_tokenizer = llguidance_hf.from_tokenizer(
|
||||||
|
tokenizer, self.vocab_size)
|
||||||
|
|
||||||
def compile_grammar(self, request_type: StructuredOutputOptions,
|
def compile_grammar(self, request_type: StructuredOutputOptions,
|
||||||
grammar_spec: str) -> StructuredOutputGrammar:
|
grammar_spec: str) -> StructuredOutputGrammar:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user