[Doc] Add note to gte-Qwen2 models (#11808)

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
Cyrus Leung 2025-01-07 21:50:58 +08:00 committed by GitHub
parent d9fa1c05ad
commit c0efe92d8b
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -430,6 +430,9 @@ You can set `--hf-overrides '{"is_causal": false}'` to change the attention mask
On the other hand, its 1.5B variant (`Alibaba-NLP/gte-Qwen2-1.5B-instruct`) uses causal attention
despite being described otherwise on its model card.
Regardless of the variant, you need to enable `--trust-remote-code` for the correct tokenizer to be
loaded. See [relevant issue on HF Transformers](https://github.com/huggingface/transformers/issues/34882).
```
If your model is not in the above list, we will try to automatically convert the model using