generation_config.json adds a mapping with the special token '<|im_end|>' to solve the problem of non-stop generation when <|im_end|> is encountered.
2
#17 opened about 5 hours ago
by
zjyhf
The tokenizer adds a special token '<|im_end|>' to solve the problem of non-stop generation when encountering <|im_end|>.
#16 opened about 5 hours ago
by
zjyhf
How to use in llama.cpp server
1
#15 opened about 23 hours ago
by
subbur
how to set context in multi-turn QA?
6
#14 opened 6 days ago
by
J22
Update README.md
#13 opened 7 days ago
by
freyacoltman
Try to run with dedicated endpoint 4x A100 320GB still get not enough hardware capacity
3
#11 opened 10 days ago
by
trungnx26

Colab Notebook
1
#10 opened 11 days ago
by
ChristophSchuhmann
Megatron LM training (fine-tuning) code ?
3
#9 opened 11 days ago
by
StephennFernandes

If i make context empty, it will output chinese.
6
#8 opened 11 days ago
by
Cometyang

Adding `safetensors` variant of this model
#7 opened 12 days ago
by
SFconvertbot

Adding `safetensors` variant of this model
#6 opened 13 days ago
by
SFconvertbot

Chat template
15
#5 opened 13 days ago
by
bartowski

Adding `safetensors` variant of this model
#4 opened 13 days ago
by
SFconvertbot

I got answer with the token "ologne" at the end
1
#3 opened 13 days ago
by
Stilgar