Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Posts
  • Docs
  • Pricing

  • Log In
  • Sign Up

nvidia
/
Llama3-ChatQA-1.5-8B

Text Generation
Transformers
Safetensors
PyTorch
English
llama
nvidia
chatqa-1.5
chatqa
llama-3
Inference Endpoints
text-generation-inference
Model card Files Files and versions Community
17
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

generation_config.json adds a mapping with the special token '<|im_end|>' to solve the problem of non-stop generation when <|im_end|> is encountered.

2
#17 opened about 5 hours ago by zjyhf

The tokenizer adds a special token '<|im_end|>' to solve the problem of non-stop generation when encountering <|im_end|>.

#16 opened about 5 hours ago by zjyhf

How to use in llama.cpp server

1
#15 opened about 23 hours ago by subbur

how to set context in multi-turn QA?

6
#14 opened 6 days ago by J22

Update README.md

#13 opened 7 days ago by freyacoltman

Try to run with dedicated endpoint 4x A100 320GB still get not enough hardware capacity

3
#11 opened 10 days ago by trungnx26

Colab Notebook

1
#10 opened 11 days ago by ChristophSchuhmann

Megatron LM training (fine-tuning) code ?

3
#9 opened 11 days ago by StephennFernandes

If i make context empty, it will output chinese.

6
#8 opened 11 days ago by Cometyang

Adding `safetensors` variant of this model

#7 opened 12 days ago by SFconvertbot

Adding `safetensors` variant of this model

#6 opened 13 days ago by SFconvertbot

Chat template

15
#5 opened 13 days ago by bartowski

Adding `safetensors` variant of this model

#4 opened 13 days ago by SFconvertbot

I got answer with the token "ologne" at the end

1
#3 opened 13 days ago by Stilgar
Company
© Hugging Face
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs