Skip to content

Directly use low bit gguf for continuous training? #10199

Answered by BarfingLemurs
FNsi asked this question in Q&A
Discussion options

You must be logged in to vote

Here's a quantized gguf conversion script: https://github.com/PygmalionAI/aphrodite-engine/blob/main/examples/gguf_to_torch.py
Use transformers for the training.

(Finetuning/Training gguf models)
#2632
l tried Q8_0 training on Openllama 3B.

https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct-QLORA_INT4_EO8
Perhaps you are looking for quantization aware training like this one?

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@FNsi
Comment options

Answer selected by FNsi
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants