-
Notifications
You must be signed in to change notification settings - Fork 71
Open
Description
Hello, I was trying to train my own LLM on the encodec tokenizer and I wanted a bit of help. The LLM does not seem to learn the tokens and a drop-in replacement on SEED tokenizer works fine. The shape of the codes is [4, 250]. How do I format this in a sequence so it has causal dependency. Currently I have been doing it like the paper states. 4 codes per frame, and frame by frame. Is there something else I should look out for? In the outputs, it just repeats the 4 codes over and over again and the generated audio is just no sound.
Metadata
Metadata
Assignees
Labels
No labels