GGUF
johnbenac's picture
Upload folder using huggingface_hub
e80739d verified

llama.cpp/example/mimi

This demonstrates running Kyutai's Mimi model via GGML.

Quickstart

Convert model to GGUF (no need to download, the script will automatically download the safetensors file)

python examples/tts/convert_mimi_to_gguf.py

# output file: kyutai-mimi.gguf

# optionally, use q8_0 quantization for faster speed
python examples/tts/convert_mimi_to_gguf.py --outtype q8_0

Then compile, run it:

cmake --build build -j --target llama-mimi

./build/bin/llama-mimi kyutai-mimi.gguf codes.txt

# output: output.wav

# alternatively, use "dummy1" to get a "wah hello there" sample output file
./build/bin/llama-mimi kyutai-mimi.gguf dummy1

Example of code file (one code per line):

1263
1597
1596
1477
1540
1720
1433
118
1066
1968
1096
232
418
566
1653
2010