init
This commit is contained in:
32
docs/HOW_TO_CONVERT_HF_MODEL.md
Normal file
32
docs/HOW_TO_CONVERT_HF_MODEL.md
Normal file
@@ -0,0 +1,32 @@
|
||||
# How to Convert 🤗 Hugging Face Model
|
||||
|
||||
Currently, Distributed Llama supports these Hugging Face models: `llama`, `mistral`, `qwen3` and `qwen3_moe`. You can try to convert any compatible Hugging Face model and run it with Distributed Llama.
|
||||
|
||||
> [!IMPORTANT]
|
||||
> All converters are in the early stages of development. After conversion, the model may not work correctly.
|
||||
|
||||
1. Download a model, for example: [Mistral-7B-v0.3](https://huggingface.co/mistralai/Mistral-7B-v0.3/tree/main).
|
||||
2. The downloaded model should contain `config.json`, `tokenizer.json`, `tokenizer_config.json` and `tokenizer.model` and safetensor files.
|
||||
3. Run the converter of the model:
|
||||
|
||||
```sh
|
||||
cd converter
|
||||
python convert-hf.py path/to/hf/model q40 mistral-7b-0.3
|
||||
```
|
||||
|
||||
4. Run the converter of the tokenizer:
|
||||
|
||||
```sh
|
||||
python convert-tokenizer-hf.py path/to/hf/model mistral-7b-0.3
|
||||
```
|
||||
|
||||
5. That's it! Now you can run the Distributed Llama.
|
||||
|
||||
```sh
|
||||
./dllama inference \
|
||||
--prompt "Hello world" \
|
||||
--steps 64 \
|
||||
--model dllama_model_mistral-7b-0.3_q40.m \
|
||||
--tokenizer dllama_tokenizer_mistral-7b-0.3.t \
|
||||
--buffer-float-type q80
|
||||
```
|
||||
Reference in New Issue
Block a user