Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 2 additions & 4 deletions tools/README.md
Original file line number Diff line number Diff line change
@@ -1,11 +1,9 @@
This needs the llama.cpp version of gguf-py to work at the moment, not the pip one as that one does not have the python quantization code yet.

# Quantizing Models
```
pip install gguf
git clone https://github.com/ggerganov/llama.cpp
pip install llama.cpp/gguf-py
```


To convert your initial source model to FP16 (or BF16), run the following command:
```
python convert.py --src E:\models\unet\flux1-dev.safetensors
Expand Down