https://huggingface.co/ibm-granite/granite-docling-258M
#1392
by
vonjack
- opened
It's queued! :D
Not sure if supported but let's hope for the best.
It will start getting processed as soon grok-2 imatrix computation completes.
You can check for progress at http://hf.tst.eu/status.html or regularly check the model
summary page at https://hf.tst.eu/model#granite-docling-258M-GGUF for quants to appear.
It is unfortinately not yet supported by llama.cpp and someone already asked for them to support it under: https://huggingface.co/ibm-granite/granite-docling-258M/discussions/16
When trying anyways llama.cpp doesn't recognize the pre-tokenizer but I would expect many other issues after that:
**************************************************************************************
** WARNING: The BPE pre-tokenizer was not recognized!
** There are 2 possible reasons for this:
** - the model has not been added to convert_hf_to_gguf_update.py yet
** - the pre-tokenization config has changed upstream
** Check your model files and convert_hf_to_gguf_update.py and update them accordingly.
** ref: https://github.com/ggml-org/llama.cpp/pull/6920
**
** chkhsh: 53e325976a6e142379c19b09afcae354f2f496f147afa8f9e189a33fe4e3024e
*************************************************************************************