https://huggingface.co/ibm-granite/granite-docling-258M

#1392
by vonjack - opened

It's queued! :D
Not sure if supported but let's hope for the best.
It will start getting processed as soon grok-2 imatrix computation completes.

You can check for progress at http://hf.tst.eu/status.html or regularly check the model
summary page at https://hf.tst.eu/model#granite-docling-258M-GGUF for quants to appear.

It is unfortinately not yet supported by llama.cpp and someone already asked for them to support it under: https://huggingface.co/ibm-granite/granite-docling-258M/discussions/16
When trying anyways llama.cpp doesn't recognize the pre-tokenizer but I would expect many other issues after that:

**************************************************************************************
** WARNING: The BPE pre-tokenizer was not recognized!
**          There are 2 possible reasons for this:
**          - the model has not been added to convert_hf_to_gguf_update.py yet
**          - the pre-tokenization config has changed upstream
**          Check your model files and convert_hf_to_gguf_update.py and update them accordingly.
** ref:     https://github.com/ggml-org/llama.cpp/pull/6920
**
** chkhsh:  53e325976a6e142379c19b09afcae354f2f496f147afa8f9e189a33fe4e3024e
*************************************************************************************

Sign up or log in to comment