Post
12
π₯ Distilling Tiny Embeddings. We're happy to build on the BERT Hash Series of models with this new set of fixed dimensional tiny embeddings models.
Ranging from 244K parameters to 970K and 50 dimensions to 128 dimensions these tiny models pack quite a punch.
Use cases include on-device semantic search, similarity comparisons, LLM chunking and Retrieval Augmented Generation (RAG). The advantage is that data never needs to leave the device while still having solid performance.
https://huggingface.co/blog/NeuML/bert-hash-embeddings
Ranging from 244K parameters to 970K and 50 dimensions to 128 dimensions these tiny models pack quite a punch.
Use cases include on-device semantic search, similarity comparisons, LLM chunking and Retrieval Augmented Generation (RAG). The advantage is that data never needs to leave the device while still having solid performance.
https://huggingface.co/blog/NeuML/bert-hash-embeddings