Custom GGUF quants of Metaβs Llama-3.2-Instruct's finetunes, where the Output Tensors are quantized to Q8_0 or F32 and the Embeddings are kept @F32
Joseph
Joseph717171
AI & ML interests
None yet
Recent Activity
liked
a model
5 days ago
manifestai/Brumby-14B-Base
liked
a model
5 days ago
nvidia/omnivinci
liked
a model
9 days ago
nvidia/Llama-3.1-8B-Instruct-FP8