Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

sergiopaniegoย 
posted an update 2 days ago
view post
Post
1531
The Christmas holidays are here! ๐ŸŽ„
Thinking about learning something new in AI?

@huggingface offers 12 FREE courses covering all the relevant topics, for every level of experience. A great challenge for the holidays (and worth saving for later ๐Ÿ™„)

Letโ€™s explore them!

๐Ÿง  ๐—Ÿ๐—Ÿ๐—  ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: large language models with HF tools
https://huggingface.co/learn/llm-course

๐Ÿค– ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: build and deploy AI agents
https://huggingface.co/learn/agents-course

๐ŸŽจ ๐——๐—ถ๐—ณ๐—ณ๐˜‚๐˜€๐—ถ๐—ผ๐—ป ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: diffusion models with ๐Ÿค— Diffusers
https://huggingface.co/learn/diffusion-course

๐Ÿ”Š ๐—”๐˜‚๐—ฑ๐—ถ๐—ผ ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: transformers for audio tasks
https://huggingface.co/learn/audio-course

๐ŸŽฎ ๐——๐—ฒ๐—ฒ๐—ฝ ๐—ฅ๐—Ÿ ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: deep reinforcement learning
https://huggingface.co/learn/deep-rl-course

๐Ÿ‘๏ธ ๐—–๐—ผ๐—บ๐—บ๐˜‚๐—ป๐—ถ๐˜๐˜† ๐—–๐—ผ๐—บ๐—ฝ๐˜‚๐˜๐—ฒ๐—ฟ ๐—ฉ๐—ถ๐˜€๐—ถ๐—ผ๐—ป ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: modern computer vision with HF
https://huggingface.co/learn/computer-vision-course

๐Ÿฆพ ๐—ฅ๐—ผ๐—ฏ๐—ผ๐˜๐—ถ๐—ฐ๐˜€ ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ (๐—Ÿ๐—ฒ๐—ฅ๐—ผ๐—ฏ๐—ผ๐˜): learning-based robotics
https://huggingface.co/learn/robotics-course

๐Ÿงฉ ๐— ๐—–๐—ฃ ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: Model Context Protocol explained
https://huggingface.co/learn/mcp-course

๐Ÿงช ๐—” ๐—ฆ๐—บ๐—ผ๐—น ๐—–๐—ผ๐˜‚๐—ฟ๐˜€๐—ฒ: post-training AI models
https://huggingface.co/learn/a-smol-course

๐Ÿ•น๏ธ ๐— ๐—Ÿ ๐—ณ๐—ผ๐—ฟ ๐—š๐—ฎ๐—บ๐—ฒ๐˜€: AI in game development
https://huggingface.co/learn/ml-for-games-course

๐ŸงŠ ๐— ๐—Ÿ ๐—ณ๐—ผ๐—ฟ ๐Ÿฏ๐——: machine learning for 3D data
https://huggingface.co/learn/ml-for-3d-course

๐Ÿ“˜ ๐—ข๐—ฝ๐—ฒ๐—ป-๐—ฆ๐—ผ๐˜‚๐—ฟ๐—ฐ๐—ฒ ๐—”๐—œ ๐—–๐—ผ๐—ผ๐—ธ๐—ฏ๐—ผ๐—ผ๐—ธ: practical AI notebooks
https://huggingface.co/learn/cookbook

All of them can be found here: https://huggingface.co/learn
MonsterMMORPGย 
posted an update 3 days ago
view post
Post
3492
Wan 2.2 Complete Training Tutorial - Text to Image, Text to Video, Image to Video, Windows & Cloud : https://youtu.be/ocEkhAsPOs4

Wan 2.2 training is now so easy. I have done over 64 different unique Wan 2.2 trainings to prepare the very best working training configurations for you. The configurations are fully working locally with as low as 6 GB GPUs. So you will be able to train your awesome Wan 2.2 image or video generation LoRAs on your Windows computer with easiness. Moreover, I have shown how to train on cloud platforms RunPod and Massed Compute so even if you have no GPU or you want faster training, you can train on cloud for very cheap prices fully privately.

Full step by step tutorial : https://youtu.be/ocEkhAsPOs4

โฑ๏ธ Video Chapters:

0:00 Introduction to Wan 2.2 Training & Capabilities
0:56 Installing & Updating Musubi Tuner Locally
2:20 Explanation of Optimized Presets & Research Logic
4:00 Differences Between T2I, T2V, and I2V Configs
5:36 Extracting Files & Running Update Batch File
6:14 Downloading Wan 2.2 Training Models via Script
7:30 Loading Configs: Selecting GPU & VRAM Options
9:33 Using nvitop to Monitor RAM & VRAM Usage
10:28 Preparing Image Dataset & Trigger Words
11:17 Generating Dataset Config & Resolution Logic
12:55 Calculating Epochs & Checkpoint Save Frequency
13:40 Troubleshooting: Fixing Missing VAE Path Error
15:12 VRAM Cache Behavior & Training Speed Analysis
15:51 Trade-offs: Learning Rate vs Resolution vs Epochs
16:29 Installing SwarmUI & Updating ComfyUI Backend
18:13 Importing Latest Presets into SwarmUI
19:25 Downloading Inference Models via Script
20:33 Generating Images with Trained Low Noise LoRA
22:22 Upscaling Workflow for High-Fidelity Results
24:15 Increasing Base Resolution to 1280x1280
27:26 Text-to-Video Generation with Lightning LoRA
30:12 Image-to-Video Generation Workflow & Settings
31:35 Restarting Backend to Clear VRAM for Model Switching
33:45 Fixing RAM Crashes with Cache-None Argument
....
  • 3 replies
ยท
ronantakizawaย 
posted an update about 23 hours ago
Jiaqi-hkustย 
posted an update 1 day ago
view post
Post
2625
We have open-sourced Robust-R1 (AAAI 2026 Oral), a new paradigm in the field of anti-degradation and robustness enhancement for multimodal large models.

Multimodal Large Language Models struggle to maintain reliable performance under extreme real-world visual degradations, which impede their practical robustness. Existing robust MLLMs predominantly rely on implicit training/adaptation that focuses solely on visual encoder generalization, suffering from limited interpretability and isolated optimization. To overcome these limitations, we propose Robust-R1, a novel framework that explicitly models visual degradations through structured reasoning chains. Our approach integrates: (i) supervised fine-tuning for degradation-aware reasoning foundations, (ii) reward-driven alignment for accurately perceiving degradation parameters, and (iii) dynamic reasoning depth scaling adapted to degradation intensity. To facilitate this approach, we introduce a specialized 11K dataset featuring realistic degradations synthesized across four critical real-world visual processing stages, each annotated with structured chains connecting degradation parameters, perceptual influence, pristine semantic reasoning chain, and conclusion. Comprehensive evaluations demonstrate state-of-the-art robustness: Robust-R1 outperforms all general and robust baselines on the real-world degradation benchmark R-Bench, while maintaining superior anti-degradation performance under multi-intensity adversarial degradations on MMMB, MMStar, and RealWorldQA.

We have made all of our papers, codes, data, model weights and demos fully open-source:
Paper: Robust-R1: Degradation-Aware Reasoning for Robust Visual Understanding (2512.17532) (help us to upvote)
GitHub code: https://github.com/jqtangust/Robust-R1 (help us to star)
HF model: https://huggingface.co/Jiaqi-hkust/Robust-R1
HF data: Jiaqi-hkust/Robust-R1
HF Space: Jiaqi-hkust/Robust-R1

We sincerely invite everyone to give it a try.

davidmezzettiย 
posted an update 2 days ago
view post
Post
2786
๐Ÿงฌโš•๏ธ๐Ÿ”ฌ Encoding the World's Medical Knowledge into 970K! We're excited to release this new series of vector embeddings models for medical literature based on our recent BERT Hash work.

And you read it right, we're talking 970,000 parameters for a surprisingly strong performing model. Enjoy!

https://huggingface.co/blog/neuml/biomedbert-hash-nano
telcomย 
posted an update 3 days ago
view post
Post
1990
arXiv CS endorsement

It's Javad, my Google Scholar Profile:
https://scholar.google.com/citations?user=bja6GwoAAAAJ&hl=en
I would like to share my articles with you on Hugging Face, I'm asking for endorsement* in Computer Science arxiv.org.

If you would like to endorse me, please visit the following URL:
https://arxiv.org/auth/endorse?x=NVUAPL
If that URL does not work for you, please visit
http://arxiv.org/auth/endorse.php
and enter the following six-digit alphanumeric string:
Endorsement Code: NVUAPL

Thanks you in advance.
Javad Taghia

* Who is qualified to endorse?

To endorse another user to submit to the cs.AI (Artificial Intelligence) subject class, an arXiv submitter must have submitted 3 papers to any of cs.AI, cs.AR, cs.CC, cs.CE, cs.CG, cs.CL, cs.CR, cs.CV, cs.CY, cs.DB, cs.DC, cs.DL, cs.DM, cs.DS, cs.ET, cs.FL, cs.GL, cs.GR, cs.GT, cs.HC, cs.IR, cs.IT, cs.LG, cs.LO, cs.MA, cs.MM, cs.MS, cs.NA, cs.NE, cs.NI, cs.OH, cs.OS, cs.PF, cs.PL, cs.RO, cs.SC, cs.SD, cs.SE, cs.SI or cs.SY earlier than three months ago and less than five years ago.

danielhanchenย 
posted an update 1 day ago
view post
Post
1193
You can now run GLM-4.7, the new 355B parameter SOTA model on your local device (128GB RAM).โœจ

The model achieves SOTA performance on coding, agentic and chat benchmarks.

GGUF: unsloth/GLM-4.7-GGUF
Guide: https://docs.unsloth.ai/models/glm-4.7
dhruv3006ย 
posted an update 1 day ago
view post
Post
1028
OpenAPI specs are a great way to describe APIs in a clear, standard format. They provide a full overview of endpoints, methods, parameters etc. which makes working with APIs easier and more consistent.

Voiden lets you turn your OpenAPI spec into organized, ready-to-use API request files.

Just import your OpenAPI file, and you can immediately browse your endpoints, grouped by tags, and start testing without any manual setup.

The generated requests come pre-configured but fully editable, so you can customize them as you want.

If you want to get started with your existing APIs or try out new ones, this can save you quite some time.

Read the docs here : https://docs.voiden.md/docs/getting-started-section/getting-started/openapi-imports/
nicolay-rย 
posted an update 2 days ago
view post
Post
1036
Time-Effective LLM Querying in Information Retrieval Tasks

๐ŸŽค Last week at Research Colloquium in Technische Universitรคt Chemnitz, we presented a framework for time-effective data handling with prompting schemas. The video of the talk is now available ๐Ÿ‘‡๏ธ

๐ŸŽฌ๏ธ Video: https://youtu.be/pa8jGOhHViI
๐ŸŒŸ Framework (bulk-chain): https://github.com/nicolay-r/bulk-chain

๐Ÿ”‘ bulk-chain solves the following problems:
โœ… Effective handling CoT schema with big amount of prompts and parameters they are based on (batching policies)
โœ… Easy-to-apply for data-iterators (datasets handling)
MikeDoesย 
posted an update 3 days ago
view post
Post
1732
How do you protect your prompts without breaking them? You need a smart sanitizer. A new system called Prฯตฯตmpt shows how.

The first, critical step in their solution is a high-performance Named Entity Recognition (NER) model to find the sensitive data. We're proud to see that these researchers, Amrita Roy Chowdhury, David Glukhov, Divyam Anshumaan, Prasad Chalasani, Nicolas Papernot, Somesh Jha, and Mihir Bellare from the University of Michigan, University of Toronto, University of Wisconsin-Madison, University of California, San Diego - Rady School of Management and Langroid Incorporated fine-tuned their NER model on 10 high-risk categories from the AI4Privacy dataset.

This is a perfect win-win. Our open-source data helps provide the foundation for the critical detection engine, which in turn enables the community to build and test better solutions like Prฯตฯตmpt's innovative use of encryption and Differential Privacy.

๐Ÿ”— Check out their paper for a deep dive into a formally private, high-utility prompt sanitizer: https://arxiv.org/pdf/2504.05147

#OpenSource
#DataPrivacy
#LLM
#Anonymization
#AIsecurity
#HuggingFace
#Ai4Privacy
#Worldslargestopensourceprivacymaskingdataset