Improve model card for UniTok with metadata and usage
#1
by
nielsr
HF Staff
- opened
This PR significantly enhances the model card for UniTok, a unified tokenizer for visual generation and understanding.
It includes:
- The appropriate
pipeline_tag: any-to-any, accurately reflecting its versatile multimodal capabilities for both generation and understanding. - The
library_name: transformers, indicating compatibility with the Hugging Face Transformers library for automated usage snippets. - Updated
licensetomit, as explicitly stated in the project's GitHub repository. - Links to the paper (UniTok: A Unified Tokenizer for Visual Generation and Understanding), the official project page (
https://foundationvision.github.io/UniTok/), and the GitHub repository (https://github.com/foundationvision/unitok). - Comprehensive information from the GitHub README, including abstract, news, performance metrics, model weights, installation instructions, and sample usage code for inference (both CLI and Python snippets for MLLM tasks), training, and further details on the Unified MLLM.
This update will greatly improve the discoverability, clarity, and usability of the model on the Hugging Face Hub.
machuofan
changed pull request status to
merged