--- license: mit task_categories: - visual-question-answering - image-classification language: - en tags: - robotics - condition-checking - multi-modal - vision-language size_categories: - n<1K --- # Condition Checking Dataset This dataset contains condition checking conversations for robotics applications, with embedded base64 images from multiple camera viewpoints. ## Dataset Description - **Task**: Visual condition checking (True/False questions about robot states) - **Modality**: Multi-modal (text + images) - **Domain**: Robotics manipulation tasks - **Format**: Conversational format suitable for VLM training ## Dataset Structure ### Data Fields - `id`: Unique identifier for each sample - `images`: Dictionary containing base64-encoded images from multiple camera viewpoints - `conversations`: List of conversation turns (human question + assistant answer) ### Camera Viewpoints The dataset includes images from 5 camera viewpoints: - `observation_images_chest` - `observation_images_left_eye` - `observation_images_left_wrist` - `observation_images_right_eye` - `observation_images_right_wrist` Each sample contains approximately 30 images total (6 per camera). ### Sample Structure ```json { "id": "frame_index_position_part", "images": { "camera_key": ["base64_image_1", "base64_image_2", ...], ... }, "conversations": [ { "from": "human", "value": "Here are the observations... condition: (object is grasped) ..." }, { "from": "gpt", "value": "True" } ] } ``` ## Usage ```python from datasets import load_dataset # Load the dataset dataset = load_dataset("jeffshen4011/condition-checking-dataset") # Access a sample sample = dataset["train"][0] print(f"Question: {sample['conversations'][0]['value'][:100]}...") print(f"Answer: {sample['conversations'][1]['value']}") print(f"Number of camera views: {len(sample['images'])}") ``` ## Dataset Statistics - **Training samples**: 9 - **Camera viewpoints**: 5 - **Images per sample**: ~30 - **Image format**: Base64-encoded PNG - **Task type**: Binary classification (True/False) ## Applications This dataset is designed for: - Training vision-language models for robotics condition checking - Multi-modal reasoning tasks - Robot state verification - Visual question answering in manipulation contexts ## Citation If you use this dataset, please cite: ```bibtex @dataset{condition_checking_dataset, title={Condition Checking Dataset for Robotics}, author={Research Team}, year={2025}, url={https://huggingface.co/datasets/jeffshen4011/condition-checking-dataset} } ```