Datasets:
ArXiv:
License:
| license: apache-2.0 | |
| # GUI Grounding Pre-training Data for OS-ATLAS | |
| This document describes the acquisition of the pre-training data used by OS-ATLAS. | |
| **Notes:** In GUI grounding data, the position of the target element is recorded in the `bbox` key, represented by `[left, top, right, bottom]`. | |
| Each value is a [0, 1] decimal number indicating the ratio of the corresponding position to the width or height of the image. | |
| The data stored in this dataset consists of raw data containing **only** element grounding information. When training a model, you need to use the corresponding prompts to process these data. | |
| The data we released is divided into three domains: mobile, desktop and web. | |
| All annotation data is stored in JSON format and each sample contains: | |
| * `img_filename`: the interface screenshot file | |
| * `instruction`: human instruction or referring expression extracted from ally tree or html | |
| * `bbox`: the bounding box of the target element corresponding to instruction | |
| Some data also contains a `data_type`, which records the type of an element in its structured information, if it can be obtained. | |
| *** | |
| ### Mobile data | |
| This part of data is stored under the *mobile_domain* directory. Our mobile grounding data consists of four parts. | |
| #### AMEX | |
| Android Multi-annotation EXpo (AMEX) is a comprehensive, large-scale dataset designed for generalist mobile GUI-control agents [1]. | |
| The annotation data is stored in | |
| -`amex_raw.json` | |
| Due to the single file size limitation of Hugging Face datasets, we stored the Amex images in *zip* format and split them into several sub-files. | |
| - `amex_images_part_aa` | |
| - `amex_images_part_ab` | |
| - `amex_images_part_ac` | |
| You need to first merge these split files back into the original file and then extract the contents. | |
| ``` | |
| cat amex_images_part_* > amex_images.zip | |
| 7z x amex_images.zip -aoa -o/path/to/extract/folder | |
| ``` | |
| #### UIBert | |
| UIBert [2] is a dataset extended from Rico dataset [3] for two tasks: similar UI component retrieval and referring expression component retrieval. | |
| The annotation data is stored in | |
| - `uibert_raw.json` | |
| The UIBert images are stored in | |
| - `UIBert.zip` | |
| #### Widget Captioning and RICOSCA | |
| Widget Captioning data are collected by [4]. | |
| RICOSCA is a dataset automatically labeled using Android VH in [5] | |
| The annotation data is stored in | |
| - `widget_captioning.json` | |
| - `ricosca.json` | |
| The rico images are stored in | |
| - `rico_imgs.zip` | |
| #### Android_world_data | |
| This part of data are sampled from a android environment for building and benchmarking autonomous computer control agents [6]. | |
| The annotation data is stored in | |
| - `aw_mobile.json` | |
| The rico images are stored in | |
| - `mobile_images.zip` | |
| *** | |
| ### Desktop data | |
| This part of data is stored under the *desktop_domain* directory. | |
| All of the desktop grounding data is collected from the real environments of personal computers running different operating systems. Each image is split into multiple sub-images to enhance data diversity. | |
| Our desktop grounding data consists of three parts: Windows, Linux and MacOS. | |
| **The image and annotation data for each operating system are stored in corresponding zip and json files.** | |
| It is worth noting that, due to the large size of the Windows image data, the split files need to be merged before extraction. | |
| ``` | |
| cat windows_image_part_* > windows_images.zip | |
| 7z x windows_images.zip -aoa -o/path/to/extract/folder | |
| ``` | |
| *** | |
| ### Web data | |
| This part of data is stored under the *web_domain* directory. | |
| Our desktop grounding data consists of two parts. | |
| #### Seeclick web data | |
| The web data from SeeClick [7] was crawled from websites provided by Common Crawl, containing more than 270k webpage screenshots and over 3 million webpage elements. | |
| The annotation data is stored in | |
| - `seeclick_web.json` | |
| The images are stored into split files and need to be merged before extraction. | |
| ``` | |
| cat seeclick_web_image_part_* > seeclick_web_images.zip | |
| 7z x seeclick_web_images.zip -aoa -o/path/to/extract/folder | |
| ``` | |
| #### Fineweb_crawled_data | |
| This part of data is crawled from web pages from the latest URLs obtained from FineWeb [8], a cleaned and deduplicated English dataset derived from Common Crawl. | |
| Since this portion of the data contains at least 1.6 million images, we have compressed them into 10 zip files, from `fineweb_3m_s11.zip` to `fineweb_3m_s52.zip`. | |
| Please extract them into the same directory. | |
| As an example, | |
| ``` | |
| 7z x fineweb_3m_s11.zip -aoa -o/same/path/to/extract/fineweb | |
| ``` | |
| The annotation data is stored in | |
| - `fineweb_3m.json` | |
| *** | |
| ### Best practices | |
| *** | |
| **The following are the open-source datasets we used as data sources. We welcome everyone to check the details and cite these sources accordingly!** | |
| [1] [AMEX: Android Multi-annotation Expo Dataset for Mobile GUI Agents](https://arxiv.org/abs/2407.17490) | |
| [2] [UIBert: Learning Generic Multimodal Representations for UI Understanding](https://arxiv.org/abs/2107.13731) | |
| [3] [Rico: A mobile app dataset for building data-driven design applications](https://dl.acm.org/doi/pdf/10.1145/3126594.3126651) | |
| [4] [Widget Captioning: Generating Natural Language Description for Mobile User Interface Elements](https://arxiv.org/pdf/2010.04295.pdf) | |
| [5] [Mapping Natural Language Instructions to Mobile UI Action Sequences](https://arxiv.org/pdf/2005.03776) | |
| [6] [ANDROIDWORLD: A Dynamic Benchmarking Environment for Autonomous Agents](https://arxiv.org/abs/2405.14573) | |
| [7] [SeeClick: Harnessing GUI Grounding for Advanced Visual GUI Agents](https://arxiv.org/abs/2401.10935) | |
| [8] [The fineweb datasets: Decanting the web for the finest text data at scale](https://arxiv.org/abs/2406.17557) |