The dataset viewer is not available for this split.
Error code: FeaturesError
Exception: ValueError
Message: Not able to read records in the JSON file at hf://datasets/bobotran/spoiler-matching@f10e71b6212adbe80c82ea240c18c831d36100f0/matching/handlabeled_only/train.json.
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 240, in compute_first_rows_from_streaming_response
iterable_dataset = iterable_dataset._resolve_features()
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 2216, in _resolve_features
features = _infer_features_from_batch(self.with_format(None)._head())
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1239, in _head
return _examples_to_batch(list(self.take(n)))
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1389, in __iter__
for key, example in ex_iterable:
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1044, in __iter__
yield from islice(self.ex_iterable, self.n)
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 282, in __iter__
for key, pa_table in self.generate_tables_fn(**self.kwargs):
File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 165, in _generate_tables
raise ValueError(f"Not able to read records in the JSON file at {file}.") from None
ValueError: Not able to read records in the JSON file at hf://datasets/bobotran/spoiler-matching@f10e71b6212adbe80c82ea240c18c831d36100f0/matching/handlabeled_only/train.json.Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Episode-Specific Spoilers
This is the spoiler matching dataset as presented in Spoiler Detection as Semantic Text Matching. It consists of comments discussing episodes from various TV shows. Unlike other spoiler datasets, this dataset assigns an episode number (and show name) to each comment, enabling matching to specific episodes and very fine-grain spoiler detection. This dataset also includes an episode summary for each (show, episode) pair. For a given show, the task is to rank the summaries for each comment.
Usage
See the Spoiler Matching repository for examples on how to train a spoiler matching model on this dataset.
Annotation
522,991 comments from 13 TV shows were scraped from episode discussion threads on Reddit. Of these comments, some are actually relevant to their respective episode discussion and others are irrelevant. A subset of these comments (11,032) were hand-labeled as irrelevant or relevant. This hand-labeled dataset was used to train an irrelevant/relevant classifier which auto-labeled the remaining comments. All relevant comments were formatted into the matching dataset.
Details
The matching folder contains the spoiler matching dataset, and the filtering folder contains intermediate data from the auto-labeling step.
matching/
This folder contains the datasets for training spoiler matching models. All comments in these files were determined to be relevant, whether that was done by a human annotator or the auto-labeler. summaries.json contains the summary for each episode. summaries.json and test.json are the same across with_autolabels and handlabeled_only.
matching/with_autolabels/
The with_autolabels folder contains the main dataset. test.json and val.json consist of hand-labeled relevants while train.json contains auto-labeled relevants. To measure the performance of spoiler matching models on unseen shows, test.json was constructed such that it consists of comments from 4 TV shows which are neither present in val.json nor train.json.
matching/handlabeled_only/
The handlabeled_only folder shares the same test.json with matching/with_autolabels/, but train.json and val.json are split 80-20 from matching/with_autolabels/val.json respectively.
filtering/
This folder contains data from the auto-labeling step.
filtering/handlabeled
This folder contains the dataset used to train the autolabeler. Comments with a 1 in the first column were hand-labeled as irrelevant. Comments with a 0 in the first column were hand-labeled as relevant. The last two columns are the show name and episode number respectively, which are not used during this step.
filtering/unlabeled
The unlabeled comments were split into two chunks to make them more manageable to load into memory during inference. All comments have a -1 in the first column to represent that they are unlabeled.
- Downloads last month
- 24