Dataset Preview
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed because of a cast error
Error code: DatasetGenerationCastError
Exception: DatasetGenerationCastError
Message: An error occurred while generating the dataset
All the data files must have the same columns, but at some point there are 5 new columns ({'value', 'sample', 'trial_type', 'duration', 'onset'}) and 9 missing columns ({'type', 'name', 'high_cutoff', 'sampling_frequency', 'low_cutoff', 'units', 'status', 'description', 'status_description'}).
This happened while the csv dataset builder was generating data using
hf://datasets/braindecode/example_dataset-eegwindows/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_events.tsv (at revision b1a5aa99a6ba22f07fdef6e6fae74c837c05b728), [/tmp/hf-datasets-cache/medium/datasets/51117127319781-config-parquet-and-info-braindecode-example_datas-4534104b/hub/datasets--braindecode--example_dataset-eegwindows/snapshots/b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_channels.tsv (origin=hf://datasets/braindecode/example_dataset-eegwindows@b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_channels.tsv), /tmp/hf-datasets-cache/medium/datasets/51117127319781-config-parquet-and-info-braindecode-example_datas-4534104b/hub/datasets--braindecode--example_dataset-eegwindows/snapshots/b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_events.tsv (origin=hf://datasets/braindecode/example_dataset-eegwindows@b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_events.tsv)]
Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1887, in _prepare_split_single
writer.write_table(table)
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 675, in write_table
pa_table = table_cast(pa_table, self._schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2272, in table_cast
return cast_table_to_schema(table, schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2218, in cast_table_to_schema
raise CastError(
datasets.table.CastError: Couldn't cast
onset: double
duration: double
trial_type: int64
sample: int64
value: int64
-- schema metadata --
pandas: '{"index_columns": [{"kind": "range", "name": null, "start": 0, "' + 821
to
{'name': Value('string'), 'type': Value('string'), 'units': Value('string'), 'low_cutoff': Value('float64'), 'high_cutoff': Value('float64'), 'description': Value('string'), 'sampling_frequency': Value('float64'), 'status': Value('string'), 'status_description': Value('float64')}
because column names don't match
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1347, in compute_config_parquet_and_info_response
parquet_operations = convert_to_parquet(builder)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 980, in convert_to_parquet
builder.download_and_prepare(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 884, in download_and_prepare
self._download_and_prepare(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 947, in _download_and_prepare
self._prepare_split(split_generator, **prepare_split_kwargs)
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1736, in _prepare_split
for job_id, done, content in self._prepare_split_single(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1889, in _prepare_split_single
raise DatasetGenerationCastError.from_cast_error(
datasets.exceptions.DatasetGenerationCastError: An error occurred while generating the dataset
All the data files must have the same columns, but at some point there are 5 new columns ({'value', 'sample', 'trial_type', 'duration', 'onset'}) and 9 missing columns ({'type', 'name', 'high_cutoff', 'sampling_frequency', 'low_cutoff', 'units', 'status', 'description', 'status_description'}).
This happened while the csv dataset builder was generating data using
hf://datasets/braindecode/example_dataset-eegwindows/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_events.tsv (at revision b1a5aa99a6ba22f07fdef6e6fae74c837c05b728), [/tmp/hf-datasets-cache/medium/datasets/51117127319781-config-parquet-and-info-braindecode-example_datas-4534104b/hub/datasets--braindecode--example_dataset-eegwindows/snapshots/b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_channels.tsv (origin=hf://datasets/braindecode/example_dataset-eegwindows@b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_channels.tsv), /tmp/hf-datasets-cache/medium/datasets/51117127319781-config-parquet-and-info-braindecode-example_datas-4534104b/hub/datasets--braindecode--example_dataset-eegwindows/snapshots/b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_events.tsv (origin=hf://datasets/braindecode/example_dataset-eegwindows@b1a5aa99a6ba22f07fdef6e6fae74c837c05b728/sourcedata/braindecode/sub-1/ses-0train/eeg/sub-1_ses-0train_task-task_run-0_desc-preproc_events.tsv)]
Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
name string | type string | units string | low_cutoff float64 | high_cutoff float64 | description string | sampling_frequency float64 | status string | status_description null |
|---|---|---|---|---|---|---|---|---|
Fz | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
FC3 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
FC1 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
FCz | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
FC2 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
FC4 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
C5 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
C3 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
C1 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
Cz | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
C2 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
C4 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
C6 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
CP3 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
CP1 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
CPz | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
CP2 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
CP4 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
P1 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
Pz | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
P2 | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
POz | EEG | V | 0 | 125 | ElectroEncephaloGram | 250 | good | null |
EOG1 | EOG | V | 0 | 125 | ElectroOculoGram | 250 | good | null |
EOG2 | EOG | V | 0 | 125 | ElectroOculoGram | 250 | good | null |
EOG3 | EOG | V | 0 | 125 | ElectroOculoGram | 250 | good | null |
stim | TRIG | V | 0 | 125 | Trigger | 250 | good | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
null | null | null | null | null | null | null | null | null |
EEG Dataset
This dataset was created using braindecode, a deep learning library for EEG/MEG/ECoG signals.
Dataset Information
| Property | Value |
|---|---|
| Recordings | 1 |
| Type | Windowed (from Raw object) |
| Channels | 26 |
| Sampling frequency | 250 Hz |
| Total duration | 0:06:26 |
| Windows/samples | 48 |
| Size | 19.22 MB |
| Format | zarr |
Quick Start
from braindecode.datasets import BaseConcatDataset
# Load from Hugging Face Hub
dataset = BaseConcatDataset.pull_from_hub("username/dataset-name")
# Access a sample
X, y, metainfo = dataset[0]
# X: EEG data [n_channels, n_times]
# y: target label
# metainfo: window indices
Training with PyTorch
from torch.utils.data import DataLoader
loader = DataLoader(dataset, batch_size=32, shuffle=True, num_workers=4)
for X, y, metainfo in loader:
# X: [batch_size, n_channels, n_times]
# y: [batch_size]
pass # Your training code
BIDS-inspired Structure
This dataset uses a BIDS-inspired organization. Metadata files follow BIDS conventions, while data is stored in Zarr format for efficient deep learning.
BIDS-style metadata:
dataset_description.json- Dataset informationparticipants.tsv- Subject metadata*_events.tsv- Trial/window events*_channels.tsv- Channel information*_eeg.json- Recording parameters
Data storage:
dataset.zarr/- Zarr format (optimized for random access)
sourcedata/braindecode/
βββ dataset_description.json
βββ participants.tsv
βββ dataset.zarr/
βββ sub-<label>/
βββ eeg/
βββ *_events.tsv
βββ *_channels.tsv
βββ *_eeg.json
Accessing Metadata
# Participants info
if hasattr(dataset, "participants"):
print(dataset.participants)
# Events for a recording
if hasattr(dataset.datasets[0], "bids_events"):
print(dataset.datasets[0].bids_events)
# Channel info
if hasattr(dataset.datasets[0], "bids_channels"):
print(dataset.datasets[0].bids_channels)
Created with braindecode
- Downloads last month
- 77