license: other
pretty_name: ConnectomeBench2
tags:
- connectomics
- proofreading
- 3d
- electron-microscopy
- mesh
size_categories:
- 100K<n<1M
configs:
- config_name: default
data_files:
- split: train
path: train/train-*.parquet
- split: validation
path: val/val-*.parquet
- split: test
path: test/test-*.parquet
ConnectomeBench2
ConnectomeBench2 is a unified benchmark for automated proofreading of connectomic neural-segmentation data. 401,170 samples across 4 species (mouse, fly, human, zebrafish) and 5 sample types (real merge edits, real split edits, synthetic adjacent / junction / synapse controls), with the associated mesh geometry and electron-microscopy (EM) renderings.
Downstream trainers should treat this dataset as the single source of truth for sample identity, labels, train/validation/test split, and which task(s) a row is valid for.
Context: Connectomic Proofreading
Connectomics scans and automatically segments neurons to create large-scale brain maps at cellular resolution. Two types of segmentation errors can occur in this process, which need to be corrected (= proofreading):
- False Splits — corrected via merge corrections
- False Merges — corrected via split corrections
Merge corrections (of false splits) are applied to multiple segments that need to be correctly merged together. Split corrections (of false merges) are applied to single segments that need to be correctly split apart.
For this reason, this dataset contains renderings of both single-segment (pre-split or post-merge) and dual-segment (post-split or pre-merge) mesh geometry, where possible. EM data is provided in dual format only — segmentation on imaging level is contiguous, so the single-version can be derived from the union of the dual.
Renderings (geometry and EM imaging data)
(top: synapse merge-pair — both masks populated; bottom: junction control — single-mask only, mask B / seg B empty)
Geometry files (the geometry and geometry_single columns) are compressed .npz payloads that decode to (3, 7, 224, 224) float16 arrays — three 2D views (front, side, top) × seven channels:
| ch | content |
|---|---|
| 0 | silhouette |
| 1 | depth |
| 2 | normal_x |
| 3 | normal_y |
| 4 | normal_z |
| 5 | mask A |
| 6 | mask B (empty in single-segment renders) |
Note that single and dual segment renders differ not only in mask channels, but also subtly differ in all other channels, due to slight differences in mesh geometry from merging/splitting.
Free split-mask labels. For split_edit rows, the dual-segment render (post-split) provides ground-truth split-mask labels (Mask A / Mask B channels) for the corresponding single-segment render (pre-split) — split-mask-generation tasks get pixel-level supervision without extra labeling.
EM coverage. EM views are not present on every sample. Coverage by sample_type (full dataset):
| sample_type | rows | has_em |
|---|---|---|
| adjacent_control | 121,333 | 100% |
| junction_control | 38,272 | 100% |
| synapse_control | 18,182 | 100% |
| merge_edit | 146,461 | 38% |
| split_edit | 77,213 | 23% |
| total | 401,170 | 63% (37% null) |
real human edits (merge_edit, split_edit) only got EM rendered on a stratified subset; synthetic controls all have EM. Filter by has_em if your task requires it.
EM imaging files (em_xy / em_xz / em_yz / em_best columns) are PNG-encoded 3-channel slices:
| ch | content |
|---|---|
| 0 | raw EM intensity |
| 1 | segment A mask |
| 2 | segment B mask |
Four imaging views per sample: three cardinal slices (xy, xz, yz) + a best slice at an oblique angle that maximizes the visible area of both segments (sum of their logs).
For single-segment tasks, segment A and B should be merged (and B zeroed). The best view may leak some dual-label information (it takes both labels into account); we advise against testing single-segment tasks on em_best.
Loading
from datasets import load_dataset
ds = load_dataset("jeffbbrown2/connectomebench2-smoke", split="train")
sample = ds[0]
# sample["em_xy"] is a PIL Image (HF auto-decodes)
# sample["geometry"] is bytes — decode with:
import io, numpy as np
geom = np.load(io.BytesIO(sample["geometry"]))["arr_0"] # shape (3, 7, 224, 224) float16
Or with raw pyarrow:
import pyarrow.parquet as pq
import numpy as np, io
df = pq.read_table("train/train-00000.parquet").to_pandas()
geom = np.load(io.BytesIO(df.iloc[0]["geometry"]))["arr_0"]
The metadata/{train,val,test}.parquet sidecars contain identifier/label/modality columns only (no image bytes) — useful for fast filtering or inspection.
Columns
Identifiers
combined_sample_hash— primary key (md5 hex 32-char off"{source_archive}|{source_archive_sample_hash}"); guaranteed unique across the dataset.source_archive_sample_hash— legacy 12-char hex hash from upstream; kept for traceability, not unique alone.source_archive— name of the originating render archive (e.g.edits_and_adj_controls_fly,junction_controls_mouse,synapse_controls_fly). 10 distinct values (5 archives × species).
Sample identity
sample_type: str— single source of truth for what kind of sample this row is. Five values:merge_edit— positive merge-correction editsplit_edit— positive split-correction editadjacent_control— synthetic negative for merge-correction (segments adjacent to genuine correction)junction_control— putative junction in proofread neuron (negative merge-error-id sample)synapse_control— synapse pair across neurons (negative merge-correction)
same_neuron: bool— derived from sample_type:Trueformerge_edit,junction_controlFalseforsplit_edit,adjacent_control,synapse_control
species: str—fly/mouse/human/zebrafish.
Image content
geometry— bytes; compressed npz (key"arr_0") decoding to(3, 7, 224, 224) float16. Null when the sample has no dual-segment render.geometry_single— same shape/dtype, single-segment version. Null when not present.em_xy/em_xz/em_yz/em_best— PIL Images (3-channel PNG,(224, 224, 3) uint8). Null when the row has no EM views.has_single_mask: bool— convenience flag.has_dual_mask: bool— convenience flag.has_em: bool— true if anyem_*column is non-null.present_slots: list[str]— modality tags actually present (e.g.["geometry", "geometry_single", "em_xy", "em_xz", "em_yz", "em_best"]).
Task routing & labels
task_routing: list[str]— which downstream task(s) this row can serve as training data for:false_split_correction— merge-correction task; fires whensample_type ∈ {merge_edit, synapse_control, adjacent_control}ANDhas_dual_mask.false_merge_identification— merge-error binary classification; fires whensample_type ∈ {split_edit, junction_control}ANDhas_single_mask.split_mask_generation— pixel-level split prediction; fires whensample_type == split_editANDhas_single_mask.
false_split_correction_label: bool=same_neuron. Populated for all rows; trainers filter bytask_routing.false_merge_identification_label: bool=not same_neuron. Populated for all rows; trainers filter bytask_routing.
Usage note. Downstream training scripts must load the appropriate geometry render per task:
- Merge Correction of false splits should use dual-segment renders
- Split Correction of false merges should use single-segment renders
- Furthermore, fuse A/B channels of EM images and discard
em_best(it sees both labels at oblique angle and can leak ground truth)
- Furthermore, fuse A/B channels of EM images and discard
Otherwise, ground-truth task or label information may leak to the model and bias performance.
Train/val/test split
split: str—train/validation/test. ~80/10/10 split assigned by spatial location of the proofreading sample (interface_point_nm), matched via cube splits (50µm cubes tiling the volume and randomly split).
Other
metadata: str— JSON-stringified original metadata struct. Parse withjson.loads. Useful keys:operation_id,source_operation_id,strategy,image_types,interface_point_nm,before_root_ids,after_root_ids, …
Counts
- 401,170 rows total · ~80/11/9 train (319,727) / validation (43,517) / test (37,926)
- 251,499 rows with EM views; all 401,170 have geometry
- ~2.2M model-level samples (EM × 4 views + geom × 3 views), or ~2.8M counting dual + single geom separately
- 506 parquet shards (~240 MB each)
Layout
README.md
shards.csv metadata across shards (path, sha256, n_samples, size)
train/train-*.parquet WebDataset-style parquet shards with image bytes
val/val-*.parquet
test/test-*.parquet
metadata/ sidecar parquets with identifiers + labels (no bytes)
train.parquet
val.parquet
test.parquet
demo.parquet stratified mini-shard (one-line preview)
figures/
channel_decomposition.png
Sources & License
Derived from the following upstream connectomic proofreading datasets:
- MICrONS (mouse cortex)
- FlyWire (Drosophila brain)
- H01 (human cortex)
- Zebrafish larval connectome
License = other; users must comply with upstream licenses (which may differ across species/sources). Final outbound license will be set after upstream license review.
Citation
If you use ConnectomeBench2, please cite:
Brown, J., Farkas, T., Razgar, G., Boyden, E. S.
ConnectomeBench2: A unified benchmark for automated connectomic proofreading.
(2026, in submission). Brown J. and Farkas T. contributed equally as first authors.
Please also cite the upstream connectome sources used by this dataset:
- MICrONS (mouse cortex): https://www.microns-explorer.org/cortical-mm3
- FlyWire (Drosophila): https://flywire.ai/
- H01 (human cortex): https://h01-release.storage.googleapis.com/landing.html
- Zebrafish (fish1): https://fish1-release.storage.googleapis.com/index.html
