Open the teaching deck, worksheet, and editable slide source.
Interactive Lab
Practice in short loops: checkpoint quiz, microtask decision, and competency progress tracking.
Checkpoint Quiz
Microtask Decision
Choose the action that best improves scientific reliability.
Progress Tracker
State is saved locally in your browser for this module.
0% complete
Annotation Challenge
Click the hotspot with the strongest evidence for the requested feature.
Selected hotspot: none
Capability target
Design and critique an ML analysis pipeline for connectomics that includes feature rationale, evaluation plan, leakage controls, and interpretation limits.
Why this module matters
ML can accelerate connectomics analysis, but naive workflows produce misleading biological claims. This module emphasizes model validity, not just model performance.
Concept set
1) Feature engineering defines the hypothesis space
Technical: feature choices encode assumptions about what variation is biologically meaningful.
Plain language: your model can only learn what your features allow.
Misconception guardrail: adding more features always improves science.
2) Evaluation must match biological use
Technical: metrics should align with downstream decisions (for example, class-specific recall for rare but critical cell types).
Plain language: high overall accuracy can still fail where it matters most.
Misconception guardrail: one summary metric is enough.
3) Leakage and shift are endemic in connectomics
Technical: spatial adjacency, reconstruction provenance, and shared preprocessing can leak signal across train/test splits.
Plain language: your model may be “cheating” without obvious signs.
Misconception guardrail: random split always gives valid generalization estimates.
Hidden curriculum scaffold
Unspoken ML norms trainees need explicitly:
justify split strategy before training.
report failure cases with examples, not only aggregate metrics.
include model-card style limitations and intended use.
Mentoring supports:
provide leakage checklist template.
require one “where model fails” figure.
review scientific usefulness, not just benchmark score.
Core workflow: connectomics ML protocol
Define task and biological decision context.
Construct feature set with rationale and preprocessing log.
Choose split strategy that blocks leakage pathways.
Train baseline + candidate models and compare error profiles.
Report metrics, limitations, and deployment constraints.
60-minute tutorial run-of-show
**00:00-08:00
Task framing and leakage examples**
**08:00-20:00
Feature rationale workshop**
**20:00-34:00
Split strategy and baseline modeling**
**34:00-46:00
Error analysis and biologically relevant metrics**
**46:00-56:00
Model-card limitation writing**
**56:00-60:00
Competency checkpoint**
Studio activity: leakage-resistant ML mini-pipeline
Scenario: You need to classify neurite fragments into coarse categories for downstream proofreading prioritization.
Tasks
Propose feature set and leakage-safe split design.
Train one baseline and one improved model (or pseudocode plan).
Report two standard metrics and one biologically targeted metric.
Draft a model limitation statement with non-supported use cases.
Expected outputs
Feature + split design sheet.
Metric table with interpretation notes.
Limitation statement.
Assessment rubric
Minimum pass
Feature and split decisions are justified.
Metrics include at least one biologically targeted criterion.
Limitation statement is specific and actionable.
Strong performance
Identifies and mitigates likely leakage channels.
Uses error analysis to propose next data improvements.
Distinguishes exploratory model from deployment-ready model.
Common failure modes
Leakage-prone random splits for spatially correlated data.
Overfocus on aggregate accuracy.
Claims of biological insight unsupported by model diagnostics.