Cross-Cutting Patterns and Lessons Learned
Objectives
Identify common themes across the seven demonstrators
Understand why physics-informed ML generalizes better with fewer parameters
See how data representation choices affect model performance
Learn how NAIC infrastructure bridges interactive and batch computing
Understand the reproducibility pattern used across all repositories
Pattern 1: Physics-Informed ML Generalizes Better with Fewer Parameters
Three use cases embed domain physics directly into their neural architectures:
UC |
Technique |
Finding |
|---|---|---|
UC2 |
Electrochemical equations in network |
12-param student beats ~50K-param Transformer on OOD |
UC3 |
Port-Hamiltonian decomposition |
Models remain valid when forces are modified |
UC7 |
PDE-structured latent spaces |
Cross-modal alignment works across discretizations |
The pattern is consistent: embedding physics into the architecture produces models that generalize better with far fewer parameters. The pure ML models win on in-distribution validation because they have orders of magnitude more parameters to memorize training patterns, but they collapse on out-of-distribution data.
This has practical implications. Smaller models:
Are cheaper to train and deploy
Have interpretable parameters (UC2’s 12 numbers can be inspected individually)
Can be validated against known physics
Degrade gracefully outside the training distribution
Pattern 2: Data Representation Matters as Much as Model Choice
Two use cases showed that how you represent the data matters as much as which model you use:
UC |
Representation Choice |
Impact |
|---|---|---|
UC5 |
Graphs instead of flat time series |
Captured spatial-temporal vessel patterns |
UC7 |
Multi-modal autoencoders with alignment |
Bridged parameter and solution spaces |
In UC5, all three GNN architectures (GCN, GraphSAGE, GAT) achieved over 92% accuracy — the performance gap between architectures was smaller than the gap between graph and non-graph representations. The structural decision dominated the model decision.
Pattern 3: NAIC Infrastructure Bridges Interactive and Batch Computing
UC1 demonstrates the clearest example:
Workflow |
Platform |
Use Case |
|---|---|---|
Interactive exploration |
NAIC Orchestrator VM |
Single-target analysis, parameter tuning |
Large-scale sweeps |
NAIC Orchestrator VM (CLI) |
42,613 experiments across all combinations |
UC2 and UC6 show how Orchestrator VMs with GPU and multi-core support handle training workloads that would be impractical on a researcher’s laptop:
UC2: GPU training for teacher and student models
UC6: 3–4x speedup from multi-core parallelized CMA-ES
The shared pattern of SSH-tunneled Jupyter access, tmux-based background training, and one-command setup scripts cuts the operational overhead for domain scientists.
Pattern 4: Reproducibility Through Self-Contained Repositories
Every completed demonstrator ships as a Git repository containing:
Component |
Purpose |
|---|---|
Data (or download scripts) |
Training and evaluation data |
Environment specifications |
|
Training code |
Scripts and utility modules |
Evaluation scripts |
Metrics computation and visualization |
Jupyter notebook |
End-to-end interactive demonstration |
Test suite |
Automated validation (CI/CD) |
All demonstrators additionally include Sphinx tutorials published on GitHub Pages, and six (all except UC4) include AGENT.md files for AI coding assistant integration.
A new researcher can go from git clone to results without external dependencies.
Pattern 5: Lean CI for Heavy Workloads
All seven repositories now include GitHub Actions CI/CD pipelines that:
Run on
python:3.11-slimDocker imagesUse lean
requirements-test.txt(no GPU frameworks)Skip tests gracefully when heavy dependencies are unavailable
Validate project structure, notebook validity, and code quality
Deploy Sphinx documentation to GitHub Pages
This approach means CI runs in seconds, not minutes, while still catching structural and code issues.
Deviations and Status
UC |
Status |
Notes |
|---|---|---|
UC1 |
Completed |
Full tutorial, CLI, dual-platform support |
UC2 |
Completed |
9-chapter tutorial, digital twin, AGENT.md |
UC3 |
Completed |
phlearn package, test suite, CI/CD pipeline |
UC4 |
Completed |
Registration pipeline, orchestrator notebook, test suite, CI/CD |
UC5 |
Completed |
Full GNN framework, YAML-based configuration |
UC6 |
Completed |
Tutorial, parallelization, published in NMI |
UC7 |
Completed |
Autoencoder framework, educational sandbox |
Note
UC7 uses an MIT license, while the other WP7 repositories follow the dual CC BY-NC 4.0 (content) + GPL-3.0-only (code) standard. This will be updated for consistency.
Keypoints
Physics-informed architectures consistently outperform pure ML on out-of-distribution data
Data representation choices can matter more than model architecture choices
NAIC infrastructure enables both interactive exploration and batch-scale experiments
Self-contained repositories with setup scripts enable
git cloneto resultsLean CI/CD pipelines validate structure and code without requiring GPU