QuForge is an end-to-end AI pipeline for superconducting qubit design exploration, reliability-aware candidate ranking, and evidence hardening.
Current design scope: fixed-frequency transmon qubits (and transmon-derived coupled metrics in this pipeline).
- ResearchGate preprint DOI: https://doi.org/10.13140/RG.2.2.24737.26726
It combines:
- physics-driven synthetic dataset generation,
- surrogate and embedding models for inverse design,
- coherence prediction with uncertainty signals,
- closed-loop selection logic,
- reliability ablations and source-holdout evidence reporting.
The goal is to make qubit design iteration faster and more systematic while being explicit about what is and is not validated.
This repository is designed for:
- research prototyping,
- pipeline benchmarking,
- candidate filtering before expensive fabrication and measurement cycles.
It is not currently claiming:
- fabrication-ready absolute coherence prediction without stronger real measured calibration.
- direct validity for non-transmon superconducting qubit modalities without dedicated re-calibration and validation.
| Path | Purpose |
|---|---|
Dataset/ |
Data generation scripts, templates, requirements, and dataset run commands |
Phase1_Surrogate/ |
Supervised surrogates for fast qubit/coupled-property prediction |
Phase2_Embedding/ |
Embedding models and nearest-neighbor inverse retrieval |
Phase3_InverseDesign/ |
Inverse design search and validation harness |
Phase4_Coherence/ |
Coherence predictor training, validation, and measurement dataset tools |
Phase4_5_PublicData/ |
Public data fetch, canonicalization, and conservative mapping |
Phase4_6_TraceCoherence/ |
Trace-based augmentation pipeline |
Phase5_ClosedLoop/ |
Closed-loop candidate scoring and fab handoff export |
Phase6_Reliability/ |
Reliability variant comparisons and stress tests |
Phase7_Evidence/ |
Multi-seed evidence hardening, holdout runs, and figure generation |
The latest strict run used:
- geometry enabled,
- Palace required (
--no-palace-fallback), - conservative public mapping with fitted/model curves excluded,
- design-disjoint source holdout with leakage audit.
| Item | Value |
|---|---|
| Single-device dataset rows | 4142 |
| Coupled-device dataset rows | 80000 |
| Strict mapped measured rows (direct) | 26 |
| Augmented measured rows (direct + synthetic) | 170 |
| Hybrid matched supervision rows | 150 |
| Eligible measured sources after strict mapping | 2 |
| Unique mapped designs | 6 |
Source-provenance note:
- Canonical ingestion includes
Zenodo:18045662,DataGov:NIST:mds2-3027,SQuADDS, andZenodo:15364358_tracefit. - Under strict mapping and fitted/model-curve exclusion, retained geometry-linked supervised rows are:
SQuADDS:5rows withT1/T2DataGov:NIST:mds2-3027:21rows withT2only- large Zenodo records:
0geometry-linked supervised rows after strict filtering/mapping.
| Metric | Result |
|---|---|
| Phase 5 targets/candidates/selected | 24 / 192 / 26 |
| Phase 7 top variant by 10-seed reliability rank | hybrid_high_trust (rank_mean=1.4, top1_count=6) |
| Holdout leakage overlap (max) | 0 |
Source holdout SQuADDS T1 MAE |
23.41 us (95% CI [11.52, 30.57]) |
Source holdout SQuADDS T2(log10) MAE |
1.40 (95% CI [0.76, 1.99]) |
Source holdout DataGov:NIST:mds2-3027 T2(log10) MAE |
0.45 (95% CI [0.37, 0.53]) |
Interpretation:
- Internal ranking behavior is meaningful and reproducible.
- Cross-source absolute transfer remains weak with current measured-data volume.
These figures are included from the latest strict large-sweep outputs.
| Topic | Supported Now | Not Supported Yet |
|---|---|---|
| Reliability-aware ranking/filtering | Yes | N/A |
| Closed-loop shortlist generation | Yes | N/A |
| Robustness comparisons across variants | Yes | N/A |
| Absolute coherence prediction for production/fab handoff | Limited | Yes, requires substantially more measured calibration data |
| Cross-lab generalization claim | Weak evidence | Yes, needs out-of-source improvement with richer measured datasets |
This project explicitly separates:
- what can be learned from simulation + public data bootstrapping,
- what requires real hardware measurement.
Important:
- Real measured coherence cannot be created by synthetic simulation alone.
- Publicly mined rows are sparse and heterogeneous after strict filtering.
- Current holdout behavior confirms a real transfer gap.
See:
Phase4_Coherence/MEASUREMENT_DATA_GUIDE.md
Dataset\.venv310\Scripts\python -m pip install -r Dataset\requirements-dataset.txt
Dataset\.venv310\Scripts\python Dataset\generate_data.py --dry-run --workdir DatasetDataset\.venv310\Scripts\python Dataset\generate_data.py --workdir Dataset --sampling-mode random --geometry-samples 1200 --junction-samples 10 --target-single-rows 10000 --max-pairs 5000 --gmsh-verbosity 0 --mesh-lc-min-um 30 --mesh-lc-max-um 120 --mesh-optimize-threshold 0.45 --no-palace-fallbackRun in order:
Dataset\.venv310\Scripts\Activate.ps1
python Dataset\generate_data.py --workdir Dataset --sampling-mode random --geometry-samples 1200 --junction-samples 10 --target-single-rows 10000 --max-pairs 5000 --gmsh-verbosity 0 --mesh-lc-min-um 30 --mesh-lc-max-um 120 --mesh-optimize-threshold 0.45 --no-palace-fallback
python Phase4_5_PublicData\build_public_canonical_dataset.py
python Phase4_5_PublicData\augment_large_raw_sources.py --input-csv Dataset\public_sources\silver\public_measurements_canonical.csv --output-csv Dataset\public_sources\silver\public_measurements_canonical.csv
python Phase4_5_PublicData\map_public_to_internal.py --max-distance 1.6 --min-confidence 0.30
python Phase4_5_PublicData\augment_anchor_conditioned_regularization.py --measurement-csv Dataset\measurement_dataset_public_bootstrap.csv --output-csv Dataset\measurement_dataset_public_bootstrap_augmented.csv --report-json Dataset\measurement_dataset_public_bootstrap_augmented.report.json --neighbors-per-anchor 24
python Phase1_Surrogate\train_surrogate.py --n-estimators 250 --max-depth 32
python Phase2_Embedding\train_phase2_embedding_nn.py --temperature 0.07
python Phase2_Embedding\validate_phase2_embedding_nn.py
python Phase3_InverseDesign\validate_phase3.py --max-queries-per-split 40
python Phase4_Coherence\train_phase4_coherence.py --label-mode hybrid --measurement-csv Dataset\measurement_dataset_public_bootstrap_augmented.csv --measured-weight 1.0 --proxy-weight 1.0 --synthetic-label-blend 0.35 --synthetic-regularization-weight 0.75 --repeated-weight-power 0.5
python Phase4_Coherence\validate_phase4_coherence.py --measurement-csv Dataset\measurement_dataset_public_bootstrap_augmented.csv
python Phase5_ClosedLoop\run_phase5_closed_loop.py
python Phase6_Reliability\run_phase6_reliability.py --measurement-csv Dataset\measurement_dataset_public_bootstrap_augmented.csv
python Phase7_Evidence\run_phase7_evidence.py --measurement-csv Dataset\measurement_dataset_public_bootstrap_augmented.csv --output-dir Phase7_Evidence\artifacts_large_sweep --seeds 42,123,777,1001,1002,1003,1004,1005,1006,1007 --holdout-mode design_disjoint --holdout-min-sources 2
# Figure scripts keep "publication" in filename for historical compatibility.
python Phase7_Evidence\generate_publication_figures.py --input-dir Phase7_Evidence\artifacts_large_sweep --output-dir Phase7_Evidence\artifacts_large_sweep\figures
python Phase7_Evidence\generate_extended_publication_figures.py --phase7-dir Phase7_Evidence\artifacts_large_sweep --phase6-dir Phase6_Reliability\artifacts --phase5-dir Phase5_ClosedLoop\artifacts --single-csv Dataset\final_dataset_single.csv --output-dir Phase7_Evidence\artifacts_large_sweep\figuresFixed evidence seed set: 42,123,777,1001,1002,1003,1004,1005,1006,1007.
- Python environment used in this project:
Dataset/.venv310. - Docker is required for Palace-based strict generation.
- WSL + Docker Desktop works for large sweeps.
- Generated runtime artifacts are intentionally git-ignored.
We are actively looking for collaboration with labs, and with individual researchers/engineers affiliated with labs, to improve real-world calibration and transfer.
Minimum schema per measured sample:
| Field | Required | Notes |
|---|---|---|
row_index or design_id |
Yes | Must map to Dataset/final_dataset_single.csv |
measured_t1_us |
Yes | Real measured T1 in microseconds |
measured_t2_us |
Yes | Real measured T2 in microseconds |
measured_freq_01_GHz |
Optional | Measured transition frequency |
measured_anharmonicity_GHz |
Optional | Measured anharmonicity |
source_name |
Strongly recommended | Lab/system identifier for source-aware calibration |
chip_id, cooldown_id, measurement_date_utc, notes |
Optional | Useful for drift and provenance analysis |
For details:
Dataset/measurement_raw.template.csvDataset/measurement_dataset.template.csvPhase4_Coherence/MEASUREMENT_DATA_GUIDE.md
- Public contribution of de-identified CSVs.
- Private collaboration under data use terms, with aggregate-only reporting.
- Joint benchmark setup with fixed train/holdout protocols.
- Email:
yousuf.tomal.0@gmail.com - Required subject line:
QuForge Collaboration - Real Measured Qubit Data - <Lab/Institution> - In the email body, include:
- data availability (approximate row count and fields),
- measurement stack (device family, readout pipeline, cooldown context),
- preferred collaboration mode (public, private, or joint benchmark).
Apache License 2.0. See LICENSE.








