cross-scenario-physics-code-transfer / code /_rev_linear_probe_matched_visual.py
physics-code-transfer-bench's picture
Initial anonymous release for NeurIPS 2026 E&D submission
189f45b verified
"""
EXP REV-LP-MV: Linear probes on matched-visual conditions (R2 highest-value fix).
Adds linear-probe baselines to the gradient figure for:
1. Velocity interpolation (matched visuals, kinematic split)
2. Elastic vs inelastic restitution split (matched visuals, dynamics-class split)
3. Standard-gravity vs low-gravity (matched visuals, dynamics shift)
Each is a logistic regression on l2-pooled V-JEPA 2 features at N in {16, 192}, 5 seeds.
"""
import json
import time
import os
from pathlib import Path
from datetime import datetime, timezone
import numpy as np
import torch
from sklearn.linear_model import LogisticRegression
from sklearn.preprocessing import StandardScaler
PROMPT_RECEIVED_TIME = datetime.now(timezone.utc).isoformat()
print(f"PROMPT_RECEIVED_TIME = {PROMPT_RECEIVED_TIME}", flush=True)
T0 = time.time()
OUT = Path("results/reviewer_response/exp_lp_matched_visual")
OUT.mkdir(parents=True, exist_ok=True)
N_LIST = [16, 192]
N_SEEDS = 5
RNG_BASE = 1234
def log(msg):
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
print(f"[{ts}] LP-MV: {msg}", flush=True)
def pool_l2(features_3d):
"""L2-pool features along temporal axis: (N, T, D) -> (N, D)."""
f = features_3d
if f.ndim == 3:
return f.mean(dim=1).numpy()
return f.numpy()
def stratified_subset(rng, y, n_per_class):
"""Indices of n_per_class examples per class."""
idxs = []
for c in np.unique(y):
cand = np.where(y == c)[0]
if len(cand) == 0:
continue
chosen = rng.choice(cand, size=min(n_per_class, len(cand)), replace=False)
idxs.extend(chosen.tolist())
return np.array(sorted(idxs))
def train_lp(X_tr, y_tr, X_te, y_te):
sc = StandardScaler().fit(X_tr)
Xs_tr = sc.transform(X_tr)
Xs_te = sc.transform(X_te)
model = LogisticRegression(max_iter=2000, C=1.0, multi_class="auto",
solver="lbfgs")
model.fit(Xs_tr, y_tr)
return float((model.predict(Xs_te) == y_te).mean())
def stats(vals):
v = np.array(vals)
return float(v.mean()), float(v.std(ddof=1) if len(v) > 1 else 0.0)
def run_split(name, X_src, y_src, X_tgt, y_tgt, n_classes):
"""Evaluate linear probe at N in N_LIST.
Source-train-only baseline: train on full source, evaluate on target (N=0).
N>0: train on full source + N stratified target examples, evaluate on remaining target.
"""
log(f"=== {name}: src={X_src.shape}, tgt={X_tgt.shape}, n_classes={n_classes}")
results = {"N0_source_only": [], "curve": {N: [] for N in N_LIST}}
# N=0: train on source, evaluate on target
for s in range(N_SEEDS):
# Subsample source to be fair (use all of it; stratification not needed here)
acc = train_lp(X_src, y_src, X_tgt, y_tgt)
results["N0_source_only"].append(acc)
log(f" N=0 src-only: {stats(results['N0_source_only'])[0]:.3f} ± {stats(results['N0_source_only'])[1]:.3f}")
# N=16,192: train on source + N stratified target, eval on remaining target
smallest = min(int(np.sum(y_tgt == c)) for c in np.unique(y_tgt))
for N in N_LIST:
per_class = max(1, N // n_classes)
# Clamp so we leave at least 30% of each target class for evaluation
per_class = min(per_class, int(0.7 * smallest))
for s in range(N_SEEDS):
rng = np.random.default_rng(RNG_BASE + s)
tgt_idx_train = stratified_subset(rng, y_tgt, per_class)
mask = np.ones(len(y_tgt), bool); mask[tgt_idx_train] = False
X_eval = X_tgt[mask]; y_eval = y_tgt[mask]
if len(y_eval) == 0:
continue
X_tr = np.concatenate([X_src, X_tgt[tgt_idx_train]], axis=0)
y_tr = np.concatenate([y_src, y_tgt[tgt_idx_train]], axis=0)
acc = train_lp(X_tr, y_tr, X_eval, y_eval)
results["curve"][N].append(acc)
if results["curve"][N]:
m, sd = stats(results["curve"][N])
log(f" N={N:>3d}: {m:.3f} ± {sd:.3f} (per_class={per_class})")
else:
log(f" N={N:>3d}: SKIPPED (insufficient target data)")
return results
# ──────────────────────────────────────────────────────────────────
# Load standard collision features and labels
# ──────────────────────────────────────────────────────────────────
log("Loading standard collision features ...")
std_feat = torch.load(
"results/acceptance_boost/exp2_cache/feat_vjepa2_collision_orig.pt",
map_location="cpu", weights_only=False)["features"]
log(f" std collision features: {tuple(std_feat.shape)}")
labels = np.load("results/kinematics_vs_mechanics/labels_collision.npz")
restitution_bin = labels["restitution_bin"] # 600-d, 3 classes
mass_bin = labels["mass_bin"] # 600-d, 3 classes
velocity_pre_scalar = labels["velocity_pre_scalar"]
restitution_scalar = labels["restitution_scalar"]
log(f" labels: restit_bin classes {sorted(set(restitution_bin))}, mass_bin classes {sorted(set(mass_bin))}")
X_std = pool_l2(std_feat) # (600, 1024)
log(f" X_std shape: {X_std.shape}")
# ──────────────────────────────────────────────────────────────────
# 1. Velocity interpolation (matched-visual kinematic split)
# train on low-velocity half, eval on high-velocity half (predict restitution_bin)
# ──────────────────────────────────────────────────────────────────
log("=== Velocity interpolation split ===")
vmed = float(np.median(velocity_pre_scalar))
log(f" velocity median = {vmed:.3f}")
mask_lo = velocity_pre_scalar < vmed
mask_hi = ~mask_lo
# direction A: train on lo, eval on hi
res_velocity_lo2hi = run_split(
"velocity lo->hi",
X_std[mask_lo], restitution_bin[mask_lo],
X_std[mask_hi], restitution_bin[mask_hi],
n_classes=3,
)
res_velocity_hi2lo = run_split(
"velocity hi->lo",
X_std[mask_hi], restitution_bin[mask_hi],
X_std[mask_lo], restitution_bin[mask_lo],
n_classes=3,
)
# ──────────────────────────────────────────────────────────────────
# 2. Elastic vs inelastic split (matched-visual dynamics-class split)
# train on elastic (restit_scalar >= 0.5), eval on inelastic (predict mass_bin)
# ──────────────────────────────────────────────────────────────────
log("=== Elastic <-> inelastic split ===")
mask_elas = restitution_scalar >= 0.5
mask_inelas = ~mask_elas
log(f" n elastic: {mask_elas.sum()}, n inelastic: {mask_inelas.sum()}")
res_elas2inelas = run_split(
"elas->inelas",
X_std[mask_elas], mass_bin[mask_elas],
X_std[mask_inelas], mass_bin[mask_inelas],
n_classes=3,
)
res_inelas2elas = run_split(
"inelas->elas",
X_std[mask_inelas], mass_bin[mask_inelas],
X_std[mask_elas], mass_bin[mask_elas],
n_classes=3,
)
# ──────────────────────────────────────────────────────────────────
# 3. Standard gravity <-> low gravity (matched-visual dynamics shift)
# Need 75-scene std subset matching low-gravity (use seed-matched first 75 by RNG)
# The exp_p1 setup matched RNG so std-grav and low-grav share per-scene physics
# We use the first 75 scenes of std-grav (seed-aligned) as the matched set.
# ──────────────────────────────────────────────────────────────────
log("=== Std gravity <-> low gravity ===")
lg_path = "results/reviewer_response/exp_p1/feat_vjepa2_lowgrav.pt"
lg_feat = torch.load(lg_path, map_location="cpu", weights_only=False)["features"]
log(f" low-grav features: {tuple(lg_feat.shape)}")
X_lg = pool_l2(lg_feat)
# Load low-grav labels from index.json
with open("kubric/output/collision_low_gravity_dataset/index.json") as fh:
lg_idx = json.load(fh)
lg_restitution_scalar = np.array([s["restitution"] for s in lg_idx])
# Use Kubric union bins -- compute on std-grav restitution scalars
restit_bin_edges = np.percentile(restitution_scalar, [33.333, 66.667])
log(f" union restit bin edges: {restit_bin_edges}")
def to_bin(scalar, edges):
return np.searchsorted(edges, scalar)
y_lg_restit = to_bin(lg_restitution_scalar, restit_bin_edges).astype(np.int64)
y_std_restit = to_bin(restitution_scalar, restit_bin_edges).astype(np.int64)
log(f" lg restit bin distribution: {np.bincount(y_lg_restit)}")
log(f" std restit bin distribution: {np.bincount(y_std_restit)}")
# Use first 75 std-grav scenes (RNG-aligned to low-grav generation)
res_std2lg = run_split(
"std->lg",
X_std[:75], y_std_restit[:75],
X_lg, y_lg_restit,
n_classes=3,
)
res_lg2std = run_split(
"lg->std",
X_lg, y_lg_restit,
X_std[:75], y_std_restit[:75],
n_classes=3,
)
# ──────────────────────────────────────────────────────────────────
# Aggregate and save
# ──────────────────────────────────────────────────────────────────
def merge_dirs(a, b):
"""Average two directional results."""
out = {"N0_source_only": [], "curve": {N: [] for N in N_LIST}}
out["N0_source_only"] = a["N0_source_only"] + b["N0_source_only"]
for N in N_LIST:
out["curve"][N] = a["curve"][N] + b["curve"][N]
return out
full = {
"velocity_lo2hi": res_velocity_lo2hi,
"velocity_hi2lo": res_velocity_hi2lo,
"velocity_mean": merge_dirs(res_velocity_lo2hi, res_velocity_hi2lo),
"elas2inelas": res_elas2inelas,
"inelas2elas": res_inelas2elas,
"elastic_mean": merge_dirs(res_elas2inelas, res_inelas2elas),
"std2lg": res_std2lg,
"lg2std": res_lg2std,
"gravity_mean": merge_dirs(res_std2lg, res_lg2std),
}
# Pretty summary
SUMMARY = ["EXP REV-LP-MV -- linear probes on matched-visual conditions (5 seeds, predict restitution/mass)",
"",
f"{'Condition':<30s} | {'N=0 (src-only)':>18s} | {'N=16':>14s} | {'N=192':>14s}",
"-" * 86]
for name, r in full.items():
if "_mean" not in name and not name in ("std2lg", "lg2std", "elas2inelas", "inelas2elas"):
continue
n0_m, n0_s = stats(r["N0_source_only"])
n16_m, n16_s = stats(r["curve"][16])
n192_m, n192_s = stats(r["curve"][192])
SUMMARY.append(f"{name:<30s} | {n0_m*100:>5.1f}% +/- {n0_s*100:>4.1f}% | {n16_m*100:>5.1f}% +/- {n16_s*100:>4.1f}% | {n192_m*100:>5.1f}% +/- {n192_s*100:>4.1f}%")
print("\n".join(SUMMARY), flush=True)
with open(OUT / "exp_lp_matched_visual_summary.txt", "w") as fh:
fh.write("\n".join(SUMMARY) + "\n")
with open(OUT / "exp_lp_matched_visual_summary.json", "w") as fh:
json.dump(full, fh, indent=2)
end_ts = datetime.now(timezone.utc).isoformat()
runtime_min = (time.time() - T0) / 60.0
print(f"\nEND_TIME = {end_ts}", flush=True)
print(f"Total runtime: {runtime_min:.2f} min", flush=True)