Unraveling the Cosmos: Key Moments from the Annual Conference
The first annual conference for the Center for Decoding the Universe, on June 5-6 (agenda), convened an impressive group of exceptional researchers. Check out the full recap, including on-demand videos and event photosâscroll down to view the photo album.
Session 1: Time-Domain Data & Anomaly Discovery
The opening session on Time-Domain Data & Anomaly Discovery brought together perspectives from astrophysics, statistics, high-energy physics, and machine learning.
Ashley Villar discussed the upcoming revolution in time-domain astrophysics, with the Rubin Observatory expected to drive the supernova discovery rate from the current ~10,000 per year to one million per year. She grouped the science challenges by anomaly type and demonstrated how each demands a different feature space. Villar highlighted âoneâinâaâmillionâ opportunities such as kilonovaâlike collapsars and PopulationâŻIII explosions, while cautioning that rare events can hide in familiar brightnessâtimescale coordinates. She concluded with a call for rapid, data-driven inference pipelines capable of triaging millions of nightly alerts without missing the exotic discoveries.
The next speaker, BenâŻNachman, reframed anomaly detection through a dual lens of latency (online vs. offline) and supervision (unsupervised to weakly supervised). Using techniques from particle physics, he contrasted density-based autoencoders, which are powerful but coordinate-dependent, with likelihood-ratio methods that remain invariant under nonlinear transformations. Nachman mentioned the case of the Gaia stellar streams, where a weakly supervised classifier automatically recovered the GDâ1 stream and its fine subâstructure, all without explicit astrophysical priors.
JoshâŻBloom then asked a metaâquestion: What do we want from anomaly detection? After showcasing a multimodal foundation model that represents light curves, spectra, and metadata into an interpretable embedding, Bloom discussed how many discoveries may be due to poor spectral reductions or photometric artifacts. He argued for pushing anomaly detection upstream, directly onto raw images and telemetry, to flag systematics before they appear as new physics. Bloom sketched a future in which automated Occam filters, key physics checks, and agentic LLMs do the first pass, leaving humans to interrogate only those outliers that survive the filters.
For the short talks, Maja Jablonska introduced SPICE. This JAX-based, fully differentiable forward model tessellates stellar photospheres and can render the spectra of spotted, pulsating, or eclipsing stars. PeterâŻMelchior then presented a restâframeâaware autoâencoder for galaxy spectra that, in just six latent parameters, flags broadâabsorptionâline quasars and starâburst outliers while remaining invariant to redshift. Next, CharlotteâŻWard introduced Scarlet2, a JAX rewrite of the Rubin deâblender that simultaneously fits Rubin, Euclid, Roman, and LS4 images. Linnea Wolniewicz then discussed a Gaussian-process pipeline to model complex stellar variability, enabling searches that capture hour-long dips and week-long exocomet occultations in Kepler light curves. Finally, Christine Ye described ResearchBench, an open benchmark that converts 22 peer-reviewed astrophysics papers into 121 autonomous reproduction tasks, with the frontier models scoring below 20%, revealing subtle but systematic reasoning gaps that future agents must address.
During the long discussion, participants debated the balance between domain knowledge and flexible ML. Several groups championed a version-controlled archive of recurring detector or reduction artifacts, against which every new alert could be cosine-matched. Resource efficiency surfaced as a priority, with one table proposing a threeâtier triage: physicsâencoded heuristics first, lightweight ML next, and only then deep learning. The room agreed that models should have saliency maps so humans can see why a detection occurred. Participants from NVIDIA and Stanford Medicine noted that low-frame-rate MRI sequences face identical rare-event problems and formed a cross-disciplinary working group. Finally, moderator PhilâŻMarshall polled the room on who would find Rubinâs first novel discovery: humans received thirtyâfive votes, ML twenty, and hybrid systems five.
The oneâminute poster session ranged from Uâmap clustering of Chandra Xâray transients to densityâablation clustering for direct darkâmatter events.
Session 2: Cosmology & Modern Inference Frameworks
The second session, Cosmology & Modern Inference Frameworks, shifted the focus from discovery to precise parameter estimation when data volumes require more than classical statistics.âŻBenâŻWandelt (in a preârecorded keynote) surveyed the state of âimplicit inferenceââ: likelihoodâfree, simulatorâbased techniques able to recover full cosmological posteriors even when the forward model includes expensive Nâbody or hydrodynamic solvers. He discussed three open bottlenecks, namely scalability, model fidelity, and misâspecification, and showed how transformerâbased halo emulators (Gotham) and diffusionâstyle field reconstructions can close each gap. Wandelt closed with a challenge: prove that tomorrowâs Rubinâscale analyses can propagate all sources of uncertainty, from subâgrid baryonic recipes to adversarial domain shifts in simulation code.
LaurenceâŻPerreaultâŻLevasseur focused on a concrete use case: joint lensâmassâŻ+âŻsourceâreconstruction for the hundreds of thousands of galaxyâscale strong lenses Rubin will deliver. She argued that classical pixelâbased inversions struggle when using 10,000+ nuisance parameters. Her group replaces the handâtuned Gaussian priors with scoreâbased generative models that learn the manifold of highâredshift galaxies. In simulation, the approach removes the sizeâbias that a zeroâcentred Gaussian prior imprints on lensâmass estimates and remains stable when the noise model itself is learned from JWST dark frames via diffusion networks.
The sessionâs third talk came from VasilisâŻSyrgkanis, who bridged theory and practice by discussing automatic debiasing. He mentioned that semiâparametric estimators often inherit the regularisation bias of the machineâlearning models that supply their nuisance functions. Syrgkanis demonstrated the recipe on causalâeffect estimation and outlined open problems at the cosmologyâstatistics interface, notably highâdimensional hierarchical models.
For the short talks, Sydney Erickson focused on a neural posterior estimator that delivers perâlens mass posteriors from HSTâlike images, which are then fed into a populationâlevel cosmology fit that simultaneously learns the lensâgalaxy mass function. TiffanyâŻFan presented a contextâaware graph neural surrogate for particleâaccelerator beams. SophiaâŻLu introduced ABI (Adaptive Bayesian Inference), a likelihoodâfree rejectionâABC variant that refines posterior quantiles via a Wasserstein distance. Guillem Megias Homar discussed early Rubin commissioning data, and Shuo Xinâs talk focused on neural rendering that employs adaptive mesh refinement.
The long discussion centered around two key issues. First, if simulatorâbased pipelines no longer fit individual images âdown to the noise,â how do we certify that the aggregated cosmological posteriors are still trustworthy? Suggestions ranged from blinded endâtoâend challenges to perâstatistic saliency checks that reveal which scales drive the inference. Second, everyone discussed the model and prior misâspecification. Consensus emerged that the community needs (i) flexible, dataâdriven priors that can be updated hierarchically as new samples roll in, and (ii) robustness tests that quantify how far wrong the forward model can be before the posteriors break. The session closed with a call to develop public misâspecification testâbeds, analogous to ImageNet adversarial suites, so methods can be stressâtested before Rubin data releases.
Oneâminute poster talks discussed diffusionâbased deconvolution for SNâŻIa host spectroscopy, negativeâweight âneuroârefinementââ for collider unfolding, and UâMAP clustering of redâshiftâspace galaxy mocks.
Session 3: Galaxies and Foundation Models
MarcâŻHuertasâCompany opened by discussing the traditional, surveyâdriven approach to decode galaxy evolution: derive physically motivated summary statistics, search for scaling relations, and tune cosmological simulations. He showed how neural networks already replace many handcrafted estimators (stellar mass, SFR, metallicity) but warned that familiar correlations still allow vastly different subâgrid physics. He argued that the next step is multiâmodal, multiâscale AI: joint embeddings of optical, radio, Xâray, and simulation data that map models and observations into a higherâdimensional space, which is the only way to break degeneracies in gas accretion, feedback, and quenching.
FrançoisâŻLanusse discussed the methodological change from singleâtask supervised CNNs to reusable foundation models. Self-supervised contrastive learning on hundreds of millions of images yields embeddings that perform as well as or better than existing models for lens finding, morphology, redshift, or stellarâmass estimation. Extending CLIPâstyle training to spectraâplusâimages (AstroCLIP) creates a shared space where a nearest neighbour in embedding space transfers physical parameters with stateâofâtheâart accuracy. He showcased a generative transformer that tokenizes every modality (HST, JWST, Euclid, SDSS) and can impute highâresolution spectra or multiâband images from sparse inputs, which is a step toward âan astronomical GPTâ that both simulates and interprets the sky.
Drawing on 15âŻyears in computer vision, DavidâŻFouhey put the spotlight on selfâsupervised depth and segmentation breakthroughs, yet cautioned that they required vast, carefully curated benchmarks. Like autonomousâdriving edgeâcases, spiral disks at zâŻââŻ10 or solar megaâflares may appear in the long tail where embeddings and uncertainty quantification may not be enough. Fouhey urged the community to invest in rigorous downstream tests to keep plausible pictures from substituting for real understanding.
A lively discussion followed on how to insert physics into learned representations, whether we will soon work only in embedding space, and how to detect novel phenomena that foundation models have never shown.
The short talk session started with PhilippâŻFrank, who introduced GeometricâŻVariationalâŻInference, which is already mapping MilkyâWay dust and blackâhole shadows. Sonia Kim demonstrated an inverseâproblem solver that is competitive with stateâofâtheâart MAP and posterior samplers. BenjaminâŻRemy showed how iterative diffusion priors sharpen sourceâseparated HST images. Rahul Mysore Venkatesh then discussed motionâcompletion transformer predicts objectâwise causal masks from raw video, hinting at similar embeddings for timeâdomain surveys. Finally, SebastianâŻWagner-Carena presented a multiâview diffusion framework that learns galaxyâandâbackground priors without labels, achieving clean component separation across resolutions.
Poster spotlights ranged from physicsâguided diffusion guidance to visionâtransformer photometric redshifts and RomanâLSST deblending methods.
The takeaway is that foundation models are rapidly becoming an integral part of survey astrophysics, but extracting physics from their embeddings will demand equal parts creativity, rigor, and domain knowledge.