federica bianco PRO
astro | data science | data for good
University of Delaware
Department of Physics and Astronomy
Biden School of Public Policy and Administration
Data Science Institute
federica b. bianco
she/her
Opportunity
the era of AI
experiment driven science -∞:1900
theory driven science 1900-1950
data driven science 1990-2010
the fourth paradigm - Jim Gray, 2009
computationally driven science 1950-1990
experiment driven science -∞:1900
theory driven science 1900-1950
data driven science 1990-2010
the fourth paradigm - Jim Gray, 2009
computationally driven science 1950-1990
AI driven science? 2010...
Input
x
y
output
data
prediction
physics
Machine Learning
Machine Learning
Input
x
y
output
function
Machine Learning
Input
x
y
output
b
m
m: slope
b: intercept
Machine Learning
Input
x
y
output
b
m
m: slope
b: intercept
parameters
x
y
learn
goal: find the right m and b that turn x into y
goal: find the right m and b that turn x into y
Machine Learning
https://symposia.obs.carnegiescience.edu/series/symposium2/ms/freedman.ps.gz
Tree models
(at the basis of Random Forest
Gradient Boosted Trees)
Machine Learning
p(class)
extracted
features vector
p(class)
pixel values tensor
Frank Rosenblatt, 1958
The Navy revealed the embryo of an electronic computer today that it expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence.
The embryo - the Weather Buerau's $2,000,000 "704" computer - learned to differentiate between left and right after 50 attempts in the Navy demonstration
July 8, 1958
GPT-3
175 Billion Parameters
3,640 PetaFLOPs days
Kaplan+ 2020
Kaplan+ 2020
x
y
A Neural Network is a kind of function that maps input to output
Input
output
hidden layers
latent space
x
y
A Neural Network is a kind of function that maps input to output
Input
output
hidden layers
latent space
Opportunity
big data in astronomy
HELL YEAH!
2025
edge computing
Will we get more data???
SKA
(2025)
edge computing
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200)Villar+ 2018
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
edge computing
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200)
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
edge computing
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200)
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200) Villar+ 2018
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200) Villar+ 2018
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200) Villar+ 2018
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
SKA
(2025)
17B stars (x10) Ivezic+19
~10 million QSO (x10) Mary Loli+21
~50k Tidal Disruption Events (from ~150) Brickman+ 2020
~10k SuperLuminous Supernovae (from ~200) Villar+ 2018
~400 strongly lensed SN Ia (from 10) Ardense+24
~50 kilonovae (from 2) Setzer+19, Andreoni+19 (+ ToO)
> 10 Interstellar Objects fom 2.... ?)
True Novelties!
Discovery
Distribution
Classification
Data Integration and Follow up
Ensamble Inference
Prediction
Discovery of Novelties
(A.K.A science!)
Discovery
Distribution
Classification
Data Integration and Follow up
Ensamble Inference
Prediction
Discovery of Novelties
(A.K.A science!)
in <60 seconds:
Difference Image Analysis
in <60 seconds:
Difference Image Analysis
Can we replace DIA with ANN?
TANSINET: Sedhagat + Mahabal 2017
in 60 seconds:
Difference Image Analysis + Bogus rejection
feature extraction + Random Forest
AUTOSCAN: Goldstein+ 2017
96% accurate
Tatiana Acero-Cuellar, UNIDEL fellow, LSSTC data science fellow
92% accurate
Tatiana Acero-Cuellar, UNIDEL fellow, LSSTC data science fellow
What is the network learning?
What can we learn from the AI?
search
template
difference
template
search
Tatiana Acero-Cuellar, UNIDEL fellow, LSSTC data science fellow
What is the network learning?
What can we learn from the AI?
Tatiana Acero-Cuellar, UNIDEL fellow, LSSTC data science fellow
Interpretable AI
Robust AI
Anomaly detection
Distribution
Classification
Data Integration and Follow up
Ensamble Inference
Prediction
Discovery of Novelties
(A.K.A science!)
Discovery
F. Förster et al 2021 AJ 161 242
AI tasks
Challange
data encoding
well... it depends
2025
(2026)
edge computing
Is the data gonna also be better?
visualizatoin and concept credit: Alex Razim
Kaicheng Zhang et al 2016 ApJ 820 67
deSoto+2024
https://plasticc.org/data-release
Boone 2017
7% of LSST data
Boone 2017
7% of LSST data
The rest
Distribution
Classification
Data Integration and Follow up
Ensamble Inference
Prediction
Discovery of Novelties
(A.K.A science!)
Discovery
Rubin will see ~1000 SN every night!
Credit: Alex Gagliano IAIFI fellow MIT/CfA
Kepler satellite EB
LSST (simulated) EB
is transient data AI ready?
is transient data AI ready?
is transient data AI ready?
is transient data AI ready?
is transient data AI ready?
is transient data AI ready?
is transient data AI ready?
GAIA
Willow Fox Fortino
UDelaware
Optimal deep learning architectures for transients' spectral classification
As seen in Muthukrishna+2019
Classification from sparse data: Lightcurves
The PLAsTiCC challenge winnre, Kyle Boone was a grad student at Berkeley, and did not sue a Neural Network!
He won $2,000
Lochner et al 2018
Text
Classification from sparse data: Lightcurves
Classification from sparse data: Lightcurves
Text
Classification from sparse data: Lightcurves
without redshift
with redshift
Classification from sparse data: Lightcurves
without redshift
with redshift
Classification from sparse data: Lightcurves
without redshift
with redshift
Methodological issues with these approaches
CNNs are not designed to ingest uncertainties. Passing them as an image layer "works" but it is not clear why since the convolution on the flux and error space are averaged after the first layer
Gaussian processes work by imposing a kernel that represents the covariance in the data (how data depend on time or time/wavelength). Imposing the same kernel for different time-domain phenomena is principally incorrect
=> bias toward known classes
Methodological issues with these approaches
Neural processes replace the imposed kernel with a learned one
Garnelo+2018
2017
Classification from sparse data: Lightcurves
Viswani 2017 Attention is all you need
Classification from sparse data: Lightcurves
Viswani 2017 Attention is all you need
Willow Fox Fortino
UDelaware
When they go high, we go low
Classification power vs spectral resolution for SNe subtypes
Time series + low resolution spectrophotometry (R~3)
Precision Photometry (broad optical bands (G, BP, and RP) with space-based precision but bright magnitude limit (g~21)
Challange
benchmark datasets
Hlozek et al, 2020
DATA CURATION IS THE BOTTLE NECK
models contributed by the community were in
- different format (spectra, lightcurves, theoretical, data-driven)
- the people that contributed the models were included in 1 paper at best
- incompleteness
- systematics
- imbalance
khakpash+ 2024 showed that the models were biased for SN Ibc
AVOCADO, SCONE, all these models are trained on a biased dataset and are being currently used for classification
Ibc data-driven templates vs PLAsTiCC
khakpash+ 2024 showed that the models were biased for SN Ibc
AVOCADO, SCONE, all these models are trained on a biased dataset and are being currently used for classification
Ibc data-driven templates vs PLAsTiCC
Ibc data-driven templates vs PLAsTiCC
AI assisted modelling
Tardis uses a neural network to replace the radiative transfer model
LOW RES SIM
HIGH RES SIM
LOW RES SIM
HIGH RES SIM
AI-AIDED HIGH RES
multimodal data analysis
and pixel to science
2022
2022
why not images too?
latent space
latent space
lightcurve latent space rep
image
latent space rep
SN 2018cow
SN2024uwq
Perley+2018
lightcurve latent space rep
image
latent space rep
SN 2018cow
Perley+2018
SN 2018cow
Perley+2018
survey optimization
Opportunity
Rubin has involved the community to an unprecedented level in survey design this is a uniquely "democratic" process!
2024
2024
Challange
little amount of data
-infinity - 1950's
theory driven: little data, mostly theory, falsifiability and all that...
-1980's - today
data driven: lots of data, drop theory and use associations, black-box modles
lots of data yet not enough for entirely automated decision making
complex theory that cannot be solved analytically
combine it with some theory
Non Linear PDEs are hard to solve!
via a modified loss function that includes residuals of the prediction and residual of the PDE
Non Linear PDEs are hard to solve!
Raissi, Perdikaris, Karniadakis 2017
late layers learn complex aggregate specialized features
early layers learn simple generalized features (like lines for CNN)
prediction "head"
original data
trained extensively on large amounts of data to solve generic problems
Foundational AI models
trained extensively on large amounts of data to solve generic problems
Foundational AI models
We use the ILSVRC-2012 ImageNet dataset with 1k classes
and 1.3M images, its superset ImageNet-21k with
21k classes and 14M images and JFT with 18k classes and
303M high-resolution images.
Typically, we pre-train ViT on large datasets, and fine-tune to (smaller) downstream tasks. For
this, we remove the pre-trained prediction head and attach a zero-initialized D × K feedforward
layer, where K is the number of downstream classe
trained extensively on large amounts of data to solve generic problems
Foundational AI models
We use the ILSVRC-2012 ImageNet dataset with 1k classes
and 1.3M images, its superset ImageNet-21k with
21k classes and 14M images and JFT with 18k classes and
303M high-resolution images.
Typically, we pre-train ViT on large datasets, and fine-tune to (smaller) downstream tasks. For
this, we remove the pre-trained prediction head and attach a zero-initialized D × K feedforward
layer, where K is the number of downstream classe
ethics of AI
Challange + Opportunity
Knowledge is power
Knowledge is power
With great power comes grteat responsibility
"Sharing is caring"
the butterfly effect
We use astrophyiscs as a neutral and safe sandbox to learn how to develop and apply powerful tool.
Deploying these tools in the real worlds can do harm.
Ethics of AI is essential training that all data scientists shoudl receive.
Joy Boulamwini
is a word I am borrowing from Margaret Atwood to describe the fact that the future is us.
However loathsome or loving we are, so will we be.
Whereas utopias are the stuff of dream dystopias are the stuff of nightmares, ustopias are what we create together when we are wide awake
US-TOPIA
thank you!
University of Delaware
Department of Physics and Astronomy
Biden School of Public Policy and Administration
Data Science Institute
federica bianco
fbianco@udel.edu
Limited Field of View: Space telescopes often have smaller fields of view compared to ground-based surveys.
Data Latency: Delays in data transmission and processing can affect rapid follow-up.
Resource Allocation: Competition for telescope time can limit observations of certain transients.... LETS NOT TRIGGER 3 ToOs ON THE SAME TRANSIENT!!
(RacusinRacusin et al., 2008et al., 2008
(RacusinRacusin et al., 2008et al., 2008
GRB 080319B, the brightest optical burst ever observed
SWIFT
rapid response
SWIFT
HST, Chandra, SPITZER
...
Kepler, K2, TESS
high precision dense time series
By federica bianco
Opportunities and Challenges of Machine Learning and AI for the next-generation astronomical survey