Deep Learning in Particle Physics
Deep Learning in Particle Physics
learning"
"Machine
Lerningg"
"neural
network"
Even bananas!
shape shifter!
Super Kamiokande (1997)
50,000 ton of ultra pure water,
"photographed" by 11,000 PMTs
smashing more than 10 million high energy protons
producing few trillion neutrinos every 1.3 seconds
NUE CREATES ELECTRON NUMU CREATES MUONS
Electron “showers/hits in cone shape” in detector Muon creates long “track/line” in detector
HOW TO IDENTIFY A NEUTRINO (IF IT
WERE A CAT)
Non ML: very hard and tedious task
Development
Workflow
(Machine
Learning)
image from:
https://towardsdatascience.com/intuitive-deep-learning-part-2-cnns-for-computer-vision-472bbb2c8060 https://towardsdatascience.com/intuitive-
deep-learning-part-2-cnns-for-computer-
vision-472bbb2c8060
APPLY THE EXACT SAME AFTER APPLYING THAT
NEURON FOR THE OTHER NEURON FOR ALL FOUR
QUADRANTS QUADRANTS, WE HAVE
FOUR DIFFERENT NUMBERS
Development
Workflow
(Machine
Learning)
image from:
https://towardsdatascience.com/intuitive-
deep-learning-part-2-cnns-for-computer-
vision-472bbb2c8060
TAKE THE MAXIMUM OF THE FOUR NUMBER TO
GET A SINGLE NUMBER
Development
Workflow
(Machine
Learning)
image from:
https://towardsdatascience.com/intuitive-
deep-learning-part-2-cnns-for-computer-
vision-472bbb2c8060
plastic with
scintillator
interspersed
between
passive
target
Each triangular hit = scintillator plane with optic fiber
"vertex": start of Detector "lights up" when charged particle passed
neutrino interaction
through it. Colors represent energy deposited
neutrino
EM-like particles,
many protons/pions
"shower"
rotated 60o from y rotated -60o from y
In experimental
particle physics,
we use simulation
(synthetic data)
where we know the
true information
(labeled data)
Useful for tuning
pattern recognition
input layer hidden layer
output layer
hit-time in x, u, v view
Supervised learning: use simulation where we know the "truth"
hidden layer
https://arxiv.org/abs/1808.08332
Journal of Instrumentation, Volume 13, Number 11, 2018
hidden layer
Softmax layer:
True z-segment: labeled data/true
information of which segment the vertex
is generated
Reconstructed z-segment: a vector of
softmax probabilities of DCNN predicted
segment
https://arxiv.org/abs/1808.08332
Journal of Instrumentation, Volume 13, Number 11, 2018
Train and prediction in different domains
Train with labeled data: in our case it is Simulation/synthetic
data (lets call it source domain)
Test with unlabeled data: in our case it is real data (lets call it
target domain)
cat not cat
from training our models in one domain and applying them in
another.
Label predictor:
output
Minimize the loss of the label
classifier so that network can
With DANN:
Label predictor:
output
Domain classifier:
works internally
Minimize the loss of the label
classifier so that network can
With DANN:
Label predictor:
output
Domain classifier:
works internally
Label predictor:
output
Domain classifier:
works internally
The network develops an insensitivity to features that are present in one domain but
not the
other, and train only on features that are common to both domains
https://arxiv.org/abs/1808.08332
Journal of Instrumentation, Volume 13, Number 11, 2018
using Caffe for the deep learning network
Blue vs black: model trained in the same domain (FSI active,Blue curve) is better
than a
model trained with an out-of domain physics model (FSI inactive, black curve)
Analysis framework developed for processing LArTPC image data. Supports C++ data structures, IO
interface, and data processing
machinery.
Directly manipulate the image data with or without OpenCV.
Interface with open source deep learning softwares including Caffe by Berkeley Lab. and TensorFlow by
Google.
ROOT format- easy to handle (for particle physics) can do other things like crop images , resize images
etc
https://doi.org/10.48550/arXiv.2103.06992
https://doi.org/10.48550/arXiv.2103.06992
https://doi.org/10.48550/arXiv.2103.06992
https://doi.org/10.48550/arXiv.2103.06992
Traditional Semantic
Reconstruction Segmentation
https://arxiv.org/abs/2008.01242
https://a3d3.ai/
https://arxiv.org/abs/2008.01242