Papers
Topics
Authors
Recent
Search
2000 character limit reached

Score-based Diffusion Models for Generating Liquid Argon Time Projection Chamber Images

Published 25 Jul 2023 in hep-ex | (2307.13687v3)

Abstract: For the first time, we show high-fidelity generation of LArTPC-like data using a generative neural network. This demonstrates that methods developed for natural images do transfer to LArTPC-produced images, which, in contrast to natural images, are globally sparse but locally dense. We present the score-based diffusion method employed. We evaluate the fidelity of the generated images using several quality metrics, including modified measures used to evaluate natural images, comparisons between high-dimensional distributions, and comparisons relevant to LArTPC experiments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. Carlo Rubbia. The liquid-argon time projection chamber: a new concept for neutrino detectors. Technical report, 1977.
  2. Fermilab proposal p-496. Technical report, 1976.
  3. Design, construction and tests of the ICARUS T600 detector. Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, 527(3):329–410, 2004.
  4. The ArgoNeuT detector in the NuMI low-energy beam line at fermilab. Journal of Instrumentation, 7(10):P10019, 2012.
  5. Design and construction of the MicroBooNE detector. Journal of Instrumentation, 12(02):P02017, 2017.
  6. A proposal for a three detector short-baseline neutrino oscillation program in the fermilab booster neutrino beam. arXiv:1503.01520, 2015.
  7. The DUNE far detector interim design report volume 1: Physics, technology and strategies. arXiv:1807.10334, 2018.
  8. Larpix: Demonstration of low-power 3d pixelated charge readout for liquid argon time projection chambers. Journal of Instrumentation, 13(10):P10007, 2018.
  9. First demonstration of a pixelated charge readout for single-phase liquid argon time projection chambers. Instruments, 4(1):9, 2020.
  10. Ionization electron signal processing in single phase lartpcs. part i. algorithm description and quantitative evaluation with microboone simulation. Journal of Instrumentation, 13(07):P07006, 2018.
  11. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10684–10695, 2022.
  12. Exploring painting synthesis with diffusion models. In 2021 IEEE 1st International Conference on Digital Twins and Parallel Intelligence (DTPI), pages 332–335. IEEE, 2021.
  13. Difffacesketch: High-fidelity face image synthesis with sketch-guided latent diffusion model. arXiv preprint arXiv:2302.06908, 2023.
  14. Pc-jedi: Diffusion for particle cloud generation in high energy physics. arXiv preprint arXiv:2303.05376, 2023.
  15. Improving generative model-based unfolding with schrödinger bridges. arXiv preprint arXiv:2308.12351, 2023.
  16. Jet diffusion versus jetgpt–modern networks for the lhc. arXiv preprint arXiv:2305.10475, 2023.
  17. Score-based generative models for calorimeter shower simulation. Phys. Rev. D, 106:092009, Nov 2022. doi: 10.1103/PhysRevD.106.092009. URL https://link.aps.org/doi/10.1103/PhysRevD.106.092009.
  18. Caloclouds: Fast geometry-independent highly-granular calorimeter simulation. arXiv preprint arXiv:2305.04847, 2023.
  19. Comparison of point cloud and image-based models for calorimeter fast simulation. arXiv preprint arXiv:2307.04780, 2023.
  20. Towards designing and exploiting generative networks for neutrino physics experiments using liquid argon time projection chambers. arXiv preprint arXiv:2204.02496, 2022.
  21. Deep generative modelling: A comparative review of vaes, gans, normalizing flows, energy-based and autoregressive models. IEEE transactions on pattern analysis and machine intelligence, 2021.
  22. Diffusion models: A comprehensive survey of methods and applications, 2023.
  23. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=PxTIG12RRHS.
  24. Improved analysis of score-based generative modeling: User-friendly bounds under minimal smoothness assumptions. In International Conference on Machine Learning, pages 4735–4763. PMLR, 2023.
  25. Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. In The Eleventh International Conference on Learning Representations, 2022.
  26. Valentin De Bortoli. Convergence of denoising diffusion models under the manifold hypothesis. Transactions on Machine Learning Research, 2022.
  27. Fast sampling of diffusion models with exponential integrator. In The Eleventh International Conference on Learning Representations, 2022.
  28. Brian DO Anderson. Reverse-time diffusion equation models. Stochastic Processes and their Applications, 12(3):313–326, 1982.
  29. Time reversal of markov processes with jumps under a finite entropy condition. Stochastic Processes and their Applications, 144:85–124, 2022. ISSN 0304-4149. doi: https://doi.org/10.1016/j.spa.2021.10.002. URL https://www.sciencedirect.com/science/article/pii/S0304414921001654.
  30. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005.
  31. Pilarnet: public dataset for particle imaging liquid argon detectors in high energy physics. arXiv preprint arXiv:2006.01993, 2020.
  32. Learning multiple layers of features from tiny images. 2009.
  33. Deep neural network for pixel-level electromagnetic particle identification in the microboone liquid argon time projection chamber. Physical Review D, 99(9):092001, 2019.
  34. Semantic segmentation with a sparse convolutional neural network for event reconstruction in microboone. Physical Review D, 103(5):052012, 2021.
  35. Search for an anomalous excess of charged-current quasielastic ν𝜈\nuitalic_ν e interactions with the microboone experiment using deep-learning-based reconstruction. Physical Review D, 105(11):112003, 2022.
  36. A kernel two-sample test. Journal of Machine Learning Research, 13(25):723–773, 2012. URL http://jmlr.org/papers/v13/gretton12a.html.
  37. Learning generative models with sinkhorn divergences. In International Conference on Artificial Intelligence and Statistics, pages 1608–1617. PMLR, 2018.
  38. Statistical bounds for entropic optimal transport: sample complexity and the central limit theorem. Advances in Neural Information Processing Systems, 32, 2019.
  39. On the empirical estimation of integral probability metrics. Electronic Journal of Statistics, 6(none):1550 – 1599, 2012. doi: 10.1214/12-EJS722. URL https://doi.org/10.1214/12-EJS722.
  40. Sample complexity of sinkhorn divergences. In Kamalika Chaudhuri and Masashi Sugiyama, editors, Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89 of Proceedings of Machine Learning Research, pages 1574–1583. PMLR, 16–18 Apr 2019. URL https://proceedings.mlr.press/v89/genevay19a.html.
  41. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
  42. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2818–2826, 2016.
  43. Metric space of collider events. Phys. Rev. Lett., 123:041801, Jul 2019. doi: 10.1103/PhysRevLett.123.041801. URL https://link.aps.org/doi/10.1103/PhysRevLett.123.041801.
  44. Understanding ddpm latent codes through optimal transport. arXiv preprint arXiv:2202.07477, 2022.
  45. Diffusion schrödinger bridge with applications to score-based generative modeling. Advances in Neural Information Processing Systems, 34:17695–17709, 2021.
  46. Genphys: From physical processes to generative models. arXiv preprint arXiv:2304.02637, 2023.
Citations (6)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.