Diffraction phase-free Bragg atom interferometry
Abstract: Bragg diffraction is a fundamental technique used to enhance the sensitivity of atom interferometers through large momentum transfer, making these devices among the most precise quantum sensors available today. To further improve their accuracy, it is necessary to achieve control over multiple interferometer paths and increase robustness against velocity spread. Optimal control theory has recently led to advancements in sensitivity and robustness under specific conditions, such as vibrations, accelerations, and other experimental challenges. In this work, we employ this tool to focus on improving the accuracy of the interferometer by minimizing the diffraction phase. We consider the finite temperature of the incoming wavepacket and the multi-path nature of high-order Bragg diffraction as showcased in a Mach-Zehnder(MZ) geometry. Our approach can achieve diffraction phases on the order of microradians or even below a microradian for a momentum width of the incoming wavepacket $\sigma_p = 0.01\hbar k$, below a milliradian for $\sigma_p= 0.1 \hbar k$ and milliradians for $\sigma_p = 0.3 \hbar k$.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.