In this work, we explore "prompt tuning," a simple yet effective mechanism for learning "soft prompts" to condition frozen language models to perform specific downstream tasks. Unlike the discrete text prompts used by GPT-3, soft prompts are learned through backpropagation and can be tuned to incorporate signals from any number of labeled examples. Our end-to-end learned approach outperforms GPT-3's few-shot learning by a large margin. More remarkably, through ablations on model size using T5, we show that prompt tuning becomes more competitive with scale: as models exceed billions of parameters, our method "closes the gap" and matches the strong performance of model tuning (where all model weights are tuned). This finding is especially relevant because large models are costly to share and serve and the ability to reuse one frozen model for multiple downstream tasks can ease this burden. Our method can be seen as a simplification of the recently proposed "prefix tuning" of Li and Liang (2021) and we provide a comparison to this and other similar approaches. Finally, we show that conditioning a frozen model with soft prompts confers benefits in robustness to domain transfer and enables efficient "prompt ensembling." We release code and model checkpoints to reproduce our experiments. 1
We report cooling of a single neutral atom to its three-dimensional vibrational ground state in an optical tweezer. After employing Raman sideband cooling for tens of milliseconds, we measure via sideband spectroscopy a three-dimensional ground-state occupation of about 90%. We further observe coherent control of the spin and motional state of the trapped atom. Our demonstration shows that an optical tweezer, formed simply by a tightly focused beam of light, creates sufficient confinement for efficient sideband cooling. This source of ground-state neutral atoms will be instrumental in numerous quantum simulation and logic applications that require a versatile platform for storing and manipulating ultracold single neutral atoms. For example, these results will improve current optical-tweezer experiments studying atom-photon coupling and Rydberg quantum logic gates, and could provide new opportunities such as rapid production of single dipolar molecules or quantum simulation in tweezer arrays.
In this work, we explore "prompt tuning", a simple yet effective mechanism for learning "soft prompts" to condition frozen language models to perform specific downstream tasks. Unlike the discrete text prompts used by GPT-3, soft prompts are learned through backpropagation and can be tuned to incorporate signal from any number of labeled examples. Our end-to-end learned approach outperforms GPT-3's "few-shot" learning by a large margin. More remarkably, through ablations on model size using T5, we show that prompt tuning becomes more competitive with scale: as models exceed billions of parameters, our method "closes the gap" and matches the strong performance of model tuning (where all model weights are tuned). This finding is especially relevant in that large models are costly to share and serve, and the ability to reuse one frozen model for multiple downstream tasks can ease this burden. Our method can be seen as a simplification of the recently proposed "prefix tuning" of Li and Liang (2021), and we provide a comparison to this and other similar approaches. Finally, we show that conditioning a frozen model with soft prompts confers benefits in robustness to domain transfer, as compared to full model tuning.
The quantum statistics of atoms is typically observed in the behavior of an ensemble via macroscopic observables. However, quantum statistics modifies the behavior of even two particles. Here, we demonstrate near-complete control over all the internal and external degrees of freedom of two laser-cooled (87)Rb atoms trapped in two optical tweezers. This controllability allows us to observe signatures of indistinguishability via two-particle interference. Our work establishes laser-cooled atoms in optical tweezers as a promising route to bottom-up engineering of scalable, low-entropy quantum systems.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.