This survey samples from the ever-growing family of adaptive resonance theory (ART) neural network models used to perform the three primary machine learning modalities, namely, unsupervised, supervised and reinforcement learning. It comprises a representative list from classic to modern ART models, thereby painting a general picture of the architectures developed by researchers over the past 30 years. The learning dynamics of these ART models are briefly described, and their distinctive characteristics such as code representation, long-term memory and corresponding geometric interpretation are discussed. Useful engineering properties of ART (speed, configurability, explainability, parallelization and hardware implementation) are examined along with current challenges. Finally, a compilation of online software libraries is provided. It is expected that this overview will be helpful to new and seasoned ART researchers. 7 Code repositories 54 8 Conclusions 55 2 Compute activation function(s): Tj = fT (x, θj, α), ∀j ∈ C. 3 Perform WTA competition: J = arg max j∈Λ (Tj). 4 Compute match function(s):7 Update category J: θ new J = fL(x, θ old J , β). 8 else 9 Deactivate category J: Λ ← Λ − {J}. 10 if Λ = {∅} then 11 Go to step 3. 12 else 13 Set J = |C| + 1. 14 Create new category: C ← C ∪ {J}. 15 Initialize new category: θ new J = fN (x, λ). 16 Set output: y where the vigilance criterion is M J ≥ ρ.Learning. If the winning category satisfies the vigilance test, then resonance occurs, and the radius R J and centroid m J of the winning node are updated as follows:where ρ ∈ (0, 1] is the vigilance parameter.Learning. If the winning category J satisfies M J ≥ ρ, then resonance occurs, and it is updated as follows:J J needs to be updated after the presentation of each sample, then γ k (new) can be estimated incrementally. Particularly, when there is a resonant committed node J, if x k is a meta-information feature, thenotherwise,