Skip to content

Latest commit

 

History

History
159 lines (135 loc) · 8.37 KB

paper.md

File metadata and controls

159 lines (135 loc) · 8.37 KB
title tags authors affiliations date bibliography
Adaptive Resonance Lib: A Python package for Adaptive Resonance Theory (ART) models
python
clustering
classification
regression
reinforcement learning
machine learning
name corresponding orcid affiliation
Niklas M. Melton
true
0000-0001-9625-7086
1
name orcid affiliation
Dustin Tanksley
0000-0002-1677-0304
1
name orcid affiliation
Donald C. Wunsch II
0000-0002-9726-9051
1
name index ror
Missouri University of Science and Technology, USA
1
00scwqd12
18 October 2024
references.bib

Summary

The Adaptive Resonance Theory Library (artlib) is a Python library that implements a wide range of Adaptive Resonance Theory (ART) algorithms. artlib currently supports eight elementary ART models and 11 compound ART models, including Fuzzy ART [@carpenter1991fuzzy], Hypersphere ART [@anagnostopoulos2000hypersphere], Ellipsoid ART [@anagnostopoulos2001a; @anagnostopoulos2001b], Gaussian ART [@williamson1996gaussian], Bayesian ART [@vigdor2007bayesian], Quadratic Neuron ART [@su2001application; @su2005new], ART1 [@carpenter1987massively], ART2 [@carpenter1987art; @carpenter1991art], ARTMAP [@carpenter1991artmap], Simplified ARTMAP [@gotarredona1998adaptive], SMART [@bartfai1994hierarchical], TopoART [@tscherepanow2010topoart], Dual Vigilance ART [@da2019dual], CVIART [@da2022icvi], BARTMAP [@xu2011bartmap; @xu2012biclustering], Fusion ART [@tan2007intelligence], FALCON [@tan2004falcon], and TD-FALCON [@tan2008integrating]. These models can be applied to tasks such as unsupervised clustering, supervised classification, regression, and reinforcement learning [@da2019survey]. This library provides an extensible and modular framework where users can integrate custom models or extend current implementations, allowing for experimentation with existing and novel machine learning techniques.

In addition to the diverse set of ART models, artlib offers implementations of visualization methods for various cluster geometries, along with pre-processing techniques such as Visual Assessment of Tendency (VAT) [@bezdek2002vat], data normalization, and complement coding.

Adaptive Resonance Theory (ART)

ART is a class of neural networks known for solving the stability-plasticity dilemma, making it particularly effective for classification, clustering, and incremental learning tasks [@grossberg1976a; @grossberg1976a; @Grossberg1980HowDA; @grossberg2013adaptive; @da2019survey]. ART models are designed to dynamically learn and adapt to new patterns without catastrophic forgetting, making them ideal for real-time systems requiring continuous learning.

Over the years, dozens of ART variations have been published [@da2019survey], extending the applicability of ART to nearly all learning regimes, including reinforcement learning [@tan2004falcon; @tan2008integrating], hierarchical clustering [@bartfai1994hierarchical], topological clustering [@tscherepanow2010topoart], and biclustering [@xu2011bartmap; @xu2012biclustering]. These numerous models provide an ART-based solution for most machine learning use cases. However, the rapid pace of bespoke model development, coupled with the challenges students face in learning ART's foundational principles, has contributed to a scarcity of open-source, user-friendly implementations for most ART variants.

The ability of ART to preserve previously learned patterns while learning new data in real-time has made it a powerful tool in domains such as robotics, medical diagnosis, and adaptive control systems. artlib aims to extend the application of these models in modern machine learning pipelines, offering a unique and approachable toolkit for leveraging ART's strengths.

Statement of Need

The Adaptive Resonance Library (artlib) is essential for researchers, developers, and educators interested in adaptive neural networks, specifically ART algorithms. While deep learning dominates machine learning, ART models offer unique advantages in incremental and real-time learning environments due to their ability to learn new data without forgetting previously learned information.

Currently, no comprehensive Python library implements a variety of ART models in an open-source, modular, and extensible manner. artlib fills this gap by offering a range of ART implementations that integrate seamlessly with machine learning workflows, including scikit-learn's Pipeline and GridSearchCV [@scikit-learn]. The library is designed for ease of use and high performance, leveraging Python's scientific stack (NumPy [@harris2020array], SciPy [@2020SciPy-NMeth], and scikit-learn [@scikit-learn]) for fast numerical computation.

The modular design of artlib enables users to create novel compound ART models, such as Dual Vigilance Fusion ART [@da2019dual; @tan2007intelligence] or Quadratic Neuron SMART [@su2001application; @su2005new; @bartfai1994hierarchical]. This flexibility offers powerful experimental and time-saving benefits, allowing researchers and practitioners to evaluate models on diverse datasets efficiently.

Additionally, the library serves as a valuable educational tool, providing well-documented code and familiar APIs to support hands-on experimentation with ART models. It is ideal for academic courses or personal projects in artificial intelligence and machine learning, making artlib a versatile resource.

artlib is actively maintained and designed for future extension, allowing users to create new ART models, adjust parameters for specific applications, and explore ART's potential for novel research problems. Its integration with popular Python libraries ensures its adaptability to current machine learning challenges.

Comparison to Existing Implementations

While there are several open-source repositories that provide python implementations of specific ART models [@birkjohann2023artpython; @aiopenlab2023art; @dilekman2022artificial; @artpy2022; @dixit2020adaptive; @inyoot2021art; @valixandra2021adaptive; @wan2022art2; @ray2023artpy], they lack modularity and are limited in scope, often implementing just one or two models. For instance, MATLAB-based ART toolboxes [@mathworks_art1s; @mathworks_fuzzyart_fuzzyartmap; @mathworks_topoart; @mathworks_art_fuzzyart_artmap] provide implementations of Fuzzy ART, TopoART, ART1, and ARTMAP models, but they lack the flexibility and modularity required for broader experimentation. The most significant existing ART implementation exists in julia and provides just five models [@Petrenko_AdaptiveResonance_jl_A_Julia_2022] but, like the previously listed MATLAB-based toolboxes, it is not easily accessible to Python-based work flows and lacks a modular design.

These existing implementations of ART models may provide standalone versions of individual models, but they are often not designed to integrate seamlessly with modern Python libraries such as scikit-learn, NumPy, and SciPy. As a result, researchers and developers working in Python-based environments face challenges when trying to incorporate ART models into their machine learning pipelines.

In contrast, artlib offers a comprehensive and modular collection of ART models, including both elementary and compound ART architectures. It is designed for interoperability with popular Python tools, enabling users to easily integrate ART models into machine learning workflows, optimize models using scikit-learn's GridSearchCV, and preprocess data using standard libraries. Further, artlib provides users the flexibility to construct their own compound ART modules (those art modules deriving properties from other, elementary modules) which may or may not exist in published literature. artlib also provides a template in the source code to encourage users to develop and experiment with their own custom ART algorithms. This flexibility and integration make artlib a powerful resource for both research and practical applications.

Acknowledgements

This research was supported by the National Science Foundation (NSF) under Award Number 2420248. The project titled EAGER: Solving Representation Learning and Catastrophic Forgetting with Adaptive Resonance Theory provided essential funding for the completion of this work.

We would also like to thank Gail Carpenter and Stephen Grossberg for their feedback regarding this project and their immense contribution to machine learning by pioneering Adaptive Resonance Theory.

References