Hopfield Neural Networks: A updated Approach for Using Associative Memory to Improve Matrices
Metadata Field | Value | Language |
---|---|---|
dc.contributor.advisor | Han, Xiaoyin | |
dc.contributor.author | Bao, Yida | |
dc.date.accessioned | 2024-07-31T19:21:52Z | |
dc.date.available | 2024-07-31T19:21:52Z | |
dc.date.issued | 2024-07-31 | |
dc.identifier.uri | https://etd.auburn.edu//handle/10415/9426 | |
dc.description.abstract | Modern Hopfield Neural Networks (MHNNs) are a class of neural networks renowned for their associative memory capabilities, which have broad applications in pattern recognition, optimization, and error correction. This dissertation explores the mathematical foundations, architectural nuances, and practical applications of MHNNs, focusing on the development of a novel energy function using the Concave-Convex Procedure (CCCP). The updated energy function enhances the network's convergence properties and robustness, addressing limitations of classical models such as low storage capacity and susceptibility to local minima.We demonstrate the efficacy of the update energy function through rigorous theoretical analysis and extensive simulation studies. Synthetic datasets with various distributional properties are generated to evaluate the network's performance in classification tasks. Our results indicate that the proposed MHNN model outperforms Original Hopfield networks and competes effectively with contemporary machine learning algorithms, including Support Vector Machines, Decision Tree, Random Forests, and Convolutional Neural Networks.In classifc Hopfield nerual network practical applications, we focus on image restoration tasks, successfully reconstructing highly corrupted images, and achieving high accuracy and low restoration error. This showcases the network's potential for real-world applications in fields such as bioinformatics, natural language processing, and healthcare diagnostics.This work not only underscores the capabilities of MHNNs in associative memory and optimization tasks but also paves the way for future innovations. By integrating advanced mathematical techniques and exploring hybrid approaches, this dissertation contributes significantly to the field of neural networks and machine learning, providing a robust framework for future research and application development. | en_US |
dc.rights | EMBARGO_NOT_AUBURN | en_US |
dc.subject | Mathematics and Statistics | en_US |
dc.title | Hopfield Neural Networks: A updated Approach for Using Associative Memory to Improve Matrices | en_US |
dc.type | PhD Dissertation | en_US |
dc.embargo.length | MONTHS_WITHHELD:36 | en_US |
dc.embargo.status | EMBARGOED | en_US |
dc.embargo.enddate | 2027-07-31 | en_US |
dc.contributor.committee | Gaillard, Philippe | |
dc.contributor.committee | Werner Van Wyk, Hans | |
dc.contributor.committee | Zheng, Jingyi | |
dc.creator.orcid | 0000-0002-7789-2735 | en_US |