A Fast Neural Architecture Search Method for Multi-Modal Classification via Knowledge Sharing
A Fast Neural Architecture Search Method for Multi-Modal Classification via Knowledge Sharing
Zhihua Cui, Shiwu Sun, Qian Guo, Xinyan Liang, Yuhua Qian, Zhixia Zhang
Proceedings of the Thirty-Fourth International Joint Conference on Artificial Intelligence
Main Track. Pages 5003-5011.
https://doi.org/10.24963/ijcai.2025/557
Neural architecture search-based multi-modal classification (NAS-MMC) aims to automatically find optimal network structures for improving the multi-modal classification performance. However, most current NAS-MMC methods are quite time-consuming during the training process. In this paper, we propose a knowledge sharing-based neural architecture search (KS-NAS) method for multi-modal classification. The KS-NAS optimizes the search process by introducing a dynamically updated knowledge base to reduce the consumption of computational resource. Specifically, during the deep evolutionary search, individuals in the initial population acquire initial parameters from a knowledge base, and then undergo training and optimization until convergence is reached, avoiding the need for training from scratch. The knowledge base is dynamically updated by aggregating the parameters of high-quality individuals trained within the population, thus progressively improving the quality of the knowledge base. As the population evolves, the knowledge base continues to optimize, ensuring that subsequent individuals can obtain higher-quality initialization parameters, which significantly accelerates the training speed of the population. Experimental results show that the KS-NAS method achieves state-of-the-art results in terms of classification performance and training efficiency across multiple popular multi-modal tasks.
Keywords:
Machine Learning: ML: Multi-modal learning
Machine Learning: ML: Classification
Machine Learning: ML: Multi-view learning
