Bhaskar Tripathi
project owner
EA for training transformers and other DNNs
We propose to use evolutionary methods (EMs) like CMA-ES, PSO, GWO, Firefly Algorithm, and ACO to optimize deep neural network (DNN) weights and architectures, focusing on optimizing node weights and evolving architectures—including Transformers and large language models (LLMs). Our team of experienced researchers will conduct rigorous experiments comparing EMs with backpropagation on standard datasets (CIFAR-10, MNIST, WikiText-2), assessing performance, convergence speed, computational cost, and scalability. We will document our methodology, provide code in an open-source repository, and demonstrate our work within a Hyperon instance.
Join the Discussion (0)
Please create account or login to post comments.