Back
He joins the ELLIS Institute Tübingen as PI and Hector Endowed Fellow and has a co-affiliation with the MPI-IS and the Tübingen AI Center as Independent Research Group Leader.
Shiwei Liu will start his new group on July 15. His WEI Lab (which stands for Wild, Efficient, and Innovative AI) will focus on empirically understanding the behavior of deep neural networks, developing deep learning algorithms and architectures that learn better, faster, and cheaper. One central theme of Shiwei’s research is to leverage, understand, and expand the role of low-dimentionality in neural networks, whose impacts span many important topics, such as efficient training/inference/scaling of large-foundation models, robustness and trustworthiness, and generative AI.
Shiwei Liu is a Royal Society Newton International Fellow at University of Oxford. He was a Postdoctoral Fellow at the University of Texas at Austin. He obtained his Ph.D. with Cum Laude from Eindhoven University of Technology in 2022. Liu has received two Rising Star Awards from the King Abdullah University of Science and Technology (KAUST) in Saudi Arabia and the Conference on Parsimony and Learning (CPAL). His Ph.D. thesis received the 2023 Best Dissertation Award from Informatics Europe.
In March 2024, Shiwei Liu gave a talk at the ELLIS Institute Scientific Symposium, held at MPI-IS, talking about sparsity in neural networks. While existing research predominantly focuses on exploiting sparsity for model compression—such as deriving sparse neural networks from pre-trained dense ones—many other promising benefits such as scalability, robustness, and fairness remain under-explored. His talk delved into these overlooked advantages. Specifically, He showcased how sparsity can boost the scalability of neural networks by efficiently training sparse models from scratch. This approach enables a significant increase in model capacity without proportionally escalating computational or memory requirements. Additionally, he explored the future implications of sparsity in the realm of large language models, discussing its potential benefits to efficient LLM scaling, lossless LLM compression, and fostering trustworthy AI.
More information