In recent years, with the advancement of multimodal foundation models (MMFMs), there has been a growing interest in enhancing their generalization abilities through continual learning (CL) to process diverse data types, from text to visuals, and continuously update their capabilities based on real-time inputs. Despite significant advancements in theoretical research and applications of continual learning, the community remains confronted with serious challenges. Our workshop aims to provide a venue where academic researchers and industry practitioners can come together to discuss the principles, limitations, and applications of multimodal foundation models in continual learning for multimedia applications and promote the understanding of multimodal foundation models in continual learning, innovative algorithms, and research on new multimodal technologies and applications.
Scope and Topics
Interested topics will include, but not be limited to:
The proposed workshop will include there invited talks and four paper presentations. The workshop will be considered for a whole-day meeting.
Time |
Programme |
08:30-08:35 |
Opening Remarks |
08:35-09:15Keynote Speaker |
Keynote: [TBD]![]() |
09:15-09:55Keynote Speaker |
Keynote: [TBD]![]() |
09:55-10:35Keynote Speaker |
Keynote: [TBD]![]() |
10:35-11:00 |
Morning Tea |
11:00-11:10 |
SR-ML: A Sequence-level Routing with Mixed Low-rank Experts Framework for Continual Learning |
11:10-11:20 |
Low Altitude-R1: Exploring the Upper Limits of Target Detection in Low-altitude Scenarios with Reinforcement Learning |
11:20-11:30 |
LaST-LoRA: Adaptive Knowledge Reuse and Latent Subspace Tracking for Continual Learning |
11:30-11:40 |
NAS-LoRA: Empowering Parameter-Efficient Fine-Tuning for Visual Foundation Models with Searchable Adaptation |
11:40-12:00 |
Panel & Closing Remarks |
Contact the Organizing Committee: woods.cl.acm.mm@gmail.com