publications

publications by categories in reversed chronological order. generated by jekyll-scholar.

2025

  1. TPAMI
    smile_upscaling.png
    Zero-Shot Sparse Mixture of Low-Rank Experts Construction From Pre-Trained Foundation Models
    Anke Tang, Li Shen, Yong Luo, and 5 more authors
    IEEE Transactions on Pattern Analysis and Machine Intelligence, 2025
  2. IJCV
    Data-adaptive weight-ensembling for multi-task model fusion
    Anke Tang, Li Shen, Yong Luo, and 4 more authors
    International Journal of Computer Vision, 2025
  3. NeurIPS
    opcm.png
    Merging models on the fly without retraining: A sequential approach to scalable continual model merging
    Anke Tang, Enneng Yang, Li Shen, and 4 more authors
    The Thirty-Ninth Annual Conference on Neural Information Processing Systems, 2025
  4. NeurIPS
    Continual Model Merging without Data: Dual Projections for Balancing Stability and Plasticity
    Enneng Yang, Anke Tang, Li Shen, and 4 more authors
    The Thirty-Ninth Annual Conference on Neural Information Processing Systems, 2025
  5. NeurIPS
    Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging
    Jinluan Yang, Dingnan Jin, Anke Tang, and 10 more authors
    The Thirty-Ninth Annual Conference on Neural Information Processing Systems, 2025
  6. ICML
    Targeted Low-rank Refinement: Enhancing Sparse Language Models with Precision
    Li Shen, Anke Tang, Yong Luo, and 3 more authors
    In Forty-second International Conference on Machine Learning, 2025
  7. ICML
    Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent
    Yongxian Wei, Anke Tang, Li Shen, and 3 more authors
    In Forty-second International Conference on Machine Learning, 2025
  8. ICLR
    Mitigating the Backdoor Effect for Multi-Task Model Merging via Safety-Aware Subspace
    Jinluan Yang, Anke Tang, Didi Zhu, and 3 more authors
    In The 13th International Conference on Learning Representations (ICLR), 2025
  9. NMI
    Learning from models beyond fine-tuning
    Hongling Zheng, Li Shen, Anke Tang, and 5 more authors
    Nature Machine Intelligence, 2025

2024

  1. Fusionbench: A comprehensive benchmark of deep model fusion
    Anke Tang, Li Shen, Yong Luo, and 3 more authors
    arXiv preprint arXiv:2406.03280, 2024
  2. Towards efficient pareto set approximation via mixture of experts based model fusion
    Anke Tang, Li Shen, Yong Luo, and 3 more authors
    arXiv preprint arXiv:2406.09770, 2024
  3. Efficient and effective weight-ensembling mixture of experts for multi-task model merging
    Li Shen, Anke Tang, Enneng Yang, and 6 more authors
    arXiv preprint arXiv:2410.21804, 2024
  4. ICML
    wemoe.png
    Merging Multi-Task Models via Weight-Ensembling Mixture of Experts
    Anke Tang, Li Shen, Yong Luo, and 3 more authors
    In The 41th International Conference on Machine Learning (ICML), 2024
  5. ICLR
    Parameter efficient multi-task model fusion with partial linearization
    Anke Tang, Li Shen, Yong Luo, and 5 more authors
    In the 12th International Conference on Learning Representations, 2024

2023

  1. Concrete subspace learning based interference elimination for multi-task model fusion
    Anke Tang, Li Shen, Yong Luo, and 4 more authors
    arXiv preprint arXiv:2312.06173, 2023
  2. IJCAI
    Improving Heterogeneous Model Reuse by Density Estimation
    Anke Tang, Yong Luo, Han Hu, and 5 more authors
    In Thirty-Second International Joint Conference on Artificial Intelligence, 2023