Kaiwen Zhou


Hi there! I am a research scientist at Huawei Noah's Ark Lab.

Prior to joining Huawei, I completed my MPhil and PhD in computer science and engineering at the Chinese University of Hong Kong in 2019 and 2022, respectively. I obtained my BSc degree in computer science and technology at Fudan University in 2017.

I am currently interested in research topics related to Optimization for Machine Learning and Out-of-Distribution Generalization.

If you are interested in internship or full-time opportunities at Huawei Noah's Ark Lab, please drop me an email.

Email: kwzhou [at] cse (dot) cuhk (dot) edu (dot) hk
Office: Huawei Industrial Base F3, Bantian, Shenzhen, China.

scholar   dblp   openreview

Publications

 (*indicates equal contribution)


  1. Enhancing Neural Subset Selection: Integrating Background Information into Set Representations
    Binghui Xie, Yatao Bian, Kaiwen Zhou, Yongqiang Chen, Peilin Zhao, Bo Han, Wei Meng, James Cheng.
    In International Conference on Learning Representations (ICLR), 2024.
  2. Enhancing Evolving Domain Generalization through Dynamic Latent Representations [paper]
    Binghui Xie, Yongqiang Chen, Jiaqi Wang, Kaiwen Zhou, Bo Han, Wei Meng, James Cheng.
    In AAAI Conference on Artificial Intelligence (AAAI), Oral 2.2%, 2024.
  3. Understanding and Improving Composite Bayesian Optimization
    Kaiwen Zhou*, Binghui Xie*, Junlong Lyu, Zhitang Chen.
    In AAAI Workshop on Learnable Optimization (AAAI LEANOPT), 2024.
  4. Understanding and Improving Feature Learning for Out-of-Distribution Generalization [paper] [github]
    Yongqiang Chen*, Wei Huang*, Kaiwen Zhou*, Yatao Bian, Bo Han, James Cheng.
    In Advances in Neural Information Processing Systems (NeurIPS), 2023;
    Also appeared in ICLR Workshop on Domain Generalization (Spotlight), 2023.
    And ICML Workshop on Spurious Correlations, Invariance and Stability, 2023.
  5. Does Invariant Graph Learning via Environment Augmentation Learn Invariance? [paper] [github]
    Yongqiang Chen, Yatao Bian, Kaiwen Zhou, Binghui Xie, Bo Han, James Cheng.
    In Advances in Neural Information Processing Systems (NeurIPS), 2023;
    Also appeared in ICLR Workshop on Domain Generalization (Spotlight), 2023.
  6. Pareto Invariant Risk Minimization: Towards Mitigating The Optimization Dilemma in Out-of-Distribution Generalization [paper] [github]
    Yongqiang Chen, Kaiwen Zhou, Yatao Bian, Binghui Xie and others.
    In International Conference on Learning Representations (ICLR), 2023;
    Also appeared in ICLR Workshop on Domain Generalization (Oral), 2023;
    And ICML Workshop on Principles of Distribution Shift, 2022.
  7. A Novel Extrapolation Technique to Accelerate WMMSE [paper]
    Kaiwen Zhou, Zhilin Chen, Guochen Liu, Zhitang Chen.
    In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023.
  8. On the Finite-Time Complexity and Practical Computation of Approximate Stationarity Concepts of Lipschitz Functions [paper]
    Lai Tian, Kaiwen Zhou, Anthony Man-Cho So.
    In International Conference on Machine Learning (ICML), 2022.
  9. Fast and Reliable Evaluation of Adversarial Robustness with Minimum-Margin Attack [paper]
    Ruize Gao, Jiongxiao Wang, Kaiwen Zhou, Feng Liu, Binghui Xie, Gang Niu, Bo Han, James Cheng.
    In International Conference on Machine Learning (ICML), 2022.
  10. Accelerating Perturbed Stochastic Iterates in Asynchronous Lock-Free Optimization [paper]
    Kaiwen Zhou, Anthony Man-Cho So, James Cheng.
    In NeurIPS Workshop on Optimization for Machine Learning (NeurIPS OPT), 2022.
  11. Practical Schemes for Finding Near-Stationary Points of Convex Finite-Sums [paper]
    Kaiwen Zhou, Lai Tian, Anthony Man-Cho So, James Cheng.
    In International Conference on Artificial Intelligence and Statistics (AISTATS), 2022.
  12. Boosting First-Order Methods by Shifting Objective: New Schemes with Faster Worst-Case Rates [paper]
    Kaiwen Zhou, Anthony Man-Cho So, James Cheng.
    In Advances in Neural Information Processing Systems (NeurIPS), 2020.
  13. Amortized Nesterov's Momentum: A Robust Momentum and Its Application to Deep Learning [paper] [code]
    Kaiwen Zhou, Yanghua Jin, Qinghua Ding, James Cheng.
    In Conference on Uncertainty in Artificial Intelligence (UAI), 2020.
  14. Tight Convergence Rate of Gradient Descent for Eigenvalue Computation [paper]
    Qinghua Ding, Kaiwen Zhou, James Cheng.
    In International Joint Conference on Artificial Intelligence (IJCAI), 2020.
  15. Convolutional Embedding for Edit Distance [paper] [github]
    Xinyan Dai, Xiao Yan, Kaiwen Zhou, Yuxuan Wang, Han Yang, James Cheng.
    In International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), 2020.
  16. Direct Acceleration of SAGA using Sampled Negative Momentum [paper]
    Kaiwen Zhou, Qinghua Ding, Fanhua Shang, James Cheng, Danli Li, Zhi-Quan Luo.
    In International Conference on Artificial Intelligence and Statistics (AISTATS), 2019.
  17. A Simple Stochastic Variance Reduced Algorithm with Fast Convergence Rates [paper]
    Kaiwen Zhou, Fanhua Shang, James Cheng.
    In International Conference on Machine Learning (ICML), 2018.
  18. Guaranteed Sufficient Decrease for Stochastic Variance Reduced Gradient Optimization [paper]
    Fanhua Shang, Yuanyuan Liu, Kaiwen Zhou, James Cheng, Kelvin Kai Wing Ng, Yuichi Yoshida.
    In International Conference on Artificial Intelligence and Statistics (AISTATS), 2018.
  19. VR-SGD: A Simple Stochastic Variance Reduction Method for Machine Learning [paper]
    Fanhua Shang, Kaiwen Zhou, Hongying Liu, James Cheng, Ivor Tsang, Lijun Zhang, Dacheng Tao, Licheng Jiao.
    In IEEE Transactions on Knowledge and Data Engineering (TKDE), 2018.
  20. ASVRG: Accelerated Proximal SVRG [paper]
    Fanhua Shang, Licheng Jiao, Kaiwen Zhou, James Cheng, Yan Ren, Yufei Jin.
    In Asian Conference on Machine Learning (ACML), 2018.

Preprints

 (*indicates equal contribution)


  1. Positional Information Matters for Invariant In-Context Learning: A Case Study of Simple Function Classes [paper]
    Yongqiang Chen, Binghui Xie, Kaiwen Zhou, Bo Han, Yatao Bian, James Cheng.
  2. Efficient Private SCO for Heavy-Tailed Data via Clipping [paper]
    Chenhan Jin, Kaiwen Zhou, Bo Han, Ming-Chang Yang, James Cheng.
  3. An Adaptive Incremental Gradient Method With Support for Non-Euclidean Norms [paper]
    Binghui Xie*, Chenhan Jin*, Kaiwen Zhou, James Cheng, Wei Meng.
  4. Local Reweighting for Adversarial Training [paper]
    Ruize Gao*, Feng Liu*, Kaiwen Zhou, Gang Niu, Bo Han, James Cheng.
  5. Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning [paper] [github]
    Xinyan Dai, Xiao Yan, Kaiwen Zhou, Han Yang, Kelvin Kai Wing Ng, James Cheng, Yu Fan.
  6. Norm-Range Partition: A Universal Catalyst for LSH based Maximum Inner Product Search (MIPS) [paper]
    Xiao Yan, Xinyan Dai, Jie Liu, Kaiwen Zhou, James Cheng.

Theses


  • Fast, Practical and Scalable First-Order Methods for Modern Machine Learning Problems [paper]
    The Chinese University of Hong Kong, PhD, 2022.
  • Accelerating Finite-sum Convex Optimization and Highly-smooth Convex Optimization [paper]
    Outstanding Thesis Award of Faculty of Engineering (Sole Winner)
    The Chinese University of Hong Kong, MPhil, 2019.

Academic Service


  • Conference refereeing: ICML 2021/22/23/24, NeurIPS 2021/22/23, AISTATS 2022/23/24, ICLR 2021/22/23/24.
  • Journal refereeing: Mathematical Programming, SIAM Journal on Optimization, Optimization Methods and Software, Mathematical Programming Computation, Transactions on Machine Learning Research.

Last update: January 17, 2024.