Zifeng Wang

I am a research scientist at Google. I received my PhD in Machine Learning at SPIRAL Group from Northeastern University, advised by Prof. Jennifer G. Dy. During my PhD, I also work closely with Prof. Stratis Ioannidis and Prof. Yanzhi Wang. I received my BS degree in Electronic Engineering from Tsinghua University. In my college years, I also worked with Prof. Jiwen Lu (Tsinghua), Prof. Jia Deng (Princeton) on computer vision and Prof. Yong Li (Tsinghua) on big data.

I am looking for self-motivated student researchers / research interns with interests and expertise in LLMs! Feel free contact me if you would like to work on cutting-edge challenges in machine learning and LLMs.

Email  /  Google Scholar  /  LinkedIn  /  Twitter  /  Resume  /  CV

profile photo
Research

  • Large language model (LLM) Alignment
  • Continual (Lifelong) learning
  • Open set recognition / novel class discovery
  • Adversarial robustness and model compression
  • Deep learning applications in computer vision, natural language understanding, document understanding, healthcare, wireless communications, etc.

News

  • 07/2023: I will be giving a talk on effective and efficient continual learning at A*STAR , thanks for hosting!
  • 05/2023: Our paper on document entity extraction is accepted at ACL 2023 (Findings)
  • 04/2022: I received the Outstanding Student Award in Research from COE, Northeastern University
  • 04/2023: Our paper on continual learning is accepted at ICML 2023 , code will be released soon!
  • 02/2023: I am glad to give talks about efficient and sparse continual learning at ContinualAI and AI Time
  • 10/2022: I received the Scholar Award for NeurIPS 2022
  • 09/2022: Our paper on efficient continual learning is accepted at NeurIPS 2022
  • 09/2022: I am glad to give a talk about prompting-based continual learning at ContinualAI
  • 09/2022: Our paper on adversarially robust pruning is accepted at ICDM 2022
  • 07/2022: Our paper on prompting-based continual learning is accepted at ECCV 2022
  • 03/2022: Our paper on prompting-based continual learning is accepted at CVPR 2022
  • Earlier news

Experiences

  • Aug 2023 - Present, Google,
    Research Scientist at Cloud AI Research
  • Sep 2018 - July 2023, Northeastern University,
    Research assistant at SPIRAL Group
  • June 2021 - Jan 2023, Google,
    Student researcher / Research Intern at Cloud AI Research
  • Feb 2017 - July 2018, Tsinghua University,
    Research assistant at i-Vision Group
  • July 2017 - Sep 2017, University of Michigan,
    Visiting researcher at Vision & Learning Lab

Selected Publications
Google Scholar for all publications.
Chain-of-Table: Evolving Tables in the Reasoning Chain for Table Understanding
Zilong Wang, Hao Zhang, Chun-Liang Li, Julian Martin Eisenschlos, Vincent Perot, Zifeng Wang, Lesly Miculicich, Yasuhisa Fujii, Jingbo Shang, Chen-Yu Lee, Tomas Pfister
International Conference on Learning Representations (ICLR), 2024.
[paper]

CHAIN-OF-TABLE enhances the reasoning capability of LLMs by leveraging tabular structures to express intermediate thoughts for table-based reasoning. It instructs LLMs to dynamically plan an operation chain according to the input table and its associated question.

QueryForm: A Simple Zero-shot Form Entity Query Framework
Zifeng Wang, Zizhao Zhang, Jacob Devlin, Chen-Yu Lee, Guolong Su, Hao Zhang, Jennifer Dy, Vincent Perot, Tomas Pfister
Findings of the Association for Computational Linguistics: ACL 2023
[paper]

QueryForm consists of a novel prompting-based framework for zero-shot document entity recognition with large language models (LLMs), and a large-scale weakly-supervised pre-training method on publicly available webpages.

DualHSIC: HSIC-Bottleneck and Alignment for Continual Learning
Zifeng Wang*, Zheng Zhan*, Yifan Gong, Yucai Shao, Stratis Ioannidis, Yanzhi Wang, Jennifer Dy
International Conference on Machine Learning (ICML), 2023.
[paper][code]

DualHSIC consists of two complementary components that stem from the Hilbert Schmidt independence criterion (HSIC): HSIC-Bottleneck for Rehearsal (HBR) lessens the inter-task interference and HSIC Alignment (HA) promotes task-invariant knowledge sharing.

SparCL: Sparse Continual Learning on the Edge
Zifeng Wang*, Zheng Zhan*, Yifan Gong, Geng Yuan, Wei Niu, Tong Jian, Bin Ren, Stratis Ioannidis, Yanzhi Wang, Jennifer Dy
Neural Information Processing Systems (NeurIPS), 2022.
[paper] [code]

SparCL explores sparsity for efficient continual learning and achieves both training acceleration and accuracy preservation through the synergy of three aspects: weight sparsity, data efficiency, and gradient sparsity.

DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning
Zifeng Wang, Zizhao Zhang, Sayna Ebrahimi, Ruoxi Sun, Han Zhang, Chen-Yu Lee, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, Tomas Pfister
European Conference on Computer Vision (ECCV), 2022.
[paper] [code]

DualPrompt presents a novel approach to attach complementary prompts to the pre-trained backbone, and then formulates the continual learning objective as learning task-invariant and task-specific “instructions".

Learning to Prompt for Continual Learning
Zifeng Wang, Zizhao Zhang, Chen-Yu Lee, Han Zhang, Ruoxi Sun, Xiaoqi Ren, Guolong Su, Vincent Perot, Jennifer Dy, Tomas Pfister
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022.
[paper] [code] [blog]

We propose a new learning paradigm for continual learning: our method learns to dynamically prompt (L2P) a pre-trained model to learn tasks sequentially under different task transitions.

Deep Bayesian Unsupervised Lifelong Learning
Tingting Zhao*, Zifeng Wang*, Aria Masoomi, Jennifer Dy
Neural Networks, 2022.
[paper] [code]

We develop a fully Bayesian inference framework for ULL with a novel end-to-end Deep Bayesian Unsupervised Lifelong Learning (DBULL) algorithm.

Revisiting Hilbert-Schmidt Information Bottleneck for Adversarial Robustness
Zifeng Wang*, Tong Jian*, Aria Masoomi, Stratis Ioannidis, Jennifer Dy
Neural Information Processing Systems (NeurIPS), 2021.
[paper] [code]
Invited oral presentation at INFORMS 2022

We investigate the HSIC (Hilbert-Schmidt independence criterion) bottleneck as a regularizer for learning an adversarially robust deep neural network classifier, both theoretically and empirically.

Improved prediction of smoking status via isoform-aware RNA-seq deep learning models
Zifeng Wang, Aria Masoomi, Zhonghui Xu, Adel Boueiz, Sool Lee, Tingting Zhao, Russell Bowler, Michael Cho, Edwin K Silverman, Craig Hersh, Jennifer Dy, Peter J Castaldi
PLOS Computational Biology, 2021.
[paper]

We propose a novel deep learning model to incorporate prior knowledge regarding the relationship of exons to transcript isoforms. We hypothesized that since smoking alters patterns of exon and isoform usage, greater predictive accuracy could be obtained by using exon and isoform-level quantifications to predict smoking status.

Learn-Prune-Share for Lifelong Learning
Zifeng Wang*, Tong Jian*, Kaushik Chowdhury, Yanzhi Wang, Jennifer Dy, Stratis Ioannidis
International Conference on Data Mining (ICDM), 2020.
[paper]

We propose a learn-prune-share (LPS) algorithm which addresses the challenges of catastrophic forgetting, parsimony, and knowledge reuse simultaneously.

Open-world class discovery with kernel networks
Zifeng Wang, Batool Salehi, Andrey Gritsenko, Kaushik Chowdhury, Stratis Ioannidis, Jennifer Dy
International Conference on Data Mining (ICDM), 2020.
Best paper candidate
[paper]

We propose Class Discovery Kernel Network with Expansion (CD-KNet-Exp), a deep learning framework for open-world class discovery problem.

Instance-wise Feature Grouping
Aria Masoomi, Chieh Wu, Tingting Zhao, Zifeng Wang, Peter Castaldi, Jennifer Dy
Neural Information Processing Systems (NeurIPS), 2020.
[paper]

We formally define two types of redundancies using information theory: Representation and Relevant redundancies. We leverage these redundancies to design a formulation for instance-wise feature group discovery and reveal a theoretical guideline to help discover the appropriate number of groups.

Finding a ‘new’ needle in the haystack: Unseen radio detection in large populations using deep learning
Andrey Gritsenko*, Zifeng Wang*, Tong Jian, Jennifer Dy, Kaushik Chowdhury, Stratis Ioannidis
IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN), 2019.
[paper] [news]
Best paper award

We propose a novel approach that facilitates new class detection without retraining a neural network, and perform extensive analysis of the proposed model both in terms of model parameters and real-world datasets.

Collaborative deep reinforcement learning for multi-object tracking
Liangliang Ren, Jiwen Lu, Zifeng Wang, Qi Tian, Jie Zhou
European Conference on Computer Vision (ECCV), 2018.
[paper]

We propose a collaborative deep reinforcement learning (C-DRL) method for multi-object tracking. Specifically, we consider each object as an agent and track it via the prediction network, and seek the optimal tracked results by exploiting the collaborative interactions of different agents and environments via the decision network

Invited Talks
Awards
  • Outstanding Student Award in Research, Northeastern University, 2023
  • Scholar Award, NeurIPS 2022
  • Best Paper Candidate, ICDM 2020
  • Best Paper Award, DySPAN 2019
  • Travel Award, DySPAN 2019
  • Travel Award, NeurIPS 2019
  • Dean's Fellowship, Northeastern University, 2018
       Highest honor awarded to new PhD students for out standing academic background.
  • Evergrande Scholarship, Tsinghua University, 2016
       Awarded to students with excellent academic performance, scientific potential and overall development.
Academic Service

Conference Reviewer: NeurIPS 21-23, ICML 21-23, ICLR 22-23, CVPR 22-23, ICCV 23, ACL ARR
PC Member: SDM 23
Journal Reviewer: TPAMI, TMLR, Neural Networks

Template Credit: Jon Barron