Belinda Zeng

Orcid: 0009-0002-5446-0523

According to our database1, Belinda Zeng authored at least 23 papers between 2020 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
VidLA: Video-Language Alignment at Scale.
CoRR, 2024

Robust Multi-Task Learning with Excess Risks.
CoRR, 2024

Better Representations via Adversarial Training in Pre-Training: A Theoretical Perspective.
CoRR, 2024

2023
Web-Scale Semantic Product Search with Large Language Models.
Proceedings of the Advances in Knowledge Discovery and Data Mining, 2023

Graph-Aware Language Model Pre-Training on a Large Graph Corpus Can Help Multiple Graph Applications.
Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023

OssCSE: Overcoming Surface Structure Bias in Contrastive Learning for Unsupervised Sentence Embedding.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Understanding and Constructing Latent Modality Structures in Multi-Modal Representation Learning.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

ForeSeer: Product Aspect Forecasting Using Temporal Graph Embedding.
Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, 2023

SST: Semantic and Structural Transformers for Hierarchy-aware Language Models in E-commerce.
Proceedings of the IEEE International Conference on Big Data, 2023

ReAugKD: Retrieval-Augmented Knowledge Distillation For Pre-trained Language Models.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2023

2022
Efficient and effective training of language and graph neural network models.
CoRR, 2022

DynaMaR: Dynamic Prompt with Mask Token Representation.
CoRR, 2022

DCAF-BERT: A Distilled Cachable Adaptable Factorized Model For Improved Ads CTR Prediction.
Proceedings of the Companion of The Web Conference 2022, Virtual Event / Lyon, France, April 25, 2022

Why do We Need Large Batchsizes in Contrastive Learning? A Gradient-Bias Perspective.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Asynchronous Convergence in Multi-Task Learning via Knowledge Distillation from Converged Tasks.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Track, 2022

DynaMaR: Dynamic Prompt with Mask Token Representation.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: EMNLP 2022 - Industry Track, Abu Dhabi, UAE, December 7, 2022

Vision-Language Pre-Training with Triple Contrastive Learning.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

Multi-modal Alignment using Representation Codebook.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

2021
Magic Pyramid: Accelerating Inference with Early Exiting and Token Pruning.
CoRR, 2021

MLIM: Vision-and-Language Model Pre-training with Masked Language and Image Modeling.
CoRR, 2021

Simpler, Faster, Stronger: Breaking The log-K Curse On Contrastive Learners With FlatNCE.
CoRR, 2021

Top-Down Attention in End-to-End Spoken Language Understanding.
Proceedings of the IEEE International Conference on Acoustics, 2021

2020
CAM: Uninteresting Speech Detector.
Proceedings of the Interspeech 2020, 2020


  Loading...