Trishul Chilimbi

Orcid: 0000-0003-4596-6784

According to our database1, Trishul Chilimbi authored at least 22 papers between 2020 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
VidLA: Video-Language Alignment at Scale.
CoRR, 2024

Robust Multi-Task Learning with Excess Risks.
CoRR, 2024

2023
Web-Scale Semantic Product Search with Large Language Models.
Proceedings of the Advances in Knowledge Discovery and Data Mining, 2023

Graph-Aware Language Model Pre-Training on a Large Graph Corpus Can Help Multiple Graph Applications.
Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023

OssCSE: Overcoming Surface Structure Bias in Contrastive Learning for Unsupervised Sentence Embedding.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Understanding and Constructing Latent Modality Structures in Multi-Modal Representation Learning.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023

SST: Semantic and Structural Transformers for Hierarchy-aware Language Models in E-commerce.
Proceedings of the IEEE International Conference on Big Data, 2023

ReAugKD: Retrieval-Augmented Knowledge Distillation For Pre-trained Language Models.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2023

2022
MiCS: Near-linear Scaling for Training Gigantic Model on Public Cloud.
Proc. VLDB Endow., 2022

SMILE: Scaling Mixture-of-Experts with Efficient Bi-level Routing.
CoRR, 2022

Efficient and effective training of language and graph neural network models.
CoRR, 2022

DynaMaR: Dynamic Prompt with Mask Token Representation.
CoRR, 2022

DCAF-BERT: A Distilled Cachable Adaptable Factorized Model For Improved Ads CTR Prediction.
Proceedings of the Companion of The Web Conference 2022, Virtual Event / Lyon, France, April 25, 2022

Why do We Need Large Batchsizes in Contrastive Learning? A Gradient-Bias Perspective.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Asynchronous Convergence in Multi-Task Learning via Knowledge Distillation from Converged Tasks.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Track, 2022

DynaMaR: Dynamic Prompt with Mask Token Representation.
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: EMNLP 2022 - Industry Track, Abu Dhabi, UAE, December 7, 2022

Vision-Language Pre-Training with Triple Contrastive Learning.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

Multi-modal Alignment using Representation Codebook.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

MICO: Selective Search with Mutual Information Co-training.
Proceedings of the 29th International Conference on Computational Linguistics, 2022

2021
Magic Pyramid: Accelerating Inference with Early Exiting and Token Pruning.
CoRR, 2021

MLIM: Vision-and-Language Model Pre-training with Masked Language and Image Modeling.
CoRR, 2021

2020
Tiering as a Stochastic Submodular Optimization Problem.
CoRR, 2020


  Loading...