Hanlin Zhang

Orcid: 0000-0002-9292-1645

Affiliations:
  • Harvard University, Cambridge, MA, USA
  • Carnegie Mellon University, Pittsburgh, PA, USA


According to our database1, Hanlin Zhang authored at least 27 papers between 2020 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
EvoLM: In Search of Lost Language Model Training Dynamics.
CoRR, June, 2025

Discovering Hierarchical Latent Capabilities of Language Models via Causal Representation Learning.
CoRR, June, 2025

Connections between Schedule-Free Optimizers, AdEMAMix, and Accelerated SGD Variants.
CoRR, February, 2025

How Does Critical Batch Size Scale in Pre-training?
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Eliminating Position Bias of Language Models: A Mechanistic Approach.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Follow My Instruction and Spill the Beans: Scalable Data Extraction from Retrieval-Augmented Generation Systems.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Mind the Gap: Examining the Self-Improvement Capabilities of Large Language Models.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

2024
Iterative Graph Self-Distillation.
IEEE Trans. Knowl. Data Eng., March, 2024


CoLoR-Filter: Conditional Loss Reduction Filtering for Targeted Language Model Pre-training.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

A Study on the Calibration of In-context Learning.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

Evaluating Step-by-Step Reasoning through Symbolic Verification.
Proceedings of the Findings of the Association for Computational Linguistics: NAACL 2024, 2024

Watermarks in the Sand: Impossibility of Strong Watermarking for Language Models.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

2023
Exploring Transformer Backbones for Heterogeneous Treatment Effect Estimation.
Trans. Mach. Learn. Res., 2023

Watermarks in the Sand: Impossibility of Strong Watermarking for Generative Models.
IACR Cryptol. ePrint Arch., 2023

Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards and Ethical Behavior in the MACHIAVELLI Benchmark.
CoRR, 2023

Do the Rewards Justify the Means? Measuring Trade-Offs Between Rewards and Ethical Behavior in the Machiavelli Benchmark.
Proceedings of the International Conference on Machine Learning, 2023

Improved Logical Reasoning of Language Models via Differentiable Symbolic Programming.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

2022
The Impact of Symbolic Representations on In-context Learning for Few-shot Reasoning.
CoRR, 2022

A Closer Look at the Calibration of Differentially Private Learners.
CoRR, 2022

Can Transformers be Strong Treatment Effect Estimators?
CoRR, 2022

Stochastic Neural Networks with Infinite Width are Deterministic.
CoRR, 2022

Toward learning human-aligned cross-domain robust models by countering misaligned features.
Proceedings of the Uncertainty in Artificial Intelligence, 2022

Towards Principled Disentanglement for Domain Generalization.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

2021
Toward Learning Human-aligned Cross-domain Robust Models by Countering Misaligned Features.
CoRR, 2021

Learning Domain Invariant Representations for Generalizable Person Re-Identification.
CoRR, 2021

2020
Towards Interpretable Natural Language Understanding with Explanations as Latent Variables.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020


  Loading...