Haoqin Tu

Orcid: 0000-0002-5627-249X

According to our database1, Haoqin Tu authored at least 30 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Knowledge or Reasoning? A Close Look at How LLMs Think Across Domains.
CoRR, June, 2025

OpenVision: A Fully-Open, Cost-Effective Family of Advanced Vision Encoders for Multimodal Learning.
CoRR, May, 2025

SFT or RL? An Early Investigation into Training R1-Like Reasoning Large Vision-Language Models.
CoRR, April, 2025

STAR-1: Safer Alignment of Reasoning LLMs with 1K Data.
CoRR, April, 2025

ViLBench: A Suite for Vision-Language Process Reward Modeling.
CoRR, March, 2025

AttnGCG: Enhancing Jailbreaking Attacks on LLMs with Attention Manipulation.
Trans. Mach. Learn. Res., 2025

Autoregressive Pretraining with Mamba in Vision.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Language Models Can See Better: Visual Contrastive Decoding For LLM Multimodal Reasoning.
Proceedings of the 2025 IEEE International Conference on Acoustics, 2025

2024
FET-LM: Flow-Enhanced Variational Autoencoder for Topic-Guided Language Modeling.
IEEE Trans. Neural Networks Learn. Syst., August, 2024

Sight Beyond Text: Multi-Modal Training Enhances LLMs in Truthfulness and Ethics.
Trans. Mach. Learn. Res., 2024

How Far Are We From AGI: Are LLMs All We Need?
Trans. Mach. Learn. Res., 2024

Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability.
CoRR, 2024

Code-Survey: An LLM-Driven Methodology for Analyzing Large-Scale Codebases.
CoRR, 2024

A Preliminary Study of o1 in Medicine: Are We Closer to an AI Doctor?
CoRR, 2024

MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?
CoRR, 2024

What If We Recaption Billions of Web Images with LLaMA-3?
CoRR, 2024

How Far Are We From AGI.
CoRR, 2024

Eagle and Finch: RWKV with Matrix-Valued States and Dynamic Recurrence.
CoRR, 2024

VHELM: A Holistic Evaluation of Vision Language Models.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Tuning LayerNorm in Attention: Towards Efficient Multi-Modal LLM Finetuning.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

How Many Are in This Image A Safety Evaluation Benchmark for Vision LLMs.
Proceedings of the Computer Vision - ECCV 2024, 2024

2023
Linguistic Steganalysis Toward Social Network.
IEEE Trans. Inf. Forensics Secur., 2023

How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs.
CoRR, 2023

ZeroGen: Zero-Shot Multimodal Controllable Text Generation with Multiple Oracles.
Proceedings of the Natural Language Processing and Chinese Computing, 2023

ReSee: Responding through Seeing Fine-grained Visual Knowledge in Open-domain Dialogue.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

2022
SeSy: Linguistic Steganalysis Framework Integrating Semantic and Syntactic Features.
IEEE Signal Process. Lett., 2022

PCAE: A framework of plug-in conditional auto-encoder for controllable text generation.
Knowl. Based Syst., 2022

An Overview on Controllable Text Generation via Variational Auto-Encoders.
CoRR, 2022

AdaVAE: Exploring Adaptive GPT-2s in Variational Auto-Encoders for Language Modeling.
CoRR, 2022

2021
Pixel-Stega: Generative Image Steganography Based on Autoregressive Models.
CoRR, 2021


  Loading...