Youliang Yuan

Orcid: 0000-0001-5896-7669

According to our database1, Youliang Yuan authored at least 26 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Towards Evaluating Proactive Risk Awareness of Multimodal Language Models.
CoRR, May, 2025

VisBias: Measuring Explicit and Implicit Social Biases in Vision Language Models.
CoRR, March, 2025

VisFactor: Benchmarking Fundamental Visual Cognition in Multimodal Large Language Models.
CoRR, February, 2025

Competing Large Language Models in Multi-Agent Gaming Environments.
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

Insight Over Sight: Exploring the Vision-Knowledge Conflicts in Multimodal LLMs.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

Can't See the Forest for the Trees: Benchmarking Multimodal Safety Awareness for Multimodal LLMs.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025

Chain-of-Jailbreak Attack for Image Generation Models via Step by Step Editing.
Proceedings of the Findings of the Association for Computational Linguistics, 2025

2024
Libra-Leaderboard: Towards Responsible AI through a Balanced Leaderboard of Safety and Capability.
CoRR, 2024

Chain-of-Jailbreak Attack for Image Generation Models via Editing Step by Step.
CoRR, 2024

Learning to Ask: When LLMs Meet Unclear Instruction.
CoRR, 2024

On the Resilience of Multi-Agent Systems with Malicious Agents.
CoRR, 2024

How Far Are We on the Decision-Making of LLMs? Evaluating LLMs' Gaming Ability in Multi-Agent Environments.
CoRR, 2024

The Earth is Flat? Unveiling Factual Errors in Large Language Models.
CoRR, 2024

A & B == B & A: Triggering Logical Reasoning Failures in Large Language Models.
CoRR, 2024

New Job, New Gender? Measuring the Social Bias in Image Generation Models.
Proceedings of the 32nd ACM International Conference on Multimedia, MM 2024, Melbourne, VIC, Australia, 28 October 2024, 2024

GPT-4 Is Too Smart To Be Safe: Stealthy Chat with LLMs via Cipher.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

On the Humanity of Conversational AI: Evaluating the Psychological Portrayal of LLMs.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Difficult Task Yes but Simple Task No: Unveiling the Laziness in Multimodal LLMs.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, 2024

LogicAsker: Evaluating and Improving the Logical Reasoning Ability of Large Language Models.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Does ChatGPT Know That It Does Not Know? Evaluating the Black-Box Calibration of ChatGPT.
Proceedings of the 2024 Joint International Conference on Computational Linguistics, 2024

All Languages Matter: On the Multilingual Safety of LLMs.
Proceedings of the Findings of the Association for Computational Linguistics, 2024

2023
Who is ChatGPT? Benchmarking LLMs' Psychological Portrayal Using PsychoBench.
CoRR, 2023

All Languages Matter: On the Multilingual Safety of Large Language Models.
CoRR, 2023

2021
Multi-intent Attention and Top-k Network with Interactive Framework for Joint Multiple Intent Detection and Slot Filling.
Proceedings of the Natural Language Processing and Chinese Computing, 2021

DCEN: A Decoupled Context Enhanced Network For Few-shot Slot Tagging.
Proceedings of the International Joint Conference on Neural Networks, 2021


  Loading...