Yingshui Tan
According to our database1,
Yingshui Tan
authored at least 30 papers
between 2019 and 2025.
Collaborative distances:
Collaborative distances:
Timeline
Legend:
Book In proceedings Article PhD thesis Dataset OtherLinks
On csauthors.net:
Bibliography
2025
MSR-Align: Policy-Grounded Multimodal Alignment for Safety-Aware Reasoning in Vision-Language Models.
CoRR, June, 2025
Reinforcement Learning Optimization for Large-Scale Learning: An Efficient and User-Friendly Scaling Library.
CoRR, June, 2025
USB: A Comprehensive and Unified Safety Evaluation Benchmark for Multimodal Large Language Models.
CoRR, May, 2025
Beyond Safe Answers: A Benchmark for Evaluating True Risk Awareness in Large Reasoning Models.
CoRR, May, 2025
DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models.
CoRR, April, 2025
CoRR, February, 2025
HiddenDetect: Detecting Jailbreak Attacks against Large Vision-Language Models via Monitoring Hidden States.
CoRR, February, 2025
ChineseSimpleVQA - "See the World, Discover Knowledge": A Chinese Factuality Evaluation for Large Vision Language Models.
CoRR, February, 2025
Equilibrate RLHF: Towards Balancing Helpfulness-Safety Trade-off in Large Language Models.
CoRR, February, 2025
Chinese SafetyQA: A Safety Short-form Factuality Benchmark for Large Language Models.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025
HiddenDetect: Detecting Jailbreak Attacks against Multimodal Large Language Models via Monitoring Hidden States.
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025
See the World, Discover Knowledge: A Chinese Factuality Evaluation for Large Vision Language Models.
Proceedings of the Findings of the Association for Computational Linguistics, 2025
2024
RapGuard: Safeguarding Multimodal Large Language Models via Rationale-aware Defensive Prompting.
CoRR, 2024
Chinese SafetyQA: A Safety Short-form Factuality Benchmark for Large Language Models.
CoRR, 2024
Enhancing Vision-Language Model Safety through Progressive Concept-Bottleneck-Driven Alignment.
CoRR, 2024
CoRR, 2024
Adaptive Dense Reward: Understanding the Gap Between Action and Reward Space in Alignment.
CoRR, 2024
2021
Neural Comput. Appl., 2021
2020
Generalizing Fault Detection Against Domain Shifts Using Stratification-Aware Cross-Validation.
CoRR, 2020
Exploiting Uncertainties from Ensemble Learners to Improve Decision-Making in Healthcare AI.
CoRR, 2020
Are Ensemble Classifiers Powerful Enough for the Detection and Diagnosis of Intermediate-Severity Faults?
CoRR, 2020
2019
Augmenting Monte Carlo Dropout Classification Models with Unsupervised Learning Tasks for Detecting and Diagnosing Out-of-Distribution Faults.
CoRR, 2019
An Encoder-Decoder Based Approach for Anomaly Detection with Application in Additive Manufacturing.
Proceedings of the 18th IEEE International Conference On Machine Learning And Applications, 2019