Yan Teng

Orcid: 0000-0002-7069-4728

Affiliations:
  • Shanghai Artificial Intelligence Laboratory, China
  • Delft University of Technology, Ethics and Philosophy of Technology, Netherlands


According to our database1, Yan Teng authored at least 25 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
LinguaSafe: A Comprehensive Multilingual Safety Benchmark for Large Language Models.
CoRR, August, 2025

The Other Mind: How Language Models Exhibit Human Temporal Cognition.
CoRR, July, 2025

Probing the Robustness of Large Language Models Safety to Latent Perturbations.
CoRR, June, 2025

Argus Inspection: Do Multimodal Large Language Models Possess the Eye of Panoptes?
CoRR, June, 2025

JailBound: Jailbreaking Internal Safety Boundaries of Vision-Language Models.
CoRR, May, 2025

SafeVid: Toward Safety Aligned Video Large Multimodal Models.
CoRR, May, 2025

Possibilities and challenges in the moral growth of large language models: a philosophical perspective.
Ethics Inf. Technol., March, 2025

WanJuanSiLu: A High-Quality Open-Source Webtext Dataset for Low-Resource Languages.
CoRR, January, 2025

Collectivism and individualism political bias in large language models: A two-step approach.
Big Data Soc., 2025

A Mousetrap: Fooling Large Reasoning Models for Jailbreak with Chain of Iterative Chaos.
Proceedings of the Findings of the Association for Computational Linguistics, 2025

HoneypotNet: Backdoor Attacks Against Model Extraction.
Proceedings of the AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25, 2025

2024
From Pixels to Principles: A Decade of Progress and Landscape in Trustworthy Computer Vision.
Sci. Eng. Ethics, June, 2024

Reflection-Bench: probing AI intelligence with reflection.
CoRR, 2024

MEOW: MEMOry Supervised LLM Unlearning Via Inverted Facts.
CoRR, 2024

ESC-Eval: Evaluating Emotion Support Conversations in Large Language Models.
CoRR, 2024

From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities.
CoRR, 2024

MLLMGuard: A Multi-dimensional Safety Evaluation Suite for Multimodal Large Language Models.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Fake Alignment: Are LLMs Really Aligned Well?
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

Flames: Benchmarking Value Alignment of LLMs in Chinese.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

ESC-Eval: Evaluating Emotion Support Conversations in Large Language Models.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

2023
Flames: Benchmarking Value Alignment of Chinese Large Language Models.
CoRR, 2023

Plug-and-Play Feature Generation for Few-Shot Medical Image Classification.
Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, 2023

2022
Beyond legislation and technological design: The importance and implications of institutional trust for privacy issues of digital contact tracing.
Frontiers Digit. Health, 2022

2021
Ethics of Trust­-Inviting Digital Systems: Blockchain, Reputation­-Based Platforms, and COVID­19 Tracing Technologies.
PhD thesis, 2021

Towards trustworthy blockchains: normative reflections on blockchain-enabled virtual institutions.
Ethics Inf. Technol., 2021


  Loading...