Xinwei Wu

Orcid: 0009-0001-2167-128X

Affiliations:
  • Tianjin University, College of Intelligence and Computing, China
  • Jilin University, College of Computer Science and Technology, Changchun, China (former)


According to our database1, Xinwei Wu authored at least 12 papers between 2021 and 2025.

Collaborative distances:
  • Dijkstra number2 of five.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2025
TaP: A Taxonomy-Guided Framework for Automated and Scalable Preference Data Generation.
CoRR, June, 2025

CONTRANS: Weak-to-Strong Alignment Engineering via Concept Transplantation.
Proceedings of the 31st International Conference on Computational Linguistics, 2025

2024
Large Language Model Safety: A Holistic Survey.
CoRR, 2024

Exploring Multilingual Human Value Concepts in Large Language Models: Is Value Alignment Consistent, Transferable and Controllable across Languages?
CoRR, 2024

IRCAN: Mitigating Knowledge Conflicts in LLM Generation via Identifying and Reweighting Context-Aware Neurons.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Exploring Multilingual Concepts of Human Values in Large Language Models: Is Value Alignment Consistent, Transferable and Controllable across Languages?
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, 2024

Mitigating Privacy Seesaw in Large Language Models: Augmented Privacy Neuron Editing via Activation Patching.
Proceedings of the Findings of the Association for Computational Linguistics, 2024

2023
Large Language Model Alignment: A Survey.
CoRR, 2023

DEPN: Detecting and Editing Privacy Neurons in Pretrained Language Models.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

2022
FewFedWeight: Few-shot Federated Learning Framework across Multiple NLP Tasks.
CoRR, 2022

Swing Distillation: A Privacy-Preserving Knowledge Distillation Framework.
CoRR, 2022

2021
Unbiased Learning to Rank in Feeds Recommendation.
Proceedings of the WSDM '21, 2021


  Loading...