Xiaoyu Li

Orcid: 0009-0002-0078-3829

According to our database1, Xiaoyu Li authored at least 31 papers between 2024 and 2025.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Exploring the Limits of KV Cache Compression in Visual Autoregressive Transformers.
CoRR, March, 2025

Time and Memory Trade-off of KV-Cache Compression in Tensor Transformer Decoding.
CoRR, March, 2025

Theoretical Guarantees for High Order Trajectory Refinement in Generative Flows.
CoRR, March, 2025

Scaling Law Phenomena Across Regression Paradigms: Multiple and Kernel Approaches.
CoRR, March, 2025

On Computational Limits of FlowAR Models: Expressivity and Efficiency.
CoRR, February, 2025

Force Matching with Relativistic Constraints: A Physics-Inspired Approach to Stable and Efficient Generative Modeling.
CoRR, February, 2025

Universal Approximation of Visual Autoregressive Transformers.
CoRR, February, 2025

Dissecting Submission Limit in Desk-Rejections: A Mathematical Analysis of Fairness in AI Conference Policies.
CoRR, February, 2025

High-Order Matching for One-Step Shortcut Diffusion Models.
CoRR, February, 2025

Neural Algorithmic Reasoning for Hypergraphs with Looped Transformers.
CoRR, January, 2025

RichSpace: Enriching Text-to-Video Prompt Space via Text Embedding Interpolation.
CoRR, January, 2025

On the Computational Capability of Graph Neural Networks: A Circuit Complexity Bound Perspective.
CoRR, January, 2025

On Computational Limits and Provably Efficient Criteria of Visual Autoregressive Models: A Fine-Grained Complexity Analysis.
CoRR, January, 2025

Circuit Complexity Bounds for Visual Autoregressive Model.
CoRR, January, 2025

NRFlow: Towards Noise-Robust Generative Modeling via High-Order Mechanism.
Proceedings of the Conference on Uncertainty in Artificial Intelligence, 2025

Fast John Ellipsoid Computation with Differential Privacy Optimization.
Proceedings of the Conference on Parsimony and Learning, 2025

The Computational Limits of State-Space Models and Mamba via the Lens of Circuit Complexity.
Proceedings of the Conference on Parsimony and Learning, 2025

Bypassing the Exponential Dependency: Looped Transformers Efficiently Learn In-context by Multi-step Gradient Descent.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2025

2024
Theoretical Constraints on the Expressive Power of RoPE-based Tensor Attention Transformers.
CoRR, 2024

Fast Gradient Computation for RoPE Attention in Almost Linear Time.
CoRR, 2024

Grams: Gradient Descent with Adaptive Momentum Scaling.
CoRR, 2024

On the Expressive Power of Modern Hopfield Networks.
CoRR, 2024

Circuit Complexity Bounds for RoPE-based Transformer Architecture.
CoRR, 2024

Advancing the Understanding of Fixed Point Iterations in Deep Neural Networks: A Detailed Analytical Study.
CoRR, 2024

Fine-grained Attention I/O Complexity: Comprehensive Analysis for Backward Passes.
CoRR, 2024

Quantum Speedups for Approximating the John Ellipsoid.
CoRR, 2024

A Tighter Complexity Analysis of SparseGPT.
CoRR, 2024

Fast John Ellipsoid Computation with Differential Privacy Optimization.
CoRR, 2024

DiHAN: A Novel Dynamic Hierarchical Graph Attention Network for Fake News Detection.
Proceedings of the 33rd ACM International Conference on Information and Knowledge Management, 2024

Faster Sampling Algorithms for Polytopes with Small Treewidth.
Proceedings of the IEEE International Conference on Big Data, 2024

Fast Second-order Method for Neural Networks under Small Treewidth Setting.
Proceedings of the IEEE International Conference on Big Data, 2024


  Loading...