Jingfeng Wu

Orcid: 0009-0008-5766-7873

According to our database1, Jingfeng Wu authored at least 49 papers between 2016 and 2025.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2025
Cloud Native System for LLM Inference Serving.
CoRR, July, 2025

Unlock the Potential of Fine-grained LLM Serving via Dynamic Module Scaling.
CoRR, July, 2025

A Simplified Analysis of SGD for Linear Regression with Weight Averaging.
CoRR, June, 2025

Improved Scaling Laws in Linear Regression via Data Reuse.
CoRR, June, 2025

Large Stepsizes Accelerate Gradient Descent for Regularized Logistic Regression.
CoRR, June, 2025

Minimax Optimal Convergence of Gradient Descent in Logistic Regression via Large and Adaptive Stepsizes.
CoRR, April, 2025

Memory-Statistics Tradeoff in Continual Learning with Structural Regularization.
CoRR, April, 2025

Implicit Bias of Gradient Descent for Non-Homogeneous Deep Networks.
CoRR, February, 2025

Benefits of Early Stopping in Gradient Descent for Overparameterized Logistic Regression.
CoRR, February, 2025

Cloudnativesim: A Toolkit for Modeling and Simulation of Cloud-Native Applications.
Softw. Pract. Exp., 2025

How Does Critical Batch Size Scale in Pre-training?
Proceedings of the Thirteenth International Conference on Learning Representations, 2025

2024
A collective AI via lifelong learning and sharing at the edge.
Nat. Mac. Intell., 2024

Context-Scaling versus Task-Scaling in In-Context Learning.
CoRR, 2024

UELLM: A Unified and Efficient Approach for LLM Inference Serving.
CoRR, 2024

In-Context Learning of a Linear Transformer Block: Benefits of the MLP Component and One-Step GD Initialization.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Scaling Laws in Linear Regression: Compute, Parameters, and Data.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

Large Stepsize Gradient Descent for Non-Homogeneous Two-Layer Networks: Margin Improvement and Fast Optimization.
Proceedings of the Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, 2024

UELLM: A Unified and Efficient Approach for Large Language Model Inference Serving.
Proceedings of the Service-Oriented Computing - 22nd International Conference, 2024

How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Risk Bounds of Accelerated SGD for Overparameterized Linear Regression.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

Large Stepsize Gradient Descent for Logistic Loss: Non-Monotonicity of the Loss Improves Optimization Efficiency.
Proceedings of the Thirty Seventh Annual Conference on Learning Theory, June 30, 2024

2023
Benign Overfitting of Constant-Stepsize SGD for Linear Regression.
J. Mach. Learn. Res., 2023

Learning High-Dimensional Single-Neuron ReLU Networks with Finite Samples.
CoRR, 2023

Private Federated Frequency Estimation: Adapting to the Hardness of the Instance.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Implicit Bias of Gradient Descent for Logistic Regression at the Edge of Stability.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Finite-Sample Analysis of Learning High-Dimensional Single ReLU Neuron.
Proceedings of the International Conference on Machine Learning, 2023

Fixed Design Analysis of Regularization-Based Continual Learning.
Proceedings of the Conference on Lifelong Learning Agents, 2023

2022
Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Last Iterate Risk Bounds of SGD with Decaying Stepsize for Overparameterized Linear Regression.
Proceedings of the International Conference on Machine Learning, 2022

Gap-Dependent Unsupervised Exploration for Reinforcement Learning.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

2021
Programmable packet scheduling with a single queue.
Proceedings of the ACM SIGCOMM 2021 Conference, Virtual Event, USA, August 23-27, 2021., 2021

Twenty Years After: Hierarchical Core-Stateless Fair Queueing.
Proceedings of the 18th USENIX Symposium on Networked Systems Design and Implementation, 2021

Ship Compute or Ship Data? Why Not Both?
Proceedings of the 18th USENIX Symposium on Networked Systems Design and Implementation, 2021

The Benefits of Implicit Regularization from SGD in Least Squares Problems.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Accommodating Picky Customers: Regret Bound and Exploration Complexity for Multi-Objective Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Direction Matters: On the Implicit Bias of Stochastic Gradient Descent with Moderate Learning Rate.
Proceedings of the 9th International Conference on Learning Representations, 2021

Lifelong Learning with Sketched Structural Regularization.
Proceedings of the Asian Conference on Machine Learning, 2021

2020
Direction Matters: On the Implicit Regularization Effect of Stochastic Gradient Descent with Moderate Learning Rate.
CoRR, 2020

On the Noisy Gradient Descent that Generalizes as SGD.
Proceedings of the 37th International Conference on Machine Learning, 2020

Obtaining Adjustable Regularization for Free via Iterate Averaging.
Proceedings of the 37th International Conference on Machine Learning, 2020

2019
The Multiplicative Noise in Stochastic Gradient Descent: Data-Dependent Regularization, Continuous and Discrete Approximation.
CoRR, 2019

The Anisotropic Noise in Stochastic Gradient Descent: Its Behavior of Escaping from Sharp Minima and Regularization Effects.
Proceedings of the 36th International Conference on Machine Learning, 2019

Automatic Cloud Segmentation Based on Fused Fully Convolutional Networks.
Proceedings of the Intelligent Computing Theories and Application, 2019

Tangent-Normal Adversarial Regularization for Semi-Supervised Learning.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019

2018
Tangent-Normal Adversarial Regularization for Semi-supervised Learning.
CoRR, 2018

The Regularization Effects of Anisotropic Noise in Stochastic Gradient Descent.
CoRR, 2018

2017
数据异常的监测技术综述 (Survey on Monitoring Techniques for Data Abnormalities).
计算机科学, 2017

2016
Research on human body composition prediction model based on Akaike Information Criterion and improved entropy method.
Proceedings of the 9th International Congress on Image and Signal Processing, 2016


  Loading...