Skip to main content

🏢 Nanjing University

Universal Online Convex Optimization with $1$ Projection per Round
·373 words·2 mins· loading · loading
Machine Learning Optimization 🏢 Nanjing University
This paper introduces a novel universal online convex optimization algorithm needing only one projection per round, achieving optimal regret bounds for various function types, including general convex…
START: A Generalized State Space Model with Saliency-Driven Token-Aware Transformation
·2428 words·12 mins· loading · loading
Computer Vision Domain Generalization 🏢 Nanjing University
START, a novel SSM-based architecture with saliency-driven token-aware transformation, achieves state-of-the-art domain generalization performance with efficient linear complexity.
SOFTS: Efficient Multivariate Time Series Forecasting with Series-Core Fusion
·5096 words·24 mins· loading · loading
AI Generated AI Applications Finance 🏢 Nanjing University
SOFTS: An efficient MLP-based model for multivariate time series forecasting using a novel STAR module for efficient channel interaction.
SCaR: Refining Skill Chaining for Long-Horizon Robotic Manipulation via Dual Regularization
·3516 words·17 mins· loading · loading
AI Applications Robotics 🏢 Nanjing University
SCaR refines skill chaining for long-horizon robotic manipulation via dual regularization, achieving higher success rates and robustness.
Reinforcement Learning Policy as Macro Regulator Rather than Macro Placer
·2290 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Nanjing University
Reinforcement learning refines existing macro placements, enhancing chip design by improving power, performance, and area (PPA) metrics and integrating the often-overlooked metric of regularity.
Prism: A Framework for Decoupling and Assessing the Capabilities of VLMs
·4267 words·21 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Nanjing University
Prism: a novel framework disentangles perception and reasoning in Vision-Language Models (VLMs) for improved model assessment and efficient VLM development.
Policy Learning from Tutorial Books via Understanding, Rehearsing and Introspecting
·3011 words·15 mins· loading · loading
Large Language Models 🏢 Nanjing University
Researchers developed Policy Learning from tutorial Books (PLfB), a novel method that trains AI agents using knowledge from tutorial books instead of relying solely on real-world data.
Online Composite Optimization Between Stochastic and Adversarial Environments
·1450 words·7 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Nanjing University
Researchers achieve optimal regret bounds in online composite optimization under stochastic and adversarial settings using a novel optimistic composite mirror descent algorithm and a universal strateg…
On the Ability of Developers' Training Data Preservation of Learnware
·449 words·3 mins· loading · loading
AI Theory Privacy 🏢 Nanjing University
Learnware systems enable model reuse; this paper proves RKME specifications protect developers’ training data while enabling effective model identification.
Neuro-Symbolic Data Generation for Math Reasoning
·1986 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Nanjing University
Neuro-symbolic framework generates high-quality mathematical datasets, enhancing LLMs’ mathematical reasoning capabilities and surpassing state-of-the-art counterparts.
Monte Carlo Tree Search based Space Transfer for Black Box Optimization
·2970 words·14 mins· loading · loading
Transfer Learning 🏢 Nanjing University
MCTS-transfer: Iteratively refining Bayesian optimization via Monte Carlo tree search for efficient black-box optimization using transfer learning.
Meta-DT: Offline Meta-RL as Conditional Sequence Modeling with World Model Disentanglement
·4081 words·20 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Nanjing University
Meta-DT: Offline meta-RL masters unseen tasks via conditional sequence modeling and world model disentanglement, showcasing superior few-shot and zero-shot generalization.
IODA: Instance-Guided One-shot Domain Adaptation for Super-Resolution
·2808 words·14 mins· loading · loading
AI Generated Computer Vision Image Generation 🏢 Nanjing University
IODA achieves efficient one-shot domain adaptation for super-resolution using a novel instance-guided strategy and image-level domain alignment, significantly improving performance with limited target…
Exploring DCN-like architecture for fast image generation with arbitrary resolution
·1909 words·9 mins· loading · loading
Computer Vision Image Generation 🏢 Nanjing University
FlowDCN: A purely convolutional generative model achieves state-of-the-art image generation speed and quality at arbitrary resolutions, surpassing transformer-based models.
Exploring and Exploiting the Asymmetric Valley of Deep Neural Networks
·3483 words·17 mins· loading · loading
Machine Learning Federated Learning 🏢 Nanjing University
Deep neural network training reveals asymmetric loss valleys, impacting model fusion and federated learning; sign consistency between noise and convergence is key.
Efficient Recurrent Off-Policy RL Requires a Context-Encoder-Specific Learning Rate
·3209 words·16 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Nanjing University
Recurrent off-policy RL, while robust, suffers from training instability. RESEL, a novel algorithm, solves this by using a context-encoder-specific learning rate, significantly improving stability an…
EASI: Evolutionary Adversarial Simulator Identification for Sim-to-Real Transfer
·2314 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Nanjing University
EASI: Evolutionary Adversarial Simulator Identification bridges the reality gap in robotics by using GAN and ES to find optimal simulator parameters, enabling seamless sim-to-real transfer with minima…
All-in-One Image Coding for Joint Human-Machine Vision with Multi-Path Aggregation
·5531 words·26 mins· loading · loading
AI Generated Computer Vision Image Coding 🏢 Nanjing University
Multi-Path Aggregation (MPA) achieves comparable performance to state-of-the-art methods in multi-task image coding, by unifying feature representations with a novel all-in-one architecture and a two-…
Adaptive Variance Reduction for Stochastic Optimization under Weaker Assumptions
·1456 words·7 mins· loading · loading
Machine Learning Optimization 🏢 Nanjing University
Adaptive STORM achieves optimal convergence rates for stochastic optimization of non-convex functions under weaker assumptions, eliminating the need for bounded gradients or function values and removi…
A Simple and Optimal Approach for Universal Online Learning with Gradient Variations
·244 words·2 mins· loading · loading
AI Theory Optimization 🏢 Nanjing University
A novel universal online learning algorithm achieves optimal gradient-variation regret across diverse function curvatures, boasting efficiency with only one gradient query per round.