AFL: A Single-Round Analytic Approach for Federated Learning with Pre-trained Models
Analytic federated learning (AFL) paradigm enabling closed-form solutions for federated learning with pre-trained models in single-round training.
Analytic federated learning (AFL) paradigm enabling closed-form solutions for federated learning with pre-trained models in single-round training.
SleepNet and DreamNet: deep learning models for visual classification via feature enrichment and reconstruction with pre-trained encoders.
Matrix Profile extension for anomaly detection in multidimensional time series from real-world applications like sensor monitoring.
DROP: distributional reinforcement learning framework with asymmetric learning rates modeling optimistic/pessimistic dopamine neuron behavior.
Pseudo-probability unlearning method for efficient privacy-preserving machine unlearning with reduced computational overhead and residual information.
Self-supervised physics-informed neural network for real-time human pose and dynamics estimation from sparse IMU sensor configurations.
Inference-time scaling method for discrete diffusion language models via trajectory refinement without retraining for reward optimization.
Negative Binomial VAE extension with discrete count-based latents for improved biological plausibility over continuous VAE representations.
Method for quantitatively estimating target task performance from unsupervised pretext tasks in semi/self-supervised learning before full training.
LNN-PINN: physics-informed neural network framework with liquid residual gating for improved predictive accuracy on complex PDE problems.
In-context learning approach for AutoML pipeline optimization beyond hyperparameter tuning, incorporating fine-tuning and ensembling techniques.
PhISM: physics-informed deep learning architecture for unsupervised hyperspectral imaging using continuous basis functions for interpretable latent representations.
LoFT method for long-tailed semi-supervised learning using foundation models with parameter-efficient fine-tuning to improve pseudo-label quality.
MDP modeling framework extending policy graphs for multi-stage stochastic programs with decision-dependent uncertainty and statistical learning.
Technique for detecting when reasoning LLMs overthink by analyzing entropy after chain-of-thought to enable early exiting.
Bayesian machine learning potentials for molecular simulations with uncertainty quantification using equivariant message passing.
Method for efficiently computing Lipschitz constant estimates for neural networks using local information to improve robustness certification.
Approximate replicability framework for machine learning algorithms that remain stable under input resampling.
Spectral clustering alternatives to Laplacian with group fairness constraints for equitable cluster representation.
GIFT unifies GRPO, DPO, and UNA in reinforcement learning framework for LLM alignment combining group-relative sampling with implicit preference learning.
LoRA-DA provides data-aware initialization for low-rank adaptation via theoretical framework and asymptotic analysis for parameter-efficient fine-tuning.
Nirvana specialized generalist model with task-aware memory mechanism for domain adaptation while preserving broad LLM capabilities.
Tensor-based Q-learning approach to handle high-dimensional reinforcement learning by exploiting problem structure without neural networks.
Adaptive symmetrization of KL divergence for learning probability distributions with normalizing flows and energy-based models.
SpecQuant framework for ultra-low-bit LLM quantization using spectral decomposition and adaptive truncation for efficient device deployment.
TREASURE foundation model for payment transaction understanding and analysis with applications to anomaly detection.
CHiQPM provides global and local interpretability for image classification in safety-critical domains with hierarchical explanations.
Adaptive Replay Buffer (ARB) dynamically prioritizes data sampling in offline-to-online reinforcement learning to balance stability and asymptotic performance.
First empirical study of machine unlearning in hybrid quantum-classical neural networks and variational quantum circuits.
Reinforcement learning framework to learn weather/climate model parametrization schemes as state-dependent functions online instead of using fixed coefficients.
Low-Rank Key-Value (LRKV) attention reduces transformer KV cache memory by exploiting redundancy across attention heads with low-rank residuals.
BadImplant introduces multi-targeted backdoor attacks against graph neural networks with injection-based mechanisms.
Explainable AI methods to improve ML reliability and prevent unexpected behavior in industrial cyber-physical systems.
SPICE uses submodular optimization and Fisher information to select training data for efficient LLM instruction tuning while addressing gradient conflicts.
Infusion framework uses influence functions to craft training data perturbations that induce targeted model behavior changes, evaluated on vision and language tasks.
Open-source foundation model for 3D molecular and materials modeling with both generative and predictive capabilities.
Interventional time series data generator for training causal foundation models on time series, extending prior-data fitted networks to temporal domains.
EvoFlows: variable-length protein sequence model using flow matching for protein engineering with native support for insertions, deletions, and mutations.
CONSERVAttack method for identifying vulnerabilities and systematic uncertainties in ML models applied to high-energy physics data analysis.
Non-parametric conformal regression method using optimal binning with CRPS loss for conditional distribution estimation.
MR-CDM: multi-resolution conditional diffusion framework for variable-length time series forecasting with hierarchical decomposition and adaptive embeddings.
Data-driven sports training framework using skeleton-based biomechanical analysis and motion modeling for personalized dart coaching.
Open-source benchmark and reproducible implementation of Matrix Profile methods for univariate and multivariate time-series anomaly detection.
On-policy self-distillation approach for LLM training combining dense teacher signals with sparse verifiable rewards from environment feedback.
NativeTernary: binary encoding format for ternary neural network weights achieving 2 bits per weight, 1.31x compression over GGUF for BitNet models.
Proposes k-maximum inner product attention mechanism for graph transformers to reduce computational complexity while maintaining expressive power.
Deep learning approach for clinical risk prediction from incomplete multimodal EHR data using point cloud paradigm to handle irregular sampling and missing modalities.
Empirical robustness analysis of TabPFN's attention mechanisms for tabular in-context learning, examining noise immunity across heterogeneous datasets.
Active inference methodology for ML-assisted data collection, using models to identify which points merit labeling under budget constraints for efficient learning.
Studies linearization of discrete transportation distance on graphs, connecting optimal transport to graph structure and providing nonasymptotic analysis.