Not All Denoising Steps Are Equal: Model Scheduling for Faster Masked Diffusion Language Models
Model scheduling for masked diffusion language models uses smaller models at early denoising steps for faster generation.
Model scheduling for masked diffusion language models uses smaller models at early denoising steps for faster generation.
Process reward models improve LLM mathematical reasoning by providing step-level feedback on intermediate errors, not just final outcomes.
Fairness-aware GNN training using contrastive learning and counterfactual augmentation to mitigate biases from graph structure.
LLM-based compression using domain-adapted LoRA for lossless and lossy text compression achieving 2x improvements.
Systematic characterization of WebGPU dispatch overhead for LLM inference across GPU vendors, backends, and browsers at batch size 1.
UI-Oceanus framework scales GUI agents via synthetic environmental dynamics and self-supervised learning instead of costly human demonstrations.
Benchmark evaluating LLM and embedding performance for drug discovery tasks, assessing advantages over traditional methods.
Frequency-aware Transformer for carbon footprint forecasting in power grids using periodic patterns and exogenous variables.
Contextual RL improves agent generalization by exposing agents to environment characteristics for better zero-shot transfer beyond training distribution.
OPRIDE method for offline preference-based RL reducing human feedback queries through efficient in-dataset exploration strategies.
Differentiable Symbolic Planning architecture combining neural networks with discrete symbolic reasoning for constraint satisfaction problems.
Framework for modeling and controlling ML model reliability under temporal distribution shift during deployment with continuous monitoring.
Contrastive prompt tuning method to optimize LLMs for generating energy-efficient code aligned with Green Software Development goals.
PRISM framework for zero-shot policy transfer in RL using interpretable concept clustering with causal validation across different algorithms.
Entropy-based analysis of combining Chain-of-Thought with RL for text-to-image generation, showing exploration-optimization tradeoffs.
Live benchmark dataset for forecasting startup success using Y Combinator batches with three-month evaluation cycles.
Dynamical systems analysis of vanishing gradient and overfitting in multi-layer perceptrons using minimal models.
Self-Directed Task Identification framework enabling models to autonomously identify target variables in zero-shot settings without pretraining.
Physics-informed deep generative models for offline RL in spaceflight to mitigate sim-to-real gap with limited real-world training data.
Open-source benchmarking of Matrix Profile methods for time-series anomaly detection on univariate and multivariate datasets.
Study comparing frontier vs smaller LLMs for mathematical proof verification, evaluating whether expensive models are necessary for proof checking.
Analysis of layer-to-layer representation changes in language models, decomposing updates into tokenwise and residual components.
VALOR framework for B2B sales revenue uplift modeling handling zero-inflated distributions with treatment-gated representation learning.
Transfer learning method for RNNs using time-warping rescaling, with theoretical analysis for linear differential equation models.
Framework for generating synthetic data under specified conditions with approximate identifiability guarantees for data distribution extrapolation.
Framework for validating assumptions in time-series causal discovery through calibrated risk assessment and effect-size diagnostics.
Re-analysis of transcription factor atlas single-cell perturbation screen with improved quality control pipeline and MORF barcode demultiplexing.
Low-precision training method for LLMs using adaptive Hadamard transforms based on outlier patterns in weights, activations, and gradients.
Research on using LLM-generated synthetic data to warm-start contextual bandits, examining alignment between LLM choices and actual user preferences.
Spectral framework for multi-scale nonlinear dimensionality reduction balancing global-local structure preservation and expressiveness-transparency.
Optimized NF4 dequantization kernels for fast LLM inference on NVIDIA GPUs, addressing FP16 conversion bottleneck.
Communication-efficient distributed learning algorithm with differential privacy using local training and gradient clipping.
ROMAN operator for time series that creates multiscale representations by building antialiased pyramids for convolutional classifiers.
VoxelCodeBench platform benchmarking code generation models for 3D spatial reasoning with execution in Unreal Engine.
Analysis of function vectors in LLMs showing they steer behavior beyond logit lens interpretability across 4,032 cross-template transfer pairs.
Complex-valued GNNs for distributed control of networked systems with basis-invariance for GPS and compass-denied environments.
Continual graph learning without storing exemplars, using class-level prototypes and analytic continual learning to address catastrophic forgetting.
AXELRAM smart SRAM architecture computing attention scores from quantized KV cache without dequantization using orthogonal-transform quantization.
Distributed GNN training with communication-free sampling and 4D hybrid parallelism for scaling mini-batch learning on large graphs.
Study of generalization limits in RLHF alignment, proposing compound jailbreaks targeting LLM safety through redistribution of existing capabilities.
Theoretical analysis of gradient descent training at edge of stability with product-stability property for convergence guarantees.
Low-rank compression of pretrained models using randomized subspace iteration for efficient SVD-based model reduction.
Physics-informed neural networks coupled with finite difference methods for thermal-hydraulic system simulation.
Muscle fatigue detection from sEMG signals using adversarial and contrastive learning with neural networks.
Mechanistic interpretability research on whether LLMs encode belief geometries like transformers trained on hidden Markov models.
Token-space attacks on reward models used in RLHF, introducing TOMPA framework for adversarial optimization beyond semantic manipulation.
Semantic communication for wireless image transmission using mixture-of-experts to adapt to diverse image contents and channel conditions.
Algebraic-geometric framework for quantum neural networks addressing barren plateaus and noise robustness.
FluxMoE system decouples expert residency to improve inference serving throughput in Mixture-of-Experts LLMs.
Technical analysis of evaluation methodology challenges for diffusion language models at scale.