When Does Context Help? A Systematic Study of Target-Conditional Molecular Property Prediction
Systematic study of target context conditioning for molecular property prediction across protein families and data regimes.
Systematic study of target context conditioning for molecular property prediction across protein families and data regimes.
TwinLoop: simulation-in-the-loop digital twin framework for online multi-agent reinforcement learning with context shifts.
Physics-driven neural network for estimating wheel polygonal roughness from vibration signals in rail vehicles.
SubFLOT: federated learning method using optimal transport for personalized submodel extraction in heterogeneous settings.
SHAPE framework for LLM reasoning using stage-aware hierarchical advantage estimation to improve process supervision efficiency.
FlowAdam: hybrid optimizer augmenting Adam with geometry-aware soft momentum injection for handling parameter couplings.
GraphWalker: graph-guided in-context learning framework for LLM-based clinical reasoning on electronic health records.
Method for improving classification calibration using generative perspective to regularize cross-entropy loss in deep networks.
Bi-Lipschitz autoencoder with injectivity guarantee for dimensionality reduction while preserving manifold geometry.
Federated learning approach for training time series foundation models using bi-level heterogeneous learning to address gradient conflicts.
Framework for extracting linearized neural network models via knowledge distillation for photonic hardware compatibility.
EmBolic: hyperbolic deep learning architecture for emotion analysis from text using Busemann energy-based attention.
Philosophical examination of machine learning through rhetoric lens, arguing ML is inherently rhetorical rather than objective.
Empirical study of Voronoi tessellations in LLM latent spaces, validating scaling laws of expressibility gaps.
Instance-adaptive variational autoencoder addressing amortization gap in latent variable models through per-instance parametrization.
MoBiE: Binarization framework for efficient inference of Mixture-of-Experts LLMs with post-training quantization techniques.
OmniTabBench: Large-scale benchmark comparing GBDTs, neural networks, and foundation models on tabular data with 100+ datasets.
STQuant framework for adaptive quantization of optimizer states during large multimodal model training to reduce memory costs.
Theoretical analysis of Bellman residual minimization for solving Markov decision processes under linear function approximation.
Neural operator enhancement method for dynamical systems combining Fourier-based operators with diffusion-based high-frequency recovery.
JAX-based differentiable framework for vertex-modeling epithelial tissue mechanics with automatic differentiation and GPU acceleration.
Decentralized multi-agent RL approach for vehicle-to-infrastructure systems using equivariant neural networks.
Efficient scaling technique for diffusion RL post-training using low-precision exploration and higher-precision training.
Neural method for learning search policies in Traveling Salesperson Problem, training models to iteratively improve solutions.
Frailty assessment framework for elderly oncology patients using multimodal wearable data and multi-instance learning.
Wearable-based stress estimation in elderly cancer patients using multimodal smartwatch and ECG data with multi-instance learning.
Transfer learning formalism using Outcome-Predictive State Representations for knowledge generalization across RL tasks.
Nonstationary classification approach using learned retrieval to condition classifiers on historical examples beyond training cutoff.
Study of expert specialization and routing behavior in sparse Mixture-of-Experts architectures for large language models at small scale.
Computer-assisted proof providing counterexample to open question about AdaBoost convergence to finite cycles.
Production application of residual reinforcement learning to automate electronic control unit calibration in vehicles.
Offline reinforcement learning approach addressing epistemic uncertainty through ensemble-based conservative value estimation.
Evaluation of ensemble deep clustering methods versus traditional approaches for disease subtype detection in electronic health records.
Theoretical analysis of multi-objective bandits comparing computational complexity to single-objective bandit problems.
Method to enhance LLM task performance by amplifying task-relevant neurons at inference time without parameter modification.
Theoretical framework addressing catastrophic forgetting in continual learning through informational structural alignment rather than external mechanisms.
Novel divisive clustering algorithm using mutual reachability minimum spanning trees to detect clusters of varying sizes and densities.
Research on using multi-turn reasoning LLMs with deep reinforcement learning for task offloading decisions in mobile edge computing systems.
Framework for generating synthetic financial time series that model both distributions and temporal dynamics using Schrödinger-Bass Bridge methods.
Hierarchical reinforcement learning framework for military aviation maintenance and logistics decisions, addressing fleet-scale decision-making under uncertainty.
Research on calibrating uncertainty quantification in LLMs for question-answering through token-level temperature scaling, addressing gaps in existing confidence measures.
Mixture proportion estimation from unlabeled data using conditional independence assumptions. Application to PU learning, label noise, and domain adaptation.
SDE-based method for constructing diffusion processes on implicit data manifolds using only point clouds. Data-driven approach without geometric primitives.
Computational complexity analysis of ML model expressiveness for complex systems. Studies how ML manages complexity through probability on sampleable distributions.
Theoretical study of differential privacy cost for language identification and generation. Establishes algorithms and lower bounds quantifying privacy-utility tradeoff.
Categorical framework formalizing deep learning model architectures using array broadcasting and morphisms. Mathematical notation for neural network composition.
Comparative analysis of SHAP explainability method applied to different ML models. Reviews interpretability for black-box model predictions.
Training method for Android UI agents improving RL efficiency using single state multiple actions paradigm to reduce sample inefficiency and emulator latency.
Graph neural networks with neural ODEs for thermal-hydraulic forecasting in nuclear reactors under partial observability. Physics-informed surrogate modeling.
Split learning framework with frequency-aware compression reducing communication overhead in distributed neural network training on resource-constrained edge devices.