TRACE: Traceroute-based Internet Route change Analysis with Ensemble Learning
arXiv paper introducing TRACE, ML pipeline for detecting Internet routing changes using traceroute latency data with ensemble learning.
arXiv paper introducing TRACE, ML pipeline for detecting Internet routing changes using traceroute latency data with ensemble learning.
arXiv paper analyzing backdoor attacks on decentralized LLM post-training via pipeline parallelism, examining vulnerabilities from malicious participants.
arXiv paper proposing fully photonic convolutional neural network with pre-trained in-situ training to overcome CMOS energy bottlenecks.
PlayGen-MoG: Mixture-of-Gaussians framework for diverse multi-agent trajectory generation in team sports with spatial coordination.
Guideline2Graph: LLM/VLM-based parser converting multimodal clinical guidelines into executable decision graphs with cross-page control flow.
Examination of confirmation bias in LLMs through rule-discovery tasks, evaluating and mitigating systematic reasoning failures.
Projection-free adaptive SGD for matrix optimization achieving dimension-independent convergence without costly quadratic projections.
Statistical survey of RLHF for LLM alignment, examining noisy subjective feedback for reward models and policy optimization.
Neural posterior estimation for Bayesian parameter inference in Li-ion batteries using probabilistic calibration.
AQVolt26: High-temperature halide dataset with r²SCAN calculations for training universal ML potentials for solid-state batteries.
Deep reinforcement learning for element-level bridge lifecycle optimization using condition state proportions.
Feature Attribution Stability Suite: Evaluates post-hoc attribution method stability under realistic input perturbations with improved metrics.
Synapse: LLM-guided two-phase retrieval and genetic algorithm optimization for job-person matching under information imbalance.
Empirical study of confidence calibration and hallucination in vision-language models for medical VQA, spanning three model families.
UniScene3D: Transformer-based 3D scene encoder jointly modeling appearance and geometry from multi-view colored pointmaps with CLIP alignment.
Optimal error algorithms for adversarial learning using randomized hypotheses, improving deterministic hypothesis bounds by factor of 1/2.
WSVD: Weighted SVD technique for efficient low-precision execution of vision-language models with reduced computational burden.
Study examining safety alignment bypass methods in LLMs via jailbreak-tuning and weight orthogonalization, analyzing disabled safety guardrails.
Trajectory-free self-test loss function for learning interacting particle system potentials from unlabeled discrete time-point data.
Gromov-Wasserstein optimal transport methods for geometry-aware multi-view data embedding under heterogeneous geometries and nonlinear distortions.
AutoVerifier: LLM-based agentic framework automating end-to-end verification of technical claims across scientific literature without domain expertise required.
RL framework using LLM-as-judge for label-free knowledge distillation and reasoning improvement in small and large language models without ground truth labels.
Placebo-anchored transport framework for meta-analysis addressing covariate shift across randomized controlled trials.
Systematic evaluation of LLM formal reasoning capabilities using Chomsky hierarchy and computation theory complexity metrics.
MOMO: Foundation model merging multi-sensor Mars remote sensing data (HiRISE, CTX, THEMIS) using Equal Validation Loss alignment.
Variational Bayesian adaptive Kalman filter for state estimation in sensor networks with packet dropouts and corrupted observations.
Transfer learning approach for loan recovery prediction addressing data scarcity and distribution shifts across portfolio domains.
Theoretical characterization of Lipschitz constants for kernel feature maps in positive definite kernels.
Procedural geometry data generation and visual grounding with vision-language models for Referring Image Segmentation in education.
Split-and-conquer framework for detecting manipulated speech regions via boundary detection and segment-level classification.
GPU-accelerated solver combining randomized subspace embedding and Nesterov acceleration for large-scale portfolio optimization.
Learning from synthetic data using provenance-based input gradient guidance to teach models discriminative regions.
Inversion-free stochastic natural gradient method for optimization on Riemannian manifolds with implicit parameter constraints.
Fredholm Integral Neural Operators framework for learning non-expansive integral operators with universal approximation proofs.
SkillRT: Compiler-inspired system treating LLM agent skills as code for consistent, efficient execution across diverse platforms.
Theoretical characterization of Gaussian universality breakdown in high-dimensional convex empirical risk minimization under non-Gaussian data.
Analysis of why discrete tokenization limits Vision-Language-Action model scaling; introduces Compression Gap information-theoretic principle.
First transferable learned membership inference attack on fine-tuned LLMs using unlimited labeled data from fine-tuning process itself.
PR3DICTR: PyTorch/MONAI-based open framework for 3D medical image classification and standardized deep learning model training.
Tsetlin Machine-based intrusion detection system for IoMT networks addressing cybersecurity vulnerabilities in medical device systems.
LLM framework for causal graph discovery using breadth-first search queries, reducing complexity from quadratic to linear query requirements.
Methods for training constrained regression trees incorporating domain-specific output constraints using mixed-integer programming and other approaches.
Integration of neural networks into combinatorial optimization for NP-hard problems, learning heuristics and optimality scores via graph convolutional networks.
Amortized inference framework training single model to predict causal mechanisms across multiple datasets for out-of-distribution generalization.
Method for detecting unauthorized training data in one-step distilled diffusion models using distributional statistics instead of memorization detection.
Zero-shot concept bottleneck models enabling interpretable predictions without target task training by leveraging pre-trained vision-language models.
Mathematical framework unifying diffusion and flow-based generative models as denoising Markov processes with rigorous theoretical foundation.
Integration of linear temporal logic specifications into RL using differentiable simulation for safe, correct-by-construction controller synthesis.
Noise-robust exploration method for RL using learning progress monitoring to escape unlearnable noise sources with improved sample efficiency.
ARMOR: one-shot post-training pruning algorithm for LLMs achieving 2:4 semi-structured sparsity with minimal performance degradation for efficient deployment.