The Planetary Cost of AI Acceleration, Part II: The 10th Planetary Boundary and the 6.5-Year Countdown
Discusses environmental and computational costs of scaling LLM agents beyond human cognitive capacity, framing AI acceleration as paradigm shift.
Discusses environmental and computational costs of scaling LLM agents beyond human cognitive capacity, framing AI acceleration as paradigm shift.
CalM: Self-supervised foundation model for calcium-imaging neural data, adaptable to multiple neuroscience analysis tasks.
Region-R1: Framework for multi-modal retrieval-augmented generation re-ranking using query-side region cropping to improve image-question relevance.
Formal verification study of 3,500 code artifacts from 7 LLMs across 500 security-critical prompts, quantifying exploitable vulnerabilities in AI-generated code.
OpenCEM: Open-source digital twin simulator and dataset integrating natural language with renewable energy microgrid dynamics for intelligent energy management.
Analyzes robustness of diffusion-based image compression to bit-flip errors, comparing against classical and learned codecs.
Qualitative case study examining Nigerian legal professionals' perceptions of AI governance, regulatory gaps, and institutional readiness.
Introduces AgriPriceBD, a benchmark dataset of 1,779 daily commodity prices from Bangladesh, comparing classical and deep learning forecasting models.
Probabilistic language tries (PLTs) unify prefix structure representation serving as lossless compressor, decision policy, and execution reuse framework.
FLeX: Fourier-based low-rank expansion for parameter-efficient cross-lingual code generation transfer from Python to Java using Code Llama 7B.
Analysis of grokking training dynamics showing spectral edge reveals functional modes invisible to mechanistic interpretability tools.
S³: stratified scaling search for test-time inference in diffusion language models using classical verifiers to improve generation without additional training.
Quantum-inspired tensor network anomaly detection (SMT-AD) using superposition of bond-dimension-1 matrix product operators with Fourier feature embeddings.
Multimodal VAE framework for survival risk modeling in multiple myeloma integrating heterogeneous omics and clinical data with improved latent regularization.
RAGEN-2 identifies reasoning collapse in RL-trained multi-turn LLM agents where models use input-agnostic templates despite stable entropy metrics.
Neural networks for identifying viscoelastic parameters in multiscale blood flow cardiovascular models using asymptotic-preserving methods.
TalkLoRA: communication-aware mixture of LoRA experts for parameter-efficient LLM fine-tuning addressing routing instability in MoE-augmented approaches.
AgentOpt framework for client-side optimization of LLM-based agents handling composition of local tools, remote APIs and diverse models with reduced costs.
GRPO preference optimization applied to small language models shows diminishing returns on hard samples, revealing capacity boundaries in math reasoning tasks.
Graph Transformer architecture combining GNNs and Transformers for multi-scale molecular property prediction with fragment-aware representation learning.
Bi-level optimization framework (BiSDG) for single domain generalization that decouples task learning from domain modeling using surrogate distributions.
Master Key Hypothesis proposes capabilities correspond to transferable directions in low-dimensional subspace; introduces UNLOCK for training-free cross-model capability transfer.
Develops universal foundation model for graph-structured biomedical data including molecular networks and regulatory circuits.
Addresses hyperparameter tuning challenges in spiking reservoir computing by introducing robustness interval concept for edge-of-chaos operation.
OT-NFM enables one-step generative modeling by learning transport maps directly instead of integrating vector fields, achieving single forward pass generation.
Proposes Neural Computers (NCs) that unify computation, memory, and I/O in learned runtime states, aiming toward fully neural computing systems that replace explicit programs.
Study on latent reasoning limits in LLMs investigating whether models discover multi-step planning strategies without supervision.
Graph embedding-based anomaly detection for microservice architectures identifying under-represented services in load testing.
Data-driven approach reducing electronics production test costs while adapting to changing defect distributions and controlling escape risk.
Conformal Margin Risk Minimization framework for robust classification under label noise without privileged knowledge.
MICA architecture for multivariate time series forecasting addressing Transformer scalability with channel-dependent attention.
Multi-GPU implementation of activation-level interpretability and steering techniques for large language models, extending single-GPU methods to distributed settings.
Novel inference-time scaling method using symbolic execution to select correct code generation solutions from LLM candidates without expensive external verifiers.
DoMinO: unified RL framework for fine-tuning discrete flow matching models viewing sampling as multi-step MDP.
Theoretical analysis of stochastic convex optimization with heavy-tailed gradients under differential privacy constraints.
Study of transformers learning analogical reasoning via copying intermediate representations using meta-learning for compositionality.
VLMShield: defense mechanism for vision-language models against malicious prompt attacks using multimodal feature extraction.
Method for post-training quantization of sparse Mixture-of-Experts models with theoretical generalization guarantees.
Framework for time-series classification using cross density ratio instead of correlation-based statistics.
Systematic study of target context conditioning for molecular property prediction across protein families and data regimes.
TwinLoop: simulation-in-the-loop digital twin framework for online multi-agent reinforcement learning with context shifts.
Physics-driven neural network for estimating wheel polygonal roughness from vibration signals in rail vehicles.
SubFLOT: federated learning method using optimal transport for personalized submodel extraction in heterogeneous settings.
SHAPE framework for LLM reasoning using stage-aware hierarchical advantage estimation to improve process supervision efficiency.
FlowAdam: hybrid optimizer augmenting Adam with geometry-aware soft momentum injection for handling parameter couplings.
GraphWalker: graph-guided in-context learning framework for LLM-based clinical reasoning on electronic health records.
Method for improving classification calibration using generative perspective to regularize cross-entropy loss in deep networks.
Bi-Lipschitz autoencoder with injectivity guarantee for dimensionality reduction while preserving manifold geometry.
Federated learning approach for training time series foundation models using bi-level heterogeneous learning to address gradient conflicts.
Framework for extracting linearized neural network models via knowledge distillation for photonic hardware compatibility.