MR-ImagenTime: Multi-Resolution Time Series Generation through Dual Image Representations
MR-CDM: Multi-resolution time series forecasting framework using hierarchical decomposition and diffusion-based generation.
MR-CDM: Multi-resolution time series forecasting framework using hierarchical decomposition and diffusion-based generation.
JoyAI-LLM Flash: Efficient Mixture-of-Experts language model in sub-50B parameter range, pretrained on 20 trillion tokens with optimized post-training.
VisionClaw: Always-on wearable AI agent on Meta Ray-Ban glasses, integrating egocentric perception with speech-driven OpenClaw task execution.
Continuous Softened Retracing reSampling method for stabilizing unsupervised self-evolution of multimodal LLMs during post-training.
k-Maximum Inner Product Attention mechanism for graph transformers, addressing quadratic complexity and analyzing expressive power of GraphGPS.
TILA: Vision-language pretraining method for analyzing temporal changes in chest X-rays rather than individual images.
Deep learning approach for in-hospital mortality prediction from incomplete multimodal EHRs using point cloud paradigm.
Empirical robustness analysis of TabPFN tabular foundation model's in-context learning under noisy conditions.
Discusses environmental and computational costs of scaling LLM agents beyond human cognitive capacity, framing AI acceleration as paradigm shift.
CalM: Self-supervised foundation model for calcium-imaging neural data, adaptable to multiple neuroscience analysis tasks.
Region-R1: Framework for multi-modal retrieval-augmented generation re-ranking using query-side region cropping to improve image-question relevance.
Formal verification study of 3,500 code artifacts from 7 LLMs across 500 security-critical prompts, quantifying exploitable vulnerabilities in AI-generated code.
OpenCEM: Open-source digital twin simulator and dataset integrating natural language with renewable energy microgrid dynamics for intelligent energy management.
Analyzes robustness of diffusion-based image compression to bit-flip errors, comparing against classical and learned codecs.
Qualitative case study examining Nigerian legal professionals' perceptions of AI governance, regulatory gaps, and institutional readiness.
Introduces AgriPriceBD, a benchmark dataset of 1,779 daily commodity prices from Bangladesh, comparing classical and deep learning forecasting models.
Probabilistic language tries (PLTs) unify prefix structure representation serving as lossless compressor, decision policy, and execution reuse framework.
FLeX: Fourier-based low-rank expansion for parameter-efficient cross-lingual code generation transfer from Python to Java using Code Llama 7B.
Analysis of grokking training dynamics showing spectral edge reveals functional modes invisible to mechanistic interpretability tools.
S³: stratified scaling search for test-time inference in diffusion language models using classical verifiers to improve generation without additional training.
Quantum-inspired tensor network anomaly detection (SMT-AD) using superposition of bond-dimension-1 matrix product operators with Fourier feature embeddings.
Multimodal VAE framework for survival risk modeling in multiple myeloma integrating heterogeneous omics and clinical data with improved latent regularization.
RAGEN-2 identifies reasoning collapse in RL-trained multi-turn LLM agents where models use input-agnostic templates despite stable entropy metrics.
Neural networks for identifying viscoelastic parameters in multiscale blood flow cardiovascular models using asymptotic-preserving methods.
TalkLoRA: communication-aware mixture of LoRA experts for parameter-efficient LLM fine-tuning addressing routing instability in MoE-augmented approaches.
AgentOpt framework for client-side optimization of LLM-based agents handling composition of local tools, remote APIs and diverse models with reduced costs.
GRPO preference optimization applied to small language models shows diminishing returns on hard samples, revealing capacity boundaries in math reasoning tasks.
Graph Transformer architecture combining GNNs and Transformers for multi-scale molecular property prediction with fragment-aware representation learning.
Bi-level optimization framework (BiSDG) for single domain generalization that decouples task learning from domain modeling using surrogate distributions.
Master Key Hypothesis proposes capabilities correspond to transferable directions in low-dimensional subspace; introduces UNLOCK for training-free cross-model capability transfer.
Develops universal foundation model for graph-structured biomedical data including molecular networks and regulatory circuits.
Addresses hyperparameter tuning challenges in spiking reservoir computing by introducing robustness interval concept for edge-of-chaos operation.
OT-NFM enables one-step generative modeling by learning transport maps directly instead of integrating vector fields, achieving single forward pass generation.
Proposes Neural Computers (NCs) that unify computation, memory, and I/O in learned runtime states, aiming toward fully neural computing systems that replace explicit programs.
Study on latent reasoning limits in LLMs investigating whether models discover multi-step planning strategies without supervision.
Graph embedding-based anomaly detection for microservice architectures identifying under-represented services in load testing.
Data-driven approach reducing electronics production test costs while adapting to changing defect distributions and controlling escape risk.
Conformal Margin Risk Minimization framework for robust classification under label noise without privileged knowledge.
MICA architecture for multivariate time series forecasting addressing Transformer scalability with channel-dependent attention.
Multi-GPU implementation of activation-level interpretability and steering techniques for large language models, extending single-GPU methods to distributed settings.
Novel inference-time scaling method using symbolic execution to select correct code generation solutions from LLM candidates without expensive external verifiers.
DoMinO: unified RL framework for fine-tuning discrete flow matching models viewing sampling as multi-step MDP.
Theoretical analysis of stochastic convex optimization with heavy-tailed gradients under differential privacy constraints.
Study of transformers learning analogical reasoning via copying intermediate representations using meta-learning for compositionality.
VLMShield: defense mechanism for vision-language models against malicious prompt attacks using multimodal feature extraction.
Method for post-training quantization of sparse Mixture-of-Experts models with theoretical generalization guarantees.
Framework for time-series classification using cross density ratio instead of correlation-based statistics.
Systematic study of target context conditioning for molecular property prediction across protein families and data regimes.
TwinLoop: simulation-in-the-loop digital twin framework for online multi-agent reinforcement learning with context shifts.
Physics-driven neural network for estimating wheel polygonal roughness from vibration signals in rail vehicles.