Nidus: Externalized Reasoning for AI-Assisted Engineering
arXiv paper on Nidus, a governance runtime using Claude, Gemini, Codex to mechanize V-model for AI-assisted software delivery.
arXiv paper on Nidus, a governance runtime using Claude, Gemini, Codex to mechanize V-model for AI-assisted software delivery.
arXiv paper proposing OmniScore, deterministic evaluation metrics for multilingual text generation as alternative to LLM judges.
arXiv paper auditing code-editing benchmarks for LLMs, finding flaws in existing evaluation methods for instructed code modification.
arXiv paper on diffusion models for medical imaging, generating paired mammogram views for cancer screening datasets.
arXiv paper on Decision Pre-Trained Transformer for in-context reinforcement learning, enabling scalable generalist agent training.
arXiv paper on CRAB method for mitigating popularity bias in generative recommendation systems via codebook rebalancing.
arXiv paper presenting π² pipeline for curating reasoning data from structured sources to improve LLM long-context reasoning.
arXiv paper on vision-language models learning from grounded video data, finding text-only bias in video benchmarks.
arXiv paper modeling prior authorization policy retrieval as MDP for adaptive decision-making in healthcare insurance.
arXiv paper on how reasoning evolves in language models through fine-tuning and RL, studied via chess task performance.
EffiPair: Relative Contrastive Feedback method for improving runtime and memory efficiency of LLM-generated code without model fine-tuning.
Compiled AI: Paradigm where LLMs generate executable code during compilation for deterministic, model-free workflow automation execution.
Planning to Explore: Curiosity-driven planning approach for LLM-based test generation using Bayesian principles to reach deep code branches.
Analysis of 10 proposed measures for evaluating qualitative interview response quality to determine predictive validity.
Adaptive Thinking Budgets: Method for allocating inference-time compute efficiently across multi-turn LLM reasoning based on turn difficulty.
Modality-aware vector-quantized VAE for reconstructing multimodal brain MRI data across different imaging modalities.
Large Sparse Reconstruction Model studies scaling transformer context windows for improved 3D object reconstruction from multiple views.
OrthoFuse: Training-free method for merging multiple adapters in diffusion models using Riemannian geometry.
Study comparing encoder and decoder-based LLMs for screening clinical narratives to automate patient recruitment for clinical trials.
RoboPlayground: Framework for democratizing robotic manipulation evaluation through structured physical domain benchmarks.
Optimization strategies using curvature-aware methods to improve convergence speed and accuracy of physics-informed neural networks.
XMark: Multi-bit watermarking method for embedding imperceptible messages in LLM-generated text for attribution and tracing.
Study on how transformer language models learn second-order generalizations about object categories from synthetic data.
Temporal extension of TabDDPM for time-series data generation, addressing temporal dependencies in diffusion-based synthetic data creation.
Region-based re-ranker for multi-modal RAG reducing visual distractors by formulating region selection as decision-making problem.
Multi-agent spec-driven development pipeline with context-grounding hooks to prevent hallucinations and architectural violations in LLM coding agents.
Formal verification of security vulnerabilities in AI-generated code across 7 frontier LLMs and 500 prompts using Z3 SMT solver.
Study on training LLMs to express uncertainty explicitly as control interface for abstention and verification tasks.
Novel autoregressive paradigm for long-sequence symbolic music generation using anchored cyclic generation.
Diagnostic RAG system for IT support with explicit diagnostic state tracking across turns to accumulate evidence and resolve hypotheses.
Multi-agent LLM system for clinician-in-the-loop gait analysis report drafting, coordinating specialized agents for multimodal data synthesis.
Training-free quantization method for 3D reconstruction models using random rotations without per-scene fine-tuning.
Study on AI's role in collective decision-making systems and procedural legitimacy conditions for participants.
Long video understanding via spatio-temporally structured intent-aware RAG, preserving video structure while retrieving query-relevant evidence.
System for adaptive LoRA hyperparameter tuning and orchestration across heterogeneous multi-tenant LLM fine-tuning workloads.
Open-source digital twin simulator integrating natural language with renewable energy microgrid dynamics and dataset.
Security study of data exfiltration attacks via backdoored tool-use LLM agents, presenting Back-Reveal attack with semantic triggers.
3D human reconstruction from single images in multi-person scenes with interaction awareness.
Open-source governance-aware agentic platform for security operations, addressing alert fatigue and cross-source event correlation with LLM assistance.
Vision-language reward model framework dynamically decomposing evaluation into interpretable dimensions via gating mechanism.
Multi-agent RAG framework using agents for IoT network intrusion detection with experience library, improving interpretability over ML approaches.
Statistical framework treating LLM evaluation as tensor completion problem, addressing uncertainty quantification in pairwise comparison leaderboards.
Empirical study on fault localization's role in LLM-based automated program repair, evaluating context requirements across 500 SWE-bench instances.
Diagnostic framework combining vision-language models with flow matching and spectral detection for veterinary pneumothorax diagnosis.
Learned elevation models as alternative to LiDAR for radio environment map estimation in wireless networks.
Singing voice conversion system using boundary-aware information bottleneck for fine-grained style control.
Analysis of transformer embedding trajectories exhibiting turbulence-like 5/3 power-law spectral scaling across languages.
FastDiSS improves few-step diffusion language models for sequence-to-sequence generation by addressing self-conditioning approximation gaps.
Context-Agent framework using dynamic discourse trees for hierarchical non-linear dialogue management in LLMs.
Empirical forensic analysis of OpenClaw agentic AI system, examining internal state reconstruction and action logging for digital investigations.