Efficient Dataset Selection for Continual Adaptation of Generative Recommenders
Dataset selection strategies for continual adaptation of generative recommenders under temporal distributional drift.
Dataset selection strategies for continual adaptation of generative recommenders under temporal distributional drift.
Geometric framework linking objective accuracy to structural recovery in prototype-based clustering via condition numbers.
Methods to mitigate distribution sharpening in math RLVR through hint synthesis and annealing strategies.
Sparse epsilon-insensitive elastic net SVM variant for noise-robust pattern classification with improved sparsity.
Symbiotic-MoE: unified pre-training framework enabling Large Multimodal Models to generate images while maintaining understanding capabilities.
LSLoRA: investigation of sensitivity-positional co-localization in GQA transformers, restricting LoRA to optimally sensitive layers.
Graph learning framework for 3D engineering AI applications including CAE and CFD predictions with explainability.
Cross-modal emotion transfer technique for emotion editing in synthetic talking face videos using generative models.
SEARL: framework for self-evolving AI agents that jointly optimize policy and tool graphs to learn from trajectories without large-scale LLMs.
Theoretical research on mean estimation under 1-bit communication constraints using adaptive randomized thresholds.
GRASS: gradient-based method for memory-efficient LLM fine-tuning using adaptive layer-wise importance sampling, balancing efficiency with model expressiveness.
Intensity Dot Product Graphs extending random dot product graphs with Poisson point process for latent positions.
Pipeline converting healthcare policy documents to executable BPMN models using LLMs for policy simulation and evaluation.
Recurrent-depth transformers enabling iterative reasoning to improve multi-hop knowledge composition in language models.
Study of accuracy-energy trade-offs in ensemble recommender systems across 93 experiments comparing multiple models to single models.
Learns first-order rules from image data without labels, automatically inventing predicates for explainable AI and enhancing LLM reasoning.
DACS mechanism for multi-agent LLM orchestration isolating per-agent context via registry and focus session modes to prevent context pollution.
GSSA-ViT framework using 3D Gaussian splatting for arbitrary-resolution weather forecasting and downscaling of atmospheric fields.
Unified framework viewing LLM post-training methods (SFT, preference optimization, RL, distillation) through off-policy and on-policy learning perspectives.
Uses deep reinforcement learning to automatically design quantum circuits for variational imaginary time evolution on NISQ devices.
Studies data mixing strategies for LLM training, questioning domain definitions, human-model alignment, and impact of domain weighting on generalization.
Discusses risks of LLM-generated peer reviews and automated editorial processes, proposing RAG-XAI detection framework for identifying machine-generated content.
DSCA method for lifelong vision-language model editing via dynamic subspace concept alignment, preventing degradation from sequential edits.
Decomposes long-context reasoning in LLMs into atomic skills, automatically identifying and improving fundamental capabilities for complex reasoning.
SearchAD large-scale dataset with 423k frames for rare image retrieval in autonomous driving scenarios across 11 established datasets.
CATMIL method for small brain structure segmentation in MRI using component-adaptive reweighting and lesion-level supervision.
First comprehensive survey of abductive reasoning in LLMs, defining taxonomy and exploring inference of plausible explanations from observations.
PrivFedTalk privacy-aware federated framework for personalized talking-head generation using diffusion models with identity-stable adapters.
LINE uses LLMs iteratively to explain individual neuron concepts in vision models without predefined vocabularies, enabling interpretability of neural networks.
DeepForestSound multi-species acoustic detector for biodiversity monitoring in African tropical forests using semi-supervised learning pipeline.
Training-free open-vocabulary semantic segmentation using global context awareness with pretrained vision and vision-language models without additional training.
Trust-adaptive differential privacy framework for data-driven systems balancing privacy and utility under heterogeneous user trust levels.
Method for 2-bit LLM quantization via optimal codebook initialization, enabling extreme compression for edge deployment with O(1) lookup dequantization.
Tempo framework compresses long videos for multimodal LLMs by query-aware selection of frames, addressing context limits and lost-in-middle problems.
Quantum-inspired ARIMA methodology for time series analysis using variational quantum circuits and swap-test-driven autocorrelation.
Diffusion model for virtual staining in histopathology that preserves cellular structures while translating immunohistochemistry images.
Study showing medical multimodal LLMs underperform traditional deep learning on image classification despite pretraining advantages.
Convolutional neural decoders for quantum error correction to accelerate decoding of quantum low-density parity-check codes.
Transformer-based floor plan generation using differentiable architectural loss functions to optimize room layouts beyond training data patterns.
RL primitive (Dataset Policy Gradient) optimizing synthetic data generators to produce targeted training examples for fine-tuning LLMs on differentiable metrics.
Explainable AI framework for onboard satellite fault detection with semantically annotated encodings for neural anomaly detectors.
RL framework extending verifiable-reward training to general reasoning tasks in LLMs using natural instruction data for causal and temporal understanding.
Unified evaluation platform for prompt injection attacks and defenses, addressing benchmark gaps in comparing robustness across diverse tasks.
Study of language generation under differential privacy constraints, proving privacy can be achieved without qualitative cost for countable language collections.
Analysis of on-policy distillation failure modes in LLM training, identifying length inflation and truncation collapse as destabilizing factors.
Survey of tabular data generation comparing GANs, diffusion models, and LLMs across sample quality, privacy, and controllability dimensions.
Meta-learning approach with uncertainty quantification for limited-data task learning, addressing out-of-distribution scenarios in safety-critical settings.
Comprehensive survey of generative AI covering large language models, architectures, deployment protocols, and real-world applications as of early 2026.
Continuous online learning framework for activity recognition systems addressing model drift and domain shift in long-term deployments.
Privacy-preserving face recognition using information-theoretic Privacy Funnel model with end-to-end trainable representation learning.