Memory-efficient Continual Learning with Prototypical Exemplar Condensation
Proposes prototypical exemplar condensation for memory-efficient continual learning, reducing stored samples per class from 20+ to single digits.
Proposes prototypical exemplar condensation for memory-efficient continual learning, reducing stored samples per class from 20+ to single digits.
ALMAB-DC framework combines active learning, multi-armed bandits, and distributed computing for expensive black-box optimization.
Investigates mechanisms of introspective awareness in LLMs, where models detect injected steering vectors with minimal false positives.
Adapts KGE metric for non-stationary geoscientific systems in water management applications.
Analyzes distributional reinforcement learning with applications to healthcare, moving beyond expectation-based objectives for uncertain domains.
Proposes hierarchical SVG tokenization approach for improved scalable vector graphics modeling with LLMs via geometric-aware token design.
ALTO system for adaptive hyperparameter tuning and orchestration of LoRA fine-tuning jobs across heterogeneous multi-tenant environments.
Introduces Gated-SwinRMT, a vision transformer combining Swin attention with Manhattan-distance decay for improved spatial modeling.
Proposes CMRM, a framework for improving classification under label noise without privileged knowledge, using quantile-calibrated regularization.
Combines LLMs with Graph Neural Networks to enhance fMRI brain network analysis by leveraging LLM representations.
Bias-constrained diffusion schedules for PDE emulation with improved reconstruction error and efficient unrolled training.
Temporal transfer learning approach for traffic optimization using real-time driving advisories for connected vehicles.
Uses cognitively motivated grammar models as behavioral biometrics for authorship verification in digital forensics.
Social recommendation model using condition-guided approach to mitigate popularity bias in recommendation systems.
Survey of detection and characterization methods for coordinated online behavior including disinformation and social movements.
Studies extrinsic gender bias in Bangla pretrained language models across four NLP classification tasks.
Method for disentangling visual concepts in image generation to enable multi-aspect creative content generation.
Automatic self-supervised learning framework for social recommendations that learns auxiliary tasks without manual design.
Extends conformal prediction framework to hierarchical classification with constrained representation complexity.
Method for constraining sequential editing of LLMs to prevent knowledge degradation using editing anchor compression.
Shows XL MIMO wireless systems exhibit universal approximation properties similar to neural networks for OTA classification.
Sparse Bayesian learning method with space power prior for block sparse signal recovery with unknown patterns.
Framework for extreme super-resolution using autoregressive chain of scale states with preference alignment.
Agentic system for generating and validating synthetic image data to address data scarcity and label noise in vision tasks.
Novel variational quantum error correction approach using state distinguishability maximization for near-term quantum devices.
Evaluates LLM reasoning capabilities in social deduction game Avalon using Bayesian inference with graph-informed models.
Neural network method for solving two-stage stochastic unit commitment optimization problems in power systems.
Benchmark for evaluating how well multimodal models describe structural properties of time series data.
arXiv paper using deep learning to infer exoplanet geometry from transit light curves.
arXiv paper on Bayesian ego-graph inference for decentralized multi-agent reinforcement learning with constrained communication.
arXiv paper on interactive program synthesis for collaborative physical task modeling from narrated demonstrations.
RESample: Data augmentation framework for Vision-Language-Action models in robotic manipulation, addressing limited distribution in demonstration datasets.
Research on representational drift in neural networks, analyzing how task-irrelevant stimuli contribute to changes in learned representations over time.
Generative View Stitching: Method enabling camera-guided video generation with bidirectional conditioning to prevent collision with generated scenes.
Methodology using flow-based approaches and non-equilibrium Monte Carlo for topology sampling in SU(3) lattice gauge theory simulations.
EGMOF: Hybrid diffusion-transformer framework for efficient generation of metal-organic frameworks for materials discovery with targeted properties.
BRIXEL: Approach to reduce computational cost of dense feature maps from vision foundation models like DINOv3 while maintaining performance.
Fed-Sparse-BNSL: Federated method for learning Bayesian network structures with differential privacy, addressing decentralized data challenges.
AV-SpeakerBench: Benchmark evaluating multimodal LLMs on fine-grained audiovisual speech understanding with 3,212 multiple-choice questions.
Research on relational visual similarity in AI vision systems, comparing current methods against human-like relational perception across different domains.
DRAM: Framework combining mechanism design and online learning for sequential multi-agent settings to ensure truthful reporting with cost-optimality.
Measurement-Consistent Langevin Corrector: Method stabilizing latent diffusion models for inverse problems by reducing discrepancy with learned reverse diffusion.
Theoretical analysis of sample complexity in symmetric composite binary quantum hypothesis testing for unknown quantum states.
ConvoLearn: Dataset of 2,134 tutor-student dialogues for fine-tuning LLM-based AI tutors, grounded in dialogic learning theory and Earth Science curriculum.
Tiled Prompts: Method addressing prompt misguidance in text-conditioned diffusion models for image and video super-resolution by handling localized details.
WeWrite: Personalized query rewriting framework for video search systems using user history to identify search intent and resolve ambiguity.
Theoretical analysis of stochastic gradient descent covariance under exchangeable mini-batch sampling and its connection to Fisher information.
PACED: LLM distillation method that weights training problems by student competence using gradient signal-to-noise ratio to improve distillation efficiency.
Framework addressing causal confusion in end-to-end autonomous driving models through causal intervention during training to improve reliability and safety.
Research on formal evaluation methods for machine learning models, focusing on test-time performance-reliability trade-offs when target KPI levels are unknown.