Ax Junlong Jia, Ziyang Chen, Xing Wu, Chaochen Gao, TingHao Yu, Feng Zhang, Songlin Hu 22d ago

PolicyLong: Towards On-Policy Context Extension

PolicyLong method for extending LLM context windows using on-policy data synthesis to align with model capabilities during training.

Ax Ivan Tjuawinata, Andre Gunawan, Anh Quan Tran, Nitish Kumar, Payal Pote, Harsh Bansal, Chu-Hung Chi, Kwok-Yan Lam, Parventanis Murthy 22d ago

A Systematic Framework for Tabular Data Disentanglement

Framework for tabular data disentanglement transforming complex attribute relationships into latent variables with reduced interdependencies.

Ax Eleni Triantafillou, Ahmed Imtiaz Humayun, Monica Ribero, Alexander Matt Turner, Michael C. Mozer, Georgios Kaissis 22d ago

Is your algorithm unlearning or untraining?

Research clarifying distinction between machine unlearning and untraining—different approaches to removing data points or behaviors from trained models.

Ax Zigeng Chen, Gongfan Fang, Xinyin Ma, Ruonan Yu, Xinchao Wang 22d ago

DMax: Aggressive Parallel Decoding for dLLMs

DMax enables efficient parallel decoding in diffusion language models through progressive self-refinement.

Ax Andrey Bocharnikov, Ivan Ermakov, Denis Kuznedelev, Vyacheslav Zhdanovskiy, Yegor Yershov 22d ago

KV Cache Offloading for Context-Intensive Tasks

KV cache offloading technique to reduce memory and latency overhead for long-context LLM inference.

Ax Mu Nan, Muquan Yu, Weijian Mai, Jacob S. Prince, Hossein Adeli, Rui Zhang, Jiahang Cao, Benjamin Becker, John A. Pyles, Margaret M. Henderson, Chunfeng Song, Nikolaus Kriegeskorte, Michael J. Tarr, Xiaoqing Hu, Andrew F. Luo 22d ago

Meta-learning In-Context Enables Training-Free Cross Subject Brain Decoding

Meta-learning approach for brain signal decoding without per-subject training.