Ax Junlong Jia, Ziyang Chen, Xing Wu, Chaochen Gao, TingHao Yu, Feng Zhang, Songlin Hu 7d ago

PolicyLong: Towards On-Policy Context Extension

PolicyLong method for extending LLM context windows using on-policy data synthesis to align with model capabilities during training.

Ax Ivan Tjuawinata, Andre Gunawan, Anh Quan Tran, Nitish Kumar, Payal Pote, Harsh Bansal, Chu-Hung Chi, Kwok-Yan Lam, Parventanis Murthy 7d ago

A Systematic Framework for Tabular Data Disentanglement

Framework for tabular data disentanglement transforming complex attribute relationships into latent variables with reduced interdependencies.

Ax Eleni Triantafillou, Ahmed Imtiaz Humayun, Monica Ribero, Alexander Matt Turner, Michael C. Mozer, Georgios Kaissis 7d ago

Is your algorithm unlearning or untraining?

Research clarifying distinction between machine unlearning and untraining—different approaches to removing data points or behaviors from trained models.

Ax Zigeng Chen, Gongfan Fang, Xinyin Ma, Ruonan Yu, Xinchao Wang 7d ago

DMax: Aggressive Parallel Decoding for dLLMs

DMax enables efficient parallel decoding in diffusion language models through progressive self-refinement.

Ax Andrey Bocharnikov, Ivan Ermakov, Denis Kuznedelev, Vyacheslav Zhdanovskiy, Yegor Yershov 7d ago

KV Cache Offloading for Context-Intensive Tasks

KV cache offloading technique to reduce memory and latency overhead for long-context LLM inference.