Ax Manish Bhatt, Sarthak Munshi, Vineeth Sai Narajala, Idan Habler, Ammar Al-Kahfah, Ken Huang, Joel Webb, Blake Gatto 8d ago

The Defense Trilemma: Why Prompt Injection Defense Wrappers Fail?

Formal proof that continuous wrapper defenses cannot protect LLMs from all prompt injection attacks, characterizing where every defense must fail.

Ax Jiang Zhou, Yunhao Wang, Xing Wu, Tinghao Yu, Feng Zhang 8d ago

WRAP++: Web discoveRy Amplified Pretraining

WRAP++ enhances LLM pretraining through synthetic data rephrasing that captures cross-document relationships beyond single-document web page rewriting.

Ax Yucheng Sheng, Jiacheng Wang, Le Liang, Hao Ye, Shi Jin 8d ago

A Graph Foundation Model for Wireless Resource Allocation

Graph foundation model for wireless network resource allocation using deep learning to solve optimization problems more efficiently than classical iterative algorithms.

Ax Oleg Platonov, Liudmila Prokhorenkova 8d ago

Cluster Attention for Graph Machine Learning

Cluster Attention mechanism for graph transformers improving receptive field while preserving graph-structure inductive biases.

Ax Henry C. Conklin, Tom Hosking, Tan Yi-Chern, Julian Gold, Jonathan D. Cohen, Thomas L. Griffiths, Max Bartolo, Seraphina Goldfarb-Tarrant 8d ago

Learning is Forgetting: LLM Training As Lossy Compression

LLM training as lossy compression framework explaining how LLMs learn by retaining task-relevant information from training data.