Training Foundation Models on a Full-Stack AMD Platform: Compute, Networking, and System Design Paper • 2511.17127 • Published 17 days ago • 1
Think-at-Hard: Selective Latent Iterations to Improve Reasoning Language Models Paper • 2511.08577 • Published 27 days ago • 104
RLVE: Scaling Up Reinforcement Learning for Language Models with Adaptive Verifiable Environments Paper • 2511.07317 • Published 28 days ago • 13
Revisiting Multimodal Positional Encoding in Vision-Language Models Paper • 2510.23095 • Published Oct 27 • 20
When "Correct" Is Not Safe: Can We Trust Functionally Correct Patches Generated by Code Agents? Paper • 2510.17862 • Published Oct 15 • 6
Glyph: Scaling Context Windows via Visual-Text Compression Paper • 2510.17800 • Published Oct 20 • 67
Efficient Parallel Samplers for Recurrent-Depth Models and Their Connection to Diffusion Language Models Paper • 2510.14961 • Published Oct 16 • 7
OmniVinci: Enhancing Architecture and Data for Omni-Modal Understanding LLM Paper • 2510.15870 • Published Oct 17 • 89
Cache-to-Cache: Direct Semantic Communication Between Large Language Models Paper • 2510.03215 • Published Oct 3 • 97
BroRL: Scaling Reinforcement Learning via Broadened Exploration Paper • 2510.01180 • Published Oct 1 • 18