-
CASA Gallery
🏠1Video Gallery for CASA: Cross-Attention via Self-Attention
-
CASA: Cross-Attention via Self-Attention for Efficient Vision-Language Fusion
Paper • 2512.19535 • Published • 10 -
kyutai/CASA-Helium1-VL-2B
Image-Text-to-Text • 3B • Updated • 146 • 5 -
kyutai/CASA-Qwen2_5-VL-3B
Image-Text-to-Text • 4B • Updated • 152 • 1
Collections
Discover the best community collections!
Collections including paper arxiv:2512.19535
-
Guided Self-Evolving LLMs with Minimal Human Supervision
Paper • 2512.02472 • Published • 50 -
DeepSearch: Overcome the Bottleneck of Reinforcement Learning with Verifiable Rewards via Monte Carlo Tree Search
Paper • 2509.25454 • Published • 140 -
Video Reasoning without Training
Paper • 2510.17045 • Published • 7 -
Agent Learning via Early Experience
Paper • 2510.08558 • Published • 269
-
A Survey on Vision-Language-Action Models: An Action Tokenization Perspective
Paper • 2507.01925 • Published • 38 -
Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning
Paper • 2507.16746 • Published • 35 -
MolmoAct: Action Reasoning Models that can Reason in Space
Paper • 2508.07917 • Published • 44 -
Discrete Diffusion VLA: Bringing Discrete Diffusion to Action Decoding in Vision-Language-Action Policies
Paper • 2508.20072 • Published • 31
-
Agentic Learner with Grow-and-Refine Multimodal Semantic Memory
Paper • 2511.21678 • Published • 11 -
QuCo-RAG: Quantifying Uncertainty from the Pre-training Corpus for Dynamic Retrieval-Augmented Generation
Paper • 2512.19134 • Published • 31 -
Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows
Paper • 2512.16969 • Published • 105 -
CASA: Cross-Attention via Self-Attention for Efficient Vision-Language Fusion
Paper • 2512.19535 • Published • 10
-
CASA Gallery
🏠1Video Gallery for CASA: Cross-Attention via Self-Attention
-
CASA: Cross-Attention via Self-Attention for Efficient Vision-Language Fusion
Paper • 2512.19535 • Published • 10 -
kyutai/CASA-Helium1-VL-2B
Image-Text-to-Text • 3B • Updated • 146 • 5 -
kyutai/CASA-Qwen2_5-VL-3B
Image-Text-to-Text • 4B • Updated • 152 • 1
-
A Survey on Vision-Language-Action Models: An Action Tokenization Perspective
Paper • 2507.01925 • Published • 38 -
Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning
Paper • 2507.16746 • Published • 35 -
MolmoAct: Action Reasoning Models that can Reason in Space
Paper • 2508.07917 • Published • 44 -
Discrete Diffusion VLA: Bringing Discrete Diffusion to Action Decoding in Vision-Language-Action Policies
Paper • 2508.20072 • Published • 31
-
Guided Self-Evolving LLMs with Minimal Human Supervision
Paper • 2512.02472 • Published • 50 -
DeepSearch: Overcome the Bottleneck of Reinforcement Learning with Verifiable Rewards via Monte Carlo Tree Search
Paper • 2509.25454 • Published • 140 -
Video Reasoning without Training
Paper • 2510.17045 • Published • 7 -
Agent Learning via Early Experience
Paper • 2510.08558 • Published • 269
-
Agentic Learner with Grow-and-Refine Multimodal Semantic Memory
Paper • 2511.21678 • Published • 11 -
QuCo-RAG: Quantifying Uncertainty from the Pre-training Corpus for Dynamic Retrieval-Augmented Generation
Paper • 2512.19134 • Published • 31 -
Probing Scientific General Intelligence of LLMs with Scientist-Aligned Workflows
Paper • 2512.16969 • Published • 105 -
CASA: Cross-Attention via Self-Attention for Efficient Vision-Language Fusion
Paper • 2512.19535 • Published • 10