H Company releases Holotron-12B, a 12B-parameter multimodal computer-use agent model post-trained from NVIDIA's Nemotron-Nano-2 VL. The model uses a hybrid State-Space Model (SSM) and attention architecture that avoids the quadratic cost of full attention, enabling a constant memory footprint per layer regardless of sequence

4m read timeFrom huggingface.co
Post cover image
Table of contents
Holotron-12B - High Throughput Computer Use AgentWhy We Built Holotron-12BConclusionWhat’s next: Scaling the Future of Agentic Intelligence with Nemotron 3 Omni

Sort: