Holo3 is a family of efficient, high-performance Vision-Language Models (VLMs) designed for autonomous agents, UI automation, and multimodal AI applications.Documentation Index
Fetch the complete documentation index at: https://hub.hcompany.ai/llms.txt
Use this file to discover all available pages before exploring further.
| Model | Active Parameters | Main Use Cases | License |
|---|---|---|---|
| Holo3-35B-A3B | 3B | High-throughput, low-latency | Apache 2.0 |
| Holo3-122B-A10B | 10B | Maximum performance, complex tasks | Research only |
Two ways to use Holo3
| Mode | Pattern | Output | When to use |
|---|---|---|---|
| Agent loop | Multi-turn: conversation + screenshots → next tool call | {note, thought, tool_call} against your tool union | Holo3 as the brain of an autonomous browser or desktop agent |
| Element localization | Single-turn: image + target description → coordinates | {x, y} in [0, 1000] | UI grounding inside any external agent or pipeline (yours or someone else’s) |
Get started
Generate an API key on Portal-H, export it, and point the OpenAI client at H Company’s endpoint:Build an autonomous agent
Holo3 can act as the brain of a browser, desktop, or mobile agent: it reads the screen, plans a move, calls a tool, observes the result, and iterates. The Agent loop guide covers the output JSON shape, chat layout, image budget, coordinate convention, and a complete loop you can drop into your harness.Element localization
Pass Holo3 a screenshot (URL or base64 data URI) and a text description of an element; get click coordinates back. Single-turn, no history, no thinking; settemperature=0.0 and enable_thinking=False. Useful as a vision tool inside any agent.
.png)
