{"id":"hone","title":"Hone","content":"Hone (formerly OpenKaito and known on-chain as Subnet-5) is a decentralized artificial intelligence subnet on the [Bittensor](https://iq.wiki/wiki/bittensor-tao) network that focuses on training, evaluating, and serving hierarchical learning and reasoning systems through distributed participation. The project emphasizes world-model learning and iterative, planner–executor reasoning loops, with performance-based evaluation by independent validators and an explicit focus on improving abstract problem-solving benchmarks. Hone operates as an open, incentive-aligned research and infrastructure effort rather than a single static model release. [\\[1\\]](#cite-id-7iVHc8UhiMIws0rl)​\n\n## Overview\n\nHone (formerly OpenKaito) is a decentralized AI and search initiative built on the [Bittensor](https://iq.wiki/wiki/bittensor-tao) network, focused on creating an open, continuously improving layer for information access and reasoning. It combines decentralized search infrastructure with ongoing model training to support transparent, composable, and scalable information retrieval.\n\nAs Hone (Subnet-5), the project centers on training general-purpose AI models using hierarchical learning and reasoning architectures. Instead of relying on conventional large-language model scaling, it develops systems in which higher-level planning components and lower-level execution components interact iteratively, enabling structured, multi-step problem-solving rather than one-pass outputs. The subnet functions as a distributed training and evaluation system. Participants contribute models and training processes, while independent [validators](https://iq.wiki/wiki/validator) assess performance using evolving datasets and benchmarks. Incentives are tied to measurable results, creating a continuous cycle of competition and improvement that refines model quality over time without centralized control.\n\nA core objective is to improve performance on difficult reasoning benchmarks, such as ARC-AGI-2, which measure abstract problem-solving ability. To achieve this, the system incorporates approaches such as joint embedding predictive architectures and hierarchical reasoning models, enabling models to build internal representations, plan solutions, and adapt to new tasks. The output is not a single fixed model but an evolving, decentralized AI system. Over time, these models are intended to be accessible through network APIs, supporting applications such as semantic search, complex analysis, and general reasoning, while remaining open, continuously updated, and collectively developed. [\\[1\\]](#cite-id-7iVHc8UhiMIws0rl) [\\[3\\]](#cite-id-kbxn7Jx69vXUaapn) \n\n![](https://ipfs.everipedia.org/ipfs/QmeFZ4nvpMjqEeV15eAMy9QbPKe19TnKvLyMH7oZP83BT4)\n\n## History\n\nHone (formerly OpenKaito) originated as Subnet-5 within the [Bittensor](https://iq.wiki/wiki/bittensor-tao) ecosystem, initially focused on decentralized search and text embedding models. On May 1, 2025, ownership of the subnet was transferred from its original creator to Latent Holdings, marking a shift in direction and scope. Following the transfer, a new development strategy was introduced in collaboration with Manifold Labs. Rather than continuing to focus primarily on search and embeddings, the subnet was repositioned toward building general-purpose reasoning systems using hierarchical AI architectures. This transition reflected a broader move from information retrieval toward more advanced AI capabilities.\n\nOn August 7, 2025, Subnet-5 was formally relaunched as Hone, reflecting its updated technical focus and research direction. From that point onward, the project centered on decentralized training of reasoning models, integrating approaches such as world-model learning and iterative reasoning within the [Bittensor](https://iq.wiki/wiki/bittensor-tao) framework. The development of Hone is led jointly by Manifold Labs and Latent Holdings. Manifold contributes experience in building and operating [Bittensor](https://iq.wiki/wiki/bittensor-tao) subnets and related infrastructure, while Latent provides strategic direction focused on open-source AI development. Together, the two organizations established Hone as a decentralized research and infrastructure effort aimed at advancing general-purpose AI systems. Since its relaunch, Hone has evolved from a specialized, embedding-focused subnet into a broader platform for distributed training and evaluation of reasoning models. [\\[1\\]](#cite-id-7iVHc8UhiMIws0rl) [\\[2\\]](#cite-id-ae2UF5X9iqXWEY6m) \n\n## Architecture\n\n### World-Model Learning\n\nThe World-Model Learning module focuses on self-supervised predictive learning, drawing from joint embedding predictive architectures (JEPA) and hierarchical extensions. The core objective is to train the model to understand data structure by predicting missing or masked information, rather than relying on labeled supervision. In practice, the model is exposed to partial inputs—such as incomplete patterns, sequences, or structured data—and must infer the latent representation of what is missing. This encourages the system to learn relationships between variables, causal structure, and context, forming a compressed internal representation often described as a “world model.”\n\nA key aspect is hierarchical representation learning. Instead of learning a single flat embedding space, the model develops multiple levels of abstraction simultaneously—for example, low-level features (patterns, tokens) and higher-level concepts (rules, relationships). These layers allow the system to generalize more effectively across domains, since it can reuse abstract knowledge rather than memorizing surface-level correlations. This module serves as the foundation of the system: it equips the model with broad, transferable knowledge and contextual understanding, which is necessary for downstream reasoning tasks. Without this step, reasoning modules tend to operate on shallow or brittle representations. [\\[1\\]](#cite-id-7iVHc8UhiMIws0rl) \n\n### Hierarchical Reasoning\n\nThe Hierarchical Reasoning module builds on top of the learned representations and introduces an explicit multi-step reasoning process inspired by hierarchical reasoning architectures. Instead of generating outputs in a single forward pass, the model is structured into two interacting components that operate over multiple iterations. The high-level component functions as a planner. It generates an abstract strategy, hypothesis, or partial solution based on the input and the current state of reasoning. The low-level component acts as an executor, taking that plan and working through concrete steps—such as filling in details, testing transformations, or constructing intermediate outputs.\n\nThese components interact in a recurrent loop. After each iteration, the executor's output is fed back into the planner, which evaluates progress and decides whether to refine the strategy, adjust assumptions, or terminate the process. This creates a dynamic reasoning cycle where the model incrementally improves its solution. This design allows the system to handle tasks that require decomposition, backtracking, and iterative refinement—capabilities that are limited in standard transformer-based models. It also enables reasoning across different time scales: slower, higher-level planning guides faster, detail-oriented execution. The hierarchical structure also improves computational efficiency. Instead of scaling a single large model to handle all reasoning in one pass, the system reuses smaller components across multiple iterations, achieving deeper reasoning without a proportional increase in model size. [\\[1\\]](#cite-id-7iVHc8UhiMIws0rl) ","summary":"Hone (formerly OpenKaito) is Bittensor’s Subnet-5, a decentralized AI subnet that trains hierarchical reasoning models via distributed miners and validator scoring, combining world-model learning with iterative planner–executor loops and open network APIs","images":[{"id":"QmSPA36RZacYmk8ADBjrm4hVkzv2G1xXKupsvgX7riwtvX","type":"image/jpeg, image/png"}],"categories":[{"id":"organizations","title":"organizations"}],"tags":[{"id":"AI"},{"id":"Organizations"},{"id":"Protocols"},{"id":"Developers"}],"media":[{"id":"QmXNCcgWfvcx8aDukbLEsb4hTRe8VWwrGFmgmfkAMWJThK","type":"GALLERY","source":"IPFS_IMG"},{"id":"QmepS3ctm8QfRXG9coeM4ZgBdgueVGHpaMUjzFDnc617jx","type":"GALLERY","source":"IPFS_IMG"},{"id":"QmeFZ4nvpMjqEeV15eAMy9QbPKe19TnKvLyMH7oZP83BT4","name":"YDzoG2wotyYnjYyc.png","caption":"","thumbnail":"QmeFZ4nvpMjqEeV15eAMy9QbPKe19TnKvLyMH7oZP83BT4","source":"IPFS_IMG"}],"metadata":[{"id":"references","value":"[\n {\n \"id\": \"7iVHc8UhiMIws0rl\",\n \"url\": \"https://subnetalpha.ai/subnet/hone/\",\n \"description\": \"Subnet overview and aims\",\n \"timestamp\": 1776863152972\n },\n {\n \"id\": \"ae2UF5X9iqXWEY6m\",\n \"url\": \"https://github.com/OpenKaito/openkaito\",\n \"description\": \"OpenKaito repository overview\",\n \"timestamp\": 1776863152972\n },\n {\n \"id\": \"kbxn7Jx69vXUaapn\",\n \"url\": \"https://wandb.ai/sn-openkaito-openkaito/sn5-validators/reports/OpenKaito-Text-Embedding-Models--Vmlldzo5NjUwODAz\",\n \"description\": \"SN5 validator dashboards (embeddings)\",\n \"timestamp\": 1776863152972\n },\n {\n \"id\": \"QyHl2IJubEPspTRX\",\n \"url\": \"https://www.coingecko.com/en/coins/openkaito\",\n \"description\": \"CoinGecko market snapshot\",\n \"timestamp\": 1776863152972\n }\n]"},{"id":"website","value":"https://www.hone.training/"},{"id":"github_profile","value":"https://github.com/manifold-inc/hone"},{"id":"coinmarketcap_url","value":"https://coinmarketcap.com/currencies/openkaito/"},{"id":"coingecko_profile","value":"https://www.coingecko.com/en/coins/openkaito"},{"id":"references","value":"https://subnetalpha.ai/subnet/hone/"},{"id":"contract_url","value":"https://taostats.io/subnets/5/chart"},{"id":"commit-message","value":"\"Added Hone decentralized AI subnet overview and organizations category\""}],"events":[{"id":"c49f8bb1-4ecb-41d6-9c4c-5c04b07ebfba","date":"2025-08","title":"Relaunched as Hone on Subnet-5","type":"CREATED","description":"Subnet-5 formally rebranded and relaunched as Hone with a focus on hierarchical reasoning and world-model learning.","link":"https://subnetalpha.ai/subnet/hone/","multiDateStart":null,"multiDateEnd":null,"continent":null,"country":null},{"id":"6d6f7572-a8a2-4e97-8e87-19d6d47b2f4c","date":"2025-05","title":"Ownership transferred to Latent Holdings","type":"DEFAULT","description":"Control of Subnet-5 transferred from the original creator to Latent Holdings, initiating a strategic shift.","link":"https://subnetalpha.ai/subnet/hone/","multiDateStart":null,"multiDateEnd":null,"continent":null,"country":null},{"id":"727095ff-ca63-434a-84e1-6dec76963c83","date":"2024-10","title":"OpenKaito validator report published on W&B","type":"DEFAULT","description":"Weights & Biases report documenting SN5 validator dashboards for text embedding models.","link":"https://wandb.ai/sn-openkaito-openkaito/sn5-validators/reports/OpenKaito-Text-Embedding-Models--Vmlldzo5NjUwODAz","multiDateStart":null,"multiDateEnd":null,"continent":null,"country":null},{"id":"59bdb3de-33c2-4a01-aa78-2dfc1cdb5449","date":"2025-10","title":"Hash Rate interview features Hone","type":"DEFAULT","description":"Mark Jeffrey’s Hash Rate podcast features Hone and Manifold Labs discussing the subnet’s direction.","link":null,"multiDateStart":null,"multiDateEnd":null,"continent":null,"country":null}],"user":{"id":"0x8af7a19a26d8fbc48defb35aefb15ec8c407f889"},"author":{"id":"0x8af7a19a26d8fbc48defb35aefb15ec8c407f889"},"operator":{"id":"0xacb6c5AD52b8f605299B0d774CE97F26e3DB80c2"},"language":"en","version":1,"linkedWikis":{"blockchains":["bittensor-tao"],"founders":[],"speakers":[]}}