Replies: 6 comments 7 replies
-
The work outlined above is in response to AEP-12 on the Akash Roadmap. The DOOOR team will be talking about their discussion and they will do a live demo at sig-providers monthly meeting April 23rd, 2025. Please track the Akash Community group calendar, and feel free to attend. |
Beta Was this translation helpful? Give feedback.
-
Hi @bruno353 - thanks for submitting this super detailed proposal and for the demo recordings. I'll start by saying that this is a very important piece of functionality that would bring a lot of value to the network, increase trust and provide a way to remove a common security based friction point for users/ companies adopting akash so I support the motivation and appreciate you talking the initiative on this. I reviewed the demos and have some questions (apologies if I'm unable to discern this from your writeup):
Thanks again! |
Beta Was this translation helpful? Give feedback.
-
Thank you for submitting your proposal to enhance the Akash Network. I appreciate the technical expertise and enthusiasm demonstrated by your team in addressing the challenges outlined here. It would be useful for the community to have some clarity on some of the technical decisions behind your proposal and its advantages over existing solutions, for example, how does it compare to some Kubernetes native solutions. As stated previously by @anilmurty this will require alignment with the Core Engineering team (@troian @chainzero @cloud-j-luna) on the impact of the required code changes. We’ll be able to have a more clear picture after the SIG Providers meeting which will allow me to provide deeper feedback and better understand the proposal. Related AEPs
This seems related to AEP-50. Related to verifiable compute AEP-29 aims to provide hardware verification. This could be relevant for the verifiable compute portion of the proposal. EDIT: Fixed the AEP links. |
Beta Was this translation helpful? Give feedback.
-
I support this proposal and am ready to help in anyway I can. |
Beta Was this translation helpful? Give feedback.
-
This proposal looks great. I also heard and saw some demos. As Anil mentioned above this looks like it would add a "piece of functionality that would bring a lot of value to the network". I would be interested in testing this when possible and also support this proposal. |
Beta Was this translation helpful? Give feedback.
-
I've created this AEP to outline the overall plan/ direction with Confidential Computing- FYI https://akash.network/roadmap/aep-65/ |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
1. Context of Proposal
DOOOR is pioneering a decentralized AI execution framework by integrating Trusted Execution Environments (TEEs) into the Akash Network, enabling secure and verifiable AI workloads. This initiative enhances Akash’s push into AI by ensuring privacy-preserving AI model execution and secure computing for LLMs (Large Language Models) and autonomous AI agents.
Our TEE integration SDK will enable AI developers and cloud providers on Akash to run workloads with cryptographic attestation, ensuring that sensitive AI models and computations are protected from unauthorized access. By supporting AMD SEV-SNP, NVIDIA H100, AWS Nitro Enclaves, and Azure Confidential Computing, DOOOR will provide an open-source framework for verifiable AI inference and execution within Akash.
2. Problem Statements
2.1 Lack of Automated TEE Management for AI Deployments
Problem: Complex and Error-Prone TEE Integration
Trusted Execution Environments (TEEs) provide an essential security layer for AI workloads, but their manual configuration and deployment on Akash are complex, inconsistent, and prone to errors. Current Akash deployments rely on standard Kubernetes-based container management, which lacks built-in attestation verification and automatic TEE enforcement. As a result, developers must manually set up, configure, and verify their workloads, leading to significant deployment overhead. This complexity discourages adoption and makes secure execution inaccessible to many AI developers.
Impact: Security Gaps and Slower Adoption
Without automated TEE integration, workloads cannot be easily verified, leading to security gaps where AI models may run on untrusted infrastructure. This discourages enterprises and privacy-conscious developers from using Akash, as they cannot ensure their data remains secure during execution. Furthermore, the time-consuming manual setup limits scalability, making it difficult for providers to offer TEE-backed services at scale.
2.2 Fragmented & Incompatible TEE Workflows Across Cloud Providers
Problem: No Unified Approach to Secure Compute on Akash
Akash aims to be a decentralized alternative to major cloud providers, but the lack of standardization in TEE workflows across different hardware and cloud environments creates fragmentation. Providers using Azure Confidential VMs, AWS Nitro Enclaves, IBM Cloud Hyper Protect, or self-hosted hardware each require different verification and attestation methods, making it difficult for developers to deploy across multiple providers. This lack of interoperability means AI workloads cannot easily migrate between providers, limiting the network’s efficiency.
Impact: Vendor Lock-in and Barriers to Adoption
Because each cloud provider requires different security protocols, developers are often locked into specific providers, reducing the decentralized value of Akash. Workloads that should be able to migrate between providers seamlessly are instead restricted to specific hardware, limiting availability, scalability, and pricing flexibility. The barriers to standardization also slow innovation, as developers must rewrite security protocols for each environment, delaying deployments and increasing costs.
2.3 No Verifiable AI Execution on Akash (GPU Attestation)
Problem: AI Inference Workloads Lack Security Guarantees
Machine learning and AI inference workloads on Akash lack any form of verifiable execution guarantees, meaning users cannot trust that their AI model is running on genuine, unmodified hardware. Unlike centralized cloud services, where security certifications and audits verify compute environments, Akash currently provides no cryptographic proof that an AI model is executing in a secure and unaltered GPU environment.
Impact: Risks of Model Theft, Data Manipulation, and AI Bias
Without hardware attestation and verifiable execution, AI model providers are at risk of model theft, where an untrusted provider extracts or manipulates a deployed model. Additionally, data processed on unverified GPUs could be compromised, leading to biased AI results, security breaches, or tampered outputs. This lack of trust deters high-value AI applications—such as financial modeling, autonomous vehicle AI, and healthcare AI—from using Akash due to security concerns.
2.4 Absence of Secure & Decentralized Key Management for AI Models
Problem: Centralized Key Storage Risks and No Encrypted Model Execution
AI workloads require secure storage and execution of private cryptographic keys, especially when handling proprietary models, encrypted datasets, or zero-knowledge computations. However, current Akash deployments lack decentralized private key management. Without an established framework for private key sharing, sealing, and attestation verification, AI models running on Akash are vulnerable to key exposure, unauthorized access, or centralized attack vectors.
Impact: Loss of Privacy and Security Vulnerabilities
Without trustless key management, sensitive AI workloads cannot be protected, making Akash an unsuitable environment for enterprise-grade AI applications. Existing cloud services use proprietary HSM (Hardware Security Modules) or centralized key vaults, but these introduce single points of failure and limit decentralization. A decentralized key bootstrap system, where AI models can securely share and verify encryption keys, is necessary to guarantee data integrity and confidentiality for sensitive computations.
2.5 No Cross-Provider TEE Attestation Mechanism
Problem: Users Cannot Verify That Their Workloads Run in a Trusted Environment
Even if AI models and workloads are deployed on Akash, there is currently no built-in way for users to verify that their workloads are running inside a TEE-protected environment. Unlike traditional cloud providers that offer security compliance guarantees (e.g., ISO 27001, SOC 2), Akash lacks a decentralized attestation mechanism that allows users to check if their AI workload is actually executing inside a trusted enclave.
Impact: No Confidence in Data Confidentiality & Execution Integrity
Without a cross-provider attestation system, AI model owners and enterprises cannot trust that data confidentiality and execution integrity are maintained. This lack of transparency prevents privacy-sensitive industries from adopting Akash, such as healthcare AI (HIPAA compliance), finance (confidential trading algorithms), and defense AI applications. Establishing hardware-backed, remote attestation across Akash providers is essential to build trust and attract high-value workloads.
3. Implementation Plan & Feature Set
Below you may find relevant Demos, Github repositories and additional technical documentation on the features, their specs and code snippets:
*Note: The Decentralized Key Sharing Management Phase is still in the design stage, and its full viability is yet to be assessed. At this stage, we commit to delivering a comprehensive report outlining the requirements, feasibility, and proposed design of the solution. Based on these findings, we will be able to better estimate the timeline, effort, and resources needed to implement the final scope effectively.
4. Projected allocation & budget
To successfully develop and deploy a fully verifiable TEE-based AI execution framework on Akash, we are requesting $331,000 in funding, carefully allocated to cover infrastructure, security audits, development, and operational costs. This budget ensures that we can build, test, and deploy the project over a period of 8 months, delivering a scalable, open-source solution that strengthens Akash’s security and AI capabilities.
**Note: This is an estimated amount, and costs may fluctuate based on audit firm rates and bug bounty program pricing.
Conclusion
The DOOOR protocol is committed to strengthening Akash’s ecosystem by building trusted execution environments (TEEs) and verifiable AI execution, ensuring secure, private, and decentralized computing. Akash was chosen because of its decentralized, cost-effective, and censorship-resistant infrastructure, making it the ideal platform for confidential AI computing. Our work will enhance security, attract enterprise AI users, and position Akash as the go-to network for privacy-first AI deployments. This project is not just something we want to build—it is something Akash needs to advance its security and attract high-value workloads.
Beyond technology, we are deeply invested in Akash’s community and want to ensure that our contributions bring direct, lasting value to providers, developers, and users. By making our TEE-based security tools open-source, providing educational resources, and incentivizing community-driven security testing, we ensure that Akash’s entire ecosystem gains long-term value. Additionally, we are committed to collaborating both on and off the field—not only by strengthening Akash’s infrastructure but also by actively engaging with its community. As we progress through testnet and mainnet launches, we will introduce initiatives that drive adoption, engagement, and long-term participation, ensuring that our success directly benefits the broader Akash community and ecosystem.
This proposal is about long-term commitment. We are dedicated builders in the Akash ecosystem, and this project is only the beginning. With this funding, we will develop the first fully verifiable TEE-based AI execution framework, setting the stage for future security innovations. We see this as an ongoing partnership where we continue to enhance Akash’s capabilities and expand privacy and security features in decentralized AI computing. Investing in this proposal is investing in Akash’s future as the leader in secure, decentralized AI infrastructure.
Appendix
Links & References
Full Demo TEE Mainnet (with explanation)
Pocket Demo TEE Mainnet
Detailed Technical Explanation with Code Snippets
GitHub Proposal Discussion
Team Background
Thiago Castroneves — CEO
Thiago de Castro Neves, Co-Founder and CEO of DOOOR, brings a unique blend of engineering, management, and blockchain expertise to the forefront of decentralized AI execution. With over 15 years of experience in the industrial engineering, he served as an engineer, manager and innovation lead for a global steel multinational, honing his skills in strategic leadership, infrastructure development, and operations management. Transitioning into the blockchain industry, Thiago played a pivotal role as Head of Business Development at Moonbeam. He led the Grants Program, interviewing 700+ teams around the world, and leading ecosystem growth initiatives, which today represents more than 60% of the overall transactions in the Moonbeam Network. His deep technical and strategic acumen now drives DOOOR’s mission to redefine secure AI execution in decentralized networks.
Bruno Laureano — CTO
Bruno Laureano dos Santos, Co-Founder and CTO of DOOOR, brings extensive expertise in AI, blockchain technology, and full-stack development, with exceptional back-end and front-end engineering skills. His ability to lead and manage high-performing technical teams has been instrumental in driving innovation, adapting to the rapidly evolving AI and blockchain landscape, and building cutting-edge decentralized solutions. With a deep understanding of trustless computing and secure AI execution, Bruno is at the forefront of DOOOR’s mission to develop scalable, secure, and efficient AI frameworks for decentralized ecosystems.
Contact Information
Background on what is Dooor and how it works with Akash
By enabling a set of tenant providers to rent computing power on a decentralized marketplace, Akash serves as a crucial layer in the Dooor protocol, allowing LLMs to run on TEE GPUs while connecting with context and fine-tuning data from data storages. This integration is key to the autonomous execution of agents with the orchestration layer. Through HTTPS calls, secure key storage (currently under development with VetKeys), and the T-ECDSA protocol for signing multi-chain messages and transactions, canisters form a network of trustless oracles that underpin the orchestration layer for the Akash chain. This creates a true on-chain environment for AI without reliance on centralized institutions. Furthermore, the smart contracts in the Aggregation Layer are mapped to the data peer tools' canisters, ensuring synchronization of AI data across chains.
The Dooor architecture introduces an innovative infrastructure that unlocks unprecedented use cases for the Akash ecosystem. By leveraging canister utilization as key components in a multi-tenant LLM management system, Dooor enables the first truly decentralized AI ecosystem on Akash, complete with real-world sub-applications.
Currently implemented with the Azle TypeScript framework, the canisters maintain their own database relationships for critical, incorruptible data points. These canisters facilitate user interactions with agents, context, and LLMs on Akash. A multiset of canisters ensures data redundancy for corruption resistance, while a master canister, governed by the Dooor DAO, ensures decentralization throughout the workflow. This architecture enables on-chain management of a swarm of community-deployed agents without intermediaries, with LLM scripts stored and executed on the computing layer.
The Dooor backend, deployed on Akash, employs a load balancer system to guarantee consistent LLM availability and fast responses for end users. The current implementation uses Flask to manage user credits and agent configurations. The developer workflow involves a Docker container that creates an image for upload, along with an
SDL
file and astartup.sh
script to initialize the framework for running LLM instances. The backend server manages the business logic.For fine-tuning operations, Unsloth is used for intensive data handling on the backend, receiving commands and data directory authorizations from the ICP canister. For running LLMs, Ollama is employed to retrieve
.gguf
models from a shared Dooor database, with the business logic also controlled by the canister.Why Not Use the Akash Computing Layer Directly?
The Akash deployment process includes a centralized step for the deployer, requiring a certificate generated locally on the user’s computer to interact with containers. This step creates a potential point of failure, as access to the certificate is necessary to provide the one-time SDL rules to the provider, which could allow unauthorized access to Dooor's provider configurations.
Using ICP canister management, this Akash certificate can be generated in-memory and the one-time SDL configuration sent to the provider without storing the certificate data. This approach ensures a fully decentralized flow. Once the configuration is completed, all communication between the canister and Akash can be securely managed using T-ECDSA.
Dooor's mission is to become an aggregation protocol for AI agents. By integrating with the Akash chain, the protocol ensures redundancy in its infrastructure resources, leveraging Akash's unique capabilities while offering competitive pricing for end users. With the load-balancer system under development by Dooor—a queue-cron system within the canister that coordinates providing bids and resources on Akash—the protocol replicates the complexity and reliability of infrastructures that are typically only achievable with centralized solutions.
How Does It Work?
EVM smart contracts handle user transactions and manage token pools. Mapping and queues allow users to create orders (such as uploading, patching, or deleting models, tools, and contexts). ICP canisters perform continuous HTTPS calls via a set of Base RPCs (acting as oracles) to gather transaction data, process it according to the defined business logic, and interact with the Akash-ICP SDK for provider management on the Akash layer.
Users or tenants specify deployment parameters, including data centers, requirements, and pricing, in a manifest file (
deploy.yaml
). This file, written in the declarative Stack Definition Language (SDL), simplifies the process of defining deployment attributes.The system is designed with idempotence in mind, ensuring that even in extreme scenarios, redundant calls can reliably guarantee the successful execution of transactions without failure.
Dooor is developing streamlined functions to convert manifests from LLM developers into readable SDL for Akash, enabling efficient deployment of virtually any tenant. This ensures flexibility and simplifies the deployment process for a wide range of use cases.
ZK, privacy and multichain operations
Dooor’s vision prioritizes privacy for developers and users, ensuring that no external entity—individual or company—can extract user code, content, or manipulate data in a malicious way (even when using TEE GPUs). To achieve this, Dooor developed the ECDSA-JWT interauth system. In this system, the canister generates a time-based expiration token via its T-ECDSA signature, enabling users to directly interact with agents and models running on the computing layer. This approach enhances the speed of data updates and retrievals while reducing the load on the canister system, as it no longer needs to approve every interaction.
The ECDSA-JWT inter-auth system moves part of the business logic to the computing layer, such as credit systems, view permissions, and ZK storage authentication. Only authorized users can access specific data, with the server provider verifying that the signature was derived and created by the canister's public key. This ensures secure and controlled access without compromising user privacy.
By enabling the provider to validate user identity and approve guard authentication to process requests, the system allows for responses to be hashed and copied at the computing layer. This ensures agent responses can be verified without requiring public access to the provider’s code.
To facilitate seamless integration with the Akash SDK, Dooor developed new interfaces for the canister. These interfaces ensure compatibility, as the default Akash SDK would otherwise disrupt canister execution.
ICP HTTPS: The HTTPS outcalls (https://internetcomputer.org/docs/current/references/https-outcalls-how-it-works) enabled by ICP are a crucial component of its integration with Akash, allowing canisters to interact with external environments. All the necessary properties to initiate bidding (https://akash.network/docs/akash-provider-service-and-associated-sub-services/bid-engine-overview/) can be implemented by the canister in a workflow that involves signing transactions and sending them to external systems. With an RPC swarm developed within the canister, the Dooor framework ensures connectivity with virtually any computing chain layer, serving as the intermediary between agents running on the computing layer and their mapped addresses on ICP. This setup allows each agent to have its own blockchain address, enabling them to perform truly autonomous operations through the canister.
Beta Was this translation helpful? Give feedback.
All reactions