OpenAI just confirmed my northern star thesis for AI today by releasing their operator agent. Not only was this my guiding thesis for $CODEC, but every other AI investment I made, including those from earlier in the year during AI mania. There’s been a lot of discussion with Codec in regards to Robotics, while that vertical will have its own narrative very soon, the underlying reason I was so bullish on Codec from day 1 is due to how its architecture powers operator agents. People still underestimate how much market share is at stake by building software that runs autonomously, outperforming human workers without the need for constant prompts or oversight. I’ve seen a lot of comparisons to $NUIT. Firstly I want to say I’m a big fan of what Nuit is building and wish nothing but for their success. If you type “nuit” into my telegram, you’ll see that back in April I said that if I had to hold one coin for multiple months it would have been Nuit due to my operator thesis. Nuit was the most promising operator project on paper, but after extensive research, I found their architecture lacked the depth needed to justify a major investment or putting my reputation behind it. With this in mind, I was already aware of the architectural gaps in existing operator agent teams and actively searching for a project that addressed them. Shortly after Codec appeared (thanks to @0xdetweiler insisting I look deeper into them) and this is the difference between the two: $CODEC vs $NUIT Codec’s architecture is built across three layers; Machine, System, and Intelligence, that separate infrastructure, environment interface, and AI logic. Each Operator agent in Codec runs in its own isolated VM or container, allowing near native performance and fault isolation. This layered design means components can scale or evolve independently without breaking the system. Nuit’s architecture takes a different path by being more monolithic. Their stack revolves around a specialized web browser agent that combines parsing, AI reasoning, and action. Meaning they deeply parse web pages into structured data for the AI to consume and relies on cloud processing for heavy AI tasks. Codec’s approach of embedding a lightweight Vision-Language-Action (VLA) model within each agent means it can run fully local. Which doesn’t require constant pinging back to the cloud for instructions, cutting out latency and avoiding dependency on uptime and bandwidth. Nuit’s agent processes tasks by first converting web pages into a semantic format and then using an LLM brain to figure out what to do, which improves over time with reinforcement learning. While effective for web automation, this flow depends on heavy cloud side AI processing and predefined page structures. Codec’s local device intelligence means decisions happen closer to the data, reducing overhead and making the system more stable to unexpected changes (no fragile scripts or DOM assumptions). Codec’s operators follow a continuous perceive–think–act loop. The machine layer streams the environment (e.g. a live app or robot feed) to the intelligence layer via the system layer’s optimized channels, giving the AI “eyes” on the current state. The agent’s VLA model then interprets the visuals and instructions together to decide on an action, which the System layer executes through keyboard/mouse events or robot control. This integrated loop means it adapts to live events, even if the UI shifts around, you won’t break the flow. To put all of this in a more simple analogy, think of Codec’s operators like a self sufficient employee who adapts to surprises on the job. Nuit’s agent is like an employee who needs to pause, describe the situation to a supervisor over the phone, and wait for instructions. Without going down too much of a technical rabbit hole, this should give you a high level idea on why I chose Codec as my primary bet on Operators. Yes Nuit has backing from YC, a stacked team and S tier github. Although Codec’s architecture has been built with horizontal scaling in mind, meaning you can deploy thousands of agents in parallel with zero shared memory or execution context between agents. Codec’s team isn’t your average devs either. Their VLA architecture opens a multitude of use cases which wasn’t possible with previous agent models due to seeing through pixels, not screenshots. I could go on but I’ll save that for future posts.
Virtual Environments for Operator Agents: $CODEC My core thesis around the explosion of AI has always centered on the rise of operator agents. But for these agents to succeed, they require deep system access, effectively granting them control over your personal computer and sensitive data, which introduces serious security concerns. We’ve already seen how companies like OpenAI and other tech giants handle user data. While most people don’t care, the individuals who stand to benefit most from operator agents, the top 1% absolutely do. Personally, there's zero chance I’m giving a company like OpenAI full access to my machine, even if it means a 10× boost in productivity. So why Codec? Codec’s architecture is centered on launching isolated, on-demand “cloud desktops” for AI agents. At its core is a Kubernetes-based orchestration service (codenamed Captain) that provisions lightweight virtual machines (VMs) inside Kubernetes pods. Each agent gets its own OS-level isolated environment (a full Linux OS instance) where it can run applications, browsers, or any code, completely sandboxed from other agents and the host. Kubernetes handles scheduling, auto-scaling, and self-healing of these agent pods, ensuring reliability and the ability to spin up/down many agent instances as load demands Trusted Execution Environments (TEEs) are used to secure these VMs, meaning the agent’s machine can be cryptographically isolated, its memory and execution can be protected from the host OS or cloud provider. This is crucial for sensitive tasks: for example, a VM running in an enclave could hold API keys or crypto wallet secrets securely. When an AI agent (an LLM-based “brain”) needs to perform actions, it sends API requests to the Captain service, which then launches or manages the agent’s VM pod. The workflow: the agent requests a machine, Captain (through Kubernetes) allocates a pod and attaches a persistent volume (for the VM’s disk). The agent can then connect into its VM (via a secure channel or streaming interface) to issue commands. Captain exposes endpoints for the agent to execute shell commands, upload/download files, retrieve logs, and even snapshot the VM for later restoration. This design gives the agent a full operating system to work in, but with controlled, audited access. Because it’s built on Kubernetes, Codec can auto-scale horizontally, if 100 agents need environments, it can schedule 100 pods across the cluster, and handle failures by restarting pods. The agent’s VM can be equipped with various MCP servers (like a “USB port” for AI). For example, Codec’s Conductor module is a container that runs a Chrome browser along with a Microsoft Playwright MCP server for browser control. This allows an AI agent to open web pages, click links, fill forms, and scrape content via standard MCP calls, as if it were a human controlling the browser. Other MCP integrations could include a filesystem/terminal MCP (to let an agent run CLI commands securely) or application-specific MCPs (for cloud APIs, databases, etc.). Essentially, Codec provides the infrastructure “wrappers” (VMs, enclaves, networking) so that high-level agent plans can safely be executed on real software and networks. Use Cases Wallet Automation: Codec can embed wallets or keys inside a TEE-protected VM, allowing an AI agent to interact with blockchain networks (trade on DeFi, manage crypto assets) without exposing secret keys. This architecture enables onchain financial agents that execute real transactions securely, something that would be very dangerous in a typical agent setup. The platform’s tagline explicitly lists support for “wallets” as a key capability. An agent could, for instance, run a CLI for an Ethereum wallet inside its enclave, sign transactions, and send them, with the assurance that if the agent misbehaves, it’s confined to its VM and the keys never leave the TEE. Browser and Web Automation: CodecFlow agents can control full web browsers in their VM. The Conductor example demonstrates an agent launching Chrome and streaming its screen to Twitch in real-time. Through the Playwright MCP, the agent can navigate websites, click buttons, and scrape data just like a human user. This is ideal for tasks like web scraping behind logins, automated web transactions, or testing web apps. Traditional frameworks usually rely on API calls or simple headless browser scripts; in contrast, CodecFlow can run a real browser with a visible UI, making it easier to handle complex web applications (e.g. with heavy JavaScript or CAPTCHA challenges) under AI control. Real-World GUI Automation (Legacy Systems): Because each agent has an actual desktop OS, it can automate legacy GUI applications or remote desktop sessions, essentially functioning like robotic process automation (RPA) but driven by AI. For example, an agent could open an Excel spreadsheet in its Windows VM, or interface with an old terminal application that has no API. Codec’s site mentions enabling “legacy automation” explicitly. This opens up using AI to operate software that isn’t accessible via modern APIs, a task that would be very hacky or unsafe without a contained environment. The included noVNC integration suggests agents can be observed or controlled via VNC, which is useful for monitoring an AI driving a GUI. Simulating SaaS Workflows: Companies often have complex processes that involve multiple SaaS applications or legacy systems. for example, an employee might take data from Salesforce, combine it with data from an internal ERP, then email a summary to a client. Codec can enable an AI agent to perform this entire sequence by actually logging into these apps through a browser or client software in its VM, much like a human would. This is like RPA, but powered by an LLM that can make decisions and handle variability. Importantly, credentials to these apps can be provided to the VM securely (and even enclosed in a TEE), so the agent can use them without ever “seeing” plaintext credentials or exposing them externally. This could accelerate automation of routine back office tasks while satisfying IT that each agent runs with least privilege and full auditability (since every action in the VM can be logged or recorded). Roadmap - Launch public demo at end of the month - Feature comparison with other similar platforms (no web3 competitor) - TAO Integration - Large Gaming Partnership In terms of originality, Codec is built on a foundation of existing technologies but integrates them in a novel way for AI agent usage. The idea of isolated execution environments is not new (containers, VMs, and TEEs are standard in cloud computing), but applying them to autonomous AI agents with a seamless API layer (MCP) is extremely novel. The platform leverages open standards and tools wherever possible: it uses MCP servers like Microsoft’s Playwright for browser control instead of reinventing that wheel, and plans to support AWS’s Firecracker micro-VMs for faster virtualization. It also forked existing solutions like noVNC for streaming desktops. Demonstrating the project is standing on the foundations of proven tech (Kubernetes, enclave hardware, open-source libraries), focusing its original development on glue logic and orchestration (the “secret sauce” is how it all works together). The combination of open-source components and a upcoming cloud service (hinted by the mention of a $CODEC token utility and public product access) means Codec will soon be accessible in multiple forms (both as a service and self-hosted). Team Moyai: 15+ years dev experience, currently leading AI development at Elixir Games. lil’km: 5+ years AI developer, currently working with HuggingFace on the LeRobot project. HuggingFace is a huge robotics company and Moyai works as head of ai at elixir games (backed by square enix and solanafdn. I’ve personally video called the entire team and really like the energy they bring. My friend who put them on my radar also met them all at Token2049 and only had good things to say. Final Thoughts There’s still a lot left to cover, which I’ll save for future updates and posts in my Telegram channel. I’ve long believed cloud infrastructure is the future for operator agents. I’ve always respected what Nuit is building, but Codec is the first project that’s shown me the full-stack conviction I was looking for. The team are clearly top tier engineers. They’ve openly said marketing isn’t their strength, which is likely why this has flown under the radar. I’ll be working closely with them to help shape the GTM strategy that actually reflects the depth of what they’re building. With a $4 mil market cap and this level of infrastructure, it feels massively underpriced. If they can deliver a usable product, I think it could easily mark the beginning of the next AI infra cycle. As always, there’s risk and while I’ve vetted the team in stealth over the past few weeks, no project is ever completely rug proof. Price targets? A lot higher.
Tldr on why I chose Codec > Nuit for Operators: Codec uses a three layer architecture (Machine, System, Intelligence) enabling isolated, high performance agents with native control. Each Codec agent runs locally using a Vision-Language-Action (VLA) loop, reducing latency and increasing reliability. Nuit’s model depends on browser parsing + cloud AI calls, which limits flexibility and introduces fragility. Codec scales horizontally across thousands of agents, with no shared state and fault tolerant modularity.
11.13K
61
The content on this page is provided by third parties. Unless otherwise stated, OKX is not the author of the cited article(s) and does not claim any copyright in the materials. The content is provided for informational purposes only and does not represent the views of OKX. It is not intended to be an endorsement of any kind and should not be considered investment advice or a solicitation to buy or sell digital assets. To the extent generative AI is utilized to provide summaries or other information, such AI generated content may be inaccurate or inconsistent. Please read the linked article for more details and information. OKX is not responsible for content hosted on third party sites. Digital asset holdings, including stablecoins and NFTs, involve a high degree of risk and can fluctuate greatly. You should carefully consider whether trading or holding digital assets is suitable for you in light of your financial condition.