Embedded LLM has unveiled TokenVisor, a new GPU monetisation and management platform co-launched with AMD, designed to help neocloud providers and enterprises manage and monetise AMD GPU clusters for Large Language Model (LLM) workloads. The announcement was made during the Advancing AI 2025 event, held on 12 June in Santa Clara, California.
Supporting monetisation and governance for GPU-powered neoclouds
TokenVisor is the first control plane built specifically for the AMD GPU-powered neocloud ecosystem. It enables users to manage LLM workloads more effectively while accelerating time-to-revenue. The platform helps providers streamline deployment, billing, and governance, offering a clear path to Return on Investment (ROI).
Key features include automated resource allocation, real-time usage monitoring, rate-limiting policies, and support for custom pricing. These capabilities allow GPU owners to efficiently commercialise their infrastructure while giving enterprise clients tools to enforce internal cost control and compliance.
Early adopters have reported significant benefits, including faster monetisation after hardware installation and better support for popular LLM and multi-modal models. The combination of technical responsiveness and comprehensive model compatibility has been cited as a major strength by users looking to quickly recover AI infrastructure investments.
Born out of community collaboration and open-source values
TokenVisor was developed in consultation with the AMD GPU neocloud community, embracing the collaborative ethos showcased at Advancing AI 2025. The platform reflects Embedded LLM’s commitment to empowering the decentralised AI ecosystem with enterprise-grade solutions.
“TokenVisor is the hypervisor for the AI Token era – unlocking decentralised GPU computing’s potential requires tools as powerful and flexible as the hardware,” said Ooi Ghee Leng, CEO of Embedded LLM. “Co-launched at Advancing AI 2025, an event that celebrates AI innovation and open-source collaboration, marks an important milestone for the AMD GPU neocloud community.”
Mahesh Balasubramanian, Senior Director of Product Marketing, Data Center GPU Business at AMD, added, “TokenVisor brings powerful new capabilities to the AMD GPU neocloud ecosystem, helping providers efficiently manage and monetise LLM workloads.”
Strengthening Singapore’s AI and cloud innovation goals
Based in Singapore, Embedded LLM is part of the country’s expanding deep tech sector and supports national goals to position itself as a hub for AI and cloud infrastructure in Southeast Asia. The launch of TokenVisor contributes to Singapore’s push for AI sovereignty and regional leadership in digital innovation.
Embedded LLM continues to develop LLM platforms aimed at making generative AI more accessible. It is an active contributor to open-source tools, including enhancements to vLLM for AMD ROCm and orchestration platforms like JamAI Base. With TokenVisor, the company offers a practical solution to monetise and manage GPU clusters in a decentralised AI environment.