TensorWave offers cloud-based access to AMD data center GPUs (such as the Instinct MI300X accelerator) for LLMs and other AI workloads. The company was the one of the first-to-market MI300X launch partners, offering enterprises the choice between bare-metal nodes and fully-managed Kubernetes clusters.
In December 2023, TensorWave announced plans to build a bare-metal specialized AI cloud called "TensorNODE," powered by GigaIO's SuperNODE infrastructure and AMD's Instinct MI300X accelerators. The deployment aims to create the first petabyte-scale GPU memory pool, supporting up to 5,760 GPUs within a single domain using GigaIO's FabreX memory fabric.
The startup's headquarters in Las Vegas allowed it to keep energy costs and overheads lower than in other major US cities. It generates revenue by renting GPU capacity by the hour (ranging from USD 1 to USD 10, depending on workload requirements) and requires a minimum six-month contract.
Key customers and partnerships
In October 2024, TensorWave partnered with US data center firm Tecfusions to lease 1 GW of AI capacity across Tecfusion's data center portfolio. Other partnerships included 1) open networking solutions provider Edgecore Networks to support the growth of AI and ML workloads (May 2024) and 2) AI inferencing startup MK1 to offer a competitive alternative to NVIDIA's dominance by optimizing AMD's cloud-native hardware for inference workloads (May 2024).
By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.