Off-Prem

Alibaba Cloud claims its modular datacenter architecture shrinks build times by 50 percent

Also reveals boosted utilization rates, upgraded IaaS and more – all in the name of AI apps


Alibaba Cloud has revealed a modular datacenter architecture it claims will help it to satisfy demand for AI infrastructure by improving performance and build times for new facilities.

Announced at its annual Apsara conference yesterday, the "CUBE DC 5.0" architecture was described as using "prefabricated modular designs" plus "advanced and proprietary technologies such as wind-liquid hybrid cooling system, all-direct current power distribution architecture and smart management system."

Alibaba Cloud hasn't explained those techs in depth, but claims the modular approach reduces deployment times by up to 50 percent compared to traditional datacenter building techniques.

The Register has asked Alibaba to explain the workings of a "wind-liquid hybrid cooling system." For what it's worth, machine translation of the term into Mandarin, then fed into search engines, produced results describing cold plate cooling – a technique that sees thin reservoirs of cooled liquids placed on hardware, with cooling achieved by circulating liquid and/or blowing air across the plates.

Whatever the term describes, CEO of Alibaba Cloud Intelligence Eddie Wu told the conference his company "is investing heavily in building an AI infrastructure for the future."

"These enhancements are not just about keeping up with AI demands but about setting a global standard for efficiency and sustainability."

Other steps towards that goal include a scheduler said to better manage hardware resources so that they achieve up to 90 percent utilization rates.

Alibaba Cloud's IaaS offering, the Enterprise Elastic Compute Service (ECS), has reached its ninth generation. Conference attendees were told it is better equipped for AI applications as it has improved recommendation engine speeds by 30 percent and database read/write queries per second by 17 percent.

Also at the conference, Alibaba Cloud announced an "Open Lake data utility" that integrates multiple big data engines so they can be used by generative AI applications. Another new offering, "DMS: OneMeta+OneOps," apparently combines and manages metadata from 40 different data sources.

It's 2024, so Alibaba Cloud also announced some AI news: the release of its Qwen 2.5 multimodal models, available in sizes from 0.5 to 72 billion parameters, supporting 29 languages and tuned for the needs of sectors including automotive and gaming. The new models are said to have "enhanced knowledge [and] stronger capabilities in math and coding."

A text-to-video AI model that works with both Chinese and English prompts, Tongyi Wanxiang, was also released.

"The new model is capable of generating high-quality videos in a wide variety of visual styles from realistic scenes to 3D animation," boasted Alibaba Cloud execs. ®

Send us news
4 Comments

Hugging Face puts the squeeze on Nvidia's software ambitions

AI model repo promises lower costs, broader compatibility for NIMs competitor

AI firms and civil society groups plead for passage of federal AI law ASAP

Congress urged to act before year's end to support US competitiveness

European datacenter energy consumption set to triple by end of decade

McKinsey warns an additional 25GW of mostly green energy will be needed

UK gov report to propose special zones for datacenters, 'AI visas'

Vendors not keen on 'lengthy bureaucracy,' and cost when they try to hire skilled foreigners

Meta spruiks benefits of open sourcing Llama models – to its own bottom line

It's not like Zuck needs the coin despite increased infrastructure spend, headcount, losses on VR

AMD teases its GPU biz 'approaching the scale' of CPU operations

Q3 profits jump 191 percent from last quarter on revenues of $6.2 billion, helped by accelerated interest in Instinct

Nvidia CEO whines Europeans aren’t buying enough GPUs

EU isn’t keeping up with US and China investments, AI arms dealer says

Amazon to cough $75B on capex in 2024, more next year

Despite extending server lifespans, AI's power demands drive more datacenter builds

Google reportedly developing an AI agent that can control your browser

Project Jarvis will apparently conduct research, purchase products, and even book a flight on your behalf

xAI picked Ethernet over InfiniBand for its H100 Colossus training cluster

Work already underway to expand system to 200,000 Nvidia Hopper chips

Fujitsu, AMD lay groundwork to pair Monaka CPUs with Instinct GPUs

Before you get too excited, Fujitsu's next-gen chips won't ship till 2027

Energy companies told to recharge for AI datacenter surge

Demand for electricity to outstrip supply soon, warns Bain