logo
ホーム ニュース

会社のニュース MinIO adds petabyte-scale MemKV cache for Nvidia GPU inference

認証
中国 Beijing Qianxing Jietong Technology Co., Ltd. 認証
中国 Beijing Qianxing Jietong Technology Co., Ltd. 認証
顧客の検討
北京Qianxing Jietongの技術Co.、株式会社の販売スタッフは非常に専門および忍耐強い。それらは引用語句をすぐに提供してもいい。プロダクトの質そして包装はまた非常によい。私達の協同は非常に滑らかである。

—— 《のFestfing DVの》 LLC

私がIntel CPUおよび東芝SSDを緊急に捜していたときに、北京Qianxing Jietongの技術Co.、株式会社からのサンディは私に多くの助けを与え、私に私がすぐに必要としたプロダクトを得た。私は実際に彼女を認める。

—— キティ円

北京Qianxing Jietongの技術Co.、株式会社のサンディは私がサーバーを買う時間の構成間違いを私に思い出させることができる非常に注意深いセールスマンである。エンジニアはまた非常に専門で、すぐにテスト プロセスを完了できる。

—— Strelkin Mikhail Vladimirovich

北京千星捷通との仕事は大変満足しています。製品の品質は素晴らしく、納期も常に守られています。営業チームはプロフェッショナルで、忍耐強く、私たちの質問にすべて丁寧に対応してくれます。彼らのサポートに心から感謝しており、長期的なパートナーシップを期待しています。強くお勧めします!

—— アフマド・ナビド

品質: 提供者との素晴らしい経験. MikroTik RB3011は既に使用されていましたが,非常に良い状態で,すべてが完璧に動作しています. コミュニケーションは迅速でスムーズでした.そして私の懸念はすぐに解決されました信頼性の高いサプライヤーです 強くお勧めします

—— ゲラン・コレシオ

オンラインです
会社 ニュース
MinIO adds petabyte-scale MemKV cache for Nvidia GPU inference

MinIO has developed a petabyte-scale MemKV caching system tailored for Nvidia GPUs, deployed on top of its AIStor object storage platform.

GPU clusters running inference require high-bandwidth memory (HBM) to store context, vectorized tokens and intermediate key-value (KV) pairs. Once GPU HBM is saturated, data cascades down to CPU DRAM and NVMe SSDs, managed by Nvidia BlueField-4 (BF4) DPUs. When these tiers reach capacity, MinIO AIStor acts as the final storage backup. Nvidia’s STX architecture governs this multi-layer cache hierarchy, and MemKV complies with the standard to deliver persistent, shared context across GPU clusters at superior scale.

最新の会社ニュース MinIO adds petabyte-scale MemKV cache for Nvidia GPU inference  0


AB Periasamy, MinIO co-founder and co-CEO, commented: “The industry has been papering over context loss for years because, at small scale, you may absorb the recompute tax. At today’s high GPU density for hyperscalers and neoclouds, this is no longer viable.

 Recomputing generated context wastes power; for clusters with thousands of GPUs, it creates fundamental structural inefficiency. Large-scale inference requires purpose-built infrastructure, and MemKV is designed specifically for this data path.”

For the first time, MinIO enables shared context pools for entire GPU clusters at microsecond-level latency matching inference workflows, avoiding millisecond delays from conventional external storage. Without sufficient cache tiers, GPUs waste resources on repeated context recalculation.

In a 128-GPU deployment with 128K-token context length, MemKV improved time-to-first-token under production loads and boosted GPU utilization from 50% to over 90%, generating an estimated $2 million annual compute cost saving.

Purpose-built for Nvidia STX architecture, MemKV supports Nvidia Dynamo and NIXL caching tools. It delivers petabytes of shared context memory at SSD-level costs, decoupling cache scaling from GPU compute resources. Its core features are listed below:
  • Native BF4 STX support: Runs as an ARM64 binary within STX infrastructure, embedded in storage rather than separate x86 storage servers.
  • End-to-end RDMA transport: Transfers KV cache between GPU memory and NVMe via RDMA, bypassing conventional file and object storage protocols.
  • GPU-optimized block size: Uses 2–16 MB blocks for GPU throughput demands, instead of legacy 4 KB storage blocks.
  • Wire-speed performance: Optimized for Nvidia Spectrum-X Ethernet and PCIe Gen6 to maximize physical fabric throughput.


MemKV directly transfers data from NVMe SSDs to AI pipelines over RDMA, eliminating HTTP overhead, file system translation and intermediate storage servers.


最新の会社ニュース MinIO adds petabyte-scale MemKV cache for Nvidia GPU inference  1

MinIO categorizes rival context memory solutions into two types: non-sharable local NVMe (G3) and general-purpose shared storage (G4). It positions MemKV as a purpose-built G3.5 tier, distinguishing itself from generic storage products.

The firm emphasizes that legacy vendors’ G3.5 offerings still retain redundant protocol nodes, metadata services and file translation layers. These layers ensure durability and consistency for training data and model weights, yet they are unnecessary for ephemeral, recomputable KV cache optimized for 2–16 MB data blocks.

Hardware RAID vendor GRAID and storage firm WEKA also provide STX-compatible KV cache solutions. A broad range of storage vendors support Nvidia STX, including Cloudian, Dell, DDN, Everpure, Hammerspace, Hitachi Vantara, HPE, Lightbits/ScaleFlux, NetApp, Nutanix, Peak:AIO, Pliops and VAST Data.

Beijing Qianxing Jietong Technology Co., Ltd.
Sandy Yang/Global Strategy Director
WhatsApp / WeChat: +86 13426366826
Email: yangyd@qianxingdata.com
Website: www.qianxingdata.com/www.storagesserver.com
Business Focus:
ICT Product Distribution/System Integration & Services/Infrastructure Solutions
With 20+ years of IT distribution experience, we partner with leading global brands to deliver reliable products and professional services.
“Using Technology to Build an Intelligent World”Your Trusted ICT Product Service Provider!
パブの時間 : 2026-05-14 13:46:14 >> ニュースのリスト
連絡先の詳細
Beijing Qianxing Jietong Technology Co., Ltd.

コンタクトパーソン: Ms. Sandy Yang

電話番号: 13426366826

私達に直接お問い合わせを送信 (0 / 3000)