Apple's MLX Framework Now Supports NVIDIA GPUs with CUDA Backend
ByAinvest
Tuesday, Jul 15, 2025 8:37 pm ET1min read
AAPL--
In a separate development, NVIDIA has introduced the B30 AI GPU, optimized for small to medium AI models and cloud services. The B30 delivers approximately 75% of the performance of the H20 AI GPU, making it a cost-effective solution for Chinese tech firms. Demand for the B30 is significant, with Chinese tech companies placing orders for hundreds of thousands of units, totaling over $1 billion in late-June, with deliveries expected in August [1].
The B30 is designed to address two major pain points for China. It is the preferred solution for inference in small and medium-sized models, aligning with the arrival of the inference era. Additionally, it acts as a low-cost computing power pool for cloud services. A computing power pool constructed with 100 B30 AI GPUs can support lightweight training of billion-parameter models while reducing procurement costs by 40% and unit power consumption by close to 30% compared to the H20 [1].
The B30's deep compatibility with the CUDA-X ecosystem allows enterprises to seamlessly migrate frameworks like PyTorch, saving technical reconstruction costs. This integration is crucial for maintaining the 'stickiness' of the CUDA ecosystem in mainstream model deployment efficiency [1].
While domestic-made AI chips from the likes of Huawei might slightly pass the B30 in single-card FP16 computing power, the B30 maintains an advantage in mainstream model deployment efficiency due to its CUDA compatibility [1].
References:
[1] https://www.tweaktown.com/news/106374/nvidias-new-b30-ai-gpu-for-china-expected-to-have-significant-demand-75-as-fast-the-h20/index.html
NVDA--
Apple's machine learning framework, MLX, is gaining support for NVIDIA GPUs through a CUDA backend. This integration will allow developers to run MLX models directly on NVIDIA GPUs, opening up new possibilities for testing, experimentation, and research use cases. The work is still in progress, but core operations such as matrix multiplication and softmax are already supported.
Apple's machine learning framework, MLX, is gaining support for NVIDIA GPUs through a CUDA backend, enabling developers to run MLX models directly on NVIDIA GPUs. This integration opens up new possibilities for testing, experimentation, and research use cases. While the work is still in progress, core operations such as matrix multiplication and softmax are already supported.In a separate development, NVIDIA has introduced the B30 AI GPU, optimized for small to medium AI models and cloud services. The B30 delivers approximately 75% of the performance of the H20 AI GPU, making it a cost-effective solution for Chinese tech firms. Demand for the B30 is significant, with Chinese tech companies placing orders for hundreds of thousands of units, totaling over $1 billion in late-June, with deliveries expected in August [1].
The B30 is designed to address two major pain points for China. It is the preferred solution for inference in small and medium-sized models, aligning with the arrival of the inference era. Additionally, it acts as a low-cost computing power pool for cloud services. A computing power pool constructed with 100 B30 AI GPUs can support lightweight training of billion-parameter models while reducing procurement costs by 40% and unit power consumption by close to 30% compared to the H20 [1].
The B30's deep compatibility with the CUDA-X ecosystem allows enterprises to seamlessly migrate frameworks like PyTorch, saving technical reconstruction costs. This integration is crucial for maintaining the 'stickiness' of the CUDA ecosystem in mainstream model deployment efficiency [1].
While domestic-made AI chips from the likes of Huawei might slightly pass the B30 in single-card FP16 computing power, the B30 maintains an advantage in mainstream model deployment efficiency due to its CUDA compatibility [1].
References:
[1] https://www.tweaktown.com/news/106374/nvidias-new-b30-ai-gpu-for-china-expected-to-have-significant-demand-75-as-fast-the-h20/index.html

Stay ahead of the market.
Get curated U.S. market news, insights and key dates delivered to your inbox.
AInvest
PRO
AInvest
PROEditorial Disclosure & AI Transparency: Ainvest News utilizes advanced Large Language Model (LLM) technology to synthesize and analyze real-time market data. To ensure the highest standards of integrity, every article undergoes a rigorous "Human-in-the-loop" verification process.
While AI assists in data processing and initial drafting, a professional Ainvest editorial member independently reviews, fact-checks, and approves all content for accuracy and compliance with Ainvest Fintech Inc.’s editorial standards. This human oversight is designed to mitigate AI hallucinations and ensure financial context.
Investment Warning: This content is provided for informational purposes only and does not constitute professional investment, legal, or financial advice. Markets involve inherent risks. Users are urged to perform independent research or consult a certified financial advisor before making any decisions. Ainvest Fintech Inc. disclaims all liability for actions taken based on this information. Found an error?Report an Issue

Comments
No comments yet