Mozi (@yeahhe)Mac mini oMLX 部署本地大模型测试,并发提速最高可达2.77 倍 中发帖

测试模型为Qwen3.5-4B-MLX-4bit (3.0 GB) 
测试平台为Mac mini M4 
 [PixPin_2026-03-10_17-47-32] 
[PixPin_2026-03-10_17-48-06] 
oMLX - LLM inference, optimized for your Mac
https://github.com/jundot/omlx
Benchmark Model: Qwen3.5-4B-MLX-4bit
================================================================================

Single Request Results
--------------------------------------------------------------------...
 
 
Back to Top