A Real-World Review of the 50,000 Yuan Apple AIPC: Better Than We Expected | M5 Max MacBook Pro Review

If you had a budget of 50,000 yuan to build a personal computer, what would you choose?
In the past, you might have put the bulk of your budget into the graphics card—after all, whether you're "playing games" or "playing games after get off work," having a powerful GPU is never a bad thing.

▲ Image | Internet
But now, the problem has become more complicated.
The budget that was previously distributed in an orderly manner among CPU, GPU, motherboard, memory, hard drive, and peripherals has suddenly been disrupted by memory, this "money-devouring behemoth".
Now, no matter what you plan to do with your computer, you'll encounter the problem of neglecting one thing for another—
Large memory, large video memory, and large hard drive are all essential, but each one will drain your wallet.
The Mac, which has emerged as a dark horse amidst the memory chaos, is precisely the best solution to the problem mentioned above.
The most powerful AI Mac to date
At the recent spring launch event, Apple unveiled the upgraded M5 MacBook Pro as expected, along with the accompanying M5 Pro and M5 Max processors.

As a result of Apple Silicon's complete overhaul of TSMC's 3nm N3P process, the two new processors certainly did not disappoint in terms of specifications.
The M5 Pro comes in two configurations: 15+16 and 18+20 cores. Both are equipped with the neural network accelerator from last year's M5, which is the "Apple version of Tensor Core".

▲ Image|Apple
The M5 Max offers 18+32 and 18+40 core options, as well as a 16-core neural network accelerator. In terms of processor size alone, both the M5 Pro and M5 Max are undoubtedly GPU-first .
This tendency is also reflected in the microarchitecture design of the new processors.
Currently, all M5 series processors have been upgraded with LPDDR5X 9600 unified memory. According to Apple, the M5 Pro has a maximum memory bandwidth of 307GB/s, while the M5 Max has 614GB/s .

▲ Image|Apple
Since both the M5 Pro and M5 Max come standard with an 18-core CPU, the difference in memory bandwidth is most likely due to the GPU specifications.
Based on pre-release predictions, this discrepancy suggests that the memory controller for the M5 series is likely located on the GPU core cluster .
This strategy is remarkably similar to the Panther Lake architecture that iFanr saw during their visit to Intel's factory last year:

The benefits of doing this are obvious—placing the GPU closer to the memory controller can effectively reduce the latency of inter-core communication of memory data, thereby indirectly improving GPU efficiency.
What are GPUs, which are faster and have more VRAM, best at? Local AI applications, of course.

This is one of the reasons why Apple mentioned "AI" so frequently on its official website this time.
Take this 14-inch MacBook Pro prototype from iFanr as an example. The one we received is the top-of-the-line 40-core GPU M5 Max version this year, paired with 128GB of unified memory and an 8TB hard drive, a performance monster costing over 55,000 yuan.

Generally speaking, when we run a local model on a Windows PC, the biggest bottleneck is often not the exorbitantly priced "motherboard memory," but rather the VRAM (video memory) inside the graphics card.
The biggest advantage of Apple's unified memory is that it can be directly accessed by the GPU.
For example, our 128GB M5 Max review unit can theoretically provide the GPU with nearly 100GB of video memory:

Now that we have such ample memory, we should, as Apple advertises, run those large-scale local AI models that we couldn't run before.
In llmfit, you can see that a 128GB M5 Max can run all models up to 125 bytes perfectly.
Only devices like the MiniMax M2.5, Qwen3, and DeepSeek v2.5 (with specs of 220b or higher) will become "barely running" (marginal).

▲ M5 Max 128GB
In comparison, a 32GB RAM M1 Max, according to LLMfit, can only run models with 2 or 4-bit quantization of around 35 bits at most .

▲ M1 Max 32GB
Considering ease of deployment and the space for context understanding, we chose to test qwen3.5-35b-a3b and qwen3-next-80b, which supports MLX, using LM Studio. Both are 8-bit quantized MoE models .

For MoE models like qwen3.5-35b-a3b, which have "a small total number of iterations and a small number of inferences," M5 Max often finishes running before it even has a chance to warm up .

▲ qwen3.5-35b-a3b
Even when faced with original text of nearly 3,000 words, after manually maximizing the model token limit, the M5 Max's first-word response time in each round of rewriting and imitation was around 1.7 seconds , or about 1.7s for TTFT and about 65tps for TPOT. There was no overflow even after accumulating nearly 10,000 words of thought and writing.

▲ qwen3.5-35b-a3b
The qwen3-next-80b, with its MLX optimization, 8-bit quantization, and larger parameter count, is even more powerful on the M5 Max.
Although it requires manually loading a model that is nearly 80GB while ignoring memory warnings, the running results are truly remarkable:
In qwen3.5-35b-a3b, it takes nearly 30 seconds to think about the same prompt word, while in qwen3-next-80b, it is almost instantaneous, with TTFT of about 3 seconds and TPOT of about 72 tps.

▲ qwen3-next-80b
This is partly because the 80b parameters are already large enough compared to the 3b active parameters, and more importantly, because it is an optimized version based on Apple's open-source MLX framework , which can maximize the advantages of Apple Silicon.
Besides the MoE model, how does M5 Max perform with dense models like Llama 3.3?

▲ Image|Tom's Guide
Although the 8-bit quantized Llama 3.3 70b model is only about 75GB in size, the huge KV cache required for the 128k context still overflows, causing LM Studio to fail to load it.
After switching to the smaller Llama 3.3 70b Q4_K_M , M5 Max finally loaded normally. After executing the above prompt, the system load was approximately 95GB, and the generation speed was 9.95 tokens/s.

In other words, when dealing with dense models of similar size, the M3 Ultra with more memory is still needed .
However, the highest resource usage we observed on M5 Max this time was not from the dense Llama 3.3, but from deepseeek-r1 running in Msty Studio:

In Msty Studio, we loaded a 75GB deepseek-r1 70b-llama-distill-q8_0 file, and in two minutes, it used 122GB of memory to write you a haiku:

▲ deepseek-r1 70b-llama-distill-q8_0
This is just the result for the local language model. Even in some traditional performance projects, M5 Max's performance has not disappointed us.
In Cinebench 2026, the M5 Max achieved a GPU score of 79,295 , which is more than 15% higher than the M4 Max and only about 5% lower than the current largest M3 Ultra.

▲ After continuous stress testing, the score dropped to around 77,000.
How would such a result be perceived in a game?
We played Cyberpunk 2077 on the M5 Max again, using the same parameters as when we reviewed the standard version of the M5 last year.
When using the default "for this Mac" preset, the M5 Max can maintain a stable frame rate of around 59 frames per second. Compared to the standard M5, the preset not only has a higher resolution and more detail, but the frame rate is also more than doubled .

After manually optimizing the settings (high detail 1.5K ray tracing FSR MetalFX super-resolution and frame generation), the M5 Max can maintain a stable 50-60 FPS in dense scenes even with the fan fully loaded.

This performance is certainly far from that of a gaming laptop, but 2077 is a very demanding game, and it's still quite surprising that the M5 Max can run it to this level in a 14-inch chassis without being plugged in.
As for other smaller, better-optimized games, such as Control: Ultimate Collection and Escape from Durkoff, as long as you don't mess with the settings, the M5 Max can generally maintain a stable 60 frames per second.
Whether it's for AI workflows or gaming, this MacBook Pro with the M5 Max chip is undoubtedly a powerful beast .
The best Apple screen to date
Besides the M5 Pro/Max, another "Pro-level" new product launched at this spring's product launch event was the long-awaited new generation Studio Display.
More specifically, it's the new Studio Display and Studio Display XDR.

After the Pro Display XDR was discontinued, the Studio Display XDR took over its mantle, becoming Apple's flagship professional monitor with a starting price of 24,999.
Our initial experience with the Studio Display XDR was consistent with that at the Apple event:
The impact of the smaller screen size is not obvious; in fact, ProMotion grabbed our attention from the very first second.

Thanks to a mini-LED panel with 2304 zones, as well as 1000 nits peak SDR brightness and 2000 nits peak HDR brightness, it's impossible to say that it's "not eye-catching".

▲ The halo effect of mini-LEDs is only visible under very extreme conditions.
Beyond the well-worn topic of wide color gamut HDR content creation, the Studio Display XDR is also a formidable player in audio-visual entertainment.
Especially if you have some HDR-enabled "blockbuster" movies on hand, the experience of using a MacBook Pro with the Studio Display XDR is unparalleled in the current Apple product line:

The same assessment applies to this year's new Studio Display.
In fact, aside from ProMotion, peak brightness, and charging power, the Studio Display's screen panel quality is completely on par with the new Studio Display XDR.

After all, Apple had given prior notice: 5K 120Hz is not something that just any processor can handle. If your Mac is using the M1 series, M2, or M3 standard version, it can only display at a maximum of 60Hz when you plug in a Studio Display XDR.
This aligns with our experience with the Studio Display XDR.
Even if your macOS version is too old, it may not be able to output a picture even if it is plugged into a monitor and can be charged.

Interestingly, when iFanr spoke with Apple staff at the launch event, they mentioned that both displays are equipped with the iPhone's SoC chip .
MacRumors, a foreign media outlet, discovered by unpacking the firmware update code for the two new displays that Apple has equipped them with A19 and A19 Pro processors respectively.

Unsurprisingly, this is intended for 5K video decoding, backlight control, Center Stage camera, and other display features.
But this has also led to more and more "processor jokes" about Apple:
At the very beginning of 2026, you will be able to buy an iPad Pro with an M5 chip, a MacBook Neo with an A18 Pro chip, and a monitor with an A19 Pro chip.
Overall, this year's Studio Display XDR is a very timely update.
Its most important significance lies in filling the gap in Apple's professional product line with ProMotion, and also in making the product interaction experience more seamless.

When Apple started talking about AI
At this spring launch event, in addition to changing its previous launch format, Apple also began to openly discuss AI.

This AI is neither the repeatedly delayed Apple Intelligence, nor the Machine Learning that Apple has repeatedly emphasized; it is simply and directly general artificial intelligence .
Judging from the current performance of its products, Apple is indeed ready when it starts talking about AI.
Back when Apple switched to Apple Silicon and a unified memory architecture in 2020, it probably didn't anticipate the explosive growth in demand for AI models and the ensuing memory crisis.

The most straightforward example is the 128GB of unified memory on this M5 Max:
- If you only consider consumer-grade DDR5 6400 memory, it's not difficult to buy 128GB, with a cost of only around 10,000 yuan , but it will never reach the bandwidth of 614GB/s.
- If you want to use your graphics card to get 128GB of VRAM, and you can't buy a professional graphics card, you'll need to buy five RTX 5090D cards . This is after ignoring the communication latency between graphics cards.
In such situations, small business teams, individual developers, AI practitioners, and others with local AI needs will find themselves in a dilemma:
Alternatively, with a limited budget, you can allocate it to memory, graphics card, CPU, hard drive, etc. when building a PC, thus diluting the overall computing performance.
Alternatively, they could grit their teeth and increase their budget, investing tens or even hundreds of thousands of yuan to enter the field of self-constructed servers.

▲ Image|Servermall
At this point, a MacBook Pro priced under 60,000 yuan, featuring 128GB of high-bandwidth memory, a top-of-the-line HDR screen and speakers, and an 8TB hard drive, has become the ultimate cost-effective choice for individual and studio users .

Even if you don't need the aforementioned "peripherals," or your local AI requirements are low, you can opt for a Mac Studio or Mac mini as a second choice.
The latter has already enjoyed its own spring during the recent "lobster boom".

▲ Image|Apple Must
While Apple Intelligence may be laughable, the potential of Apple Silicon and unified memory in this "big AI era" is only the tip of the iceberg.
#Welcome to follow iFanr's official WeChat account: iFanr (WeChat ID: ifanr), where more exciting content will be presented to you as soon as possible.
ifanr | Original Link · View Comments · Sina Weibo











https://www.huntress.com/blog/openclaw-github-ghostsocks-infostealer


























