The M1 MacBook Professional is an outdated however nonetheless succesful system in 2026.
Kyle Kucharski/ZDNET
Comply with ZDNET: Add us as a most popular supply on Google.
ZDNET’s key takeaways
- Ollama makes it pretty simple to obtain open-source LLMs.
- Even small fashions can run painfully gradual.
- Do not do that with no new machine with 32GB of RAM.
As a reporter masking synthetic intelligence for over a decade now, I’ve at all times identified that working synthetic intelligence brings all types of laptop engineering challenges. For one factor, the big language fashions preserve getting greater, and so they preserve demanding an increasing number of DRAM reminiscence to run their mannequin “parameters,” or “neural weights.”
Additionally: Tips on how to set up an LLM on MacOS (and why you must)
I’ve identified all that, however I needed to get a really feel for it firsthand. I needed to run a big language mannequin on my house laptop.
Now, downloading and working an AI mannequin can contain plenty of work to arrange the “setting.” So, impressed by my colleague Jack Wallen’s protection of the open-source device Ollama, I downloaded the MacOS binary of Ollama as my gateway to native AI.
Ollama is comparatively simple to make use of, and it has carried out good work integrating with LangChain, Codex, and extra, which implies it’s changing into a device for bringing collectively plenty of elements of AI, which is thrilling.
Causes to maintain it native
Working LLMs regionally, fairly than simply typing into ChatGPT or Perplexity on-line, has plenty of enchantment for not simply programmers, however any info employee.
First, as an info employee, you’ll be extra fascinating within the job market if you are able to do one thing like obtain a mannequin and run it fairly than typing into the web immediate similar to each free person of ChatGPT. We’re speaking fundamental skilled improvement right here.
Second, with a neighborhood occasion of an LLM, you may preserve your delicate information from leaving your machine. That needs to be of apparent significance to any info employee, not simply coders. In my case, my undertaking objective was to make use of native fashions as a technique to mine my very own trove of articles over time, as a form of report on what I’ve written, together with issues I may need forgotten about. I favored the thought of protecting all of the information native fairly than importing them to a cloud service.
Additionally: I attempted vibe coding an app as a newbie – here is what Cursor and Replit taught me
Third, you may keep away from charges charged by OpenAI, Google, Anthropic, and the remaining. As I wrote lately, costs are set to rise for utilizing LLMs on-line, so now is an efficient time to consider methods to do the majority of your work offline, by yourself machine, the place the meter will not be always working.
(Disclosure: Ziff Davis, ZDNET’s mother or father firm, filed an April 2025 lawsuit towards OpenAI, alleging it infringed Ziff Davis copyrights in coaching and working its AI techniques.)
Fourth, you will have much more management. For instance, if you happen to do wish to do programming, you may tweak LLMs, generally known as fine-tuning them, to get extra targeted outcomes. And you need to use varied regionally put in instruments resembling LangChain, Anthropic’s Claude Code device, OpenAI’s Codex coding device, and extra.
Additionally: Why you may pay extra for AI in 2026, and three money-saving tricks to strive
Even if you happen to simply wish to do information-worker duties resembling producing stories, doing so with a neighborhood cache of paperwork or a neighborhood database will be carried out with larger management than importing stuff to the bot.
Naked-minimum bare-metal
I set out on this experiment with a bare-minimum machine, so far as what it takes to run an LLM. I needed to search out out what would occur if somebody who does not always purchase new machines tried to do that at house on the identical laptop they use for on a regular basis duties.
My MacBook Professional is three years outdated and has 16 gigabytes of RAM and a terabyte laborious drive that is three-quarters full, working not the newest MacOS, however MacOS Sonoma. It is the 2021 mannequin, mannequin quantity MK193LL/A, and so, whereas it was prime of the road once I purchased it at Finest Purchase in January of 2023 in a close-out sale, it was already changing into yesterday’s greatest mannequin again then.
Additionally: 5 causes I take advantage of native AI on my desktop – as a substitute of ChatGPT, Gemini, or Claude
I do know, I do know: That is past the everyday helpful lifetime of machines and past anybody’s depreciation schedule. However, the MacBook was a terrific improve on the time, and it has continued to carry out fantastically each day for the everyday information-worker duties: calendar, tons of e-mail, tons of internet sites, video post-production, podcast audio recording, and extra. I by no means have any complaints. Hey, if it ain’t broke, proper?
So the query was, how would this venerable however nonetheless mighty machine deal with a really totally different new form of workload?
Beginning Ollama
The beginning-up display screen for Ollama appears to be like like ChatGPT, with a pleasant immediate to kind into, a “plus” signal to add a doc, and a drop-down menu of fashions you may set up regionally, together with fashionable ones resembling Qwen.
If you happen to simply begin typing on the immediate, Ollama will mechanically attempt to obtain no matter mannequin is exhibiting within the drop-down menu. So, do not do any typing except you wish to go together with mannequin roulette.
As an alternative, I appeared by the fashions within the drop-down listing, and I spotted that a few of these fashions weren’t native — they have been within the cloud. Ollama runs a cloud service if you would like its infrastructure as a substitute of your personal. That may be helpful if you wish to use a lot bigger fashions that might overly tax your personal infrastructure.
Per the pricing web page, Ollama affords some entry to the cloud within the free account, with the flexibility to run a number of cloud fashions lined by the “Professional” plan at $20 per 30 days, and much more utilization within the “Max” plan at $100 per 30 days.
Additionally: This app makes utilizing Ollama native AI on MacOS gadgets really easy
Sticking with regionally working choices, I made a decision to take a look at the broader listing of fashions within the mannequin listing maintained by Ollama.
At random, I selected glm-4.7-flash, from the Chinese language AI startup Z.ai. Weighing in at 30 billion “parameters,” or neural weights, GLM-4.7-flash can be a “small” massive language mannequin by at the moment’s requirements, however not tiny, as there are open-source fashions with fewer than a billion parameters. (A billion parameters was huge, not so way back!)
The listing provides you the terminal instructions to obtain the chosen mannequin from the Mac terminal, simply by copying and pasting on the immediate, resembling:
ollama run glm-4.7-flash
Be aware of disk house. Glm-4.7-flash weighs in at 19 gigabytes of disk utilization, and bear in mind, that is small!
In my expertise, downloading fashions appears pretty swift, although not lightning quick. On a gigabit-speed cable modem to my house workplace supplied by Spectrum in New York Metropolis, the mannequin was downloading at a price of 45 megabytes per second at one level, although it later dropped to a slower price of throughput.
Attending to know the mannequin
My first immediate was pretty simple: “What sort of massive language mannequin are you?”
I sat waiting for some time as the primary few characters materialized in response: “[Light bulb icon] Considering — Let me analyze what makes me a” and that was it.
Additionally: My go-to LLM device simply dropped an excellent easy Mac and PC app for native AI – why you must strive it
Ten minutes later, it hadn’t gotten a lot farther.
Let me analyze what makes me a big language mannequin and the best way to clarify this to the person.
First, I want to think about my elementary nature as an AI system. I ought to clarify that I am designed to grasp and generate human language by patterns in massive datasets. The secret’s to be clear
And all the pieces on the Mac had turn into noticeably sluggish.
Forty-five minutes later, glm-4.7-flash was nonetheless producing ideas about considering: “Let me construction this rationalization to first state clearly…,” and so forth.
Trapped in immediate creep
An hour and 16 minutes later — the mannequin “thought” for five,197.3 seconds — I lastly had a solution to my question about what sort of language mannequin glm-4.7-flash was. The reply turned out to not be all that attention-grabbing for on a regular basis spent. It did not inform me a lot about glm that I could not have divined alone, nor something important in regards to the distinction between glm and different massive language fashions:
I figured I used to be carried out with glm at this level. Sadly, Ollama gives no directions for eradicating a mannequin as soon as it is put in regionally. The fashions are stored in a hidden folder “.ollama” within the present person listing on MacOS, inside one other folder referred to as “fashions.” Contained in the fashions folder are two folders, “blobs” and “manifests.” The majority of a mannequin is within the blobs folder. Contained in the manifests is a folder “library” containing a folder named for every mannequin you’ve got downloaded, and inside that, a “newest” folder.
Utilizing the terminal, I deleted the contents of blobs and deleted the contents of every mannequin folder, and that solved the matter. (Jack later knowledgeable me that the terminal command to eliminate any mannequin is “ollama rm
Jack had additionally really useful OpenAI’s current open-source mannequin, gpt-oss, within the 20-billion-parameter taste, “20b,” which he stated was markedly sooner working regionally than others he’d tried. So, I went subsequent to that within the listing.
Additionally: That is the quickest native AI I’ve tried, and it isn’t even shut – the best way to get it
This time, after about six minutes, gpt-oss:20b produced — at a tempo not snail-like, however not swift both — the response that it’s “ChatGPT, powered by OpenAI’s GPT-4 household,” and so forth.
That response was adopted by a pleasant desk of particulars. (Oddly, gpt-oss:20b informed me it had “roughly 175 billion parameters,” which suggests gpt-oss:20b does not totally grasp its personal 20b id.)
At any price, this was high-quality for a easy immediate. But it surely was already clear that I used to be going to have issues with anything extra bold. The sensation of ready for the reply was gradual sufficient — a form of immediate creep, you would possibly say — that I did not dare enterprise so as to add any extra complexity, resembling importing a complete trove of writings.
We’ll want a more moderen machine
OpenAI’s precise ChatGPT on-line service (working GPT5.2) tells me {that a} minimal configuration for a pc working gpt-oss:20b is admittedly 32 gigabytes of DRAM. The M1 Professional silicon of the MacBook has an built-in GPU, and ChatGPT approvingly identified that Ollama has supplied the gpt-oss:20b model with help for the Mac GPU, a library generally known as a “llama.cpp backend.”
Additionally: I attempted the one agentic browser that runs native AI – and located just one draw back
So, all the pieces needs to be OK, however I actually do want extra DRAM than simply 16 gigs. And I must commerce up from the now five-year-old M1 to an M4 or M5. It is fairly fascinating to me, with three many years of writing about computer systems, that for an info employee, we’re speaking about 32 gigabytes because the minimal cheap configuration.
As I discussed lately, DRAM is skyrocketing in worth as a result of all these cloud information facilities are consuming an increasing number of DRAM to run massive language fashions. So, it is me towards the cloud distributors, you could possibly say, and I am going to in all probability be dipping into the bank card to commerce as much as a brand new laptop. (Apple will give me about $599 for my M1 MacBook as a trade-in.)
Whereas my fledgling native Ollama effort did not yield success, it has given me a newfound appreciation for simply how memory-intensive AI is. I at all times knew that from years of reporting on AI, however I now really feel it in my bones, that sense when the response to the immediate takes ceaselessly scrolling throughout the display screen.



