1 min read

Running Small LLMs (oxymoron) Locally With…

Running Small LLMs (oxymoron) Locally With…

Running Small LLMs (oxymoron) Locally With an Integrated GPU on a Laptop.

This is what running local models on a consumer-grade laptop without a discrete GPU looks like... Still acceptable for many use cases in my opinion, but I definitely want to upgrade to a machine with a dedicated GPU.

Framework 13

OS: Linux

Hardware: AMD Ryzen 7040Series AMD® Ryzen 7 7840u w/ radeon 780m graphics × 16

Run in Terminal:
ollama run --verbose llama3.2:1b


The "verbose" specification provides the statistics output.

#stats #statistics #education #analytics #econometrics #foss #opensource #learning #nostr #ai #llm #llms #framework #framework13 #frameworklaptop #nostr #foss #fossanalytics #software #linux

This post and comments are published on Nostr.