Email or username:

Password:

Forgot your password?
Simon Willison

Thanks to the combo of Ollama and the llm-ollama plugin you can now run Meta's Llama 3.2 Vision image model (7.9GB) on a Mac and use it to run prompts against images simonwillison.net/2024/Nov/13/

If you have Ollama installed you can fetch the 11B model (7.9 GB) like this:

ollama pull llama3.2-vision

Or the larger 90B model (55GB) like this:

ollama pull llama3.2-vision:90b

I was delighted to learn that Sukhbinder Singh had already contributed support for LLM attachments to Sergey Alexandrov's llm-ollama plugin, which means the following works once you've pulled the models:

llm install --upgrade llm-ollama
llm -m llama3.2-vision:latest 'describe' \
  -a https://static.simonwillison.net/static/2024/pelican.jpg
A photograph of a California Brown Pelican in a harbor
$ llm -m llama3.2-vision:latest 'describe' \
  -a https://static.simonwillison.net/static/2024/pelican.jpg 
This image features a brown pelican standing on rocks, facing the camera and positioned to the left of center. The bird's long beak is a light brown color with a darker tip, while its white neck is adorned with gray feathers that continue down to its body. Its legs are also gray.

In the background, out-of-focus boats and water are visible, providing context for the pelican's environment.
6 comments
Jan

@simon Curious how you’re running Ollama - is it just in your laptop or you have some beefy server running it?

Simon Willison

@jmalonzo just on my laptop, it's a M2 Max with 64GB of RAM so it's pretty good for small and medium sized models

Any model that needs more than 32GB of RAM tends to mean I can't easily run other RAM-hungry apps like VS Code and Firefox

Jeff Triplett

@simon the 90B (55GB) might confuse people.

You do need ~88GB of RAM, not counting your context window, just to run the 90B model size. So 128 GB of RAM, or else you are going to get 1 token per 30 to 45 seconds or more of output while everything swaps around.

That small model is going to run very, very well on any M-series Mac with enough RAM.

Simon Willison

@webology thanks, updated that to say "Or the larger 90B model (55GB download, likely needs ~88GB of RAM) like this:"

Jeff Triplett

@simon I hate that the memory size is so complicated. (that's my best rough guess btw)

Check out Matt's work youtube.com/@technovangelist

Check out this video youtube.com/watch?v=QfFRNF5AhM and one of Matt's about context windows. They are the best I have seen for explaining how much RAM one needs just to run one.

What annoys me is models are measured in Billions and context sizes in Kilobytes, and both are deceptive. i.e. 128K context size = 64 GB of RAM. NOT OBVIOUS at all.

@simon I hate that the memory size is so complicated. (that's my best rough guess btw)

Check out Matt's work youtube.com/@technovangelist

Check out this video youtube.com/watch?v=QfFRNF5AhM and one of Matt's about context windows. They are the best I have seen for explaining how much RAM one needs just to run one.

Go Up