Run Google Gemma2:2b AI on any machine
A short tutorial on running Gemma2:2B on all platforms.
Desktop (Windows / macOS / Linux)
Unless you have graphics card or M series apple chip inference will be slow.
Install ollama.
Open a terminal or command prompt
Pull & Run the Gemma 2B model:
ollama run gemma:2b
Start chatting with the model
To exit, type “/bye” or press Ctrl+C
Mobile (Android / iOS / iPadOS)
It will be quite a bit slower on mobile.
Android
- Download the model - https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_M.gguf
- Install Maid app from play store - https://play.google.com/store/apps/details?id=com.danemadsen.maid&pcampaignid=pcampaignidMKT-Other-global-all-co-prtnr-py-PartBadge-Mar2515-1
- Select llama.cpp if that’s not default.
- Click on top-right -> Model Settings -> Load Model
iOS / iPadOS
- Follow instructions from this site - https://llmfarm.site/
- After you install LLM Farm app you should download gemma2:2b Q4_K_M for fastest inference.
- iOS setup is more complicated because apple… I wish you luck.