Pre-Event Setup Instructions

Greetings, fellow explorer!

Thanks for registering for the event Pie & AI Pune: Your Personal AI – Run Models Locally.

To get the most from this event, please perform these quick setup steps before the event:

Setup Steps

1 Check your system RAM size

Check your system information to find out how much RAM it has.

Note: To run Open Source models locally, your system should preferably have a minimum of 8GB RAM. If you don't have enough RAM, you can instead run Ollama's cloud models.

2 Create Ollama Account & Install Application

Create an account on Ollama here: https://ollama.com/signup

Also download and install the Ollama desktop application for your operating system.
The Ollama desktop application and Ollama Command Line Inteface (CLI) are now available on your system.

3 Sign into your Ollama Account

Start the Ollama desktop application. From the Settings of your Ollama desktop application, sign into your Ollama account.

On successful sign in, you might be prompted to open the desktop application. Accept/OK the prompt.

Ollama Settings Screen

Ollama Settings Screen

4 Install A Model

Model Download and Installation with Prompt in Ollama application

i Click New Chat on the desktop application if another chat is open.Otherwise, skip to the next step.

ii Enter a model name phi3:3.8b.

iii Enter any text prompt. Here's an example:

"Which models on Ollama can be run on systems with less than 5GB RAM?"

iv Submit the prompt.

Note: You might notice a progress bar in the desktop application indicating download and installation of the phi3:3.8b model, before you see a response to the prompt. Wait for the download and installation to finish, and the complete response to be displayed.

+ Troubleshooting: Model installation failure due network connection errors

If you see an error like "max retries exceeded" or "dial tcp: lookup... no such host", this indicates a network connectivity issue.

Network error during model installation in Ollama desktop app

Model installation error in Ollama desktop application

Solutions:

  • Check your internet connection and ensure it's stable.
  • Disable VPN or proxy if you're using one.
  • Check whether your firewall settings are blocking Ollama from downloading models.
  • Try again after a few minutes. The model registry might be temporarily unavailable.
  • If the problem persists, try using the CLI method: ollama pull phi3:3.8b.

5 Verify Model Installation

Open a command prompt or terminal on your system and run the following command:

ollama list

You should see the model phi3:3.8b listed in the output.

Phi3 Installation Verification Screenshot

Verification of Phi3 Model Installation

Congratulations! You have installed your first Open Source model phi3:3.8b on Ollama to run on your system.

6 Install Gemma3 Model

Now install another Open Source model Gemma3:4b by running this command in the command prompt or terminal:

ollama pull gemma3:4b

Note: You might see the model installation in progress. Wait for the installation to finish.
This is another way of installing Open Source models on Ollama.

+ Troubleshooting: Model installation errors

Same as in the Ollama desktop application, the "max retries exceeded" errors in Ollama CLI also indicate network connectivity issues.:

Network error during model installation in Ollama CLI

Model installation error in Ollama CLI

Solutions:

See the troubleshooting steps for Ollama desktop application.

7 Verify Model Installation

Run the following command again in the command prompt or terminal to confirm that gemma3:4b is now added to the installed model list:

ollama list

8 Create Hugging Face Account

Create a Hugging Face account here: https://huggingface.co/join

That's it! You are ready to run models locally on Ollama desktop and CLI.

Explore the next steps