\ \ Did you know you can run powerful AI models right on your computer? It’s true! Today, I will show you how easy it is to get started with Llama 3.2 and OpenWebUI.
https://youtu.be/woUB_vhoML0?si=cpnjLLPb1KpcacEQ&embedable=true
\ \
Running Llama 3.2 Locally: A Comprehensive Guide Introduction to Llama 3.2Llama 3.2 is the latest iteration of Meta’s open-source language model, offering enhanced capabilities for text and image processing. It is designed to run efficiently on local devices, making it ideal for applications that require privacy and low latency. The model comes in various sizes, including 1B, 3B, and 11B parameters. In this tutorial, I’m going to use the 1B model, but you can download any you like.
\
\ Before setting up Llama 3.2 locally, ensure you have the following:
\
Step-by-Step Installation Guide Step 1: Install Docker\ Docker is a tool that allows developers to package applications and their dependencies into a standardized unit called a container, which can run consistently across different computing environments. Unlike virtual machines, containers are lightweight and share the host system’s operating system, making them more efficient and faster to start.
\
You can download the Docker Desktop application from the Docker website https://www.docker.com/ or via the terminal. Docker is available for Mac, Windows, and Linux.
\
How to Install Docker on UbuntuDownload the package https://desktop.docker.com/linux/main/amd64/docker-desktop-amd64.deb
\ Then use these commands:
sudo apt-get update sudo apt-get install [path to the docker package].deb\ More details: https://docs.docker.com/desktop/install/linux/ubuntu/
\
Step 2: Install OllamaOllama is essential for running large language models like Llama 3.2 locally.
Follow these steps to install it:
I’m going to use the terminal and this command:
curl -fsSL https://ollama.com/install.sh | sh\
Step 3: Download a New Llama 3.2 model from Ollama websiteOnce Ollama is set up, you can download Llama 3.2 models. All the available models you can find here https://ollama.com/library
Since I want to use 1B parameters model I’ll be using this command:
\
ollama run llama3.2:1b\
Step 4: Install OpenWebUI with Default ConfigurationOpenWebUI is a self-hosted, extensible web interface designed to interact entirely offline with large language models (LLMs). It offers a user-friendly experience similar to ChatGPT, supports integration with various LLMs, such as those compatible with OpenAI and Ollama, and provides features like markdown support, model management, and multi-user access.
\ Go to https://docs.openwebui.com/getting-started/ and find the section Quick Start with Docker. Copy the code and then run it in the terminal.
\ I’m going to run this code:
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v ollama:/root/.ollama -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:ollama\ We will run a Docker command. It is used to run a containerized version of the OpenWebUI application.
Once the installation is finished, open your browser and go to http://0.0.0.0:3000/. If everything is okay, you will see the website. Create an account, and then you will see the welcome screen, which is similar to ChatGPT.
\
Troubleshooting Common Issue
WebUI models not being pulled/appliedSome of you could face an issue when the models are not available in the drop-down. To fix it, follow these instructions:
\
2. Click on “Admin Settings”.
3. Click on “Models.”
4. In the input field, write the name of the model you would like to pull and click on the “download” icon.
\ All the names you can find on https://ollama.com/library/llama3.2
\ This should fix the issue.
\
ConclusionRunning Llama 3.2 locally provides significant advantages regarding privacy and control over AI applications. But to have a smooth experience, you would need a powerful computer. :)
\ If you like this tutorial, please follow me on YouTube.
\ Thanks! :)
All Rights Reserved. Copyright , Central Coast Communications, Inc.