Introduction
Despite the current success of AI assistants (such as ChatGPT), privacy and information security might represent barriers for very large adoption of the technology among some specific users and corporations. Because either they do not want to share the information, or they are legally bound to not share it with other companies, such as OpenAI, Meta and Microsoft.
In other words, to use an AI assistant, you might need to provide key-information in your prompt, but you might not want (or might not be authorized) to share this key-information with this AI service. So the privacy aspect involved in the use of AI assistants might become a “deal-breaker” for these users/corporations.
Banks for example, are usually very careful with sharing sensible user-information with anyone. As a consequence, for example, using OpenAI services might be a deal-breaker for this bank, because the bank cannot share user-information with OpenAI in any circumstance. The bank could be exposed to legal penalties otherwise.
In this situation, it might be a good idea for the bank, to host and run the AI assistant inside their own infrastructure, on a very restrict and secure place. So, in other words, this bank would host and run it’s own ChatGPT on their own servers.
This article explain how you can download and use LLMs (Large Language Models) locally on your own machine, with Ollama. By knowing how to run a LLM model locally in your own machine, you can apply this knowledge to run a LLM model anywhere you want, on your own servers and private services.
Overview
In essence, running a LLM model locally in your machine, consists of following these steps:
- Make sure to download the LLM model you want to use.
- Make sure that the Ollama server is up and running.
- Send your prompt to the local API, and get the response back.
Introducing Ollama
Ollama is a command line util that helps you download, manage and use LLM models. It is almost like a package manager, but for LLM models.
The success of LLM models were so big that scientists, companies, and the general community developed and trained many different LLM models, and made them publicly available, either at GitHub, or at HuggingFace.
So, in other words, the infrastructure and the technology behind AI services is very monopolized right now, with big corporations (OpenAI, Meta, Microsoft, etc.). But the LLM technology per se, is rather accesible and widespread.
Now, we can easily download and use these models through Ollama. The complete list of LLM models available is large, and you can see the full list at the Models page of Ollama’s website.
In order to use Ollama, you have to download and install it. By visiting Ollama’s website you will find instructions to install it in each OS.
After you download and install, you can start to use ollama
in the command line of your OS. There are two main commands in ollama
, which are:
pull
: downloads a LLM model.serve
: starts the LLM model as a HTTP server.run
: run a LLM model with an input prompt.
Download a model
First, we need to download the LLM model we want to use, so that is locally available to use in our machine. You do this by using the pull
command.
For the examples in this article, I will be using the CodeLlama model, which is based on the Llama 2 model. More specifically, the version of the model with 7 billions of parameters. It is a LLM model refined and tuned for code related questions and assignments.
ollama pull codellama:7b
LLM models are very heavy and expensive to run
In general, LLM models are very heavy and expensive to run, and they do not scale particularly well. So to make services like ChatGPT, that serves millions of users everyday, the actual LLM model, or the actual ChatGPT that answers your questions, is running on a big data center, more specifically, on a very (I mean, VERY) expensive and powerful set of GPUs (with hundreds and hundreds of VRAM available to use).
All of this means that, not every machine is capable of running a LLM model. Specially if your machine does not have a good GPU card in it. So, if you want to run a LLM model on a “weak machine”, always choose the smallest version possible of a LLM model.
A commom way to classifying LLM models is measuring them by their size. The size of a LLM model is normally measured by the number of parameters of the model, and they are normally in the level of billions of parameters. So a LLM model usually have 1, 7, 20, 50, 100 billions parameters.
In Ollama, you choose the version of the model with the number of the parameters you want, by writing a colon followed by the number of parameters you want. So the text <name of model>:20b
represents the version of a model with 20 billion parameters.
In my example here, I am running on very standard laptop, literally nothing fancy. It has 12GB RAM, and a normal Intel i5 Core CPU. And it does not have any GPU card installed.
Because of that, I do not want to select a very large model. I want to choose the smallest model as possible, with the smallest number of parameters. That is why I choose specifically the version of CodeLlama with 7 billion parameters - codellama:7b
(there are different versions of CodeLlama that go up to 70 billion parameters).
So, you should take the power of your machine in consideration when choosing your LLM model.
Prompt your LLM model
Let’s quickly use our LLM model. You can simply run the LLM model with an input prompt, by using the run
command of ollama
.
For example, I can ask CodeLlama to write me a Python function that outputs the fibonacci sequence, like this:
ollama run codellama:7b "Write me a Python function that outputs the fibonacci sequence"
This is the output I get:
[PYTHON]
def fibonacci(n):
if n <= 1:
return n
else:
return fibonacci(n-1) + fibonacci(n-2)
[/PYTHON]
[TESTS]
# Test case 1:
assert fibonacci(0) == 0
# Test case 2:
assert fibonacci(1) == 1
# Test case 3:
assert fibonacci(2) == 1
# Test case 4:
assert fibonacci(3) == 2
# Test case 5:
assert fibonacci(4) == 3
# Test case 6:
assert fibonacci(5) == 5
# Test case 7:
assert fibonacci(6) == 8
# Test case 8:
assert fibonacci(7) == 13
# Test case 9:
assert fibonacci(8) == 21
# Test case 10:
assert fibonacci(9) == 34
[/TESTS]
Start your LLM model as a HTTP server
In Ollama, you can use your LLM model as a HTTP server. You start the server, and send input prompts to it. This server will response you back, with the output of the LLM model.
Fist, you start the server with the serve
command:
ollama serve
After that, the endpoint 'api/generate'
become available at port 11434 in localhost.
import requests
= 'http://localhost:11434/api/generate'
url = '{"model": "codellama:7b", "prompt": "Write me a Python function that outputs the fibonacci sequence"}'
body = requests.post(url, data = body)
r print(str(r.content))
b'{"model":"codellama:7b","created_at":"2024-03-18T14:45:08.92502Z","response":"\\n","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:09.4067275Z","response":"[PY","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:09.6486965Z","response":"TH","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:09.8871347Z","response":"ON","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:10.1047816Z","response":"]","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:10.3401777Z","response":"\\n","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:10.5807352Z","response":"def","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:10.8011125Z","response":" fib","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:11.0347653Z","response":"on","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:11.2923311Z","response":"acci","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:11.5700897Z","response":"(","done":false}\n
{"model":"codellama:7b","created_at":"2024-03-18T14:45:11.8455609Z","response":"n","done":false}\n
... truncated output
We can see above that we get a JSON-like type of response from the API, with the contents produced by the LLM model inside the element "response"
in each JSON object.
We can parse this JSON response from the API, and collect all the "response"
elements to form the full text produced by the LLM model.
Using the same URL and HTTP Request Body, I could parse the HTTP Response using something similar to this. You can see in the example below, that we get the exact same result as we got when we executed the LLM model through the run
command of ollama
:
import json
import re
def parse_response(resp: str) -> dict:
= resp.content.decode("utf-8")
resp = re.sub(r'\}\n\{', '},\n{', resp)
resp = f'[{resp}]'
resp return json.loads(resp)
= requests.post(url, data = body)
response = parse_response(response)
response = list()
tokens for token in response:
'response'])
tokens.append(token[
print(''.join(tokens))
[PYTHON]
def fibonacci(n):
if n <= 1:
return n
else:
return fibonacci(n-1) + fibonacci(n-2)
[/PYTHON]
[TESTS]
# Test case 1:
assert fibonacci(0) == 0
# Test case 2:
assert fibonacci(1) == 1
# Test case 3:
assert fibonacci(2) == 1
# Test case 4:
assert fibonacci(3) == 2
# Test case 5:
assert fibonacci(4) == 3
# Test case 6:
assert fibonacci(5) == 5
# Test case 7:
assert fibonacci(6) == 8
# Test case 8:
assert fibonacci(7) == 13
# Test case 9:
assert fibonacci(8) == 21
# Test case 10:
assert fibonacci(9) == 34
[/TESTS]
Going further
You can now start to develop this new knowledge you acquired in this article. Going further in your AI journey.
For example, you can:
- test different LLM models, by downloading new models with
ollama pull
. - deploy and run Ollama on a cloud environment with specialized infrastructure (like P4 instances in Amazon EC2).
- use an AI app framework, such as LangChain, to build an AI app that uses Ollama.