In the past years, organizations such as OpenAI, Anthropic, Mistral and others have provided us access to low cost yet high performance large language models (LLMs) without the stress of managing the infrastructure. Despite the organizations' promise not to exploit the user's information it handled from their respective models, some businesses remain wary and prefer on-premises systems that process and retain data.
Aside from transparency concerns, people are also skeptical of network latency and downtime associated with using LLMs from these established organizations. To address these issues, people started to initiate and run LLMs locally on personal computers and/or managed servers.
In this article, we'll introduce a remarkable tool designed to run open-source LLMs locally—this is Ollama.
Open-source LLMs are often distributed in the GGUF format, formerly known as GGML. Running these types of models, however, involves significant work since the community has to create a low-level software program to allow their execution. Unfortunately, using this program requires a deep understanding of command lines and specific terminology, deeming it inconvenient to use for the average user.
Ollama, on the other hand, has streamlined the process of running LLMs, making it more accessible to a wider audience. It is an easy-to-use tool that allows users to interact and run open-source LLMs locally, within their desktops, supporting a large model library containing models such as Lama 2, Mistal 7B, and Openchat just to name a few. By bringing it locally to your machine, you control the entire flow which gives a lot of potential. The best part of using Ollama is that it is FREE!
As of this posting, Ollama is only available for macOS and Linux. A Windows OS version is still in the works and will be available soon, so be sure to check their website at ollama.ai.
Head over to Ollama's official website to download and start the installation process.
Check out this link from Mike Bird on another example of a model you can try on Ollama: https://twitter.com/MikeBirdTech/status/1721911715187626260
Like any other programs, Ollama has its fair share of pros and cons that must be discussed to set proper expectations to people who are not familiar with the tool.
For those seeking data privacy with hands-on AI experimentation, Ollama is shaping up to be an invaluable tool. As it continues developing amid the open-source AI movement, we can expect an even more refined user experience and functionality.
The whole point of this article is to show the value for running highly capable LLMs on your local machine to avoid having to use externally hosted proprietary LLMs which you don't have control over. Ollama offers the convenience of having your own LLM that's free such as the one's offered by OpenAI, Mistral, and the likes.
For those seeking data privacy with hands-on AI experimentation, Ollama is shaping up to be an invaluable tool. As it continues developing amid the open-source AI movement, we can expect an even more refined user experience and functionality.
Overall, Ollama has effectively lowered the barrier to locally leveraging powerful language models. This enables broader AI literacy and creativity at the edge. The project shows promising potential to democratize access to the latest advancements in natural language processing.
Sources: