[ad_1]
Image by Author
New open source models like LLaMA 2 have become quite advanced and are free to use. You can use them commercially or fine-tune them on your own data to develop specialized versions. With their ease of use, you can now run them locally on your own device.
In this post, we will learn how to download the necessary files and the LLaMA 2 model to run the CLI program and interact with an AI assistant. The setup is simple enough that even non-technical users or students can get it running by following a few basic steps.
To install llama.cpp locally, the simplest method is to download the pre-built executable from the llama.cpp releases.
To install it on Windows 11 with the NVIDIA GPU, we need to first download the llama-master-eb542d3-bin-win-cublas-[version]-x64.zip
file. After downloading, extract it in the directory of your choice. It is recommended to create a new folder and extract all the files in it.
Next, we will download the cuBLAS drivers cudart-llama-bin-win-[version]-x64.zip
and extract them in the main directory. For using the GPU acceleration, you have two options: cuBLAS
for NVIDIA GPUs and clBLAS
for AMD GPUs.
Note: The [version] is the version of the CUDA installed on your local system. You can check it by running
nvcc --version
in the terminal.
To begin, create a folder named “Models” in the main directory. Within the Models folder, create a new folder named “llama2_7b”. Next, download the LLaMA 2 model file from the Hugging Face hub. You can choose any version you prefer, but for this guide, we will be downloading the llama-2-7b-chat.Q5_K_M.gguf
file. Once the download is complete, move the file into the “llama2_7b” folder you just created.
Note: To avoid any errors, please make sure to download only the
.gguf
model files before running the mode.
You can now open the terminal in the main directory. By right clicking and selecting “Open in Terminal” option. You can also open PowerShell and the us “cd” to change directory.
Copy and paste the command below and press “Enter”. We are executing the main.exe
file with model directory location, gpu, color, and system prompt arguments.
./main.exe -m .\Models\llama2_7b\llama-2-7b-chat.Q5_K_M.gguf -i --n-gpu-layers 32 -ins --color -p "<<SYS>> As an AI assistant, your core values include being supportive, considerate, and truthful, ensuring that every interaction is guided by clarity and helpfulness. <</SYS>>"
Our llama.ccp CLI program has been successfully initialized with the system prompt. It tells us it’s a helpful AI assistant and shows various commands to use.
Let’s test out the LLaMA 2 in the PowerShell by providing the prompt. We have asked a simple question about the age of the earth.
The answer is accurate. Let’s ask a follow up question about earth.
As you can see, the model has provided us with multiple interesting facts about our planet.
You can ask the AI assistant to generate code and an explanation in the terminal, which you can easily copy and use in your IDE.
Perfect.
Running Llama 2 locally provides a powerful yet easy-to-use chatbot experience that is customized to your needs. By following this simple guide, you can learn to build your own private chatbot set up in no time without needing to rely on paid services.
The main benefits of running LlaMA 2 locally are full control over your data and conversations as well as no usage limits. You can chat with your bot as much as you want and even tweak it to improve responses.
While less convenient than an instantly available cloud AI API, local setup brings peace of mind regarding data privacy.
Abid Ali Awan (@1abidaliawan) is a certified data scientist professional who loves building machine learning models. Currently, he is focusing on content creation and writing technical blogs on machine learning and data science technologies. Abid holds a Master’s degree in Technology Management and a bachelor’s degree in Telecommunication Engineering. His vision is to build an AI product using a graph neural network for students struggling with mental illness.
[ad_2]
Source link