From the course: Large Language Models on AWS: Building and Deploying Open-Source LLMs
Unlock this course with a free trial
Join today to access over 24,600 courses taught by industry experts.
Compiling llama.cpp demo
From the course: Large Language Models on AWS: Building and Deploying Open-Source LLMs
Compiling llama.cpp demo
- [Instructor] Here I have llama.cpp downloaded. How do I know that this code base is installed? Typically what I do, if I forget, is I look at the prompt, right? And I can see that it's in master because I'm using OMIZSH. But I also could do a get remote dash V. And this tells me, oh, okay, good. I'm inside of a repo. Here's exactly the origin, and I can go from there. Now, what I typically would recommend is to optimize the compile if you're going to be working with this. In my particular situation here, what I would typically do, is actually optimize the compile for my architecture. So in this case, let's break through the command. So first step, the time command is typically something I'll do when I do a first compile, just to make sure I have a sanity check and I can look at how long it took to compile. Now if we just run it real quick, you can see here it says, okay, you know, 0.166 total. So it's already been compiled. I don't have to do anything different. But in the case of…
Contents
-
-
-
(Locked)
Implications of Amdahl’s law: A walkthrough4m 5s
-
(Locked)
Compiling llama.cpp demo4m 17s
-
(Locked)
GGUF file format3m 18s
-
(Locked)
Python UV scripting3m 55s
-
Python UV packaging overview1m 59s
-
(Locked)
Key concepts in llama.cpp walkthrough4m 37s
-
(Locked)
GGUF quantized llama.cpp end-to-end demo4m 3s
-
(Locked)
Llama.cpp on AWS G5 demo4m 20s
-
(Locked)
-