In this Hugging Face pipeline tutorial for beginners we'll use Llama 2 by Meta. We will load Llama 2 and run the code in the free Colab Notebook. You'll learn how to chat with Llama 2 (the most hyped open source llm) easily thanks to the Hugging Face library.
Happy watching!
+++ Useful Links +++
Colab Notebook from this video: https://colab.research.google.com/dri...
Get your Llama-2 for the project: https://huggingface.co/meta-llama/Lla...
+++ Hugging Face +++
Blog post for running Llama 2 (Llama 2 is here) by Hugging Face: https://huggingface.co/blog/llama2
Hugging Face pipelines documentation: https://huggingface.co/docs/transform...
Hugging Face pipeline tasks list: https://huggingface.co/docs/transform...
+++ Chat with Llamas +++
Chat with Llama-2 7B on Hugging Face: https://huggingface.co/spaces/hugging...
Chat with Llama-2 70B on Hugging Face: https://huggingface.co/spaces/ysharma...
Chapters:
0:00 Intro
1:35 Changing runtime type to GPU on Colab
2:10 Install Hugging Face & PyTorch
2:35 Get access to Llama 2
3:28 Login into Hugging Face in Colab
4:17 Loading Llama 2 Model and Tokenizer into Colab
5:20 Creating the Hugging Face Pipeline
6:23 Generate Llama 2 responses
8:07 Llama Response #1
8:30 Llama Response #2
10:10 Llama Response #3
10:55 Llama Response #4
11:30 Llama Response #5
11:55 Conversational Llama 2
12:34 Outro
Connect with me:
LinkedIn: / kris-ograbek
Medium: / kris-ograbek