Skip to content

All course material for the LangChain course from Aurelio AI.

Notifications You must be signed in to change notification settings

aurelio-labs/langchain-course

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

85 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LangChain Course

Welcome to the LangChain course by Aurelio AI!

Getting Started

Python Environment (IMPORTANT)

This course repo contains everything you need to install an exact duplicate Python environment as used during the course creation.

Installing Python Venvs

The Python packages are managed using the uv package manager, and so we must install uv as a prerequisite for the course. We do so by following the installation guide. For Mac users, as of 22 Oct 2024 enter the following in your terminal:

curl -LsSf https://astral.sh/uv/install.sh | sh

Once uv is installed and available in your terminal you can navigate to the course root directory and execute:

uv python install 3.12.7
uv venv --python 3.12.7
uv sync

❗️ You may need to restart the terminal if the uv command is not recognized by your terminal.

With that we have our chapter venv installed. When working through the code for a specific chapter, always create a new venv to avoid dependency hell.

Using Venv in VS Code / Cursor

To use our new venv in VS Code or Cursor we simply execute:

cd example-chapter
cursor .  # run via Cursor
code .    # run via VS Code

This command will open a new code window, from here you open the relevant files (like Jupyter notebook files), click on the top-right Select Environment, click Python Environments..., and choose the top .venv environment provided.

Uninstalling Venvs

Naturally, we might not want to keep all of these venvs clogging up the memory on our system, so after completing the course we recommend removing the venv with:

deactivate
rm -rf .venv -r

Ollama

The course can be run using OpenAI or Ollama. If using Ollama, you must go to ollama.com and install Ollama for your respective OS (MacOS is recommended).

Whenever an LLM is used via Ollama you must:

  1. Ensure Ollama is running by executing ollama serve in your terminal or running the Ollama application. Make sure to keep note of the port the server is running on, by default Ollama runs on http://localhost:11434

  2. Download the LLM being used in your current example using ollama pull. For example, to download Llama 3.2 3B, we execute ollama pull llama 3.2:3b in our terminal.

About

All course material for the LangChain course from Aurelio AI.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published