Enqurious logo
Back to blog
Guides & Tutorials

Basics of Langchain

Basics of Langchain blog cover image
Large Language Model
GenAI
Langchain
Burhanuddin NahargarwalaJr. Data Engineer

I asked ChatGPT a business-specific question, and unfortunately, it couldn’t provide a satisfactory answer. Wondering why?

Let's pose a question to ChatGPT: "List all products with limited stocks."

Business_related_question_to_gpt.png

The response is vague and unsatisfactory. But why is that? What other options do we have other than ChatGPT that can help us? Let’s explore all possibilities, starting with understanding ChatGPT:

The GenAI Revolution and ChatGPT: A Prelude

In the dynamic realm of GenAI, one name stands out: ChatGPT. A familiar companion in our digital interactions, ChatGPT is where we turn with our lots of questions, expecting precise and informed answers. Its understanding of context became unparalleled, enabling it to seamlessly navigate complex conversations with users. It could effortlessly switch between topics, understand humor, and adapt its tone to match the preferences of its questioner. So don’t underestimate the power of common GPT, I mean ChatGpt.

power_of_genai_meme.png

When we ask a question to ChatGPT, instead of directly passing it to ChatGPT, it is first converted into tokens via a tokenizer.

Large Language Models receive a text as input and generate a text as output. However, being statistical models, they work much better with numbers than text sequences. That’s why every input to the model is processed by a tokenizer, before being used by the core model.

A token is a chunk of text consisting of a variable number of characters, so the tokenizer's main task is splitting the input into an array of tokens. Then, each token is mapped to a token index, which is the integer encoding of the original text chunk. Click here to see how the text gets converted into tokens.

tokenizer.png

Challenges come in the situation when the inquiries are not just general but deeply rooted in specific contexts, like queries related to an e-commerce business, real-time queries, etc.

Imagine you've launched an e-commerce venture. Your goal is to offer an unparalleled customer experience with instant query resolution. But can ChatGPT, a generalist in its intelligence, be able to handle the business-specific questions that come its way?

We just saw that it’s not that efficient. Why? ChatGPT, despite its brilliance, is a general-purpose model. It's designed for broad queries and not context-specific questions critical to your business. This is where LangChain enters, a superhero with the powers to solve specialized queries 🤽.

Introduction to LangChain

LangChain is the bridge that connects ChatGPT's general intelligence with the specific needs of your context. It empowers ChatGPT to access external tools, from Wikipedia and search engines to databases, CSV files, and documents.

With LangChain, you can feed ChatGPT the context it needs. Imagine ChatGPT, now equipped with LangChain, being able to pull data from your business database or analyze customer trends from a CSV file. Suddenly, the answers become tailored, precise, and contextually aware.

Developers around the globe have harnessed the power of LangChain and Streamlit to build exceptional chatbots. Take a look at this remarkable example and see how LangChain elevates ChatGPT to new heights.

LangChain is not just a tool; it's a revolutionary framework transforming the way we interact with Large Language models (LLMs). At its heart, LangChain allows us to orchestrate complex AI tasks with remarkable simplicity and efficiency. Let's unpack the key components that make LangChain a game-changer in the world of AI.

Imagine LangChain as a sophisticated puzzle, where each piece is a component that, when connected, forms a more powerful and complex AI application. This "chaining" is what gives LangChain its unique ability to handle advanced use cases involving LLMs. Chains may consist of multiple components from several modules:

  1. Prompt Templates: Prompt templates are templates for different types of prompts. Like “chatbot” style templates, ELI5 (Explain Like I'm 5) question-answering, etc

  2. LLMs: Large language models like GPT-3, Google PaLM 2, Llama 2, etc

  3. Chains: Chains in LangChain are used to execute a sequence of commands that involves LLMs.

  4. Agents: Agents use LLMs to decide what actions should be taken. Tools like web search or calculators can be used, and all are packaged into a logical loop of operations.

  5. Memory: Short-term memory, long-term memory.

LLMs:

In LangChain, there are broadly two types of large language models (LLMs). Let’s understand it via the given image:

Langchain types of LLM.jpg

  1. The first model is LLM, which is designed to process and generate language based on string inputs and outputs.

  2. ChatModels in LangChain are more complex and designed to handle conversational contexts. The input for a ChatModel is a list of ChatMessages, not just a single string. The output of a ChatModel is a single ChatMessage, which is the model's response to the ongoing conversation.


    A chat message has two required components:

    1. content: This is the content of the message.

    2. role: This is the role of the entity from which the chat message is coming from.

Create an LLM object

Step 1: Install LangChain

First, you'll need to install the LangChain library. You can do this by running the following command in your Python environment:

!pip install langchain

Using LangChain will usually require integrations with one or more model providers, data stores, APIs, etc. For this example, we'll use OpenAI's model APIs. You can go with other models also, such as HuggingFace, Llama, etc.

Step 2: Install OpenAI Python Package

Next, install the OpenAI Python package, which will allow you to interact with OpenAI's API:

! pip install openai==0.27.9

Step 3: Obtain OpenAI API Key

To use OpenAI's models, you'll need an API key. Follow these steps to get one:

  1. Create/Open OpenAI account:

    • If you don't already have an OpenAI account, create one.

    • Visit OpenAI's website and sign up or log in.

  2. Generate API Key:

    1. Once logged in, navigate to the API section, where you can generate a new API key.

    2. Follow the on-screen instructions to create a key.

  3. Free Credit for New Users:

    1. As a new user, you'll typically receive free credits to get started. This allows you to experiment with the API without incurring an initial cost.

After setting up LangChain and obtaining your OpenAI API key, you can start using LangChain with OpenAI's LLMs.

Let’s import the OpenAI class from the langchain package:

from langchain.llms import OpenAI
import os

# First export the OPENAI_API_KEY in environment variable
Os.environ["OPENAI_API_KEY"] = " sk-XXXX"

# Create LLM model
llm = OpenAI()
llm('Python is founded by whom and in which year?')

o/p: '\n\nPython was founded by Guido van Rossum in 1991.'

This is how it worked. The input that we passed is a single string, and as an output, we got a string.

OpenAI has two important parameters that are:

  1. temperature: It controls the randomness and creativity of the model’s output. Its value ranges from 0 to 1.

    1. The low temperature (close to 0) results in more deterministic, predictable, and conservative outputs, suitable for tasks where consistency and accuracy are more important than creativity. Example: A temperature of 0.1 might be used for factual data retrieval or business analytics.

    2. High temperatures lead to more varied, random, and creative outputs. Ideal for tasks requiring innovation, such as creative writing, brainstorming, or generating diverse ideas. Example: A temperature of 0.9 might be used for creative story writing or generating unique ideas.

  2. model_name: The model_name parameter allows you to specify which of OpenAI's LLMs you want to use. Different models have varying capabilities, sizes, and computational requirements.

# Create LLM model
llm = OpenAI(temperature=0.5, model_name='gpt-3.5-turbo')

Similarly, we can create Chat models, for that import ChatOpenAI from langchain.chat_models:

# We can pass the list of messages, as
from langchain.chat_models import ChatOpenAI
from langchain.schema.messages import HumanMessage, SystemMessage

messages = [
    SystemMessage(content="You're a helpful assistant. In case if you don't know the answer then reply Sorry I don't know the answer!"),
    HumanMessage(content='What is object oriented programming?')
]

# Create LLM chat model
chat_llm = ChatOpenAI()
chat_llm.invoke(messages)

o/p: AIMessage(content='Object-oriented programming (OOP) is a programming…)

In addition to passing a list of HumanMessage and SystemMessage, you can also pass a simple string to the ChatOpenAI model. This is useful for simpler interactions where you don't need the structure of HumanMessage or SystemMessage.

response = chat_llm.invoke("Who is the father of computer?")
print(response)

o/p: AIMessage(content='The father of computer is considered to be Alan Turing, an English mathematician, logician, and computer scientist who laid the foundations for modern computing.')

Conclusion:

LangChain is a key tool that makes advanced AI models like ChatGPT more useful for specific tasks. We've looked at how LangChain helps these models understand and respond better in different situations, by using various settings and approaches. Going forward, we'll explore more about how LangChain works, including its use of templates, chains, and agents, to make AI interactions even smarter and more relevant. Essentially, LangChain is a powerful way to make AI more tailored and effective for our needs.