id
stringlengths
6
6
text
stringlengths
20
17.2k
title
stringclasses
1 value
148678
import { CheerioWebBaseLoader } from "@langchain/community/document_loaders/web/cheerio"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { HuggingFaceTransformersEmbeddings } from "@langchain/community/embeddings/hf_...
148683
/* eslint-disable import/first */ /* eslint-disable arrow-body-style */ /* eslint-disable import/no-duplicates */ import { ChatOpenAI } from "@langchain/openai"; const chat = new ChatOpenAI({ model: "gpt-3.5-turbo-1106", temperature: 0.2, }); import { CheerioWebBaseLoader } from "@langchain/community/document_lo...
148687
import { ChatOpenAI } from "@langchain/openai"; import { createSqlQueryChain } from "langchain/chains/sql_db"; import { SqlDatabase } from "langchain/sql_db"; import { DataSource } from "typeorm"; import { QuerySqlTool } from "langchain/tools/sql"; import { PromptTemplate } from "@langchain/core/prompts"; import { Stri...
148692
import { StringOutputParser } from "@langchain/core/output_parsers"; import { ChatPromptTemplate, PromptTemplate } from "@langchain/core/prompts"; import { RunnableSequence } from "@langchain/core/runnables"; import { ChatOpenAI } from "@langchain/openai"; import { createSqlQueryChain } from "langchain/chains/sql_db"; ...
148694
import { ChatOpenAI } from "@langchain/openai"; import { createSqlQueryChain } from "langchain/chains/sql_db"; import { SqlDatabase } from "langchain/sql_db"; import { DataSource } from "typeorm"; import { QuerySqlTool } from "langchain/tools/sql"; const datasource = new DataSource({ type: "sqlite", database: "../...
148698
import { MemoryVectorStore } from "langchain/vectorstores/memory"; import { SemanticSimilarityExampleSelector } from "@langchain/core/example_selectors"; import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { FewShotPromptTemplate, PromptTemplate } from "@langchain/core/prompts"; import { createSqlQ...
148700
import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { SemanticSimilarityExampleSelector } from "@langchain/core/example_selectors"; import { FewShotPromptTemplate, PromptTemplate, ChatPromptTemplate, SystemMessagePromptTemplate, MessagesPlaceholder, } from "@langchain/core/prompts"; im...
148703
import { ChatPromptTemplate, MessagesPlaceholder, } from "@langchain/core/prompts"; import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { AgentExecutor, createOpenAIToolsAgent } from "langchain/agents"; import { SqlToolkit } from "langchain/agents/toolkits/sql"; import { SqlDatabase } from "lan...
148723
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { ConversationalRetrievalQAChain } from "langchain/chains"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import { BufferMemory } from "langchain/mem...
148724
import { LLMChain } from "langchain/chains"; import { ChatOpenAI } from "@langchain/openai"; import { ChatPromptTemplate } from "@langchain/core/prompts"; // We can also construct an LLMChain from a ChatPromptTemplate and a chat model. const chat = new ChatOpenAI({ temperature: 0 }); const chatPrompt = ChatPromptTempl...
148742
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { LLMChain } from "langchain/chains"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import { BufferMemory } from "langchain/memory"; import * as fs f...
148743
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { ConversationalRetrievalQAChain } from "langchain/chains"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import { BufferMemory } from "langchain/mem...
148763
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { ConversationalRetrievalQAChain } from "langchain/chains"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import { BufferMemory } from "langchain/mem...
148770
import { OpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { ConversationalRetrievalQAChain } from "langchain/chains"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import * as fs from "fs"; /* Initialize the LL...
148773
import { OpenAI } from "@langchain/openai"; import { ConversationChain } from "langchain/chains"; const model = new OpenAI({}); const chain = new ConversationChain({ llm: model }); const res1 = await chain.invoke({ input: "Hi! I'm Jim." }); console.log({ res1 }); const res2 = await chain.invoke({ input: "What's my nam...
148778
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import * as fs from "fs"; import { formatDocumentsAsString } from "langchain/util/document"; import { PromptTe...
148779
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { ConversationalRetrievalQAChain } from "langchain/chains"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { BufferMemory } from "langchain/memory"; const CUSTOM_QUESTION_GENERATOR_CHAIN_PROMPT = `Given the following ...
148781
import { OpenAI } from "@langchain/openai"; import { loadQAStuffChain, loadQAMapReduceChain } from "langchain/chains"; import { Document } from "@langchain/core/documents"; // This first example uses the `StuffDocumentsChain`. const llmA = new OpenAI({}); const chainA = loadQAStuffChain(llmA); const docs = [ new Doc...
148783
import { z } from "zod"; import { zodToJsonSchema } from "zod-to-json-schema"; import { ChatOpenAI } from "@langchain/openai"; import { ChatPromptTemplate } from "@langchain/core/prompts"; import { JsonOutputToolsParser } from "@langchain/core/output_parsers/openai_tools"; const EXTRACTION_TEMPLATE = `Extract and save...
148784
import { TokenTextSplitter } from "@langchain/textsplitters"; import fs from "fs"; import path from "path"; import { Document } from "@langchain/core/documents"; export const run = async () => { /* Split text */ const text = fs.readFileSync( path.resolve(__dirname, "../../state_of_the_union.txt"), "utf8" ...
148785
import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; export const run = async () => { const text = `Hi.\n\nI'm Harrison.\n\nHow? Are? You?\nOkay then f f f f. This is a weird text to write, but gotta test the splittingggg some how.\n\n Bye!\n\n-H.`; const splitter = new RecursiveChara...
148786
import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; import { Document } from "@langchain/core/documents"; const text = `Some other considerations include: - Do you deploy your backend and frontend together, or separately? - Do you deploy your backend co-located with your database, or separately...
148787
import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; const text = ` --- sidebar_position: 1 --- # Document transformers Once you've loaded documents, you'll often want to transform them to better suit your application. The simplest example is you may want to split a long document into smaller ch...
148789
import { RecursiveCharacterTextSplitter } from "@langchain/textsplitters"; const text = `<!DOCTYPE html> <html> <head> <title>🦜️🔗 LangChain</title> <style> body { font-family: Arial, sans-serif; } h1 { color: darkblue; } </style> </head> <body> <div> ...
148790
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { CharacterTextSplitter } from "@langchain/textsplitters"; import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { ChatPromptTemplate } from "@langchain/core/prompts"; import { createStuffDocumentsChain } from "langchain/chai...
148793
import { CharacterTextSplitter } from "@langchain/textsplitters"; import { Document } from "@langchain/core/documents"; export const run = async () => { /* Split text */ const text = "foo bar baz 123"; const splitter = new CharacterTextSplitter({ separator: " ", chunkSize: 7, chunkOverlap: 3, }); ...
148801
import { OpenAIEmbeddings } from "@langchain/openai"; import { AstraDBVectorStore, AstraLibArgs, } from "@langchain/community/vectorstores/astradb"; const astraConfig: AstraLibArgs = { token: process.env.ASTRA_DB_APPLICATION_TOKEN as string, endpoint: process.env.ASTRA_DB_ENDPOINT as string, collection: proc...
148803
import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { OpenAIEmbeddings } from "@langchain/openai"; // Create a vector store through any method, here from texts as an example const vectorStore = await HNSWLib.fromTexts( ["Hello world", "Bye bye", "hello nice world"], [{ id: 2 }, { id: 1 }, {...
148808
import { USearch } from "@langchain/community/vectorstores/usearch"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TextLoader } from "langchain/document_loaders/fs/text"; // Create docs with a loader const loader = new TextLoader("src/document_loaders/example_data/example.txt"); const docs = await loa...
148813
import { MongoDBAtlasVectorSearch } from "@langchain/mongodb"; import { CohereEmbeddings } from "@langchain/cohere"; import { MongoClient } from "mongodb"; const client = new MongoClient(process.env.MONGODB_ATLAS_URI || ""); const namespace = "langchain.test"; const [dbName, collectionName] = namespace.split("."); con...
148816
import { HNSWLib } from "@langchain/community/vectorstores/hnswlib"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TextLoader } from "langchain/document_loaders/fs/text"; // Create docs with a loader const loader = new TextLoader("src/document_loaders/example_data/example.txt"); const docs = await loa...
148820
// If you want to import the browser version, use the following line instead: // import { CloseVectorWeb } from "@langchain/community/vectorstores/closevector/web"; import { CloseVectorNode } from "@langchain/community/vectorstores/closevector/node"; import { OpenAIEmbeddings } from "@langchain/openai"; // Create a ve...
148823
import { FaissStore } from "@langchain/community/vectorstores/faiss"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TextLoader } from "langchain/document_loaders/fs/text"; // Create docs with a loader const loader = new TextLoader("src/document_loaders/example_data/example.txt"); const docs = await lo...
148824
import { MongoDBAtlasVectorSearch } from "@langchain/mongodb"; import { CohereEmbeddings } from "@langchain/cohere"; import { MongoClient } from "mongodb"; const client = new MongoClient(process.env.MONGODB_ATLAS_URI || ""); const namespace = "langchain.test"; const [dbName, collectionName] = namespace.split("."); con...
148828
import { MemoryVectorStore } from "langchain/vectorstores/memory"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TextLoader } from "langchain/document_loaders/fs/text"; // Create docs with a loader const loader = new TextLoader("src/document_loaders/example_data/example.txt"); const docs = await loade...
148832
import { MemoryVectorStore } from "langchain/vectorstores/memory"; import { OpenAIEmbeddings } from "@langchain/openai"; const vectorStore = await MemoryVectorStore.fromTexts( ["Hello world", "Bye bye", "hello nice world"], [{ id: 2 }, { id: 1 }, { id: 3 }], new OpenAIEmbeddings() ); const resultOne = await vec...
148842
import { FaissStore } from "@langchain/community/vectorstores/faiss"; import { OpenAIEmbeddings } from "@langchain/openai"; // Create a vector store through any method, here from texts as an example const vectorStore = await FaissStore.fromTexts( ["Hello world", "Bye bye", "hello nice world"], [{ id: 2 }, { id: 1 ...
148844
import { OpenAIEmbeddings } from "@langchain/openai"; import { Neo4jVectorStore } from "@langchain/community/vectorstores/neo4j_vector"; /** * `similaritySearch` Method with Metadata Filtering: * * Description: * This method facilitates advanced similarity searches within a Neo4j vector index, leveraging both text...
148854
import { Chroma } from "@langchain/community/vectorstores/chroma"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TextLoader } from "langchain/document_loaders/fs/text"; // Create docs with a loader const loader = new TextLoader("src/document_loaders/example_data/example.txt"); const docs = await loade...
148861
import { createClient } from "redis"; import { OpenAIEmbeddings } from "@langchain/openai"; import { RedisVectorStore } from "@langchain/redis"; import { Document } from "@langchain/core/documents"; const client = createClient({ url: process.env.REDIS_URL ?? "redis://localhost:6379", }); await client.connect(); con...
148862
import { createClient } from "redis"; import { OpenAIEmbeddings } from "@langchain/openai"; import { RedisVectorStore } from "@langchain/redis"; import { Document } from "@langchain/core/documents"; const client = createClient({ url: process.env.REDIS_URL ?? "redis://localhost:6379", }); await client.connect(); con...
148863
import { createClient } from "redis"; import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { RedisVectorStore } from "@langchain/redis"; import { ChatPromptTemplate } from "@langchain/core/prompts"; import { createStuffDocumentsChain } from "langchain/chains/combine_documents"; import { createRetrie...
148864
import { createClient } from "redis"; import { OpenAIEmbeddings } from "@langchain/openai"; import { RedisVectorStore } from "@langchain/redis"; import { Document } from "@langchain/core/documents"; const client = createClient({ url: process.env.REDIS_URL ?? "redis://localhost:6379", }); await client.connect(); con...
148885
/* eslint-disable @typescript-eslint/no-non-null-assertion */ import { Pinecone } from "@pinecone-database/pinecone"; import { Document } from "@langchain/core/documents"; import { OpenAIEmbeddings } from "@langchain/openai"; import { PineconeStore } from "@langchain/pinecone"; // Instantiate a new Pinecone client, wh...
148886
/* eslint-disable @typescript-eslint/no-non-null-assertion */ import { Pinecone } from "@pinecone-database/pinecone"; import { OpenAIEmbeddings } from "@langchain/openai"; import { PineconeStore } from "@langchain/pinecone"; // Instantiate a new Pinecone client, which will automatically read the // env vars: PINECONE_...
148887
/* eslint-disable @typescript-eslint/no-non-null-assertion */ import { Pinecone } from "@pinecone-database/pinecone"; import { OpenAIEmbeddings } from "@langchain/openai"; import { PineconeStore } from "@langchain/pinecone"; // Instantiate a new Pinecone client, which will automatically read the // env vars: PINECONE_...
148893
import { QdrantVectorStore } from "@langchain/qdrant"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TextLoader } from "langchain/document_loaders/fs/text"; // Create docs with a loader const loader = new TextLoader("src/document_loaders/example_data/example.txt"); const docs = await loader.load(); c...
148904
import { Client, ClientOptions } from "@elastic/elasticsearch"; import { OpenAI, OpenAIEmbeddings } from "@langchain/openai"; import { VectorDBQAChain } from "langchain/chains"; import { ElasticClientArgs, ElasticVectorSearch, } from "@langchain/community/vectorstores/elasticsearch"; import { Document } from "@lan...
148918
# 🦜️🔗 LangChain ⚡ Build context-aware reasoning applications ⚡ [![Release Notes](https://img.shields.io/github/release/langchain-ai/langchain?style=flat-square)](https://github.com/langchain-ai/langchain/releases) [![CI](https://github.com/langchain-ai/langchain/actions/workflows/check_diffs.yml/badge.svg)](https:/...
148924
FROM python:3.11 RUN pip install langchain
148926
{ "cells": [ { "cell_type": "markdown", "id": "70b333e6", "metadata": {}, "source": [ "[![View Article](https://img.shields.io/badge/View%20Article-blue)](https://www.mongodb.com/developer/products/atlas/advanced-rag-langchain-mongodb/)\n" ] }, { "cell_type": "markdown", "id": "d84a72ea"...
148927
" <td>{'id': 4465, 'rating': 7.6, 'votes': 744}</td>\n", " <td>[USA]</td>\n", " <td>None</td>\n", " <td>Young Pauline is left a lot of money when her ...</td>\n", " <td>The Perils of Pauline</td>\n", " <td>[English]</td>\n", " <td>NaN</td>\n",...
148950
# SQL Database Chain This example demonstrates the use of the `SQLDatabaseChain` for answering questions over a SQL database. Under the hood, LangChain uses SQLAlchemy to connect to SQL databases. The `SQLDatabaseChain` can therefore be used with any SQL dialect supported by SQLAlchemy, such as MS SQL, MySQL, MariaDB...
148955
</CodeOutputBlock> ## Adding Memory How to add memory to a SQLDatabaseChain: ```python from langchain_openai import OpenAI from langchain_community.utilities import SQLDatabase from langchain_experimental.sql import SQLDatabaseChain ``` Set up the SQLDatabase and LLM ```python db = SQLDatabase.from_uri("sqlite:///...
148959
'Examples of tracks by Bach include "American Woman", "Concerto for 2 Violins in D Minor, BWV 1043: I. Vivace", "Aria Mit 30 Veränderungen, BWV 988 \'Goldberg Variations\': Aria", "Suite for Solo Cello No. 1 in G Major, BWV 1007: I. Prélude", and "Toccata and Fugue in D Minor, BWV 565: I. Toccata".' ``` </CodeOutputBl...
148961
Requirement already satisfied: pyyaml in /workspace/langchain/.venv/lib/python3.9/site-packages (6.0) Requirement already satisfied: chromadb in /workspace/langchain/.venv/lib/python3.9/site-packages (0.3.21) Requirement already satisfied: pandas>=1.3 in /workspace/langchain/.venv/lib/python3.9/site-packages (f...
148962
Requirement already satisfied: packaging>=20.9 in /workspace/langchain/.venv/lib/python3.9/site-packages (from huggingface-hub>=0.4.0->sentence-transformers>=2.2.2->chromadb) (23.1) Requirement already satisfied: anyio<5,>=3.4.0 in /workspace/langchain/.venv/lib/python3.9/site-packages (from starlette<0.27.0,>=0.26...
148965
```python YAML_EXAMPLES = """ - input: How many customers are not from Brazil? table_info: | CREATE TABLE "Customer" ( "CustomerId" INTEGER NOT NULL, "FirstName" NVARCHAR(40) NOT NULL, "LastName" NVARCHAR(20) NOT NULL, "Company" NVARCHAR(80), "Address" NVARCHAR(70), "City" NVAR...
148973
{ "cells": [ { "cell_type": "markdown", "id": "71a43144", "metadata": {}, "source": [ "# Structure answers with OpenAI functions\n", "\n", "OpenAI functions allows for structuring of response output. This is often useful in question answering when you want to not only get the final answer but...
148975
" HumanMessagePromptTemplate.from_template(\"{context}\"),\n", " HumanMessagePromptTemplate.from_template(\"Question: {question}\"),\n", " HumanMessage(\n", " content=\"Tips: Make sure to answer in the correct format. Return all of the countries mentioned in the sources in uppercase characte...
149010
{ "cells": [ { "cell_type": "code", "execution_count": 2, "id": "c48812ed-35bd-4fbe-9a2c-6c7335e5645e", "metadata": {}, "outputs": [], "source": [ "from langchain_anthropic import ChatAnthropic\n", "from langchain_core.runnables import ConfigurableField\n", "from langchain_core.tools im...
149021
" return_values={\"output\": llm_output.split(\"Final Answer:\")[-1].strip()},\n", " log=llm_output,\n", " )\n", " # Parse out the action and action input\n", " regex = r\"Action\\s*\\d*\\s*:(.*?)\\nAction\\s*\\d*\\s*Input\\s*\\d*\\s*:[\\s]*(.*)\"\n...
149027
{ "cells": [ { "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ "# Human input chat model\n", "\n", "Along with HumanInputLLM, LangChain also provides a pseudo chat model class that can be used for testing, debugging, or educational purposes. This allows you to mock out...
149040
{ "cells": [ { "cell_type": "markdown", "id": "10f50955-be55-422f-8c62-3a32f8cf02ed", "metadata": {}, "source": [ "# RAG application running locally on Intel Xeon CPU using langchain and open-source models" ] }, { "cell_type": "markdown", "id": "48113be6-44bb-4aac-aed3-76a1365b9561", ...
149041
"**For that, importing Chroma vector database from langchain. Also, importing open source GPT4All for embedding models**" ] }, { "cell_type": "code", "execution_count": 8, "id": "9ff99dd7-9d47-4239-ba0a-d775792334ba", "metadata": {}, "outputs": [], "source": [ "from langchain_chroma import ...
149044
] }, { "cell_type": "code", "execution_count": 22, "id": "4654e5b7-635f-4767-8b31-4c430164cdd5", "metadata": {}, "outputs": [], "source": [ "retriever = vectorstore.as_retriever()\n", "qa_chain = (\n", " {\"context\": retriever | format_docs, \"question\": RunnablePassthrough()}\n",...
149079
"Created a chunk of size 1405, which is longer than the specified 1000\n", "Created a chunk of size 2221, which is longer than the specified 1000\n", "Created a chunk of size 1128, which is longer than the specified 1000\n", "Created a chunk of size 1021, which is longer than the specified 1000\n", ...
149080
"### turn on below for custom filtering\n", "# retriever.search_kwargs['filter'] = filter" ] }, { "cell_type": "code", "execution_count": 20, "metadata": { "tags": [] }, "outputs": [], "source": [ "from langchain.chains import ConversationalRetrievalChain\n", "from langchain_ope...
149085
"from langchain_core.pydantic_v1 import BaseModel, Field\n", "from langchain_core.runnables import RunnablePassthrough\n", "from langchain_core.utils.function_calling import convert_to_openai_tool\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings\n", "from langgraph.prebuilt import ToolInvo...
149087
"workflow.add_node(\"generate\", generate) # generatae\n", "workflow.add_node(\"transform_query\", transform_query) # transform_query\n", "workflow.add_node(\"prepare_for_final_grade\", prepare_for_final_grade) # passthrough\n", "\n", "# Build graph\n", "workflow.set_entry_point(\"retrieve\")\n",...
149088
{ "cells": [ { "cell_type": "markdown", "id": "68b24990", "metadata": {}, "source": [ "# Combine agents and vector stores\n", "\n", "This notebook covers how to combine agents and vector stores. The use case for this is that you've ingested your data into a vector store and want to interact w...
149095
{ "cells": [ { "cell_type": "markdown", "metadata": {}, "source": [ "# Human input LLM\n", "\n", "Similar to the fake LLM, LangChain provides a pseudo LLM class that can be used for testing, debugging, or educational purposes. This allows you to mock out calls to the LLM and simulate how a human...
149097
{ "cells": [ { "cell_type": "markdown", "id": "517a9fd4", "metadata": {}, "source": [ "# BabyAGI with Tools\n", "\n", "This notebook builds on top of [baby agi](baby_agi.html), but shows how you can swap out the execution chain. The previous execution chain was just an LLM which made stuff up...
149110
"File \u001b[0;32m~/langchain/langchain/chains/base.py:140\u001b[0m, in \u001b[0;36mChain.__call__\u001b[0;34m(self, inputs, return_only_outputs, callbacks)\u001b[0m\n\u001b[1;32m 138\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m (\u001b[38;5;167;01mKeyboardInterrupt\u001b[39;00m, \u001b[38;5;167;01mException\u001b...
149111
"File \u001b[0;32m~/langchain/langchain/agents/agent.py:953\u001b[0m, in \u001b[0;36mAgentExecutor._call\u001b[0;34m(self, inputs, run_manager)\u001b[0m\n\u001b[1;32m 951\u001b[0m \u001b[38;5;66;03m# We now enter the agent loop (until it returns something).\u001b[39;00m\n\u001b[1;32m 952\u001b[0m \u001b[38;5;28;0...
149121
"metadata": {}, "outputs": [], "source": [ "from langchain.agents import Tool\n", "from langchain_experimental.utilities import PythonREPL\n", "\n", "python_repl = PythonREPL()\n", "\n", "repl_tool = Tool(\n", " name=\"python_repl\",\n", " description=\"A Python shell. Use th...
149122
{ "cells": [ { "cell_type": "markdown", "id": "ba5f8741", "metadata": {}, "source": [ "# Custom agent with tool retrieval\n", "\n", "The novel idea introduced in this notebook is the idea of using retrieval to select the set of tools to use to answer an agent query. This is useful when you ha...
149184
"1 None None \n", "2 None None \n", "3 None None \n", "4 Illinois J. Math. 52 (2008) no.2, 681-689 ...
149185
"compressor = LLMLinguaCompressor(model_name=\"openai-community/gpt2\", device_map=\"cpu\")\n", "compression_retriever = ContextualCompressionRetriever(\n", " base_compressor=compressor, base_retriever=retriever\n", ")" ] }, { "cell_type": "markdown", "metadata": { "id": "Sm5QZdshwJLN" ...
149203
{ "cells": [ { "cell_type": "markdown", "id": "707d13a7", "metadata": {}, "source": [ "# Databricks\n", "\n", "This notebook covers how to connect to the [Databricks runtimes](https://docs.databricks.com/runtime/index.html) and [Databricks SQL](https://www.databricks.com/product/databricks-sq...
149208
{ "cells": [ { "attachments": { "semantic-chunking-rag.png": { "" } }, "cell_type": "markdown", "metadata": {}, "source": [ "# Retrieval Augmented Generation (RAG)\n", "\n", "This notebook demonstrates an example of using [LangChain](https://www.langchain.com/) to delvelop a R...
149209
"cell_type": "markdown", "metadata": {}, "source": [ "## Document Q&A" ] }, { "cell_type": "code", "execution_count": null, "metadata": {}, "outputs": [], "source": [ "# Ask a question about the document\n", "\n", "rag_chain.invoke(\"<your question>\")" ] }, { "cell...
149232
"# RAG pipeline\n", "chain = (\n", " {\n", " \"text_context\": text_retriever | RunnableLambda(split_image_text_types),\n", " \"image_context\": image_retriever | RunnableLambda(split_image_text_types),\n", " \"question\": RunnablePassthrough(),\n", " }\n", " | ...
149244
" f\"Retry. You are required to fix the parsing errors: {error} \\n\\n You must invoke the provided tool.\",\n", " )\n", " ]\n", " return {\n", " \"messages\": messages,\n", " \"context\": inputs[\"context\"],\n", " }\n", "\n", "\n", "# This w...
149249
"from langchain_openai import ChatOpenAI\n", "\n", "# Prompt\n", "template = \"\"\"Answer the question based only on the following context:\n", "{context}\n", "\n", "Question: {question}\n", "\"\"\"\n", "prompt = ChatPromptTemplate.from_template(template)\n", "\n", "# LLM API\n",...
149258
{ "cells": [ { "attachments": {}, "cell_type": "markdown", "id": "fc935871-7640-41c6-b798-58514d860fe0", "metadata": {}, "source": [ "## LLaMA2 chat with SQL\n", "\n", "Open source, local LLMs are great to consider for any application that demands data privacy.\n", "\n", "SQL is on...
149259
"SQL Query: {query}\n", "SQL Response: {response}\"\"\"\n", "prompt_response = ChatPromptTemplate.from_messages(\n", " [\n", " (\n", " \"system\",\n", " \"Given an input question and SQL response, convert it to a natural language answer. No pre-amble.\",\n", "...
149260
{ "cells": [ { "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ "# SalesGPT - Context-Aware AI Sales Assistant With Knowledge Base and Ability Generate Stripe Payment Links\n", "\n", "This notebook demonstrates an implementation of a **Context-Aware** AI Sales agent wit...
149276
"def _parse(text):\n", " return text.strip('\"').strip(\"**\")" ] }, { "cell_type": "code", "execution_count": 13, "id": "c9c34bef", "metadata": {}, "outputs": [], "source": [ "rewriter = rewrite_prompt | ChatOpenAI(temperature=0) | StrOutputParser() | _parse" ] }, { "cell_...
149300
"PREFIX wd: <http://www.wikidata.org/entity/>\n", "PREFIX wdt: <http://www.wikidata.org/prop/direct/>\n", "PREFIX p: <http://www.wikidata.org/prop/>\n", "PREFIX ps: <http://www.wikidata.org/prop/statement/>\n", "\n", "When generating sparql:\n", "* Try to avoid \"count\" and \"filter\" queries i...
149314
"### Initialize the Chroma Client, persist_directory is optinal if you want to save the VectorDB to disk and reload it using same code and path" ] }, { "cell_type": "code", "execution_count": 9, "id": "35953afc-fb35-4dc9-842c-b756b80f4ec4", "metadata": {}, "outputs": [], "source": [ "collec...
149321
{ "cells": [ { "cell_type": "markdown", "id": "fa6802ac", "metadata": {}, "source": [ "# Shared memory across agents and tools\n", "\n", "This notebook goes over adding memory to **both** an Agent and its tools. Before going through this notebook, please walk through the following notebooks, ...
149333
"prompt = hub.pull(\"hwchase17/react\")\n", "agent = create_react_agent(model, tools, prompt)\n", "agent_executor = AgentExecutor(agent=agent, tools=tools, memory=memory)" ] }, { "cell_type": "code", "execution_count": 11, "id": "970d23df", "metadata": {}, "outputs": [ { "name": "...
149604
# Conceptual guide import ThemedImage from '@theme/ThemedImage'; import useBaseUrl from '@docusaurus/useBaseUrl'; This section contains introductions to key parts of LangChain. ## Architecture LangChain as a framework consists of a number of packages. ### `langchain-core` This package contains base abstractions of...
149606
### Prompt templates <span data-heading-keywords="prompt,prompttemplate,chatprompttemplate"></span> Prompt templates help to translate user input and parameters into instructions for a language model. This can be used to guide a model's response, helping it understand the context and generate relevant and coherent lan...
149607
| Name | Supports Streaming | Has Format Instructions | Calls LLM | Input Type | Output Type | Description ...
149608
The base Embeddings class in LangChain provides two methods: one for embedding documents and one for embedding a query. The former takes as input multiple texts, while the latter takes a single text. The reason for having these as two separate methods is that some embedding providers have different embedding methods fo...
149609
There is a legacy `agent` concept in LangChain that we are moving towards deprecating: `AgentExecutor`. AgentExecutor was essentially a runtime for agents. It was a great place to get started, however, it was not flexible enough as you started to have more customized agents. In order to solve that we built LangGraph to...
149610
There are ways to do this [using callbacks](/docs/concepts/#callbacks-1), or by constructing your chain in such a way that it passes intermediate values to the end with something like chained [`.assign()`](/docs/how_to/passthrough/) calls, but LangChain also includes an `.astream_events()` method that combines the flex...
149611
class Joke(BaseModel): """Joke to tell user.""" setup: str = Field(description="The setup of the joke") punchline: str = Field(description="The punchline to the joke") rating: Optional[int] = Field(description="How funny the joke is, from 1 to 10") structured_llm = llm.with_structured_output(Joke) st...