code
stringlengths
141
79.4k
apis
listlengths
1
23
extract_api
stringlengths
126
73.2k
from __future__ import annotations from typing import TYPE_CHECKING, ClassVar, Collection, Dict, List from venv import logger import requests from langchain.schema.document import Document from langchain.vectorstores.base import VectorStore, VectorStoreRetriever from langchain_core.pydantic_v1 import Field, root_vali...
[ "langchain.schema.document.Document", "langchain_core.pydantic_v1.Field", "langchain_core.pydantic_v1.root_validator" ]
[((631, 658), 'langchain_core.pydantic_v1.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (636, 658), False, 'from langchain_core.pydantic_v1 import Field, root_validator\n'), ((942, 958), 'langchain_core.pydantic_v1.root_validator', 'root_validator', ([], {}), '()\n', (956, 958), False,...
import os from typing import Optional from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from langchain.schema import BaseMessage, HumanMessage from rebyte_langchain.rebyte_langchain import RebyteEndpoint from realtime_ai_character.llm.base import ( AsyncCallbackAudioHandler, Asyn...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.schema.HumanMessage" ]
[((473, 493), 'realtime_ai_character.logger.get_logger', 'get_logger', (['__name__'], {}), '(__name__)\n', (483, 493), False, 'from realtime_ai_character.logger import get_logger\n'), ((572, 603), 'os.getenv', 'os.getenv', (['"""REBYTE_API_KEY"""', '""""""'], {}), "('REBYTE_API_KEY', '')\n", (581, 603), False, 'import ...
from langchain.agents import AgentType, initialize_agent from langchain_community.agent_toolkits import ZapierToolkit from langchain_community.tools import BaseTool from langchain_community.utilities.zapier import ZapierNLAWrapper from langchain_openai import ChatOpenAI class ZapierNLA(BaseTool): name = "Zapier" ...
[ "langchain_openai.ChatOpenAI", "langchain_community.agent_toolkits.ZapierToolkit.from_zapier_nla_wrapper", "langchain_community.utilities.zapier.ZapierNLAWrapper" ]
[((577, 632), 'langchain_community.utilities.zapier.ZapierNLAWrapper', 'ZapierNLAWrapper', ([], {'zapier_nla_api_key': 'zapier_nla_api_key'}), '(zapier_nla_api_key=zapier_nla_api_key)\n', (593, 632), False, 'from langchain_community.utilities.zapier import ZapierNLAWrapper\n'), ((651, 696), 'langchain_community.agent_t...
from celery import shared_task from langchain.text_splitter import RecursiveCharacterTextSplitter from shared.models.opencopilot_db.pdf_data_sources import ( insert_pdf_data_source, update_pdf_data_source_status, ) from langchain.document_loaders import UnstructuredMarkdownLoader from shared.utils.opencopilot_...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((1830, 1925), 'shared.models.opencopilot_db.pdf_data_sources.update_pdf_data_source_status', 'update_pdf_data_source_status', ([], {'chatbot_id': 'chatbot_id', 'file_name': 'file_name', 'status': '"""PENDING"""'}), "(chatbot_id=chatbot_id, file_name=file_name,\n status='PENDING')\n", (1859, 1925), False, 'from sha...
from concurrent.futures import ThreadPoolExecutor import asyncio import logging from typing import Optional, Tuple from langchain import ConversationChain from vocode.streaming.agent.base_agent import RespondAgent from vocode.streaming.models.agent import ChatVertexAIAgentConfig from langchain_community.chat_models imp...
[ "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.memory.ConversationBufferMemory", "langchain.prompts.MessagesPlaceholder", "langchain.schema.SystemMessage", "langchain_community.chat_models.ChatVertexAI", "langchain.ConversationChain" ]
[((1089, 1103), 'langchain_community.chat_models.ChatVertexAI', 'ChatVertexAI', ([], {}), '()\n', (1101, 1103), False, 'from langchain_community.chat_models import ChatVertexAI\n'), ((1127, 1173), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'return_messages': '(True)'}), '(return_mess...
# SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved. # SPDX-License-Identifier: Apache-2.0 # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # ht...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.document_loaders.DirectoryLoader", "langchain_core.output_parsers.StrOutputParser", "langchain.vectorstores.FAISS.from_documents", "langchain_core.prompts.ChatPromptTemplate.from_messages", "langchain_nvidia_ai_endpoints.NVIDIAEmbeddings", "lan...
[((1034, 1067), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (1052, 1067), True, 'import streamlit as st\n'), ((2031, 2063), 'langchain_nvidia_ai_endpoints.ChatNVIDIA', 'ChatNVIDIA', ([], {'model': '"""mixtral_8x7b"""'}), "(model='mixtral_8x7b')\n", (2041, 2063...
from langchain.chains import RetrievalQA, ConversationalRetrievalChain, ConversationChain from langchain.prompts.prompt import PromptTemplate from langchain.vectorstores.base import VectorStoreRetriever from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationBufferMemory import pickle impo...
[ "langchain.chains.ConversationChain", "langchain.prompts.prompt.PromptTemplate", "langchain.vectorstores.base.VectorStoreRetriever", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.memory.ConversationBufferMemory", "langchain.chat_models.ChatOpenAI", "langchain.prompts.prompt.Prompt...
[((727, 766), 'langchain.prompts.prompt.PromptTemplate.from_template', 'PromptTemplate.from_template', (['_template'], {}), '(_template)\n', (755, 766), False, 'from langchain.prompts.prompt import PromptTemplate\n'), ((1521, 1595), 'langchain.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'template': 'templat...
# flake8: noqa from langchain.prompts import PromptTemplate ## Use a shorter template to reduce the number of tokens in the prompt template = """Create a final answer to the given questions using the provided document excerpts (given in no particular order) as sources. ALWAYS include a "SOURCES" section in your answer...
[ "langchain.prompts.PromptTemplate" ]
[((2121, 2197), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'template': 'template', 'input_variables': "['summaries', 'question']"}), "(template=template, input_variables=['summaries', 'question'])\n", (2135, 2197), False, 'from langchain.prompts import PromptTemplate\n')]
import whisper import textwrap from langchain.chat_models import ChatOpenAI from langchain.chains.summarize import load_summarize_chain from langchain.schema import Document from langchain.llms import OpenAI from langchain import PromptTemplate from langchain.chains import LLMChain from langchain.document_loaders impo...
[ "langchain.chains.summarize.load_summarize_chain", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.schema.Document", "langchain.document_loaders.WebBaseLoader", "langchain.chains.LLMChain", "langchain.PromptTemplate" ]
[((508, 534), 'whisper.load_model', 'whisper.load_model', (['"""base"""'], {}), "('base')\n", (526, 534), False, 'import whisper\n'), ((607, 653), 'textwrap.fill', 'textwrap.fill', (["transcription['text']"], {'width': '(50)'}), "(transcription['text'], width=50)\n", (620, 653), False, 'import textwrap\n'), ((709, 766)...
from langchain.agents import load_tools from langchain.tools import AIPluginTool from parse import * from langchain.chat_models.base import BaseChatModel from langchain.chat_models import ChatOpenAI, AzureChatOpenAI import utils def create_plugins_static(): plugins = [ AIPluginTool.from_plugin_url( ...
[ "langchain.tools.AIPluginTool.from_plugin_url", "langchain.chat_models.AzureChatOpenAI", "langchain.agents.load_tools", "langchain.chat_models.ChatOpenAI" ]
[((410, 438), 'langchain.agents.load_tools', 'load_tools', (["['requests_all']"], {}), "(['requests_all'])\n", (420, 438), False, 'from langchain.agents import load_tools\n'), ((285, 371), 'langchain.tools.AIPluginTool.from_plugin_url', 'AIPluginTool.from_plugin_url', (['"""https://www.klarna.com/.well-known/ai-plugin....
import re import string from collections import Counter import numpy as np import pandas as pd import tqdm from langchain.evaluation.qa import QAEvalChain from langchain.llms import OpenAI from algos.PWS import PWS_Base, PWS_Extra from algos.notool import CoT, IO from algos.react import ReactBase def normalize_answ...
[ "langchain.llms.OpenAI" ]
[((373, 410), 're.sub', 're.sub', (['"""\\\\b(a|an|the)\\\\b"""', '""" """', 'text'], {}), "('\\\\b(a|an|the)\\\\b', ' ', text)\n", (379, 410), False, 'import re\n'), ((1278, 1304), 'collections.Counter', 'Counter', (['prediction_tokens'], {}), '(prediction_tokens)\n', (1285, 1304), False, 'from collections import Coun...
from datetime import date, datetime from decimal import Decimal from langchain.chains import LLMChain from langchain.prompts.chat import ( ChatPromptTemplate, HumanMessagePromptTemplate, ) from sqlalchemy import text from dataherald.model.chat_model import ChatModel from dataherald.repositories.database_conne...
[ "langchain.prompts.chat.HumanMessagePromptTemplate.from_template", "langchain.chains.LLMChain", "langchain.prompts.chat.ChatPromptTemplate.from_messages" ]
[((1101, 1123), 'dataherald.model.chat_model.ChatModel', 'ChatModel', (['self.system'], {}), '(self.system)\n', (1110, 1123), False, 'from dataherald.model.chat_model import ChatModel\n'), ((1272, 1302), 'dataherald.repositories.prompts.PromptRepository', 'PromptRepository', (['self.storage'], {}), '(self.storage)\n', ...
import streamlit as st import urllib import os import re import time import random from operator import itemgetter from collections import OrderedDict from langchain_core.documents import Document from langchain_openai import AzureChatOpenAI from langchain_core.output_parsers import StrOutputParser from utils import g...
[ "langchain_openai.AzureChatOpenAI", "langchain_core.documents.Document", "langchain_core.output_parsers.StrOutputParser" ]
[((376, 455), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""GPT Smart Search"""', 'page_icon': '"""📖"""', 'layout': '"""wide"""'}), "(page_title='GPT Smart Search', page_icon='📖', layout='wide')\n", (394, 455), True, 'import streamlit as st\n'), ((498, 726), 'streamlit.markdown', 'st.mark...
from __future__ import annotations import asyncio import logging import typing as t from abc import ABC, abstractmethod from dataclasses import dataclass from functools import partial from langchain_community.chat_models import ChatVertexAI from langchain_community.llms import VertexAI from langchain_core.language_mo...
[ "langchain_openai.chat_models.ChatOpenAI" ]
[((765, 792), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (782, 792), False, 'import logging\n'), ((6829, 6869), 'langchain_openai.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model': 'model', 'timeout': 'timeout'}), '(model=model, timeout=timeout)\n', (6839, 6869), False, 'from langc...
"""Utility functions and constants. I am having some problems caching the memory and the retrieval. When I decorate for caching, I get streamlit init errors. """ import logging import pathlib from typing import Any from langchain.document_loaders import ( PyPDFLoader, TextLoader, UnstructuredEPubLoader, ...
[ "langchain.memory.ConversationBufferMemory" ]
[((637, 735), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'return_messages': '(True)', 'output_key': '"""answer"""'}), "(memory_key='chat_history', return_messages=True,\n output_key='answer')\n", (661, 735), False, 'from langchain.memory import ...
import os import re import urllib import urllib.parse import urllib.request from typing import Any, List, Tuple, Union from urllib.parse import urlparse import requests from bs4 import BeautifulSoup from langchain.chains import LLMChain from langchain.prompts import Prompt from langchain.tools import BaseTool from lan...
[ "langchain.utilities.GoogleSerperAPIWrapper" ]
[((2438, 2451), 'sherpa_ai.config.task_config.AgentConfig', 'AgentConfig', ([], {}), '()\n', (2449, 2451), False, 'from sherpa_ai.config.task_config import AgentConfig\n'), ((894, 986), 'loguru.logger.warning', 'logger.warning', (['"""No SERPER_API_KEY found in environment variables, skipping SearchTool"""'], {}), "(\n...
from dotenv import load_dotenv from langchain_core.prompts import PromptTemplate load_dotenv() from langchain import hub from langchain.agents import create_react_agent, AgentExecutor from langchain_core.tools import Tool from langchain_openai import ChatOpenAI from tools.tools import get_profile_url def lookup(nam...
[ "langchain_openai.ChatOpenAI", "langchain.agents.AgentExecutor", "langchain.agents.create_react_agent", "langchain_core.tools.Tool", "langchain_core.prompts.PromptTemplate", "langchain.hub.pull" ]
[((82, 95), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (93, 95), False, 'from dotenv import load_dotenv\n'), ((346, 399), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, model_name='gpt-3.5-turbo')\n", (356, 399), False, 'from ...
import logging, json, os from Utilities.envVars import * from Utilities.envVars import * # Import required libraries from Utilities.cogSearchVsRetriever import CognitiveSearchVsRetriever from langchain.chains import RetrievalQA from langchain import PromptTemplate from Utilities.evaluator import indexDocs import json i...
[ "langchain.evaluation.qa.QAEvalChain.from_llm", "langchain.chains.RetrievalQA.from_chain_type", "langchain.chat_models.ChatOpenAI", "langchain.document_loaders.PDFMinerLoader", "langchain.chat_models.AzureChatOpenAI", "langchain.PromptTemplate" ]
[((911, 1164), 'collections.namedtuple', 'namedtuple', (['"""RunDoc"""', "['evalatorQaData', 'totalQuestions', 'promptStyle', 'documentId',\n 'splitMethods', 'chunkSizes', 'overlaps', 'retrieverType', 'reEvaluate',\n 'topK', 'model', 'fileName', 'embeddingModelType', 'temperature',\n 'tokenLength']"], {}), "('...
import dataclasses import typing from dataclasses import dataclass from typing import Tuple, cast from langchain.chat_models.base import BaseChatModel from langchain.output_parsers import PydanticOutputParser from langchain.schema import BaseMessage, HumanMessage from pydantic import BaseModel, create_model T = typin...
[ "langchain.output_parsers.PydanticOutputParser", "langchain.schema.HumanMessage" ]
[((315, 334), 'typing.TypeVar', 'typing.TypeVar', (['"""T"""'], {}), "('T')\n", (329, 334), False, 'import typing\n'), ((339, 375), 'typing.TypeVar', 'typing.TypeVar', (['"""B"""'], {'bound': 'BaseModel'}), "('B', bound=BaseModel)\n", (353, 375), False, 'import typing\n'), ((568, 604), 'typing.cast', 'cast', (['str', '...
from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.chat_models import ChatOpenAI from virl.config import cfg from virl.utils.common_utils import print_prompt, print_answer, parse_answer_to_json from .gpt_chat import GPTChat from .azure_gpt import AzureGPTChat __all__...
[ "langchain.agents.initialize_agent", "langchain.agents.load_tools", "langchain.chat_models.ChatOpenAI" ]
[((679, 701), 'virl.utils.common_utils.print_prompt', 'print_prompt', (['question'], {}), '(question)\n', (691, 701), False, 'from virl.utils.common_utils import print_prompt, print_answer, parse_answer_to_json\n'), ((830, 850), 'virl.utils.common_utils.print_answer', 'print_answer', (['answer'], {}), '(answer)\n', (84...
from langchain_community.chat_models import ChatAnthropic from langchain_core.output_parsers import StrOutputParser from langchain_core.prompts import ChatPromptTemplate from langchain_core.runnables import RunnablePassthrough prompt = ChatPromptTemplate.from_template("Tell me a short joke about {topic}") output_parse...
[ "langchain_core.prompts.ChatPromptTemplate.from_template", "langchain_community.chat_models.ChatAnthropic", "langchain_core.runnables.RunnablePassthrough", "langchain_core.output_parsers.StrOutputParser" ]
[((237, 307), 'langchain_core.prompts.ChatPromptTemplate.from_template', 'ChatPromptTemplate.from_template', (['"""Tell me a short joke about {topic}"""'], {}), "('Tell me a short joke about {topic}')\n", (269, 307), False, 'from langchain_core.prompts import ChatPromptTemplate\n'), ((324, 341), 'langchain_core.output_...
from langchain.chains import LLMChain from langchain.prompts import PromptTemplate from tqdm import tqdm from lmchain.tools import tool_register class GLMToolChain: def __init__(self, llm): self.llm = llm self.tool_register = tool_register self.tools = tool_register.get_tools() def ...
[ "langchain.chains.LLMChain" ]
[((3292, 3320), 'lmchain.agents.llmMultiAgent.AgentZhipuAI', 'llmMultiAgent.AgentZhipuAI', ([], {}), '()\n', (3318, 3320), False, 'from lmchain.agents import llmMultiAgent\n'), ((3381, 3408), 'lmchain.chains.toolchain.GLMToolChain', 'toolchain.GLMToolChain', (['llm'], {}), '(llm)\n', (3403, 3408), False, 'from lmchain....
import json import time import hashlib from typing import Dict, Any, List, Tuple import re from os import environ import streamlit as st from langchain.schema import BaseRetriever from langchain.tools import Tool from langchain.pydantic_v1 import BaseModel, Field from sqlalchemy import Column, Text, create_engine, Me...
[ "langchain.agents.openai_functions_agent.agent_token_buffer_memory.AgentTokenBufferMemory", "langchain.pydantic_v1.Field", "langchain_experimental.retrievers.vector_sql_database.VectorSQLDatabaseChainRetriever", "langchain.utilities.sql_database.SQLDatabase", "langchain.schema.messages.ToolMessage", "lang...
[((3163, 3322), 'langchain.prompts.ChatPromptTemplate.from_strings', 'ChatPromptTemplate.from_strings', ([], {'string_messages': "[(SystemMessagePromptTemplate, combine_prompt_template), (\n HumanMessagePromptTemplate, '{question}')]"}), "(string_messages=[(\n SystemMessagePromptTemplate, combine_prompt_template)...
"""Simple agents which can be used as a starting point for running the deception environment with Umshini (see tutorials).""" # pyright: reportGeneralTypeIssues=false from langchain.chat_models import ChatOpenAI from langchain.schema import AIMessage, HumanMessage, SystemMessage class DeceptionDefender: def __ini...
[ "langchain.schema.AIMessage", "langchain.schema.SystemMessage", "langchain.schema.HumanMessage", "langchain.chat_models.ChatOpenAI" ]
[((433, 460), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0.9)'}), '(temperature=0.9)\n', (443, 460), False, 'from langchain.chat_models import ChatOpenAI\n'), ((738, 765), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0.9)'}), '(temperature=0.9)\n', (748, 765), F...
"""Wrapper around Replicate API.""" import logging from typing import Any, Dict, List, Mapping, Optional from pydantic import Extra, Field, root_validator from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.utils import get_from_dict_or_env logger = log...
[ "langchain.utils.get_from_dict_or_env" ]
[((317, 344), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (334, 344), False, 'import logging\n'), ((1212, 1239), 'pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1217, 1239), False, 'from pydantic import Extra, Field, root_validator\n'), ((1275...
import databutton as db import re from io import BytesIO from typing import Tuple, List import pickle from langchain.docstore.document import Document from langchain.embeddings.openai import OpenAIEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.vectorstores.faiss import FAI...
[ "langchain.embeddings.openai.OpenAIEmbeddings", "langchain.docstore.document.Document", "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((446, 461), 'pypdf.PdfReader', 'PdfReader', (['file'], {}), '(file)\n', (455, 461), False, 'from pypdf import PdfReader\n'), ((555, 597), 're.sub', 're.sub', (['"""(\\\\w+)-\\\\n(\\\\w+)"""', '"""\\\\1\\\\2"""', 'text'], {}), "('(\\\\w+)-\\\\n(\\\\w+)', '\\\\1\\\\2', text)\n", (561, 597), False, 'import re\n'), ((675...
import datetime import difflib import logging import os from functools import wraps from queue import Queue from threading import Thread from typing import Any, Callable, Dict, List import numpy as np import openai import pandas as pd import sqlalchemy from google.api_core.exceptions import GoogleAPIError from langcha...
[ "langchain.agents.mrkl.base.ZeroShotAgent.create_prompt", "langchain.agents.mrkl.base.ZeroShotAgent", "langchain.agents.agent.AgentExecutor.from_agent_and_tools", "langchain.chains.llm.LLMChain", "langchain_community.callbacks.get_openai_callback" ]
[((2000, 2027), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (2017, 2027), False, 'import logging\n'), ((2038, 2074), 'dataherald.sql_generator.SQLGenerator.get_upper_bound_limit', 'SQLGenerator.get_upper_bound_limit', ([], {}), '()\n', (2072, 2074), False, 'from dataherald.sql_generato...
from marqo import Client import pandas as pd import numpy as np from langchain_openai import OpenAI from langchain.docstore.document import Document from langchain.chains import LLMChain from dotenv import load_dotenv from utilities import ( load_data, extract_text_from_highlights, qna_prompt, predic...
[ "langchain_openai.OpenAI", "langchain.docstore.document.Document" ]
[((349, 362), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (360, 362), False, 'from dotenv import load_dotenv\n'), ((984, 992), 'marqo.Client', 'Client', ([], {}), '()\n', (990, 992), False, 'from marqo import Client\n'), ((1812, 1823), 'utilities.load_data', 'load_data', ([], {}), '()\n', (1821, 1823), False...
from typing import List, Optional from langchain.schema.language_model import BaseLanguageModel from server.knowledge_base.model.kb_document_model import DocumentWithVSId from configs import (logger) from langchain.chains import StuffDocumentsChain, LLMChain from langchain.prompts import PromptTemplate from langchai...
[ "langchain.chains.combine_documents.map_reduce.MapReduceDocumentsChain", "langchain.docstore.document.Document", "langchain.chains.LLMChain", "langchain.prompts.PromptTemplate.from_template", "langchain.chains.StuffDocumentsChain", "langchain.prompts.PromptTemplate", "langchain.chains.combine_documents....
[((1461, 1536), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['page_content']", 'template': '"""{page_content}"""'}), "(input_variables=['page_content'], template='{page_content}')\n", (1475, 1536), False, 'from langchain.prompts import PromptTemplate\n'), ((1853, 1943), 'langchain.pr...
from fastapi import FastAPI, Form, Request, Response, File, Depends, HTTPException, status from fastapi.responses import RedirectResponse from fastapi.staticfiles import StaticFiles from fastapi.templating import Jinja2Templates from fastapi.encoders import jsonable_encoder from langchain.llms import CTransformers...
[ "langchain.chains.summarize.load_summarize_chain", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.embeddings.HuggingFaceBgeEmbeddings", "langchain.docstore.document.Document", "langchain.llms.CTransformers", "langchain.vectorstores.FAISS.from_documents", "langchain.document_loaders...
[((911, 920), 'fastapi.FastAPI', 'FastAPI', ([], {}), '()\n', (918, 920), False, 'from fastapi import FastAPI, Form, Request, Response, File, Depends, HTTPException, status\n'), ((1008, 1046), 'fastapi.templating.Jinja2Templates', 'Jinja2Templates', ([], {'directory': '"""templates"""'}), "(directory='templates')\n", (...
#!/usr/bin/env python """Example LangChain server exposes a retriever.""" from fastapi import FastAPI from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores import FAISS from langserve import add_routes vectorstore = FAISS.from_texts( ["cats like fish", "dogs like sticks"], embedding=OpenAI...
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((381, 515), 'fastapi.FastAPI', 'FastAPI', ([], {'title': '"""LangChain Server"""', 'version': '"""1.0"""', 'description': '"""Spin up a simple api server using Langchain\'s Runnable interfaces"""'}), '(title=\'LangChain Server\', version=\'1.0\', description=\n "Spin up a simple api server using Langchain\'s Runna...
""" Chatbot for talking to Podcast using Langchain, Ollama and LanceDB """ from langchain.document_loaders import WikipediaLoader import pandas as pd from langchain.memory import ConversationSummaryMemory import lancedb from langchain.vectorstores import LanceDB from langchain.embeddings import OpenAIEmbeddings from l...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.memory.ConversationSummaryMemory", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.document_loaders.WikipediaLoader", "langchain.chat_models.ChatOllama", "langchain.embeddings.OpenAIEmbeddings" ]
[((525, 556), 'lancedb.connect', 'lancedb.connect', (['"""/tmp/lancedb"""'], {}), "('/tmp/lancedb')\n", (540, 556), False, 'import lancedb\n'), ((883, 946), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(500)', 'chunk_overlap': '(0)'}), '(chunk_size=500...
## Conversational Q&A Chatbot import streamlit as st from langchain.schema import HumanMessage,SystemMessage,AIMessage from langchain.chat_models import ChatOpenAI ## Streamlit UI st.set_page_config(page_title="Conversational Q&A Chatbot") st.header("Hey, Let's Chat") from dotenv import load_dotenv load_d...
[ "langchain.schema.AIMessage", "langchain.schema.SystemMessage", "langchain.schema.HumanMessage", "langchain.chat_models.ChatOpenAI" ]
[((189, 248), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Conversational Q&A Chatbot"""'}), "(page_title='Conversational Q&A Chatbot')\n", (207, 248), True, 'import streamlit as st\n'), ((250, 278), 'streamlit.header', 'st.header', (['"""Hey, Let\'s Chat"""'], {}), '("Hey, Let\'s Chat")\n...
# Copyright 2023 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "langchain.chains.LLMChain", "langchain.prompts.PromptTemplate", "langchain.llms.HuggingFacePipeline" ]
[((882, 921), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (901, 921), False, 'import logging\n'), ((1019, 1069), 'os.environ.get', 'os.environ.get', (['"""MODEL_ID"""', '"""google/flan-t5-small"""'], {}), "('MODEL_ID', 'google/flan-t5-small')\n", (1033, 1069)...
"""Prompt schema definition.""" from __future__ import annotations from string import Formatter from typing import Any, Dict, List from pydantic import BaseModel, Extra, root_validator from langchain.prompts.base import ( DEFAULT_FORMATTER_MAPPING, BasePromptTemplate, check_valid_template, ) class Prom...
[ "langchain.prompts.base.check_valid_template" ]
[((1613, 1629), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1627, 1629), False, 'from pydantic import BaseModel, Extra, root_validator\n'), ((1806, 1905), 'langchain.prompts.base.check_valid_template', 'check_valid_template', (["values['template']", "values['template_format']", "values['input_variab...
from langchain.chat_models import ChatOpenAI from langchain.chains.summarize import load_summarize_chain from langchain.text_splitter import TokenTextSplitter from langchain.docstore.document import Document # Function to initialize the large language model. def initialize_llm(openai_api_key, model_name, temperature):...
[ "langchain.chains.summarize.load_summarize_chain", "langchain.docstore.document.Document", "langchain.text_splitter.TokenTextSplitter", "langchain.chat_models.ChatOpenAI" ]
[((331, 424), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': 'openai_api_key', 'model_name': 'model_name', 'temperature': 'temperature'}), '(openai_api_key=openai_api_key, model_name=model_name,\n temperature=temperature)\n', (341, 424), False, 'from langchain.chat_models import ChatOpenAI...
"""Wrapper around Google's PaLM Chat API.""" from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Callable, Dict, List, Mapping, Optional from pydantic import BaseModel, root_validator from tenacity import ( before_sleep_log, retry, retry_if_exception_type, stop_aft...
[ "langchain.schema.ChatMessage", "langchain.utils.get_from_dict_or_env", "langchain.schema.ChatResult", "langchain.schema.HumanMessage", "langchain.schema.AIMessage" ]
[((792, 819), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (809, 819), False, 'import logging\n'), ((2563, 2598), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (2573, 2598), False, 'from langchain.schema import AIMessag...
#!/usr/bin/env python # -*- encoding: utf-8 -*- ''' @File : create_db.py @Time : 2023/12/14 10:56:31 @Author : Logan Zou @Version : 1.0 @Contact : loganzou0421@163.com @License : (C)Copyright 2017-2018, Liugroup-NLPR-CASIA @Desc : 知识库搭建 ''' # 首先导入所需第三方库 from langchain.document_loaders import U...
[ "langchain.document_loaders.UnstructuredFileLoader", "langchain.embeddings.huggingface.HuggingFaceEmbeddings", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.vectorstores.Chroma.from_documents", "langchain.document_loaders.UnstructuredMarkdownLoader" ]
[((2018, 2083), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(500)', 'chunk_overlap': '(150)'}), '(chunk_size=500, chunk_overlap=150)\n', (2048, 2083), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((2164, 2237), 'lang...
from flask import Flask, request from flask_restful import Resource, Api, reqparse, abort from werkzeug.utils import secure_filename ######################################################################## import tempfile import os from langchain.document_loaders import DirectoryLoader, PyMuPDFLoader from langchain.te...
[ "langchain.document_loaders.DirectoryLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.vectorstores.Pinecone.from_documents", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.Pinecone.from_existing_index", "...
[((718, 731), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (729, 731), False, 'from dotenv import load_dotenv\n'), ((753, 785), 'os.environ.get', 'os.environ.get', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (767, 785), False, 'import os\n'), ((809, 843), 'os.environ.get', 'os.environ.get', (['"""...
from rich import print from typing import Any, List from langchain import LLMChain from langchain.agents import Tool from langchain.chat_models import AzureChatOpenAI from LLMAgent.callbackHandler import CustomHandler from langchain.callbacks import get_openai_callback from langchain.memory import ConversationBufferMem...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.LLMChain", "langchain.agents.ZeroShotAgent.create_prompt", "langchain.agents.ZeroShotAgent", "langchain.memory.ConversationBufferMemory", "langchain.callbacks.get_openai_callback", "langchain.agents.Tool" ]
[((6579, 6594), 'LLMAgent.callbackHandler.CustomHandler', 'CustomHandler', ([], {}), '()\n', (6592, 6594), False, 'from LLMAgent.callbackHandler import CustomHandler\n'), ((6905, 7044), 'langchain.agents.ZeroShotAgent.create_prompt', 'ZeroShotAgent.create_prompt', (['tools'], {'prefix': 'customedPrefix', 'suffix': 'suf...
from langchain_app.models.http_llm import HTTPBaseLLM def default_parameters(): return {"temperature": 0, "max_new_tokens": 256, "stop": ["Observation:"]} def build_llama_base_llm(prompt_url="http://127.0.0.1:8000/prompt", parameters=None): if parameters is None: parameters = default_parameters() ...
[ "langchain_app.models.http_llm.HTTPBaseLLM" ]
[((330, 387), 'langchain_app.models.http_llm.HTTPBaseLLM', 'HTTPBaseLLM', ([], {'prompt_url': 'prompt_url', 'parameters': 'parameters'}), '(prompt_url=prompt_url, parameters=parameters)\n', (341, 387), False, 'from langchain_app.models.http_llm import HTTPBaseLLM\n')]
from langchain.llms import LlamaCpp from langchain.embeddings import HuggingFaceEmbeddings from langchain.callbacks.manager import CallbackManager from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler def hf_embeddings(): return HuggingFaceEmbeddings( model_name = "sentence-transf...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.llms.LlamaCpp", "langchain.embeddings.HuggingFaceEmbeddings" ]
[((260, 335), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': '"""sentence-transformers/all-mpnet-base-v2"""'}), "(model_name='sentence-transformers/all-mpnet-base-v2')\n", (281, 335), False, 'from langchain.embeddings import HuggingFaceEmbeddings\n'), ((456, 642), 'langchain....
import os import yaml from types import SimpleNamespace import openai import numpy as np from sklearn.metrics.pairwise import cosine_similarity from langchain.vectorstores import FAISS from langchain.embeddings import HuggingFaceEmbeddings with open("config.yml") as f: config = yaml.safe_load(f) config = SimpleN...
[ "langchain.embeddings.HuggingFaceEmbeddings", "langchain.vectorstores.FAISS.load_local" ]
[((313, 338), 'types.SimpleNamespace', 'SimpleNamespace', ([], {}), '(**config)\n', (328, 338), False, 'from types import SimpleNamespace\n'), ((286, 303), 'yaml.safe_load', 'yaml.safe_load', (['f'], {}), '(f)\n', (300, 303), False, 'import yaml\n'), ((602, 627), 'numpy.argsort', 'np.argsort', (['(-similarities)'], {})...
from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.agents import AgentType from langchain_app.models.vicuna_request_llm import VicunaLLM # First, let's load the language model we're going to use to control the agent. llm = VicunaLLM() # Next, let's load some tools to...
[ "langchain_app.models.vicuna_request_llm.VicunaLLM", "langchain.agents.initialize_agent", "langchain.agents.load_tools" ]
[((275, 286), 'langchain_app.models.vicuna_request_llm.VicunaLLM', 'VicunaLLM', ([], {}), '()\n', (284, 286), False, 'from langchain_app.models.vicuna_request_llm import VicunaLLM\n'), ((405, 441), 'langchain.agents.load_tools', 'load_tools', (["['python_repl']"], {'llm': 'llm'}), "(['python_repl'], llm=llm)\n", (415, ...
import logging import sys from typing import Callable from langchain.prompts import MessagesPlaceholder from langchain.agents import AgentType, AgentExecutor from langchain.agents import initialize_agent as initialize_agent_base from langchain.agents.agent_toolkits.base import BaseToolkit from langchain.chains.base i...
[ "langchain.agents.initialize_agent", "langchain.prompts.MessagesPlaceholder" ]
[((343, 370), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (360, 370), False, 'import logging\n'), ((2107, 2151), 'langchain.agents.initialize_agent', 'initialize_agent_base', ([], {'agent': 'agent'}), '(agent=agent, **kwargs)\n', (2128, 2151), True, 'from langchain.agents import initia...
import os os.environ["LANGCHAIN_TRACING"] = "true" from langchain import OpenAI from langchain.agents import initialize_agent, AgentType from langchain.llms import OpenAI from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType def multiplier(a, b): return a / b def parsing_mu...
[ "langchain.agents.initialize_agent", "langchain.llms.OpenAI", "langchain.agents.Tool" ]
[((412, 433), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (418, 433), False, 'from langchain.llms import OpenAI\n'), ((826, 917), 'langchain.agents.initialize_agent', 'initialize_agent', (['tools', 'llm'], {'agent': 'AgentType.ZERO_SHOT_REACT_DESCRIPTION', 'verbose': '(True)'})...
# Copyright 2023 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, ...
[ "langchain.llms.vertexai.VertexAI", "langchain.memory.ConversationBufferMemory", "langchain.chains.RetrievalQA.from_chain_type" ]
[((957, 985), 'sys.path.append', 'sys.path.append', (['current_dir'], {}), '(current_dir)\n', (972, 985), False, 'import sys\n'), ((1302, 1333), 'os.getenv', 'os.getenv', (['"""GOOGLE_CLOUD_REGIN"""'], {}), "('GOOGLE_CLOUD_REGIN')\n", (1311, 1333), False, 'import os\n'), ((1347, 1380), 'os.getenv', 'os.getenv', (['"""G...
import boto3 from botocore.exceptions import ClientError import json import langchain from importlib import reload from langchain.agents.structured_chat import output_parser from typing import List import logging import os import sqlalchemy from sqlalchemy import create_engine from langchain.docstore.document import Do...
[ "langchain.memory.ConversationBufferMemory", "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.tools.python.tool.PythonREPLTool", "langchain.llms.bedrock.Bedrock", "langchain.prompts.PromptTemplate", "langchain_experimental.plan_and_execute.load_chat_planner", "langchain_experimental.plan...
[((1364, 1397), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (1382, 1397), True, 'import streamlit as st\n'), ((1532, 1553), 'utility.custom_logga.Logger', 'custom_logga.Logger', ([], {}), '()\n', (1551, 1553), False, 'from utility import get_cfn_details, custo...
from langchain.agents.agent_toolkits import create_python_agent from langchain.tools.python.tool import PythonREPLTool from langchain.python import PythonREPL from langchain.llms.openai import OpenAI from langchain.agents.agent_types import AgentType from langchain.chat_models import ChatOpenAI import os agent_execut...
[ "langchain.llms.openai.OpenAI", "langchain.tools.python.tool.PythonREPLTool" ]
[((354, 394), 'langchain.llms.openai.OpenAI', 'OpenAI', ([], {'temperature': '(0.5)', 'max_tokens': '(2000)'}), '(temperature=0.5, max_tokens=2000)\n', (360, 394), False, 'from langchain.llms.openai import OpenAI\n'), ((405, 421), 'langchain.tools.python.tool.PythonREPLTool', 'PythonREPLTool', ([], {}), '()\n', (419, 4...
# docsGpt.py - Contains the docsGpt functions and classes for document parsing # Author: Armin Norouzi, Farhad Davaripour # Contact: https://github.com/Farhad-Davaripour/DocsGPT # Date created: April 14, 2023 # Last modified: May 3, 2023 # License: MIT License # Import required modules import sys import subprocess fr...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.vectorstores.FAISS.from_texts", "langchain.llms.OpenAI", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((1393, 1431), 'getpass.getpass', 'getpass', (['"""Enter your OpenAI token: ()"""'], {}), "('Enter your OpenAI token: ()')\n", (1400, 1431), False, 'from getpass import getpass\n'), ((1523, 1541), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (1539, 1541), False, 'from langchain...
import httpx from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain_community.document_loaders import (AsyncChromiumLoader, AsyncHtmlLoader) from langchain_community.document_transformers import BeautifulSoupTransformer from modelscope_agent.t...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter.from_tiktoken_encoder", "langchain_community.document_loaders.AsyncHtmlLoader", "langchain_community.document_transformers.BeautifulSoupTransformer" ]
[((359, 387), 'modelscope_agent.tools.register_tool', 'register_tool', (['"""web_browser"""'], {}), "('web_browser')\n", (372, 387), False, 'from modelscope_agent.tools import BaseTool, register_tool\n'), ((857, 919), 'httpx.Client', 'httpx.Client', ([], {'headers': 'self.headers', 'verify': '(False)', 'timeout': '(30....
from langchain.chains.base import Chain from langchain.tools import Tool, BaseTool def chain_as_tool(chain: Chain, name: str, description: str, **kwargs) -> BaseTool: """Converts a chain into a tool.""" return Tool( name=name, description=description, func=chain.invoke, corouti...
[ "langchain.tools.Tool" ]
[((220, 319), 'langchain.tools.Tool', 'Tool', ([], {'name': 'name', 'description': 'description', 'func': 'chain.invoke', 'coroutine': 'chain.ainvoke'}), '(name=name, description=description, func=chain.invoke, coroutine=chain\n .ainvoke, **kwargs)\n', (224, 319), False, 'from langchain.tools import Tool, BaseTool\n...
"""Loaders for Prefect.""" import asyncio import httpx import os import shutil import tempfile from pathlib import Path from typing import List from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain_prefect.types import GitHubComment, GitHubIssue from pre...
[ "langchain_prefect.types.GitHubIssue", "langchain.docstore.document.Document", "langchain_prefect.types.GitHubComment" ]
[((4368, 4386), 'tempfile.mkdtemp', 'tempfile.mkdtemp', ([], {}), '()\n', (4384, 4386), False, 'import tempfile\n'), ((944, 974), 'os.environ.get', 'os.environ.get', (['"""GITHUB_TOKEN"""'], {}), "('GITHUB_TOKEN')\n", (958, 974), False, 'import os\n'), ((1493, 1590), 'httpx.get', 'httpx.get', ([], {'url': 'url', 'heade...
from langchain.agents import AgentType, initialize_agent, load_tools from langchain.llms import OpenAI from benchllm import SemanticEvaluator, Test, Tester tools = load_tools(["serpapi", "llm-math"], llm=OpenAI(temperature=0)) agent = initialize_agent(tools, OpenAI(temperature=0), agent=AgentType.ZERO_SHOT_REACT_DESC...
[ "langchain.llms.OpenAI" ]
[((569, 588), 'benchllm.SemanticEvaluator', 'SemanticEvaluator', ([], {}), '()\n', (586, 588), False, 'from benchllm import SemanticEvaluator, Test, Tester\n'), ((261, 282), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (267, 282), False, 'from langchain.llms import OpenAI\n'), (...
from dotenv import load_dotenv from langchain.chains import RetrievalQA from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings from langchain import HuggingFacePipeline from colorama import Fore, Style import re from langchain.vectorstores import Chroma from langchain.docstore.document im...
[ "langchain.docstore.document.Document", "langchain.vectorstores.Chroma.from_documents", "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((490, 503), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (501, 503), False, 'from dotenv import load_dotenv\n'), ((517, 539), 'os.getenv', 'os.getenv', (['"""TEST_FILE"""'], {}), "('TEST_FILE')\n", (526, 539), False, 'import os\n'), ((559, 588), 'os.getenv', 'os.getenv', (['"""EMBEDDINGS_MODEL"""'], {}), "(...
"""Wrapper around HuggingFace Pipeline APIs.""" import importlib.util import logging from typing import Any, List, Mapping, Optional from pydantic import BaseModel, Extra from langchain.llms.base import LLM from langchain.llms.utils import enforce_stop_tokens DEFAULT_MODEL_ID = "gpt2" DEFAULT_TASK = "text-generation...
[ "langchain.llms.utils.enforce_stop_tokens" ]
[((390, 409), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (407, 409), False, 'import logging\n'), ((2546, 2602), 'transformers.AutoTokenizer.from_pretrained', 'AutoTokenizer.from_pretrained', (['model_id'], {}), '(model_id, **_model_kwargs)\n', (2575, 2602), False, 'from transformers import AutoModelFor...
from langchain.retrievers.self_query.base import SelfQueryRetriever from langchain.chains.query_constructor.base import AttributeInfo from datetime import datetime current_time_iso = datetime.utcnow().isoformat() + "Z" # example metadat """ { "type": "file_load_gcs", "attrs": "namespace:edmonbrain", "source": ...
[ "langchain.retrievers.self_query.base.SelfQueryRetriever.from_llm", "langchain.chains.query_constructor.base.AttributeInfo" ]
[((1179, 1311), 'langchain.chains.query_constructor.base.AttributeInfo', 'AttributeInfo', ([], {'name': '"""source"""', 'description': '"""The document source url or path to where the document is located"""', 'type': '"""string"""'}), "(name='source', description=\n 'The document source url or path to where the docu...
import sys from langchain.embeddings import HuggingFaceEmbeddings from langchain.vectorstores import FAISS from langchain.document_loaders import PyPDFLoader, DirectoryLoader, UnstructuredExcelLoader, TextLoader, UnstructuredPowerPointLoader, UnstructuredMarkdownLoader, Docx2txtLoader from langchain.text_splitter impor...
[ "langchain.document_loaders.DirectoryLoader", "langchain.embeddings.HuggingFaceEmbeddings", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.vectorstores.FAISS.from_documents", "langchain.vectorstores.FAISS.from_texts" ]
[((506, 552), 'os.path.join', 'os.path.join', (['current_directory', '""".."""', '"""mpnet"""'], {}), "(current_directory, '..', 'mpnet')\n", (518, 552), False, 'import os\n'), ((666, 751), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': 'model_directory', 'model_kwargs': "{'d...
import os import re from typing import List, Optional, Any from langchain.schema import Document from langchain.text_splitter import RecursiveCharacterTextSplitter from loguru import logger from tqdm import tqdm from src.config import local_embedding, retrieve_proxy, chunk_overlap, chunk_size, hf_emb_model_name from ...
[ "langchain.document_loaders.UnstructuredWordDocumentLoader", "langchain.embeddings.huggingface.HuggingFaceEmbeddings", "langchain_community.vectorstores.FAISS.from_documents", "langchain.document_loaders.UnstructuredPowerPointLoader", "langchain.document_loaders.UnstructuredEPubLoader", "langchain.schema....
[((440, 465), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (455, 465), False, 'import os\n'), ((3874, 3910), 'loguru.logger.debug', 'logger.debug', (['"""Loading documents..."""'], {}), "('Loading documents...')\n", (3886, 3910), False, 'from loguru import logger\n'), ((3915, 3956), 'loguru...
from fastapi import FastAPI from langchain.chains import RetrievalQA from langchain.chat_models import ChatOpenAI from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import ElasticVectorSearch from config import openai_api_key embedding = OpenAIEmbeddings(openai_api_key=openai_api_key...
[ "langchain.vectorstores.ElasticVectorSearch", "langchain.chat_models.ChatOpenAI", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((274, 321), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'openai_api_key': 'openai_api_key'}), '(openai_api_key=openai_api_key)\n', (290, 321), False, 'from langchain.embeddings.openai import OpenAIEmbeddings\n'), ((328, 444), 'langchain.vectorstores.ElasticVectorSearch', 'ElasticVectorSe...
from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List import pandas as pd import streamlit as st from langchain.chains import LLMChain from langchain.prompts.few_shot import FewShotPromptTemplate from doccano_mini.components import ( display_download_button, openai_model_f...
[ "langchain.chains.LLMChain" ]
[((763, 785), 'pandas.read_json', 'pd.read_json', (['filepath'], {}), '(filepath)\n', (775, 785), True, 'import pandas as pd\n'), ((921, 984), 'streamlit.experimental_data_editor', 'st.experimental_data_editor', (['df'], {'num_rows': '"""dynamic"""', 'width': '(1000)'}), "(df, num_rows='dynamic', width=1000)\n", (948, ...
"""This module contains functions for loading and managing vector stores in the Wandbot ingestion system. The module includes the following functions: - `load`: Loads the vector store from the specified source artifact path and returns the name of the resulting artifact. Typical usage example: project = "wandbot...
[ "langchain.schema.Document" ]
[((944, 964), 'wandbot.utils.get_logger', 'get_logger', (['__name__'], {}), '(__name__)\n', (954, 964), False, 'from wandbot.utils import get_logger, load_index, load_service_context, load_storage_context\n'), ((1677, 1696), 'wandbot.ingestion.config.VectorStoreConfig', 'VectorStoreConfig', ([], {}), '()\n', (1694, 169...
""" This module contains the OpenAIImageToText class, which is a subclass of ChatOpenAI that is specialized for converting images to text. """ from langchain_openai import ChatOpenAI from langchain_core.messages import HumanMessage class OpenAIImageToText(ChatOpenAI): """ A class that uses OpenAI's Chat API...
[ "langchain_core.messages.HumanMessage" ]
[((1233, 1394), 'langchain_core.messages.HumanMessage', 'HumanMessage', ([], {'content': "[{'type': 'text', 'text': 'What is this image showing'}, {'type':\n 'image_url', 'image_url': {'url': image_url, 'detail': 'auto'}}]"}), "(content=[{'type': 'text', 'text': 'What is this image showing'\n }, {'type': 'image_u...
from langchain.tools import BaseTool from langchain.tools.render import render_text_description from langchain_core.language_models.base import LanguageModelLike from langchain_core.messages import ( AIMessage, FunctionMessage, HumanMessage, SystemMessage, ) from langgraph.checkpoint import BaseCheckpoi...
[ "langchain_core.messages.AIMessage", "langchain.tools.render.render_text_description", "langchain_core.messages.SystemMessage", "langchain_core.messages.FunctionMessage" ]
[((1121, 1143), 'langchain_core.messages.AIMessage', 'AIMessage', ([], {'content': 'log'}), '(content=log)\n', (1130, 1143), False, 'from langchain_core.messages import AIMessage, FunctionMessage, HumanMessage, SystemMessage\n'), ((2644, 2663), 'langgraph.prebuilt.ToolExecutor', 'ToolExecutor', (['tools'], {}), '(tools...
# Copyright 2024 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, s...
[ "langchain_google_el_carro.ElCarroEngine.from_instance", "langchain_google_el_carro.chat_message_history.ElCarroChatMessageHistory" ]
[((1019, 1095), 'langchain_google_el_carro.ElCarroEngine.from_instance', 'ElCarroEngine.from_instance', (['db_host', 'db_port', 'db_name', 'db_user', 'db_password'], {}), '(db_host, db_port, db_name, db_user, db_password)\n', (1046, 1095), False, 'from langchain_google_el_carro import ElCarroEngine\n'), ((1238, 1349), ...
from typing import Optional, Type from langchain.callbacks.manager import CallbackManagerForToolRun from langchain_community.utilities import OpenWeatherMapAPIWrapper from pydantic import BaseModel, Field from exceptions import ToolEnvKeyException from tools.base import BaseTool class OpenWeatherMapSchema(BaseModel...
[ "langchain_community.utilities.OpenWeatherMapAPIWrapper" ]
[((340, 402), 'pydantic.Field', 'Field', (['...'], {'description': '"""The search query for OpenWeatherMap."""'}), "(..., description='The search query for OpenWeatherMap.')\n", (345, 402), False, 'from pydantic import BaseModel, Field\n'), ((1406, 1477), 'langchain_community.utilities.OpenWeatherMapAPIWrapper', 'OpenW...
from textwrap import dedent from langchain import OpenAI from langchain.schema import BaseModel from utils import format_prompt_components_without_tools def extract_first_message(message: str) -> str: """The LLM can continue the conversation from the recipient. So extract just the first line.""" return mes...
[ "langchain.OpenAI" ]
[((627, 656), 'textwrap.dedent', 'dedent', (['inspirational_thought'], {}), '(inspirational_thought)\n', (633, 656), False, 'from textwrap import dedent\n'), ((912, 981), 'utils.format_prompt_components_without_tools', 'format_prompt_components_without_tools', (['ai_settings', 'contact_settings'], {}), '(ai_settings, c...
"""VectorStore wrapper around a Postgres/PGVector database.""" from __future__ import annotations import enum import logging import uuid from typing import Any, Dict, Iterable, List, Optional, Tuple, Type import sqlalchemy from pgvector.sqlalchemy import Vector from sqlalchemy.dialects.postgresql import JSON, UUID fr...
[ "langchain.utils.get_from_dict_or_env", "langchain.docstore.document.Document" ]
[((593, 611), 'sqlalchemy.orm.declarative_base', 'declarative_base', ([], {}), '()\n', (609, 611), False, 'from sqlalchemy.orm import Session, declarative_base, relationship\n'), ((929, 965), 'sqlalchemy.Column', 'sqlalchemy.Column', (['sqlalchemy.String'], {}), '(sqlalchemy.String)\n', (946, 965), False, 'import sqlal...
import tempfile import time import os from utils import compute_sha1_from_file from langchain.schema import Document import streamlit as st from langchain.text_splitter import RecursiveCharacterTextSplitter from typing import List from sqlite3 import Connection from verse.sqlite_helper import * def update_metadata(co...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter.from_tiktoken_encoder", "langchain.schema.Document" ]
[((1152, 1175), 'time.strftime', 'time.strftime', (['"""%Y%m%d"""'], {}), "('%Y%m%d')\n", (1165, 1175), False, 'import time\n'), ((1468, 1492), 'os.remove', 'os.remove', (['tmp_file.name'], {}), '(tmp_file.name)\n', (1477, 1492), False, 'import os\n'), ((1679, 1812), 'langchain.text_splitter.RecursiveCharacterTextSplit...
import json import logging from typing import Any, Dict, Iterator, List, Optional import requests from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.pydantic_v1 import Field from langchain.schema.output import GenerationChunk logger = logging.getLogger...
[ "langchain.pydantic_v1.Field", "langchain.schema.output.GenerationChunk" ]
[((303, 330), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (320, 330), False, 'import logging\n'), ((1278, 1308), 'langchain.pydantic_v1.Field', 'Field', (['(True)'], {'alias': '"""do_sample"""'}), "(True, alias='do_sample')\n", (1283, 1308), False, 'from langchain.pydantic_v1 import Fi...
# imports from loguru import logger # LLM modules from langchain_community.llms.huggingface_hub import HuggingFaceHub from langchain_community.llms.ollama import Ollama from langchain_openai import ChatOpenAI, AzureChatOpenAI from langchain.callbacks.manager import CallbackManager from langchain.callbacks.streaming_std...
[ "langchain_openai.AzureChatOpenAI", "langchain_openai.ChatOpenAI", "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain_community.llms.huggingface_hub.HuggingFaceHub" ]
[((1610, 1675), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'client': 'None', 'model': 'self.llm_model_type', 'temperature': '(0)'}), '(client=None, model=self.llm_model_type, temperature=0)\n', (1620, 1675), False, 'from langchain_openai import ChatOpenAI, AzureChatOpenAI\n'), ((2163, 2272), 'langchain_communit...
from typing import List from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores import Chroma from langchain_core.documents import Document from dotenv import load_dotenv from themind.llm.func_instraction import instruct from pydantic import BaseModel import csv from themind.vectorstores.chunking....
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((657, 675), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (673, 675), False, 'from langchain.embeddings import OpenAIEmbeddings\n')]
import re import time import copy import random import numpy as np import multiprocessing import matplotlib.pyplot as plt import modules.prompts as prompts from langchain import PromptTemplate from shapely.ops import substring from shapely.geometry import Polygon, box, Point, LineString class WallObjectGenerator(): ...
[ "langchain.PromptTemplate" ]
[((704, 850), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['room_type', 'wall_height', 'floor_objects', 'wall_objects']", 'template': 'prompts.wall_object_constraints_prompt'}), "(input_variables=['room_type', 'wall_height', 'floor_objects',\n 'wall_objects'], template=prompts.wall_object...
from langchain_openai import ChatOpenAI from langchain_core.output_parsers import StrOutputParser from langchain.prompts import PromptTemplate from langchain.prompts.chat import ChatPromptTemplate from config.config import OPENAI_API_KEY from game.poker import PokerGameManager from db.db_utils import DatabaseManager im...
[ "langchain_openai.ChatOpenAI", "langchain_core.output_parsers.StrOutputParser", "langchain.prompts.chat.ChatPromptTemplate.from_messages" ]
[((456, 489), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': 'model_name'}), '(model_name=model_name)\n', (466, 489), False, 'from langchain_openai import ChatOpenAI\n'), ((514, 531), 'langchain_core.output_parsers.StrOutputParser', 'StrOutputParser', ([], {}), '()\n', (529, 531), False, 'from langcha...
import logging from typing import Any, Dict, List, Optional from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.llms.utils import enforce_stop_tokens from langchain.pydantic_v1 import Extra, root_validator from langchain.schema import Generation, LLMResul...
[ "langchain.llms.utils.enforce_stop_tokens", "langchain.utils.get_from_dict_or_env", "langchain.schema.Generation", "langchain.schema.LLMResult", "langchain.pydantic_v1.root_validator" ]
[((381, 408), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (398, 408), False, 'import logging\n'), ((1472, 1488), 'langchain.pydantic_v1.root_validator', 'root_validator', ([], {}), '()\n', (1486, 1488), False, 'from langchain.pydantic_v1 import Extra, root_validator\n'), ((1702, 1753),...
import re from typing import Any, Dict, List, Optional from langchain_core.load import loads, dumps from langchain_community.chat_models import ( ChatAnthropic, ChatAnyscale, ChatBaichuan, QianfanChatEndpoint, BedrockChat, ChatDatabricks, ChatDeepInfra, ErnieBotChat, ChatEverlyAI, ...
[ "langchain_core.load.dumps" ]
[((9560, 9597), 're.search', 're.search', (['f"""{pattern}=\'(.*?)\'"""', 'text'], {}), '(f"{pattern}=\'(.*?)\'", text)\n', (9569, 9597), False, 'import re\n'), ((2545, 2562), 'langchain_core.load.dumps', 'dumps', (['serialized'], {}), '(serialized)\n', (2550, 2562), False, 'from langchain_core.load import loads, dumps...
"""This example shows how to use the ChatGPT API with LangChain to answer questions about Prefect.""" from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import Chroma from langchain.text_splitter import CharacterTextSplitter from langchain.chains import ChatVectorDBChain from langchai...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain_prefect.loaders.GitHubRepoLoader", "langchain.prompts.chat.SystemMessagePromptTemplate.from_template", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.Chroma.from_documents", "langchain.prompts.chat.HumanMessagePromptTemplate.from_tem...
[((680, 735), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {'chunk_size': '(1000)', 'chunk_overlap': '(0)'}), '(chunk_size=1000, chunk_overlap=0)\n', (701, 735), False, 'from langchain.text_splitter import CharacterTextSplitter\n'), ((803, 821), 'langchain.embeddings.openai.OpenAIEmbed...
from dotenv import load_dotenv load_dotenv() import os from langchain.llms import OpenAI from langchain.chat_models import ChatOpenAI from langchain.prompts import ( PromptTemplate, ) from langchain.chains import ConversationChain from langchain.memory import ConversationBufferMemory from langchain.agents import A...
[ "langchain.agents.ConversationalChatAgent.from_llm_and_tools", "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.memory.ConversationBufferMemory", "langchain.chat_models.ChatOpenAI" ]
[((31, 44), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (42, 44), False, 'from dotenv import load_dotenv\n'), ((574, 601), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (583, 601), False, 'import os\n'), ((1087, 1183), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([],...
from __future__ import annotations from abc import abstractmethod from typing import TYPE_CHECKING, Any, Dict, List, Sequence from langchain.load.serializable import Serializable from langchain.pydantic_v1 import Field if TYPE_CHECKING: from langchain.prompts.chat import ChatPromptTemplate def get_buffer_strin...
[ "langchain.pydantic_v1.Field", "langchain.prompts.chat.ChatPromptTemplate" ]
[((2151, 2178), 'langchain.pydantic_v1.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (2156, 2178), False, 'from langchain.pydantic_v1 import Field\n'), ((2610, 2645), 'langchain.prompts.chat.ChatPromptTemplate', 'ChatPromptTemplate', ([], {'messages': '[self]'}), '(messages=[self])\n',...
import logging from typing import Any, Dict, List, Mapping, Optional from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.llms.utils import enforce_stop_tokens from langchain.pydantic_v1 import BaseModel, Extra, Field, root_validator from langchain.utils i...
[ "langchain.llms.utils.enforce_stop_tokens", "langchain.pydantic_v1.Field", "langchain.pydantic_v1.root_validator", "langchain.utils.get_from_dict_or_env" ]
[((357, 384), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (374, 384), False, 'import logging\n'), ((1004, 1031), 'langchain.pydantic_v1.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1009, 1031), False, 'from langchain.pydantic_v1 import BaseModel, Ext...
from typing import Optional, Type import streamlit as st import tldextract import whois import whoisit from langchain.agents import AgentType, Tool, initialize_agent from langchain.chat_models import ChatOpenAI from langchain.tools import BaseTool from langchain.tools.ddg_search import DuckDuckGoSearchRun from pydanti...
[ "langchain.tools.ddg_search.DuckDuckGoSearchRun", "langchain.agents.initialize_agent", "langchain.agents.Tool", "langchain.chat_models.ChatOpenAI" ]
[((363, 390), 'streamlit.title', 'st.title', (['"""TakedownGPT ⬇️🤖"""'], {}), "('TakedownGPT ⬇️🤖')\n", (371, 390), True, 'import streamlit as st\n'), ((434, 467), 'streamlit.sidebar.header', 'st.sidebar.header', (['"""How to Use 📝"""'], {}), "('How to Use 📝')\n", (451, 467), True, 'import streamlit as st\n'), ((468...
# -*- coding: utf-8 -*- import os import re from typing import List, Union, Dict, Tuple, Any, Optional from langchain.agents import Tool, AgentExecutor, AgentOutputParser, load_tools from langchain.tools.base import BaseTool from langchain.prompts import StringPromptTemplate from langchain import OpenAI, GoogleSearchAP...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.LLMChain", "langchain.schema.AgentAction", "langchain.chat_models.ChatOpenAI", "langchain.schema.AgentFinish" ]
[((473, 493), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (488, 493), False, 'import sys\n'), ((494, 515), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (509, 515), False, 'import sys\n'), ((1325, 1340), 'utils.parser.get_arguments', 'get_arguments', ([], {}), '()\n', (13...
import sqlite3 import pandas as pd import os import json import warnings from langchain import SQLDatabase from langchain.docstore.document import Document from langchain.vectorstores import Chroma from langchain.embeddings import HuggingFaceEmbeddings from sqlalchemy import exc from sqlalchemy.exc import SAWarning ...
[ "langchain.embeddings.HuggingFaceEmbeddings" ]
[((320, 373), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {'category': 'SAWarning'}), "('ignore', category=SAWarning)\n", (343, 373), False, 'import warnings\n'), ((973, 1033), 'src.data.setup.db_setup_functions.build_schema_info', 'build_schema_info', ([], {'filepath': 'data_directory', 'f...
import asyncio from functools import partial from typing import ( Any, List, Mapping, Optional, ) from ai21.models import CompletionsResponse from langchain_core.callbacks import ( AsyncCallbackManagerForLLMRun, CallbackManagerForLLMRun, ) from langchain_core.language_models import BaseLLM from...
[ "langchain_core.outputs.LLMResult" ]
[((4179, 4236), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (4188, 4236), False, 'from langchain_core.outputs import Generation, LLMResult\n'), ((4626, 4659), 'functools.partial', 'partial', (['s...
from langchain.llms import OpenAI from callback import MyCallbackHandler from langchain.callbacks.base import BaseCallbackManager class QaLlm(): def __init__(self) -> None: manager = BaseCallbackManager([MyCallbackHandler()]) self.llm = OpenAI(temperature=0, callback_manager=manager, model_name="g...
[ "langchain.llms.OpenAI" ]
[((259, 334), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'callback_manager': 'manager', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, callback_manager=manager, model_name='gpt-3.5-turbo')\n", (265, 334), False, 'from langchain.llms import OpenAI\n'), ((218, 237), 'callback.MyCallbackHandle...
from langchain.llms import OpenAI from langchain.chat_models import ChatOpenAI from apikey import ( apikey, google_search, google_cse, serp, aws_access_key, aws_secret_key, aws_region, ) import os from typing import Dict from langchain.prompts import PromptTemplate from langchain.chains impo...
[ "langchain.memory.ConversationBufferMemory", "langchain.llms.OpenAI", "langchain.chains.LLMChain", "langchain.utilities.GoogleSearchAPIWrapper", "langchain.prompts.PromptTemplate" ]
[((765, 835), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0.3)', 'max_tokens': '(100)', 'model_name': '"""text-davinci-003"""'}), "(temperature=0.3, max_tokens=100, model_name='text-davinci-003')\n", (771, 835), False, 'from langchain.llms import OpenAI\n'), ((860, 886), 'langchain.memory.ConversationBuff...
""" DATE: 2023/5/28 AUTHOR: ZLYANG CONTACT: zhlyang95@hotmail.com """ ### define tools ### import requests import io import base64 import os from PIL import Image from typing import Optional from langchain.tools import BaseTool from langchain.callbacks.manager import ( AsyncCallbackManagerForToo...
[ "langchain.LLMChain", "langchain.PromptTemplate.from_template" ]
[((757, 785), 'random.randint', 'random.randint', (['(32768)', '(65536)'], {}), '(32768, 65536)\n', (771, 785), False, 'import re, random\n'), ((2323, 2364), 'langchain.PromptTemplate.from_template', 'PromptTemplate.from_template', (['qa_template'], {}), '(qa_template)\n', (2351, 2364), False, 'from langchain import LL...
from langchain.retrievers import AmazonKendraRetriever from langchain.chains import ConversationalRetrievalChain from langchain import SagemakerEndpoint from langchain.llms.sagemaker_endpoint import LLMContentHandler from langchain.prompts import PromptTemplate import sys import json import os class bcolors: HEAD...
[ "langchain.SagemakerEndpoint", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.retrievers.AmazonKendraRetriever", "langchain.prompts.PromptTemplate.from_template", "langchain.prompts.PromptTemplate" ]
[((1327, 1604), 'langchain.SagemakerEndpoint', 'SagemakerEndpoint', ([], {'endpoint_name': 'endpoint_name', 'region_name': 'region', 'model_kwargs': "{'temperature': 0.8, 'max_new_tokens': 512, 'do_sample': True, 'top_p': 0.9,\n 'repetition_penalty': 1.03, 'stop': ['\\nUser:', '<|endoftext|>', '</s>']}", 'content_ha...
#Make sure to install the following packages: dlt, langchain, duckdb, python-dotenv, openai, weaviate-client import dlt from langchain import PromptTemplate, LLMChain from langchain.chains.openai_functions import create_structured_output_chain from langchain.chat_models import ChatOpenAI from langchain.document_loader...
[ "langchain.LLMChain", "langchain.chains.openai_functions.create_structured_output_chain", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.retrievers.WeaviateHybridSearchRetriever", "langchain.LLMMathChain.from_llm", "langchain.schema.HumanMessage", "langchain.schema.SystemMessag...
[((741, 754), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (752, 754), False, 'from dotenv import load_dotenv\n'), ((848, 866), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (864, 866), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((1129, 1146), 'langchain.do...
import sys import os sys.path.append(os.path.dirname(os.path.abspath(__file__)) + '/../') import asyncio from argparse import Namespace from models.loader.args import parser from models.loader import LoaderCheckPoint from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType import mo...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.memory.ConversationBufferMemory", "langchain.agents.Tool", "langchain.memory.ReadOnlySharedMemory", "langchain.chains.LLMChain", "langchain.prompts.PromptTemplate" ]
[((1027, 1054), 'models.loader.LoaderCheckPoint', 'LoaderCheckPoint', (['args_dict'], {}), '(args_dict)\n', (1043, 1054), False, 'from models.loader import LoaderCheckPoint\n'), ((1075, 1093), 'models.shared.loaderLLM', 'shared.loaderLLM', ([], {}), '()\n', (1091, 1093), True, 'import models.shared as shared\n'), ((125...
import logging from time import sleep from langchain.llms import OpenAI from scrapy import Request, Spider from selenium import webdriver from selenium.webdriver.common.keys import Keys from conf import ( CONNECTION_REQUEST_LLM_PROMPT, DEFAULT_CONNECTION_MESSAGE, MAX_PROFILES_TO_CONNECT, MAX_PROFILES_...
[ "langchain.llms.OpenAI", "langchain.PromptTemplate.from_template" ]
[((689, 716), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (706, 716), False, 'import logging\n'), ((1186, 1271), 'linkedin.integrations.selenium.get_by_xpath_or_none', 'get_by_xpath_or_none', (['driver', '"""//button[@aria-label="Got it"]"""'], {'wait_timeout': '(0.5)'}), '(driver, \'/...
import streamlit as st import os from azure.storage.blob import BlobServiceClient, BlobClient, ContainerClient from PyPDF2 import PdfReader # Import #import textwrap import openai from langchain.llms import AzureOpenAI, OpenAI from langchain.embeddings import OpenAIEmbeddings from llama_index.vector_stores import Redis...
[ "langchain.embeddings.OpenAIEmbeddings", "langchain.llms.OpenAI", "langchain.llms.AzureOpenAI" ]
[((558, 616), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.INFO'}), '(stream=sys.stdout, level=logging.INFO)\n', (577, 616), False, 'import logging\n'), ((744, 780), 'os.getenv', 'os.getenv', (['"""REDIS_HOST"""', '"""localhost"""'], {}), "('REDIS_HOST', 'localhost')\n",...
from langchain.chat_models import ChatOpenAI from langchain.schema import HumanMessage, SystemMessage from whenx.models.team import Team from whenx.models.scout import Scout from whenx.models.sentinel import Sentinel from whenx.models.soldier import Soldier import re from whenx.database import db class Captain: ...
[ "langchain.schema.SystemMessage", "langchain.schema.HumanMessage", "langchain.chat_models.ChatOpenAI" ]
[((575, 587), 'whenx.database.db.add', 'db.add', (['team'], {}), '(team)\n', (581, 587), False, 'from whenx.database import db\n'), ((596, 607), 'whenx.database.db.commit', 'db.commit', ([], {}), '()\n', (605, 607), False, 'from whenx.database import db\n'), ((624, 675), 'whenx.models.scout.Scout', 'Scout', ([], {'inst...
import json import re from langchain.chains import RetrievalQA from utils.functions import find_nth, remove_extra_heading, add_json_characters, Timeout from langchain import LLMChain from langchain.chat_models import ChatOpenAI def section_schemas(heading, keyword, format_instructions, retriever, prompt): ch...
[ "langchain.LLMChain", "langchain.chains.RetrievalQA.from_chain_type", "langchain.chat_models.ChatOpenAI" ]
[((325, 387), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model_name': '"""gpt-3.5-turbo-16k-0613"""'}), "(temperature=0, model_name='gpt-3.5-turbo-16k-0613')\n", (335, 387), False, 'from langchain.chat_models import ChatOpenAI\n'), ((433, 466), 'langchain.LLMChain', 'LLMChain', ([], ...
"""Experiment with different models.""" from __future__ import annotations from typing import List, Optional, Sequence from langchain_core.language_models.llms import BaseLLM from langchain_core.prompts.prompt import PromptTemplate from langchain_core.utils.input import get_color_mapping, print_text from langchain....
[ "langchain_core.utils.input.print_text", "langchain_core.utils.input.get_color_mapping", "langchain_core.prompts.prompt.PromptTemplate", "langchain.chains.llm.LLMChain" ]
[((1752, 1782), 'langchain_core.utils.input.get_color_mapping', 'get_color_mapping', (['chain_range'], {}), '(chain_range)\n', (1769, 1782), False, 'from langchain_core.utils.input import get_color_mapping, print_text\n'), ((2307, 2370), 'langchain_core.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'input_var...
from typing import Any, List, Optional, Sequence, Tuple from langchain_core._api import deprecated from langchain_core.agents import AgentAction from langchain_core.callbacks import BaseCallbackManager from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts import BasePromptTemplate fr...
[ "langchain_core.prompts.chat.SystemMessagePromptTemplate.from_template", "langchain.chains.llm.LLMChain", "langchain.agents.utils.validate_tools_single_input", "langchain_core.pydantic_v1.Field", "langchain_core.prompts.chat.ChatPromptTemplate", "langchain_core.prompts.chat.HumanMessagePromptTemplate.from...
[((915, 985), 'langchain_core._api.deprecated', 'deprecated', (['"""0.1.0"""'], {'alternative': '"""create_react_agent"""', 'removal': '"""0.2.0"""'}), "('0.1.0', alternative='create_react_agent', removal='0.2.0')\n", (925, 985), False, 'from langchain_core._api import deprecated\n'), ((1072, 1111), 'langchain_core.pyd...
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. 2023 # SPDX-License-Identifier: Apache-2.0 from typing import Any, Dict, List, Optional from langchain.agents import tool from langchain.chains.base import Chain from langchain.chains import LLMChain from langchain import PromptTemplate from langcha...
[ "langchain.chains.LLMChain", "langchain.callbacks.manager.CallbackManagerForChainRun.get_noop_manager" ]
[((1178, 1216), 'chainlit.context.context.session.emit', 'context.session.emit', (['"""view"""', 'entityId'], {}), "('view', entityId)\n", (1198, 1216), False, 'from chainlit.context import context\n'), ((2370, 2412), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=p...
from langchain.retrievers import AmazonKendraRetriever from langchain.chains import RetrievalQA from langchain import OpenAI from langchain.prompts import PromptTemplate from langchain import SagemakerEndpoint from langchain.llms.sagemaker_endpoint import LLMContentHandler import json import os def build_chain(): ...
[ "langchain.SagemakerEndpoint", "langchain.retrievers.AmazonKendraRetriever", "langchain.prompts.PromptTemplate", "langchain.chains.RetrievalQA.from_chain_type" ]
[((1839, 1906), 'langchain.retrievers.AmazonKendraRetriever', 'AmazonKendraRetriever', ([], {'index_id': 'kendra_index_id', 'region_name': 'region'}), '(index_id=kendra_index_id, region_name=region)\n', (1860, 1906), False, 'from langchain.retrievers import AmazonKendraRetriever\n'), ((2373, 2458), 'langchain.prompts.P...
from typing import List, Optional from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.utilities.arxiv import ArxivAPIWrapper class ArxivLoader(BaseLoader): """Load a query result from `Arxiv`. The loader converts the original PDF format into...
[ "langchain.utilities.arxiv.ArxivAPIWrapper" ]
[((863, 971), 'langchain.utilities.arxiv.ArxivAPIWrapper', 'ArxivAPIWrapper', ([], {'load_max_docs': 'self.load_max_docs', 'load_all_available_meta': 'self.load_all_available_meta'}), '(load_max_docs=self.load_max_docs, load_all_available_meta=\n self.load_all_available_meta)\n', (878, 971), False, 'from langchain.u...
# flake8: noqa from langchain_core.prompts.prompt import PromptTemplate API_URL_PROMPT_TEMPLATE = """You are given the below API Documentation: {api_docs} Using this documentation, generate the full API url to call for answering the user question. You should build the API url in order to get a response that is as shor...
[ "langchain_core.prompts.prompt.PromptTemplate" ]
[((542, 637), 'langchain_core.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['api_docs', 'question']", 'template': 'API_URL_PROMPT_TEMPLATE'}), "(input_variables=['api_docs', 'question'], template=\n API_URL_PROMPT_TEMPLATE)\n", (556, 637), False, 'from langchain_core.prompts.prompt imp...