code stringlengths 141 79.4k | apis listlengths 1 23 | extract_api stringlengths 126 73.2k |
|---|---|---|
from __future__ import annotations
from typing import TYPE_CHECKING, ClassVar, Collection, Dict, List
from venv import logger
import requests
from langchain.schema.document import Document
from langchain.vectorstores.base import VectorStore, VectorStoreRetriever
from langchain_core.pydantic_v1 import Field, root_vali... | [
"langchain.schema.document.Document",
"langchain_core.pydantic_v1.Field",
"langchain_core.pydantic_v1.root_validator"
] | [((631, 658), 'langchain_core.pydantic_v1.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (636, 658), False, 'from langchain_core.pydantic_v1 import Field, root_validator\n'), ((942, 958), 'langchain_core.pydantic_v1.root_validator', 'root_validator', ([], {}), '()\n', (956, 958), False,... |
import os
from typing import Optional
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
from langchain.schema import BaseMessage, HumanMessage
from rebyte_langchain.rebyte_langchain import RebyteEndpoint
from realtime_ai_character.llm.base import (
AsyncCallbackAudioHandler,
Asyn... | [
"langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler",
"langchain.schema.HumanMessage"
] | [((473, 493), 'realtime_ai_character.logger.get_logger', 'get_logger', (['__name__'], {}), '(__name__)\n', (483, 493), False, 'from realtime_ai_character.logger import get_logger\n'), ((572, 603), 'os.getenv', 'os.getenv', (['"""REBYTE_API_KEY"""', '""""""'], {}), "('REBYTE_API_KEY', '')\n", (581, 603), False, 'import ... |
from langchain.agents import AgentType, initialize_agent
from langchain_community.agent_toolkits import ZapierToolkit
from langchain_community.tools import BaseTool
from langchain_community.utilities.zapier import ZapierNLAWrapper
from langchain_openai import ChatOpenAI
class ZapierNLA(BaseTool):
name = "Zapier"
... | [
"langchain_openai.ChatOpenAI",
"langchain_community.agent_toolkits.ZapierToolkit.from_zapier_nla_wrapper",
"langchain_community.utilities.zapier.ZapierNLAWrapper"
] | [((577, 632), 'langchain_community.utilities.zapier.ZapierNLAWrapper', 'ZapierNLAWrapper', ([], {'zapier_nla_api_key': 'zapier_nla_api_key'}), '(zapier_nla_api_key=zapier_nla_api_key)\n', (593, 632), False, 'from langchain_community.utilities.zapier import ZapierNLAWrapper\n'), ((651, 696), 'langchain_community.agent_t... |
from celery import shared_task
from langchain.text_splitter import RecursiveCharacterTextSplitter
from shared.models.opencopilot_db.pdf_data_sources import (
insert_pdf_data_source,
update_pdf_data_source_status,
)
from langchain.document_loaders import UnstructuredMarkdownLoader
from shared.utils.opencopilot_... | [
"langchain.text_splitter.RecursiveCharacterTextSplitter"
] | [((1830, 1925), 'shared.models.opencopilot_db.pdf_data_sources.update_pdf_data_source_status', 'update_pdf_data_source_status', ([], {'chatbot_id': 'chatbot_id', 'file_name': 'file_name', 'status': '"""PENDING"""'}), "(chatbot_id=chatbot_id, file_name=file_name,\n status='PENDING')\n", (1859, 1925), False, 'from sha... |
from concurrent.futures import ThreadPoolExecutor
import asyncio
import logging
from typing import Optional, Tuple
from langchain import ConversationChain
from vocode.streaming.agent.base_agent import RespondAgent
from vocode.streaming.models.agent import ChatVertexAIAgentConfig
from langchain_community.chat_models imp... | [
"langchain.prompts.HumanMessagePromptTemplate.from_template",
"langchain.memory.ConversationBufferMemory",
"langchain.prompts.MessagesPlaceholder",
"langchain.schema.SystemMessage",
"langchain_community.chat_models.ChatVertexAI",
"langchain.ConversationChain"
] | [((1089, 1103), 'langchain_community.chat_models.ChatVertexAI', 'ChatVertexAI', ([], {}), '()\n', (1101, 1103), False, 'from langchain_community.chat_models import ChatVertexAI\n'), ((1127, 1173), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'return_messages': '(True)'}), '(return_mess... |
# SPDX-FileCopyrightText: Copyright (c) 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
# SPDX-License-Identifier: Apache-2.0
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# ht... | [
"langchain.text_splitter.CharacterTextSplitter",
"langchain.document_loaders.DirectoryLoader",
"langchain_core.output_parsers.StrOutputParser",
"langchain.vectorstores.FAISS.from_documents",
"langchain_core.prompts.ChatPromptTemplate.from_messages",
"langchain_nvidia_ai_endpoints.NVIDIAEmbeddings",
"lan... | [((1034, 1067), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (1052, 1067), True, 'import streamlit as st\n'), ((2031, 2063), 'langchain_nvidia_ai_endpoints.ChatNVIDIA', 'ChatNVIDIA', ([], {'model': '"""mixtral_8x7b"""'}), "(model='mixtral_8x7b')\n", (2041, 2063... |
from langchain.chains import RetrievalQA, ConversationalRetrievalChain, ConversationChain
from langchain.prompts.prompt import PromptTemplate
from langchain.vectorstores.base import VectorStoreRetriever
from langchain.chat_models import ChatOpenAI
from langchain.memory import ConversationBufferMemory
import pickle
impo... | [
"langchain.chains.ConversationChain",
"langchain.prompts.prompt.PromptTemplate",
"langchain.vectorstores.base.VectorStoreRetriever",
"langchain.chains.ConversationalRetrievalChain.from_llm",
"langchain.memory.ConversationBufferMemory",
"langchain.chat_models.ChatOpenAI",
"langchain.prompts.prompt.Prompt... | [((727, 766), 'langchain.prompts.prompt.PromptTemplate.from_template', 'PromptTemplate.from_template', (['_template'], {}), '(_template)\n', (755, 766), False, 'from langchain.prompts.prompt import PromptTemplate\n'), ((1521, 1595), 'langchain.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'template': 'templat... |
# flake8: noqa
from langchain.prompts import PromptTemplate
## Use a shorter template to reduce the number of tokens in the prompt
template = """Create a final answer to the given questions using the provided document excerpts (given in no particular order) as sources. ALWAYS include a "SOURCES" section in your answer... | [
"langchain.prompts.PromptTemplate"
] | [((2121, 2197), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'template': 'template', 'input_variables': "['summaries', 'question']"}), "(template=template, input_variables=['summaries', 'question'])\n", (2135, 2197), False, 'from langchain.prompts import PromptTemplate\n')] |
import whisper
import textwrap
from langchain.chat_models import ChatOpenAI
from langchain.chains.summarize import load_summarize_chain
from langchain.schema import Document
from langchain.llms import OpenAI
from langchain import PromptTemplate
from langchain.chains import LLMChain
from langchain.document_loaders impo... | [
"langchain.chains.summarize.load_summarize_chain",
"langchain.llms.OpenAI",
"langchain.chat_models.ChatOpenAI",
"langchain.schema.Document",
"langchain.document_loaders.WebBaseLoader",
"langchain.chains.LLMChain",
"langchain.PromptTemplate"
] | [((508, 534), 'whisper.load_model', 'whisper.load_model', (['"""base"""'], {}), "('base')\n", (526, 534), False, 'import whisper\n'), ((607, 653), 'textwrap.fill', 'textwrap.fill', (["transcription['text']"], {'width': '(50)'}), "(transcription['text'], width=50)\n", (620, 653), False, 'import textwrap\n'), ((709, 766)... |
from langchain.agents import load_tools
from langchain.tools import AIPluginTool
from parse import *
from langchain.chat_models.base import BaseChatModel
from langchain.chat_models import ChatOpenAI, AzureChatOpenAI
import utils
def create_plugins_static():
plugins = [
AIPluginTool.from_plugin_url(
... | [
"langchain.tools.AIPluginTool.from_plugin_url",
"langchain.chat_models.AzureChatOpenAI",
"langchain.agents.load_tools",
"langchain.chat_models.ChatOpenAI"
] | [((410, 438), 'langchain.agents.load_tools', 'load_tools', (["['requests_all']"], {}), "(['requests_all'])\n", (420, 438), False, 'from langchain.agents import load_tools\n'), ((285, 371), 'langchain.tools.AIPluginTool.from_plugin_url', 'AIPluginTool.from_plugin_url', (['"""https://www.klarna.com/.well-known/ai-plugin.... |
import re
import string
from collections import Counter
import numpy as np
import pandas as pd
import tqdm
from langchain.evaluation.qa import QAEvalChain
from langchain.llms import OpenAI
from algos.PWS import PWS_Base, PWS_Extra
from algos.notool import CoT, IO
from algos.react import ReactBase
def normalize_answ... | [
"langchain.llms.OpenAI"
] | [((373, 410), 're.sub', 're.sub', (['"""\\\\b(a|an|the)\\\\b"""', '""" """', 'text'], {}), "('\\\\b(a|an|the)\\\\b', ' ', text)\n", (379, 410), False, 'import re\n'), ((1278, 1304), 'collections.Counter', 'Counter', (['prediction_tokens'], {}), '(prediction_tokens)\n', (1285, 1304), False, 'from collections import Coun... |
from datetime import date, datetime
from decimal import Decimal
from langchain.chains import LLMChain
from langchain.prompts.chat import (
ChatPromptTemplate,
HumanMessagePromptTemplate,
)
from sqlalchemy import text
from dataherald.model.chat_model import ChatModel
from dataherald.repositories.database_conne... | [
"langchain.prompts.chat.HumanMessagePromptTemplate.from_template",
"langchain.chains.LLMChain",
"langchain.prompts.chat.ChatPromptTemplate.from_messages"
] | [((1101, 1123), 'dataherald.model.chat_model.ChatModel', 'ChatModel', (['self.system'], {}), '(self.system)\n', (1110, 1123), False, 'from dataherald.model.chat_model import ChatModel\n'), ((1272, 1302), 'dataherald.repositories.prompts.PromptRepository', 'PromptRepository', (['self.storage'], {}), '(self.storage)\n', ... |
import streamlit as st
import urllib
import os
import re
import time
import random
from operator import itemgetter
from collections import OrderedDict
from langchain_core.documents import Document
from langchain_openai import AzureChatOpenAI
from langchain_core.output_parsers import StrOutputParser
from utils import g... | [
"langchain_openai.AzureChatOpenAI",
"langchain_core.documents.Document",
"langchain_core.output_parsers.StrOutputParser"
] | [((376, 455), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""GPT Smart Search"""', 'page_icon': '"""📖"""', 'layout': '"""wide"""'}), "(page_title='GPT Smart Search', page_icon='📖', layout='wide')\n", (394, 455), True, 'import streamlit as st\n'), ((498, 726), 'streamlit.markdown', 'st.mark... |
from __future__ import annotations
import asyncio
import logging
import typing as t
from abc import ABC, abstractmethod
from dataclasses import dataclass
from functools import partial
from langchain_community.chat_models import ChatVertexAI
from langchain_community.llms import VertexAI
from langchain_core.language_mo... | [
"langchain_openai.chat_models.ChatOpenAI"
] | [((765, 792), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (782, 792), False, 'import logging\n'), ((6829, 6869), 'langchain_openai.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model': 'model', 'timeout': 'timeout'}), '(model=model, timeout=timeout)\n', (6839, 6869), False, 'from langc... |
"""Utility functions and constants.
I am having some problems caching the memory and the retrieval. When
I decorate for caching, I get streamlit init errors.
"""
import logging
import pathlib
from typing import Any
from langchain.document_loaders import (
PyPDFLoader,
TextLoader,
UnstructuredEPubLoader,
... | [
"langchain.memory.ConversationBufferMemory"
] | [((637, 735), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'return_messages': '(True)', 'output_key': '"""answer"""'}), "(memory_key='chat_history', return_messages=True,\n output_key='answer')\n", (661, 735), False, 'from langchain.memory import ... |
import os
import re
import urllib
import urllib.parse
import urllib.request
from typing import Any, List, Tuple, Union
from urllib.parse import urlparse
import requests
from bs4 import BeautifulSoup
from langchain.chains import LLMChain
from langchain.prompts import Prompt
from langchain.tools import BaseTool
from lan... | [
"langchain.utilities.GoogleSerperAPIWrapper"
] | [((2438, 2451), 'sherpa_ai.config.task_config.AgentConfig', 'AgentConfig', ([], {}), '()\n', (2449, 2451), False, 'from sherpa_ai.config.task_config import AgentConfig\n'), ((894, 986), 'loguru.logger.warning', 'logger.warning', (['"""No SERPER_API_KEY found in environment variables, skipping SearchTool"""'], {}), "(\n... |
from dotenv import load_dotenv
from langchain_core.prompts import PromptTemplate
load_dotenv()
from langchain import hub
from langchain.agents import create_react_agent, AgentExecutor
from langchain_core.tools import Tool
from langchain_openai import ChatOpenAI
from tools.tools import get_profile_url
def lookup(nam... | [
"langchain_openai.ChatOpenAI",
"langchain.agents.AgentExecutor",
"langchain.agents.create_react_agent",
"langchain_core.tools.Tool",
"langchain_core.prompts.PromptTemplate",
"langchain.hub.pull"
] | [((82, 95), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (93, 95), False, 'from dotenv import load_dotenv\n'), ((346, 399), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, model_name='gpt-3.5-turbo')\n", (356, 399), False, 'from ... |
import logging, json, os
from Utilities.envVars import *
from Utilities.envVars import *
# Import required libraries
from Utilities.cogSearchVsRetriever import CognitiveSearchVsRetriever
from langchain.chains import RetrievalQA
from langchain import PromptTemplate
from Utilities.evaluator import indexDocs
import json
i... | [
"langchain.evaluation.qa.QAEvalChain.from_llm",
"langchain.chains.RetrievalQA.from_chain_type",
"langchain.chat_models.ChatOpenAI",
"langchain.document_loaders.PDFMinerLoader",
"langchain.chat_models.AzureChatOpenAI",
"langchain.PromptTemplate"
] | [((911, 1164), 'collections.namedtuple', 'namedtuple', (['"""RunDoc"""', "['evalatorQaData', 'totalQuestions', 'promptStyle', 'documentId',\n 'splitMethods', 'chunkSizes', 'overlaps', 'retrieverType', 'reEvaluate',\n 'topK', 'model', 'fileName', 'embeddingModelType', 'temperature',\n 'tokenLength']"], {}), "('... |
import dataclasses
import typing
from dataclasses import dataclass
from typing import Tuple, cast
from langchain.chat_models.base import BaseChatModel
from langchain.output_parsers import PydanticOutputParser
from langchain.schema import BaseMessage, HumanMessage
from pydantic import BaseModel, create_model
T = typin... | [
"langchain.output_parsers.PydanticOutputParser",
"langchain.schema.HumanMessage"
] | [((315, 334), 'typing.TypeVar', 'typing.TypeVar', (['"""T"""'], {}), "('T')\n", (329, 334), False, 'import typing\n'), ((339, 375), 'typing.TypeVar', 'typing.TypeVar', (['"""B"""'], {'bound': 'BaseModel'}), "('B', bound=BaseModel)\n", (353, 375), False, 'import typing\n'), ((568, 604), 'typing.cast', 'cast', (['str', '... |
from langchain.agents import load_tools
from langchain.agents import initialize_agent
from langchain.chat_models import ChatOpenAI
from virl.config import cfg
from virl.utils.common_utils import print_prompt, print_answer, parse_answer_to_json
from .gpt_chat import GPTChat
from .azure_gpt import AzureGPTChat
__all__... | [
"langchain.agents.initialize_agent",
"langchain.agents.load_tools",
"langchain.chat_models.ChatOpenAI"
] | [((679, 701), 'virl.utils.common_utils.print_prompt', 'print_prompt', (['question'], {}), '(question)\n', (691, 701), False, 'from virl.utils.common_utils import print_prompt, print_answer, parse_answer_to_json\n'), ((830, 850), 'virl.utils.common_utils.print_answer', 'print_answer', (['answer'], {}), '(answer)\n', (84... |
from langchain_community.chat_models import ChatAnthropic
from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate
from langchain_core.runnables import RunnablePassthrough
prompt = ChatPromptTemplate.from_template("Tell me a short joke about {topic}")
output_parse... | [
"langchain_core.prompts.ChatPromptTemplate.from_template",
"langchain_community.chat_models.ChatAnthropic",
"langchain_core.runnables.RunnablePassthrough",
"langchain_core.output_parsers.StrOutputParser"
] | [((237, 307), 'langchain_core.prompts.ChatPromptTemplate.from_template', 'ChatPromptTemplate.from_template', (['"""Tell me a short joke about {topic}"""'], {}), "('Tell me a short joke about {topic}')\n", (269, 307), False, 'from langchain_core.prompts import ChatPromptTemplate\n'), ((324, 341), 'langchain_core.output_... |
from langchain.chains import LLMChain
from langchain.prompts import PromptTemplate
from tqdm import tqdm
from lmchain.tools import tool_register
class GLMToolChain:
def __init__(self, llm):
self.llm = llm
self.tool_register = tool_register
self.tools = tool_register.get_tools()
def ... | [
"langchain.chains.LLMChain"
] | [((3292, 3320), 'lmchain.agents.llmMultiAgent.AgentZhipuAI', 'llmMultiAgent.AgentZhipuAI', ([], {}), '()\n', (3318, 3320), False, 'from lmchain.agents import llmMultiAgent\n'), ((3381, 3408), 'lmchain.chains.toolchain.GLMToolChain', 'toolchain.GLMToolChain', (['llm'], {}), '(llm)\n', (3403, 3408), False, 'from lmchain.... |
import json
import time
import hashlib
from typing import Dict, Any, List, Tuple
import re
from os import environ
import streamlit as st
from langchain.schema import BaseRetriever
from langchain.tools import Tool
from langchain.pydantic_v1 import BaseModel, Field
from sqlalchemy import Column, Text, create_engine, Me... | [
"langchain.agents.openai_functions_agent.agent_token_buffer_memory.AgentTokenBufferMemory",
"langchain.pydantic_v1.Field",
"langchain_experimental.retrievers.vector_sql_database.VectorSQLDatabaseChainRetriever",
"langchain.utilities.sql_database.SQLDatabase",
"langchain.schema.messages.ToolMessage",
"lang... | [((3163, 3322), 'langchain.prompts.ChatPromptTemplate.from_strings', 'ChatPromptTemplate.from_strings', ([], {'string_messages': "[(SystemMessagePromptTemplate, combine_prompt_template), (\n HumanMessagePromptTemplate, '{question}')]"}), "(string_messages=[(\n SystemMessagePromptTemplate, combine_prompt_template)... |
"""Simple agents which can be used as a starting point for running the deception environment with Umshini (see tutorials)."""
# pyright: reportGeneralTypeIssues=false
from langchain.chat_models import ChatOpenAI
from langchain.schema import AIMessage, HumanMessage, SystemMessage
class DeceptionDefender:
def __ini... | [
"langchain.schema.AIMessage",
"langchain.schema.SystemMessage",
"langchain.schema.HumanMessage",
"langchain.chat_models.ChatOpenAI"
] | [((433, 460), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0.9)'}), '(temperature=0.9)\n', (443, 460), False, 'from langchain.chat_models import ChatOpenAI\n'), ((738, 765), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0.9)'}), '(temperature=0.9)\n', (748, 765), F... |
"""Wrapper around Replicate API."""
import logging
from typing import Any, Dict, List, Mapping, Optional
from pydantic import Extra, Field, root_validator
from langchain.callbacks.manager import CallbackManagerForLLMRun
from langchain.llms.base import LLM
from langchain.utils import get_from_dict_or_env
logger = log... | [
"langchain.utils.get_from_dict_or_env"
] | [((317, 344), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (334, 344), False, 'import logging\n'), ((1212, 1239), 'pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1217, 1239), False, 'from pydantic import Extra, Field, root_validator\n'), ((1275... |
import databutton as db
import re
from io import BytesIO
from typing import Tuple, List
import pickle
from langchain.docstore.document import Document
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain.vectorstores.faiss import FAI... | [
"langchain.embeddings.openai.OpenAIEmbeddings",
"langchain.docstore.document.Document",
"langchain.text_splitter.RecursiveCharacterTextSplitter"
] | [((446, 461), 'pypdf.PdfReader', 'PdfReader', (['file'], {}), '(file)\n', (455, 461), False, 'from pypdf import PdfReader\n'), ((555, 597), 're.sub', 're.sub', (['"""(\\\\w+)-\\\\n(\\\\w+)"""', '"""\\\\1\\\\2"""', 'text'], {}), "('(\\\\w+)-\\\\n(\\\\w+)', '\\\\1\\\\2', text)\n", (561, 597), False, 'import re\n'), ((675... |
import datetime
import difflib
import logging
import os
from functools import wraps
from queue import Queue
from threading import Thread
from typing import Any, Callable, Dict, List
import numpy as np
import openai
import pandas as pd
import sqlalchemy
from google.api_core.exceptions import GoogleAPIError
from langcha... | [
"langchain.agents.mrkl.base.ZeroShotAgent.create_prompt",
"langchain.agents.mrkl.base.ZeroShotAgent",
"langchain.agents.agent.AgentExecutor.from_agent_and_tools",
"langchain.chains.llm.LLMChain",
"langchain_community.callbacks.get_openai_callback"
] | [((2000, 2027), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (2017, 2027), False, 'import logging\n'), ((2038, 2074), 'dataherald.sql_generator.SQLGenerator.get_upper_bound_limit', 'SQLGenerator.get_upper_bound_limit', ([], {}), '()\n', (2072, 2074), False, 'from dataherald.sql_generato... |
from marqo import Client
import pandas as pd
import numpy as np
from langchain_openai import OpenAI
from langchain.docstore.document import Document
from langchain.chains import LLMChain
from dotenv import load_dotenv
from utilities import (
load_data,
extract_text_from_highlights,
qna_prompt,
predic... | [
"langchain_openai.OpenAI",
"langchain.docstore.document.Document"
] | [((349, 362), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (360, 362), False, 'from dotenv import load_dotenv\n'), ((984, 992), 'marqo.Client', 'Client', ([], {}), '()\n', (990, 992), False, 'from marqo import Client\n'), ((1812, 1823), 'utilities.load_data', 'load_data', ([], {}), '()\n', (1821, 1823), False... |
from typing import List, Optional
from langchain.schema.language_model import BaseLanguageModel
from server.knowledge_base.model.kb_document_model import DocumentWithVSId
from configs import (logger)
from langchain.chains import StuffDocumentsChain, LLMChain
from langchain.prompts import PromptTemplate
from langchai... | [
"langchain.chains.combine_documents.map_reduce.MapReduceDocumentsChain",
"langchain.docstore.document.Document",
"langchain.chains.LLMChain",
"langchain.prompts.PromptTemplate.from_template",
"langchain.chains.StuffDocumentsChain",
"langchain.prompts.PromptTemplate",
"langchain.chains.combine_documents.... | [((1461, 1536), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['page_content']", 'template': '"""{page_content}"""'}), "(input_variables=['page_content'], template='{page_content}')\n", (1475, 1536), False, 'from langchain.prompts import PromptTemplate\n'), ((1853, 1943), 'langchain.pr... |
from fastapi import FastAPI, Form, Request, Response, File, Depends, HTTPException, status
from fastapi.responses import RedirectResponse
from fastapi.staticfiles import StaticFiles
from fastapi.templating import Jinja2Templates
from fastapi.encoders import jsonable_encoder
from langchain.llms import CTransformers... | [
"langchain.chains.summarize.load_summarize_chain",
"langchain.text_splitter.RecursiveCharacterTextSplitter",
"langchain.embeddings.HuggingFaceBgeEmbeddings",
"langchain.docstore.document.Document",
"langchain.llms.CTransformers",
"langchain.vectorstores.FAISS.from_documents",
"langchain.document_loaders... | [((911, 920), 'fastapi.FastAPI', 'FastAPI', ([], {}), '()\n', (918, 920), False, 'from fastapi import FastAPI, Form, Request, Response, File, Depends, HTTPException, status\n'), ((1008, 1046), 'fastapi.templating.Jinja2Templates', 'Jinja2Templates', ([], {'directory': '"""templates"""'}), "(directory='templates')\n", (... |
#!/usr/bin/env python
"""Example LangChain server exposes a retriever."""
from fastapi import FastAPI
from langchain.embeddings import OpenAIEmbeddings
from langchain.vectorstores import FAISS
from langserve import add_routes
vectorstore = FAISS.from_texts(
["cats like fish", "dogs like sticks"], embedding=OpenAI... | [
"langchain.embeddings.OpenAIEmbeddings"
] | [((381, 515), 'fastapi.FastAPI', 'FastAPI', ([], {'title': '"""LangChain Server"""', 'version': '"""1.0"""', 'description': '"""Spin up a simple api server using Langchain\'s Runnable interfaces"""'}), '(title=\'LangChain Server\', version=\'1.0\', description=\n "Spin up a simple api server using Langchain\'s Runna... |
"""
Chatbot for talking to Podcast using Langchain, Ollama and LanceDB
"""
from langchain.document_loaders import WikipediaLoader
import pandas as pd
from langchain.memory import ConversationSummaryMemory
import lancedb
from langchain.vectorstores import LanceDB
from langchain.embeddings import OpenAIEmbeddings
from l... | [
"langchain.text_splitter.RecursiveCharacterTextSplitter",
"langchain.memory.ConversationSummaryMemory",
"langchain.chains.ConversationalRetrievalChain.from_llm",
"langchain.document_loaders.WikipediaLoader",
"langchain.chat_models.ChatOllama",
"langchain.embeddings.OpenAIEmbeddings"
] | [((525, 556), 'lancedb.connect', 'lancedb.connect', (['"""/tmp/lancedb"""'], {}), "('/tmp/lancedb')\n", (540, 556), False, 'import lancedb\n'), ((883, 946), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(500)', 'chunk_overlap': '(0)'}), '(chunk_size=500... |
## Conversational Q&A Chatbot
import streamlit as st
from langchain.schema import HumanMessage,SystemMessage,AIMessage
from langchain.chat_models import ChatOpenAI
## Streamlit UI
st.set_page_config(page_title="Conversational Q&A Chatbot")
st.header("Hey, Let's Chat")
from dotenv import load_dotenv
load_d... | [
"langchain.schema.AIMessage",
"langchain.schema.SystemMessage",
"langchain.schema.HumanMessage",
"langchain.chat_models.ChatOpenAI"
] | [((189, 248), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Conversational Q&A Chatbot"""'}), "(page_title='Conversational Q&A Chatbot')\n", (207, 248), True, 'import streamlit as st\n'), ((250, 278), 'streamlit.header', 'st.header', (['"""Hey, Let\'s Chat"""'], {}), '("Hey, Let\'s Chat")\n... |
# Copyright 2023 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"langchain.chains.LLMChain",
"langchain.prompts.PromptTemplate",
"langchain.llms.HuggingFacePipeline"
] | [((882, 921), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (901, 921), False, 'import logging\n'), ((1019, 1069), 'os.environ.get', 'os.environ.get', (['"""MODEL_ID"""', '"""google/flan-t5-small"""'], {}), "('MODEL_ID', 'google/flan-t5-small')\n", (1033, 1069)... |
"""Prompt schema definition."""
from __future__ import annotations
from string import Formatter
from typing import Any, Dict, List
from pydantic import BaseModel, Extra, root_validator
from langchain.prompts.base import (
DEFAULT_FORMATTER_MAPPING,
BasePromptTemplate,
check_valid_template,
)
class Prom... | [
"langchain.prompts.base.check_valid_template"
] | [((1613, 1629), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1627, 1629), False, 'from pydantic import BaseModel, Extra, root_validator\n'), ((1806, 1905), 'langchain.prompts.base.check_valid_template', 'check_valid_template', (["values['template']", "values['template_format']", "values['input_variab... |
from langchain.chat_models import ChatOpenAI
from langchain.chains.summarize import load_summarize_chain
from langchain.text_splitter import TokenTextSplitter
from langchain.docstore.document import Document
# Function to initialize the large language model.
def initialize_llm(openai_api_key, model_name, temperature):... | [
"langchain.chains.summarize.load_summarize_chain",
"langchain.docstore.document.Document",
"langchain.text_splitter.TokenTextSplitter",
"langchain.chat_models.ChatOpenAI"
] | [((331, 424), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': 'openai_api_key', 'model_name': 'model_name', 'temperature': 'temperature'}), '(openai_api_key=openai_api_key, model_name=model_name,\n temperature=temperature)\n', (341, 424), False, 'from langchain.chat_models import ChatOpenAI... |
"""Wrapper around Google's PaLM Chat API."""
from __future__ import annotations
import logging
from typing import TYPE_CHECKING, Any, Callable, Dict, List, Mapping, Optional
from pydantic import BaseModel, root_validator
from tenacity import (
before_sleep_log,
retry,
retry_if_exception_type,
stop_aft... | [
"langchain.schema.ChatMessage",
"langchain.utils.get_from_dict_or_env",
"langchain.schema.ChatResult",
"langchain.schema.HumanMessage",
"langchain.schema.AIMessage"
] | [((792, 819), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (809, 819), False, 'import logging\n'), ((2563, 2598), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (2573, 2598), False, 'from langchain.schema import AIMessag... |
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
'''
@File : create_db.py
@Time : 2023/12/14 10:56:31
@Author : Logan Zou
@Version : 1.0
@Contact : loganzou0421@163.com
@License : (C)Copyright 2017-2018, Liugroup-NLPR-CASIA
@Desc : 知识库搭建
'''
# 首先导入所需第三方库
from langchain.document_loaders import U... | [
"langchain.document_loaders.UnstructuredFileLoader",
"langchain.embeddings.huggingface.HuggingFaceEmbeddings",
"langchain.text_splitter.RecursiveCharacterTextSplitter",
"langchain.vectorstores.Chroma.from_documents",
"langchain.document_loaders.UnstructuredMarkdownLoader"
] | [((2018, 2083), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(500)', 'chunk_overlap': '(150)'}), '(chunk_size=500, chunk_overlap=150)\n', (2048, 2083), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((2164, 2237), 'lang... |
from flask import Flask, request
from flask_restful import Resource, Api, reqparse, abort
from werkzeug.utils import secure_filename
########################################################################
import tempfile
import os
from langchain.document_loaders import DirectoryLoader, PyMuPDFLoader
from langchain.te... | [
"langchain.document_loaders.DirectoryLoader",
"langchain.text_splitter.RecursiveCharacterTextSplitter",
"langchain.chains.ConversationalRetrievalChain.from_llm",
"langchain.vectorstores.Pinecone.from_documents",
"langchain.chat_models.ChatOpenAI",
"langchain.vectorstores.Pinecone.from_existing_index",
"... | [((718, 731), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (729, 731), False, 'from dotenv import load_dotenv\n'), ((753, 785), 'os.environ.get', 'os.environ.get', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (767, 785), False, 'import os\n'), ((809, 843), 'os.environ.get', 'os.environ.get', (['"""... |
from rich import print
from typing import Any, List
from langchain import LLMChain
from langchain.agents import Tool
from langchain.chat_models import AzureChatOpenAI
from LLMAgent.callbackHandler import CustomHandler
from langchain.callbacks import get_openai_callback
from langchain.memory import ConversationBufferMem... | [
"langchain.agents.AgentExecutor.from_agent_and_tools",
"langchain.LLMChain",
"langchain.agents.ZeroShotAgent.create_prompt",
"langchain.agents.ZeroShotAgent",
"langchain.memory.ConversationBufferMemory",
"langchain.callbacks.get_openai_callback",
"langchain.agents.Tool"
] | [((6579, 6594), 'LLMAgent.callbackHandler.CustomHandler', 'CustomHandler', ([], {}), '()\n', (6592, 6594), False, 'from LLMAgent.callbackHandler import CustomHandler\n'), ((6905, 7044), 'langchain.agents.ZeroShotAgent.create_prompt', 'ZeroShotAgent.create_prompt', (['tools'], {'prefix': 'customedPrefix', 'suffix': 'suf... |
from langchain_app.models.http_llm import HTTPBaseLLM
def default_parameters():
return {"temperature": 0, "max_new_tokens": 256, "stop": ["Observation:"]}
def build_llama_base_llm(prompt_url="http://127.0.0.1:8000/prompt", parameters=None):
if parameters is None:
parameters = default_parameters()
... | [
"langchain_app.models.http_llm.HTTPBaseLLM"
] | [((330, 387), 'langchain_app.models.http_llm.HTTPBaseLLM', 'HTTPBaseLLM', ([], {'prompt_url': 'prompt_url', 'parameters': 'parameters'}), '(prompt_url=prompt_url, parameters=parameters)\n', (341, 387), False, 'from langchain_app.models.http_llm import HTTPBaseLLM\n')] |
from langchain.llms import LlamaCpp
from langchain.embeddings import HuggingFaceEmbeddings
from langchain.callbacks.manager import CallbackManager
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
def hf_embeddings():
return HuggingFaceEmbeddings(
model_name = "sentence-transf... | [
"langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler",
"langchain.llms.LlamaCpp",
"langchain.embeddings.HuggingFaceEmbeddings"
] | [((260, 335), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': '"""sentence-transformers/all-mpnet-base-v2"""'}), "(model_name='sentence-transformers/all-mpnet-base-v2')\n", (281, 335), False, 'from langchain.embeddings import HuggingFaceEmbeddings\n'), ((456, 642), 'langchain.... |
import os
import yaml
from types import SimpleNamespace
import openai
import numpy as np
from sklearn.metrics.pairwise import cosine_similarity
from langchain.vectorstores import FAISS
from langchain.embeddings import HuggingFaceEmbeddings
with open("config.yml") as f:
config = yaml.safe_load(f)
config = SimpleN... | [
"langchain.embeddings.HuggingFaceEmbeddings",
"langchain.vectorstores.FAISS.load_local"
] | [((313, 338), 'types.SimpleNamespace', 'SimpleNamespace', ([], {}), '(**config)\n', (328, 338), False, 'from types import SimpleNamespace\n'), ((286, 303), 'yaml.safe_load', 'yaml.safe_load', (['f'], {}), '(f)\n', (300, 303), False, 'import yaml\n'), ((602, 627), 'numpy.argsort', 'np.argsort', (['(-similarities)'], {})... |
from langchain.agents import load_tools
from langchain.agents import initialize_agent
from langchain.agents import AgentType
from langchain_app.models.vicuna_request_llm import VicunaLLM
# First, let's load the language model we're going to use to control the agent.
llm = VicunaLLM()
# Next, let's load some tools to... | [
"langchain_app.models.vicuna_request_llm.VicunaLLM",
"langchain.agents.initialize_agent",
"langchain.agents.load_tools"
] | [((275, 286), 'langchain_app.models.vicuna_request_llm.VicunaLLM', 'VicunaLLM', ([], {}), '()\n', (284, 286), False, 'from langchain_app.models.vicuna_request_llm import VicunaLLM\n'), ((405, 441), 'langchain.agents.load_tools', 'load_tools', (["['python_repl']"], {'llm': 'llm'}), "(['python_repl'], llm=llm)\n", (415, ... |
import logging
import sys
from typing import Callable
from langchain.prompts import MessagesPlaceholder
from langchain.agents import AgentType, AgentExecutor
from langchain.agents import initialize_agent as initialize_agent_base
from langchain.agents.agent_toolkits.base import BaseToolkit
from langchain.chains.base i... | [
"langchain.agents.initialize_agent",
"langchain.prompts.MessagesPlaceholder"
] | [((343, 370), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (360, 370), False, 'import logging\n'), ((2107, 2151), 'langchain.agents.initialize_agent', 'initialize_agent_base', ([], {'agent': 'agent'}), '(agent=agent, **kwargs)\n', (2128, 2151), True, 'from langchain.agents import initia... |
import os
os.environ["LANGCHAIN_TRACING"] = "true"
from langchain import OpenAI
from langchain.agents import initialize_agent, AgentType
from langchain.llms import OpenAI
from langchain.agents import initialize_agent, Tool
from langchain.agents import AgentType
def multiplier(a, b):
return a / b
def parsing_mu... | [
"langchain.agents.initialize_agent",
"langchain.llms.OpenAI",
"langchain.agents.Tool"
] | [((412, 433), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (418, 433), False, 'from langchain.llms import OpenAI\n'), ((826, 917), 'langchain.agents.initialize_agent', 'initialize_agent', (['tools', 'llm'], {'agent': 'AgentType.ZERO_SHOT_REACT_DESCRIPTION', 'verbose': '(True)'})... |
# Copyright 2023 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | [
"langchain.llms.vertexai.VertexAI",
"langchain.memory.ConversationBufferMemory",
"langchain.chains.RetrievalQA.from_chain_type"
] | [((957, 985), 'sys.path.append', 'sys.path.append', (['current_dir'], {}), '(current_dir)\n', (972, 985), False, 'import sys\n'), ((1302, 1333), 'os.getenv', 'os.getenv', (['"""GOOGLE_CLOUD_REGIN"""'], {}), "('GOOGLE_CLOUD_REGIN')\n", (1311, 1333), False, 'import os\n'), ((1347, 1380), 'os.getenv', 'os.getenv', (['"""G... |
import boto3
from botocore.exceptions import ClientError
import json
import langchain
from importlib import reload
from langchain.agents.structured_chat import output_parser
from typing import List
import logging
import os
import sqlalchemy
from sqlalchemy import create_engine
from langchain.docstore.document import Do... | [
"langchain.memory.ConversationBufferMemory",
"langchain.prompts.ChatPromptTemplate.from_messages",
"langchain.tools.python.tool.PythonREPLTool",
"langchain.llms.bedrock.Bedrock",
"langchain.prompts.PromptTemplate",
"langchain_experimental.plan_and_execute.load_chat_planner",
"langchain_experimental.plan... | [((1364, 1397), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (1382, 1397), True, 'import streamlit as st\n'), ((1532, 1553), 'utility.custom_logga.Logger', 'custom_logga.Logger', ([], {}), '()\n', (1551, 1553), False, 'from utility import get_cfn_details, custo... |
from langchain.agents.agent_toolkits import create_python_agent
from langchain.tools.python.tool import PythonREPLTool
from langchain.python import PythonREPL
from langchain.llms.openai import OpenAI
from langchain.agents.agent_types import AgentType
from langchain.chat_models import ChatOpenAI
import os
agent_execut... | [
"langchain.llms.openai.OpenAI",
"langchain.tools.python.tool.PythonREPLTool"
] | [((354, 394), 'langchain.llms.openai.OpenAI', 'OpenAI', ([], {'temperature': '(0.5)', 'max_tokens': '(2000)'}), '(temperature=0.5, max_tokens=2000)\n', (360, 394), False, 'from langchain.llms.openai import OpenAI\n'), ((405, 421), 'langchain.tools.python.tool.PythonREPLTool', 'PythonREPLTool', ([], {}), '()\n', (419, 4... |
# docsGpt.py - Contains the docsGpt functions and classes for document parsing
# Author: Armin Norouzi, Farhad Davaripour
# Contact: https://github.com/Farhad-Davaripour/DocsGPT
# Date created: April 14, 2023
# Last modified: May 3, 2023
# License: MIT License
# Import required modules
import sys
import subprocess
fr... | [
"langchain.text_splitter.CharacterTextSplitter",
"langchain.vectorstores.FAISS.from_texts",
"langchain.llms.OpenAI",
"langchain.embeddings.openai.OpenAIEmbeddings"
] | [((1393, 1431), 'getpass.getpass', 'getpass', (['"""Enter your OpenAI token: ()"""'], {}), "('Enter your OpenAI token: ()')\n", (1400, 1431), False, 'from getpass import getpass\n'), ((1523, 1541), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (1539, 1541), False, 'from langchain... |
import httpx
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain_community.document_loaders import (AsyncChromiumLoader,
AsyncHtmlLoader)
from langchain_community.document_transformers import BeautifulSoupTransformer
from modelscope_agent.t... | [
"langchain.text_splitter.RecursiveCharacterTextSplitter.from_tiktoken_encoder",
"langchain_community.document_loaders.AsyncHtmlLoader",
"langchain_community.document_transformers.BeautifulSoupTransformer"
] | [((359, 387), 'modelscope_agent.tools.register_tool', 'register_tool', (['"""web_browser"""'], {}), "('web_browser')\n", (372, 387), False, 'from modelscope_agent.tools import BaseTool, register_tool\n'), ((857, 919), 'httpx.Client', 'httpx.Client', ([], {'headers': 'self.headers', 'verify': '(False)', 'timeout': '(30.... |
from langchain.chains.base import Chain
from langchain.tools import Tool, BaseTool
def chain_as_tool(chain: Chain, name: str, description: str, **kwargs) -> BaseTool:
"""Converts a chain into a tool."""
return Tool(
name=name,
description=description,
func=chain.invoke,
corouti... | [
"langchain.tools.Tool"
] | [((220, 319), 'langchain.tools.Tool', 'Tool', ([], {'name': 'name', 'description': 'description', 'func': 'chain.invoke', 'coroutine': 'chain.ainvoke'}), '(name=name, description=description, func=chain.invoke, coroutine=chain\n .ainvoke, **kwargs)\n', (224, 319), False, 'from langchain.tools import Tool, BaseTool\n... |
"""Loaders for Prefect."""
import asyncio
import httpx
import os
import shutil
import tempfile
from pathlib import Path
from typing import List
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain_prefect.types import GitHubComment, GitHubIssue
from pre... | [
"langchain_prefect.types.GitHubIssue",
"langchain.docstore.document.Document",
"langchain_prefect.types.GitHubComment"
] | [((4368, 4386), 'tempfile.mkdtemp', 'tempfile.mkdtemp', ([], {}), '()\n', (4384, 4386), False, 'import tempfile\n'), ((944, 974), 'os.environ.get', 'os.environ.get', (['"""GITHUB_TOKEN"""'], {}), "('GITHUB_TOKEN')\n", (958, 974), False, 'import os\n'), ((1493, 1590), 'httpx.get', 'httpx.get', ([], {'url': 'url', 'heade... |
from langchain.agents import AgentType, initialize_agent, load_tools
from langchain.llms import OpenAI
from benchllm import SemanticEvaluator, Test, Tester
tools = load_tools(["serpapi", "llm-math"], llm=OpenAI(temperature=0))
agent = initialize_agent(tools, OpenAI(temperature=0), agent=AgentType.ZERO_SHOT_REACT_DESC... | [
"langchain.llms.OpenAI"
] | [((569, 588), 'benchllm.SemanticEvaluator', 'SemanticEvaluator', ([], {}), '()\n', (586, 588), False, 'from benchllm import SemanticEvaluator, Test, Tester\n'), ((261, 282), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (267, 282), False, 'from langchain.llms import OpenAI\n'), (... |
from dotenv import load_dotenv
from langchain.chains import RetrievalQA
from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings
from langchain import HuggingFacePipeline
from colorama import Fore, Style
import re
from langchain.vectorstores import Chroma
from langchain.docstore.document im... | [
"langchain.docstore.document.Document",
"langchain.vectorstores.Chroma.from_documents",
"langchain.text_splitter.RecursiveCharacterTextSplitter"
] | [((490, 503), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (501, 503), False, 'from dotenv import load_dotenv\n'), ((517, 539), 'os.getenv', 'os.getenv', (['"""TEST_FILE"""'], {}), "('TEST_FILE')\n", (526, 539), False, 'import os\n'), ((559, 588), 'os.getenv', 'os.getenv', (['"""EMBEDDINGS_MODEL"""'], {}), "(... |
"""Wrapper around HuggingFace Pipeline APIs."""
import importlib.util
import logging
from typing import Any, List, Mapping, Optional
from pydantic import BaseModel, Extra
from langchain.llms.base import LLM
from langchain.llms.utils import enforce_stop_tokens
DEFAULT_MODEL_ID = "gpt2"
DEFAULT_TASK = "text-generation... | [
"langchain.llms.utils.enforce_stop_tokens"
] | [((390, 409), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (407, 409), False, 'import logging\n'), ((2546, 2602), 'transformers.AutoTokenizer.from_pretrained', 'AutoTokenizer.from_pretrained', (['model_id'], {}), '(model_id, **_model_kwargs)\n', (2575, 2602), False, 'from transformers import AutoModelFor... |
from langchain.retrievers.self_query.base import SelfQueryRetriever
from langchain.chains.query_constructor.base import AttributeInfo
from datetime import datetime
current_time_iso = datetime.utcnow().isoformat() + "Z"
# example metadat
"""
{
"type": "file_load_gcs",
"attrs": "namespace:edmonbrain",
"source": ... | [
"langchain.retrievers.self_query.base.SelfQueryRetriever.from_llm",
"langchain.chains.query_constructor.base.AttributeInfo"
] | [((1179, 1311), 'langchain.chains.query_constructor.base.AttributeInfo', 'AttributeInfo', ([], {'name': '"""source"""', 'description': '"""The document source url or path to where the document is located"""', 'type': '"""string"""'}), "(name='source', description=\n 'The document source url or path to where the docu... |
import sys
from langchain.embeddings import HuggingFaceEmbeddings
from langchain.vectorstores import FAISS
from langchain.document_loaders import PyPDFLoader, DirectoryLoader, UnstructuredExcelLoader, TextLoader, UnstructuredPowerPointLoader, UnstructuredMarkdownLoader, Docx2txtLoader
from langchain.text_splitter impor... | [
"langchain.document_loaders.DirectoryLoader",
"langchain.embeddings.HuggingFaceEmbeddings",
"langchain.text_splitter.RecursiveCharacterTextSplitter",
"langchain.vectorstores.FAISS.from_documents",
"langchain.vectorstores.FAISS.from_texts"
] | [((506, 552), 'os.path.join', 'os.path.join', (['current_directory', '""".."""', '"""mpnet"""'], {}), "(current_directory, '..', 'mpnet')\n", (518, 552), False, 'import os\n'), ((666, 751), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': 'model_directory', 'model_kwargs': "{'d... |
import os
import re
from typing import List, Optional, Any
from langchain.schema import Document
from langchain.text_splitter import RecursiveCharacterTextSplitter
from loguru import logger
from tqdm import tqdm
from src.config import local_embedding, retrieve_proxy, chunk_overlap, chunk_size, hf_emb_model_name
from ... | [
"langchain.document_loaders.UnstructuredWordDocumentLoader",
"langchain.embeddings.huggingface.HuggingFaceEmbeddings",
"langchain_community.vectorstores.FAISS.from_documents",
"langchain.document_loaders.UnstructuredPowerPointLoader",
"langchain.document_loaders.UnstructuredEPubLoader",
"langchain.schema.... | [((440, 465), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (455, 465), False, 'import os\n'), ((3874, 3910), 'loguru.logger.debug', 'logger.debug', (['"""Loading documents..."""'], {}), "('Loading documents...')\n", (3886, 3910), False, 'from loguru import logger\n'), ((3915, 3956), 'loguru... |
from fastapi import FastAPI
from langchain.chains import RetrievalQA
from langchain.chat_models import ChatOpenAI
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.vectorstores import ElasticVectorSearch
from config import openai_api_key
embedding = OpenAIEmbeddings(openai_api_key=openai_api_key... | [
"langchain.vectorstores.ElasticVectorSearch",
"langchain.chat_models.ChatOpenAI",
"langchain.embeddings.openai.OpenAIEmbeddings"
] | [((274, 321), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'openai_api_key': 'openai_api_key'}), '(openai_api_key=openai_api_key)\n', (290, 321), False, 'from langchain.embeddings.openai import OpenAIEmbeddings\n'), ((328, 444), 'langchain.vectorstores.ElasticVectorSearch', 'ElasticVectorSe... |
from abc import ABC, abstractmethod
from pathlib import Path
from typing import Dict, List
import pandas as pd
import streamlit as st
from langchain.chains import LLMChain
from langchain.prompts.few_shot import FewShotPromptTemplate
from doccano_mini.components import (
display_download_button,
openai_model_f... | [
"langchain.chains.LLMChain"
] | [((763, 785), 'pandas.read_json', 'pd.read_json', (['filepath'], {}), '(filepath)\n', (775, 785), True, 'import pandas as pd\n'), ((921, 984), 'streamlit.experimental_data_editor', 'st.experimental_data_editor', (['df'], {'num_rows': '"""dynamic"""', 'width': '(1000)'}), "(df, num_rows='dynamic', width=1000)\n", (948, ... |
"""This module contains functions for loading and managing vector stores in the Wandbot ingestion system.
The module includes the following functions:
- `load`: Loads the vector store from the specified source artifact path and returns the name of the resulting artifact.
Typical usage example:
project = "wandbot... | [
"langchain.schema.Document"
] | [((944, 964), 'wandbot.utils.get_logger', 'get_logger', (['__name__'], {}), '(__name__)\n', (954, 964), False, 'from wandbot.utils import get_logger, load_index, load_service_context, load_storage_context\n'), ((1677, 1696), 'wandbot.ingestion.config.VectorStoreConfig', 'VectorStoreConfig', ([], {}), '()\n', (1694, 169... |
"""
This module contains the OpenAIImageToText class,
which is a subclass of ChatOpenAI that is specialized for converting images to text.
"""
from langchain_openai import ChatOpenAI
from langchain_core.messages import HumanMessage
class OpenAIImageToText(ChatOpenAI):
"""
A class that uses OpenAI's Chat API... | [
"langchain_core.messages.HumanMessage"
] | [((1233, 1394), 'langchain_core.messages.HumanMessage', 'HumanMessage', ([], {'content': "[{'type': 'text', 'text': 'What is this image showing'}, {'type':\n 'image_url', 'image_url': {'url': image_url, 'detail': 'auto'}}]"}), "(content=[{'type': 'text', 'text': 'What is this image showing'\n }, {'type': 'image_u... |
from langchain.tools import BaseTool
from langchain.tools.render import render_text_description
from langchain_core.language_models.base import LanguageModelLike
from langchain_core.messages import (
AIMessage,
FunctionMessage,
HumanMessage,
SystemMessage,
)
from langgraph.checkpoint import BaseCheckpoi... | [
"langchain_core.messages.AIMessage",
"langchain.tools.render.render_text_description",
"langchain_core.messages.SystemMessage",
"langchain_core.messages.FunctionMessage"
] | [((1121, 1143), 'langchain_core.messages.AIMessage', 'AIMessage', ([], {'content': 'log'}), '(content=log)\n', (1130, 1143), False, 'from langchain_core.messages import AIMessage, FunctionMessage, HumanMessage, SystemMessage\n'), ((2644, 2663), 'langgraph.prebuilt.ToolExecutor', 'ToolExecutor', (['tools'], {}), '(tools... |
# Copyright 2024 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, s... | [
"langchain_google_el_carro.ElCarroEngine.from_instance",
"langchain_google_el_carro.chat_message_history.ElCarroChatMessageHistory"
] | [((1019, 1095), 'langchain_google_el_carro.ElCarroEngine.from_instance', 'ElCarroEngine.from_instance', (['db_host', 'db_port', 'db_name', 'db_user', 'db_password'], {}), '(db_host, db_port, db_name, db_user, db_password)\n', (1046, 1095), False, 'from langchain_google_el_carro import ElCarroEngine\n'), ((1238, 1349), ... |
from typing import Optional, Type
from langchain.callbacks.manager import CallbackManagerForToolRun
from langchain_community.utilities import OpenWeatherMapAPIWrapper
from pydantic import BaseModel, Field
from exceptions import ToolEnvKeyException
from tools.base import BaseTool
class OpenWeatherMapSchema(BaseModel... | [
"langchain_community.utilities.OpenWeatherMapAPIWrapper"
] | [((340, 402), 'pydantic.Field', 'Field', (['...'], {'description': '"""The search query for OpenWeatherMap."""'}), "(..., description='The search query for OpenWeatherMap.')\n", (345, 402), False, 'from pydantic import BaseModel, Field\n'), ((1406, 1477), 'langchain_community.utilities.OpenWeatherMapAPIWrapper', 'OpenW... |
from textwrap import dedent
from langchain import OpenAI
from langchain.schema import BaseModel
from utils import format_prompt_components_without_tools
def extract_first_message(message: str) -> str:
"""The LLM can continue the conversation from the recipient. So extract just the first line."""
return mes... | [
"langchain.OpenAI"
] | [((627, 656), 'textwrap.dedent', 'dedent', (['inspirational_thought'], {}), '(inspirational_thought)\n', (633, 656), False, 'from textwrap import dedent\n'), ((912, 981), 'utils.format_prompt_components_without_tools', 'format_prompt_components_without_tools', (['ai_settings', 'contact_settings'], {}), '(ai_settings, c... |
"""VectorStore wrapper around a Postgres/PGVector database."""
from __future__ import annotations
import enum
import logging
import uuid
from typing import Any, Dict, Iterable, List, Optional, Tuple, Type
import sqlalchemy
from pgvector.sqlalchemy import Vector
from sqlalchemy.dialects.postgresql import JSON, UUID
fr... | [
"langchain.utils.get_from_dict_or_env",
"langchain.docstore.document.Document"
] | [((593, 611), 'sqlalchemy.orm.declarative_base', 'declarative_base', ([], {}), '()\n', (609, 611), False, 'from sqlalchemy.orm import Session, declarative_base, relationship\n'), ((929, 965), 'sqlalchemy.Column', 'sqlalchemy.Column', (['sqlalchemy.String'], {}), '(sqlalchemy.String)\n', (946, 965), False, 'import sqlal... |
import tempfile
import time
import os
from utils import compute_sha1_from_file
from langchain.schema import Document
import streamlit as st
from langchain.text_splitter import RecursiveCharacterTextSplitter
from typing import List
from sqlite3 import Connection
from verse.sqlite_helper import *
def update_metadata(co... | [
"langchain.text_splitter.RecursiveCharacterTextSplitter.from_tiktoken_encoder",
"langchain.schema.Document"
] | [((1152, 1175), 'time.strftime', 'time.strftime', (['"""%Y%m%d"""'], {}), "('%Y%m%d')\n", (1165, 1175), False, 'import time\n'), ((1468, 1492), 'os.remove', 'os.remove', (['tmp_file.name'], {}), '(tmp_file.name)\n', (1477, 1492), False, 'import os\n'), ((1679, 1812), 'langchain.text_splitter.RecursiveCharacterTextSplit... |
import json
import logging
from typing import Any, Dict, Iterator, List, Optional
import requests
from langchain.callbacks.manager import CallbackManagerForLLMRun
from langchain.llms.base import LLM
from langchain.pydantic_v1 import Field
from langchain.schema.output import GenerationChunk
logger = logging.getLogger... | [
"langchain.pydantic_v1.Field",
"langchain.schema.output.GenerationChunk"
] | [((303, 330), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (320, 330), False, 'import logging\n'), ((1278, 1308), 'langchain.pydantic_v1.Field', 'Field', (['(True)'], {'alias': '"""do_sample"""'}), "(True, alias='do_sample')\n", (1283, 1308), False, 'from langchain.pydantic_v1 import Fi... |
# imports
from loguru import logger
# LLM modules
from langchain_community.llms.huggingface_hub import HuggingFaceHub
from langchain_community.llms.ollama import Ollama
from langchain_openai import ChatOpenAI, AzureChatOpenAI
from langchain.callbacks.manager import CallbackManager
from langchain.callbacks.streaming_std... | [
"langchain_openai.AzureChatOpenAI",
"langchain_openai.ChatOpenAI",
"langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler",
"langchain_community.llms.huggingface_hub.HuggingFaceHub"
] | [((1610, 1675), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'client': 'None', 'model': 'self.llm_model_type', 'temperature': '(0)'}), '(client=None, model=self.llm_model_type, temperature=0)\n', (1620, 1675), False, 'from langchain_openai import ChatOpenAI, AzureChatOpenAI\n'), ((2163, 2272), 'langchain_communit... |
from typing import List
from langchain.embeddings import OpenAIEmbeddings
from langchain.vectorstores import Chroma
from langchain_core.documents import Document
from dotenv import load_dotenv
from themind.llm.func_instraction import instruct
from pydantic import BaseModel
import csv
from themind.vectorstores.chunking.... | [
"langchain.embeddings.OpenAIEmbeddings"
] | [((657, 675), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (673, 675), False, 'from langchain.embeddings import OpenAIEmbeddings\n')] |
import re
import time
import copy
import random
import numpy as np
import multiprocessing
import matplotlib.pyplot as plt
import modules.prompts as prompts
from langchain import PromptTemplate
from shapely.ops import substring
from shapely.geometry import Polygon, box, Point, LineString
class WallObjectGenerator():
... | [
"langchain.PromptTemplate"
] | [((704, 850), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['room_type', 'wall_height', 'floor_objects', 'wall_objects']", 'template': 'prompts.wall_object_constraints_prompt'}), "(input_variables=['room_type', 'wall_height', 'floor_objects',\n 'wall_objects'], template=prompts.wall_object... |
from langchain_openai import ChatOpenAI
from langchain_core.output_parsers import StrOutputParser
from langchain.prompts import PromptTemplate
from langchain.prompts.chat import ChatPromptTemplate
from config.config import OPENAI_API_KEY
from game.poker import PokerGameManager
from db.db_utils import DatabaseManager
im... | [
"langchain_openai.ChatOpenAI",
"langchain_core.output_parsers.StrOutputParser",
"langchain.prompts.chat.ChatPromptTemplate.from_messages"
] | [((456, 489), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': 'model_name'}), '(model_name=model_name)\n', (466, 489), False, 'from langchain_openai import ChatOpenAI\n'), ((514, 531), 'langchain_core.output_parsers.StrOutputParser', 'StrOutputParser', ([], {}), '()\n', (529, 531), False, 'from langcha... |
import logging
from typing import Any, Dict, List, Optional
from langchain.callbacks.manager import CallbackManagerForLLMRun
from langchain.llms.base import LLM
from langchain.llms.utils import enforce_stop_tokens
from langchain.pydantic_v1 import Extra, root_validator
from langchain.schema import Generation, LLMResul... | [
"langchain.llms.utils.enforce_stop_tokens",
"langchain.utils.get_from_dict_or_env",
"langchain.schema.Generation",
"langchain.schema.LLMResult",
"langchain.pydantic_v1.root_validator"
] | [((381, 408), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (398, 408), False, 'import logging\n'), ((1472, 1488), 'langchain.pydantic_v1.root_validator', 'root_validator', ([], {}), '()\n', (1486, 1488), False, 'from langchain.pydantic_v1 import Extra, root_validator\n'), ((1702, 1753),... |
import re
from typing import Any, Dict, List, Optional
from langchain_core.load import loads, dumps
from langchain_community.chat_models import (
ChatAnthropic,
ChatAnyscale,
ChatBaichuan,
QianfanChatEndpoint,
BedrockChat,
ChatDatabricks,
ChatDeepInfra,
ErnieBotChat,
ChatEverlyAI,
... | [
"langchain_core.load.dumps"
] | [((9560, 9597), 're.search', 're.search', (['f"""{pattern}=\'(.*?)\'"""', 'text'], {}), '(f"{pattern}=\'(.*?)\'", text)\n', (9569, 9597), False, 'import re\n'), ((2545, 2562), 'langchain_core.load.dumps', 'dumps', (['serialized'], {}), '(serialized)\n', (2550, 2562), False, 'from langchain_core.load import loads, dumps... |
"""This example shows how to use the ChatGPT API
with LangChain to answer questions about Prefect."""
from langchain.embeddings.openai import OpenAIEmbeddings
from langchain.vectorstores import Chroma
from langchain.text_splitter import CharacterTextSplitter
from langchain.chains import ChatVectorDBChain
from langchai... | [
"langchain.text_splitter.CharacterTextSplitter",
"langchain_prefect.loaders.GitHubRepoLoader",
"langchain.prompts.chat.SystemMessagePromptTemplate.from_template",
"langchain.chat_models.ChatOpenAI",
"langchain.vectorstores.Chroma.from_documents",
"langchain.prompts.chat.HumanMessagePromptTemplate.from_tem... | [((680, 735), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {'chunk_size': '(1000)', 'chunk_overlap': '(0)'}), '(chunk_size=1000, chunk_overlap=0)\n', (701, 735), False, 'from langchain.text_splitter import CharacterTextSplitter\n'), ((803, 821), 'langchain.embeddings.openai.OpenAIEmbed... |
from dotenv import load_dotenv
load_dotenv()
import os
from langchain.llms import OpenAI
from langchain.chat_models import ChatOpenAI
from langchain.prompts import (
PromptTemplate,
)
from langchain.chains import ConversationChain
from langchain.memory import ConversationBufferMemory
from langchain.agents import A... | [
"langchain.agents.ConversationalChatAgent.from_llm_and_tools",
"langchain.agents.AgentExecutor.from_agent_and_tools",
"langchain.memory.ConversationBufferMemory",
"langchain.chat_models.ChatOpenAI"
] | [((31, 44), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (42, 44), False, 'from dotenv import load_dotenv\n'), ((574, 601), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (583, 601), False, 'import os\n'), ((1087, 1183), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([],... |
from __future__ import annotations
from abc import abstractmethod
from typing import TYPE_CHECKING, Any, Dict, List, Sequence
from langchain.load.serializable import Serializable
from langchain.pydantic_v1 import Field
if TYPE_CHECKING:
from langchain.prompts.chat import ChatPromptTemplate
def get_buffer_strin... | [
"langchain.pydantic_v1.Field",
"langchain.prompts.chat.ChatPromptTemplate"
] | [((2151, 2178), 'langchain.pydantic_v1.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (2156, 2178), False, 'from langchain.pydantic_v1 import Field\n'), ((2610, 2645), 'langchain.prompts.chat.ChatPromptTemplate', 'ChatPromptTemplate', ([], {'messages': '[self]'}), '(messages=[self])\n',... |
import logging
from typing import Any, Dict, List, Mapping, Optional
from langchain.callbacks.manager import CallbackManagerForLLMRun
from langchain.llms.base import LLM
from langchain.llms.utils import enforce_stop_tokens
from langchain.pydantic_v1 import BaseModel, Extra, Field, root_validator
from langchain.utils i... | [
"langchain.llms.utils.enforce_stop_tokens",
"langchain.pydantic_v1.Field",
"langchain.pydantic_v1.root_validator",
"langchain.utils.get_from_dict_or_env"
] | [((357, 384), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (374, 384), False, 'import logging\n'), ((1004, 1031), 'langchain.pydantic_v1.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1009, 1031), False, 'from langchain.pydantic_v1 import BaseModel, Ext... |
from typing import Optional, Type
import streamlit as st
import tldextract
import whois
import whoisit
from langchain.agents import AgentType, Tool, initialize_agent
from langchain.chat_models import ChatOpenAI
from langchain.tools import BaseTool
from langchain.tools.ddg_search import DuckDuckGoSearchRun
from pydanti... | [
"langchain.tools.ddg_search.DuckDuckGoSearchRun",
"langchain.agents.initialize_agent",
"langchain.agents.Tool",
"langchain.chat_models.ChatOpenAI"
] | [((363, 390), 'streamlit.title', 'st.title', (['"""TakedownGPT ⬇️🤖"""'], {}), "('TakedownGPT ⬇️🤖')\n", (371, 390), True, 'import streamlit as st\n'), ((434, 467), 'streamlit.sidebar.header', 'st.sidebar.header', (['"""How to Use 📝"""'], {}), "('How to Use 📝')\n", (451, 467), True, 'import streamlit as st\n'), ((468... |
# -*- coding: utf-8 -*-
import os
import re
from typing import List, Union, Dict, Tuple, Any, Optional
from langchain.agents import Tool, AgentExecutor, AgentOutputParser, load_tools
from langchain.tools.base import BaseTool
from langchain.prompts import StringPromptTemplate
from langchain import OpenAI, GoogleSearchAP... | [
"langchain.agents.AgentExecutor.from_agent_and_tools",
"langchain.LLMChain",
"langchain.schema.AgentAction",
"langchain.chat_models.ChatOpenAI",
"langchain.schema.AgentFinish"
] | [((473, 493), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (488, 493), False, 'import sys\n'), ((494, 515), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (509, 515), False, 'import sys\n'), ((1325, 1340), 'utils.parser.get_arguments', 'get_arguments', ([], {}), '()\n', (13... |
import sqlite3
import pandas as pd
import os
import json
import warnings
from langchain import SQLDatabase
from langchain.docstore.document import Document
from langchain.vectorstores import Chroma
from langchain.embeddings import HuggingFaceEmbeddings
from sqlalchemy import exc
from sqlalchemy.exc import SAWarning
... | [
"langchain.embeddings.HuggingFaceEmbeddings"
] | [((320, 373), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {'category': 'SAWarning'}), "('ignore', category=SAWarning)\n", (343, 373), False, 'import warnings\n'), ((973, 1033), 'src.data.setup.db_setup_functions.build_schema_info', 'build_schema_info', ([], {'filepath': 'data_directory', 'f... |
import asyncio
from functools import partial
from typing import (
Any,
List,
Mapping,
Optional,
)
from ai21.models import CompletionsResponse
from langchain_core.callbacks import (
AsyncCallbackManagerForLLMRun,
CallbackManagerForLLMRun,
)
from langchain_core.language_models import BaseLLM
from... | [
"langchain_core.outputs.LLMResult"
] | [((4179, 4236), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (4188, 4236), False, 'from langchain_core.outputs import Generation, LLMResult\n'), ((4626, 4659), 'functools.partial', 'partial', (['s... |
from langchain.llms import OpenAI
from callback import MyCallbackHandler
from langchain.callbacks.base import BaseCallbackManager
class QaLlm():
def __init__(self) -> None:
manager = BaseCallbackManager([MyCallbackHandler()])
self.llm = OpenAI(temperature=0, callback_manager=manager, model_name="g... | [
"langchain.llms.OpenAI"
] | [((259, 334), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'callback_manager': 'manager', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, callback_manager=manager, model_name='gpt-3.5-turbo')\n", (265, 334), False, 'from langchain.llms import OpenAI\n'), ((218, 237), 'callback.MyCallbackHandle... |
from langchain.llms import OpenAI
from langchain.chat_models import ChatOpenAI
from apikey import (
apikey,
google_search,
google_cse,
serp,
aws_access_key,
aws_secret_key,
aws_region,
)
import os
from typing import Dict
from langchain.prompts import PromptTemplate
from langchain.chains impo... | [
"langchain.memory.ConversationBufferMemory",
"langchain.llms.OpenAI",
"langchain.chains.LLMChain",
"langchain.utilities.GoogleSearchAPIWrapper",
"langchain.prompts.PromptTemplate"
] | [((765, 835), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0.3)', 'max_tokens': '(100)', 'model_name': '"""text-davinci-003"""'}), "(temperature=0.3, max_tokens=100, model_name='text-davinci-003')\n", (771, 835), False, 'from langchain.llms import OpenAI\n'), ((860, 886), 'langchain.memory.ConversationBuff... |
"""
DATE: 2023/5/28
AUTHOR: ZLYANG
CONTACT: zhlyang95@hotmail.com
"""
### define tools ###
import requests
import io
import base64
import os
from PIL import Image
from typing import Optional
from langchain.tools import BaseTool
from langchain.callbacks.manager import (
AsyncCallbackManagerForToo... | [
"langchain.LLMChain",
"langchain.PromptTemplate.from_template"
] | [((757, 785), 'random.randint', 'random.randint', (['(32768)', '(65536)'], {}), '(32768, 65536)\n', (771, 785), False, 'import re, random\n'), ((2323, 2364), 'langchain.PromptTemplate.from_template', 'PromptTemplate.from_template', (['qa_template'], {}), '(qa_template)\n', (2351, 2364), False, 'from langchain import LL... |
from langchain.retrievers import AmazonKendraRetriever
from langchain.chains import ConversationalRetrievalChain
from langchain import SagemakerEndpoint
from langchain.llms.sagemaker_endpoint import LLMContentHandler
from langchain.prompts import PromptTemplate
import sys
import json
import os
class bcolors:
HEAD... | [
"langchain.SagemakerEndpoint",
"langchain.chains.ConversationalRetrievalChain.from_llm",
"langchain.retrievers.AmazonKendraRetriever",
"langchain.prompts.PromptTemplate.from_template",
"langchain.prompts.PromptTemplate"
] | [((1327, 1604), 'langchain.SagemakerEndpoint', 'SagemakerEndpoint', ([], {'endpoint_name': 'endpoint_name', 'region_name': 'region', 'model_kwargs': "{'temperature': 0.8, 'max_new_tokens': 512, 'do_sample': True, 'top_p': 0.9,\n 'repetition_penalty': 1.03, 'stop': ['\\nUser:', '<|endoftext|>', '</s>']}", 'content_ha... |
#Make sure to install the following packages: dlt, langchain, duckdb, python-dotenv, openai, weaviate-client
import dlt
from langchain import PromptTemplate, LLMChain
from langchain.chains.openai_functions import create_structured_output_chain
from langchain.chat_models import ChatOpenAI
from langchain.document_loader... | [
"langchain.LLMChain",
"langchain.chains.openai_functions.create_structured_output_chain",
"langchain.prompts.HumanMessagePromptTemplate.from_template",
"langchain.retrievers.WeaviateHybridSearchRetriever",
"langchain.LLMMathChain.from_llm",
"langchain.schema.HumanMessage",
"langchain.schema.SystemMessag... | [((741, 754), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (752, 754), False, 'from dotenv import load_dotenv\n'), ((848, 866), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (864, 866), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((1129, 1146), 'langchain.do... |
import sys
import os
sys.path.append(os.path.dirname(os.path.abspath(__file__)) + '/../')
import asyncio
from argparse import Namespace
from models.loader.args import parser
from models.loader import LoaderCheckPoint
from langchain.agents import initialize_agent, Tool
from langchain.agents import AgentType
import mo... | [
"langchain.agents.AgentExecutor.from_agent_and_tools",
"langchain.memory.ConversationBufferMemory",
"langchain.agents.Tool",
"langchain.memory.ReadOnlySharedMemory",
"langchain.chains.LLMChain",
"langchain.prompts.PromptTemplate"
] | [((1027, 1054), 'models.loader.LoaderCheckPoint', 'LoaderCheckPoint', (['args_dict'], {}), '(args_dict)\n', (1043, 1054), False, 'from models.loader import LoaderCheckPoint\n'), ((1075, 1093), 'models.shared.loaderLLM', 'shared.loaderLLM', ([], {}), '()\n', (1091, 1093), True, 'import models.shared as shared\n'), ((125... |
import logging
from time import sleep
from langchain.llms import OpenAI
from scrapy import Request, Spider
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
from conf import (
CONNECTION_REQUEST_LLM_PROMPT,
DEFAULT_CONNECTION_MESSAGE,
MAX_PROFILES_TO_CONNECT,
MAX_PROFILES_... | [
"langchain.llms.OpenAI",
"langchain.PromptTemplate.from_template"
] | [((689, 716), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (706, 716), False, 'import logging\n'), ((1186, 1271), 'linkedin.integrations.selenium.get_by_xpath_or_none', 'get_by_xpath_or_none', (['driver', '"""//button[@aria-label="Got it"]"""'], {'wait_timeout': '(0.5)'}), '(driver, \'/... |
import streamlit as st
import os
from azure.storage.blob import BlobServiceClient, BlobClient, ContainerClient
from PyPDF2 import PdfReader
# Import
#import textwrap
import openai
from langchain.llms import AzureOpenAI, OpenAI
from langchain.embeddings import OpenAIEmbeddings
from llama_index.vector_stores import Redis... | [
"langchain.embeddings.OpenAIEmbeddings",
"langchain.llms.OpenAI",
"langchain.llms.AzureOpenAI"
] | [((558, 616), 'logging.basicConfig', 'logging.basicConfig', ([], {'stream': 'sys.stdout', 'level': 'logging.INFO'}), '(stream=sys.stdout, level=logging.INFO)\n', (577, 616), False, 'import logging\n'), ((744, 780), 'os.getenv', 'os.getenv', (['"""REDIS_HOST"""', '"""localhost"""'], {}), "('REDIS_HOST', 'localhost')\n",... |
from langchain.chat_models import ChatOpenAI
from langchain.schema import HumanMessage, SystemMessage
from whenx.models.team import Team
from whenx.models.scout import Scout
from whenx.models.sentinel import Sentinel
from whenx.models.soldier import Soldier
import re
from whenx.database import db
class Captain:
... | [
"langchain.schema.SystemMessage",
"langchain.schema.HumanMessage",
"langchain.chat_models.ChatOpenAI"
] | [((575, 587), 'whenx.database.db.add', 'db.add', (['team'], {}), '(team)\n', (581, 587), False, 'from whenx.database import db\n'), ((596, 607), 'whenx.database.db.commit', 'db.commit', ([], {}), '()\n', (605, 607), False, 'from whenx.database import db\n'), ((624, 675), 'whenx.models.scout.Scout', 'Scout', ([], {'inst... |
import json
import re
from langchain.chains import RetrievalQA
from utils.functions import find_nth, remove_extra_heading, add_json_characters, Timeout
from langchain import LLMChain
from langchain.chat_models import ChatOpenAI
def section_schemas(heading, keyword, format_instructions, retriever, prompt):
ch... | [
"langchain.LLMChain",
"langchain.chains.RetrievalQA.from_chain_type",
"langchain.chat_models.ChatOpenAI"
] | [((325, 387), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model_name': '"""gpt-3.5-turbo-16k-0613"""'}), "(temperature=0, model_name='gpt-3.5-turbo-16k-0613')\n", (335, 387), False, 'from langchain.chat_models import ChatOpenAI\n'), ((433, 466), 'langchain.LLMChain', 'LLMChain', ([], ... |
"""Experiment with different models."""
from __future__ import annotations
from typing import List, Optional, Sequence
from langchain_core.language_models.llms import BaseLLM
from langchain_core.prompts.prompt import PromptTemplate
from langchain_core.utils.input import get_color_mapping, print_text
from langchain.... | [
"langchain_core.utils.input.print_text",
"langchain_core.utils.input.get_color_mapping",
"langchain_core.prompts.prompt.PromptTemplate",
"langchain.chains.llm.LLMChain"
] | [((1752, 1782), 'langchain_core.utils.input.get_color_mapping', 'get_color_mapping', (['chain_range'], {}), '(chain_range)\n', (1769, 1782), False, 'from langchain_core.utils.input import get_color_mapping, print_text\n'), ((2307, 2370), 'langchain_core.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'input_var... |
from typing import Any, List, Optional, Sequence, Tuple
from langchain_core._api import deprecated
from langchain_core.agents import AgentAction
from langchain_core.callbacks import BaseCallbackManager
from langchain_core.language_models import BaseLanguageModel
from langchain_core.prompts import BasePromptTemplate
fr... | [
"langchain_core.prompts.chat.SystemMessagePromptTemplate.from_template",
"langchain.chains.llm.LLMChain",
"langchain.agents.utils.validate_tools_single_input",
"langchain_core.pydantic_v1.Field",
"langchain_core.prompts.chat.ChatPromptTemplate",
"langchain_core.prompts.chat.HumanMessagePromptTemplate.from... | [((915, 985), 'langchain_core._api.deprecated', 'deprecated', (['"""0.1.0"""'], {'alternative': '"""create_react_agent"""', 'removal': '"""0.2.0"""'}), "('0.1.0', alternative='create_react_agent', removal='0.2.0')\n", (925, 985), False, 'from langchain_core._api import deprecated\n'), ((1072, 1111), 'langchain_core.pyd... |
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. 2023
# SPDX-License-Identifier: Apache-2.0
from typing import Any, Dict, List, Optional
from langchain.agents import tool
from langchain.chains.base import Chain
from langchain.chains import LLMChain
from langchain import PromptTemplate
from langcha... | [
"langchain.chains.LLMChain",
"langchain.callbacks.manager.CallbackManagerForChainRun.get_noop_manager"
] | [((1178, 1216), 'chainlit.context.context.session.emit', 'context.session.emit', (['"""view"""', 'entityId'], {}), "('view', entityId)\n", (1198, 1216), False, 'from chainlit.context import context\n'), ((2370, 2412), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=p... |
from langchain.retrievers import AmazonKendraRetriever
from langchain.chains import RetrievalQA
from langchain import OpenAI
from langchain.prompts import PromptTemplate
from langchain import SagemakerEndpoint
from langchain.llms.sagemaker_endpoint import LLMContentHandler
import json
import os
def build_chain():
... | [
"langchain.SagemakerEndpoint",
"langchain.retrievers.AmazonKendraRetriever",
"langchain.prompts.PromptTemplate",
"langchain.chains.RetrievalQA.from_chain_type"
] | [((1839, 1906), 'langchain.retrievers.AmazonKendraRetriever', 'AmazonKendraRetriever', ([], {'index_id': 'kendra_index_id', 'region_name': 'region'}), '(index_id=kendra_index_id, region_name=region)\n', (1860, 1906), False, 'from langchain.retrievers import AmazonKendraRetriever\n'), ((2373, 2458), 'langchain.prompts.P... |
from typing import List, Optional
from langchain.docstore.document import Document
from langchain.document_loaders.base import BaseLoader
from langchain.utilities.arxiv import ArxivAPIWrapper
class ArxivLoader(BaseLoader):
"""Load a query result from `Arxiv`.
The loader converts the original PDF format into... | [
"langchain.utilities.arxiv.ArxivAPIWrapper"
] | [((863, 971), 'langchain.utilities.arxiv.ArxivAPIWrapper', 'ArxivAPIWrapper', ([], {'load_max_docs': 'self.load_max_docs', 'load_all_available_meta': 'self.load_all_available_meta'}), '(load_max_docs=self.load_max_docs, load_all_available_meta=\n self.load_all_available_meta)\n', (878, 971), False, 'from langchain.u... |
# flake8: noqa
from langchain_core.prompts.prompt import PromptTemplate
API_URL_PROMPT_TEMPLATE = """You are given the below API Documentation:
{api_docs}
Using this documentation, generate the full API url to call for answering the user question.
You should build the API url in order to get a response that is as shor... | [
"langchain_core.prompts.prompt.PromptTemplate"
] | [((542, 637), 'langchain_core.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['api_docs', 'question']", 'template': 'API_URL_PROMPT_TEMPLATE'}), "(input_variables=['api_docs', 'question'], template=\n API_URL_PROMPT_TEMPLATE)\n", (556, 637), False, 'from langchain_core.prompts.prompt imp... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.