code
stringlengths
141
79.4k
apis
listlengths
1
23
extract_api
stringlengths
126
73.2k
"""Test Upstash Redis cache functionality.""" import uuid import pytest import langchain from langchain.cache import UpstashRedisCache from langchain.schema import Generation, LLMResult from tests.unit_tests.llms.fake_chat_model import FakeChatModel from tests.unit_tests.llms.fake_llm import FakeLLM URL = "<UPSTASH_...
[ "langchain.llm_cache.clear", "langchain.schema.Generation", "langchain.llm_cache.redis.flushall", "langchain.llm_cache.redis.pttl", "langchain.llm_cache._key", "langchain.llm_cache.lookup" ]
[((436, 473), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (456, 473), False, 'import pytest\n'), ((809, 846), 'pytest.mark.requires', 'pytest.mark.requires', (['"""upstash_redis"""'], {}), "('upstash_redis')\n", (829, 846), False, 'import pytest\n'), ((2491, 252...
import os import langchain import streamlit as st from collections import defaultdict from urllib.error import URLError from dotenv import load_dotenv load_dotenv() if os.environ.get("QNA_DEBUG") == "true": langchain.debug = True from qna.llm import make_qna_chain, get_llm from qna.db import get_cache, get_vecto...
[ "langchain.llm_cache.clear" ]
[((152, 165), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (163, 165), False, 'from dotenv import load_dotenv\n'), ((170, 197), 'os.environ.get', 'os.environ.get', (['"""QNA_DEBUG"""'], {}), "('QNA_DEBUG')\n", (184, 197), False, 'import os\n'), ((489, 500), 'qna.db.get_cache', 'get_cache', ([], {}), '()\n', (...
''' Create Vector Store from all documents in a folder, currently supports .pptx, .docx, .pdf files. Created by Ric Zhou on 2021-03-27 ''' from langchain.document_loaders import (UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredPDFLoader) import glob import langchain.text_splitte...
[ "langchain.document_loaders.UnstructuredWordDocumentLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.UnstructuredPowerPointLoader", "langchain.vectorstores.FAISS.save_local", "langchain.document_loaders.PyPDFLoader", "langchain.embeddings.OpenAIEmbeddings" ]
[((604, 617), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (615, 617), False, 'from dotenv import load_dotenv\n'), ((618, 633), 'GlobalClasses.GlobalContext', 'GlobalContext', ([], {}), '()\n', (631, 633), False, 'from GlobalClasses import GlobalContext\n'), ((939, 1017), 'langchain.text_splitter.RecursiveCha...
import os import key import tabulate # Set API key os.environ["OPENAI_API_KEY"] = key.OPENAI_API_KEY # Import langchain from langchain.chains import RetrievalQA from langchain.chat_models import ChatOpenAI from langchain.document_loaders import CSVLoader from langchain.indexes import VectorstoreIndexCreator from langc...
[ "langchain.document_loaders.CSVLoader", "langchain.indexes.VectorstoreIndexCreator", "langchain.chat_models.ChatOpenAI" ]
[((465, 508), 'langchain.document_loaders.CSVLoader', 'CSVLoader', ([], {'file_path': 'file', 'encoding': '"""utf-8"""'}), "(file_path=file, encoding='utf-8')\n", (474, 508), False, 'from langchain.document_loaders import CSVLoader\n'), ((708, 735), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature':...
import openai import langchain as lc from langchain.llms import OpenAI import gradio as gr # 设置OpenAI API密钥 openai.api_key = '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' # 初始化Langchain的OpenAI LLM llm = OpenAI(api_key=openai.api_key) # 定义一个函数来处理上传的文档并生成响应 def process_document(document): # 这里可以添加代码来处理文档,...
[ "langchain.llms.OpenAI" ]
[((213, 243), 'langchain.llms.OpenAI', 'OpenAI', ([], {'api_key': 'openai.api_key'}), '(api_key=openai.api_key)\n', (219, 243), False, 'from langchain.llms import OpenAI\n'), ((508, 536), 'gradio.inputs.File', 'gr.inputs.File', ([], {'label': '"""上传文档"""'}), "(label='上传文档')\n", (522, 536), True, 'import gradio as gr\n'...
import streamlit as st # import langchain # langchain.debug=True from crewai import Agent from utils.tools import ( retrieval_tool, search_tool, get_current_stock_price ) from langchain_openai import ChatOpenAI # from dotenv import load_dotenv # load_dotenv() # openai_api_key = os.environ.get("OPENAI_API...
[ "langchain_openai.ChatOpenAI" ]
[((405, 499), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo-16k"""', 'openai_api_key': "ss.configurations['openai_api_key']"}), "(model='gpt-3.5-turbo-16k', openai_api_key=ss.configurations[\n 'openai_api_key'])\n", (415, 499), False, 'from langchain_openai import ChatOpenAI\n'), ((57...
import langchain import os import openai from InstructorEmbedding import INSTRUCTOR from langchain_community.embeddings import HuggingFaceInstructEmbeddings from langchain_community.vectorstores import Chroma from langchain.chains import VectorDBQA, RetrievalQA, ConversationalRetrievalChain from custom_retrival ...
[ "langchain_community.llms.HuggingFaceHub" ]
[((2377, 2404), 'streamlit.chat_message', 'st.chat_message', (['"""assitant"""'], {}), "('assitant')\n", (2392, 2404), True, 'import streamlit as st\n'), ((2498, 2519), 'streamlit.chat_input', 'st.chat_input', (['"""user"""'], {}), "('user')\n", (2511, 2519), True, 'import streamlit as st\n'), ((1212, 1338), 'langchain...
import os import pandas as pd import math from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import Chroma from langchain.text_splitter import CharacterTextSplitter from langchain import OpenAI, VectorDBQA, OpenAI from langchain.llms import OpenAIChat from langchain.document_loaders im...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.vectorstores.Chroma.from_documents", "langchain.document_loaders.DataFrameLoader", "langchain.OpenAI", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((527, 555), 'sys.modules.pop', 'sys.modules.pop', (['"""pysqlite3"""'], {}), "('pysqlite3')\n", (542, 555), False, 'import sys\n'), ((558, 587), 'streamlit.title', 'st.title', (['"""GPT module (TEST)"""'], {}), "('GPT module (TEST)')\n", (566, 587), True, 'import streamlit as st\n'), ((606, 660), 'streamlit.text_inpu...
# Python built-in module import os import time import json # Python installed module import tiktoken import langchain from spacy.lang.en import English class SentencizerSplitter(object): def __init__(self, config_dict): self.total_tokens = config_dict["embedding"]["total_tokens"] self.approx_tota...
[ "langchain.schema.document.Document" ]
[((511, 520), 'spacy.lang.en.English', 'English', ([], {}), '()\n', (518, 520), False, 'from spacy.lang.en import English\n'), ((586, 653), 'tiktoken.encoding_for_model', 'tiktoken.encoding_for_model', (["config_dict['embedding']['model_name']"], {}), "(config_dict['embedding']['model_name'])\n", (613, 653), False, 'im...
import os import json import openai from utils import * import random import langchain from langchain import PromptTemplate from langchain.llms import OpenAI, OpenAIChat from langchain import LLMChain from re import compile from datetime import datetime from typing import NamedTuple from openai import Embedding #set ...
[ "langchain.llms.OpenAIChat", "langchain.LLMChain", "langchain.PromptTemplate" ]
[((1869, 1883), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1881, 1883), False, 'from datetime import datetime\n'), ((2826, 2890), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_text', 'input_variables': "['Memory']"}), "(template=prompt_text, input_variables=['Memory'])\n", (28...
import langchain class ChromaDB: def __init__(self, path): self.db = langchain.ChromaDB(path) def index_url(self, url): self.db.index_url(url)
[ "langchain.ChromaDB" ]
[((82, 106), 'langchain.ChromaDB', 'langchain.ChromaDB', (['path'], {}), '(path)\n', (100, 106), False, 'import langchain\n')]
import langchain_helper as lch import streamlit as st st.title("Pets name generator") animal_type = st.sidebar.selectbox("What is your pet?", ("Cat", "Dog", "Cow", "Hamnster")) pet_color = st.sidebar.text_area(f"What color is your {animal_type}?", max_chars=15) if pet_color: response = lch.generate_pet_names(an...
[ "langchain_helper.langchain_agent", "langchain_helper.generate_pet_names" ]
[((55, 86), 'streamlit.title', 'st.title', (['"""Pets name generator"""'], {}), "('Pets name generator')\n", (63, 86), True, 'import streamlit as st\n'), ((102, 178), 'streamlit.sidebar.selectbox', 'st.sidebar.selectbox', (['"""What is your pet?"""', "('Cat', 'Dog', 'Cow', 'Hamnster')"], {}), "('What is your pet?', ('C...
# Copyright (c) Khulnasoft Platforms, Inc. and affiliates. # This software may be used and distributed according to the terms of the Llmk 2 Community License Agreement. import langchain from langchain.llms import Replicate from flask import Flask from flask import request import os import requests import json class ...
[ "langchain.llms.Replicate" ]
[((1513, 1619), 'langchain.llms.Replicate', 'Replicate', ([], {'model': 'llmk2_13b_chat', 'model_kwargs': "{'temperature': 0.01, 'top_p': 1, 'max_new_tokens': 500}"}), "(model=llmk2_13b_chat, model_kwargs={'temperature': 0.01, 'top_p':\n 1, 'max_new_tokens': 500})\n", (1522, 1619), False, 'from langchain.llms import...
import os import langchain from config import * from util import * from langchain.llms import OpenAI, Cohere, HuggingFaceHub from langchain.chat_models import ChatOpenAI from langchain.agents import AgentType, initialize_agent, load_tools from typing import Optional, Type from langchain.callbacks.manager import AsyncCa...
[ "langchain.agents.initialize_agent", "langchain.llms.OpenAI", "langchain.agents.Tool", "langchain.chat_models.ChatOpenAI" ]
[((786, 807), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (792, 807), False, 'from langchain.llms import OpenAI, Cohere, HuggingFaceHub\n'), ((815, 840), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (825, 840), False, 'fro...
"""Interface with the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain_core.load.dump import dumps from langchain_core.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_k...
[ "langchain_core.load.load.loads", "langchain_core.load.dump.dumps", "langchainhub.Client" ]
[((679, 711), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (685, 711), False, 'from langchainhub import Client\n'), ((1912, 1925), 'langchain_core.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1917, 1925), False, 'from langchain_core.load.dump imp...
import langchain from dotenv import load_dotenv from langchain_openai import ChatOpenAI, OpenAI from langchain.schema import HumanMessage, AIMessage, SystemMessage from langchain.prompts import PromptTemplate, FewShotPromptTemplate from langchain.output_parsers import CommaSeparatedListOutputParser from langchain.cache...
[ "langchain_openai.ChatOpenAI", "langchain.cache.InMemoryCache", "langchain.prompts.PromptTemplate" ]
[((423, 436), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (434, 436), False, 'from dotenv import load_dotenv\n'), ((459, 474), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (472, 474), False, 'from langchain.cache import InMemoryCache\n'), ((508, 541), 'langchain_openai.ChatOpenAI', 'Ch...
""" Simple flask application to demo model inference """ # Loads env variable when running locally from dotenv import load_dotenv load_dotenv() ## Add parent directory to path for aws_helpers import sys sys.path.append('..') # Imports from flask import Flask, request, render_template, Response import json import os...
[ "langchain_inference.run_chain" ]
[((132, 145), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (143, 145), False, 'from dotenv import load_dotenv\n'), ((206, 227), 'sys.path.append', 'sys.path.append', (['""".."""'], {}), "('..')\n", (221, 227), False, 'import sys\n'), ((460, 511), 'os.path.join', 'os.path.join', (['"""data"""', '"""documents""...
import json from pathlib import Path from typing import Dict, List import langchain import numpy as np import typer from langchain.cache import SQLiteCache from langchain.llms import OpenAI from tqdm import tqdm langchain.llm_cache = SQLiteCache(database_path=".langchain.db") def _is_daster_empl(title: str) -> bool...
[ "langchain.llms.OpenAI", "langchain.cache.SQLiteCache" ]
[((236, 278), 'langchain.cache.SQLiteCache', 'SQLiteCache', ([], {'database_path': '""".langchain.db"""'}), "(database_path='.langchain.db')\n", (247, 278), False, 'from langchain.cache import SQLiteCache\n'), ((905, 952), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0.1)', 'model_name': '"""gpt-4-32k"""'}...
import langchain.vectorstores.opensearch_vector_search as ovs from opensearchpy import OpenSearch, RequestsHttpConnection, AWSV4SignerAuth, helpers from langchain.vectorstores import OpenSearchVectorSearch def create_ovs_client( collection_id, index_name, region, boto3_session, bedrock_embeddings...
[ "langchain.vectorstores.OpenSearchVectorSearch" ]
[((470, 515), 'opensearchpy.AWSV4SignerAuth', 'AWSV4SignerAuth', (['credentials', 'region', 'service'], {}), '(credentials, region, service)\n', (485, 515), False, 'from opensearchpy import OpenSearch, RequestsHttpConnection, AWSV4SignerAuth, helpers\n'), ((543, 724), 'opensearchpy.OpenSearch', 'OpenSearch', ([], {'hos...
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by redu...
[ "langchain.load.load.loads", "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps" ]
[((1586, 1613), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1603, 1613), False, 'import logging\n'), ((5793, 5811), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (5809, 5811), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), (...
import argparse import json import os from typing import Union, List, Literal from langchain.indexes import SQLRecordManager, index, IndexingResult import langchain_community.vectorstores as vectorstores from langchain_core.embeddings import Embeddings from langchain_openai import OpenAIEmbeddings from langchain_core...
[ "langchain_openai.OpenAIEmbeddings", "langchain.indexes.index", "langchain.indexes.SQLRecordManager" ]
[((1067, 1115), 'langchain_openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'model': '"""text-embedding-3-small"""'}), "(model='text-embedding-3-small')\n", (1083, 1115), False, 'from langchain_openai import OpenAIEmbeddings\n'), ((3445, 3480), 'os.makedirs', 'os.makedirs', (['db_path'], {'exist_ok': '(True)'}), '(d...
import textwrap import streamlit as st import langchain_helper as lch st.title("YouTube Assistant") with st.sidebar: with st.form(key='my_form'): youtube_url = st.sidebar.text_area( label="What is the YouTube video URL?", max_chars=50 ) query = st.sidebar.text_are...
[ "langchain_helper.create_db_from_youtube_video_url", "langchain_helper.get_response_from_query" ]
[((73, 102), 'streamlit.title', 'st.title', (['"""YouTube Assistant"""'], {}), "('YouTube Assistant')\n", (81, 102), True, 'import streamlit as st\n'), ((130, 152), 'streamlit.form', 'st.form', ([], {'key': '"""my_form"""'}), "(key='my_form')\n", (137, 152), True, 'import streamlit as st\n'), ((176, 250), 'streamlit.si...
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks....
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages....
[((923, 960), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (928, 960), False, 'from pydantic import Field, root_validator\n'), ((1034, 1067), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (103...
from langchain.chains.router import MultiPromptChain from langchain.chains.router.llm_router import LLMRouterChain, RouterOutputParser from langchain.prompts import PromptTemplate from langchain.chat_models import ChatOpenAI from langchain.prompts import ChatPromptTemplate from langchain.chains import LLMChain from ap...
[ "langchain.chains.router.llm_router.LLMRouterChain.from_llm", "langchain.chat_models.ChatOpenAI", "langchain.chains.router.MultiPromptChain", "langchain.chains.router.llm_router.RouterOutputParser", "langchain.chains.LLMChain", "langchain.prompts.ChatPromptTemplate.from_template" ]
[((3977, 4033), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model': 'ChatGPTModel.GPT3.value'}), '(temperature=0, model=ChatGPTModel.GPT3.value)\n', (3987, 4033), False, 'from langchain.chat_models import ChatOpenAI\n'), ((4531, 4574), 'langchain.prompts.ChatPromptTemplate.from_templa...
from __future__ import annotations import asyncio import functools import logging import os import warnings from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( Any, AsyncGenerator, Dict, Generator, List, Optional, Type, Type...
[ "langchain.schema.get_buffer_string", "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainT...
[((1114, 1141), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1131, 1141), False, 'import logging\n'), ((1286, 1329), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1296, 1329), False, 'from contextvars i...
import argparse import json import logging import os import pathlib from typing import Dict, List, Union, Optional import langchain import pandas as pd import tiktoken import wandb from langchain import LLMChain, FAISS from langchain.cache import SQLiteCache from langchain.chains import HypotheticalDocumentEmbedder fr...
[ "langchain.docstore.document.Document", "langchain.text_splitter.MarkdownTextSplitter", "langchain.cache.SQLiteCache", "langchain.chat_models.ChatOpenAI", "langchain.document_loaders.NotebookLoader", "langchain.text_splitter.PythonCodeTextSplitter", "langchain.text_splitter.TokenTextSplitter", "langch...
[((902, 943), 'langchain.cache.SQLiteCache', 'SQLiteCache', ([], {'database_path': '"""langchain.db"""'}), "(database_path='langchain.db')\n", (913, 943), False, 'from langchain.cache import SQLiteCache\n'), ((954, 981), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (971, 981), False, 'i...
import langchain from dotenv import load_dotenv from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from rmrkl import ChatZeroShotAgent, RetryAgentExecutor from .prompt import FORMAT_INSTRUCTIONS, QUESTION_PROMPT, SUFFIX from .tools import make_tools, Doc, Text,search_texts, load_texts imp...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((329, 342), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (340, 342), False, 'from dotenv import load_dotenv\n'), ((2064, 2077), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (2074, 2077), False, 'import time\n'), ((1632, 1784), 'rmrkl.ChatZeroShotAgent.from_llm_and_tools', 'ChatZeroShotAgent.from_llm_...
"""A tracer that runs evaluators over completed runs.""" from __future__ import annotations import logging from concurrent.futures import Future, ThreadPoolExecutor from typing import Any, Dict, List, Optional, Sequence, Set, Union from uuid import UUID import langsmith from langsmith import schemas as langsmith_sche...
[ "langchain.callbacks.tracers.langchain.get_client", "langchain.callbacks.manager.tracing_v2_enabled" ]
[((553, 580), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (570, 580), False, 'import logging\n'), ((2572, 2588), 'uuid.UUID', 'UUID', (['example_id'], {}), '(example_id)\n', (2576, 2588), False, 'from uuid import UUID\n'), ((2678, 2707), 'langchain.callbacks.tracers.langchain.get_clien...
"""A Tracer implementation that records to LangChain endpoint.""" from __future__ import annotations import logging import os from concurrent.futures import Future, ThreadPoolExecutor, wait from datetime import datetime from typing import Any, Dict, List, Optional, Set, Union from uuid import UUID from langchainplus_...
[ "langchain.schema.messages.messages_to_dict", "langchainplus_sdk.LangChainPlusClient", "langchain.env.get_runtime_environment" ]
[((633, 660), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (650, 660), False, 'import logging\n'), ((1895, 1928), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(1)'}), '(max_workers=1)\n', (1913, 1928), False, 'from concurrent.futures import Future...
""" Tools and utilities to work with https://github.com/unitedstates/congress fetch bill metadata to data/118/bills ``` usc-run govinfo --bulkdata=BILLSTATUS --congress=118 ``` fetch bill text to data/govinfo/BILLS ``` usc-run govinfo --bulkdata=BILLS --congress=118 ``` fetch plaw text to data/govinfo/PLAWS ``` us...
[ "langchain.schema.Document" ]
[((809, 836), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (826, 836), False, 'import logging\n'), ((14811, 14841), 'bs4.BeautifulSoup', 'BeautifulSoup', (['text_xml', '"""xml"""'], {}), "(text_xml, 'xml')\n", (14824, 14841), False, 'from bs4 import BeautifulSoup\n'), ((16122, 16131), '...
import streamlit as st st.set_page_config(layout="wide") import streamlit.components.v1 as components from streamlit_extras.stateful_button import button import os from dotenv import load_dotenv load_dotenv() import pandas as pd import numpy as np import time as time from tqdm import tqdm import base64 from pathlib im...
[ "langchain.chat_models.ChatOpenAI", "langchain.cache.SQLiteCache" ]
[((23, 56), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (41, 56), True, 'import streamlit as st\n'), ((195, 208), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (206, 208), False, 'from dotenv import load_dotenv\n'), ((637, 682), 'langchain.cache.SQLit...
import os import json import time from typing import List import faiss import pypdf import random import itertools import text_utils import pandas as pd import altair as alt import streamlit as st from io import StringIO from llama_index import Document from langchain.llms import Anthropic from langchain.chains import ...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.retrievers.SVMRetriever.from_texts", "langchain.embeddings.HuggingFaceEmbeddings", "langchain.chains.RetrievalQA.from_chain_type", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chat_models.ChatOpenAI", "langchain.vectorst...
[((13312, 13350), 'streamlit.sidebar.image', 'st.sidebar.image', (['"""img/diagnostic.jpg"""'], {}), "('img/diagnostic.jpg')\n", (13328, 13350), True, 'import streamlit as st\n'), ((15130, 15159), 'streamlit.header', 'st.header', (['"""`Auto-evaluator`"""'], {}), "('`Auto-evaluator`')\n", (15139, 15159), True, 'import ...
# general imports from constants import * # streamlit imports import streamlit as st from utils import * from streamlit_lottie import st_lottie # llama index imports import openai from llama_index import ( VectorStoreIndex, download_loader, ServiceContext, set_global_service_context, ) from llama_inde...
[ "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((1017, 1080), 'llama_index.llms.OpenAI', 'OpenAI', ([], {'model': '"""gpt-4-1106-preview"""', 'system_prompt': 'system_prompt'}), "(model='gpt-4-1106-preview', system_prompt=system_prompt)\n", (1023, 1080), False, 'from llama_index.llms import OpenAI\n'), ((1187, 1248), 'llama_index.ServiceContext.from_defaults', 'Se...
# basic import pandas as pd # import langchain from langchain.chains import LLMChain, HypotheticalDocumentEmbedder from langchain import PromptTemplate # import from other files from config import Settings from base import BASEVectorSearch settings = Settings() # ベクトル検索クラス class BasicSearch(BASEVectorSearch): #...
[ "langchain.chains.LLMChain", "langchain.chains.HypotheticalDocumentEmbedder", "langchain.PromptTemplate" ]
[((254, 264), 'config.Settings', 'Settings', ([], {}), '()\n', (262, 264), False, 'from config import Settings\n'), ((966, 1036), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['question']", 'template': 'prompt_template'}), "(input_variables=['question'], template=prompt_template)\n", (980, 10...
#%% import pandas as pd from utils import get_random_string from dotenv import load_dotenv import os import langchain from langchain.prompts import PromptTemplate from langchain.llms import OpenAI from langchain.chat_models import ChatOpenAI from openai import OpenAI import json import requests import datetime import...
[ "langchain.prompts.PromptTemplate.from_template", "langchain.chat_models.ChatOpenAI" ]
[((347, 360), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (358, 360), False, 'from dotenv import load_dotenv\n'), ((368, 416), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (378, 416), False, 'from l...
import os #from dotenv import load_dotenv import openai import langchain os.environ["OPENAI_API_KEY"] ="" os.environ["SQL_SERVER_USERNAME"] = "" os.environ["SQL_SERVER_ENDPOINT"] = "" os.environ["SQL_SERVER_PASSWORD"] = "" os.environ["SQL_SERVER_DATABASE"] = "" from sqlalchemy import create_engine from sqlalchemy....
[ "langchain.agents.create_sql_agent", "langchain.agents.agent_toolkits.SQLDatabaseToolkit", "langchain.llms.OpenAI", "langchain.sql_database.SQLDatabase.from_uri" ]
[((785, 808), 'sqlalchemy.engine.url.URL.create', 'URL.create', ([], {}), '(**db_config)\n', (795, 808), False, 'from sqlalchemy.engine.url import URL\n'), ((814, 842), 'langchain.sql_database.SQLDatabase.from_uri', 'SQLDatabase.from_uri', (['db_url'], {}), '(db_url)\n', (834, 842), False, 'from langchain.sql_database ...
"""A tracer that runs evaluators over completed runs.""" from __future__ import annotations import logging import threading import weakref from concurrent.futures import Future, ThreadPoolExecutor, wait from typing import Any, Dict, List, Optional, Sequence, Tuple, Union, cast from uuid import UUID import langsmith f...
[ "langchain.callbacks.tracers.langchain._get_executor", "langchain.callbacks.tracers.langchain.get_client", "langchain.callbacks.manager.tracing_v2_enabled" ]
[((672, 699), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (689, 699), False, 'import logging\n'), ((755, 772), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (770, 772), False, 'import weakref\n'), ((3430, 3447), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (3445, 3...
import langchain from langchain.llms import OpenAI from langchain.chat_models import ChatOpenAI from langchain.prompts import load_prompt from langchain.output_parsers import OutputFixingParser # output_parser = DatetimeOutputParser() # # misformatted = result.content from langchain.schema import AIMessage, HumanMessa...
[ "langchain.prompts.PromptTemplate", "langchain.llms.OpenAI", "langchain.prompts.load_prompt", "langchain.chat_models.ChatOpenAI" ]
[((1062, 1070), 'langchain.llms.OpenAI', 'OpenAI', ([], {}), '()\n', (1068, 1070), False, 'from langchain.llms import OpenAI\n'), ((1078, 1112), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': 'api_key'}), '(openai_api_key=api_key)\n', (1088, 1112), False, 'from langchain.chat_models import Ch...
import os import re from uuid import UUID from typing import Any, Dict, List, Optional, Union import asyncio import langchain import streamlit as st from langchain.schema import LLMResult from langchain.chat_models import ChatOpenAI from langchain.agents import Tool from langchain.agents import AgentType from langcha...
[ "langchain.agents.initialize_agent", "langchain.memory.ConversationBufferMemory", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.agents.Tool" ]
[((815, 826), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (824, 826), False, 'import os\n'), ((6031, 6120), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""', 'temperature': '(0)', 'openai_api_key': 'openai_api_key'}), "(model_name='gpt-3.5-turbo', temperature=0, openai_api_key...
import os import weaviate import key_config import langchain from langchain.vectorstores import Weaviate from langchain.chains import ConversationalRetrievalChain from langchain.memory import ConversationSummaryMemory from langchain.chat_models import ChatOpenAI from langchain.embeddings import OpenAIEmbeddings clien...
[ "langchain.memory.ConversationSummaryMemory", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.Weaviate", "langchain.embeddings.OpenAIEmbeddings" ]
[((438, 486), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'model': '"""text-embedding-ada-002"""'}), "(model='text-embedding-ada-002')\n", (454, 486), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((496, 566), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""g...
# MIT License # # Copyright (c) 2024, Justin Randall, Smart Interactive Transformations Inc. # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitatio...
[ "langchain_core.prompts.HumanMessagePromptTemplate.from_template", "langchain_core.messages.SystemMessage" ]
[((2556, 2594), 'langchain_core.messages.SystemMessage', 'SystemMessage', ([], {'content': 'self.sys_prompt'}), '(content=self.sys_prompt)\n', (2569, 2594), False, 'from langchain_core.messages import SystemMessage\n'), ((2612, 2675), 'langchain_core.prompts.HumanMessagePromptTemplate.from_template', 'HumanMessagePromp...
from approaches.index.store.cosmos_index_store import CosmosIndexStore from llama_index import StorageContext from approaches.index.store.cosmos_doc_store import CosmosDocumentStore from llama_index import load_index_from_storage import os import openai from langchain.chat_models import AzureChatOpenAI from langchain....
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((832, 845), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (843, 845), False, 'from dotenv import load_dotenv\n'), ((1039, 1074), 'os.environ.get', 'os.environ.get', (['"""AZURE_OPENAI_BASE"""'], {}), "('AZURE_OPENAI_BASE')\n", (1053, 1074), False, 'import os\n'), ((1098, 1153), 'os.environ.get', 'os.environ....
from abc import ABC, abstractmethod from typing import List, Optional from pydantic import BaseModel, Extra, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import ( AIMessage, BaseLanguageMod...
[ "langchain.schema.ChatResult", "langchain.schema.ChatGeneration", "langchain.schema.HumanMessage", "langchain.schema.AIMessage", "langchain.schema.LLMResult", "langchain.callbacks.get_callback_manager" ]
[((568, 605), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (573, 605), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((696, 739), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager'}), '(default_factory=get_ca...
import boto3 import os import json # from langchain_experimental.agents.agent_toolkits import create_pandas_dataframe_agent from langchain_community.llms import Bedrock from llama_index.node_parser import SimpleNodeParser from llama_index.embeddings import LangchainEmbedding from langchain_community.embeddings import ...
[ "langchain_community.llms.Bedrock" ]
[((974, 1051), 'logging.basicConfig', 'logging.basicConfig', ([], {'filename': '"""rag.log"""', 'encoding': '"""utf-8"""', 'level': 'logging.INFO'}), "(filename='rag.log', encoding='utf-8', level=logging.INFO)\n", (993, 1051), False, 'import logging\n'), ((1113, 1128), 'boto3.Session', 'boto3.Session', ([], {}), '()\n'...
import pandas as pd energy_data = pd.read_csv('EnergyUsage.csv') summary_stats = energy_data.describe() from langchain import LangchainAgent # Initialize Langchain agent agent = LangchainAgent() # Connect to LLama2 agent.connect_to_llama2() # Send data to LLama2 agent.send_data_to_llama2(summary_stats) # Optional...
[ "langchain.LangchainAgent" ]
[((35, 65), 'pandas.read_csv', 'pd.read_csv', (['"""EnergyUsage.csv"""'], {}), "('EnergyUsage.csv')\n", (46, 65), True, 'import pandas as pd\n'), ((181, 197), 'langchain.LangchainAgent', 'LangchainAgent', ([], {}), '()\n', (195, 197), False, 'from langchain import LangchainAgent\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator impor...
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.load.dump.dumpd", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.RunInfo", "langchain.schema.AIMessage", "langchain.llm_cache.lookup...
[((2353, 2390), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2358, 2390), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2464, 2497), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, ...
"""Create a ChatVectorDBChain for question/answering.""" from langchain.callbacks.base import AsyncCallbackManager from langchain.callbacks.tracers import LangChainTracer from langchain.chains import ConversationalRetrievalChain from langchain.chains.chat_vector_db.prompts import PromptTemplate from langchain.chains.ll...
[ "langchain.chains.question_answering.load_qa_chain", "langchain.callbacks.tracers.LangChainTracer", "langchain.callbacks.base.AsyncCallbackManager", "langchain.llms.OpenAI", "langchain.chains.llm.LLMChain", "langchain.chains.chat_vector_db.prompts.PromptTemplate", "langchain.chains.chat_vector_db.prompt...
[((668, 759), 'langchain.chains.chat_vector_db.prompts.PromptTemplate', 'PromptTemplate', ([], {'template': 'mi_qa_prompt_template', 'input_variables': "['context', 'question']"}), "(template=mi_qa_prompt_template, input_variables=['context',\n 'question'])\n", (682, 759), False, 'from langchain.chains.chat_vector_d...
import logging import os import pprint import uuid from typing import List import chromadb import gradio as gr import requests import zhipuai from bs4 import BeautifulSoup from dotenv import load_dotenv, find_dotenv # Import langchain stuff from langchain.chains import ConversationalRetrievalChain from langchain.docum...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain_community.vectorstores.chroma.Chroma.from_documents", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.memory.ConversationBufferMemory", "langchain_community.vectorstores.chroma.Chroma", "langchain_core.prompts.PromptTemplate"...
[((1392, 1490), 'llms.zhipuai_llm.ZhipuAILLM', 'ZhipuAILLM', ([], {'model': '"""chatglm_turbo"""', 'temperature': '(0.9)', 'top_p': '(0.1)', 'zhipuai_api_key': 'zhipuai.api_key'}), "(model='chatglm_turbo', temperature=0.9, top_p=0.1,\n zhipuai_api_key=zhipuai.api_key)\n", (1402, 1490), False, 'from llms.zhipuai_llm ...
import langchain llm_huggingface=HuggingFaceHub(repo_id="google/flan-t5-large",model_kwargs={"temperature":0,"max_length":64}) capital_prompt = PromptTemplate(input_variables=["country"], template="Tell me the capital of {country}") famous_prompt = PromptTemplate(input_variables=["famous"], template="Tell me the capit...
[ "langchain.chains.SimpleSequentialChain", "langchain.LLMChain" ]
[((353, 415), 'langchain.LLMChain', 'langchain.LLMChain', ([], {'llm': 'llm_huggingface', 'prompt': 'capital_prompt'}), '(llm=llm_huggingface, prompt=capital_prompt)\n', (371, 415), False, 'import langchain\n'), ((431, 492), 'langchain.LLMChain', 'langchain.LLMChain', ([], {'llm': 'llm_huggingface', 'prompt': 'famous_p...
"""An example of how to test Python code generating prompts""" import re # Brining some "prompt generator" classes from promptimize.prompt_cases import LangchainPromptCase # Bringing some useful eval function that help evaluating and scoring responses # eval functions have a handle on the prompt object and are expect...
[ "langchain.output_parsers.ResponseSchema", "langchain.output_parsers.StructuredOutputParser.from_response_schemas", "langchain.PromptTemplate" ]
[((1146, 1208), 'langchain.output_parsers.StructuredOutputParser.from_response_schemas', 'StructuredOutputParser.from_response_schemas', (['response_schemas'], {}), '(response_schemas)\n', (1190, 1208), False, 'from langchain.output_parsers import StructuredOutputParser, ResponseSchema\n'), ((2218, 2382), 'langchain.Pr...
""" The ``mlflow.langchain`` module provides an API for logging and loading LangChain models. This module exports multivariate LangChain models in the langchain flavor and univariate LangChain models in the pyfunc flavor: LangChain (native) format This is the main flavor that can be accessed with LangChain APIs. :...
[ "langchain.agents.initialize_agent", "langchain.chains.loading.load_chain" ]
[((2012, 2046), 'logging.getLogger', 'logging.getLogger', (['mlflow.__name__'], {}), '(mlflow.__name__)\n', (2029, 2046), False, 'import logging\n'), ((11731, 11807), 'mlflow.utils.environment._validate_env_arguments', '_validate_env_arguments', (['conda_env', 'pip_requirements', 'extra_pip_requirements'], {}), '(conda...
# Import the necessary libraries import random import time from llama_index.llms import OpenAI import streamlit as st from llama_index import VectorStoreIndex, ServiceContext, StorageContext, set_global_service_context from langchain.embeddings.huggingface import HuggingFaceEmbeddings from llama_index.embeddings import...
[ "langchain_openai.ChatOpenAI", "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((855, 895), 'streamlit.title', 'st.title', (['"""🦜🔗 Tourism Assistant Chatbot"""'], {}), "('🦜🔗 Tourism Assistant Chatbot')\n", (863, 895), True, 'import streamlit as st\n'), ((5721, 5781), 'llama_index.set_global_service_context', 'set_global_service_context', (['st.session_state.service_context'], {}), '(st.sess...
import langchain_helper as lch import streamlit as st import textwrap st.title("Youtube Assistant") with st.sidebar: with st.form(key="my_from"): youtube_url = st.text_area( label = "What is the youtube url?", max_chars=50, ) query = st.text_area( label ...
[ "langchain_helper.get_response_from_query", "langchain_helper.create_vector_db_from_youtube" ]
[((71, 100), 'streamlit.title', 'st.title', (['"""Youtube Assistant"""'], {}), "('Youtube Assistant')\n", (79, 100), True, 'import streamlit as st\n'), ((517, 563), 'langchain_helper.create_vector_db_from_youtube', 'lch.create_vector_db_from_youtube', (['youtube_url'], {}), '(youtube_url)\n', (550, 563), True, 'import ...
# This code sets up the necessary components, interacts with the LangChain tool and ChatOpenAI model to perform text summarization, # and provides a user interface for input and output. from langchain.document_loaders import UnstructuredFileLoader # Importing necessary modules from langchain.document_loaders import ...
[ "langchain.chains.summarize.load_summarize_chain", "langchain.document_loaders.UnstructuredFileLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.chat_models.ChatOpenAI", "langchain.document_loaders.UnstructuredPDFLoader", "langchain.prompts.PromptTemplate" ]
[((5769, 5891), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Positive summarizer"""', 'page_icon': '"""📖"""', 'layout': '"""wide"""', 'initial_sidebar_state': '"""collapsed"""'}), "(page_title='Positive summarizer', page_icon='📖', layout=\n 'wide', initial_sidebar_state='collapsed')\n...
from mp_api.client import MPRester from emmet.core.summary import HasProps import openai import langchain from langchain_openai import ChatOpenAI from langchain.agents import initialize_agent from langchain.agents import Tool, tool from langchain.prompts.few_shot import FewShotPromptTemplate from langchain.prompts.prom...
[ "langchain_openai.ChatOpenAI", "langchain.prompts.prompt.PromptTemplate", "langchain_openai.OpenAIEmbeddings", "langchain.agents.Tool", "langchain.prompts.few_shot.FewShotPromptTemplate" ]
[((991, 1024), 're.compile', 're.compile', (['"""([A-Z][a-z]*)(\\\\d*)"""'], {}), "('([A-Z][a-z]*)(\\\\d*)')\n", (1001, 1024), False, 'import re\n'), ((2083, 2198), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': 'self.model', 'temperature': '(0.7)', 'n': '(1)', 'best_of': '(5)', 'top_p': '(1.0)', 'sto...
import streamlit as st from streamlit_chat import message import pandas as pd from langchain.llms import OpenAI import os from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationSummaryBufferMemory import plotly.express from streamlit_searchbox import st_searchbox from typing import List, ...
[ "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((1329, 1342), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (1340, 1342), False, 'from dotenv import load_dotenv\n'), ((1378, 1486), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""PubMeta.ai"""', 'page_icon': '"""⚕️"""', 'layout': '"""wide"""', 'initial_sidebar_state': '"""auto"""...
from typing import Any, Dict, List, Optional from langchain import PromptTemplate ,LLMChain import langchain from langchain.chat_models import ChatOpenAI ,AzureChatOpenAI from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler import sys import re import argparse import os print(sys.path) sys.p...
[ "langchain.LLMChain", "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.prompts.chat.ChatPromptTemplate", "langchain.schema.SystemMessage", "langchain.prompts.chat.HumanMessagePromptTemplate.from_template", "langchain.PromptTemplate" ]
[((315, 335), 'sys.path.append', 'sys.path.append', (['"""."""'], {}), "('.')\n", (330, 335), False, 'import sys\n'), ((3893, 3960), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_template', 'input_variables': "['essay']"}), "(template=prompt_template, input_variables=['essay'])\n", (3907, 3960...
import streamlit as st from dotenv import load_dotenv load_dotenv() import os import tempfile from llama_index import SimpleDirectoryReader, StorageContext, LLMPredictor from llama_index import VectorStoreIndex from llama_index import ServiceContext from llama_index.embeddings.langchain import LangchainEmbedding from...
[ "langchain.embeddings.CohereEmbeddings", "langchain.chat_models.ChatOpenAI" ]
[((55, 68), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (66, 68), False, 'from dotenv import load_dotenv\n'), ((860, 890), 'llama_index.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {}), '()\n', (888, 890), False, 'from llama_index import SimpleDirectoryReader, StorageContext, LLMPredic...
import os import langchain from langchain.utilities import SerpAPIWrapper from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType from langchain.chat_models import ChatOpenAI os.environ['OPENAI_API_KEY'] = "" os.environ['SERPAPI_API_KEY'] = "" llm = ChatOpenAI(temperature=0, model=...
[ "langchain.agents.initialize_agent", "langchain.utilities.SerpAPIWrapper", "langchain.agents.Tool", "langchain.chat_models.ChatOpenAI" ]
[((288, 341), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model': '"""gpt-3.5-turbo-0613"""'}), "(temperature=0, model='gpt-3.5-turbo-0613')\n", (298, 341), False, 'from langchain.chat_models import ChatOpenAI\n'), ((352, 368), 'langchain.utilities.SerpAPIWrapper', 'SerpAPIWrapper', (...
import streamlit as st from dotenv import load_dotenv import os from htmlTemplates import css, bot_template, user_template import langchain from langchain.document_loaders import GitLoader from langchain.text_splitter import ( RecursiveCharacterTextSplitter, Language, ) from langchain.text_splitter import Recur...
[ "langchain.prompts.chat.SystemMessagePromptTemplate.from_template", "langchain.memory.ConversationBufferMemory", "langchain.document_loaders.GitLoader", "langchain.vectorstores.DeepLake", "langchain.chat_models.ChatOpenAI", "langchain.text_splitter.RecursiveCharacterTextSplitter.from_language", "langcha...
[((822, 895), 'langchain.document_loaders.GitLoader', 'GitLoader', ([], {'clone_url': 'github_url', 'repo_path': 'local_path', 'branch': 'repo_branch'}), '(clone_url=github_url, repo_path=local_path, branch=repo_branch)\n', (831, 895), False, 'from langchain.document_loaders import GitLoader\n'), ((2876, 2915), 'langch...
import json from llama_index.core.service_context_elements.llm_predictor import LLMPredictor from llama_index.core.utilities.sql_wrapper import SQLDatabase from llama_index.core.response_synthesizers import get_response_synthesizer from llama_index.embeddings.langchain import LangchainEmbedding from llama_index.core.re...
[ "langchain.agents.initialize_agent", "langchain_community.chat_models.ChatOpenAI" ]
[((3043, 3075), 'app.database.dbc.get_llm_by_name', 'dbc.get_llm_by_name', (['db', 'llmName'], {}), '(db, llmName)\n', (3062, 3075), False, 'from app.database import dbc\n'), ((4079, 4112), 'app.database.dbc.get_project_by_name', 'dbc.get_project_by_name', (['db', 'name'], {}), '(db, name)\n', (4102, 4112), False, 'fro...
import langchain_helper as lch import streamlit as st st.title("Pets name generator") user_input_animal_type = st.sidebar.selectbox("What is your pet?", ("Cat", "Dog", "Cow", "Hamster")) if user_input_animal_type == "Cat": pet_color = st.sidebar.text_area("What's the color of your cat?", max_chars=15) if user_i...
[ "langchain_helper.generate_pet_name" ]
[((55, 86), 'streamlit.title', 'st.title', (['"""Pets name generator"""'], {}), "('Pets name generator')\n", (63, 86), True, 'import streamlit as st\n'), ((113, 188), 'streamlit.sidebar.selectbox', 'st.sidebar.selectbox', (['"""What is your pet?"""', "('Cat', 'Dog', 'Cow', 'Hamster')"], {}), "('What is your pet?', ('Ca...
"""LangChain agent The agent chooses a sequence of actions to respond to a human's question. It has access to a set of tools. The agent memorizes the conversation history and can use it to make decisions. """ from typing import Optional from dotenv import load_dotenv from langchain.agents import AgentExecutor from l...
[ "langchain_core.prompts.MessagesPlaceholder", "langchain.agents.AgentExecutor", "langchain.memory.ConversationBufferMemory", "langchain.agents.format_scratchpad.format_log_to_str", "langchain.agents.output_parsers.JSONAgentOutputParser" ]
[((956, 969), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (967, 969), False, 'from dotenv import load_dotenv\n'), ((2930, 2956), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {}), '()\n', (2954, 2956), False, 'from langchain.memory import ConversationBufferMemory\n'), ((3315, ...
# This is an example of integrating a LLM with streamlit import streamlit as st import os import openai import langchain from langchain.llms import OpenAI from langchain import PromptTemplate #from dotenv import load_dotenv # Specify the path to the .env file #dotenv_path = os.path.join(os.path.dirname(__file__), '.en...
[ "langchain.llms.OpenAI", "langchain.PromptTemplate" ]
[((390, 459), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Globalize Email"""', 'page_icon': '""":robot:"""'}), "(page_title='Globalize Email', page_icon=':robot:')\n", (408, 459), True, 'import streamlit as st\n'), ((462, 489), 'streamlit.header', 'st.header', (['"""Globalize Text"""'], {...
import sys sys.stdout.reconfigure(encoding="utf-8") sys.stdin.reconfigure(encoding="utf-8") import streamlit as st import streamlit.components.v1 as components import re import random CODE_BUILD_KG = """ # 准备 GraphStore os.environ['NEBULA_USER'] = "root" os.environ['NEBULA_PASSWORD'] = "nebula" # default passwor...
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((12, 52), 'sys.stdout.reconfigure', 'sys.stdout.reconfigure', ([], {'encoding': '"""utf-8"""'}), "(encoding='utf-8')\n", (34, 52), False, 'import sys\n'), ((53, 92), 'sys.stdin.reconfigure', 'sys.stdin.reconfigure', ([], {'encoding': '"""utf-8"""'}), "(encoding='utf-8')\n", (74, 92), False, 'import sys\n'), ((2988, 3...
from langchain.agents import AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.tools import Tool, StructuredTool from langchain.prompts import StringPromptTemplate from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.llms import VertexAI from typing imp...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.schema.AgentAction", "langchain.llms.VertexAI", "langchain.schema.AgentFinish", "langchain.callbacks.FileCallbackHandler", "langchain.chains.LLMChain" ]
[((1046, 1098), 'os.makedirs', 'os.makedirs', (['f"""./results/{timestamp}"""'], {'exist_ok': '(True)'}), "(f'./results/{timestamp}', exist_ok=True)\n", (1057, 1098), False, 'import os\n'), ((1332, 1364), 'logging.getLogger', 'logging.getLogger', (['"""info_logger"""'], {}), "('info_logger')\n", (1349, 1364), False, 'i...
import langchain from dotenv import load_dotenv from langchain.chains import FlareChain from langchain.chat_models import ChatOpenAI from langchain.embeddings import OpenAIEmbeddings from langchain.llms import OpenAI from langchain.vectorstores import FAISS langchain.verbose = True load_dotenv() # FAISSで保存されたベクトルを読み...
[ "langchain.vectorstores.FAISS.load_local", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.chains.FlareChain.from_llm", "langchain.embeddings.OpenAIEmbeddings" ]
[((285, 298), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (296, 298), False, 'from dotenv import load_dotenv\n'), ((336, 354), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (352, 354), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((360, 403), 'langchain.vect...
import langchain from langchain.llms import OpenAI from langchain.chat_models import ChatOpenAI from langchain.output_parsers import OutputFixingParser # output_parser = DatetimeOutputParser() # # misformatted = result.content from langchain.schema import AIMessage, HumanMessage, SystemMessage from langchain.cache imp...
[ "langchain.output_parsers.PydanticOutputParser", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.prompts.ChatPromptTemplate.from_messages" ]
[((1020, 1028), 'langchain.llms.OpenAI', 'OpenAI', ([], {}), '()\n', (1026, 1028), False, 'from langchain.llms import OpenAI\n'), ((1036, 1070), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': 'api_key'}), '(openai_api_key=api_key)\n', (1046, 1070), False, 'from langchain.chat_models import Ch...
import os from dotenv import load_dotenv import streamlit as st from langchain.chains import LLMChain from langchain import PromptTemplate from genai.credentials import Credentials from genai.extensions.langchain import LangChainInterface from genai.schemas import GenerateParams load_dotenv() api_key = os.getenv("GE...
[ "langchain.chains.LLMChain", "langchain.PromptTemplate" ]
[((283, 296), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (294, 296), False, 'from dotenv import load_dotenv\n'), ((307, 335), 'os.getenv', 'os.getenv', (['"""GENAI_KEY"""', 'None'], {}), "('GENAI_KEY', None)\n", (316, 335), False, 'import os\n'), ((351, 379), 'os.getenv', 'os.getenv', (['"""GENAI_API"""', '...
from __future__ import annotations from abc import ABC, abstractmethod from typing import Type import langchain from langchain import LLMChain, LLMMathChain, PromptTemplate from langchain.cache import InMemoryCache from langchain.chains import SequentialChain from langchain.output_parsers import PydanticOutputParser ...
[ "langchain.LLMChain", "langchain.output_parsers.PydanticOutputParser", "langchain.chains.SequentialChain", "langchain.LLMMathChain.from_llm", "langchain.cache.InMemoryCache", "langchain.PromptTemplate" ]
[((1259, 1311), 'langchain.output_parsers.PydanticOutputParser', 'PydanticOutputParser', ([], {'pydantic_object': 'RoomLoopAnswer'}), '(pydantic_object=RoomLoopAnswer)\n', (1279, 1311), False, 'from langchain.output_parsers import PydanticOutputParser\n'), ((1527, 1571), 'langchain.output_parsers.PydanticOutputParser',...
import os from datetime import datetime, timezone from dotenv import load_dotenv from langchain.agents import AgentType, initialize_agent from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationBufferMemory from langchain.prompts import MessagesPlaceholder from src.xm_group_tools import ...
[ "langchain.agents.initialize_agent", "langchain.memory.ConversationBufferMemory", "langchain.prompts.MessagesPlaceholder", "langchain.chat_models.ChatOpenAI" ]
[((458, 471), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (469, 471), False, 'from dotenv import load_dotenv\n'), ((482, 509), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (491, 509), False, 'import os\n'), ((742, 770), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([...
"""Push and pull to the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain.load.dump import dumps from langchain.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_key: Opti...
[ "langchain.load.load.loads", "langchainhub.Client", "langchain.load.dump.dumps" ]
[((671, 703), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (677, 703), False, 'from langchainhub import Client\n'), ((1886, 1899), 'langchain.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1891, 1899), False, 'from langchain.load.dump import dumps\...
"""Base interface that all chains should implement.""" import inspect import json import logging import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import Field, root_validator, validator import langchain from lang...
[ "langchain.schema.RunInfo", "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.load.dump.dumpd", "langchain.callbacks.manager.CallbackManager.configure" ]
[((702, 729), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (719, 729), False, 'import logging\n'), ((2435, 2468), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2440, 2468), False, 'from pydantic import Field, root_validator...
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import functools import inspect import logging import uuid from enum import Enum from typing import ( TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, ...
[ "langchain.schema.messages.messages_from_dict", "langchain._api.warn_deprecated", "langchain.schema.runnable.config.get_executor_for_config", "langchain.evaluation.schema.EvaluatorType", "langchain.smith.evaluation.name_generation.random_name", "langchain.smith.evaluation.StringRunEvaluatorChain.from_run_...
[((1724, 1751), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1741, 1751), False, 'import logging\n'), ((33983, 34008), 'langchain.callbacks.tracers.evaluation.wait_for_all_evaluators', 'wait_for_all_evaluators', ([], {}), '()\n', (34006, 34008), False, 'from langchain.callbacks.tracers...
from pydantic import BaseModel, Field import os from langchain import OpenAI from langchain.chat_models import ChatOpenAI from langchain.agents import initialize_agent, Tool # from langchain.chains import PALChain from langchain.chains.conversation.memory import ConversationBufferMemory from langchain import Pr...
[ "langchain.agents.AgentExecutor", "langchain.memory.ConversationBufferMemory", "langchain.prompts.MessagesPlaceholder", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.FAISS.from_documents", "langchain.agents.agent_toolkits.create_retriever_tool", "langchain.schema.messages.SystemMessage", ...
[((3412, 3472), 'streamlit.markdown', 'st.markdown', (['hide_share_button_style'], {'unsafe_allow_html': '(True)'}), '(hide_share_button_style, unsafe_allow_html=True)\n', (3423, 3472), True, 'import streamlit as st\n'), ((3474, 3537), 'streamlit.markdown', 'st.markdown', (['hide_star_and_github_style'], {'unsafe_allow...
from langchain.embeddings import HuggingFaceHubEmbeddings from langchain.embeddings.openai import OpenAIEmbeddings from llama_index import ServiceContext, LLMPredictor from llama_index.embeddings.langchain import LangchainEmbedding from langchain import OpenAI, HuggingFaceHub from langchain.chat_models import ChatOpenA...
[ "langchain.embeddings.HuggingFaceHubEmbeddings", "langchain.HuggingFaceHub", "langchain.OpenAI", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((1146, 1255), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'llm_predictor': 'llm_predictor', 'embed_model': 'embed_model', 'chunk_size_limit': '(512)'}), '(llm_predictor=llm_predictor, embed_model=\n embed_model, chunk_size_limit=512)\n', (1174, 1255), False, 'from llama_index ...
"""Schemas for the langchainplus API.""" from __future__ import annotations import logging import os from concurrent.futures import Future, ThreadPoolExecutor, wait from datetime import datetime from typing import Dict, List, Optional, Union, cast from uuid import UUID, uuid4 from pydantic import Field, PrivateAttr, ...
[ "langchainplus_sdk.utils.get_runtime_environment" ]
[((533, 560), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (550, 560), False, 'import logging\n'), ((683, 716), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(1)'}), '(max_workers=1)\n', (701, 716), False, 'from concurrent.futures import Future, Th...
import streamlit as st import openai import langchain # from langchain import PromptTemplate, LLMChain # from langchain.llms import OpenAI # # Set your OpenAI API key # openai_api_key = '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' # # Create a sidebar for language selection # st.sidebar.title('Translation A...
[ "langchain.LLMChain", "langchain.llms.OpenAI", "langchain.PromptTemplate" ]
[((3099, 3134), 'streamlit.sidebar.title', 'st.sidebar.title', (['"""Translation App"""'], {}), "('Translation App')\n", (3115, 3134), True, 'import streamlit as st\n'), ((3216, 3265), 'streamlit.sidebar.selectbox', 'st.sidebar.selectbox', (['"""Input Language"""', 'languages'], {}), "('Input Language', languages)\n", ...
# Import os to set API key import os import langchain # Bring in streamlit for UI/app interface import streamlit as st from langchain.callbacks import get_openai_callback from langchain.utilities import GoogleSerperAPIWrapper from common.SerperSearchRetriever import SerperSearchRetriever from st_pages import add_inde...
[ "langchain.llms.openai.OpenAI", "langchain.chains.FlareChain.from_llm", "langchain.utilities.GoogleSerperAPIWrapper", "langchain.callbacks.get_openai_callback" ]
[((511, 524), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (522, 524), False, 'from dotenv import load_dotenv\n'), ((526, 543), 'st_pages.add_indentation', 'add_indentation', ([], {}), '()\n', (541, 543), False, 'from st_pages import add_indentation\n'), ((910, 954), 'streamlit.title', 'st.title', (['"""🦜🔗�...
import streamlit as st import langchain_helper st.title("Learning Helper for CBSE Class XI") subject = st.sidebar.selectbox("Pick your subject", ("maths", "physics","chemistry")) #st.sidebar.t if subject: response = langchain_helper.generate_subject_help_links(subject) st.header(f'Helpful Links for {subject}...
[ "langchain_helper.generate_subject_help_links" ]
[((48, 93), 'streamlit.title', 'st.title', (['"""Learning Helper for CBSE Class XI"""'], {}), "('Learning Helper for CBSE Class XI')\n", (56, 93), True, 'import streamlit as st\n'), ((105, 181), 'streamlit.sidebar.selectbox', 'st.sidebar.selectbox', (['"""Pick your subject"""', "('maths', 'physics', 'chemistry')"], {})...
import langchain.llms from langchain import GoogleSearchAPIWrapper, LLMChain from langchain.agents import initialize_agent, AgentType, Tool, ZeroShotAgent, AgentExecutor from langchain.schema import BaseMemory def setup_agent(llm: langchain.llms.BaseLLM, memory: BaseMemory): search = GoogleSearchAPIWrapper() ...
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.LLMChain", "langchain.agents.ZeroShotAgent.create_prompt", "langchain.agents.ZeroShotAgent", "langchain.GoogleSearchAPIWrapper", "langchain.agents.Tool" ]
[((291, 315), 'langchain.GoogleSearchAPIWrapper', 'GoogleSearchAPIWrapper', ([], {}), '()\n', (313, 315), False, 'from langchain import GoogleSearchAPIWrapper, LLMChain\n'), ((833, 948), 'langchain.agents.ZeroShotAgent.create_prompt', 'ZeroShotAgent.create_prompt', (['tools'], {'prefix': 'prefix', 'suffix': 'suffix', '...
import itertools from langchain.cache import InMemoryCache, SQLiteCache import langchain import pandas as pd from certa.utils import merge_sources from certa.explain import CertaExplainer from datetime import datetime import os import ellmer.models import ellmer.metrics from time import sleep, time import json import t...
[ "langchain.cache.InMemoryCache", "langchain.cache.SQLiteCache" ]
[((8572, 8636), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Run saliency experiments."""'}), "(description='Run saliency experiments.')\n", (8595, 8636), False, 'import argparse\n'), ((598, 613), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (611, 613), False, 'f...
# Databricks notebook source # MAGIC %md # MAGIC # 3. Chatbotの作成とデプロイ # MAGIC # MAGIC <br/> # MAGIC <img src="https://github.com/naoyaabe-db/public_demo_images/blob/3380b6d73937cd95efae845799c37de910b7394c/rag_demo_images/diagram_notebook3.png?raw=true" style="float: right" width="1000px"> # MAGIC <br/> # MAGIC # MAGIC...
[ "langchain.chains.RetrievalQA.from_chain_type", "langchain.vectorstores.DatabricksVectorSearch", "langchain.prompts.PromptTemplate", "langchain.chat_models.ChatDatabricks" ]
[((5946, 5996), 'mlflow.deployments.get_deploy_client', 'mlflow.deployments.get_deploy_client', (['"""databricks"""'], {}), "('databricks')\n", (5982, 5996), False, 'import mlflow\n'), ((6909, 6974), 'langchain.chat_models.ChatDatabricks', 'ChatDatabricks', ([], {'endpoint': 'chat_model_endpoint_name', 'max_tokens': '(...
# Import Langchain modules from langchain.document_loaders import PyPDFLoader from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores import FAISS from langchain.chains import RetrievalQA from langchain.llms import OpenAI # Impo...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.llms.OpenAI", "langchain.vectorstores.FAISS.from_documents", "langchain.document_loaders.PyPDFLoader", "langchain.embeddings.OpenAIEmbeddings" ]
[((573, 606), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (596, 606), False, 'import warnings\n'), ((712, 808), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO', 'format': '"""%(asctime)s - %(levelname)s - %(message)s"""'}), "(level=logging.IN...
import streamlit as st import dotenv import langchain import json from cassandra.cluster import Session from cassandra.query import PreparedStatement from langchain.agents.agent_toolkits import create_retriever_tool, create_conversational_retrieval_agent from langchain.chat_models import ChatOpenAI from langchain.emb...
[ "langchain.chat_models.ChatOpenAI", "langchain.schema.Document", "langchain.agents.agent_toolkits.create_conversational_retrieval_agent", "langchain.agents.agent_toolkits.create_retriever_tool", "langchain.schema.SystemMessage", "langchain.embeddings.OpenAIEmbeddings" ]
[((5021, 5054), 'streamlit.set_page_config', 'st.set_page_config', ([], {'layout': '"""wide"""'}), "(layout='wide')\n", (5039, 5054), True, 'import streamlit as st\n'), ((5462, 5502), 'streamlit.chat_input', 'st.chat_input', ([], {'placeholder': '"""Ask chatbot"""'}), "(placeholder='Ask chatbot')\n", (5475, 5502), True...
import os import langchain import streamlit as st from dotenv import load_dotenv from langchain.chat_models import ChatOpenAI from langchain.cache import InMemoryCache from langchain.prompts import SystemMessagePromptTemplate, HumanMessagePromptTemplate, AIMessagePromptTemplate, ChatPromptTemplate, PromptTemplate # C...
[ "langchain.cache.InMemoryCache", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.chat_models.ChatOpenAI", "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.prompts.SystemMessagePromptTemplate.from_template" ]
[((326, 345), 'dotenv.load_dotenv', 'load_dotenv', (['""".env"""'], {}), "('.env')\n", (337, 345), False, 'from dotenv import load_dotenv\n'), ((356, 377), 'os.environ.get', 'os.environ.get', (['"""key"""'], {}), "('key')\n", (370, 377), False, 'import os\n'), ((416, 431), 'langchain.cache.InMemoryCache', 'InMemoryCach...
import streamlit as st import langchain_helper st.title("Restaurant Name Generator") cuisine = st.sidebar.selectbox("Pick a Cuisine",("Indian","Italian","Mexican","Arabic","American")) if cuisine: response = langchain_helper.generate_restaurant_name_and_items(cuisine) st.header(response['resta...
[ "langchain_helper.generate_restaurant_name_and_items" ]
[((51, 88), 'streamlit.title', 'st.title', (['"""Restaurant Name Generator"""'], {}), "('Restaurant Name Generator')\n", (59, 88), True, 'import streamlit as st\n'), ((102, 200), 'streamlit.sidebar.selectbox', 'st.sidebar.selectbox', (['"""Pick a Cuisine"""', "('Indian', 'Italian', 'Mexican', 'Arabic', 'American')"], {...
import os import dotenv dotenv.load_dotenv() ### Load the credentials api_key = os.getenv("API_KEY", None) ibm_cloud_url = os.getenv("IBM_CLOUD_URL", None) project_id = os.getenv("PROJECT_ID", None) HUGGINGFACEHUB_API_TOKEN = os.getenv("HUGGINGFACEHUB_API_TOKEN", None) min_new_tokens=1 max_new_tokens=300 temperature...
[ "langchain.embeddings.HuggingFaceHubEmbeddings" ]
[((24, 44), 'dotenv.load_dotenv', 'dotenv.load_dotenv', ([], {}), '()\n', (42, 44), False, 'import dotenv\n'), ((81, 107), 'os.getenv', 'os.getenv', (['"""API_KEY"""', 'None'], {}), "('API_KEY', None)\n", (90, 107), False, 'import os\n'), ((124, 156), 'os.getenv', 'os.getenv', (['"""IBM_CLOUD_URL"""', 'None'], {}), "('...
from typing import Any, Dict, List, Optional from .few_shot_agent import FewShotAgent from .few_shot_agent import FewShotAgentExecutor from langchain import LLMChain from langchain.tools.base import BaseTool from typing import Any, Callable, Dict, List, Optional, Sequence, Tuple, Union from .prompts import * import nes...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((2122, 2142), 'nest_asyncio.apply', 'nest_asyncio.apply', ([], {}), '()\n', (2140, 2142), False, 'import nest_asyncio\n'), ((772, 804), 'langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler', 'StreamingStdOutCallbackHandler', ([], {}), '()\n', (802, 804), False, 'from langchain.callbacks.streaming_stdo...
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator impor...
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMRes...
[((2302, 2339), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2307, 2339), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2413, 2446), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, ...
from fastapi import FastAPI from langchain import ConversationChain from langchain.chat_models import ChatOpenAI from scripts.utils import MEMORY from scripts.doc_loader import load_document from lanarky import LangchainRouter from starlette.requests import Request from starlette.templating import Jinja2Templates from...
[ "langchain.chat_models.ChatOpenAI" ]
[((352, 369), 'config.set_environment', 'set_environment', ([], {}), '()\n', (367, 369), False, 'from config import set_environment\n'), ((377, 386), 'fastapi.FastAPI', 'FastAPI', ([], {}), '()\n', (384, 386), False, 'from fastapi import FastAPI\n'), ((570, 619), 'starlette.templating.Jinja2Templates', 'Jinja2Templates...
# %% import torch import os from llama_index.core import VectorStoreIndex, SimpleDirectoryReader from llama_index.llms.huggingface import HuggingFaceInferenceAPI from llama_index.core import Settings from langchain.embeddings.huggingface import HuggingFaceEmbeddings from llama_index.embeddings.langchain import Langch...
[ "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((1018, 1239), 'llama_index.core.PromptTemplate', 'PromptTemplate', (['"""Your job is to summarize different sections of the document given to you.Write a response that appropriately completes the request given to you.\n\n### Instruction:\n{query_str}\n\n### Response:"""'], {}), '(\n """Your job is to summarize dif...
import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Callable, Dict, List, Optional, Sequence import langchain from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, import_pandas, ...
[ "langchain.callbacks.utils.import_spacy", "langchain.callbacks.utils.import_pandas", "langchain.callbacks.utils.import_textstat", "langchain.callbacks.utils.flatten_dict" ]
[((1047, 1114), 'comet_ml.Experiment', 'comet_ml.Experiment', ([], {'workspace': 'workspace', 'project_name': 'project_name'}), '(workspace=workspace, project_name=project_name)\n', (1066, 1114), False, 'import comet_ml\n'), ((1249, 1266), 'langchain.callbacks.utils.import_textstat', 'import_textstat', ([], {}), '()\n'...
import streamlit as st from langchain import PromptTemplate from utils.studio_style import apply_studio_style from utils.studio_style import keyword_label, sentiment_label from utils import langchain from utils import bedrock from utils import config from datetime import datetime import pandas as pd import json import ...
[ "langchain.PromptTemplate" ]
[((329, 402), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Summarize Product Reviews"""', 'page_icon': '"""🛒"""'}), "(page_title='Summarize Product Reviews', page_icon='🛒')\n", (347, 402), True, 'import streamlit as st\n'), ((415, 438), 'utils.config.get_background', 'config.get_backgrou...
from langchain.llms import OpenAI from typing import Any, Dict, List, Optional import langchain from langchain import PromptTemplate ,LLMChain from langchain.chains.question_answering import load_qa_chain from langchain.chains.qa_with_sources import load_qa_with_sources_chain from langchain.chains import RetrievalQAWit...
[ "langchain.chains.question_answering.load_qa_chain", "langchain.LLMChain", "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.llms.LlamaCpp", "langchain.PromptTemplate" ]
[((404, 429), 'sys.path.append', 'sys.path.append', (['"""utils/"""'], {}), "('utils/')\n", (419, 429), False, 'import sys\n'), ((4364, 4456), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_template_llama2', 'input_variables': "['context', 'question']"}), "(template=prompt_template_llama2, inpu...
docs = """When and under what conditions can I apply to your graduate programs? Graduate student admissions are made in the fall and spring semesters specified in the academic calendar. Minimum application requirements: To have the undergraduate degree required in the program application requirements To have at least ...
[ "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.chat_models.ChatOpenAI", "langchain.cache.SQLiteCache" ]
[((2056, 2075), 'dotenv.load_dotenv', 'load_dotenv', (['""".env"""'], {}), "('.env')\n", (2067, 2075), False, 'from dotenv import load_dotenv\n'), ((2086, 2118), 'os.environ.get', 'os.environ.get', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (2100, 2118), False, 'import os\n'), ((2200, 2242), 'langchain.cac...
from typing import List, Optional, Tuple, Dict, Callable, Any, Union from functools import reduce import os import os from pathlib import Path import re from .utils import maybe_is_text, maybe_is_truncated from .qaprompts import ( summary_prompt, qa_prompt, search_prompt, citation_prompt, ...
[ "langchain.callbacks.get_openai_callback", "langchain.embeddings.openai.OpenAIEmbeddings", "langchain.chat_models.ChatOpenAI", "langchain.cache.SQLiteCache" ]
[((906, 929), 'langchain.cache.SQLiteCache', 'SQLiteCache', (['CACHE_PATH'], {}), '(CACHE_PATH)\n', (917, 929), False, 'from langchain.cache import SQLiteCache\n'), ((839, 866), 'os.path.dirname', 'os.path.dirname', (['CACHE_PATH'], {}), '(CACHE_PATH)\n', (854, 866), False, 'import os\n'), ((784, 795), 'pathlib.Path.ho...
import io import json import time from queue import Queue from typing import Dict, List import numpy as np import tiktoken from anyio.from_thread import start_blocking_portal from django.conf import settings from langchain.schema import AIMessage, HumanMessage from openai import OpenAI from pinecone import QueryRespon...
[ "langchain.schema.AIMessage", "langchain.schema.HumanMessage" ]
[((1714, 1744), 'openai.OpenAI', 'OpenAI', ([], {'api_key': 'openai_api_key'}), '(api_key=openai_api_key)\n', (1720, 1744), False, 'from openai import OpenAI\n'), ((3257, 3288), 'json.dumps', 'json.dumps', (['sanitized_reference'], {}), '(sanitized_reference)\n', (3267, 3288), False, 'import json\n'), ((3431, 3467), 't...
import json import os import langchain.memory.entity from langchain.chat_models import AzureChatOpenAI from flask import Flask, request import httpx from dotenv import load_dotenv from langchain.memory import ConversationSummaryBufferMemory, ConversationBufferWindowMemory from langchain.prompts.prompt import PromptTem...
[ "langchain.prompts.prompt.PromptTemplate", "langchain.LLMChain", "langchain.memory.ConversationBufferWindowMemory", "langchain.memory.ConversationSummaryBufferMemory", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.prompts.MessagesPlaceholder", "langchain.prompts.SystemMessageP...
[((560, 575), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (565, 575), False, 'from flask import Flask, request\n'), ((576, 615), 'dotenv.load_dotenv', 'load_dotenv', ([], {'dotenv_path': '"""./config.env"""'}), "(dotenv_path='./config.env')\n", (587, 615), False, 'from dotenv import load_dotenv\n'), ((2...
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import asyncio import functools import itertools import logging import uuid from enum import Enum from typing import ( Any, Callable, Coroutine, Dict, Iterator, List, Optional, Seque...
[ "langchain.schema.messages.messages_from_dict", "langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.chat_models.openai.ChatOpenAI", "langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandl...
[((1370, 1397), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1387, 1397), False, 'import logging\n'), ((1708, 1725), 'urllib.parse.urlparse', 'urlparse', (['api_url'], {}), '(api_url)\n', (1716, 1725), False, 'from urllib.parse import urlparse, urlunparse\n'), ((24648, 24668), 'asyncio...