id
stringlengths
6
6
text
stringlengths
20
17.2k
title
stringclasses
1 value
157653
"""Integration tests for the langchain tracer module.""" import asyncio import os from aiohttp import ClientSession from langchain_core.callbacks.manager import atrace_as_chain_group, trace_as_chain_group from langchain_core.prompts import PromptTemplate from langchain_core.tracers.context import tracing_v2_enabled ...
157675
"""Test AzureChatOpenAI wrapper.""" import os from typing import Any import pytest from langchain_core.callbacks import CallbackManager from langchain_core.messages import BaseMessage, HumanMessage from langchain_core.outputs import ChatGeneration, ChatResult, LLMResult from langchain_community.chat_models import Az...
157826
def test_chroma_update_document() -> None: """Test the update_document function in the Chroma class.""" # Make a consistent embedding embedding = ConsistentFakeEmbeddings() # Initial document content and id initial_content = "foo" document_id = "doc1" # Create an instance of Document with ...
157835
import importlib import os import time import uuid from typing import TYPE_CHECKING, List import numpy as np import pytest from langchain_core.documents import Document from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.vectorstores.pinecone import Pinecone if TYPE_CHECKING: imp...
157882
@pytest.mark.skipif(not singlestoredb_installed, reason="singlestoredb not installed") def test_singlestoredb_filter_metadata_7(texts: List[str]) -> None: """Test filtering by float""" table_name = "test_singlestoredb_filter_metadata_7" drop(table_name) docs = [ Document( page_conten...
157899
def test_add_texts_with_given_embedding(self, weaviate_url: str) -> None: texts = ["foo", "bar", "baz"] embedding = FakeEmbeddings() docsearch = Weaviate.from_texts( texts, embedding=embedding, weaviate_url=weaviate_url ) docsearch.add_texts(["foo"]) output ...
157903
"""Test Deep Lake functionality.""" import pytest from langchain_core.documents import Document from pytest import FixtureRequest from langchain_community.vectorstores import DeepLake from tests.integration_tests.vectorstores.fake_embeddings import FakeEmbeddings @pytest.fixture def deeplake_datastore() -> DeepLake...
158189
"""Test HuggingFace Pipeline wrapper.""" from pathlib import Path from langchain_community.llms.huggingface_pipeline import HuggingFacePipeline from langchain_community.llms.loading import load_llm from tests.integration_tests.llms.utils import assert_llm_equality def test_huggingface_pipeline_text_generation() -> ...
158221
def test_write_retrieve_keywords() -> None: from langchain_openai import OpenAIEmbeddings greetings = Document( id="greetings", page_content="Typical Greetings", metadata={ METADATA_LINKS_KEY: [ Link.incoming(kind="parent", tag="parent"), ], ...
158282
class AzureCosmosDBSemanticCache(BaseCache): """Cache that uses Cosmos DB Mongo vCore vector-store backend""" DEFAULT_DATABASE_NAME = "CosmosMongoVCoreCacheDB" DEFAULT_COLLECTION_NAME = "CosmosMongoVCoreCacheColl" def __init__( self, cosmosdb_connection_string: str, database_na...
158284
class SingleStoreDBSemanticCache(BaseCache): """Cache that uses SingleStore DB as a backend""" def __init__( self, embedding: Embeddings, *, cache_table_prefix: str = "cache_", search_threshold: float = 0.2, **kwargs: Any, ): """Initialize with necess...
158298
from typing import List, Optional import aiohttp import requests from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class RemoteLangChainRetriever(BaseR...
158301
"""Wrapper around Embedchain Retriever.""" from __future__ import annotations from typing import Any, Iterable, List, Optional from langchain_core.callbacks import CallbackManagerForRetrieverRun from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class EmbedchainRetrie...
158307
from typing import Any, List, Optional import aiohttp import requests from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class ChaindeskRetriever(BaseRe...
158309
class WebResearchRetriever(BaseRetriever): """`Google Search API` retriever.""" # Inputs vectorstore: VectorStore = Field( ..., description="Vector store for storing web pages" ) llm_chain: LLMChain search: GoogleSearchAPIWrapper = Field(..., description="Google Search API Wrapper") ...
158310
from typing import Any, Dict, List, cast from langchain_core.callbacks import CallbackManagerForRetrieverRun from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever from pydantic import Field class LlamaIndexRetriever(BaseRetriever): """`LlamaIndex` retriever. It is...
158315
from typing import List, Optional import aiohttp import requests from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class DataberryRetriever(BaseRetriev...
158322
from typing import Any, List from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever from langchain_community.utilities import YouSearchAPIWrapper class You...
158324
import os import re from typing import Any, Dict, List, Literal, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class AskNewsRetriever(Base...
158326
from __future__ import annotations import json from typing import Any, Dict, List, Optional import aiohttp import requests from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers ...
158328
# Unexpected keyword argument "extra" for "__init_subclass__" of "object" class RetrieveResult(BaseModel, extra="allow"): # type: ignore[call-arg] """`Amazon Kendra Retrieve API` search result. It is composed of: * relevant passages or text excerpts given an input query. """ QueryId: str ...
158422
# flake8: noqa QUERY_CHECKER = """ {query} Double check the {dialect} query above for common mistakes, including: - Using NOT IN with NULL values - Using UNION when UNION ALL should have been used - Using BETWEEN for exclusive ranges - Data type mismatch in predicates - Properly quoting identifiers - Using the correct ...
158610
"""HuggingFace sentence_transformer embedding models.""" from langchain_community.embeddings.huggingface import HuggingFaceEmbeddings SentenceTransformerEmbeddings = HuggingFaceEmbeddings
158647
@deprecated( since="0.0.9", removal="1.0", alternative_import="langchain_openai.OpenAIEmbeddings", ) class OpenAIEmbeddings(BaseModel, Embeddings): """OpenAI embedding models. To use, you should have the ``openai`` python package installed, and the environment variable ``OPENAI_API_KEY`` set wi...
158648
@pre_init def validate_environment(cls, values: Dict) -> Dict: """Validate that api key and python package exists in environment.""" values["openai_api_key"] = get_from_dict_or_env( values, "openai_api_key", "OPENAI_API_KEY" ) values["openai_api_base"] = values["openai_ap...
158651
"""Azure OpenAI embeddings wrapper.""" from __future__ import annotations import os import warnings from typing import Any, Callable, Dict, Optional, Union from langchain_core._api.deprecation import deprecated from langchain_core.utils import get_from_dict_or_env from pydantic import Field, model_validator from typ...
158655
# This file is adapted from # https://github.com/langchain-ai/langchain/blob/master/libs/community/langchain_community/embeddings/huggingface.py from typing import Any, Dict, List, Optional from langchain_core.embeddings import Embeddings from pydantic import BaseModel, ConfigDict, Field DEFAULT_BGE_MODEL = "BAAI/bg...
158657
import warnings from typing import Any, Dict, List, Optional import requests from langchain_core._api import deprecated, warn_deprecated from langchain_core.embeddings import Embeddings from pydantic import BaseModel, ConfigDict, Field, SecretStr DEFAULT_MODEL_NAME = "sentence-transformers/all-mpnet-base-v2" DEFAULT_...
158658
el, Embeddings): """HuggingFace sentence_transformers embedding models. To use, you should have the ``sentence_transformers`` python package installed. To use Nomic, make sure the version of ``sentence_transformers`` >= 2.3.0. Bge Example: .. code-block:: python from langchain_com...
158661
"""Wrapper around text2vec embedding models.""" from typing import Any, List, Optional from langchain_core.embeddings import Embeddings from pydantic import BaseModel class Text2vecEmbeddings(Embeddings, BaseModel): """text2vec embedding models. Install text2vec first, run 'pip install -U text2vec'. Th...
158675
from __future__ import annotations import logging from typing import Any, Dict, List, Optional from langchain_core.embeddings import Embeddings from langchain_core.utils import convert_to_secret_str, get_from_dict_or_env, pre_init from pydantic import BaseModel, Field, SecretStr logger = logging.getLogger(__name__) ...
158711
"""Callback Handler for LLMonitor`. #### Parameters: - `app_id`: The app id of the app you want to report to. Defaults to `None`, which means that `LLMONITOR_APP_ID` will be used. - `api_url`: The url of the LLMonitor API. Defaults to `None`, which means that either `LLMONITOR_API_U...
158720
"""Callback handler for Context AI""" import os from typing import Any, Dict, List from uuid import UUID from langchain_core.callbacks import BaseCallbackHandler from langchain_core.messages import BaseMessage from langchain_core.outputs import LLMResult from langchain_core.utils import guard_import def import_cont...
158723
def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run on agent action.""" self.step += 1 self.tool_starts += 1 self.starts += 1 resp: Dict[str, Any] = {} resp.update( { "action": "on_agent_action", "tool"...
158787
stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> Iterator[ChatGenerationChunk]: default_chunk_class = AIMessageChunk self.client.arun( [convert_message_to_dict(m) for m in messages], self.spark_user...
158829
"""**Chat Models** are a variation on language models. While Chat Models use language models under the hood, the interface they expose is a bit different. Rather than expose a "text in, text out" API, they expose an interface where "chat messages" are the inputs and outputs. **Class hierarchy:** .. code-block:: ...
158830
_module_lookup = { "AzureChatOpenAI": "langchain_community.chat_models.azure_openai", "BedrockChat": "langchain_community.chat_models.bedrock", "ChatAnthropic": "langchain_community.chat_models.anthropic", "ChatAnyscale": "langchain_community.chat_models.anyscale", "ChatBaichuan": "langchain_communi...
158844
class ChatOpenAI(BaseChatModel): """`OpenAI` Chat large language models API. To use, you should have the ``openai`` python package installed, and the environment variable ``OPENAI_API_KEY`` set with your API key. Any parameters that are valid to be passed to the openai.create call can be passed in...
158850
class AzureChatOpenAI(ChatOpenAI): """`Azure OpenAI` Chat Completion API. To use this class you must have a deployed model on Azure OpenAI. Use `deployment_name` in the constructor to refer to the "Model deployment name" in the Azure portal. In addition, you should have the ``openai`` python packa...
158894
@deprecated( since="0.3.1", removal="1.0.0", alternative_import="langchain_ollama.ChatOllama", ) class ChatOllama(BaseChatModel, _OllamaCommon): """Ollama locally runs large language models. To use, follow the instructions at https://ollama.ai/. Example: .. code-block:: python ...
158999
"""SQLAlchemy wrapper around a database.""" from __future__ import annotations from typing import Any, Dict, Iterable, List, Literal, Optional, Sequence, Union import sqlalchemy from langchain_core._api import deprecated from langchain_core.utils import get_from_env from sqlalchemy import ( MetaData, Table, ...
159019
import logging from typing import Any logger = logging.getLogger(__name__) def __getattr__(name: str) -> Any: if name in "PythonREPL": raise AssertionError( "PythonREPL has been deprecated from langchain_community due to being " "flagged by security scanners. See: " "h...
159055
async def adelete_by_metadata_filter( self, filter: dict[str, Any], *, batch_size: int = 50, ) -> int: """Delete all documents matching a certain metadata filtering condition. This operation does not use the vector embeddings in any way, it simply removes all...
159065
from __future__ import annotations import asyncio import base64 import itertools import json import logging import time import uuid from typing import ( TYPE_CHECKING, Any, Callable, ClassVar, Collection, Dict, Iterable, List, Literal, Optional, Tuple, Type, Union, ...
159066
class AzureSearch(VectorStore): """`Azure Cognitive Search` vector store.""" def __init__( self, azure_search_endpoint: str, azure_search_key: str, index_name: str, embedding_function: Union[Callable, Embeddings], search_type: str = "hybrid", semantic_con...
159070
async def asemantic_hybrid_search_with_score( self, query: str, k: int = 4, score_type: Literal["score", "reranker_score"] = "score", *, score_threshold: Optional[float] = None, **kwargs: Any, ) -> List[Tuple[Document, float]]: """ Returns the ...
159084
class Aerospike(VectorStore): """`Aerospike` vector store. To use, you should have the ``aerospike_vector_search`` python package installed. """ def __init__( self, client: Client, embedding: Union[Embeddings, Callable], namespace: str, index_name: Optional[str]...
159088
from __future__ import annotations import logging from typing import Any, Iterable, List, Optional, Tuple, Union from uuid import uuid4 import numpy as np from langchain_core._api.deprecation import deprecated from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from langchai...
159089
def __init__( self, embedding_function: Embeddings, collection_name: str = "LangChainCollection", collection_description: str = "", collection_properties: Optional[dict[str, Any]] = None, connection_args: Optional[dict[str, Any]] = None, consistency_level: str = "...
159090
def _create_collection( self, embeddings: list, metadatas: Optional[list[dict]] = None ) -> None: from pymilvus import ( Collection, CollectionSchema, DataType, FieldSchema, MilvusException, ) from pymilvus.orm.types import ...
159116
@deprecated( since="0.0.18", removal="1.0", alternative_import="langchain_pinecone.Pinecone" ) class Pinecone(VectorStore): """`Pinecone` vector store. To use, you should have the ``pinecone-client`` python package installed. This version of Pinecone is deprecated. Please use `langchain_pinecone.Pinec...
159118
import asyncio from typing import Any, Dict, Iterable, List, Optional, Tuple import numpy as np from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from langchain_core.vectorstores import VectorStore from langchain_community.vectorstores.utils import maximal_marginal_relevan...
159142
def similarity_search( self, query: str, k: int = 4, filter: Optional[dict] = None, **kwargs: Any, ) -> List[Document]: """Return documents most similar to query. Args: query: Text to look up documents similar to. k: Number of Document...
159150
def similarity_search_with_score( self, query: str, k: int = 4, filter: Optional[MetadataFilter] = None, search_params: Optional[common_types.SearchParams] = None, offset: int = 0, score_threshold: Optional[float] = None, consistency: Optional[common_types...
159151
def similarity_search_with_score_by_vector( self, embedding: List[float], k: int = 4, filter: Optional[MetadataFilter] = None, search_params: Optional[common_types.SearchParams] = None, offset: int = 0, score_threshold: Optional[float] = None, consistency:...
159156
@classmethod def construct_instance( cls: Type[Qdrant], texts: List[str], embedding: Embeddings, location: Optional[str] = None, url: Optional[str] = None, port: Optional[int] = 6333, grpc_port: int = 6334, prefer_grpc: bool = False, https: Opt...
159157
@classmethod async def aconstruct_instance( cls: Type[Qdrant], texts: List[str], embedding: Embeddings, location: Optional[str] = None, url: Optional[str] = None, port: Optional[int] = 6333, grpc_port: int = 6334, prefer_grpc: bool = False, htt...
159181
@deprecated( since="0.0.25", removal="1.0", alternative_import="langchain_mongodb.MongoDBAtlasVectorSearch", ) class MongoDBAtlasVectorSearch(VectorStore): """`MongoDB Atlas Vector Search` vector store. To use, you should have both: - the ``pymongo`` python package installed - a connection ...
159204
def _search_with_score_and_embeddings_by_vector( self, embedding: List[float], k: int = DEFAULT_TOP_K, filter: Optional[Dict[str, Any]] = None, brute_force: bool = False, fraction_lists_to_search: Optional[float] = None, ) -> List[Tuple[Document, List[float], float]]:...
159208
def _construct_documents_from_results_without_score( self, results: Dict[str, List[Dict[str, str]]] ) -> List[Document]: """Helper to convert Marqo results into documents. Args: results (List[dict]): A marqo results object with the 'hits'. include_scores (bool, optio...
159212
class DocumentDBVectorSearch(VectorStore): """`Amazon DocumentDB (with MongoDB compatibility)` vector store. Please refer to the official Vector Search documentation for more details: https://docs.aws.amazon.com/documentdb/latest/developerguide/vector-search.html To use, you should have both: - the...
159213
def _similarity_search_without_score( self, embeddings: List[float], k: int = 4, ef_search: int = 40, filter: Optional[Dict[str, Any]] = None, ) -> List[Document]: """Returns a list of documents. Args: embeddings: The query vector k: t...
159216
def similarity_search( self, query: str, k: int = 4, **kwargs: Any ) -> List[Document]: """Return docs most similar to query.""" documents = self.similarity_search_with_score(query=query, k=k) return [doc for doc, _ in documents] def similarity_search_with_score( self, q...
159227
add_texts( self, texts: Iterable[str], metadatas: Optional[List[dict]] = None, ids: Optional[List[str]] = None, **kwargs: Any, ) -> List[str]: """Run more texts through the embeddings and add to the vectorstore. Args: texts: Iterable of strings to...
159231
_select_relevance_score_fn(self) -> Callable[[float], float]: """ The 'correct' relevance function may differ depending on a few things, including: - the distance / similarity metric used by the VectorStore - the scale of your embeddings (OpenAI's are unit normed. Many others are...
159246
def similarity_search( self, query: str, k: int = 4, filter: Optional[Dict[str, Any]] = None, *, query_type: Optional[str] = None, **kwargs: Any, ) -> List[Document]: """Return docs most similar to query. Args: query: Text to look ...
159270
class Weaviate(VectorStore): """`Weaviate` vector store. To use, you should have the ``weaviate-client`` python package installed. Example: .. code-block:: python import weaviate from langchain_community.vectorstores import Weaviate client = weaviate.Client(ur...
159271
def max_marginal_relevance_search_by_vector( self, embedding: List[float], k: int = 4, fetch_k: int = 20, lambda_mult: float = 0.5, **kwargs: Any, ) -> List[Document]: """Return docs selected using the maximal marginal relevance. Maximal marginal rele...
159286
from __future__ import annotations import logging from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple, Union import numpy as np try: import deeplake from deeplake import VectorStore as DeepLakeVectorStore from deeplake.core.fast_forwarding import version_compare from deeplake.util...
159287
def add_texts( self, texts: Iterable[str], metadatas: Optional[List[dict]] = None, ids: Optional[List[str]] = None, **kwargs: Any, ) -> List[str]: """Run more texts through the embeddings and add to the vectorstore. Examples: >>> ids = deeplake_ve...
159288
def _search( self, query: Optional[str] = None, embedding: Optional[Union[List[float], np.ndarray]] = None, embedding_function: Optional[Callable] = None, k: int = 4, distance_metric: Optional[str] = None, use_maximal_marginal_relevance: bool = False, fetc...
159289
def similarity_search_by_vector( self, embedding: Union[List[float], np.ndarray], k: int = 4, **kwargs: Any, ) -> List[Document]: """ Return docs most similar to embedding vector. Examples: >>> # Search using an embedding >>> data = ve...
159300
def _similarity_search_with_score( self, embeddings: List[float], k: int = 4, pre_filter: Optional[Dict] = None, with_embedding: bool = False, ) -> List[Tuple[Document, float]]: query = "SELECT " # If limit_offset_clause is not specified, add TOP clause ...
159309
@override def similarity_search_by_vector( self, embedding: List[float], k: int = 4, **kwargs: Any ) -> List[Document]: """Returns the k most similar documents to the given embedding vector Args: embedding: The embedding vector to search for k: The number of simi...
159313
@classmethod def from_texts( cls, texts: List[str], embedding: Optional[Embeddings] = None, metadatas: Optional[List[dict]] = None, user_id: Optional[str] = None, app_id: Optional[str] = None, number_of_docs: Optional[int] = None, pat: Optional[str] = ...
159318
from __future__ import annotations import base64 import logging import uuid from typing import ( TYPE_CHECKING, Any, Callable, Dict, Iterable, List, Optional, Tuple, Type, ) import numpy as np from langchain_core._api import deprecated from langchain_core.documents import Document ...
159319
def add_texts( self, texts: Iterable[str], metadatas: Optional[List[dict]] = None, ids: Optional[List[str]] = None, **kwargs: Any, ) -> List[str]: """Run more texts through the embeddings and add to the vectorstore. Args: texts (Iterable[str]): Te...
159320
def similarity_search_by_image( self, uri: str, k: int = DEFAULT_K, filter: Optional[Dict[str, str]] = None, **kwargs: Any, ) -> List[Document]: """Search for similar images based on the given image URI. Args: uri (str): URI of the image to search...
159321
def update_document(self, document_id: str, document: Document) -> None: """Update a document in the collection. Args: document_id (str): ID of the document to update. document (Document): Document to update. """ return self.update_documents([document_id], [docum...
159327
def add_texts( self, texts: Iterable[str], metadatas: Optional[List[dict]] = None, ids: Optional[List[str]] = None, **kwargs: Any, ) -> List[str]: """Run more texts through the embeddings and add to the vectorstore. Args: texts: Iterable of string...
159333
import json import logging import numbers from hashlib import sha1 from typing import Any, Dict, Iterable, List, Optional, Tuple from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from langchain_core.vectorstores import VectorStore logger = logging.getLogger() class Aliba...
159342
def __init__( self, embedding: Embeddings, *, distance_strategy: DistanceStrategy = DEFAULT_DISTANCE_STRATEGY, table_name: str = "embeddings", content_field: str = "content", metadata_field: str = "metadata", vector_field: str = "vector", id_field:...
159344
def similarity_search( self, query: str, k: int = 4, filter: Optional[dict] = None, search_strategy: SearchStrategy = SearchStrategy.VECTOR_ONLY, filter_threshold: float = 0, text_weight: float = 0.5, vector_weight: float = 0.5, vector_select_count...
159345
def similarity_search_with_score( self, query: str, k: int = 4, filter: Optional[dict] = None, search_strategy: SearchStrategy = SearchStrategy.VECTOR_ONLY, filter_threshold: float = 1, text_weight: float = 0.5, vector_weight: float = 0.5, vector_s...
159354
from __future__ import annotations import logging import uuid from typing import TYPE_CHECKING, Any, Iterable, List, Optional, Tuple, Union, cast import numpy as np from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from langchain_core.utils.iter import batch_iterate from l...
159362
from __future__ import annotations import logging from typing import Any, Dict, List, Optional from langchain_core.embeddings import Embeddings from langchain_community.vectorstores.milvus import Milvus logger = logging.getLogger(__name__) class Zilliz(Milvus): """`Zilliz` vector store. You need to have ...
159372
def add_texts( self, texts: Iterable[str], metadatas: Optional[List[Dict[str, Any]]] = None, ids: Optional[List[str]] = None, batch_size: Optional[int] = None, **kwargs: Any, ) -> List[str]: """Run texts through the embeddings and persist in vectorstore. ...
159375
@deprecated( "0.0.27", alternative="Use ElasticsearchStore class in langchain-elasticsearch package", pending=True, ) class ElasticVectorSearch(VectorStore): """ ElasticVectorSearch uses the brute force method of searching on vectors. Recommended to use ElasticsearchStore instead, which gives ...
159376
@classmethod def from_texts( cls, texts: List[str], embedding: Embeddings, metadatas: Optional[List[dict]] = None, ids: Optional[List[str]] = None, index_name: Optional[str] = None, refresh_indices: bool = True, **kwargs: Any, ) -> ElasticVectorSea...
159388
class Redis(VectorStore): """Redis vector database. Deployment Options: Below, we will use a local deployment as an example. However, Redis can be deployed in all of the following ways: - [Redis Cloud](https://redis.com/redis-enterprise-cloud/overview/) - [Docker (Redis Stack)](https:/...
159390
ssmethod def from_texts( cls: Type[Redis], texts: List[str], embedding: Embeddings, metadatas: Optional[List[dict]] = None, index_name: Optional[str] = None, index_schema: Optional[Union[Dict[str, ListOfDict], str, os.PathLike]] = None, vector_schema: Optional...
159392
similarity_search_by_vector( self, embedding: List[float], k: int = 4, filter: Optional[RedisFilterExpression] = None, return_metadata: bool = True, distance_threshold: Optional[float] = None, **kwargs: Any, ) -> List[Document]: """Run similarity searc...
159394
s RedisVectorStoreRetriever(VectorStoreRetriever): """Retriever for Redis VectorStore.""" vectorstore: Redis """Redis VectorStore.""" search_type: str = "similarity" """Type of search to perform. Can be either 'similarity', 'similarity_distance_threshold', 'similarity_score_threshold' ...
159402
from abc import ABC from typing import TYPE_CHECKING, Any, Iterable, List, Optional, Tuple, Type import numpy as np from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from langchain_core.vectorstores import VectorStore from pydantic import Field from langchain_community.vec...
159461
class ConfluenceLoader(BaseLoader): """Load `Confluence` pages. Port of https://llamahub.ai/l/confluence This currently supports username/api_key, Oauth2 login or personal access token authentication. Specify a list page_ids and/or space_key to load in the corresponding pages into Document obj...
159493
from __future__ import annotations from pathlib import Path from typing import ( TYPE_CHECKING, Any, Iterator, List, Literal, Optional, Sequence, Union, ) from langchain_core.documents import Document from langchain_community.document_loaders.base import BaseBlobParser, BaseLoader fro...
159501
"""Loader that uses unstructured to load files.""" from __future__ import annotations import logging import os from abc import ABC, abstractmethod from pathlib import Path from typing import IO, Any, Callable, Iterator, List, Optional, Sequence, Union from langchain_core._api.deprecation import deprecated from langc...
159504
import json import logging import time from typing import Iterator, List import requests from langchain_core.documents import Document from langchain_community.document_loaders.base import BaseLoader logger = logging.getLogger(__name__) class CubeSemanticLoader(BaseLoader): """Load `Cube semantic layer` metada...
159511
from langchain_community.document_loaders.notiondb import ( NotionDBLoader, ) from langchain_community.document_loaders.obs_directory import ( OBSDirectoryLoader, ) from langchain_community.document_loaders.obs_file import ( OBSFileLoader, ) from langchain_community.docum...
159518
import os import tempfile from typing import List from langchain_core.documents import Document from langchain_community.document_loaders.base import BaseLoader from langchain_community.document_loaders.unstructured import UnstructuredFileLoader class AzureBlobStorageFileLoader(BaseLoader): """Load from `Azure ...