| | """A simple script to run a Flow that can be used for development and debugging.""" |
| |
|
| | import os |
| |
|
| | import hydra |
| |
|
| | import aiflows |
| | from aiflows.backends.api_info import ApiInfo |
| | from aiflows.utils.general_helpers import read_yaml_file, quick_load_api_keys |
| |
|
| | from aiflows import logging |
| | from aiflows.flow_cache import CACHING_PARAMETERS, clear_cache |
| |
|
| | from aiflows.utils import serve_utils |
| | from aiflows.workers import run_dispatch_worker_thread |
| | from aiflows.messages import FlowMessage |
| | from aiflows.interfaces import KeyInterface |
| | from aiflows.utils.colink_utils import start_colink_server |
| | from aiflows.workers import run_dispatch_worker_thread |
| |
|
| | CACHING_PARAMETERS.do_caching = False |
| | |
| |
|
| |
|
| |
|
| |
|
| | dependencies = [ |
| | {"url": "aiflows/VectorStoreFlowModule", "revision": os.getcwd()} |
| | ] |
| |
|
| | from aiflows import flow_verse |
| | flow_verse.sync_dependencies(dependencies) |
| | if __name__ == "__main__": |
| | |
| | |
| | |
| | cl = start_colink_server() |
| |
|
| |
|
| | |
| | root_dir = "." |
| | cfg_path = os.path.join(root_dir, "demo.yaml") |
| | cfg = read_yaml_file(cfg_path) |
| | |
| | |
| | |
| | api_information = [ApiInfo(backend_used="openai", |
| | api_key = os.getenv("OPENAI_API_KEY"))] |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | quick_load_api_keys(cfg, api_information, key="api_infos") |
| |
|
| | |
| | |
| | serve_utils.serve_flow( |
| | cl = cl, |
| | flow_class_name="flow_modules.aiflows.VectorStoreFlowModule.ChromaDBFlow", |
| | flow_endpoint="ChromaDBFlow", |
| | ) |
| | |
| | |
| | run_dispatch_worker_thread(cl) |
| | |
| |
|
| | |
| | proxy_flow_cdb= serve_utils.get_flow_instance( |
| | cl=cl, |
| | flow_endpoint="ChromaDBFlow", |
| | user_id="local", |
| | config_overrides = cfg["chroma_demo_flow"] |
| | ) |
| | |
| | |
| | serve_utils.serve_flow( |
| | cl = cl, |
| | flow_class_name="flow_modules.aiflows.VectorStoreFlowModule.VectorStoreFlow", |
| | flow_endpoint="VectorStoreFlow", |
| | ) |
| | |
| | |
| | |
| | run_dispatch_worker_thread(cl) |
| |
|
| | |
| | proxy_flow_vs= serve_utils.get_flow_instance( |
| | cl=cl, |
| | flow_endpoint="VectorStoreFlow", |
| | user_id="local", |
| | config_overrides = cfg["vector_store_demo_flow"], |
| | ) |
| | |
| | |
| | |
| | data_write = {"id": 1, "operation": "write", "content": "The Capital of Switzerland is Bern"} |
| | data_read1 = {"id": 1, "operation": "read", "content": "Switzerland"} |
| | data_read2 = {"id": 3, "operation": "read", "content": "What did the author do growing up?"} |
| | |
| | data = [data_read2,data_write,data_read1] |
| | |
| | futures = [] |
| | |
| | print("##########CHROMA DB DEMO###############") |
| | for dp in data: |
| | |
| | input_message = FlowMessage( |
| | data=data_write, |
| | ) |
| | futures.append(proxy_flow_cdb.get_reply_future(input_message)) |
| | |
| | replies = [ft.get_data() for ft in futures] |
| | for dp,rp in zip(data, replies): |
| | print("~~~~~ Message Sent~~~~~") |
| | print(dp) |
| | print("~~~~~ Replies ~~~~~") |
| | print(rp) |
| | |
| | |
| | |
| | print("##########VECTOR STORE DEMO##############") |
| | for dp in data: |
| | |
| | input_message = FlowMessage( |
| | data=data_write, |
| | ) |
| | futures.append(proxy_flow_vs.get_reply_future(input_message)) |
| | |
| | replies = [ft.get_data() for ft in futures] |
| | for dp,rp in zip(data, replies): |
| | print("~~~~~ Message Sent~~~~~") |
| | print(dp) |
| | print("~~~~~ Replies ~~~~~") |
| | print(rp) |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |