code
stringlengths
141
97.3k
apis
listlengths
1
24
extract_api
stringlengths
113
214k
from typing import Any, List, Sequence, Tuple, Union from langchain_core._api import deprecated from langchain_core.agents import AgentAction, AgentFinish from langchain_core.callbacks import Callbacks from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts.base import BasePromptTempla...
[ "langchain_core.prompts.chat.AIMessagePromptTemplate.from_template", "langchain.agents.format_scratchpad.format_xml", "langchain.agents.output_parsers.XMLAgentOutputParser", "langchain_core._api.deprecated", "langchain_core.prompts.chat.ChatPromptTemplate.from_template" ]
[((875, 943), 'langchain_core._api.deprecated', 'deprecated', (['"""0.1.0"""'], {'alternative': '"""create_xml_agent"""', 'removal': '"""0.2.0"""'}), "('0.1.0', alternative='create_xml_agent', removal='0.2.0')\n", (885, 943), False, 'from langchain_core._api import deprecated\n'), ((1644, 1696), 'langchain_core.prompts...
from typing import Any, List, Sequence, Tuple, Union from langchain_core._api import deprecated from langchain_core.agents import AgentAction, AgentFinish from langchain_core.callbacks import Callbacks from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts.base import BasePromptTempla...
[ "langchain_core.prompts.chat.AIMessagePromptTemplate.from_template", "langchain.agents.format_scratchpad.format_xml", "langchain.agents.output_parsers.XMLAgentOutputParser", "langchain_core._api.deprecated", "langchain_core.prompts.chat.ChatPromptTemplate.from_template" ]
[((875, 943), 'langchain_core._api.deprecated', 'deprecated', (['"""0.1.0"""'], {'alternative': '"""create_xml_agent"""', 'removal': '"""0.2.0"""'}), "('0.1.0', alternative='create_xml_agent', removal='0.2.0')\n", (885, 943), False, 'from langchain_core._api import deprecated\n'), ((1644, 1696), 'langchain_core.prompts...
"""**Graphs** provide a natural language interface to graph databases.""" import warnings from typing import Any from langchain_core._api import LangChainDeprecationWarning from langchain.utils.interactive_env import is_interactive_env def __getattr__(name: str) -> Any: from langchain_community import graphs ...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((378, 398), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (396, 398), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((408, 773), 'warnings.warn', 'warnings.warn', (['f"""Importing graphs from langchain is deprecated. Importing from langchai...
"""**Graphs** provide a natural language interface to graph databases.""" import warnings from typing import Any from langchain_core._api import LangChainDeprecationWarning from langchain.utils.interactive_env import is_interactive_env def __getattr__(name: str) -> Any: from langchain_community import graphs ...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((378, 398), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (396, 398), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((408, 773), 'warnings.warn', 'warnings.warn', (['f"""Importing graphs from langchain is deprecated. Importing from langchai...
"""**Graphs** provide a natural language interface to graph databases.""" import warnings from typing import Any from langchain_core._api import LangChainDeprecationWarning from langchain.utils.interactive_env import is_interactive_env def __getattr__(name: str) -> Any: from langchain_community import graphs ...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((378, 398), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (396, 398), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((408, 773), 'warnings.warn', 'warnings.warn', (['f"""Importing graphs from langchain is deprecated. Importing from langchai...
"""**Graphs** provide a natural language interface to graph databases.""" import warnings from typing import Any from langchain_core._api import LangChainDeprecationWarning from langchain.utils.interactive_env import is_interactive_env def __getattr__(name: str) -> Any: from langchain_community import graphs ...
[ "langchain.utils.interactive_env.is_interactive_env" ]
[((378, 398), 'langchain.utils.interactive_env.is_interactive_env', 'is_interactive_env', ([], {}), '()\n', (396, 398), False, 'from langchain.utils.interactive_env import is_interactive_env\n'), ((408, 773), 'warnings.warn', 'warnings.warn', (['f"""Importing graphs from langchain is deprecated. Importing from langchai...
"""Chain that makes API calls and summarizes the responses to answer a question.""" from __future__ import annotations from typing import Any, Dict, List, Optional, Sequence, Tuple from urllib.parse import urlparse from langchain_community.utilities.requests import TextRequestsWrapper from langchain_core.callbacks im...
[ "langchain.chains.llm.LLMChain", "langchain_core.callbacks.AsyncCallbackManagerForChainRun.get_noop_manager", "langchain_community.utilities.requests.TextRequestsWrapper", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager", "langchain_core.pydantic_v1.root_validator", "langchain_core.p...
[((979, 992), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (987, 992), False, 'from urllib.parse import urlparse\n'), ((2555, 2574), 'langchain_core.pydantic_v1.Field', 'Field', ([], {'exclude': '(True)'}), '(exclude=True)\n', (2560, 2574), False, 'from langchain_core.pydantic_v1 import Field, root_va...
"""Chain that makes API calls and summarizes the responses to answer a question.""" from __future__ import annotations from typing import Any, Dict, List, Optional, Sequence, Tuple from urllib.parse import urlparse from langchain_community.utilities.requests import TextRequestsWrapper from langchain_core.callbacks im...
[ "langchain.chains.llm.LLMChain", "langchain_core.callbacks.AsyncCallbackManagerForChainRun.get_noop_manager", "langchain_community.utilities.requests.TextRequestsWrapper", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager", "langchain_core.pydantic_v1.root_validator", "langchain_core.p...
[((979, 992), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (987, 992), False, 'from urllib.parse import urlparse\n'), ((2555, 2574), 'langchain_core.pydantic_v1.Field', 'Field', ([], {'exclude': '(True)'}), '(exclude=True)\n', (2560, 2574), False, 'from langchain_core.pydantic_v1 import Field, root_va...
"""Chain that makes API calls and summarizes the responses to answer a question.""" from __future__ import annotations from typing import Any, Dict, List, Optional, Sequence, Tuple from urllib.parse import urlparse from langchain_community.utilities.requests import TextRequestsWrapper from langchain_core.callbacks im...
[ "langchain.chains.llm.LLMChain", "langchain_core.callbacks.AsyncCallbackManagerForChainRun.get_noop_manager", "langchain_community.utilities.requests.TextRequestsWrapper", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager", "langchain_core.pydantic_v1.root_validator", "langchain_core.p...
[((979, 992), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (987, 992), False, 'from urllib.parse import urlparse\n'), ((2555, 2574), 'langchain_core.pydantic_v1.Field', 'Field', ([], {'exclude': '(True)'}), '(exclude=True)\n', (2560, 2574), False, 'from langchain_core.pydantic_v1 import Field, root_va...
"""Chain that makes API calls and summarizes the responses to answer a question.""" from __future__ import annotations from typing import Any, Dict, List, Optional, Sequence, Tuple from urllib.parse import urlparse from langchain_community.utilities.requests import TextRequestsWrapper from langchain_core.callbacks im...
[ "langchain.chains.llm.LLMChain", "langchain_core.callbacks.AsyncCallbackManagerForChainRun.get_noop_manager", "langchain_community.utilities.requests.TextRequestsWrapper", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager", "langchain_core.pydantic_v1.root_validator", "langchain_core.p...
[((979, 992), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (987, 992), False, 'from urllib.parse import urlparse\n'), ((2555, 2574), 'langchain_core.pydantic_v1.Field', 'Field', ([], {'exclude': '(True)'}), '(exclude=True)\n', (2560, 2574), False, 'from langchain_core.pydantic_v1 import Field, root_va...
"""Hypothetical Document Embeddings. https://arxiv.org/abs/2212.10496 """ from __future__ import annotations from typing import Any, Dict, List, Optional import numpy as np from langchain_core.callbacks import CallbackManagerForChainRun from langchain_core.embeddings import Embeddings from langchain_core.language_mo...
[ "langchain.chains.llm.LLMChain", "langchain.chains.hyde.prompts.PROMPT_MAP.keys", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager" ]
[((3148, 3180), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt)\n', (3156, 3180), False, 'from langchain.chains.llm import LLMChain\n'), ((2258, 2303), 'langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager', 'CallbackManagerForChainRun.get...
"""Hypothetical Document Embeddings. https://arxiv.org/abs/2212.10496 """ from __future__ import annotations from typing import Any, Dict, List, Optional import numpy as np from langchain_core.callbacks import CallbackManagerForChainRun from langchain_core.embeddings import Embeddings from langchain_core.language_mo...
[ "langchain.chains.llm.LLMChain", "langchain.chains.hyde.prompts.PROMPT_MAP.keys", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager" ]
[((3148, 3180), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt)\n', (3156, 3180), False, 'from langchain.chains.llm import LLMChain\n'), ((2258, 2303), 'langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager', 'CallbackManagerForChainRun.get...
"""Hypothetical Document Embeddings. https://arxiv.org/abs/2212.10496 """ from __future__ import annotations from typing import Any, Dict, List, Optional import numpy as np from langchain_core.callbacks import CallbackManagerForChainRun from langchain_core.embeddings import Embeddings from langchain_core.language_mo...
[ "langchain.chains.llm.LLMChain", "langchain.chains.hyde.prompts.PROMPT_MAP.keys", "langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager" ]
[((3148, 3180), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt)\n', (3156, 3180), False, 'from langchain.chains.llm import LLMChain\n'), ((2258, 2303), 'langchain_core.callbacks.CallbackManagerForChainRun.get_noop_manager', 'CallbackManagerForChainRun.get...
"""Attempt to implement MRKL systems as described in arxiv.org/pdf/2205.00445.pdf.""" from __future__ import annotations from typing import Any, Callable, List, NamedTuple, Optional, Sequence from langchain_core._api import deprecated from langchain_core.callbacks import BaseCallbackManager from langchain_core.langua...
[ "langchain.agents.mrkl.output_parser.MRKLOutputParser", "langchain.chains.LLMChain", "langchain.agents.utils.validate_tools_single_input", "langchain_core.prompts.PromptTemplate", "langchain.agents.tools.Tool", "langchain_core._api.deprecated", "langchain_core.prompts.PromptTemplate.from_template", "l...
[((1278, 1348), 'langchain_core._api.deprecated', 'deprecated', (['"""0.1.0"""'], {'alternative': '"""create_react_agent"""', 'removal': '"""0.2.0"""'}), "('0.1.0', alternative='create_react_agent', removal='0.2.0')\n", (1288, 1348), False, 'from langchain_core._api import deprecated\n'), ((5068, 5104), 'langchain_core...
"""Attempt to implement MRKL systems as described in arxiv.org/pdf/2205.00445.pdf.""" from __future__ import annotations from typing import Any, Callable, List, NamedTuple, Optional, Sequence from langchain_core._api import deprecated from langchain_core.callbacks import BaseCallbackManager from langchain_core.langua...
[ "langchain.agents.mrkl.output_parser.MRKLOutputParser", "langchain.chains.LLMChain", "langchain.agents.utils.validate_tools_single_input", "langchain_core.prompts.PromptTemplate", "langchain.agents.tools.Tool", "langchain_core._api.deprecated", "langchain_core.prompts.PromptTemplate.from_template", "l...
[((1278, 1348), 'langchain_core._api.deprecated', 'deprecated', (['"""0.1.0"""'], {'alternative': '"""create_react_agent"""', 'removal': '"""0.2.0"""'}), "('0.1.0', alternative='create_react_agent', removal='0.2.0')\n", (1288, 1348), False, 'from langchain_core._api import deprecated\n'), ((5068, 5104), 'langchain_core...
import streamlit as st import datetime import os import psycopg2 from dotenv import load_dotenv from langchain.prompts import PromptTemplate from langchain.docstore.document import Document def log(message): current_time = datetime.datetime.now() milliseconds = current_time.microsecond // 1000 timestamp ...
[ "langchain.docstore.document.Document", "langchain.prompts.PromptTemplate" ]
[((2668, 2806), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['input_question', 'table_info', 'columns_info', 'top_k', 'no_answer_text']", 'template': '_postgres_prompt'}), "(input_variables=['input_question', 'table_info',\n 'columns_info', 'top_k', 'no_answer_text'], template=_po...
import os import pandas as pd from langchain.chains import LLMChain from langchain.llms import OpenAI from langchain.prompts import PromptTemplate import mlflow assert ( "OPENAI_API_KEY" in os.environ ), "Please set the OPENAI_API_KEY environment variable to run this example." def build_and_evalute_model_with_...
[ "langchain.llms.OpenAI", "langchain.prompts.PromptTemplate", "langchain.chains.LLMChain" ]
[((1832, 1932), 'mlflow.load_table', 'mlflow.load_table', (['"""eval_results_table.json"""'], {'extra_columns': "['run_id', 'params.prompt_template']"}), "('eval_results_table.json', extra_columns=['run_id',\n 'params.prompt_template'])\n", (1849, 1932), False, 'import mlflow\n'), ((349, 367), 'mlflow.start_run', 'm...
import os import voyager.utils as U from langchain.chat_models import ChatOpenAI from langchain.embeddings.openai import OpenAIEmbeddings from langchain.schema import HumanMessage, SystemMessage from langchain.vectorstores import Chroma from voyager.prompts import load_prompt from voyager.control_primitives import lo...
[ "langchain.embeddings.openai.OpenAIEmbeddings", "langchain.schema.HumanMessage", "langchain.chat_models.ChatOpenAI" ]
[((583, 678), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': 'model_name', 'temperature': 'temperature', 'request_timeout': 'request_timout'}), '(model_name=model_name, temperature=temperature, request_timeout=\n request_timout)\n', (593, 678), False, 'from langchain.chat_models import ChatOpe...
from langflow import CustomComponent from langchain.agents import AgentExecutor, create_json_agent from langflow.field_typing import ( BaseLanguageModel, ) from langchain_community.agent_toolkits.json.toolkit import JsonToolkit class JsonAgentComponent(CustomComponent): display_name = "JsonAgent" descript...
[ "langchain.agents.create_json_agent" ]
[((657, 700), 'langchain.agents.create_json_agent', 'create_json_agent', ([], {'llm': 'llm', 'toolkit': 'toolkit'}), '(llm=llm, toolkit=toolkit)\n', (674, 700), False, 'from langchain.agents import AgentExecutor, create_json_agent\n')]
import os from fedml.serving import FedMLPredictor from fedml.serving import FedMLInferenceRunner from langchain import PromptTemplate, LLMChain from langchain.llms import HuggingFacePipeline import torch from transformers import ( AutoConfig, AutoModelForCausalLM, AutoTokenizer, TextGenerationPipeline,...
[ "langchain.LLMChain", "langchain.PromptTemplate" ]
[((2184, 2209), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (2207, 2209), False, 'import argparse\n'), ((2559, 2588), 'fedml.serving.FedMLInferenceRunner', 'FedMLInferenceRunner', (['chatbot'], {}), '(chatbot)\n', (2579, 2588), False, 'from fedml.serving import FedMLInferenceRunner\n'), ((70...
from langchain.utilities import BashProcess from langchain.agents import load_tools def get_built_in_tools(tools: list[str]): bash = BashProcess() load_tools(["human"]) return [bash]
[ "langchain.agents.load_tools", "langchain.utilities.BashProcess" ]
[((139, 152), 'langchain.utilities.BashProcess', 'BashProcess', ([], {}), '()\n', (150, 152), False, 'from langchain.utilities import BashProcess\n'), ((158, 179), 'langchain.agents.load_tools', 'load_tools', (["['human']"], {}), "(['human'])\n", (168, 179), False, 'from langchain.agents import load_tools\n')]
# # Copyright 2016 The BigDL Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in ...
[ "langchain.llms.utils.enforce_stop_tokens" ]
[((5354, 5476), 'transformers.pipeline', 'hf_pipeline', ([], {'task': 'task', 'model': 'model', 'tokenizer': 'tokenizer', 'device': '"""cpu"""', 'model_kwargs': '_model_kwargs'}), "(task=task, model=model, tokenizer=tokenizer, device='cpu',\n model_kwargs=_model_kwargs, **_pipeline_kwargs)\n", (5365, 5476), True, 'f...
""" This module provides an implementation for generating question data from documents. Supported types of document sources include: - plain text - unstructured files: Text, PDF, PowerPoint, HTML, Images, Excel spreadsheets, Word documents, Markdown, etc. - documents from Google Drive (provide file...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.schema.Document", "langchain.document_loaders.GoogleDriveLoader", "langchain.document_loaders.UnstructuredFileLoader" ]
[((7802, 7926), 'yival.data_generators.base_data_generator.BaseDataGenerator.register_data_generator', 'BaseDataGenerator.register_data_generator', (['"""document_data_generator"""', 'DocumentDataGenerator', 'DocumentDataGeneratorConfig'], {}), "('document_data_generator',\n DocumentDataGenerator, DocumentDataGenera...
from typing import AsyncGenerator, Optional, Tuple from langchain import ConversationChain import logging from typing import Optional, Tuple from pydantic.v1 import SecretStr from vocode.streaming.agent.base_agent import RespondAgent from vocode.streaming.agent.utils import get_sentence_from_buffer from langchain im...
[ "langchain.schema.AIMessage", "langchain.ConversationChain", "langchain.schema.HumanMessage", "langchain_community.chat_models.ChatAnthropic", "langchain.prompts.MessagesPlaceholder", "langchain.memory.ConversationBufferMemory", "langchain.prompts.HumanMessagePromptTemplate.from_template" ]
[((2147, 2238), 'langchain_community.chat_models.ChatAnthropic', 'ChatAnthropic', ([], {'model_name': 'agent_config.model_name', 'anthropic_api_key': 'anthropic_api_key'}), '(model_name=agent_config.model_name, anthropic_api_key=\n anthropic_api_key)\n', (2160, 2238), False, 'from langchain_community.chat_models imp...
from typing import Any, Dict from langchain.base_language import BaseLanguageModel from langchain.prompts import ( ChatPromptTemplate, HumanMessagePromptTemplate, MessagesPlaceholder, SystemMessagePromptTemplate, ) from langchain.chains import ConversationChain from real_agents.adapters.exe...
[ "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.prompts.SystemMessagePromptTemplate.from_template", "langchain.prompts.MessagesPlaceholder", "langchain.chains.ConversationChain" ]
[((894, 940), 'real_agents.adapters.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'return_messages': '(True)'}), '(return_messages=True)\n', (918, 940), False, 'from real_agents.adapters.memory import ConversationBufferMemory\n'), ((1746, 1824), 'langchain.chains.ConversationChain', 'ConversationC...
import os from dotenv import load_dotenv, find_dotenv from langchain import HuggingFaceHub from langchain import PromptTemplate, LLMChain, OpenAI from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.chains.summarize import load_summarize_chain from langchain.document_loaders import YoutubeL...
[ "langchain.PromptTemplate", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.YoutubeLoader.from_youtube_url", "langchain.chains.summarize.load_summarize_chain", "langchain.LLMChain", "langchain.OpenAI", "langchain.HuggingFaceHub" ]
[((955, 1048), 'langchain.HuggingFaceHub', 'HuggingFaceHub', ([], {'repo_id': 'repo_id', 'model_kwargs': "{'temperature': 0.1, 'max_new_tokens': 500}"}), "(repo_id=repo_id, model_kwargs={'temperature': 0.1,\n 'max_new_tokens': 500})\n", (969, 1048), False, 'from langchain import HuggingFaceHub\n'), ((1305, 1368), 'l...
from dotenv import load_dotenv from langchain import OpenAI from langchain.document_loaders.csv_loader import CSVLoader load_dotenv() filepath = "academy/academy.csv" loader = CSVLoader(filepath) data = loader.load() print(data) llm = OpenAI(temperature=0) from langchain.agents import create_csv_agent agent = crea...
[ "langchain.document_loaders.csv_loader.CSVLoader", "langchain.agents.create_csv_agent", "langchain.OpenAI" ]
[((122, 135), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (133, 135), False, 'from dotenv import load_dotenv\n'), ((179, 198), 'langchain.document_loaders.csv_loader.CSVLoader', 'CSVLoader', (['filepath'], {}), '(filepath)\n', (188, 198), False, 'from langchain.document_loaders.csv_loader import CSVLoader\n'...
from waifu.llm.Brain import Brain from waifu.llm.VectorDB import VectorDB from waifu.llm.SentenceTransformer import STEmbedding from langchain.chat_models import ChatOpenAI from langchain.embeddings import OpenAIEmbeddings from typing import Any, List, Mapping, Optional from langchain.schema import BaseMessage import o...
[ "langchain.embeddings.OpenAIEmbeddings", "langchain.chat_models.ChatOpenAI" ]
[((576, 690), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': 'api_key', 'model_name': 'model', 'streaming': 'stream', 'callbacks': '[callback]', 'temperature': '(0.85)'}), '(openai_api_key=api_key, model_name=model, streaming=stream,\n callbacks=[callback], temperature=0.85)\n', (586, 690)...
import re from typing import Union from langchain.schema import AgentAction, AgentFinish, OutputParserException from src.agents.agent import AgentOutputParser class ReActOutputParser(AgentOutputParser): """Output parser for the ReAct agent.""" def parse(self, text: str) -> Union[AgentAction, AgentFinish]: ...
[ "langchain.schema.AgentAction", "langchain.schema.AgentFinish", "langchain.schema.OutputParserException" ]
[((685, 726), 're.search', 're.search', (['"""(.*?)\\\\[(.*?)\\\\]"""', 'action_str'], {}), "('(.*?)\\\\[(.*?)\\\\]', action_str)\n", (694, 726), False, 'import re\n'), ((444, 504), 'langchain.schema.OutputParserException', 'OutputParserException', (['f"""Could not parse LLM Output: {text}"""'], {}), "(f'Could not pars...
import re from langchain.agents import AgentOutputParser from langchain.schema import AgentAction, AgentFinish, OutputParserException from typing import Union from cat.mad_hatter.mad_hatter import MadHatter from cat.log import log class ChooseProcedureOutputParser(AgentOutputParser): def parse(self, llm_output:...
[ "langchain.schema.AgentFinish", "langchain.schema.OutputParserException" ]
[((936, 975), 're.search', 're.search', (['regex', 'llm_output', 're.DOTALL'], {}), '(regex, llm_output, re.DOTALL)\n', (945, 975), False, 'import re\n'), ((1551, 1562), 'cat.mad_hatter.mad_hatter.MadHatter', 'MadHatter', ([], {}), '()\n', (1560, 1562), False, 'from cat.mad_hatter.mad_hatter import MadHatter\n'), ((101...
from typing import List, Union from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from pydantic import BaseModel, Extra, validator from mindsdb.integrations.handlers.rag_handler.settings import ( DEFAULT_EMBEDDINGS_MODEL, RAGBaseParameters, ) EVAL_COLUMN_NAMES = ( "question",...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((1785, 1845), 'pydantic.validator', 'validator', (['"""generation_evaluation_metrics"""'], {'allow_reuse': '(True)'}), "('generation_evaluation_metrics', allow_reuse=True)\n", (1794, 1845), False, 'from pydantic import BaseModel, Extra, validator\n'), ((2190, 2249), 'pydantic.validator', 'validator', (['"""retrieval_...
from typing import List, Optional, Mapping, Any from functools import partial from langchain.llms.base import LLM from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from transformers import AutoModel, AutoTokenizer from peft i...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((3052, 3122), 'transformers.AutoTokenizer.from_pretrained', 'AutoTokenizer.from_pretrained', (['self.model_path'], {'trust_remote_code': '(True)'}), '(self.model_path, trust_remote_code=True)\n', (3081, 3122), False, 'from transformers import AutoModel, AutoTokenizer\n'), ((3401, 3471), 'transformers.AutoTokenizer.fr...
# Copyright Vespa.ai. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root. from datasets import load_dataset import json import unicodedata def remove_control_characters(s): return "".join(ch for ch in s if unicodedata.category(ch)[0]!="C") from langchain.text_splitter import Recur...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((362, 477), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1536)', 'chunk_overlap': '(0)', 'length_function': 'len', 'is_separator_regex': '(False)'}), '(chunk_size=1536, chunk_overlap=0,\n length_function=len, is_separator_regex=False)\n', (392, 4...
from time import sleep import copy import redis import json import pickle import traceback from flask import Response, request, stream_with_context from typing import Dict, Union import os from langchain.schema import HumanMessage, SystemMessage from backend.api.language_model import get_llm from backend.main import ...
[ "langchain.schema.HumanMessage", "langchain.schema.SystemMessage" ]
[((11305, 11357), 'backend.main.app.route', 'app.route', (['"""/api/chat_xlang_webot"""'], {'methods': "['POST']"}), "('/api/chat_xlang_webot', methods=['POST'])\n", (11314, 11357), False, 'from backend.main import app, message_id_register, message_pool, logger\n'), ((2664, 2689), 'real_agents.web_agent.WebBrowsingExec...
import os import json from langchain.schema import messages_from_dict, messages_to_dict from langchain.memory import ( ConversationBufferMemory, ChatMessageHistory, ) class YeagerAIContext: """Context for the @yeager.ai agent.""" def __init__(self, username: str, session_id: str, session_path: str):...
[ "langchain.memory.ConversationBufferMemory", "langchain.memory.ChatMessageHistory", "langchain.schema.messages_to_dict", "langchain.schema.messages_from_dict" ]
[((472, 492), 'langchain.memory.ChatMessageHistory', 'ChatMessageHistory', ([], {}), '()\n', (490, 492), False, 'from langchain.memory import ConversationBufferMemory, ChatMessageHistory\n'), ((527, 597), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', ...
import argparse import os import subprocess import time import gradio as gr from huggingface_hub import snapshot_download from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain_community.document_loaders import ( Docx2txtLoader, PyPDFLoader, TextLoader, YoutubeLoader, ) from ...
[ "langchain_community.document_loaders.Docx2txtLoader", "langchain_community.document_loaders.PyPDFLoader", "langchain_community.document_loaders.TextLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain_community.vectorstores.Chroma.from_documents", "langchain_community.document_lo...
[((934, 946), 'chat_with_mlx.models.utils.model_info', 'model_info', ([], {}), '()\n', (944, 946), False, 'from chat_with_mlx.models.utils import model_info\n'), ((991, 1040), 'openai.OpenAI', 'OpenAI', ([], {'api_key': '"""EMPTY"""', 'base_url': 'openai_api_base'}), "(api_key='EMPTY', base_url=openai_api_base)\n", (99...
import time import numpy as np import torch from torch.nn import functional as F ########## # Functions for IMDB demo notebook. # Data source: Stanford AI Lab https://ai.stanford.edu/~amaas/data/sentiment/ ########## # Output words instead of scores. def sentiment_score_to_name(score: float): if score > 0: ...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((1670, 1681), 'time.time', 'time.time', ([], {}), '()\n', (1679, 1681), False, 'import time\n'), ((2431, 2473), 'torch.nn.functional.normalize', 'F.normalize', (['review_embeddings'], {'p': '(2)', 'dim': '(1)'}), '(review_embeddings, p=2, dim=1)\n', (2442, 2473), True, 'from torch.nn import functional as F\n'), ((253...
from langchain.chains import LLMChain from langchain_core.prompts import PromptTemplate from langchain_openai import ChatOpenAI from output_parsers import summary_parser, ice_breaker_parser, topics_of_interest_parser llm = ChatOpenAI(temperature=0, model_name="gpt-3.5-turbo") llm_creative = ChatOpenAI(temperature=1, ...
[ "langchain.chains.LLMChain", "langchain_openai.ChatOpenAI" ]
[((225, 278), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, model_name='gpt-3.5-turbo')\n", (235, 278), False, 'from langchain_openai import ChatOpenAI\n'), ((294, 347), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '...
# coding: UTF-8 import gc import glob import torch import time import os import json from collections import defaultdict from langchain.embeddings.huggingface import HuggingFaceEmbeddings from langchain.schema import Document from langchain.vectorstores import FAISS from tqdm import tqdm import config import re base_...
[ "langchain.vectorstores.FAISS.from_documents", "langchain.schema.Document" ]
[((2132, 2158), 're.findall', 're.findall', (['pattern_1', 'key'], {}), '(pattern_1, key)\n', (2142, 2158), False, 'import re\n'), ((1175, 1221), 'langchain.schema.Document', 'Document', ([], {'page_content': 'strs', 'metadata': 'metadata'}), '(page_content=strs, metadata=metadata)\n', (1183, 1221), False, 'from langch...
# -*- coding: UTF-8 -*- """ @Project : AI-Vtuber @File : claude_model.py @Author : HildaM @Email : Hilda_quan@163.com @Date : 2023/06/17 下午 4:44 @Description : 本地向量数据库模型设置 """ from langchain.embeddings import HuggingFaceEmbeddings import os # 项目根路径 TEC2VEC_MODELS_PATH = os.getcwd() + "\\" + "data" + "\\" + ...
[ "langchain.embeddings.HuggingFaceEmbeddings" ]
[((468, 542), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': '(TEC2VEC_MODELS_PATH + DEFAULT_MODEL_NAME)'}), '(model_name=TEC2VEC_MODELS_PATH + DEFAULT_MODEL_NAME)\n', (489, 542), False, 'from langchain.embeddings import HuggingFaceEmbeddings\n'), ((908, 934), 'os.path.exists...
import datetime import json import pkgutil import time import uuid import os import copy from dataclasses import asdict import datasets as ds from cot.config import Config from cot.utils.schemas.cot import features as cot_features # disable transformation (e.g. map) caching # https://huggingface.co/docs/datasets/v2....
[ "langchain.chat_models.ChatOpenAI", "langchain.utils.get_from_dict_or_env", "langchain.Cohere", "langchain.Prompt", "langchain.llms.utils.enforce_stop_tokens", "langchain.OpenAI", "langchain.HuggingFaceHub" ]
[((383, 403), 'datasets.disable_caching', 'ds.disable_caching', ([], {}), '()\n', (401, 403), True, 'import datasets as ds\n'), ((428, 472), 'pkgutil.get_data', 'pkgutil.get_data', (['__name__', '"""fragments.json"""'], {}), "(__name__, 'fragments.json')\n", (444, 472), False, 'import pkgutil\n'), ((873, 893), 'dataset...
import os import threading import time from contextlib import ExitStack from pathlib import Path from typing import cast, Optional import yaml from dotenv import load_dotenv from firebase_admin import auth from langchain.text_splitter import CharacterTextSplitter from llama_index import SimpleDirectoryReader from read...
[ "langchain.text_splitter.CharacterTextSplitter" ]
[((664, 677), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (675, 677), False, 'from dotenv import load_dotenv\n'), ((687, 707), 'realtime_ai_character.logger.get_logger', 'get_logger', (['__name__'], {}), '(__name__)\n', (697, 707), False, 'from realtime_ai_character.logger import get_logger\n'), ((917, 944),...
import sys from dotenv import load_dotenv from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.llms import OpenAI from commands import chrome_click_on_link, chrome_get_the_links_on_the_page, chrome_open_url, chrome_read_the_page, computer_applescript_action, say_text ...
[ "langchain.llms.OpenAI", "langchain.agents.initialize_agent" ]
[((350, 363), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (361, 363), False, 'from dotenv import load_dotenv\n'), ((394, 415), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (400, 415), False, 'from langchain.llms import OpenAI\n'), ((613, 692), 'langchain.agents.initia...
import os from typing import Optional from langchain import LLMChain, OpenAI, PromptTemplate from langchain.chains.base import Chain from langchain.chains.conversation.memory import ConversationBufferMemory from langchain.llms.base import BaseLLM from langchain.llms.loading import load_llm DEFAULT_LLM = None # Defau...
[ "langchain.PromptTemplate", "langchain.chains.conversation.memory.ConversationBufferMemory", "langchain.llms.loading.load_llm", "langchain.LLMChain", "langchain.OpenAI" ]
[((825, 914), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'TEMPLATE', 'input_variables': "['query', 'df_head', 'df_columns']"}), "(template=TEMPLATE, input_variables=['query', 'df_head',\n 'df_columns'])\n", (839, 914), False, 'from langchain import LLMChain, OpenAI, PromptTemplate\n'), ((1501, 1...
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Project : AI. @by PyCharm # @File : promptwatch # @Time : 2023/7/13 10:03 # @Author : betterme # @WeChat : meutils # @Software : PyCharm # @Description : import os from meutils.pipe import * from langchain import OpenAI, LLMChain,...
[ "langchain.PromptTemplate.from_template", "langchain.OpenAI" ]
[((417, 467), 'langchain.PromptTemplate.from_template', 'PromptTemplate.from_template', (['"""这是个prompt: {input}"""'], {}), "('这是个prompt: {input}')\n", (445, 467), False, 'from langchain import OpenAI, LLMChain, PromptTemplate\n'), ((486, 552), 'promptwatch.register_prompt_template', 'register_prompt_template', (['"""n...
import sys from typing import Any import readline from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationBufferMemory import colorama from callbacks import handlers from config import config from i18n import text from utils import utils from agent.agent import create_agent from walrus....
[ "langchain.memory.ConversationBufferMemory" ]
[((442, 455), 'config.config.init', 'config.init', ([], {}), '()\n', (453, 455), False, 'from config import config\n'), ((460, 475), 'colorama.init', 'colorama.init', ([], {}), '()\n', (473, 475), False, 'import colorama\n'), ((623, 653), 'i18n.text.init_system_messages', 'text.init_system_messages', (['llm'], {}), '(l...
"""Wrapper around Cohere APIs.""" from __future__ import annotations import logging from typing import Any, Callable, Dict, List, Optional from pydantic import Extra, root_validator from tenacity import ( before_sleep_log, retry, retry_if_exception_type, stop_after_attempt, wait_exponential, ) fr...
[ "langchain.llms.utils.enforce_stop_tokens", "langchain.utils.get_from_dict_or_env" ]
[((531, 558), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (548, 558), False, 'import logging\n'), ((3018, 3034), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (3032, 3034), False, 'from pydantic import Extra, root_validator\n'), ((3195, 3259), 'langchain.utils.get_from...
"""Wrapper around GooseAI API.""" import logging from typing import Any, Dict, List, Mapping, Optional from pydantic import Extra, Field, root_validator from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.utils import get_from_dict_or_env logger = loggi...
[ "langchain.utils.get_from_dict_or_env" ]
[((315, 342), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (332, 342), False, 'import logging\n'), ((1675, 1702), 'pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1680, 1702), False, 'from pydantic import Extra, Field, root_validator\n'), ((1836...
"""Wrapper around Anyscale""" from typing import Any, Dict, List, Mapping, Optional import requests from pydantic import Extra, root_validator from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.llms.utils import enforce_stop_tokens from langchain.utils ...
[ "langchain.llms.utils.enforce_stop_tokens", "langchain.utils.get_from_dict_or_env" ]
[((1679, 1695), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1693, 1695), False, 'from pydantic import Extra, root_validator\n'), ((1862, 1938), 'langchain.utils.get_from_dict_or_env', 'get_from_dict_or_env', (['values', '"""anyscale_service_url"""', '"""ANYSCALE_SERVICE_URL"""'], {}), "(values, 'any...
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Project : AI. @by PyCharm # @File : chatpicture # @Time : 2023/8/23 13:56 # @Author : betterme # @WeChat : meutils # @Software : PyCharm # @Description : 增加代理 根据意图选择 OCR类型 from meutils.pipe import * from meutils.ai_cv.ocr_api impor...
[ "langchain.chat_models.ChatOpenAI" ]
[((732, 744), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {}), '()\n', (742, 744), False, 'from langchain.chat_models import ChatOpenAI\n'), ((526, 549), 'meutils.ai_cv.ocr_api.OCR.basic_accurate', 'OCR.basic_accurate', (['img'], {}), '(img)\n', (544, 549), False, 'from meutils.ai_cv.ocr_api import OCR\n'), ...
from typing import List from pydantic import BaseModel, Field from langchain.agents import AgentExecutor, Tool from langchain.llms.base import BaseLLM from .agent.base import AutonomousAgent class ExecutionAgent(BaseModel): agent: AgentExecutor = Field(...) @classmethod def from_llm(cls, llm: BaseLLM, o...
[ "langchain.agents.AgentExecutor.from_agent_and_tools" ]
[((254, 264), 'pydantic.Field', 'Field', (['...'], {}), '(...)\n', (259, 264), False, 'from pydantic import BaseModel, Field\n'), ((533, 610), 'langchain.agents.AgentExecutor.from_agent_and_tools', 'AgentExecutor.from_agent_and_tools', ([], {'agent': 'agent', 'tools': 'tools', 'verbose': 'verbose'}), '(agent=agent, too...
# process_text.py from lib.chat.setup import openai_embeddings from langchain.docstore.document import Document from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.vectorstores import Chroma from langchain.document_loaders.csv_loader import CSVLoader import requests import json import cha...
[ "langchain.vectorstores.Chroma.from_documents", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.csv_loader.CSVLoader", "langchain.docstore.document.Document", "langchain.vectorstores.Chroma" ]
[((612, 703), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(4096)', 'chunk_overlap': '(256)', 'length_function': 'len'}), '(chunk_size=4096, chunk_overlap=256,\n length_function=len)\n', (642, 703), False, 'from langchain.text_splitter import Recurs...
import re from typing import List from langchain.text_splitter import RecursiveCharacterTextSplitter, TextSplitter from langchain.docstore.document import Document as LCDocument class MarkDownSplitter(TextSplitter): '''To split markdown''' def split_text(self, text: str) -> List[str]: if self.count_t...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((2824, 2848), 're.sub', 're.sub', (['"""<(.*?)>"""', '""""""', 'l'], {}), "('<(.*?)>', '', l)\n", (2830, 2848), False, 'import re\n'), ((514, 628), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': 'self._chunk_size', 'chunk_overlap': '(0)', 'length_functi...
import os import json from dotenv import load_dotenv from langchain.agents import Tool from langchain.chat_models import ChatOpenAI from ai.ai_functions import get_company_info, get_intro_response from consts import company_handbook_faiss_path, llm_model_type, demo_company_name from utils import calculate_vesting # L...
[ "langchain.chat_models.ChatOpenAI" ]
[((339, 352), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (350, 352), False, 'from dotenv import load_dotenv\n'), ((393, 420), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (402, 420), False, 'import os\n'), ((427, 494), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([...
from langchain.prompts import PromptTemplate _symptom_extract_template = """Consider the following conversation patient note: Patient note: {note} Choose on of the symptoms to be the chief complaint (it is usually the first symptom mentioned). Provide your response strictly in the following format, replacing only th...
[ "langchain.prompts.PromptTemplate.from_template" ]
[((830, 885), 'langchain.prompts.PromptTemplate.from_template', 'PromptTemplate.from_template', (['_symptom_extract_template'], {}), '(_symptom_extract_template)\n', (858, 885), False, 'from langchain.prompts import PromptTemplate\n'), ((904, 957), 'langchain.prompts.PromptTemplate.from_template', 'PromptTemplate.from_...
import re import string import traceback from collections import Counter import numpy as np import pandas as pd import tqdm from langchain.evaluation.qa import QAEvalChain from langchain.llms import OpenAI from algos.PWS import PWS_Base, PWS_Extra from algos.notool import CoT, IO from algos.react import ReactBase fr...
[ "langchain.llms.OpenAI" ]
[((432, 469), 're.sub', 're.sub', (['"""\\\\b(a|an|the)\\\\b"""', '""" """', 'text'], {}), "('\\\\b(a|an|the)\\\\b', ' ', text)\n", (438, 469), False, 'import re\n'), ((1337, 1363), 'collections.Counter', 'Counter', (['prediction_tokens'], {}), '(prediction_tokens)\n', (1344, 1363), False, 'from collections import Coun...
import json from langchain.schema.messages import SystemMessage from langchain.output_parsers.json import parse_partial_json from creator.code_interpreter import CodeInterpreter, language_map from creator.config.library import config from creator.utils import load_system_prompt, remove_tips from creator.llm.llm_creat...
[ "langchain.output_parsers.json.parse_partial_json", "langchain.schema.messages.SystemMessage" ]
[((389, 444), 'creator.utils.load_system_prompt', 'load_system_prompt', (['config.tips_for_testing_prompt_path'], {}), '(config.tips_for_testing_prompt_path)\n', (407, 444), False, 'from creator.utils import load_system_prompt, remove_tips\n'), ((459, 513), 'creator.utils.load_system_prompt', 'load_system_prompt', (['c...
import requests from typing import Any, Dict, Optional from langchain.chains.api.prompt import API_RESPONSE_PROMPT, API_URL_PROMPT from langchain.chains import APIChain from langchain.prompts import BasePromptTemplate from langchain.base_language import BaseLanguageModel from langchain.chains.llm import LLMChain from...
[ "langchain.chains.llm.LLMChain" ]
[((1139, 1179), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'api_url_prompt'}), '(llm=llm, prompt=api_url_prompt)\n', (1147, 1179), False, 'from langchain.chains.llm import LLMChain\n'), ((1207, 1252), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'api_respons...
from langchain.agents.tools import Tool from langchain.chains import LLMMathChain from langchain.chat_models import ChatOpenAI from langchain.llms import OpenAI from langchain_experimental.plan_and_execute import ( PlanAndExecute, load_agent_executor, load_chat_planner, ) llm = OpenAI(temperature=0) llm_ma...
[ "langchain_experimental.plan_and_execute.load_chat_planner", "langchain.chat_models.ChatOpenAI", "langchain.llms.OpenAI", "langchain.chains.LLMMathChain.from_llm", "langchain.agents.tools.Tool", "langchain_experimental.plan_and_execute.load_agent_executor", "langchain_experimental.plan_and_execute.PlanA...
[((292, 313), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (298, 313), False, 'from langchain.llms import OpenAI\n'), ((331, 375), 'langchain.chains.LLMMathChain.from_llm', 'LLMMathChain.from_llm', ([], {'llm': 'llm', 'verbose': '(True)'}), '(llm=llm, verbose=True)\n', (352, 375...
"""Functionality for loading chains.""" import json from pathlib import Path from typing import Any, Union import yaml from langchain.chains.api.base import APIChain from langchain.chains.base import Chain from langchain.chains.combine_documents.map_reduce import MapReduceDocumentsChain from langchain.chains.combine_...
[ "langchain.chains.llm.LLMChain", "langchain.chains.qa_with_sources.base.QAWithSourcesChain", "langchain.chains.api.base.APIChain", "langchain.chains.qa_with_sources.vector_db.VectorDBQAWithSourcesChain", "langchain.chains.hyde.base.HypotheticalDocumentEmbedder", "langchain.chains.pal.base.PALChain", "la...
[((2165, 2207), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt, **config)\n', (2173, 2207), False, 'from langchain.chains.llm import LLMChain\n'), ((2853, 2945), 'langchain.chains.hyde.base.HypotheticalDocumentEmbedder', 'HypotheticalDocumentEmbedder', ([...
"""Wrapper around HuggingFace APIs.""" from typing import Any, Dict, List, Mapping, Optional import requests from pydantic import Extra, root_validator from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.llms.utils import enforce_stop_tokens from langcha...
[ "langchain.llms.utils.enforce_stop_tokens", "langchain.utils.get_from_dict_or_env" ]
[((1661, 1677), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1675, 1677), False, 'from pydantic import Extra, root_validator\n'), ((1848, 1936), 'langchain.utils.get_from_dict_or_env', 'get_from_dict_or_env', (['values', '"""huggingfacehub_api_token"""', '"""HUGGINGFACEHUB_API_TOKEN"""'], {}), "(valu...
"""Clear Weaviate index.""" import logging import os import weaviate from langchain.embeddings import OpenAIEmbeddings from langchain.indexes import SQLRecordManager, index from langchain.vectorstores import Weaviate logger = logging.getLogger(__name__) WEAVIATE_URL = os.environ["WEAVIATE_URL"] WEAVIATE_API_KEY = os...
[ "langchain.embeddings.OpenAIEmbeddings", "langchain.indexes.SQLRecordManager", "langchain.indexes.index" ]
[((228, 255), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (245, 255), False, 'import logging\n'), ((893, 984), 'langchain.indexes.SQLRecordManager', 'SQLRecordManager', (['f"""weaviate/{WEAVIATE_DOCS_INDEX_NAME}"""'], {'db_url': 'RECORD_MANAGER_DB_URL'}), "(f'weaviate/{WEAVIATE_DOCS_IN...
import logging logging.basicConfig(level=logging.CRITICAL) import os from pathlib import Path import openai from dotenv import load_dotenv from langchain.chat_models import ChatOpenAI from llama_index import ( GPTVectorStoreIndex, LLMPredictor, ServiceContext, StorageContext, download_loader, ...
[ "langchain.chat_models.ChatOpenAI" ]
[((16, 59), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.CRITICAL'}), '(level=logging.CRITICAL)\n', (35, 59), False, 'import logging\n'), ((444, 457), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (455, 457), False, 'from dotenv import load_dotenv\n'), ((644, 729), 'llama_index.Service...
"""This is the logic for ingesting PDF and DOCX files into LangChain.""" import os from pathlib import Path from langchain.text_splitter import RecursiveCharacterTextSplitter from pdfminer.high_level import extract_text from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores import FAISS from dote...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.embeddings.OpenAIEmbeddings" ]
[((374, 387), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (385, 387), False, 'from dotenv import load_dotenv\n'), ((408, 437), 'os.getenv', 'os.getenv', (['"""OPENAI_API_TOKEN"""'], {}), "('OPENAI_API_TOKEN')\n", (417, 437), False, 'import os\n'), ((1354, 1445), 'langchain.text_splitter.RecursiveCharacterTex...
import os from langchain.llms.bedrock import Bedrock from langchain import PromptTemplate def get_llm(): model_kwargs = { "maxTokenCount": 1024, "stopSequences": [], "temperature": 0, "topP": 0.9 } llm = Bedrock( # credentials_profile_name=os.environ...
[ "langchain.PromptTemplate.from_template" ]
[((844, 882), 'langchain.PromptTemplate.from_template', 'PromptTemplate.from_template', (['template'], {}), '(template)\n', (872, 882), False, 'from langchain import PromptTemplate\n'), ((437, 470), 'os.environ.get', 'os.environ.get', (['"""BWB_REGION_NAME"""'], {}), "('BWB_REGION_NAME')\n", (451, 470), False, 'import ...
from langchain import PromptTemplate, LLMChain from langchain.document_loaders import TextLoader from langchain.embeddings import LlamaCppEmbeddings from langchain.llms import GPT4All from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.callbacks.base import CallbackManager from langchain.c...
[ "langchain.llms.GPT4All", "langchain.PromptTemplate", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.vectorstores.faiss.FAISS.load_local", "langchain.embeddings.LlamaCppEmbeddings", "langchain.document_loaders.TextLoader", "langchain.callbacks.streaming_stdout.StreamingStdOutCallba...
[((968, 1007), 'langchain.document_loaders.TextLoader', 'TextLoader', (['"""./docs/shortened_sotu.txt"""'], {}), "('./docs/shortened_sotu.txt')\n", (978, 1007), False, 'from langchain.document_loaders import TextLoader\n'), ((1021, 1062), 'langchain.embeddings.LlamaCppEmbeddings', 'LlamaCppEmbeddings', ([], {'model_pat...
import os from langchain.embeddings import OpenAIEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.vectorstores import Chroma class Database: def __init__(self, directory): self.embeddings = OpenAIEmbeddings() self.text_splitter = RecursiveCharacterTex...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.embeddings.OpenAIEmbeddings", "langchain.vectorstores.Chroma.from_texts" ]
[((251, 269), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (267, 269), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((299, 363), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(500)', 'chunk_overlap': '...
import os import re import time from typing import Any from dotenv import load_dotenv from langchain.callbacks.base import BaseCallbackHandler from langchain.chat_models import ChatOpenAI from langchain.schema import LLMResult from slack_bolt import App from slack_bolt.adapter.socket_mode import SocketModeHandler CHA...
[ "langchain.chat_models.ChatOpenAI" ]
[((347, 360), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (358, 360), False, 'from dotenv import load_dotenv\n'), ((392, 518), 'slack_bolt.App', 'App', ([], {'signing_secret': "os.environ['SLACK_SIGNING_SECRET']", 'token': "os.environ['SLACK_BOT_TOKEN']", 'process_before_response': '(True)'}), "(signing_secr...
import json from typing import Optional, Any from langchain.schema import AIMessage from langchain.schema.runnable import RunnableSerializable, RunnableConfig from pydantic import BaseModel class FunctionCall(BaseModel): name: str arguments: dict[str, Any] class ParseFunctionCall(RunnableSerializable[AIMes...
[ "langchain.schema.AIMessage" ]
[((682, 700), 'langchain.schema.AIMessage', 'AIMessage', ([], {}), '(**input)\n', (691, 700), False, 'from langchain.schema import AIMessage\n'), ((1460, 1478), 'langchain.schema.AIMessage', 'AIMessage', ([], {}), '(**input)\n', (1469, 1478), False, 'from langchain.schema import AIMessage\n'), ((945, 970), 'json.loads'...
from langchain.chains.router import MultiPromptChain from langchain.chat_models import ChatOpenAI from dotenv import load_dotenv import os # A template for working with LangChain multi prompt chain. # It's a great way to let the large language model choose which prompts suits the question. # Load env files load_doten...
[ "langchain.chains.router.MultiPromptChain.from_prompts", "langchain.chat_models.ChatOpenAI" ]
[((310, 323), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (321, 323), False, 'from dotenv import load_dotenv\n'), ((341, 373), 'os.environ.get', 'os.environ.get', (['"""openai_api_key"""'], {}), "('openai_api_key')\n", (355, 373), False, 'import os\n'), ((1461, 1551), 'langchain.chat_models.ChatOpenAI', 'Cha...
from dotenv import load_dotenv from src.slackbot import SlackBot from src.handlers import create_handlers import asyncio from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler handler = StreamingStdOutCallbackHandler() # Load environment variables load_dotenv() # Load custom tools import src.c...
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler" ]
[((211, 243), 'langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler', 'StreamingStdOutCallbackHandler', ([], {}), '()\n', (241, 243), False, 'from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n'), ((273, 286), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (284, 286)...
#!/usr/bin/env python """A more complex example that shows how to configure index name at run time.""" from typing import Any, Iterable, List, Optional, Type from fastapi import FastAPI from langchain.embeddings import OpenAIEmbeddings from langchain.schema import Document from langchain.schema.embeddings import Embed...
[ "langchain.embeddings.OpenAIEmbeddings", "langchain.schema.runnable.ConfigurableFieldSingleOption" ]
[((893, 1027), 'fastapi.FastAPI', 'FastAPI', ([], {'title': '"""LangChain Server"""', 'version': '"""1.0"""', 'description': '"""Spin up a simple api server using Langchain\'s Runnable interfaces"""'}), '(title=\'LangChain Server\', version=\'1.0\', description=\n "Spin up a simple api server using Langchain\'s Runn...
from itertools import chain import pandas as pd from datasets import Dataset from joblib import Parallel, delayed from langchain.text_splitter import RecursiveCharacterTextSplitter from tqdm import tqdm def sl_hf_dataset_for_tokenizer( sl, sl_dataset_name, tokenizer, max_length, margin=192, min_length=7 ): "...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((1048, 1158), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(max_length - margin)', 'chunk_overlap': '(0)', 'length_function': 'token_len'}), '(chunk_size=max_length - margin,\n chunk_overlap=0, length_function=token_len)\n', (1078, 1158), False, '...
from contextlib import contextmanager import uuid import os import tiktoken from . import S2_tools as scholar import csv import sys import requests # pdf loader from langchain.document_loaders import OnlinePDFLoader ## paper questioning tools from llama_index import Document from llama_index.vector_stores import Pi...
[ "langchain.document_loaders.OnlinePDFLoader" ]
[((768, 796), 'os.mkdir', 'os.mkdir', (['workspace_dir_name'], {}), '(workspace_dir_name)\n', (776, 796), False, 'import os\n'), ((5950, 5986), 'tiktoken.encoding_for_model', 'tiktoken.encoding_for_model', (['"""gpt-4"""'], {}), "('gpt-4')\n", (5977, 5986), False, 'import tiktoken\n'), ((7532, 7548), 'os.listdir', 'os....
import streamlit as st from langchain.prompts import PromptTemplate chat_template = PromptTemplate( input_variables=['transcript','summary','chat_history','user_message', 'sentiment_report'], template=''' You are an AI chatbot intended to discuss about the user's audio transcription. \nT...
[ "langchain.prompts.PromptTemplate" ]
[((88, 562), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['transcript', 'summary', 'chat_history', 'user_message', 'sentiment_report']", 'template': '"""\n You are an AI chatbot intended to discuss about the user\'s audio transcription.\n \nTRANSCRIPT: "{transcript}"\n ...
from enum import Enum from typing import Callable, Tuple from langchain.agents.agent import AgentExecutor from langchain.agents.tools import BaseTool, Tool class ToolScope(Enum): GLOBAL = "global" SESSION = "session" SessionGetter = Callable[[], Tuple[str, AgentExecutor]] def tool( name: str, des...
[ "langchain.agents.tools.Tool" ]
[((1245, 1306), 'langchain.agents.tools.Tool', 'Tool', ([], {'name': 'self.name', 'description': 'self.description', 'func': 'func'}), '(name=self.name, description=self.description, func=func)\n', (1249, 1306), False, 'from langchain.agents.tools import BaseTool, Tool\n')]
import asyncio from langchain.document_loaders import PyPDFLoader, DirectoryLoader from langchain import PromptTemplate from langchain.embeddings import HuggingFaceEmbeddings from langchain.vectorstores import FAISS from langchain.llms import CTransformers from langchain.chains import RetrievalQA import chainlit...
[ "langchain.vectorstores.FAISS.load_local", "langchain.embeddings.HuggingFaceEmbeddings", "langchain.llms.CTransformers", "langchain.PromptTemplate" ]
[((808, 900), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'custom_prompt_template', 'input_variables': "['context', 'question']"}), "(template=custom_prompt_template, input_variables=['context',\n 'question'])\n", (822, 900), False, 'from langchain import PromptTemplate\n'), ((1522, 1635), 'langc...
from langchain.chains import LLMChain from langchain.prompts import PromptTemplate from langchain.chat_models import ChatOpenAI from dotenv import load_dotenv import os from langchain.chains import SimpleSequentialChain # Create a .env file in the root of your project and add your OpenAI API key to it # Load env files...
[ "langchain.prompts.PromptTemplate", "langchain.chains.SimpleSequentialChain", "langchain.chat_models.ChatOpenAI", "langchain.chains.LLMChain" ]
[((321, 334), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (332, 334), False, 'from dotenv import load_dotenv\n'), ((352, 384), 'os.environ.get', 'os.environ.get', (['"""openai_api_key"""'], {}), "('openai_api_key')\n", (366, 384), False, 'import os\n'), ((469, 524), 'langchain.chat_models.ChatOpenAI', 'ChatO...
import os import re from dotenv import load_dotenv from langchain.chat_models import ChatOpenAI from slack_bolt import App from slack_bolt.adapter.socket_mode import SocketModeHandler load_dotenv() # ボットトークンを使ってアプリを初期化します app = App(token=os.environ.get("SLACK_BOT_TOKEN")) @app.event("app_mention") def handle_menti...
[ "langchain.chat_models.ChatOpenAI" ]
[((186, 199), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (197, 199), False, 'from dotenv import load_dotenv\n'), ((378, 412), 're.sub', 're.sub', (['"""<@.*>"""', '""""""', "event['text']"], {}), "('<@.*>', '', event['text'])\n", (384, 412), False, 'import re\n'), ((424, 532), 'langchain.chat_models.ChatOpe...
""" This module contains the function to classify the user query. """ import json from langchain.prompts import ChatPromptTemplate from langchain.chains import create_extraction_chain from langchain.chat_models import ChatOpenAI from langchain.document_loaders import TextLoader from langchain.document_loaders import ...
[ "langchain.prompts.ChatPromptTemplate.from_template", "langchain.chat_models.ChatOpenAI" ]
[((582, 627), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model': 'config.model'}), '(temperature=0, model=config.model)\n', (592, 627), False, 'from langchain.chat_models import ChatOpenAI\n'), ((650, 853), 'langchain.prompts.ChatPromptTemplate.from_template', 'ChatPromptTemplate.fro...
from langchain.chat_models import ChatOpenAI from langchain.prompts import ChatPromptTemplate from langchain.schema import StrOutputParser from langchain.schema.runnable import Runnable from langchain.schema.runnable.config import RunnableConfig import chainlit as cl @cl.on_chat_start async def on_chat_start(): ...
[ "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.schema.StrOutputParser", "langchain.chat_models.ChatOpenAI" ]
[((328, 398), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_base': '"""http://localhost:8888/v1"""', 'streaming': '(True)'}), "(openai_api_base='http://localhost:8888/v1', streaming=True)\n", (338, 398), False, 'from langchain.chat_models import ChatOpenAI\n'), ((411, 600), 'langchain.prompts.Chat...
import os import streamlit as st from PyPDF2 import PdfReader, PdfWriter from langchain.text_splitter import CharacterTextSplitter from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import FAISS from langchain.chains.question_answering import load_qa_chain from langchain.llms i...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.vectorstores.FAISS.from_texts", "langchain.llms.OpenAI", "langchain.callbacks.get_openai_callback", "langchain.chains.question_answering.load_qa_chain", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((481, 579), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {'separator': '"""\n"""', 'chunk_size': '(1000)', 'chunk_overlap': '(200)', 'length_function': 'len'}), "(separator='\\n', chunk_size=1000, chunk_overlap=200,\n length_function=len)\n", (502, 579), False, 'from langchain.tex...
"""Example of observing LLM calls made by via callable OpenAI LLM.""" from langchain.llms import OpenAI from langchain_prefect.plugins import RecordLLMCalls llm = OpenAI(temperature=0.9) with RecordLLMCalls(): llm("What would be a good name for a company that makes colorful socks?")
[ "langchain.llms.OpenAI", "langchain_prefect.plugins.RecordLLMCalls" ]
[((166, 189), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0.9)'}), '(temperature=0.9)\n', (172, 189), False, 'from langchain.llms import OpenAI\n'), ((196, 212), 'langchain_prefect.plugins.RecordLLMCalls', 'RecordLLMCalls', ([], {}), '()\n', (210, 212), False, 'from langchain_prefect.plugins import Record...
from langchain.agents import load_tools from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType from langchain.utilities import SerpAPIWrapper from langchain_app.models.vicuna_request_llm import VicunaLLM # First, let's load the language model we're going to use to control the agent...
[ "langchain.agents.initialize_agent", "langchain.utilities.SerpAPIWrapper", "langchain.agents.load_tools", "langchain_app.models.vicuna_request_llm.VicunaLLM", "langchain.agents.Tool" ]
[((328, 339), 'langchain_app.models.vicuna_request_llm.VicunaLLM', 'VicunaLLM', ([], {}), '()\n', (337, 339), False, 'from langchain_app.models.vicuna_request_llm import VicunaLLM\n'), ((419, 448), 'langchain.utilities.SerpAPIWrapper', 'SerpAPIWrapper', ([], {'params': 'params'}), '(params=params)\n', (433, 448), False...
# /app/src/tools/setup.py import logging from langchain.pydantic_v1 import BaseModel, Field from langchain.tools import BaseTool from langchain_community.tools import DuckDuckGoSearchResults from src.tools.doc_search import DocumentSearch logger = logging.getLogger(__name__) class SearchWebInput(BaseModel): qu...
[ "langchain.pydantic_v1.Field", "langchain_community.tools.DuckDuckGoSearchResults" ]
[((251, 278), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (268, 278), False, 'import logging\n'), ((331, 368), 'langchain.pydantic_v1.Field', 'Field', ([], {'description': '"""The search query"""'}), "(description='The search query')\n", (336, 368), False, 'from langchain.pydantic_v1 i...
from llama_index.callbacks import CallbackManager, LlamaDebugHandler, CBEventType from llama_index import ListIndex, ServiceContext, SimpleDirectoryReader, VectorStoreIndex ''' Title of the page: A simple Python implementation of the ReAct pattern for LLMs Name of the website: LlamaIndex (GPT Index) is a data framewor...
[ "langchain.chat_models.ChatOpenAI" ]
[((676, 718), 'llama_index.callbacks.LlamaDebugHandler', 'LlamaDebugHandler', ([], {'print_trace_on_end': '(True)'}), '(print_trace_on_end=True)\n', (693, 718), False, 'from llama_index.callbacks import CallbackManager, LlamaDebugHandler, CBEventType\n'), ((738, 768), 'llama_index.callbacks.CallbackManager', 'CallbackM...
from typing import Any, Callable from pandas import DataFrame from exact_rag.config import EmbeddingType, Embeddings, DatabaseType, Databases from langchain_openai import OpenAIEmbeddings from langchain_community.embeddings import OllamaEmbeddings from langchain.vectorstores.chroma import Chroma from langchain.vector...
[ "langchain_community.document_loaders.DataFrameLoader", "langchain.text_splitter.RecursiveCharacterTextSplitter.from_tiktoken_encoder", "langchain.indexes.SQLRecordManager", "langchain.indexes.index" ]
[((2955, 3052), 'langchain.indexes.SQLRecordManager', 'SQLRecordManager', (['database_model.sql_namespace'], {'db_url': 'f"""sqlite:///{database_model.sql_url}"""'}), "(database_model.sql_namespace, db_url=\n f'sqlite:///{database_model.sql_url}')\n", (2971, 3052), False, 'from langchain.indexes import SQLRecordMana...
import base64 from enum import Enum import json import time import logging from pywebagent.env.browser import BrowserEnv from langchain.schema import HumanMessage, SystemMessage from langchain.chat_models import ChatOpenAI logger = logging.getLogger(__name__) TASK_STATUS = Enum("TASK_STATUS", "IN_PROGRESS SUCCESS FA...
[ "langchain.schema.SystemMessage", "langchain.schema.HumanMessage", "langchain.chat_models.ChatOpenAI" ]
[((233, 260), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (250, 260), False, 'import logging\n'), ((277, 326), 'enum.Enum', 'Enum', (['"""TASK_STATUS"""', '"""IN_PROGRESS SUCCESS FAILED"""'], {}), "('TASK_STATUS', 'IN_PROGRESS SUCCESS FAILED')\n", (281, 326), False, 'from enum import E...
"""Load markdown, html, text from files, clean up, split, ingest into Pinecone.""" import pinecone import tiktoken from langchain.document_loaders import ReadTheDocsLoader from langchain.embeddings import OpenAIEmbeddings from langchain.text_splitter import NLTKTextSplitter from langchain.vectorstores.pinecone import P...
[ "langchain.text_splitter.NLTKTextSplitter.from_tiktoken_encoder", "langchain.document_loaders.ReadTheDocsLoader", "langchain.embeddings.OpenAIEmbeddings", "langchain.vectorstores.pinecone.Pinecone.from_documents" ]
[((402, 445), 'langchain.document_loaders.ReadTheDocsLoader', 'ReadTheDocsLoader', (['"""hasura.io/docs/latest/"""'], {}), "('hasura.io/docs/latest/')\n", (419, 445), False, 'from langchain.document_loaders import ReadTheDocsLoader\n'), ((500, 573), 'langchain.text_splitter.NLTKTextSplitter.from_tiktoken_encoder', 'NLT...
from typing import List, Optional, Any, Dict from langchain.llms.base import LLM from langchain.utils import get_from_dict_or_env from pydantic import Extra, root_validator from sam.gpt.quora import PoeClient, PoeResponse # token = "KaEMfvDPEXoS115jzAFRRg%3D%3D" # prompt = "write a java function that prints the nt...
[ "langchain.utils.get_from_dict_or_env" ]
[((573, 589), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (587, 589), False, 'from pydantic import Extra, root_validator\n'), ((663, 714), 'langchain.utils.get_from_dict_or_env', 'get_from_dict_or_env', (['values', '"""token"""', '"""POE_COOKIE"""'], {}), "(values, 'token', 'POE_COOKIE')\n", (683, 71...
from __future__ import annotations from typing import List, Optional from pydantic import ValidationError from langchain.chains.llm import LLMChain from langchain.chat_models.base import BaseChatModel from langchain.experimental.autonomous_agents.autogpt.output_parser import ( AutoGPTOutputParser, BaseAutoGP...
[ "langchain.chains.llm.LLMChain", "langchain.tools.human.tool.HumanInputRun", "langchain.schema.AIMessage", "langchain.schema.HumanMessage", "langchain.experimental.autonomous_agents.autogpt.prompt.AutoGPTPrompt", "langchain.schema.Document", "langchain.experimental.autonomous_agents.autogpt.output_parse...
[((1753, 1918), 'langchain.experimental.autonomous_agents.autogpt.prompt.AutoGPTPrompt', 'AutoGPTPrompt', ([], {'ai_name': 'ai_name', 'ai_role': 'ai_role', 'tools': 'tools', 'input_variables': "['memory', 'messages', 'goals', 'user_input']", 'token_counter': 'llm.get_num_tokens'}), "(ai_name=ai_name, ai_role=ai_role, t...
"""Main entrypoint for the app.""" import asyncio import os from operator import itemgetter from typing import List, Optional, Sequence, Tuple, Union from uuid import UUID from fastapi import Depends, FastAPI, Request from fastapi.middleware.cors import CORSMiddleware from langchain.callbacks.manager import CallbackMa...
[ "langchain.schema.runnable.ConfigurableField", "langchain.schema.messages.HumanMessage", "langchain.prompts.PromptTemplate.from_template", "langchain.chat_models.ChatOpenAI", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.AsyncHtmlLoader", "langchain.schema.output_...
[((4923, 4932), 'fastapi.FastAPI', 'FastAPI', ([], {}), '()\n', (4930, 4932), False, 'from fastapi import Depends, FastAPI, Request\n'), ((12927, 12987), 'os.path.isfile', 'os.path.isfile', (["os.environ['GOOGLE_APPLICATION_CREDENTIALS']"], {}), "(os.environ['GOOGLE_APPLICATION_CREDENTIALS'])\n", (12941, 12987), False,...
from langchain.chat_models import ChatOpenAI from langchain_experimental.plan_and_execute import PlanAndExecute, load_agent_executor, load_chat_planner from langchain.llms import OpenAI from langchain import SerpAPIWrapper from langchain.agents.tools import Tool from langchain import LLMMathChain search = SerpAPIWrapp...
[ "langchain_experimental.plan_and_execute.load_chat_planner", "langchain.chat_models.ChatOpenAI", "langchain.LLMMathChain.from_llm", "langchain.llms.OpenAI", "langchain.agents.tools.Tool", "langchain_experimental.plan_and_execute.load_agent_executor", "langchain.SerpAPIWrapper", "langchain_experimental...
[((308, 324), 'langchain.SerpAPIWrapper', 'SerpAPIWrapper', ([], {}), '()\n', (322, 324), False, 'from langchain import SerpAPIWrapper\n'), ((331, 352), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)'}), '(temperature=0)\n', (337, 352), False, 'from langchain.llms import OpenAI\n'), ((370, 414), 'langchai...
"""Map-reduce chain. Splits up a document, sends the smaller parts to the LLM with one prompt, then combines the results with another one. """ from __future__ import annotations from typing import Any, Dict, List, Mapping, Optional from langchain.callbacks.manager import CallbackManagerForChainRun, Callbacks from la...
[ "langchain.chains.llm.LLMChain", "langchain.chains.combine_documents.map_reduce.MapReduceDocumentsChain", "langchain.callbacks.manager.CallbackManagerForChainRun.get_noop_manager", "langchain.chains.ReduceDocumentsChain", "langchain.docstore.document.Document", "langchain.chains.combine_documents.stuff.St...
[((1734, 1787), 'langchain.chains.llm.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt', 'callbacks': 'callbacks'}), '(llm=llm, prompt=prompt, callbacks=callbacks)\n', (1742, 1787), False, 'from langchain.chains.llm import LLMChain\n'), ((1810, 1930), 'langchain.chains.combine_documents.stuff.StuffDocuments...
import os import re import argparse import json import boto3 from bs4 import BeautifulSoup from langchain.document_loaders import PDFMinerPDFasHTMLLoader from langchain.docstore.document import Document from langchain.text_splitter import RecursiveCharacterTextSplitter,CharacterTextSplitter import statistics smr_clien...
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.PDFMinerPDFasHTMLLoader" ]
[((324, 357), 'boto3.client', 'boto3.client', (['"""sagemaker-runtime"""'], {}), "('sagemaker-runtime')\n", (336, 357), False, 'import boto3\n'), ((8397, 8430), 'langchain.document_loaders.PDFMinerPDFasHTMLLoader', 'PDFMinerPDFasHTMLLoader', (['pdf_path'], {}), '(pdf_path)\n', (8420, 8430), False, 'from langchain.docum...
from langchain.vectorstores import Chroma from langchain.embeddings import OpenAIEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.llms import OpenAI from langchain.chains import VectorDBQA from langchain.document_loaders import TextLoader from typing import List from langchai...
[ "langchain.vectorstores.Chroma.from_documents", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.llms.OpenAI", "langchain.embeddings.OpenAIEmbeddings", "langchain.document_loaders.TextLoader" ]
[((515, 541), 'langchain.document_loaders.TextLoader', 'TextLoader', (['self.file_path'], {}), '(self.file_path)\n', (525, 541), False, 'from langchain.document_loaders import TextLoader\n'), ((886, 950), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1...
import dataclasses import json import numpy as np import os import requests import sys from typing import List from langchain.text_splitter import CharacterTextSplitter from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores import Chroma from langchain.chat_models import ChatOpenAI from l...
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.Chroma.from_documents", "langchain.prompts.PromptTemplate", "langchain.schema.Document", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((1563, 1648), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_template', 'input_variables': "['context', 'question']"}), "(template=prompt_template, input_variables=['context',\n 'question'])\n", (1577, 1648), False, 'from langchain.prompts import PromptTemplate\n'), ((1786, 1811), ...
import logging from pathlib import Path from typing import List, Optional, Tuple from dotenv import load_dotenv load_dotenv() from queue import Empty, Queue from threading import Thread import gradio as gr from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from langchain.chat_models imp...
[ "langchain.schema.AIMessage", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.schema.HumanMessage", "langchain.schema.SystemMessage" ]
[((114, 127), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (125, 127), False, 'from dotenv import load_dotenv\n'), ((604, 698), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""[%(asctime)s %(levelname)s]: %(message)s"""', 'level': 'logging.INFO'}), "(format='[%(asctime)s %(levelname)s]: %(me...
import logging from langchain.chains import RetrievalQA from neogpt.prompts.prompt import get_prompt def local_retriever(db, llm, persona="default"): """ Fn: local_retriever Description: The function sets up the local retrieval-based question-answering system. Args: db (object): The database...
[ "langchain.chains.RetrievalQA.from_chain_type" ]
[((466, 493), 'neogpt.prompts.prompt.get_prompt', 'get_prompt', ([], {'persona': 'persona'}), '(persona=persona)\n', (476, 493), False, 'from neogpt.prompts.prompt import get_prompt\n'), ((590, 768), 'langchain.chains.RetrievalQA.from_chain_type', 'RetrievalQA.from_chain_type', ([], {'llm': 'llm', 'retriever': 'local_r...
from langchain import PromptTemplate PROMPT = """ 你需要扮演一个优秀的关键信息提取助手,从人类的对话中提取关键性内容(最多5个关键词),以协助其他助手更精准地回答问题。 注意:你不需要做任何解释说明,只需严格按照示例的格式输出关键词。 示例: 人类:我有一个服装厂,是否可以应用你们的装箱算法改善装载率呢? AI: 服装厂, 装箱算法, 装载率 现在开始: 人类:{query} AI: """ def information_extraction_raw_prompt(): return PromptTemplate(template=PROMPT, input_v...
[ "langchain.PromptTemplate" ]
[((281, 339), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'PROMPT', 'input_variables': "['query']"}), "(template=PROMPT, input_variables=['query'])\n", (295, 339), False, 'from langchain import PromptTemplate\n'), ((397, 455), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'PROMPT',...