o
    Jjg                     @  s   d Z ddlmZ ddlmZmZmZmZ ddlm	Z	 ddl
mZ ddlmZ ddlmZ ddlmZ dd	lmZ dd
lmZ ddlmZmZ ddlmZ e	ddddG dd deZdS )zCUse a single chain to route an input to one of multiple llm chains.    )annotations)AnyDictListOptional)
deprecated)BaseLanguageModel)PromptTemplate)ConversationChain)Chain)LLMChain)MultiRouteChain)LLMRouterChainRouterOutputParser)MULTI_PROMPT_ROUTER_TEMPLATEz0.2.12z1.0zUse RunnableLambda to select from multiple prompt templates. See example in API reference: https://api.python.langchain.com/en/latest/chains/langchain.chains.router.multi_prompt.MultiPromptChain.html)sinceremovalmessagec                   @  s0   e Zd ZdZedddZe	ddddZdS )MultiPromptChaina	  A multi-route chain that uses an LLM router chain to choose amongst prompts.

    This class is deprecated. See below for a replacement, which offers several
    benefits, including streaming and batch support.

    Below is an example implementation:

        .. code-block:: python

            from operator import itemgetter
            from typing import Literal
            from typing_extensions import TypedDict

            from langchain_core.output_parsers import StrOutputParser
            from langchain_core.prompts import ChatPromptTemplate
            from langchain_core.runnables import RunnableLambda, RunnablePassthrough
            from langchain_openai import ChatOpenAI

            llm = ChatOpenAI(model="gpt-4o-mini")

            prompt_1 = ChatPromptTemplate.from_messages(
                [
                    ("system", "You are an expert on animals."),
                    ("human", "{query}"),
                ]
            )
            prompt_2 = ChatPromptTemplate.from_messages(
                [
                    ("system", "You are an expert on vegetables."),
                    ("human", "{query}"),
                ]
            )

            chain_1 = prompt_1 | llm | StrOutputParser()
            chain_2 = prompt_2 | llm | StrOutputParser()

            route_system = "Route the user's query to either the animal or vegetable expert."
            route_prompt = ChatPromptTemplate.from_messages(
                [
                    ("system", route_system),
                    ("human", "{query}"),
                ]
            )


            class RouteQuery(TypedDict):
                """Route query to destination."""
                destination: Literal["animal", "vegetable"]


            route_chain = (
                route_prompt
                | llm.with_structured_output(RouteQuery)
                | itemgetter("destination")
            )

            chain = {
                "destination": route_chain,  # "animal" or "vegetable"
                "query": lambda x: x["query"],  # pass through input query
            } | RunnableLambda(
                # if animal, chain_1. otherwise, chain_2.
                lambda x: chain_1 if x["destination"] == "animal" else chain_2,
            )

            chain.invoke({"query": "what color are carrots"})
    return	List[str]c                 C  s   dgS )Ntext )selfr   r   \/var/www/html/zoom/venv/lib/python3.10/site-packages/langchain/chains/router/multi_prompt.pyoutput_keys`   s   zMultiPromptChain.output_keysNllmr   prompt_infosList[Dict[str, str]]default_chainOptional[Chain]kwargsr   c                 K  s   dd |D }d |}tj|d}t|dgt d}t||}	i }
|D ]}|d }|d }t|dgd	}t||d
}||
|< q%|pHt|dd}| d|	|
|d|S )zCConvenience constructor for instantiating from destination prompts.c                 S  s"   g | ]}|d   d|d  qS )namez: descriptionr   ).0pr   r   r   
<listcomp>m   s   " z1MultiPromptChain.from_prompts.<locals>.<listcomp>
)destinationsinput)templateinput_variablesoutput_parserr"   prompt_template)r*   r+   )r   promptr   )r   
output_key)router_chaindestination_chainsr   Nr   )	joinr   formatr	   r   r   from_llmr   r
   )clsr   r   r   r!   r(   destinations_strrouter_templaterouter_promptr0   r1   p_infor"   r-   r.   chain_default_chainr   r   r   from_promptsd   s4   	

zMultiPromptChain.from_prompts)r   r   )N)
r   r   r   r   r   r    r!   r   r   r   )__name__
__module____qualname____doc__propertyr   classmethodr<   r   r   r   r   r      s    
Cr   N)r@   
__future__r   typingr   r   r   r   langchain_core._apir   langchain_core.language_modelsr   langchain_core.promptsr	   langchain.chainsr
   langchain.chains.baser   langchain.chains.llmr   langchain.chains.router.baser   "langchain.chains.router.llm_routerr   r   +langchain.chains.router.multi_prompt_promptr   r   r   r   r   r   <module>   s$    	