Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

W-16295334: Updated naming fix for demo folder #29

Merged
merged 1 commit into from
Aug 14, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions demo/mulechain-ai-connector-demo/README.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
MuleChain AI Connector Demo
MuleSoft AI Chain Connector Demo
====================================
Anypoint Studio demo for MuleChain AI Connector.
Anypoint Studio demo for MuleSoft AI Chain Connector.


Prerequisites
Expand Down Expand Up @@ -43,4 +43,4 @@ You can use postman to trigger curls under the web server http://localhost:8081
* POST - /toolsMemory (Tools implementation with memory)
* POST - /chains (Tools implementation of Langchain with legacy chains)

_Please refer to [MuleChain AI Connector_Postman Collection](000_mulechain-ai-connector.postman_collection.json) available in source of this demo. You may have to provide own paths for various files._
_Please refer to [MuleSoft_AI_Chain_Connector_Postman_Collection](000_mulechain-ai-connector.postman_collection.json) available in source of this demo. You may have to provide own paths for various files._
Original file line number Diff line number Diff line change
@@ -1,20 +1,20 @@
<?xml version="1.0" encoding="UTF-8"?>

<mule xmlns:ee="http://www.mulesoft.org/schema/mule/ee/core" xmlns:mulechain="http://www.mulesoft.org/schema/mule/mulechain"
<mule xmlns:ee="http://www.mulesoft.org/schema/mule/ee/core" xmlns:ms-aichain="http://www.mulesoft.org/schema/mule/ms-aichain"
xmlns:http="http://www.mulesoft.org/schema/mule/http"
xmlns="http://www.mulesoft.org/schema/mule/core" xmlns:doc="http://www.mulesoft.org/schema/mule/documentation" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.mulesoft.org/schema/mule/core http://www.mulesoft.org/schema/mule/core/current/mule.xsd
http://www.mulesoft.org/schema/mule/http http://www.mulesoft.org/schema/mule/http/current/mule-http.xsd
http://www.mulesoft.org/schema/mule/mulechain http://www.mulesoft.org/schema/mule/mulechain/current/mule-mulechain.xsd
http://www.mulesoft.org/schema/mule/ms-aichain http://www.mulesoft.org/schema/mule/ms-aichain/current/mule-ms-aichain.xsd
http://www.mulesoft.org/schema/mule/ee/core http://www.mulesoft.org/schema/mule/ee/core/current/mule-ee.xsd">
<http:listener-config name="HTTP_Listener_config" doc:name="HTTP Listener config" doc:id="817b3a99-cf1c-41ea-a8bd-10932e93a6e3" >
<http:listener-connection host="0.0.0.0" port="8081" />
</http:listener-config>
<mulechain:llm-configuration-config name="MISTRAL_AI" llmType="MISTRAL_AI" configType="Configuration Json" doc:name="MuleChain AI Llm configuration" doc:id="bf1ef7ec-4aa1-41c8-a184-a13ca165c925" filePath='#[mule.home ++ "/apps/" ++ app.name ++ "/envVars.json"]' modelName="mistral-large-latest" temperature="0.1"/>
<mulechain:llm-configuration-config name="OPENAI" llmType="OPENAI" configType="Configuration Json" doc:name="MuleChain AI Llm configuration" doc:id="edb0d5a6-97c5-4d93-8098-4e197e563827" filePath='#[mule.home ++ "/apps/" ++ app.name ++ "/envVars.json"]' temperature="0.1" modelName="gpt-3.5-turbo"/>
<mulechain:llm-configuration-config name="OPENAI-GPT4-TURBO" llmType="OPENAI" configType="Configuration Json" doc:name="MuleChain AI Llm configuration" doc:id="74b85066-1569-4f10-a06b-e49e854eeef2" filePath='#[mule.home ++ "/apps/" ++ app.name ++ "/envVars.json"]' modelName="gpt-4o" />
<ms-aichain:config name="MISTRAL_AI" llmType="MISTRAL_AI" configType="Configuration Json" doc:name="MuleSoft AI Chain configuration" doc:id="bf1ef7ec-4aa1-41c8-a184-a13ca165c925" filePath='#[mule.home ++ "/apps/" ++ app.name ++ "/envVars.json"]' modelName="mistral-large-latest" temperature="0.1"/>
<ms-aichain:config name="OPENAI" llmType="OPENAI" configType="Configuration Json" doc:name="MuleSoft AI Chain configuration" doc:id="edb0d5a6-97c5-4d93-8098-4e197e563827" filePath='#[mule.home ++ "/apps/" ++ app.name ++ "/envVars.json"]' temperature="0.1" modelName="gpt-3.5-turbo"/>
<ms-aichain:config name="OPENAI-GPT4-TURBO" llmType="OPENAI" configType="Configuration Json" doc:name="MuleSoft AI Chain configuration" doc:id="74b85066-1569-4f10-a06b-e49e854eeef2" filePath='#[mule.home ++ "/apps/" ++ app.name ++ "/envVars.json"]' modelName="gpt-4o" />
<flow name="PromptTemplate" doc:id="cff3a8ed-3799-424a-becf-9d7387729bd0" >
<http:listener doc:name="Listener" doc:id="dd18126e-81f5-48ef-8f35-9dd19afdfaf0" config-ref="HTTP_Listener_config" path="/agent"/>
<mulechain:agent-define-prompt-template doc:name="Agent define prompt template" doc:id="c4f5bb4b-1bd4-49a3-8612-29f1dacbbef6" config-ref="OPENAI" template="You are a customer satisfaction agent, who analyses the customer feedback in the dataset. Answer via json output and add a type for the result only with positive or negative as well as the complete answer" instructions="If the customer feedback in the dataset is negative, open a service satisfaction case and apologize to the customer. If the customer feedback in the dataset is positive, thank the customer and wish them a nice day. Don't repeat the feedback and be more direct starting the conversation with formal greetings" dataset="#[payload.data]"/>
<ms-aichain:agent-define-prompt-template doc:name="Agent define prompt template" doc:id="c4f5bb4b-1bd4-49a3-8612-29f1dacbbef6" config-ref="OPENAI" template="You are a customer satisfaction agent, who analyses the customer feedback in the dataset. Answer via json output and add a type for the result only with positive or negative as well as the complete answer" instructions="If the customer feedback in the dataset is negative, open a service satisfaction case and apologize to the customer. If the customer feedback in the dataset is positive, thank the customer and wish them a nice day. Don't repeat the feedback and be more direct starting the conversation with formal greetings" dataset="#[payload.data]"/>
<ee:transform doc:name="formatResponse" doc:id="368aa9bc-4876-4926-99a1-75ab0b24e813" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -26,7 +26,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="ChatAnswerPrompt" doc:id="c3de899b-5e08-493f-be6a-f7cda10e8214" >
<http:listener doc:name="Listener" doc:id="ec2df0e9-8ff7-4f86-ba6a-5bcd64a092ba" config-ref="HTTP_Listener_config" path="/chat"/>
<mulechain:chat-answer-prompt doc:name="Chat answer prompt" doc:id="e2706dfd-79e6-463f-9dab-ecef6f576c22" config-ref="OPENAI" prompt="#[payload.data]"/>
<ms-aichain:chat-answer-prompt doc:name="Chat answer prompt" doc:id="e2706dfd-79e6-463f-9dab-ecef6f576c22" config-ref="OPENAI" prompt="#[payload.data]"/>
<ee:transform doc:name="formatResponse" doc:id="a8371667-3524-491c-9745-ad97f2f3f429" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -38,7 +38,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="ChatAnswerPromptWithMemory" doc:id="0036faaa-b891-42c8-96e3-a7f08660e41d" >
<http:listener doc:name="Listener" doc:id="a3d66872-10a6-4eb9-916a-436f25d92841" config-ref="HTTP_Listener_config" path="/chatMemory"/>
<mulechain:chat-answer-prompt-with-memory doc:name="Chat answer prompt with memory" doc:id="d012af47-a41d-46aa-8476-39e1a2f48d92" config-ref="OPENAI" data="#[payload.data]" memoryName="#[payload.memory]" dbFilePath="#[payload.memoryFilePath]" maxMessages="#[payload.maxMessages]"/>
<ms-aichain:chat-answer-prompt-with-memory doc:name="Chat answer prompt with memory" doc:id="d012af47-a41d-46aa-8476-39e1a2f48d92" config-ref="OPENAI" data="#[payload.data]" memoryName="#[payload.memory]" dbFilePath="#[payload.memoryFilePath]" maxMessages="#[payload.maxMessages]"/>
<ee:transform doc:name="formatResponse" doc:id="0e8fc3af-6745-4f7d-b576-c7eb1082e3f4" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -50,7 +50,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="EmbeddingNewStore" doc:id="7ec661b2-2e96-4d29-bef8-13afdc4df66f" >
<http:listener doc:name="Listener" doc:id="f57d8296-4ebf-4c85-b6bc-b154cc34d1bc" config-ref="HTTP_Listener_config" path="/embeddingNew"/>
<mulechain:embedding-new-store doc:name="Embedding new store" doc:id="8f4b53ec-f238-4af0-bc9b-f18bc8ef579a" storeName='#["embedding.store"]'/>
<ms-aichain:embedding-new-store doc:name="Embedding new store" doc:id="8f4b53ec-f238-4af0-bc9b-f18bc8ef579a" storeName='#["embedding.store"]'/>
<ee:transform doc:name="formatResponse" doc:id="f4347a0b-58c7-405b-93be-275e8533e376" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -62,7 +62,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="EmbeddingAddFile" doc:id="074f8a86-b931-477e-b3cc-3741a63a1822" >
<http:listener doc:name="Listener" doc:id="d1d00ff4-7c9d-4548-92ca-8a0fa50b3124" config-ref="HTTP_Listener_config" path="/embeddingAdd" />
<mulechain:embedding-add-document-to-store doc:name="Embedding add document to store" doc:id="7f5feb33-8302-4c06-a358-a23267b808fb" storeName='#["embedding.store"]' contextPath="#[payload.filePath]" fileType="#[payload.fileType]"/>
<ms-aichain:embedding-add-document-to-store doc:name="Embedding add document to store" doc:id="7f5feb33-8302-4c06-a358-a23267b808fb" storeName='#["embedding.store"]' contextPath="#[payload.filePath]" fileType="#[payload.fileType]"/>
<ee:transform doc:name="formatResponse" doc:id="6ad0e8d7-e75b-483d-84ed-e93b93aadb12" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -74,7 +74,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="EmbeddingAddFolder" doc:id="b84bd92d-eea3-4691-bd6a-17c03d932789" >
<http:listener doc:name="Listener" doc:id="88456faf-719b-42fa-bd1e-8716c0ccf94c" config-ref="HTTP_Listener_config" path="/embeddingAddFolder" />
<mulechain:embedding-add-folder-to-store doc:name="Embedding add folder to store" doc:id="c113688a-3f1e-44d6-8b5c-a60a65abfdea" storeName='#["embedding.store"]' contextPath="#[payload.folderPath]"/>
<ms-aichain:embedding-add-folder-to-store doc:name="Embedding add folder to store" doc:id="c113688a-3f1e-44d6-8b5c-a60a65abfdea" storeName='#["embedding.store"]' contextPath="#[payload.folderPath]"/>
<ee:transform doc:name="formatResponse" doc:id="d44d7f03-09a1-402e-af5d-f759abf12455" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -86,7 +86,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="EmbeddingQueryStore" doc:id="4dc447bb-2257-42c1-80e3-6f7c4d507737" >
<http:listener doc:name="Listener" doc:id="e16e5d95-c2d8-4778-99dc-da74ae55e6e1" config-ref="HTTP_Listener_config" path="/queryStore" />
<mulechain:embedding-query-from-store doc:name="Embedding query from store" doc:id="1ee361ea-e62a-4e0f-9c74-0363f8721052" storeName='#["embedding.store"]' question="#[payload.question]" maxResults="3" minScore="0.7" getLatest="true"/>
<ms-aichain:embedding-query-from-store doc:name="Embedding query from store" doc:id="1ee361ea-e62a-4e0f-9c74-0363f8721052" storeName='#["embedding.store"]' question="#[payload.question]" maxResults="3" minScore="0.7" getLatest="true"/>
<ee:transform doc:name="formatResponse" doc:id="b83f08cf-5b42-4025-be79-906b121f1b3e" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -98,7 +98,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="EmbeddingGetInfo" doc:id="3d54ea4d-b932-42c3-8ab1-1a24675a5caf" >
<http:listener doc:name="Listener" doc:id="a4f76b29-9203-4b65-81aa-d5be3fff4789" config-ref="HTTP_Listener_config" path="/embeddingInfo" />
<mulechain:embedding-get-info-from-store doc:name="Embedding get info from store" doc:id="ae5b31c6-d99d-4ab8-848c-03124785184b" config-ref="OPENAI" storeName='#["embedding.store"]' data="#[payload.question]" getLatest="true"/>
<ms-aichain:embedding-get-info-from-store doc:name="Embedding get info from store" doc:id="ae5b31c6-d99d-4ab8-848c-03124785184b" config-ref="OPENAI" storeName='#["embedding.store"]' data="#[payload.question]" getLatest="true"/>
<ee:transform doc:name="formatResponse" doc:id="889f816f-bbdf-4386-b707-a3f216fb857d" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -110,7 +110,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="EmbeddingAdd5" doc:id="347caa3d-ed7b-4ba8-bf74-4f184e8727d3" >
<http:listener doc:name="Listener" doc:id="6720bf73-6b8c-4bc7-8153-f21a5df4f7d8" config-ref="HTTP_Listener_config" path="/embeddingInfoLegacy" />
<mulechain:embedding-get-info-from-store-legacy doc:name="Embedding get info from store legacy" doc:id="1f9e0410-4357-4ad8-bf17-29112f20e2f6" config-ref="OPENAI" storeName='#["embedding.store"]' data="#[payload.question]"/>
<ms-aichain:embedding-get-info-from-store-legacy doc:name="Embedding get info from store legacy" doc:id="1f9e0410-4357-4ad8-bf17-29112f20e2f6" config-ref="OPENAI" storeName='#["embedding.store"]' data="#[payload.question]"/>
<ee:transform doc:name="formatResponse" doc:id="e092b180-27ca-4665-9d7b-966a1951de6a" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -122,7 +122,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="ImageGenerate" doc:id="d9dfb328-65d4-499b-8eaf-282e21a245cf" >
<http:listener doc:name="Listener" doc:id="ebd04ee3-0e40-456e-ad6e-8be37cd9f706" config-ref="HTTP_Listener_config" path="/image"/>
<mulechain:image-generate doc:name="Image generate" doc:id="e698de71-aaf5-4da3-a202-7d4d5dec5814" config-ref="OPENAI-GPT4-TURBO" data="#[payload.prompt]" />
<ms-aichain:image-generate doc:name="Image generate" doc:id="e698de71-aaf5-4da3-a202-7d4d5dec5814" config-ref="OPENAI-GPT4-TURBO" data="#[payload.prompt]" />
<ee:transform doc:name="formatResponse" doc:id="cc5a0538-d2a7-41b9-b98b-d64103f881bb">
<ee:message>
<ee:set-payload><![CDATA[%dw 2.0
Expand All @@ -142,7 +142,7 @@ output application/json
</flow>
<flow name="ImageRead" doc:id="e1603839-8651-4fff-9436-9cb66e41cd8c" >
<http:listener doc:name="Listener" doc:id="5248fffd-9678-432f-987f-b2ca0e0382b2" config-ref="HTTP_Listener_config" path="/imageRead"/>
<mulechain:image-read doc:name="Image read" doc:id="a6a11fd7-6f48-4063-820c-1d9d488dd033" config-ref="OPENAI-GPT4-TURBO" data="#[payload.prompt]" contextURL="#[payload.imageUrl]" />
<ms-aichain:image-read doc:name="Image read" doc:id="a6a11fd7-6f48-4063-820c-1d9d488dd033" config-ref="OPENAI-GPT4-TURBO" data="#[payload.prompt]" contextURL="#[payload.imageUrl]" />
<ee:transform doc:name="formatResponse" doc:id="d8b1c90a-e5f4-478b-ac2b-4c6e8362a6cb" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -154,7 +154,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="RAG" doc:id="3831de2a-109f-4864-96ea-4a9f4e1255f0" >
<http:listener doc:name="Listener" doc:id="e3db6396-bea6-4fcc-8ade-fe07dcad2150" config-ref="HTTP_Listener_config" path="/rag"/>
<mulechain:rag-load-document doc:name="Rag load document" doc:id="66b215ae-fab5-4da9-ac91-ebe1c60d6a7b" config-ref="OPENAI-GPT4-TURBO" data="#[payload.data]" contextPath="#[payload.filePath]" fileType="#[payload.fileType]"/>
<ms-aichain:rag-load-document doc:name="Rag load document" doc:id="66b215ae-fab5-4da9-ac91-ebe1c60d6a7b" config-ref="OPENAI-GPT4-TURBO" data="#[payload.data]" contextPath="#[payload.filePath]" fileType="#[payload.fileType]"/>
<ee:transform doc:name="formatResponse" doc:id="fe0fa9da-9c53-4870-8e11-40e14a31907f" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -164,9 +164,9 @@ read(payload,"application/json")]]></ee:set-payload>
</ee:message>
</ee:transform>
</flow>
<flow name="mulechain-testingFlow" doc:id="b074f6aa-3a5b-40b2-9147-6e18b02c019d" >
<flow name="ms-aichain-testingFlow" doc:id="b074f6aa-3a5b-40b2-9147-6e18b02c019d" >
<http:listener doc:name="Listener" doc:id="15e4941a-468b-48ab-b21f-61fbb38544ca" config-ref="HTTP_Listener_config" path="/sentiments"/>
<mulechain:sentiment-analyze doc:name="Sentiment analyze" doc:id="f3d0d5dc-aab8-4986-85b5-2860435657f8" config-ref="OPENAI" data="#[payload.data]"/>
<ms-aichain:sentiment-analyze doc:name="Sentiment analyze" doc:id="f3d0d5dc-aab8-4986-85b5-2860435657f8" config-ref="OPENAI" data="#[payload.data]"/>
<ee:transform doc:name="formatResponse" doc:id="efa3de35-88e7-4f2a-80e8-0b0a8e4b8015" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -178,7 +178,7 @@ read(payload,"application/json")]]></ee:set-payload>
</flow>
<flow name="ToolsAIService" doc:id="b4f5e268-13f0-40d6-a03a-47dbaac8d4f7" >
<http:listener doc:name="Listener" doc:id="c551c245-3152-4795-a857-989fbee12adc" config-ref="HTTP_Listener_config" path="/aiservice"/>
<mulechain:tools-use-ai-service doc:name="Tools use ai service" doc:id="e7a3fcc4-653e-45c8-8f95-7b22ba9e4a63" config-ref="OPENAI-GPT4-TURBO" data='#[payload.question ++ ", dont assume things and invent answers. Answer with the most probable statement."]' toolConfig='#[mule.home ++ "/apps/" ++ app.name ++ "/tools.config.json"]'/>
<ms-aichain:tools-use-ai-service doc:name="Tools use ai service" doc:id="e7a3fcc4-653e-45c8-8f95-7b22ba9e4a63" config-ref="OPENAI-GPT4-TURBO" data='#[payload.question ++ ", dont assume things and invent answers. Answer with the most probable statement."]' toolConfig='#[mule.home ++ "/apps/" ++ app.name ++ "/tools.config.json"]'/>
<ee:transform doc:name="formatResponse" doc:id="e808ec02-e4f7-4cfc-93cc-6ee9b8650568" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -188,9 +188,9 @@ read(payload,"application/json")]]></ee:set-payload>
</ee:message>
</ee:transform>
</flow>
<flow name="mulechain-testingFlow1" doc:id="c5236fd3-d578-4f91-b1f9-16c334e1cd7c" >
<flow name="ms-aichain-testingFlow1" doc:id="c5236fd3-d578-4f91-b1f9-16c334e1cd7c" >
<http:listener doc:name="Listener" doc:id="30727c05-9524-4011-9d3a-6658f4706d4e" config-ref="HTTP_Listener_config" path="/chains"/>
<mulechain:tools-use-ai-service-legacy doc:name="Tools use ai service legacy" doc:id="51aa98dc-0ad2-42ad-9fc0-0883bdb766a4" config-ref="OPENAI-GPT4-TURBO" data='#[payload.question ++ ", dont assume things and invent answers. Answer with the most probable statement."]' toolConfig='#[mule.home ++ "/apps/" ++ app.name ++ "/tools.config.json"]' />
<ms-aichain:tools-use-ai-service-legacy doc:name="Tools use ai service legacy" doc:id="51aa98dc-0ad2-42ad-9fc0-0883bdb766a4" config-ref="OPENAI-GPT4-TURBO" data='#[payload.question ++ ", dont assume things and invent answers. Answer with the most probable statement."]' toolConfig='#[mule.home ++ "/apps/" ++ app.name ++ "/tools.config.json"]' />
<ee:transform doc:name="formatResponse" doc:id="fc9bea8b-7a58-449e-a519-e7dd893d4064" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand All @@ -200,9 +200,9 @@ read(payload,"application/json")]]></ee:set-payload>
</ee:message>
</ee:transform>
</flow>
<flow name="mulechain-testingFlow2" doc:id="2f02e10e-6422-4b94-ab34-9584e5ee0ed6" >
<flow name="ms-aichain-testingFlow2" doc:id="2f02e10e-6422-4b94-ab34-9584e5ee0ed6" >
<http:listener doc:name="Listener" doc:id="0f940ad8-5210-4433-9651-79af3fbd88c4" config-ref="HTTP_Listener_config" path="/scanned"/>
<mulechain:image-read-scanned-documents doc:name="Image read scanned documents" doc:id="45154a42-ce31-4b34-aadc-dc3ad10fcdc0" config-ref="OPENAI-GPT4-TURBO" data="#[payload.prompt]" filePath="#[payload.filePath]"/>
<ms-aichain:image-read-scanned-documents doc:name="Image read scanned documents" doc:id="45154a42-ce31-4b34-aadc-dc3ad10fcdc0" config-ref="OPENAI-GPT4-TURBO" data="#[payload.prompt]" filePath="#[payload.filePath]"/>
<ee:transform doc:name="formatResponse" doc:id="020fe15b-c8e8-4c04-a4e3-812d536b4771" >
<ee:message >
<ee:set-payload ><![CDATA[%dw 2.0
Expand Down