From 0d13ca2060122a8b696f9d9df837141fb7a31b20 Mon Sep 17 00:00:00 2001 From: Ian Arawjo Date: Wed, 3 May 2023 10:13:35 -0400 Subject: [PATCH] Pull from llmItemsCurrState when sending responses --- chain-forge/src/PromptNode.js | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/chain-forge/src/PromptNode.js b/chain-forge/src/PromptNode.js index f0bbc4f..e7290f5 100644 --- a/chain-forge/src/PromptNode.js +++ b/chain-forge/src/PromptNode.js @@ -136,7 +136,7 @@ const PromptNode = ({ data, id }) => { console.log('Connected!'); // Check that there is at least one LLM selected: - if (llmItems.length === 0) { + if (llmItemsCurrState.length === 0) { alert('Please select at least one LLM to prompt.') return; } @@ -191,7 +191,7 @@ const PromptNode = ({ data, id }) => { headers: {'Content-Type': 'application/json', 'Access-Control-Allow-Origin': '*'}, body: JSON.stringify({ id: id, - llm: llmItems.map(item => item.model), + llm: llmItemsCurrState.map(item => item.model), prompt: py_prompt_template, vars: pulled_data, params: {