mirror of
https://github.com/ianarawjo/ChainForge.git
synced 2025-03-14 08:16:37 +00:00
Add newest GPT-4 models (#153)
This commit is contained in:
parent
c5088527bb
commit
0b32792d07
@ -1,15 +1,15 @@
|
||||
{
|
||||
"files": {
|
||||
"main.css": "/static/css/main.8665fcca.css",
|
||||
"main.js": "/static/js/main.ae4e3642.js",
|
||||
"main.js": "/static/js/main.38f29425.js",
|
||||
"static/js/787.4c72bb55.chunk.js": "/static/js/787.4c72bb55.chunk.js",
|
||||
"index.html": "/index.html",
|
||||
"main.8665fcca.css.map": "/static/css/main.8665fcca.css.map",
|
||||
"main.ae4e3642.js.map": "/static/js/main.ae4e3642.js.map",
|
||||
"main.38f29425.js.map": "/static/js/main.38f29425.js.map",
|
||||
"787.4c72bb55.chunk.js.map": "/static/js/787.4c72bb55.chunk.js.map"
|
||||
},
|
||||
"entrypoints": [
|
||||
"static/css/main.8665fcca.css",
|
||||
"static/js/main.ae4e3642.js"
|
||||
"static/js/main.38f29425.js"
|
||||
]
|
||||
}
|
@ -1 +1 @@
|
||||
<!doctype html><html lang="en"><head><meta charset="utf-8"/><script async src="https://www.googletagmanager.com/gtag/js?id=G-RN3FDBLMCR"></script><script>function gtag(){dataLayer.push(arguments)}window.dataLayer=window.dataLayer||[],gtag("js",new Date),gtag("config","G-RN3FDBLMCR")</script><link rel="icon" href="/favicon.ico"/><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="theme-color" content="#000000"/><meta name="description" content="A visual programming environment for prompt engineering"/><link rel="apple-touch-icon" href="/logo192.png"/><link rel="manifest" href="/manifest.json"/><title>ChainForge</title><script defer="defer" src="/static/js/main.ae4e3642.js"></script><link href="/static/css/main.8665fcca.css" rel="stylesheet"></head><body><noscript>You need to enable JavaScript to run this app.</noscript><div id="root"></div></body></html>
|
||||
<!doctype html><html lang="en"><head><meta charset="utf-8"/><script async src="https://www.googletagmanager.com/gtag/js?id=G-RN3FDBLMCR"></script><script>function gtag(){dataLayer.push(arguments)}window.dataLayer=window.dataLayer||[],gtag("js",new Date),gtag("config","G-RN3FDBLMCR")</script><link rel="icon" href="/favicon.ico"/><meta name="viewport" content="width=device-width,initial-scale=1"/><meta name="theme-color" content="#000000"/><meta name="description" content="A visual programming environment for prompt engineering"/><link rel="apple-touch-icon" href="/logo192.png"/><link rel="manifest" href="/manifest.json"/><title>ChainForge</title><script defer="defer" src="/static/js/main.38f29425.js"></script><link href="/static/css/main.8665fcca.css" rel="stylesheet"></head><body><noscript>You need to enable JavaScript to run this app.</noscript><div id="root"></div></body></html>
|
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@ -41,8 +41,8 @@ const ChatGPTSettings = {
|
||||
"type": "string",
|
||||
"title": "Model Version",
|
||||
"description": "Select an OpenAI model to query. For more details on the differences, see the OpenAI API documentation.",
|
||||
"enum": ["gpt-3.5-turbo", "gpt-3.5-turbo-0613", "gpt-3.5-turbo-0301", "gpt-3.5-turbo-16k", "gpt-3.5-turbo-16k-0613", "gpt-4", "gpt-4-0613", "gpt-4-0314", "gpt-4-32k", "gpt-4-32k-0613", "gpt-4-32k-0314", "text-davinci-003", "text-davinci-002", "code-davinci-002"],
|
||||
"default": "gpt-3.5-turbo"
|
||||
"enum": ["gpt-3.5-turbo", "gpt-3.5-turbo-1106", "gpt-3.5-turbo-0613", "gpt-3.5-turbo-0301", "gpt-3.5-turbo-16k", "gpt-3.5-turbo-16k-0613", "gpt-4", "gpt-4-1106-preview","gpt-4-32k", "gpt-4-0613", "gpt-4-0314", "gpt-4-32k-0613", "gpt-4-32k-0314", "gpt-3.5-turbo-instruct", "text-davinci-003", "text-davinci-002", "code-davinci-002"],
|
||||
"default": "gpt-3.5-turbo",
|
||||
},
|
||||
"system_msg": {
|
||||
"type": "string",
|
||||
@ -60,6 +60,13 @@ const ChatGPTSettings = {
|
||||
"maximum": 2,
|
||||
"multipleOf": 0.01
|
||||
},
|
||||
"response_format": {
|
||||
"type": "string",
|
||||
"title": "response_format",
|
||||
"enum": ["text", "json_object"],
|
||||
"description": "An object specifying the format that the model must output. Currently, can only be text or JSON. Only works with newest GPT models. IMPORTANT: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. Without this, the model may generate an unending stream of whitespace until the generation reaches the token limit, resulting in a long-running and seemingly 'stuck' request.",
|
||||
"default": "text",
|
||||
},
|
||||
"functions": {
|
||||
"type": "string",
|
||||
"title": "functions",
|
||||
@ -87,6 +94,12 @@ const ChatGPTSettings = {
|
||||
"description": "Up to 4 sequences where the API will stop generating further tokens. Enclose stop sequences in double-quotes \"\" and use whitespace to separate them.",
|
||||
"default": ""
|
||||
},
|
||||
"seed": {
|
||||
"type": "integer",
|
||||
"title": "seed",
|
||||
"description": "If specified, the OpenAI API will make a best effort to sample deterministically, such that repeated requests with the same seed and parameters should return the same result. Determinism is not guaranteed.",
|
||||
"allow_empty_str": true,
|
||||
},
|
||||
"max_tokens": {
|
||||
"type": "integer",
|
||||
"title": "max_tokens",
|
||||
@ -110,7 +123,7 @@ const ChatGPTSettings = {
|
||||
"maximum": 2,
|
||||
"multipleOf": 0.005
|
||||
},
|
||||
"logit_bias": {
|
||||
"logit_bias": {
|
||||
"type": "string",
|
||||
"title": "logit_bias",
|
||||
"description": "Modify the likelihood of specified tokens appearing in the completion. Accepts a json object that maps tokens (specified by their token ID in the tokenizer) to an associated bias value from -100 to 100. Mathematically, the bias is added to the logits generated by the model prior to sampling. The exact effect will vary per model, but values between -1 and 1 should decrease or increase likelihood of selection; values like -100 or 100 should result in a ban or exclusive selection of the relevant token."
|
||||
@ -134,6 +147,9 @@ const ChatGPTSettings = {
|
||||
"ui:help": "Defaults to 1.0. Leave at default if you prefer to set top_p.",
|
||||
"ui:widget": "range"
|
||||
},
|
||||
"response_format": {
|
||||
"ui:help": "Defaults to text.",
|
||||
},
|
||||
"functions": {
|
||||
"ui:help": "Leave blank to not specify any functions. NOTE: JSON schema MUST NOT have trailing commas.",
|
||||
"ui:widget": "textarea",
|
||||
@ -160,6 +176,9 @@ const ChatGPTSettings = {
|
||||
"max_tokens": {
|
||||
"ui:help": "Defaults to infinity."
|
||||
},
|
||||
"seed": {
|
||||
"ui:help": "Defaults to blank (no seed)."
|
||||
},
|
||||
"logit_bias": {
|
||||
"ui:widget": "textarea",
|
||||
"ui:help": "Defaults to none."
|
||||
@ -180,8 +199,11 @@ const ChatGPTSettings = {
|
||||
'stop': (str) => {
|
||||
if (str.trim().length === 0) return [];
|
||||
return str.match(/"((?:[^"\\]|\\.)*)"/g).map(s => s.substring(1, s.length-1)); // split on double-quotes but exclude escaped double-quotes inside the group
|
||||
}
|
||||
}
|
||||
},
|
||||
'response_format': (str) => {
|
||||
return { type: str };
|
||||
},
|
||||
}
|
||||
};
|
||||
|
||||
const GPT4Settings = {
|
||||
@ -273,13 +295,13 @@ const ClaudeSettings = {
|
||||
"default": -1
|
||||
},
|
||||
"top_p": {
|
||||
"type": "number",
|
||||
"title": "top_p",
|
||||
"description": "Does nucleus sampling, in which we compute the cumulative distribution over all the options for each subsequent token in decreasing probability order and cut it off once it reaches a particular probability specified by top_p. Defaults to -1, which disables it. Note that you should either alter temperature or top_p, but not both.",
|
||||
"default": -1,
|
||||
"minimum": -1,
|
||||
"maximum": 1,
|
||||
"multipleOf": 0.001,
|
||||
"type": "number",
|
||||
"title": "top_p",
|
||||
"description": "Does nucleus sampling, in which we compute the cumulative distribution over all the options for each subsequent token in decreasing probability order and cut it off once it reaches a particular probability specified by top_p. Defaults to -1, which disables it. Note that you should either alter temperature or top_p, but not both.",
|
||||
"default": -1,
|
||||
"minimum": -1,
|
||||
"maximum": 1,
|
||||
"multipleOf": 0.001,
|
||||
},
|
||||
}
|
||||
},
|
||||
@ -342,11 +364,7 @@ const PaLM2Settings = {
|
||||
"title": "Model",
|
||||
"description": "Select a PaLM model to query. For more details on the differences, see the Google PaLM API documentation.",
|
||||
"enum": ["text-bison-001", "chat-bison-001"],
|
||||
"default": "chat-bison-001",
|
||||
"shortname_map": {
|
||||
"text-bison-001": "PaLM2-text",
|
||||
"chat-bison-001": "PaLM2-chat",
|
||||
}
|
||||
"default": "chat-bison-001"
|
||||
},
|
||||
"temperature": {
|
||||
"type": "number",
|
||||
@ -618,19 +636,9 @@ const HuggingFaceTextInferenceSettings = {
|
||||
"model": {
|
||||
"type": "string",
|
||||
"title": "Model",
|
||||
"description": "Select a suggested HuggingFace-hosted model to query using the Inference API. For more details, check out https://huggingface.co/inference-api",
|
||||
"enum": ["mistralai/Mistral-7B-Instruct-v0.1", "HuggingFaceH4/zephyr-7b-beta", "tiiuae/falcon-7b-instruct", "microsoft/DialoGPT-large", "bigscience/bloom-560m", "gpt2", "bigcode/santacoder", "bigcode/starcoder", "Other (HuggingFace)"],
|
||||
"default": "tiiuae/falcon-7b-instruct",
|
||||
"shortname_map": {
|
||||
"mistralai/Mistral-7B-Instruct-v0.1": "Mistral-7B",
|
||||
"HuggingFaceH4/zephyr-7b-beta": "Zephyr-7B",
|
||||
"tiiuae/falcon-7b-instruct": "Falcon-7B",
|
||||
"microsoft/DialoGPT-large": "DialoGPT",
|
||||
"bigscience/bloom-560m": "Bloom560M",
|
||||
"gpt2": "GPT-2",
|
||||
"bigcode/santacoder": "santacoder",
|
||||
"bigcode/starcoder": "starcoder"
|
||||
}
|
||||
"description": "Select a suggested HuggingFace-hosted model to query using the Inference API. For more details, check out https://huggingface.co/inference-api",
|
||||
"enum": ["tiiuae/falcon-7b-instruct", "microsoft/DialoGPT-large", "bigscience/bloom-560m", "gpt2", "bigcode/santacoder", "bigcode/starcoder", "Other (HuggingFace)"],
|
||||
"default": "tiiuae/falcon-7b-instruct",
|
||||
},
|
||||
"custom_model": {
|
||||
"type": "string",
|
||||
@ -673,9 +681,9 @@ const HuggingFaceTextInferenceSettings = {
|
||||
"type": "number",
|
||||
"title": "top_p",
|
||||
"description": "Sets the maximum cumulative probability of tokens to sample from (from 0 to 1.0). Set to -1 to remain unspecified.",
|
||||
"default": -1,
|
||||
"minimum": -1,
|
||||
"maximum": 1,
|
||||
"default": -1,
|
||||
"minimum": -1,
|
||||
"maximum": 1,
|
||||
"multipleOf": 0.001,
|
||||
},
|
||||
"repetition_penalty": {
|
||||
@ -783,14 +791,6 @@ const AlephAlphaLuminousSettings = {
|
||||
"luminous-supreme-control",
|
||||
],
|
||||
default: "luminous-base",
|
||||
shortname_map: {
|
||||
"luminous-extended": "luminous-ext",
|
||||
"luminous-extended-control": "luminous-ext-ctrl",
|
||||
"luminous-base-control": "luminous-base-ctrl",
|
||||
"luminous-base": "luminous-base",
|
||||
"luminous-supreme": "luminous-supr",
|
||||
"luminous-supreme-control": "luminous-supr-ctrl",
|
||||
}
|
||||
},
|
||||
temperature: {
|
||||
type: "number",
|
||||
@ -1097,7 +1097,7 @@ export const getTemperatureSpecForModel = (modelName) => {
|
||||
};
|
||||
|
||||
export const postProcessFormData = (settingsSpec, formData) => {
|
||||
// Strip all 'model' and 'shortname' props in the submitted form, as these are passed elsewhere or unnecessary for the backend
|
||||
// Strip all 'model' and 'shortname' props in the submitted form, as these are passed elsewhere or unecessary for the backend
|
||||
const skip_keys = {'model': true, 'shortname': true};
|
||||
|
||||
let new_data = {};
|
||||
|
@ -166,6 +166,8 @@ export async function call_chatgpt(prompt: string, model: LLM, n: number = 1, te
|
||||
let modelname: string = model.toString();
|
||||
if (params?.stop !== undefined && (!Array.isArray(params.stop) || params.stop.length === 0))
|
||||
delete params.stop;
|
||||
if (params?.seed !== undefined && (params.seed.toString().length === 0))
|
||||
delete params?.seed;
|
||||
if (params?.functions !== undefined && (!Array.isArray(params.functions) || params.functions.length === 0))
|
||||
delete params?.functions;
|
||||
if (params?.function_call !== undefined && ((!(typeof params.function_call === 'string')) || params.function_call.trim().length === 0))
|
||||
|
Loading…
x
Reference in New Issue
Block a user