Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
System message as part of Conversation
Browse files
src/lib/components/InferencePlayground/InferencePlayground.svelte
CHANGED
|
@@ -20,16 +20,17 @@
|
|
| 20 |
|
| 21 |
export let models: ModelEntryWithTokenizer[];
|
| 22 |
|
| 23 |
-
const
|
|
|
|
| 24 |
|
| 25 |
let conversation: Conversation = {
|
| 26 |
model: models[0],
|
| 27 |
config: defaultGenerationConfig,
|
| 28 |
-
messages: [{ ...
|
|
|
|
| 29 |
streaming: true,
|
| 30 |
};
|
| 31 |
|
| 32 |
-
let systemMessage: ChatCompletionInputMessage = { role: "system", content: "" };
|
| 33 |
let hfToken: string | undefined = import.meta.env.VITE_HF_TOKEN;
|
| 34 |
let viewCode = false;
|
| 35 |
let showTokenModal = false;
|
|
@@ -41,7 +42,7 @@
|
|
| 41 |
$: systemPromptSupported = isSystemPromptSupported(conversation.model);
|
| 42 |
$: {
|
| 43 |
if (!systemPromptSupported) {
|
| 44 |
-
systemMessage = { role: "system", content: "" };
|
| 45 |
}
|
| 46 |
}
|
| 47 |
|
|
@@ -61,8 +62,8 @@
|
|
| 61 |
}
|
| 62 |
|
| 63 |
function reset() {
|
| 64 |
-
systemMessage.content = "";
|
| 65 |
-
conversation.messages = [{ ...
|
| 66 |
}
|
| 67 |
|
| 68 |
function abort() {
|
|
@@ -98,12 +99,11 @@
|
|
| 98 |
conversation.messages = [...conversation.messages];
|
| 99 |
}
|
| 100 |
},
|
| 101 |
-
abortController
|
| 102 |
-
systemMessage
|
| 103 |
);
|
| 104 |
} else {
|
| 105 |
waitForNonStreaming = true;
|
| 106 |
-
const newMessage = await handleNonStreamingResponse(hf, conversation
|
| 107 |
// check if the user did not abort the request
|
| 108 |
if (waitForNonStreaming) {
|
| 109 |
conversation.messages = [...conversation.messages, newMessage];
|
|
@@ -162,7 +162,7 @@
|
|
| 162 |
placeholder={systemPromptSupported
|
| 163 |
? "Enter a custom prompt"
|
| 164 |
: "System prompt is not supported with the chosen model."}
|
| 165 |
-
bind:value={systemMessage.content}
|
| 166 |
class="absolute inset-x-0 bottom-0 h-full resize-none bg-transparent px-3 pt-10 text-sm outline-none"
|
| 167 |
></textarea>
|
| 168 |
</div>
|
|
|
|
| 20 |
|
| 21 |
export let models: ModelEntryWithTokenizer[];
|
| 22 |
|
| 23 |
+
const startMessageUser: ChatCompletionInputMessage = { role: "user", content: "" };
|
| 24 |
+
const startMessageSystem: ChatCompletionInputMessage = { role: "system", content: "" };
|
| 25 |
|
| 26 |
let conversation: Conversation = {
|
| 27 |
model: models[0],
|
| 28 |
config: defaultGenerationConfig,
|
| 29 |
+
messages: [{ ...startMessageUser }],
|
| 30 |
+
systemMessage: startMessageSystem,
|
| 31 |
streaming: true,
|
| 32 |
};
|
| 33 |
|
|
|
|
| 34 |
let hfToken: string | undefined = import.meta.env.VITE_HF_TOKEN;
|
| 35 |
let viewCode = false;
|
| 36 |
let showTokenModal = false;
|
|
|
|
| 42 |
$: systemPromptSupported = isSystemPromptSupported(conversation.model);
|
| 43 |
$: {
|
| 44 |
if (!systemPromptSupported) {
|
| 45 |
+
conversation.systemMessage = { role: "system", content: "" };
|
| 46 |
}
|
| 47 |
}
|
| 48 |
|
|
|
|
| 62 |
}
|
| 63 |
|
| 64 |
function reset() {
|
| 65 |
+
conversation.systemMessage.content = "";
|
| 66 |
+
conversation.messages = [{ ...startMessageUser }];
|
| 67 |
}
|
| 68 |
|
| 69 |
function abort() {
|
|
|
|
| 99 |
conversation.messages = [...conversation.messages];
|
| 100 |
}
|
| 101 |
},
|
| 102 |
+
abortController
|
|
|
|
| 103 |
);
|
| 104 |
} else {
|
| 105 |
waitForNonStreaming = true;
|
| 106 |
+
const newMessage = await handleNonStreamingResponse(hf, conversation);
|
| 107 |
// check if the user did not abort the request
|
| 108 |
if (waitForNonStreaming) {
|
| 109 |
conversation.messages = [...conversation.messages, newMessage];
|
|
|
|
| 162 |
placeholder={systemPromptSupported
|
| 163 |
? "Enter a custom prompt"
|
| 164 |
: "System prompt is not supported with the chosen model."}
|
| 165 |
+
bind:value={conversation.systemMessage.content}
|
| 166 |
class="absolute inset-x-0 bottom-0 h-full resize-none bg-transparent px-3 pt-10 text-sm outline-none"
|
| 167 |
></textarea>
|
| 168 |
</div>
|
src/lib/components/InferencePlayground/inferencePlaygroundUtils.ts
CHANGED
|
@@ -11,17 +11,17 @@ export async function handleStreamingResponse(
|
|
| 11 |
hf: HfInference,
|
| 12 |
conversation: Conversation,
|
| 13 |
onChunk: (content: string) => void,
|
| 14 |
-
abortController: AbortController
|
| 15 |
-
systemMessage?: ChatCompletionInputMessage
|
| 16 |
): Promise<void> {
|
|
|
|
| 17 |
const messages = [
|
| 18 |
-
...(isSystemPromptSupported(
|
| 19 |
...conversation.messages,
|
| 20 |
];
|
| 21 |
let out = "";
|
| 22 |
for await (const chunk of hf.chatCompletionStream(
|
| 23 |
{
|
| 24 |
-
model:
|
| 25 |
messages,
|
| 26 |
temperature: conversation.config.temperature,
|
| 27 |
max_tokens: conversation.config.maxTokens,
|
|
@@ -37,16 +37,16 @@ export async function handleStreamingResponse(
|
|
| 37 |
|
| 38 |
export async function handleNonStreamingResponse(
|
| 39 |
hf: HfInference,
|
| 40 |
-
conversation: Conversation
|
| 41 |
-
systemMessage?: ChatCompletionInputMessage
|
| 42 |
): Promise<ChatCompletionInputMessage> {
|
|
|
|
| 43 |
const messages = [
|
| 44 |
-
...(isSystemPromptSupported(
|
| 45 |
...conversation.messages,
|
| 46 |
];
|
| 47 |
|
| 48 |
const response = await hf.chatCompletion({
|
| 49 |
-
model:
|
| 50 |
messages,
|
| 51 |
temperature: conversation.config.temperature,
|
| 52 |
max_tokens: conversation.config.maxTokens,
|
|
|
|
| 11 |
hf: HfInference,
|
| 12 |
conversation: Conversation,
|
| 13 |
onChunk: (content: string) => void,
|
| 14 |
+
abortController: AbortController
|
|
|
|
| 15 |
): Promise<void> {
|
| 16 |
+
const { model, systemMessage } = conversation;
|
| 17 |
const messages = [
|
| 18 |
+
...(isSystemPromptSupported(model) && systemMessage.content?.length ? [systemMessage] : []),
|
| 19 |
...conversation.messages,
|
| 20 |
];
|
| 21 |
let out = "";
|
| 22 |
for await (const chunk of hf.chatCompletionStream(
|
| 23 |
{
|
| 24 |
+
model: model.id,
|
| 25 |
messages,
|
| 26 |
temperature: conversation.config.temperature,
|
| 27 |
max_tokens: conversation.config.maxTokens,
|
|
|
|
| 37 |
|
| 38 |
export async function handleNonStreamingResponse(
|
| 39 |
hf: HfInference,
|
| 40 |
+
conversation: Conversation
|
|
|
|
| 41 |
): Promise<ChatCompletionInputMessage> {
|
| 42 |
+
const { model, systemMessage } = conversation;
|
| 43 |
const messages = [
|
| 44 |
+
...(isSystemPromptSupported(model) && systemMessage.content?.length ? [systemMessage] : []),
|
| 45 |
...conversation.messages,
|
| 46 |
];
|
| 47 |
|
| 48 |
const response = await hf.chatCompletion({
|
| 49 |
+
model: model.id,
|
| 50 |
messages,
|
| 51 |
temperature: conversation.config.temperature,
|
| 52 |
max_tokens: conversation.config.maxTokens,
|
src/lib/components/InferencePlayground/types.ts
CHANGED
|
@@ -6,6 +6,7 @@ export type Conversation = {
|
|
| 6 |
model: ModelEntryWithTokenizer;
|
| 7 |
config: GenerationConfig;
|
| 8 |
messages: ChatCompletionInputMessage[];
|
|
|
|
| 9 |
streaming: boolean;
|
| 10 |
};
|
| 11 |
|
|
|
|
| 6 |
model: ModelEntryWithTokenizer;
|
| 7 |
config: GenerationConfig;
|
| 8 |
messages: ChatCompletionInputMessage[];
|
| 9 |
+
systemMessage: ChatCompletionInputMessage;
|
| 10 |
streaming: boolean;
|
| 11 |
};
|
| 12 |
|