Update app.py
Browse files
app.py
CHANGED
|
@@ -2,7 +2,7 @@ import gradio as gr
|
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 3 |
|
| 4 |
# モデルとトークナイザーの読み込み
|
| 5 |
-
model_name = "
|
| 6 |
tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False) # Slow tokenizerを使用
|
| 7 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 8 |
|
|
@@ -32,7 +32,7 @@ def respond(input_text, system_message, max_tokens, temperature, top_p):
|
|
| 32 |
|
| 33 |
# Gradioインターフェースの作成
|
| 34 |
with gr.Blocks() as demo:
|
| 35 |
-
gr.Markdown("##
|
| 36 |
|
| 37 |
# 追加の入力フィールドをリストで設定
|
| 38 |
additional_inputs = [
|
|
|
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 3 |
|
| 4 |
# モデルとトークナイザーの読み込み
|
| 5 |
+
model_name = "llm-jp/llm-jp-3-3.7b-instruct"
|
| 6 |
tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False) # Slow tokenizerを使用
|
| 7 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 8 |
|
|
|
|
| 32 |
|
| 33 |
# Gradioインターフェースの作成
|
| 34 |
with gr.Blocks() as demo:
|
| 35 |
+
gr.Markdown("## llmjpチャットボット")
|
| 36 |
|
| 37 |
# 追加の入力フィールドをリストで設定
|
| 38 |
additional_inputs = [
|