Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,12 +1,12 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
from concurrent.futures import ThreadPoolExecutor
|
| 3 |
import fine_tuning
|
| 4 |
-
|
| 5 |
|
| 6 |
# -----------------------------
|
| 7 |
# Load fine-tuned model
|
| 8 |
# -----------------------------
|
| 9 |
-
model, tokenizer, device = load_and_train()
|
| 10 |
|
| 11 |
# -----------------------------
|
| 12 |
# RAG Backend (Stub Example)
|
|
@@ -31,7 +31,7 @@ def combined_generate(prompt, max_tokens):
|
|
| 31 |
|
| 32 |
futures = {
|
| 33 |
executor.submit(fine_tuning.generate_answer, model, tokenizer, device, prompt, max_tokens): "Fine-tuned",
|
| 34 |
-
executor.submit(
|
| 35 |
}
|
| 36 |
|
| 37 |
answers = {"Fine-tuned": "", "RAG": ""}
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
from concurrent.futures import ThreadPoolExecutor
|
| 3 |
import fine_tuning
|
| 4 |
+
import rag
|
| 5 |
|
| 6 |
# -----------------------------
|
| 7 |
# Load fine-tuned model
|
| 8 |
# -----------------------------
|
| 9 |
+
model, tokenizer, device = fine_tuning.load_and_train()
|
| 10 |
|
| 11 |
# -----------------------------
|
| 12 |
# RAG Backend (Stub Example)
|
|
|
|
| 31 |
|
| 32 |
futures = {
|
| 33 |
executor.submit(fine_tuning.generate_answer, model, tokenizer, device, prompt, max_tokens): "Fine-tuned",
|
| 34 |
+
executor.submit(rag.generate_answer, prompt): "RAG",
|
| 35 |
}
|
| 36 |
|
| 37 |
answers = {"Fine-tuned": "", "RAG": ""}
|