Spaces:
Sleeping
Sleeping
fix load rag
Browse files
app.py
CHANGED
|
@@ -250,6 +250,9 @@ if "save_dir" not in st.session_state:
|
|
| 250 |
if "uploaded_files" not in st.session_state:
|
| 251 |
st.session_state.uploaded_files = set()
|
| 252 |
|
|
|
|
|
|
|
|
|
|
| 253 |
@st.dialog("Setup Gemini")
|
| 254 |
def vote():
|
| 255 |
st.markdown(
|
|
@@ -347,8 +350,10 @@ def compute_rag_chain(_model, _embd, docs_texts):
|
|
| 347 |
if st.session_state.uploaded_files:
|
| 348 |
if st.session_state.gemini_api is not None:
|
| 349 |
if st.session_state.rag is None:
|
|
|
|
| 350 |
docs_texts = [d.page_content for d in documents]
|
| 351 |
st.session_state.rag = compute_rag_chain(st.session_state.model, st.session_state.embd, docs_texts)
|
|
|
|
| 352 |
|
| 353 |
if st.session_state.gemini_api is not None:
|
| 354 |
if st.session_state.llm is None:
|
|
@@ -372,20 +377,23 @@ for message in st.session_state.chat_history:
|
|
| 372 |
st.write(message["content"])
|
| 373 |
|
| 374 |
prompt = st.chat_input("Bạn muốn hỏi gì?")
|
| 375 |
-
|
| 376 |
-
|
| 377 |
-
|
| 378 |
-
|
| 379 |
-
|
| 380 |
-
st.write(prompt)
|
| 381 |
-
|
| 382 |
-
with st.chat_message("assistant"):
|
| 383 |
-
if st.session_state.rag is not None:
|
| 384 |
-
respone = st.session_state.rag.invoke(prompt)
|
| 385 |
-
st.write(respone)
|
| 386 |
-
else:
|
| 387 |
-
ans = st.session_state.llm.invoke(prompt)
|
| 388 |
-
respone = ans.content
|
| 389 |
-
st.write(respone)
|
| 390 |
|
| 391 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 250 |
if "uploaded_files" not in st.session_state:
|
| 251 |
st.session_state.uploaded_files = set()
|
| 252 |
|
| 253 |
+
if "processing" not in st.session_state:
|
| 254 |
+
st.session_state.processing = False
|
| 255 |
+
|
| 256 |
@st.dialog("Setup Gemini")
|
| 257 |
def vote():
|
| 258 |
st.markdown(
|
|
|
|
| 350 |
if st.session_state.uploaded_files:
|
| 351 |
if st.session_state.gemini_api is not None:
|
| 352 |
if st.session_state.rag is None:
|
| 353 |
+
st.session_state.processing = True
|
| 354 |
docs_texts = [d.page_content for d in documents]
|
| 355 |
st.session_state.rag = compute_rag_chain(st.session_state.model, st.session_state.embd, docs_texts)
|
| 356 |
+
st.session_state.processing = False
|
| 357 |
|
| 358 |
if st.session_state.gemini_api is not None:
|
| 359 |
if st.session_state.llm is None:
|
|
|
|
| 377 |
st.write(message["content"])
|
| 378 |
|
| 379 |
prompt = st.chat_input("Bạn muốn hỏi gì?")
|
| 380 |
+
|
| 381 |
+
if not st.session_state.processing:
|
| 382 |
+
if st.session_state.gemini_api:
|
| 383 |
+
if prompt:
|
| 384 |
+
st.session_state.chat_history.append({"role": "user", "content": prompt})
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 385 |
|
| 386 |
+
with st.chat_message("user"):
|
| 387 |
+
st.write(prompt)
|
| 388 |
+
|
| 389 |
+
with st.chat_message("assistant"):
|
| 390 |
+
if st.session_state.rag is not None:
|
| 391 |
+
respone = st.session_state.rag.invoke(prompt)
|
| 392 |
+
st.write(respone)
|
| 393 |
+
else:
|
| 394 |
+
ans = st.session_state.llm.invoke(prompt)
|
| 395 |
+
respone = ans.content
|
| 396 |
+
st.write(respone)
|
| 397 |
+
|
| 398 |
+
st.session_state.chat_history.append({"role": "assistant", "content": respone})
|
| 399 |
+
|