Spaces:
Sleeping
Sleeping
Update chatbot.py
Browse files- chatbot.py +6 -4
chatbot.py
CHANGED
@@ -4,9 +4,11 @@ import json
|
|
4 |
from groq import Groq
|
5 |
from transformers import AutoTokenizer
|
6 |
from langchain_core.runnables import RunnablePassthrough
|
7 |
-
from langchain.memory import
|
8 |
-
|
9 |
-
from
|
|
|
|
|
10 |
from langchain_community.vectorstores import FAISS
|
11 |
from deep_translator import GoogleTranslator
|
12 |
|
@@ -31,7 +33,7 @@ class Comsatsbot:
|
|
31 |
return len(tokenizer.encode(text, add_special_tokens=False))
|
32 |
|
33 |
# Use updated memory setup
|
34 |
-
self.memory =
|
35 |
|
36 |
|
37 |
self.chats_collection = chats_collection
|
|
|
4 |
from groq import Groq
|
5 |
from transformers import AutoTokenizer
|
6 |
from langchain_core.runnables import RunnablePassthrough
|
7 |
+
from langchain.memory import ConversationBufferMemory
|
8 |
+
|
9 |
+
from langchain_openai import ChatOpenAI
|
10 |
+
from langchain.document_loaders import CSVLoader
|
11 |
+
|
12 |
from langchain_community.vectorstores import FAISS
|
13 |
from deep_translator import GoogleTranslator
|
14 |
|
|
|
33 |
return len(tokenizer.encode(text, add_special_tokens=False))
|
34 |
|
35 |
# Use updated memory setup
|
36 |
+
self.memory = ConversationBufferMemory(memory_key="chat_history", return_messages=True, max_token_limit=3000, tokenizer=llama_token_count)
|
37 |
|
38 |
|
39 |
self.chats_collection = chats_collection
|