File size: 3,420 Bytes
dfe3441
 
 
 
 
 
 
0112b0e
 
dfe3441
 
 
 
0112b0e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bbcc74a
 
 
 
 
 
 
 
 
0112b0e
 
bbcc74a
 
 
 
 
0112b0e
 
bbcc74a
 
288b789
 
 
 
 
 
 
 
 
 
 
 
 
483a574
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bbcc74a
483a574
 
0112b0e
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
from openai import AsyncOpenAI  # importing openai for API usage
import chainlit as cl  # importing chainlit for our app
from chainlit.prompt import Prompt, PromptMessage  # importing prompt tools
from chainlit.playground.providers import ChatOpenAI  # importing ChatOpenAI tools
from dotenv import load_dotenv
import asyncio
import datetime
from aimakerspace.text_utils import TextFileLoader, CharacterTextSplitter
from aimakerspace.vectordatabase import VectorDatabase
from aimakerspace.openai_utils.prompts import (
    UserRolePrompt,
    SystemRolePrompt,
)

load_dotenv()

text_loader = TextFileLoader("data/KingLear.txt")
documents = text_loader.load_documents()

text_splitter = CharacterTextSplitter()
split_documents = text_splitter.split_texts(documents)

#openai.api_key = getpass("OpenAI API Key: ")
#os.environ["OPENAI_API_KEY"] = openai.api_key

vector_db = VectorDatabase()
vector_db = asyncio.run(vector_db.abuild_from_list(split_documents))

from aimakerspace.openai_utils.prompts import (
    UserRolePrompt,
    SystemRolePrompt,
    AssistantRolePrompt,
)

RAQA_PROMPT_TEMPLATE = """
Use the provided context to answer the user's query. 

You may not answer the user's query unless there is specific context in the following text.

If you do not know the answer, or cannot answer, please respond with "I don't know".

Context:
{context}
"""

raqa_prompt = SystemRolePrompt(RAQA_PROMPT_TEMPLATE)

USER_PROMPT_TEMPLATE = """
User Query:
{user_query}
"""

user_prompt = UserRolePrompt(USER_PROMPT_TEMPLATE)

@cl.on_chat_start
async def start_chat():    
    settings = {
        "model": "gpt-3.5-turbo",
        "temperature": 0,
        "max_tokens": 1000,
        "top_p": 1,
        "frequency_penalty": 0,
        "presence_penalty": 0,
    }
    cl.user_session.set("settings", settings)


@cl.on_message
async def main(message: cl.Message):
    settings = cl.user_session.get("settings")

    client = AsyncOpenAI()

    context_list = vector_db.search_by_text(message.content, k=4)
    context_prompt = ""
    for context in context_list:
        context_prompt += context[0] + "\n"
    formatted_system_prompt = raqa_prompt.create_message(context=context_prompt)
    formatted_user_prompt = user_prompt.create_message(user_query=message.content)
    print(formatted_system_prompt)
    print(formatted_user_prompt)
    prompt = Prompt(
        provider=ChatOpenAI.id,
        messages=[
            PromptMessage(
                role="system",
                template=RAQA_PROMPT_TEMPLATE,
                formatted=formatted_system_prompt['content'],
            ),
            PromptMessage(
                role="user",
                template=USER_PROMPT_TEMPLATE,
                formatted=formatted_user_prompt['content'],
            ),
        ],
        inputs={"context": context_prompt,
            "user_query": message.content},
        settings=settings,
    )

    msg = cl.Message(content="")
    
    async for stream_resp in await client.chat.completions.create(
        messages=[m.to_openai() for m in prompt.messages], stream=True, **settings
    ):
        token = stream_resp.choices[0].delta.content
        if not token:
            token = ""
        await msg.stream_token(token)
        
    # Update the prompt object with the completion
    prompt.completion = msg.content
    msg.prompt = prompt

    # Send and close the message stream
    await msg.send()