Compare commits
2 Commits
820d4aed21
...
39707e2ce3
| Author | SHA1 | Date | |
|---|---|---|---|
| 39707e2ce3 | |||
| 84961e5d0d |
@ -1,3 +1,4 @@
|
|||||||
|
import sys
|
||||||
import argparse
|
import argparse
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from itertools import zip_longest
|
from itertools import zip_longest
|
||||||
@ -51,7 +52,7 @@ def add_file_as_code(question_parts: list[str], file: str) -> None:
|
|||||||
|
|
||||||
def create_message(chat: ChatDB, args: argparse.Namespace) -> Message:
|
def create_message(chat: ChatDB, args: argparse.Namespace) -> Message:
|
||||||
"""
|
"""
|
||||||
Creates a new message from the given arguments and writes it
|
Create a new message from the given arguments and write it
|
||||||
to the cache directory.
|
to the cache directory.
|
||||||
"""
|
"""
|
||||||
question_parts = []
|
question_parts = []
|
||||||
@ -73,10 +74,37 @@ def create_message(chat: ChatDB, args: argparse.Namespace) -> Message:
|
|||||||
tags=args.output_tags, # FIXME
|
tags=args.output_tags, # FIXME
|
||||||
ai=args.AI,
|
ai=args.AI,
|
||||||
model=args.model)
|
model=args.model)
|
||||||
chat.cache_add([message])
|
# only write the new message to the cache,
|
||||||
|
# don't add it to the internal list
|
||||||
|
chat.cache_write([message])
|
||||||
return message
|
return message
|
||||||
|
|
||||||
|
|
||||||
|
def make_request(ai: AI, chat: ChatDB, message: Message, args: argparse.Namespace) -> None:
|
||||||
|
"""
|
||||||
|
Make an AI request with the given AI, chat history, message and arguments.
|
||||||
|
Write the response(s) to the cache directory, without appending it to the
|
||||||
|
given chat history. Then print the response(s).
|
||||||
|
"""
|
||||||
|
# print history and message question before making the request
|
||||||
|
ai.print()
|
||||||
|
chat.print(paged=False)
|
||||||
|
print(message.to_str())
|
||||||
|
response: AIResponse = ai.request(message,
|
||||||
|
chat,
|
||||||
|
args.num_answers, # FIXME
|
||||||
|
args.output_tags) # FIXME
|
||||||
|
# only write the response messages to the cache,
|
||||||
|
# don't add them to the internal list
|
||||||
|
chat.cache_write(response.messages)
|
||||||
|
for idx, msg in enumerate(response.messages):
|
||||||
|
print(f"=== ANSWER {idx+1} ===")
|
||||||
|
print(msg.answer)
|
||||||
|
if response.tokens:
|
||||||
|
print("===============")
|
||||||
|
print(response.tokens)
|
||||||
|
|
||||||
|
|
||||||
def question_cmd(args: argparse.Namespace, config: Config) -> None:
|
def question_cmd(args: argparse.Namespace, config: Config) -> None:
|
||||||
"""
|
"""
|
||||||
Handler for the 'question' command.
|
Handler for the 'question' command.
|
||||||
@ -95,28 +123,29 @@ def question_cmd(args: argparse.Namespace, config: Config) -> None:
|
|||||||
|
|
||||||
# create the correct AI instance
|
# create the correct AI instance
|
||||||
ai: AI = create_ai(args, config)
|
ai: AI = create_ai(args, config)
|
||||||
|
|
||||||
|
# === ASK ===
|
||||||
if args.ask:
|
if args.ask:
|
||||||
ai.print()
|
make_request(ai, chat, message, args)
|
||||||
chat.print(paged=False)
|
# === REPEAT ===
|
||||||
response: AIResponse = ai.request(message,
|
|
||||||
chat,
|
|
||||||
args.num_answers, # FIXME
|
|
||||||
args.output_tags) # FIXME
|
|
||||||
chat.msg_update([response.messages[0]])
|
|
||||||
chat.cache_add(response.messages[1:])
|
|
||||||
for idx, msg in enumerate(response.messages):
|
|
||||||
print(f"=== ANSWER {idx+1} ===")
|
|
||||||
print(msg.answer)
|
|
||||||
if response.tokens:
|
|
||||||
print("===============")
|
|
||||||
print(response.tokens)
|
|
||||||
elif args.repeat is not None:
|
elif args.repeat is not None:
|
||||||
lmessage = chat.msg_latest()
|
lmessage = chat.msg_latest(loc='cache')
|
||||||
assert lmessage
|
if lmessage is None:
|
||||||
# TODO: repeat either the last question or the
|
print("No message found to repeat!")
|
||||||
# one(s) given in 'args.repeat' (overwrite
|
sys.exit(1)
|
||||||
# existing ones if 'args.overwrite' is True)
|
else:
|
||||||
pass
|
print(f"Repeating message '{lmessage.msg_id()}':")
|
||||||
|
# overwrite the latest message if requested or empty
|
||||||
|
if lmessage.answer is None or args.overwrite is True:
|
||||||
|
lmessage.clear_answer()
|
||||||
|
make_request(ai, chat, lmessage, args)
|
||||||
|
# otherwise create a new one
|
||||||
|
else:
|
||||||
|
args.ask = [lmessage.question]
|
||||||
|
message = create_message(chat, args)
|
||||||
|
make_request(ai, chat, message, args)
|
||||||
|
|
||||||
|
# === PROCESS ===
|
||||||
elif args.process is not None:
|
elif args.process is not None:
|
||||||
# TODO: process either all questions without an
|
# TODO: process either all questions without an
|
||||||
# answer or the one(s) given in 'args.process'
|
# answer or the one(s) given in 'args.process'
|
||||||
|
|||||||
@ -540,6 +540,9 @@ class Message():
|
|||||||
if self.tags:
|
if self.tags:
|
||||||
self.tags = rename_tags(self.tags, tags_rename)
|
self.tags = rename_tags(self.tags, tags_rename)
|
||||||
|
|
||||||
|
def clear_answer(self) -> None:
|
||||||
|
self.answer = None
|
||||||
|
|
||||||
def msg_id(self) -> str:
|
def msg_id(self) -> str:
|
||||||
"""
|
"""
|
||||||
Returns an ID that is unique throughout all messages in the same (DB) directory.
|
Returns an ID that is unique throughout all messages in the same (DB) directory.
|
||||||
|
|||||||
@ -3,10 +3,15 @@ import unittest
|
|||||||
import argparse
|
import argparse
|
||||||
import tempfile
|
import tempfile
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from unittest.mock import MagicMock
|
from unittest import mock
|
||||||
from chatmastermind.commands.question import create_message
|
from unittest.mock import MagicMock, call, ANY
|
||||||
|
from typing import Optional
|
||||||
|
from chatmastermind.configuration import Config
|
||||||
|
from chatmastermind.commands.question import create_message, question_cmd
|
||||||
|
from chatmastermind.tags import Tag
|
||||||
from chatmastermind.message import Message, Question, Answer
|
from chatmastermind.message import Message, Question, Answer
|
||||||
from chatmastermind.chat import ChatDB
|
from chatmastermind.chat import Chat, ChatDB
|
||||||
|
from chatmastermind.ai import AI, AIResponse, Tokens
|
||||||
|
|
||||||
|
|
||||||
class TestMessageCreate(unittest.TestCase):
|
class TestMessageCreate(unittest.TestCase):
|
||||||
@ -16,10 +21,10 @@ class TestMessageCreate(unittest.TestCase):
|
|||||||
"""
|
"""
|
||||||
def setUp(self) -> None:
|
def setUp(self) -> None:
|
||||||
# create ChatDB structure
|
# create ChatDB structure
|
||||||
self.db_path = tempfile.TemporaryDirectory()
|
self.db_dir = tempfile.TemporaryDirectory()
|
||||||
self.cache_path = tempfile.TemporaryDirectory()
|
self.cache_dir = tempfile.TemporaryDirectory()
|
||||||
self.chat = ChatDB.from_dir(cache_path=Path(self.cache_path.name),
|
self.chat = ChatDB.from_dir(cache_path=Path(self.cache_dir.name),
|
||||||
db_path=Path(self.db_path.name))
|
db_path=Path(self.db_dir.name))
|
||||||
# create some messages
|
# create some messages
|
||||||
self.message_text = Message(Question("What is this?"),
|
self.message_text = Message(Question("What is this?"),
|
||||||
Answer("It is pure text"))
|
Answer("It is pure text"))
|
||||||
@ -74,6 +79,7 @@ Aaaand again some text."""
|
|||||||
os.remove(self.source_file1.name)
|
os.remove(self.source_file1.name)
|
||||||
os.remove(self.source_file2.name)
|
os.remove(self.source_file2.name)
|
||||||
os.remove(self.source_file3.name)
|
os.remove(self.source_file3.name)
|
||||||
|
os.remove(self.source_file4.name)
|
||||||
|
|
||||||
def message_list(self, tmp_dir: tempfile.TemporaryDirectory) -> list[Path]:
|
def message_list(self, tmp_dir: tempfile.TemporaryDirectory) -> list[Path]:
|
||||||
# exclude '.next'
|
# exclude '.next'
|
||||||
@ -81,10 +87,10 @@ Aaaand again some text."""
|
|||||||
|
|
||||||
def test_message_file_created(self) -> None:
|
def test_message_file_created(self) -> None:
|
||||||
self.args.ask = ["What is this?"]
|
self.args.ask = ["What is this?"]
|
||||||
cache_dir_files = self.message_list(self.cache_path)
|
cache_dir_files = self.message_list(self.cache_dir)
|
||||||
self.assertEqual(len(cache_dir_files), 0)
|
self.assertEqual(len(cache_dir_files), 0)
|
||||||
create_message(self.chat, self.args)
|
create_message(self.chat, self.args)
|
||||||
cache_dir_files = self.message_list(self.cache_path)
|
cache_dir_files = self.message_list(self.cache_dir)
|
||||||
self.assertEqual(len(cache_dir_files), 1)
|
self.assertEqual(len(cache_dir_files), 1)
|
||||||
message = Message.from_file(cache_dir_files[0])
|
message = Message.from_file(cache_dir_files[0])
|
||||||
self.assertIsInstance(message, Message)
|
self.assertIsInstance(message, Message)
|
||||||
@ -193,3 +199,135 @@ This is embedded source code.
|
|||||||
It is embedded code
|
It is embedded code
|
||||||
```
|
```
|
||||||
"""))
|
"""))
|
||||||
|
|
||||||
|
|
||||||
|
class TestQuestionCmd(unittest.TestCase):
|
||||||
|
|
||||||
|
def setUp(self) -> None:
|
||||||
|
# create DB and cache
|
||||||
|
self.db_dir = tempfile.TemporaryDirectory()
|
||||||
|
self.cache_dir = tempfile.TemporaryDirectory()
|
||||||
|
# create configuration
|
||||||
|
self.config = Config()
|
||||||
|
self.config.cache = self.cache_dir.name
|
||||||
|
self.config.db = self.db_dir.name
|
||||||
|
# create a mock argparse.Namespace
|
||||||
|
self.args = argparse.Namespace(
|
||||||
|
ask=['What is the meaning of life?'],
|
||||||
|
num_answers=1,
|
||||||
|
output_tags=['science'],
|
||||||
|
AI='openai',
|
||||||
|
model='gpt-3.5-turbo',
|
||||||
|
or_tags=None,
|
||||||
|
and_tags=None,
|
||||||
|
exclude_tags=None,
|
||||||
|
source_text=None,
|
||||||
|
source_code=None,
|
||||||
|
create=None,
|
||||||
|
repeat=None,
|
||||||
|
process=None
|
||||||
|
)
|
||||||
|
# create a mock AI instance
|
||||||
|
self.ai = MagicMock(spec=AI)
|
||||||
|
self.ai.request.side_effect = self.mock_request
|
||||||
|
|
||||||
|
def input_message(self, args: argparse.Namespace) -> Message:
|
||||||
|
"""
|
||||||
|
Create the expected input message for a question using the
|
||||||
|
given arguments.
|
||||||
|
"""
|
||||||
|
# NOTE: we only use the first question from the "ask" list
|
||||||
|
# -> message creation using "question.create_message()" is
|
||||||
|
# tested above
|
||||||
|
# the answer is always empty for the input message
|
||||||
|
return Message(Question(args.ask[0]),
|
||||||
|
tags=args.output_tags,
|
||||||
|
ai=args.AI,
|
||||||
|
model=args.model)
|
||||||
|
|
||||||
|
def mock_request(self,
|
||||||
|
question: Message,
|
||||||
|
chat: Chat,
|
||||||
|
num_answers: int = 1,
|
||||||
|
otags: Optional[set[Tag]] = None) -> AIResponse:
|
||||||
|
"""
|
||||||
|
Mock the 'ai.request()' function
|
||||||
|
"""
|
||||||
|
question.answer = Answer("Answer 0")
|
||||||
|
question.tags = otags
|
||||||
|
question.ai = 'FakeAI'
|
||||||
|
question.model = 'FakeModel'
|
||||||
|
answers: list[Message] = [question]
|
||||||
|
for n in range(1, num_answers):
|
||||||
|
answers.append(Message(question=question.question,
|
||||||
|
answer=Answer(f"Answer {n}"),
|
||||||
|
tags=otags,
|
||||||
|
ai='FakeAI',
|
||||||
|
model='FakeModel'))
|
||||||
|
return AIResponse(answers, Tokens(10, 10, 20))
|
||||||
|
|
||||||
|
def message_list(self, tmp_dir: tempfile.TemporaryDirectory) -> list[Path]:
|
||||||
|
# exclude '.next'
|
||||||
|
return sorted([f for f in Path(tmp_dir.name).glob('*.[ty]*')])
|
||||||
|
|
||||||
|
@mock.patch('chatmastermind.commands.question.create_ai')
|
||||||
|
def test_ask_single_answer(self, mock_create_ai: MagicMock) -> None:
|
||||||
|
"""
|
||||||
|
Test single answer with no errors
|
||||||
|
"""
|
||||||
|
mock_create_ai.return_value = self.ai
|
||||||
|
expected_question = self.input_message(self.args)
|
||||||
|
expected_responses = self.mock_request(expected_question,
|
||||||
|
Chat([]),
|
||||||
|
self.args.num_answers,
|
||||||
|
self.args.output_tags).messages
|
||||||
|
|
||||||
|
# execute the command
|
||||||
|
question_cmd(self.args, self.config)
|
||||||
|
|
||||||
|
# check for correct request call
|
||||||
|
self.ai.request.assert_called_once_with(expected_question,
|
||||||
|
ANY,
|
||||||
|
self.args.num_answers,
|
||||||
|
self.args.output_tags)
|
||||||
|
# check for the expected message files
|
||||||
|
chat = ChatDB.from_dir(Path(self.cache_dir.name),
|
||||||
|
Path(self.db_dir.name))
|
||||||
|
cached_msg = chat.msg_gather(loc='cache')
|
||||||
|
self.assertEqual(len(self.message_list(self.cache_dir)), 1)
|
||||||
|
self.assertSequenceEqual(cached_msg, expected_responses)
|
||||||
|
|
||||||
|
@mock.patch('chatmastermind.commands.question.ChatDB.from_dir')
|
||||||
|
@mock.patch('chatmastermind.commands.question.create_ai')
|
||||||
|
def test_ask_single_answer_mocked(self, mock_create_ai: MagicMock, mock_from_dir: MagicMock) -> None:
|
||||||
|
"""
|
||||||
|
Test single answer with no errors (mocked ChatDB version)
|
||||||
|
"""
|
||||||
|
chat = MagicMock(spec=ChatDB)
|
||||||
|
mock_from_dir.return_value = chat
|
||||||
|
|
||||||
|
mock_create_ai.return_value = self.ai
|
||||||
|
expected_question = self.input_message(self.args)
|
||||||
|
expected_responses = self.mock_request(expected_question,
|
||||||
|
Chat([]),
|
||||||
|
self.args.num_answers,
|
||||||
|
self.args.output_tags).messages
|
||||||
|
|
||||||
|
# execute the command
|
||||||
|
question_cmd(self.args, self.config)
|
||||||
|
|
||||||
|
# check for correct request call
|
||||||
|
self.ai.request.assert_called_once_with(expected_question,
|
||||||
|
chat,
|
||||||
|
self.args.num_answers,
|
||||||
|
self.args.output_tags)
|
||||||
|
|
||||||
|
# check for the correct ChatDB calls:
|
||||||
|
# - initial question has been written (prior to the actual request)
|
||||||
|
# - responses have been written (after the request)
|
||||||
|
chat.cache_write.assert_has_calls([call([expected_question]),
|
||||||
|
call(expected_responses)],
|
||||||
|
any_order=False)
|
||||||
|
|
||||||
|
# check that the messages have not been added to the internal message list
|
||||||
|
chat.cache_add.assert_not_called()
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user