Compare commits

..

1 Commits

Author SHA1 Message Date
b2401d57ae configuration: implemented new Config format 2023-09-07 08:15:37 +02:00
6 changed files with 120 additions and 47 deletions

View File

@ -3,8 +3,7 @@ Creates different AI instances, based on the given configuration.
""" """
import argparse import argparse
from typing import cast from .configuration import Config
from .configuration import Config, OpenAIConfig, default_ai_ID
from .ai import AI, AIError from .ai import AI, AIError
from .ais.openai import OpenAI from .ais.openai import OpenAI
@ -13,14 +12,9 @@ def create_ai(args: argparse.Namespace, config: Config) -> AI:
""" """
Creates an AI subclass instance from the given args and configuration. Creates an AI subclass instance from the given args and configuration.
""" """
if args.ai: if args.ai == 'openai':
ai_conf = config.ais[args.ai] # FIXME: create actual 'OpenAIConfig' and set values from 'args'
elif default_ai_ID in config.ais: # FIXME: use actual name from config
ai_conf = config.ais[default_ai_ID] return OpenAI("openai", config.openai)
else:
raise AIError("No AI name given and no default exists")
if ai_conf.name == 'openai':
return OpenAI(cast(OpenAIConfig, ai_conf))
else: else:
raise AIError(f"AI '{args.ai}' is not supported") raise AIError(f"AI '{args.ai}' is not supported")

View File

@ -18,7 +18,7 @@ class OpenAI(AI):
""" """
def __init__(self, config: OpenAIConfig) -> None: def __init__(self, config: OpenAIConfig) -> None:
self.ID = config.ID self.ai_type = config.ai_type
self.name = config.name self.name = config.name
self.config = config self.config = config

View File

@ -9,8 +9,6 @@ OpenAIConfigInst = TypeVar('OpenAIConfigInst', bound='OpenAIConfig')
supported_ais: list[str] = ['openai'] supported_ais: list[str] = ['openai']
default_ai_ID: str = 'default'
default_config_path = '.config.yaml'
class ConfigError(Exception): class ConfigError(Exception):
@ -22,7 +20,7 @@ class AIConfig:
""" """
The base class of all AI configurations. The base class of all AI configurations.
""" """
ID: str ai_type: str
name: str name: str
@ -33,11 +31,11 @@ class OpenAIConfig(AIConfig):
""" """
# all members have default values, so we can easily create # all members have default values, so we can easily create
# a default configuration # a default configuration
ID: str = 'default' ai_type: str = 'openai'
name: str = 'openai' name: str = 'openai_1'
api_key: str = '0123456789'
system: str = 'You are an assistant' system: str = 'You are an assistant'
model: str = 'gpt-3.5-turbo-16k' api_key: str = '0123456789'
model: str = 'gpt-3.5'
temperature: float = 1.0 temperature: float = 1.0
max_tokens: int = 4000 max_tokens: int = 4000
top_p: float = 1.0 top_p: float = 1.0
@ -50,8 +48,8 @@ class OpenAIConfig(AIConfig):
Create OpenAIConfig from a dict. Create OpenAIConfig from a dict.
""" """
return cls( return cls(
ID='openai', ai_type='openai',
name='openai', name=str(source['name']),
system=str(source['system']), system=str(source['system']),
api_key=str(source['api_key']), api_key=str(source['api_key']),
model=str(source['model']), model=str(source['model']),
@ -66,24 +64,24 @@ class OpenAIConfig(AIConfig):
return asdict(self) return asdict(self)
def ai_config_instance(name: str, conf_dict: Optional[dict[str, Any]] = None) -> AIConfig: def ai_type_instance(ai_type: str, conf_dict: Optional[dict[str, Any]] = None) -> AIConfig:
""" """
Creates an AIConfig instance of the given name. Creates an AIConfig instance of the given type.
""" """
if name.lower() == 'openai': if ai_type.lower() == 'openai':
if conf_dict is None: if conf_dict is None:
return OpenAIConfig() return OpenAIConfig()
else: else:
return OpenAIConfig.from_dict(conf_dict) return OpenAIConfig.from_dict(conf_dict)
else: else:
raise ConfigError(f"AI '{name}' is not supported") raise ConfigError(f"AI type '{ai_type}' is not supported")
def create_default_ai_configs() -> dict[str, AIConfig]: def create_default_ai_configs() -> dict[str, AIConfig]:
""" """
Create a dict containing default configurations for all supported AIs. Create a dict containing default configurations for all supported AIs.
""" """
return {ai_config_instance(name).ID: ai_config_instance(name) for name in supported_ais} return {ai_type_instance(ai_type).name: ai_type_instance(ai_type) for ai_type in supported_ais}
@dataclass @dataclass
@ -103,9 +101,9 @@ class Config:
""" """
# create the correct AI type instances # create the correct AI type instances
ais: dict[str, AIConfig] = {} ais: dict[str, AIConfig] = {}
for ID, conf in source['ais'].items(): for name, conf in source['ais'].items():
ai_conf = ai_config_instance(conf['name'], conf) ai_conf = ai_type_instance(conf['type'], conf)
ais[ID] = ai_conf ais[name] = ai_conf
return cls( return cls(
db=str(source['db']), db=str(source['db']),
ais=ais ais=ais
@ -123,16 +121,16 @@ class Config:
def from_file(cls: Type[ConfigInst], path: str) -> ConfigInst: def from_file(cls: Type[ConfigInst], path: str) -> ConfigInst:
with open(path, 'r') as f: with open(path, 'r') as f:
source = yaml.load(f, Loader=yaml.FullLoader) source = yaml.load(f, Loader=yaml.FullLoader)
# add the AI ID to the config (for easy internal access) # add the AI name to the config (for easy internal access)
for ID, conf in source['ais'].items(): for name, conf in source['ais'].items():
conf['ID'] = ID conf['name'] = name
return cls.from_dict(source) return cls.from_dict(source)
def to_file(self, file_path: Path) -> None: def to_file(self, file_path: Path) -> None:
# remove the AI name from the config (for a cleaner format) # remove the AI name from the config (for a cleaner format)
data = self.as_dict() data = self.as_dict()
for conf in data['ais'].values(): for ai_name, ai_conf in data['ais'].items():
del (conf['ID']) del (ai_conf['name'])
with open(file_path, 'w') as f: with open(file_path, 'w') as f:
yaml.dump(data, f, sort_keys=False) yaml.dump(data, f, sort_keys=False)

View File

@ -6,19 +6,61 @@ import sys
import argcomplete import argcomplete
import argparse import argparse
from pathlib import Path from pathlib import Path
from .configuration import Config, default_config_path from .utils import terminal_width, print_tag_args, print_chat_hist, ChatType
from .storage import save_answers, create_chat_hist
from .api_client import ai, openai_api_key, print_models
from .configuration import Config
from .chat import ChatDB from .chat import ChatDB
from .message import Message, MessageFilter, MessageError, Question from .message import Message, MessageFilter, MessageError, Question
from .ai_factory import create_ai from .ai_factory import create_ai
from .ai import AI, AIResponse from .ai import AI, AIResponse
from itertools import zip_longest
from typing import Any from typing import Any
default_config = '.config.yaml'
def tags_completer(prefix: str, parsed_args: Any, **kwargs: Any) -> list[str]: def tags_completer(prefix: str, parsed_args: Any, **kwargs: Any) -> list[str]:
config = Config.from_file(parsed_args.config) config = Config.from_file(parsed_args.config)
return list(Message.tags_from_dir(Path(config.db), prefix=prefix)) return list(Message.tags_from_dir(Path(config.db), prefix=prefix))
def create_question_with_hist(args: argparse.Namespace,
config: Config,
) -> tuple[ChatType, str, list[str]]:
"""
Creates the "AI request", including the question and chat history as determined
by the specified tags.
"""
tags = args.or_tags or []
xtags = args.exclude_tags or []
otags = args.output_tags or []
if not args.source_code_only:
print_tag_args(tags, xtags, otags)
question_parts = []
question_list = args.question if args.question is not None else []
source_list = args.source if args.source is not None else []
for question, source in zip_longest(question_list, source_list, fillvalue=None):
if question is not None and source is not None:
with open(source) as r:
question_parts.append(f"{question}\n\n```\n{r.read().strip()}\n```")
elif question is not None:
question_parts.append(question)
elif source is not None:
with open(source) as r:
question_parts.append(f"```\n{r.read().strip()}\n```")
full_question = '\n\n'.join(question_parts)
chat = create_chat_hist(full_question, tags, xtags, config,
match_all_tags=True if args.and_tags else False, # FIXME
with_tags=False,
with_file=False)
return chat, full_question, tags
def tags_cmd(args: argparse.Namespace, config: Config) -> None: def tags_cmd(args: argparse.Namespace, config: Config) -> None:
""" """
Handler for the 'tags' command. Handler for the 'tags' command.
@ -32,12 +74,17 @@ def tags_cmd(args: argparse.Namespace, config: Config) -> None:
# TODO: add renaming # TODO: add renaming
def config_cmd(args: argparse.Namespace) -> None: def config_cmd(args: argparse.Namespace, config: Config) -> None:
""" """
Handler for the 'config' command. Handler for the 'config' command.
""" """
if args.create: if args.list_models:
Config.create_default(Path(args.create)) print_models()
elif args.print_model:
print(config.openai.model)
elif args.model:
config.openai.model = args.model
config.to_file(args.config)
def question_cmd(args: argparse.Namespace, config: Config) -> None: def question_cmd(args: argparse.Namespace, config: Config) -> None:
@ -81,6 +128,25 @@ def question_cmd(args: argparse.Namespace, config: Config) -> None:
pass pass
def ask_cmd(args: argparse.Namespace, config: Config) -> None:
"""
Handler for the 'ask' command.
"""
if args.max_tokens:
config.openai.max_tokens = args.max_tokens
if args.temperature:
config.openai.temperature = args.temperature
if args.model:
config.openai.model = args.model
chat, question, tags = create_question_with_hist(args, config)
print_chat_hist(chat, False, args.source_code_only)
otags = args.output_tags or []
answers, usage = ai(chat, config, args.num_answers)
save_answers(question, answers, tags, otags, config)
print("-" * terminal_width())
print(f"Usage: {usage}")
def hist_cmd(args: argparse.Namespace, config: Config) -> None: def hist_cmd(args: argparse.Namespace, config: Config) -> None:
""" """
Handler for the 'hist' command. Handler for the 'hist' command.
@ -116,7 +182,7 @@ def print_cmd(args: argparse.Namespace, config: Config) -> None:
def create_parser() -> argparse.ArgumentParser: def create_parser() -> argparse.ArgumentParser:
parser = argparse.ArgumentParser( parser = argparse.ArgumentParser(
description="ChatMastermind is a Python application that automates conversation with AI") description="ChatMastermind is a Python application that automates conversation with AI")
parser.add_argument('-C', '--config', help='Config file name.', default=default_config_path) parser.add_argument('-C', '--config', help='Config file name.', default=default_config)
# subcommand-parser # subcommand-parser
cmdparser = parser.add_subparsers(dest='command', cmdparser = parser.add_subparsers(dest='command',
@ -161,6 +227,22 @@ def create_parser() -> argparse.ArgumentParser:
question_cmd_parser.add_argument('-S', '--source-code-only', help='Add pure source code to the chat history', question_cmd_parser.add_argument('-S', '--source-code-only', help='Add pure source code to the chat history',
action='store_true') action='store_true')
# 'ask' command parser
ask_cmd_parser = cmdparser.add_parser('ask', parents=[tag_parser],
help="Ask a question.",
aliases=['a'])
ask_cmd_parser.set_defaults(func=ask_cmd)
ask_cmd_parser.add_argument('-q', '--question', nargs='+', help='Question to ask',
required=True)
ask_cmd_parser.add_argument('-m', '--max-tokens', help='Max tokens to use', type=int)
ask_cmd_parser.add_argument('-T', '--temperature', help='Temperature to use', type=float)
ask_cmd_parser.add_argument('-M', '--model', help='Model to use')
ask_cmd_parser.add_argument('-n', '--num-answers', help='Number of answers to produce', type=int,
default=1)
ask_cmd_parser.add_argument('-s', '--source', nargs='+', help='Source add content of a file to the query')
ask_cmd_parser.add_argument('-S', '--source-code-only', help='Add pure source code to the chat history',
action='store_true')
# 'hist' command parser # 'hist' command parser
hist_cmd_parser = cmdparser.add_parser('hist', parents=[tag_parser], hist_cmd_parser = cmdparser.add_parser('hist', parents=[tag_parser],
help="Print chat history.", help="Print chat history.",
@ -196,7 +278,7 @@ def create_parser() -> argparse.ArgumentParser:
action='store_true') action='store_true')
config_group.add_argument('-m', '--print-model', help="Print the currently configured model", config_group.add_argument('-m', '--print-model', help="Print the currently configured model",
action='store_true') action='store_true')
config_group.add_argument('-c', '--create', help="Create config with default settings in the given file") config_group.add_argument('-M', '--model', help="Set model in the config file")
# 'print' command parser # 'print' command parser
print_cmd_parser = cmdparser.add_parser('print', print_cmd_parser = cmdparser.add_parser('print',
@ -215,12 +297,11 @@ def main() -> int:
parser = create_parser() parser = create_parser()
args = parser.parse_args() args = parser.parse_args()
command = parser.parse_args() command = parser.parse_args()
config = Config.from_file(args.config)
if command.func == config_cmd: openai_api_key(config.openai.api_key)
command.func(command)
else: command.func(command, config)
config = Config.from_file(args.config)
command.func(command, config)
return 0 return 0

View File

@ -12,7 +12,7 @@ setup(
long_description=long_description, long_description=long_description,
long_description_content_type="text/markdown", long_description_content_type="text/markdown",
url="https://github.com/ok2/ChatMastermind", url="https://github.com/ok2/ChatMastermind",
packages=find_packages() + ["chatmastermind.ais"], packages=find_packages(),
classifiers=[ classifiers=[
"Development Status :: 3 - Alpha", "Development Status :: 3 - Alpha",
"Environment :: Console", "Environment :: Console",
@ -32,7 +32,7 @@ setup(
"openai", "openai",
"PyYAML", "PyYAML",
"argcomplete", "argcomplete",
"pytest", "pytest"
], ],
python_requires=">=3.9", python_requires=">=3.9",
test_suite="tests", test_suite="tests",