Compare commits
3 Commits
b2401d57ae
...
675c506969
| Author | SHA1 | Date | |
|---|---|---|---|
| 675c506969 | |||
| 0be70f716e | |||
| 74a26b8c2f |
@ -3,7 +3,8 @@ Creates different AI instances, based on the given configuration.
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
from .configuration import Config
|
from typing import cast
|
||||||
|
from .configuration import Config, OpenAIConfig, default_ai_ID
|
||||||
from .ai import AI, AIError
|
from .ai import AI, AIError
|
||||||
from .ais.openai import OpenAI
|
from .ais.openai import OpenAI
|
||||||
|
|
||||||
@ -12,9 +13,14 @@ def create_ai(args: argparse.Namespace, config: Config) -> AI:
|
|||||||
"""
|
"""
|
||||||
Creates an AI subclass instance from the given args and configuration.
|
Creates an AI subclass instance from the given args and configuration.
|
||||||
"""
|
"""
|
||||||
if args.ai == 'openai':
|
if args.ai:
|
||||||
# FIXME: create actual 'OpenAIConfig' and set values from 'args'
|
ai_conf = config.ais[args.ai]
|
||||||
# FIXME: use actual name from config
|
elif default_ai_ID in config.ais:
|
||||||
return OpenAI("openai", config.openai)
|
ai_conf = config.ais[default_ai_ID]
|
||||||
|
else:
|
||||||
|
raise AIError("No AI name given and no default exists")
|
||||||
|
|
||||||
|
if ai_conf.name == 'openai':
|
||||||
|
return OpenAI(cast(OpenAIConfig, ai_conf))
|
||||||
else:
|
else:
|
||||||
raise AIError(f"AI '{args.ai}' is not supported")
|
raise AIError(f"AI '{args.ai}' is not supported")
|
||||||
|
|||||||
0
chatmastermind/ais/__init__.py
Normal file
0
chatmastermind/ais/__init__.py
Normal file
@ -18,7 +18,7 @@ class OpenAI(AI):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, config: OpenAIConfig) -> None:
|
def __init__(self, config: OpenAIConfig) -> None:
|
||||||
self.ai_type = config.ai_type
|
self.ID = config.ID
|
||||||
self.name = config.name
|
self.name = config.name
|
||||||
self.config = config
|
self.config = config
|
||||||
|
|
||||||
|
|||||||
@ -9,6 +9,8 @@ OpenAIConfigInst = TypeVar('OpenAIConfigInst', bound='OpenAIConfig')
|
|||||||
|
|
||||||
|
|
||||||
supported_ais: list[str] = ['openai']
|
supported_ais: list[str] = ['openai']
|
||||||
|
default_ai_ID: str = 'default'
|
||||||
|
default_config_path = '.config.yaml'
|
||||||
|
|
||||||
|
|
||||||
class ConfigError(Exception):
|
class ConfigError(Exception):
|
||||||
@ -20,7 +22,7 @@ class AIConfig:
|
|||||||
"""
|
"""
|
||||||
The base class of all AI configurations.
|
The base class of all AI configurations.
|
||||||
"""
|
"""
|
||||||
ai_type: str
|
ID: str
|
||||||
name: str
|
name: str
|
||||||
|
|
||||||
|
|
||||||
@ -31,11 +33,11 @@ class OpenAIConfig(AIConfig):
|
|||||||
"""
|
"""
|
||||||
# all members have default values, so we can easily create
|
# all members have default values, so we can easily create
|
||||||
# a default configuration
|
# a default configuration
|
||||||
ai_type: str = 'openai'
|
ID: str = 'default'
|
||||||
name: str = 'openai_1'
|
name: str = 'openai'
|
||||||
system: str = 'You are an assistant'
|
|
||||||
api_key: str = '0123456789'
|
api_key: str = '0123456789'
|
||||||
model: str = 'gpt-3.5'
|
system: str = 'You are an assistant'
|
||||||
|
model: str = 'gpt-3.5-turbo-16k'
|
||||||
temperature: float = 1.0
|
temperature: float = 1.0
|
||||||
max_tokens: int = 4000
|
max_tokens: int = 4000
|
||||||
top_p: float = 1.0
|
top_p: float = 1.0
|
||||||
@ -48,8 +50,8 @@ class OpenAIConfig(AIConfig):
|
|||||||
Create OpenAIConfig from a dict.
|
Create OpenAIConfig from a dict.
|
||||||
"""
|
"""
|
||||||
return cls(
|
return cls(
|
||||||
ai_type='openai',
|
ID='openai',
|
||||||
name=str(source['name']),
|
name='openai',
|
||||||
system=str(source['system']),
|
system=str(source['system']),
|
||||||
api_key=str(source['api_key']),
|
api_key=str(source['api_key']),
|
||||||
model=str(source['model']),
|
model=str(source['model']),
|
||||||
@ -64,24 +66,24 @@ class OpenAIConfig(AIConfig):
|
|||||||
return asdict(self)
|
return asdict(self)
|
||||||
|
|
||||||
|
|
||||||
def ai_type_instance(ai_type: str, conf_dict: Optional[dict[str, Any]] = None) -> AIConfig:
|
def ai_config_instance(name: str, conf_dict: Optional[dict[str, Any]] = None) -> AIConfig:
|
||||||
"""
|
"""
|
||||||
Creates an AIConfig instance of the given type.
|
Creates an AIConfig instance of the given name.
|
||||||
"""
|
"""
|
||||||
if ai_type.lower() == 'openai':
|
if name.lower() == 'openai':
|
||||||
if conf_dict is None:
|
if conf_dict is None:
|
||||||
return OpenAIConfig()
|
return OpenAIConfig()
|
||||||
else:
|
else:
|
||||||
return OpenAIConfig.from_dict(conf_dict)
|
return OpenAIConfig.from_dict(conf_dict)
|
||||||
else:
|
else:
|
||||||
raise ConfigError(f"AI type '{ai_type}' is not supported")
|
raise ConfigError(f"AI '{name}' is not supported")
|
||||||
|
|
||||||
|
|
||||||
def create_default_ai_configs() -> dict[str, AIConfig]:
|
def create_default_ai_configs() -> dict[str, AIConfig]:
|
||||||
"""
|
"""
|
||||||
Create a dict containing default configurations for all supported AIs.
|
Create a dict containing default configurations for all supported AIs.
|
||||||
"""
|
"""
|
||||||
return {ai_type_instance(ai_type).name: ai_type_instance(ai_type) for ai_type in supported_ais}
|
return {ai_config_instance(name).ID: ai_config_instance(name) for name in supported_ais}
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
@ -101,9 +103,9 @@ class Config:
|
|||||||
"""
|
"""
|
||||||
# create the correct AI type instances
|
# create the correct AI type instances
|
||||||
ais: dict[str, AIConfig] = {}
|
ais: dict[str, AIConfig] = {}
|
||||||
for name, conf in source['ais'].items():
|
for ID, conf in source['ais'].items():
|
||||||
ai_conf = ai_type_instance(conf['type'], conf)
|
ai_conf = ai_config_instance(conf['name'], conf)
|
||||||
ais[name] = ai_conf
|
ais[ID] = ai_conf
|
||||||
return cls(
|
return cls(
|
||||||
db=str(source['db']),
|
db=str(source['db']),
|
||||||
ais=ais
|
ais=ais
|
||||||
@ -121,16 +123,16 @@ class Config:
|
|||||||
def from_file(cls: Type[ConfigInst], path: str) -> ConfigInst:
|
def from_file(cls: Type[ConfigInst], path: str) -> ConfigInst:
|
||||||
with open(path, 'r') as f:
|
with open(path, 'r') as f:
|
||||||
source = yaml.load(f, Loader=yaml.FullLoader)
|
source = yaml.load(f, Loader=yaml.FullLoader)
|
||||||
# add the AI name to the config (for easy internal access)
|
# add the AI ID to the config (for easy internal access)
|
||||||
for name, conf in source['ais'].items():
|
for ID, conf in source['ais'].items():
|
||||||
conf['name'] = name
|
conf['ID'] = ID
|
||||||
return cls.from_dict(source)
|
return cls.from_dict(source)
|
||||||
|
|
||||||
def to_file(self, file_path: Path) -> None:
|
def to_file(self, file_path: Path) -> None:
|
||||||
# remove the AI name from the config (for a cleaner format)
|
# remove the AI name from the config (for a cleaner format)
|
||||||
data = self.as_dict()
|
data = self.as_dict()
|
||||||
for ai_name, ai_conf in data['ais'].items():
|
for conf in data['ais'].values():
|
||||||
del (ai_conf['name'])
|
del (conf['ID'])
|
||||||
with open(file_path, 'w') as f:
|
with open(file_path, 'w') as f:
|
||||||
yaml.dump(data, f, sort_keys=False)
|
yaml.dump(data, f, sort_keys=False)
|
||||||
|
|
||||||
|
|||||||
@ -6,61 +6,19 @@ import sys
|
|||||||
import argcomplete
|
import argcomplete
|
||||||
import argparse
|
import argparse
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from .utils import terminal_width, print_tag_args, print_chat_hist, ChatType
|
from .configuration import Config, default_config_path
|
||||||
from .storage import save_answers, create_chat_hist
|
|
||||||
from .api_client import ai, openai_api_key, print_models
|
|
||||||
from .configuration import Config
|
|
||||||
from .chat import ChatDB
|
from .chat import ChatDB
|
||||||
from .message import Message, MessageFilter, MessageError, Question
|
from .message import Message, MessageFilter, MessageError, Question
|
||||||
from .ai_factory import create_ai
|
from .ai_factory import create_ai
|
||||||
from .ai import AI, AIResponse
|
from .ai import AI, AIResponse
|
||||||
from itertools import zip_longest
|
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
default_config = '.config.yaml'
|
|
||||||
|
|
||||||
|
|
||||||
def tags_completer(prefix: str, parsed_args: Any, **kwargs: Any) -> list[str]:
|
def tags_completer(prefix: str, parsed_args: Any, **kwargs: Any) -> list[str]:
|
||||||
config = Config.from_file(parsed_args.config)
|
config = Config.from_file(parsed_args.config)
|
||||||
return list(Message.tags_from_dir(Path(config.db), prefix=prefix))
|
return list(Message.tags_from_dir(Path(config.db), prefix=prefix))
|
||||||
|
|
||||||
|
|
||||||
def create_question_with_hist(args: argparse.Namespace,
|
|
||||||
config: Config,
|
|
||||||
) -> tuple[ChatType, str, list[str]]:
|
|
||||||
"""
|
|
||||||
Creates the "AI request", including the question and chat history as determined
|
|
||||||
by the specified tags.
|
|
||||||
"""
|
|
||||||
tags = args.or_tags or []
|
|
||||||
xtags = args.exclude_tags or []
|
|
||||||
otags = args.output_tags or []
|
|
||||||
|
|
||||||
if not args.source_code_only:
|
|
||||||
print_tag_args(tags, xtags, otags)
|
|
||||||
|
|
||||||
question_parts = []
|
|
||||||
question_list = args.question if args.question is not None else []
|
|
||||||
source_list = args.source if args.source is not None else []
|
|
||||||
|
|
||||||
for question, source in zip_longest(question_list, source_list, fillvalue=None):
|
|
||||||
if question is not None and source is not None:
|
|
||||||
with open(source) as r:
|
|
||||||
question_parts.append(f"{question}\n\n```\n{r.read().strip()}\n```")
|
|
||||||
elif question is not None:
|
|
||||||
question_parts.append(question)
|
|
||||||
elif source is not None:
|
|
||||||
with open(source) as r:
|
|
||||||
question_parts.append(f"```\n{r.read().strip()}\n```")
|
|
||||||
|
|
||||||
full_question = '\n\n'.join(question_parts)
|
|
||||||
chat = create_chat_hist(full_question, tags, xtags, config,
|
|
||||||
match_all_tags=True if args.and_tags else False, # FIXME
|
|
||||||
with_tags=False,
|
|
||||||
with_file=False)
|
|
||||||
return chat, full_question, tags
|
|
||||||
|
|
||||||
|
|
||||||
def tags_cmd(args: argparse.Namespace, config: Config) -> None:
|
def tags_cmd(args: argparse.Namespace, config: Config) -> None:
|
||||||
"""
|
"""
|
||||||
Handler for the 'tags' command.
|
Handler for the 'tags' command.
|
||||||
@ -74,17 +32,12 @@ def tags_cmd(args: argparse.Namespace, config: Config) -> None:
|
|||||||
# TODO: add renaming
|
# TODO: add renaming
|
||||||
|
|
||||||
|
|
||||||
def config_cmd(args: argparse.Namespace, config: Config) -> None:
|
def config_cmd(args: argparse.Namespace) -> None:
|
||||||
"""
|
"""
|
||||||
Handler for the 'config' command.
|
Handler for the 'config' command.
|
||||||
"""
|
"""
|
||||||
if args.list_models:
|
if args.create:
|
||||||
print_models()
|
Config.create_default(Path(args.create))
|
||||||
elif args.print_model:
|
|
||||||
print(config.openai.model)
|
|
||||||
elif args.model:
|
|
||||||
config.openai.model = args.model
|
|
||||||
config.to_file(args.config)
|
|
||||||
|
|
||||||
|
|
||||||
def question_cmd(args: argparse.Namespace, config: Config) -> None:
|
def question_cmd(args: argparse.Namespace, config: Config) -> None:
|
||||||
@ -128,25 +81,6 @@ def question_cmd(args: argparse.Namespace, config: Config) -> None:
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
def ask_cmd(args: argparse.Namespace, config: Config) -> None:
|
|
||||||
"""
|
|
||||||
Handler for the 'ask' command.
|
|
||||||
"""
|
|
||||||
if args.max_tokens:
|
|
||||||
config.openai.max_tokens = args.max_tokens
|
|
||||||
if args.temperature:
|
|
||||||
config.openai.temperature = args.temperature
|
|
||||||
if args.model:
|
|
||||||
config.openai.model = args.model
|
|
||||||
chat, question, tags = create_question_with_hist(args, config)
|
|
||||||
print_chat_hist(chat, False, args.source_code_only)
|
|
||||||
otags = args.output_tags or []
|
|
||||||
answers, usage = ai(chat, config, args.num_answers)
|
|
||||||
save_answers(question, answers, tags, otags, config)
|
|
||||||
print("-" * terminal_width())
|
|
||||||
print(f"Usage: {usage}")
|
|
||||||
|
|
||||||
|
|
||||||
def hist_cmd(args: argparse.Namespace, config: Config) -> None:
|
def hist_cmd(args: argparse.Namespace, config: Config) -> None:
|
||||||
"""
|
"""
|
||||||
Handler for the 'hist' command.
|
Handler for the 'hist' command.
|
||||||
@ -182,7 +116,7 @@ def print_cmd(args: argparse.Namespace, config: Config) -> None:
|
|||||||
def create_parser() -> argparse.ArgumentParser:
|
def create_parser() -> argparse.ArgumentParser:
|
||||||
parser = argparse.ArgumentParser(
|
parser = argparse.ArgumentParser(
|
||||||
description="ChatMastermind is a Python application that automates conversation with AI")
|
description="ChatMastermind is a Python application that automates conversation with AI")
|
||||||
parser.add_argument('-C', '--config', help='Config file name.', default=default_config)
|
parser.add_argument('-C', '--config', help='Config file name.', default=default_config_path)
|
||||||
|
|
||||||
# subcommand-parser
|
# subcommand-parser
|
||||||
cmdparser = parser.add_subparsers(dest='command',
|
cmdparser = parser.add_subparsers(dest='command',
|
||||||
@ -227,22 +161,6 @@ def create_parser() -> argparse.ArgumentParser:
|
|||||||
question_cmd_parser.add_argument('-S', '--source-code-only', help='Add pure source code to the chat history',
|
question_cmd_parser.add_argument('-S', '--source-code-only', help='Add pure source code to the chat history',
|
||||||
action='store_true')
|
action='store_true')
|
||||||
|
|
||||||
# 'ask' command parser
|
|
||||||
ask_cmd_parser = cmdparser.add_parser('ask', parents=[tag_parser],
|
|
||||||
help="Ask a question.",
|
|
||||||
aliases=['a'])
|
|
||||||
ask_cmd_parser.set_defaults(func=ask_cmd)
|
|
||||||
ask_cmd_parser.add_argument('-q', '--question', nargs='+', help='Question to ask',
|
|
||||||
required=True)
|
|
||||||
ask_cmd_parser.add_argument('-m', '--max-tokens', help='Max tokens to use', type=int)
|
|
||||||
ask_cmd_parser.add_argument('-T', '--temperature', help='Temperature to use', type=float)
|
|
||||||
ask_cmd_parser.add_argument('-M', '--model', help='Model to use')
|
|
||||||
ask_cmd_parser.add_argument('-n', '--num-answers', help='Number of answers to produce', type=int,
|
|
||||||
default=1)
|
|
||||||
ask_cmd_parser.add_argument('-s', '--source', nargs='+', help='Source add content of a file to the query')
|
|
||||||
ask_cmd_parser.add_argument('-S', '--source-code-only', help='Add pure source code to the chat history',
|
|
||||||
action='store_true')
|
|
||||||
|
|
||||||
# 'hist' command parser
|
# 'hist' command parser
|
||||||
hist_cmd_parser = cmdparser.add_parser('hist', parents=[tag_parser],
|
hist_cmd_parser = cmdparser.add_parser('hist', parents=[tag_parser],
|
||||||
help="Print chat history.",
|
help="Print chat history.",
|
||||||
@ -278,7 +196,7 @@ def create_parser() -> argparse.ArgumentParser:
|
|||||||
action='store_true')
|
action='store_true')
|
||||||
config_group.add_argument('-m', '--print-model', help="Print the currently configured model",
|
config_group.add_argument('-m', '--print-model', help="Print the currently configured model",
|
||||||
action='store_true')
|
action='store_true')
|
||||||
config_group.add_argument('-M', '--model', help="Set model in the config file")
|
config_group.add_argument('-c', '--create', help="Create config with default settings in the given file")
|
||||||
|
|
||||||
# 'print' command parser
|
# 'print' command parser
|
||||||
print_cmd_parser = cmdparser.add_parser('print',
|
print_cmd_parser = cmdparser.add_parser('print',
|
||||||
@ -297,11 +215,12 @@ def main() -> int:
|
|||||||
parser = create_parser()
|
parser = create_parser()
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
command = parser.parse_args()
|
command = parser.parse_args()
|
||||||
config = Config.from_file(args.config)
|
|
||||||
|
|
||||||
openai_api_key(config.openai.api_key)
|
if command.func == config_cmd:
|
||||||
|
command.func(command)
|
||||||
command.func(command, config)
|
else:
|
||||||
|
config = Config.from_file(args.config)
|
||||||
|
command.func(command, config)
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|||||||
4
setup.py
4
setup.py
@ -12,7 +12,7 @@ setup(
|
|||||||
long_description=long_description,
|
long_description=long_description,
|
||||||
long_description_content_type="text/markdown",
|
long_description_content_type="text/markdown",
|
||||||
url="https://github.com/ok2/ChatMastermind",
|
url="https://github.com/ok2/ChatMastermind",
|
||||||
packages=find_packages(),
|
packages=find_packages() + ["chatmastermind.ais"],
|
||||||
classifiers=[
|
classifiers=[
|
||||||
"Development Status :: 3 - Alpha",
|
"Development Status :: 3 - Alpha",
|
||||||
"Environment :: Console",
|
"Environment :: Console",
|
||||||
@ -32,7 +32,7 @@ setup(
|
|||||||
"openai",
|
"openai",
|
||||||
"PyYAML",
|
"PyYAML",
|
||||||
"argcomplete",
|
"argcomplete",
|
||||||
"pytest"
|
"pytest",
|
||||||
],
|
],
|
||||||
python_requires=">=3.9",
|
python_requires=">=3.9",
|
||||||
test_suite="tests",
|
test_suite="tests",
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user