mirror of
https://github.com/danielmiessler/fabric
synced 2024-11-10 07:10:31 +00:00
196 lines
7.3 KiB
Python
Executable File
196 lines
7.3 KiB
Python
Executable File
from .utils import Standalone, Update, Setup, Alias, run_electron_app
|
|
import argparse
|
|
import sys
|
|
import os
|
|
|
|
|
|
script_directory = os.path.dirname(os.path.realpath(__file__))
|
|
|
|
|
|
def main():
|
|
parser = argparse.ArgumentParser(
|
|
description="An open source framework for augmenting humans using AI."
|
|
)
|
|
parser.add_argument("--text", "-t", help="Text to extract summary from")
|
|
parser.add_argument(
|
|
"--copy", "-C", help="Copy the response to the clipboard", action="store_true"
|
|
)
|
|
parser.add_argument(
|
|
'--agents', '-a',
|
|
help="Use praisonAI to create an AI agent and then use it. ex: 'write me a movie script'", action="store_true"
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--output",
|
|
"-o",
|
|
help="Save the response to a file",
|
|
nargs="?",
|
|
const="analyzepaper.txt",
|
|
default=None,
|
|
)
|
|
parser.add_argument('--session', '-S',
|
|
help="Continue your previous conversation. Default is your previous conversation", nargs="?", const="default")
|
|
parser.add_argument('--clearsession', help="deletes indicated session. Use 'all' to delete all dessions")
|
|
parser.add_argument(
|
|
"--gui", help="Use the GUI (Node and npm need to be installed)", action="store_true")
|
|
parser.add_argument(
|
|
"--stream",
|
|
"-s",
|
|
help="Use this option if you want to see the results in realtime. NOTE: You will not be able to pipe the output into another command.",
|
|
action="store_true",
|
|
)
|
|
parser.add_argument(
|
|
"--list", "-l", help="List available patterns", action="store_true"
|
|
)
|
|
parser.add_argument(
|
|
'--temp', help="set the temperature for the model. Default is 0", default=0, type=float)
|
|
parser.add_argument(
|
|
'--top_p', help="set the top_p for the model. Default is 1", default=1, type=float)
|
|
parser.add_argument(
|
|
'--frequency_penalty', help="set the frequency penalty for the model. Default is 0.1", default=0.1, type=float)
|
|
parser.add_argument(
|
|
'--presence_penalty', help="set the presence penalty for the model. Default is 0.1", default=0.1, type=float)
|
|
parser.add_argument(
|
|
"--update", "-u", help="Update patterns. NOTE: This will revert the default model to gpt4-turbo. please run --changeDefaultModel to once again set default model", action="store_true")
|
|
parser.add_argument("--pattern", "-p", help="The pattern (prompt) to use")
|
|
parser.add_argument(
|
|
"--setup", help="Set up your fabric instance", action="store_true"
|
|
)
|
|
parser.add_argument('--changeDefaultModel',
|
|
help="Change the default model. For a list of available models, use the --listmodels flag.")
|
|
|
|
parser.add_argument(
|
|
"--model", "-m", help="Select the model to use"
|
|
)
|
|
parser.add_argument(
|
|
"--listmodels", help="List all available models", action="store_true"
|
|
)
|
|
parser.add_argument('--remoteOllamaServer',
|
|
help='The URL of the remote ollamaserver to use. ONLY USE THIS if you are using a local ollama server in an non-deault location or port')
|
|
parser.add_argument('--context', '-c',
|
|
help="Use Context file (context.md) to add context to your pattern", action="store_true")
|
|
|
|
args = parser.parse_args()
|
|
home_holder = os.path.expanduser("~")
|
|
config = os.path.join(home_holder, ".config", "fabric")
|
|
config_patterns_directory = os.path.join(config, "patterns")
|
|
config_context = os.path.join(config, "context.md")
|
|
env_file = os.path.join(config, ".env")
|
|
if not os.path.exists(config):
|
|
os.makedirs(config)
|
|
if args.setup:
|
|
Setup().run()
|
|
Alias().execute()
|
|
sys.exit()
|
|
if not os.path.exists(env_file) or not os.path.exists(config_patterns_directory):
|
|
print("Please run --setup to set up your API key and download patterns.")
|
|
sys.exit()
|
|
if not os.path.exists(config_patterns_directory):
|
|
Update()
|
|
Alias()
|
|
sys.exit()
|
|
if args.changeDefaultModel:
|
|
Setup().default_model(args.changeDefaultModel)
|
|
sys.exit()
|
|
if args.gui:
|
|
run_electron_app()
|
|
sys.exit()
|
|
if args.update:
|
|
Update()
|
|
Alias()
|
|
sys.exit()
|
|
if args.context:
|
|
if not os.path.exists(os.path.join(config, "context.md")):
|
|
print("Please create a context.md file in ~/.config/fabric")
|
|
sys.exit()
|
|
if args.agents:
|
|
standalone = Standalone(args)
|
|
text = "" # Initialize text variable
|
|
# Check if an argument was provided to --agents
|
|
if args.text:
|
|
text = args.text
|
|
else:
|
|
text = standalone.get_cli_input()
|
|
if text:
|
|
standalone = Standalone(args)
|
|
standalone.agents(text)
|
|
sys.exit()
|
|
if args.session:
|
|
from .helper import Session
|
|
session = Session()
|
|
if args.session == "default":
|
|
session_file = session.find_most_recent_file()
|
|
if session_file is None:
|
|
args.session = "default"
|
|
else:
|
|
args.session = session_file.split("/")[-1]
|
|
if args.clearsession:
|
|
from .helper import Session
|
|
session = Session()
|
|
session.clear_session(args.clearsession)
|
|
if args.clearsession == "all":
|
|
print(f"All sessions cleared")
|
|
else:
|
|
print(f"Session {args.clearsession} cleared")
|
|
sys.exit()
|
|
standalone = Standalone(args, args.pattern)
|
|
if args.list:
|
|
try:
|
|
direct = sorted(os.listdir(config_patterns_directory))
|
|
for d in direct:
|
|
print(d)
|
|
sys.exit()
|
|
except FileNotFoundError:
|
|
print("No patterns found")
|
|
sys.exit()
|
|
if args.listmodels:
|
|
gptmodels, localmodels, claudemodels = standalone.fetch_available_models()
|
|
print("GPT Models:")
|
|
for model in gptmodels:
|
|
print(model)
|
|
print("\nLocal Models:")
|
|
for model in localmodels:
|
|
print(model)
|
|
print("\nClaude Models:")
|
|
for model in claudemodels:
|
|
print(model)
|
|
sys.exit()
|
|
if args.text is not None:
|
|
text = args.text
|
|
else:
|
|
text = standalone.get_cli_input()
|
|
if args.stream and not args.context:
|
|
if args.remoteOllamaServer:
|
|
standalone.streamMessage(text, host=args.remoteOllamaServer)
|
|
else:
|
|
standalone.streamMessage(text)
|
|
sys.exit()
|
|
if args.stream and args.context:
|
|
with open(config_context, "r") as f:
|
|
context = f.read()
|
|
if args.remoteOllamaServer:
|
|
standalone.streamMessage(
|
|
text, context=context, host=args.remoteOllamaServer)
|
|
else:
|
|
standalone.streamMessage(text, context=context)
|
|
sys.exit()
|
|
elif args.context:
|
|
with open(config_context, "r") as f:
|
|
context = f.read()
|
|
if args.remoteOllamaServer:
|
|
standalone.sendMessage(
|
|
text, context=context, host=args.remoteOllamaServer)
|
|
else:
|
|
standalone.sendMessage(text, context=context)
|
|
sys.exit()
|
|
else:
|
|
if args.remoteOllamaServer:
|
|
standalone.sendMessage(text, host=args.remoteOllamaServer)
|
|
else:
|
|
standalone.sendMessage(text)
|
|
sys.exit()
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|