feat: import order

feature/stdout-breaker
adldotori 1 year ago
parent 8df3214024
commit 9eefef4019

@ -1,36 +1,29 @@
from typing import Dict, List, TypedDict
import re
import uvicorn
from typing import Dict, List, TypedDict
import torch
import uvicorn
from fastapi import FastAPI
from fastapi.staticfiles import StaticFiles
from pydantic import BaseModel
from env import settings
from core.prompts.error import ERROR_PROMPT
from core.agents.manager import AgentManager
from core.handlers.base import BaseHandler, FileHandler, FileType
from core.handlers.dataframe import CsvToDataframe
from core.handlers.image import ImageCaptioning
from core.prompts.error import ERROR_PROMPT
from core.tools.base import BaseToolSet
from core.tools.terminal import Terminal
from core.tools.cpu import ExitConversation, RequestsGet, WineDB
from core.tools.editor import CodeEditor
from core.tools.cpu import (
RequestsGet,
WineDB,
ExitConversation,
)
from core.tools.gpu import (
ImageEditing,
InstructPix2Pix,
Text2Image,
VisualQuestionAnswering,
)
from core.handlers.base import BaseHandler, FileHandler, FileType
from core.handlers.image import ImageCaptioning
from core.handlers.dataframe import CsvToDataframe
from core.tools.terminal import Terminal
from core.upload import StaticUploader
from env import settings
from logger import logger
app = FastAPI()

@ -1,14 +1,13 @@
from langchain.chat_models.base import BaseChatModel
from langchain.output_parsers.base import BaseOutputParser
from env import settings
from core.prompts.input import EVAL_PREFIX, EVAL_SUFFIX
from core.tools.base import BaseToolSet
from core.tools.factory import ToolsFactory
from env import settings
from .llm import ChatOpenAI
from .chat_agent import ConversationalChatAgent
from .llm import ChatOpenAI
from .parser import EvalOutputParser

@ -3,8 +3,8 @@ from typing import Any, Dict, List, Optional, Union
from langchain.callbacks.base import BaseCallbackHandler
from langchain.schema import AgentAction, AgentFinish, LLMResult
from logger import logger
from ansi import ANSI, Color, Style
from logger import logger
class EVALCallbackHandler(BaseCallbackHandler):

@ -5,15 +5,6 @@ import logging
import sys
from typing import Any, Callable, Dict, List, Mapping, Optional, Tuple
from pydantic import BaseModel, Extra, Field, root_validator
from tenacity import (
before_sleep_log,
retry,
retry_if_exception_type,
stop_after_attempt,
wait_exponential,
)
from langchain.chat_models.base import BaseChatModel
from langchain.schema import (
AIMessage,
@ -25,6 +16,14 @@ from langchain.schema import (
SystemMessage,
)
from langchain.utils import get_from_dict_or_env
from pydantic import BaseModel, Extra, Field, root_validator
from tenacity import (
before_sleep_log,
retry,
retry_if_exception_type,
stop_after_attempt,
wait_exponential,
)
from logger import logger
@ -128,7 +127,7 @@ class ChatOpenAI(BaseChatModel, BaseModel):
"""Whether to stream the results or not."""
n: int = 1
"""Number of chat completions to generate for each prompt."""
max_tokens: int = 256
max_tokens: int = 2048
"""Maximum number of tokens to generate."""
class Config:
@ -226,7 +225,6 @@ class ChatOpenAI(BaseChatModel, BaseModel):
def _generate(
self, messages: List[BaseMessage], stop: Optional[List[str]] = None
) -> ChatResult:
message_dicts, params = self._create_message_dicts(messages, stop)
logger.debug("Messages:\n")
for item in message_dicts:

@ -1,18 +1,17 @@
from typing import Dict
from langchain.agents.tools import BaseTool
from langchain.agents.agent import Agent, AgentExecutor
from langchain.agents.tools import BaseTool
from langchain.callbacks import set_handler
from langchain.callbacks.base import CallbackManager
from langchain.chains.conversation.memory import ConversationBufferMemory
from langchain.memory.chat_memory import BaseChatMemory
from langchain.callbacks.base import CallbackManager
from langchain.callbacks import set_handler
from core.tools.base import BaseToolSet
from core.tools.factory import ToolsFactory
from .callback import EVALCallbackHandler
from .builder import AgentBuilder
from .callback import EVALCallbackHandler
callback_manager = CallbackManager([EVALCallbackHandler()])
set_handler(EVALCallbackHandler())

@ -1,8 +1,9 @@
import os
import uuid
import requests
from typing import Dict
from enum import Enum
from typing import Dict
import requests
class FileType(Enum):

@ -1,9 +1,6 @@
import torch
from PIL import Image
from transformers import (
BlipProcessor,
BlipForConditionalGeneration,
)
from transformers import BlipForConditionalGeneration, BlipProcessor
from core.prompts.file import IMAGE_PROMPT

@ -1,8 +1,8 @@
from typing import Callable, Tuple
from enum import Enum
from typing import Callable, Tuple
from langchain.agents.tools import Tool, BaseTool
from langchain.agents.agent import AgentExecutor
from langchain.agents.tools import BaseTool, Tool
class ToolScope(Enum):

@ -1,15 +1,13 @@
from env import settings
import requests
from llama_index.readers.database import DatabaseReader
from llama_index import GPTSimpleVectorIndex
from bs4 import BeautifulSoup
from llama_index import GPTSimpleVectorIndex
from llama_index.readers.database import DatabaseReader
from .base import tool, BaseToolSet, ToolScope, SessionGetter
from env import settings
from logger import logger
from .base import BaseToolSet, SessionGetter, ToolScope, tool
class RequestsGet(BaseToolSet):
@tool(

@ -1,4 +1,4 @@
from core.tools.base import tool, BaseToolSet
from core.tools.base import BaseToolSet, tool
from logger import logger
from .patch import CodePatcher

@ -56,8 +56,12 @@ test.py|7,5|9,13|news_titles = []
test.py|11,16|11,16|_titles
"""
import os
from pathlib import Path
from typing import Tuple
from env import settings
class Position:
separator = ","
@ -76,7 +80,7 @@ class PatchCommand:
separator = "|"
def __init__(self, filepath: str, start: Position, end: Position, content: str):
self.filepath: str = filepath
self.filepath: str = str(Path(settings["PLAYGROUND_DIR"]) / Path(filepath))
self.start: Position = start
self.end: Position = end
self.content: str = content
@ -92,6 +96,13 @@ class PatchCommand:
return sum([len(line) for line in lines])
def execute(self) -> Tuple[int, int]:
# make sure the directory exists
if not str(Path(self.filepath).resolve()).startswith(
str(Path(settings["PLAYGROUND_DIR"]).resolve())
):
return "You can't write file outside of current directory."
os.makedirs(os.path.dirname(self.filepath), exist_ok=True)
lines = self.read_lines()
before = sum([len(line) for line in lines])

@ -3,7 +3,7 @@ read protocol:
<filepath>|<start line>-<end line>
"""
from typing import Optional, List, Tuple
from typing import List, Optional, Tuple
class Line:

@ -6,6 +6,7 @@ write protocol:
"""
import os
from pathlib import Path
from env import settings

@ -1,4 +1,5 @@
from typing import Optional
from langchain.agents import load_tools
from langchain.agents.tools import BaseTool
from langchain.llms.base import BaseLLM

@ -1,29 +1,26 @@
import os
import uuid
import numpy as np
import torch
from PIL import Image
from transformers import (
CLIPSegProcessor,
CLIPSegForImageSegmentation,
)
from transformers import (
BlipProcessor,
BlipForQuestionAnswering,
)
from diffusers import (
StableDiffusionPipeline,
EulerAncestralDiscreteScheduler,
StableDiffusionInpaintPipeline,
StableDiffusionInstructPix2PixPipeline,
StableDiffusionPipeline,
)
from PIL import Image
from transformers import (
BlipForQuestionAnswering,
BlipProcessor,
CLIPSegForImageSegmentation,
CLIPSegProcessor,
)
from diffusers import EulerAncestralDiscreteScheduler
from utils import get_new_image_name
from logger import logger
from utils import get_new_image_name
from .base import tool, BaseToolSet
from .base import BaseToolSet, tool
class MaskFormer(BaseToolSet):

@ -1,12 +1,11 @@
import subprocess
from typing import Dict, List
from tempfile import TemporaryFile
from typing import Dict, List
from core.tools.base import BaseToolSet, SessionGetter, ToolScope, tool
from core.tools.terminal.syscall import SyscallTracer
from env import settings
from logger import logger
from core.tools.base import tool, BaseToolSet, ToolScope, SessionGetter
from core.tools.terminal.syscall import SyscallTracer
class Terminal(BaseToolSet):
@ -17,9 +16,8 @@ class Terminal(BaseToolSet):
name="Terminal",
description="Executes commands in a terminal."
"If linux errno occurs, we have to solve the problem with the terminal. "
"It can't execute interactive operations or blocking operations. "
"Input should be valid commands, "
"and the output will be any output from running that command.",
"Input must be one valid command. "
"Output will be any output from running that command.",
scope=ToolScope.SESSION,
)
def execute(self, commands: str, get_session: SessionGetter) -> str:
@ -45,9 +43,6 @@ class Terminal(BaseToolSet):
except Exception as e:
output = str(e)
if len(output) > 1000:
output = output[:1000] + "..."
logger.debug(
f"\nProcessed Terminal, Input Commands: {commands} "
f"Output Answer: {output}"

@ -1,16 +1,16 @@
from typing import Tuple, Optional
import signal
from typing import Optional, Tuple
from ptrace.debugger import (
PtraceDebugger,
PtraceProcess,
ProcessExit,
ProcessSignal,
NewProcessEvent,
ProcessExecution,
ProcessExit,
ProcessSignal,
PtraceDebugger,
PtraceProcess,
)
from ptrace.syscall import PtraceSyscall
from ptrace.func_call import FunctionCallOptions
from ptrace.syscall import PtraceSyscall
from ptrace.tools import signal_to_exitcode
@ -61,7 +61,7 @@ class SyscallTracer:
break
try:
self.wait_syscall_with_timeout(5)
self.wait_syscall_with_timeout(60)
except ProcessExit as event:
if event.exitcode is not None:
exitcode = event.exitcode

@ -1,7 +1,9 @@
import os
import boto3
from env import DotEnv
from .base import AbstractUploader

@ -2,6 +2,7 @@ import os
import shutil
from env import DotEnv
from .base import AbstractUploader

@ -10,7 +10,7 @@ services:
volumes:
- ./static/:/app/static/
ports:
- "3000:3000"
- "4500:4500"
- "7000:7000"
- "8000:8000" # eval port
env_file:
@ -26,7 +26,7 @@ services:
- ../.cache/huggingface/:/root/.cache/huggingface/
- ./static/:/app/static/
ports:
- "3000:3000"
- "4500:4500"
- "7000:7000"
- "8000:8000" # eval port
env_file:

@ -1,4 +1,5 @@
import logging
from env import settings
logger = logging.getLogger()

@ -1,9 +1,9 @@
import os
import random
import torch
import uuid
import numpy as np
import numpy as np
import torch
os.makedirs("image", exist_ok=True)
os.makedirs("audio", exist_ok=True)

Loading…
Cancel
Save