~ | improve compatibility with lower python versions

remove tiktoken
pull/1272/head
abc 7 months ago
parent 79761c6e42
commit 2345588d38

@ -4,7 +4,7 @@ import logging
from fastapi import FastAPI, Response, Request
from fastapi.responses import StreamingResponse
from typing import List, Union, Any, Dict, AnyStr
from ._tokenizer import tokenize
#from ._tokenizer import tokenize
from .. import BaseProvider
import time
@ -95,8 +95,8 @@ class Api:
completion_timestamp = int(time.time())
if not stream:
prompt_tokens, _ = tokenize(''.join([message['content'] for message in messages]))
completion_tokens, _ = tokenize(response)
#prompt_tokens, _ = tokenize(''.join([message['content'] for message in messages]))
#completion_tokens, _ = tokenize(response)
json_data = {
'id': f'chatcmpl-{completion_id}',
@ -114,9 +114,9 @@ class Api:
}
],
'usage': {
'prompt_tokens': prompt_tokens,
'completion_tokens': completion_tokens,
'total_tokens': prompt_tokens + completion_tokens,
'prompt_tokens': 0, #prompt_tokens,
'completion_tokens': 0, #completion_tokens,
'total_tokens': 0, #prompt_tokens + completion_tokens,
},
}

@ -1,9 +1,9 @@
import tiktoken
from typing import Union
# import tiktoken
# from typing import Union
def tokenize(text: str, model: str = 'gpt-3.5-turbo') -> Union[int, str]:
encoding = tiktoken.encoding_for_model(model)
encoded = encoding.encode(text)
num_tokens = len(encoded)
# def tokenize(text: str, model: str = 'gpt-3.5-turbo') -> Union[int, str]:
# encoding = tiktoken.encoding_for_model(model)
# encoded = encoding.encode(text)
# num_tokens = len(encoded)
return num_tokens, encoded
# return num_tokens, encoded
Loading…
Cancel
Save