|
@@ -55,7 +55,7 @@ def num_tokens_from_messages(messages, model):
|
|
|
except KeyError:
|
|
except KeyError:
|
|
|
logger.debug("Warning: model not found. Using cl100k_base encoding.")
|
|
logger.debug("Warning: model not found. Using cl100k_base encoding.")
|
|
|
encoding = tiktoken.get_encoding("cl100k_base")
|
|
encoding = tiktoken.get_encoding("cl100k_base")
|
|
|
- if model == "gpt-3.5-turbo":
|
|
|
|
|
|
|
+ if model == "gpt-3.5-turbo" or model == "gpt-35-turbo":
|
|
|
return num_tokens_from_messages(messages, model="gpt-3.5-turbo-0301")
|
|
return num_tokens_from_messages(messages, model="gpt-3.5-turbo-0301")
|
|
|
elif model == "gpt-4":
|
|
elif model == "gpt-4":
|
|
|
return num_tokens_from_messages(messages, model="gpt-4-0314")
|
|
return num_tokens_from_messages(messages, model="gpt-4-0314")
|
|
@@ -76,4 +76,4 @@ def num_tokens_from_messages(messages, model):
|
|
|
if key == "name":
|
|
if key == "name":
|
|
|
num_tokens += tokens_per_name
|
|
num_tokens += tokens_per_name
|
|
|
num_tokens += 3 # every reply is primed with <|start|>assistant<|message|>
|
|
num_tokens += 3 # every reply is primed with <|start|>assistant<|message|>
|
|
|
- return num_tokens
|
|
|
|
|
|
|
+ return num_tokens
|