From ab096d37ffefb0cfbbae5335dd5020980a88ab53 Mon Sep 17 00:00:00 2001 From: Sourcery AI <> Date: Thu, 5 Oct 2023 16:17:17 +0000 Subject: [PATCH] 'Refactored by Sourcery' --- chatgpt/ChatGPT_Flask_Demo_v2/ChatGPT_Flask_Demo/__init__.py | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/chatgpt/ChatGPT_Flask_Demo_v2/ChatGPT_Flask_Demo/__init__.py b/chatgpt/ChatGPT_Flask_Demo_v2/ChatGPT_Flask_Demo/__init__.py index d0fa16f..f495a81 100644 --- a/chatgpt/ChatGPT_Flask_Demo_v2/ChatGPT_Flask_Demo/__init__.py +++ b/chatgpt/ChatGPT_Flask_Demo_v2/ChatGPT_Flask_Demo/__init__.py @@ -3,14 +3,13 @@ import openai import tiktoken -openai.api_key = os.getenv("OPENAI_API_KEY") +openai.api_key = os.getenv("OPENAI_API_KEY") openai.organization = os.getenv("OPENAI_ORGANIZATION") system_message = {"role": "system", "content": "You are a helpful assistant."} max_response_tokens = 250 token_limit= 4096 -conversation=[] -conversation.append(system_message) +conversation = [system_message] def num_tokens_from_messages(messages, model="gpt-3.5-turbo-0301"): encoding = tiktoken.encoding_for_model(model)