From 0e07ae77cf190f90962e1f2032f94e2923d72941 Mon Sep 17 00:00:00 2001 From: "grandsonneo@gmail.com" Date: Fri, 21 Jul 2023 22:17:53 -0400 Subject: Use the temperature parameter --- g4f/Provider/Providers/AItianhu.py | 4 ++-- g4f/Provider/Providers/Ails.py | 2 +- g4f/Provider/Providers/ChatgptLogin.py | 2 +- g4f/Provider/Providers/DfeHub.py | 8 ++++---- g4f/Provider/Providers/EasyChat.py | 8 ++++---- g4f/Provider/Providers/Lockchat.py | 2 +- 6 files changed, 13 insertions(+), 13 deletions(-) (limited to 'g4f') diff --git a/g4f/Provider/Providers/AItianhu.py b/g4f/Provider/Providers/AItianhu.py index 5a151e1e..0bdaa09a 100644 --- a/g4f/Provider/Providers/AItianhu.py +++ b/g4f/Provider/Providers/AItianhu.py @@ -22,8 +22,8 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs): "prompt": base, "options": {}, "systemMessage": "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully. Respond using markdown.", - "temperature": 0.8, - "top_p": 1 + "temperature": kwargs.get("temperature", 0.8), + "top_p": kwargs.get("top_p", 1) } response = requests.post(url, headers=headers, json=data) if response.status_code == 200: diff --git a/g4f/Provider/Providers/Ails.py b/g4f/Provider/Providers/Ails.py index f6174f85..60d3603e 100644 --- a/g4f/Provider/Providers/Ails.py +++ b/g4f/Provider/Providers/Ails.py @@ -75,7 +75,7 @@ def _create_completion(model: str, messages: list, temperature: float = 0.6, str json_data = json.dumps(separators=(',', ':'), obj={ 'model': 'gpt-3.5-turbo', - 'temperature': 0.6, + 'temperature': temperature, 'stream': True, 'messages': messages} | sig) diff --git a/g4f/Provider/Providers/ChatgptLogin.py b/g4f/Provider/Providers/ChatgptLogin.py index 3659ee17..0fdbab8e 100644 --- a/g4f/Provider/Providers/ChatgptLogin.py +++ b/g4f/Provider/Providers/ChatgptLogin.py @@ -75,7 +75,7 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs): 'userName': '
User:
', 'aiName': '
AI:
', 'model': 'gpt-3.5-turbo', - 'temperature': 0.8, + 'temperature': kwargs.get('temperature', 0.8), 'maxTokens': 1024, 'maxResults': 1, 'apiKey': '', diff --git a/g4f/Provider/Providers/DfeHub.py b/g4f/Provider/Providers/DfeHub.py index afc9861a..e3ff8045 100644 --- a/g4f/Provider/Providers/DfeHub.py +++ b/g4f/Provider/Providers/DfeHub.py @@ -32,10 +32,10 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs): json_data = { 'messages': messages, 'model': 'gpt-3.5-turbo', - 'temperature': 0.5, - 'presence_penalty': 0, - 'frequency_penalty': 0, - 'top_p': 1, + 'temperature': kwargs.get('temperature', 0.5), + 'presence_penalty': kwargs.get('presence_penalty', 0), + 'frequency_penalty': kwargs.get('frequency_penalty', 0), + 'top_p': kwargs.get('top_p', 1), "stream": True, } response = requests.post('https://chat.dfehub.com/api/openai/v1/chat/completions', diff --git a/g4f/Provider/Providers/EasyChat.py b/g4f/Provider/Providers/EasyChat.py index a59ea072..909428fa 100644 --- a/g4f/Provider/Providers/EasyChat.py +++ b/g4f/Provider/Providers/EasyChat.py @@ -34,10 +34,10 @@ def _create_completion(model: str, messages: list, stream: bool, **kwargs): 'messages': messages, 'stream': True, 'model': model, - 'temperature': 0.5, - 'presence_penalty': 0, - 'frequency_penalty': 0, - 'top_p': 1, + 'temperature': kwargs.get('temperature', 0.5), + 'presence_penalty': kwargs.get('presence_penalty', 0), + 'frequency_penalty': kwargs.get('frequency_penalty', 0), + 'top_p': kwargs.get('top_p', 1), } response = requests.post('https://free.easychat.work/api/openai/v1/chat/completions', diff --git a/g4f/Provider/Providers/Lockchat.py b/g4f/Provider/Providers/Lockchat.py index 489356ce..dd1edb84 100644 --- a/g4f/Provider/Providers/Lockchat.py +++ b/g4f/Provider/Providers/Lockchat.py @@ -11,7 +11,7 @@ working = False def _create_completion(model: str, messages: list, stream: bool, temperature: float = 0.7, **kwargs): payload = { - "temperature": 0.7, + "temperature": temperature, "messages": messages, "model": model, "stream": True, -- cgit v1.2.3