From 3b97e69ba48189c2550230077d415ede861202a9 Mon Sep 17 00:00:00 2001 From: repsac-by Date: Thu, 1 Aug 2024 22:56:30 +0300 Subject: Update the models of PerplexityLabs provider --- g4f/Provider/PerplexityLabs.py | 15 ++++++++++++--- 1 file changed, 12 insertions(+), 3 deletions(-) diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py index 0a298e55..7f4587e1 100644 --- a/g4f/Provider/PerplexityLabs.py +++ b/g4f/Provider/PerplexityLabs.py @@ -11,11 +11,20 @@ API_URL = "https://www.perplexity.ai/socket.io/" WS_URL = "wss://www.perplexity.ai/socket.io/" class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin): - url = "https://labs.perplexity.ai" + url = "https://labs.perplexity.ai" working = True default_model = "mixtral-8x7b-instruct" models = [ - "llama-3-sonar-large-32k-online", "llama-3-sonar-small-32k-online", "llama-3-sonar-large-32k-chat", "llama-3-sonar-small-32k-chat", "llama-3-8b-instruct", "llama-3-70b-instruct", "gemma-2-9b-it", "gemma-2-27b-it", "nemotron-4-340b-instruct", "mixtral-8x7b-instruct", + "llama-3.1-sonar-large-128k-online", + "llama-3.1-sonar-small-128k-online", + "llama-3.1-sonar-large-128k-chat", + "llama-3.1-sonar-small-128k-chat", + "llama-3.1-8b-instruct", + "llama-3.1-70b-instruct", + "gemma-2-9b-it", + "gemma-2-27b-it", + "nemotron-4-340b-instruct", + "mixtral-8x7b-instruct" ] @classmethod @@ -54,7 +63,7 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin): data=post_data ) as response: await raise_for_status(response) - assert await response.text() == "OK" + assert await response.text() == "OK" async with session.ws_connect(f"{WS_URL}?EIO=4&transport=websocket&sid={sid}", autoping=False) as ws: await ws.send_str("2probe") assert(await ws.receive_str() == "3probe") -- cgit v1.2.3