From f797ec828ce3b5965796b437878e14b6eec6235b Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Sat, 16 Mar 2024 20:02:15 +0100 Subject: Fix GigaChat exceptions, Raise missing browser error --- g4f/Provider/HuggingFace.py | 10 ++++------ 1 file changed, 4 insertions(+), 6 deletions(-) (limited to 'g4f/Provider/HuggingFace.py') diff --git a/g4f/Provider/HuggingFace.py b/g4f/Provider/HuggingFace.py index a73411ce..647780fd 100644 --- a/g4f/Provider/HuggingFace.py +++ b/g4f/Provider/HuggingFace.py @@ -7,6 +7,7 @@ from ..typing import AsyncResult, Messages from .base_provider import AsyncGeneratorProvider, ProviderModelMixin from .helper import get_connector from ..errors import RateLimitError, ModelNotFoundError +from ..requests.raise_for_status import raise_for_status class HuggingFace(AsyncGeneratorProvider, ProviderModelMixin): url = "https://huggingface.co/chat" @@ -44,12 +45,9 @@ class HuggingFace(AsyncGeneratorProvider, ProviderModelMixin): connector=get_connector(connector, proxy) ) as session: async with session.post(f"{api_base.rstrip('/')}/models/{model}", json=payload) as response: - if response.status == 429: - raise RateLimitError("Rate limit reached. Set a api_key") - elif response.status == 404: + if response.status == 404: raise ModelNotFoundError(f"Model is not supported: {model}") - elif response.status != 200: - raise RuntimeError(f"Response {response.status}: {await response.text()}") + await raise_for_status(response) if stream: first = True async for line in response.content: @@ -68,7 +66,7 @@ def format_prompt(messages: Messages) -> str: system_messages = [message["content"] for message in messages if message["role"] == "system"] question = " ".join([messages[-1]["content"], *system_messages]) history = "".join([ - f"[INST]{messages[idx-1]['content']} [/INST] {message}" + f"[INST]{messages[idx-1]['content']} [/INST] {message['content']}" for idx, message in enumerate(messages) if message["role"] == "assistant" ]) -- cgit v1.2.3