From 4988db080e991ebd5514816a81e0600994214876 Mon Sep 17 00:00:00 2001 From: Celina Hanouti Date: Wed, 5 Feb 2025 15:28:33 +0100 Subject: [PATCH] rename function --- .../inference/_providers/fal_ai.py | 7 ++----- .../inference/_providers/hf_inference.py | 15 +++++++-------- .../inference/_providers/replicate.py | 4 ++-- .../inference/_providers/sambanova.py | 4 ++-- .../inference/_providers/together.py | 4 ++-- 5 files changed, 15 insertions(+), 19 deletions(-) diff --git a/src/huggingface_hub/inference/_providers/fal_ai.py b/src/huggingface_hub/inference/_providers/fal_ai.py index fcb2187b1c..aebb0e6d1b 100644 --- a/src/huggingface_hub/inference/_providers/fal_ai.py +++ b/src/huggingface_hub/inference/_providers/fal_ai.py @@ -59,7 +59,7 @@ def prepare_request( raise ValueError( "You must provide an api_key to work with fal.ai API or log in with `huggingface-cli login`." ) - mapped_model = self.map_model(model) + mapped_model = self._map_model(model) headers = { **build_hf_headers(token=api_key), **headers, @@ -85,10 +85,7 @@ def prepare_request( headers=headers, ) - def map_model( - self, - model: Optional[str], - ) -> str: + def _map_model(self, model: Optional[str]) -> str: """Default implementation for mapping model HF model IDs to provider model IDs.""" if model is None: raise ValueError("Please provide a HF model ID supported by fal.ai.") diff --git a/src/huggingface_hub/inference/_providers/hf_inference.py b/src/huggingface_hub/inference/_providers/hf_inference.py index 4e7955d622..06e7713c3e 100644 --- a/src/huggingface_hub/inference/_providers/hf_inference.py +++ b/src/huggingface_hub/inference/_providers/hf_inference.py @@ -79,24 +79,23 @@ def prepare_request( ) -> RequestParameters: if extra_payload is None: extra_payload = {} - model = self.map_model(model) - url = self.build_url(model) - data, json = self._prepare_payload(inputs, parameters=parameters, model=model, extra_payload=extra_payload) + mapped_model = self._map_model(model) + url = self.build_url(mapped_model) + data, json = self._prepare_payload( + inputs, parameters=parameters, model=mapped_model, extra_payload=extra_payload + ) headers = self.prepare_headers(headers=headers, api_key=api_key) return RequestParameters( url=url, task=self.task, - model=model, + model=mapped_model, json=json, data=data, headers=headers, ) - def map_model( - self, - model: Optional[str], - ) -> str: + def _map_model(self, model: Optional[str]) -> str: return model if model is not None else get_recommended_model(self.task) def build_url(self, model: str) -> str: diff --git a/src/huggingface_hub/inference/_providers/replicate.py b/src/huggingface_hub/inference/_providers/replicate.py index 57d4e025f1..94ffb23593 100644 --- a/src/huggingface_hub/inference/_providers/replicate.py +++ b/src/huggingface_hub/inference/_providers/replicate.py @@ -71,7 +71,7 @@ def prepare_request( base_url = BASE_URL logger.info("Calling Replicate provider directly.") - mapped_model = self.map_model(model) + mapped_model = self._map_model(model) url = _build_url(base_url, mapped_model) headers = { @@ -91,7 +91,7 @@ def prepare_request( headers=headers, ) - def map_model(self, model: Optional[str]) -> str: + def _map_model(self, model: Optional[str]) -> str: """Default implementation for mapping model HF model IDs to provider model IDs.""" if model is None: raise ValueError("Please provide a HF model ID supported by Replicate.") diff --git a/src/huggingface_hub/inference/_providers/sambanova.py b/src/huggingface_hub/inference/_providers/sambanova.py index b212241668..960561ce05 100644 --- a/src/huggingface_hub/inference/_providers/sambanova.py +++ b/src/huggingface_hub/inference/_providers/sambanova.py @@ -63,7 +63,7 @@ def prepare_request( logger.info("Calling Sambanova provider directly.") headers = {**build_hf_headers(token=api_key), **headers} - mapped_model = self.map_model(model) + mapped_model = self._map_model(model) payload = { "messages": inputs, **{k: v for k, v in parameters.items() if v is not None}, @@ -79,7 +79,7 @@ def prepare_request( headers=headers, ) - def map_model(self, model: Optional[str]) -> str: + def _map_model(self, model: Optional[str]) -> str: """Default implementation for mapping model HF model IDs to provider model IDs.""" if model is None: raise ValueError("Please provide a HF model ID supported by Sambanova.") diff --git a/src/huggingface_hub/inference/_providers/together.py b/src/huggingface_hub/inference/_providers/together.py index 057190f6b8..33da61e801 100644 --- a/src/huggingface_hub/inference/_providers/together.py +++ b/src/huggingface_hub/inference/_providers/together.py @@ -105,7 +105,7 @@ def prepare_request( else: base_url = BASE_URL logger.info("Calling Together provider directly.") - mapped_model = self.map_model(model) + mapped_model = self._map_model(model) if "model" in parameters: parameters["model"] = mapped_model payload = self._prepare_payload(inputs, parameters=parameters) @@ -119,7 +119,7 @@ def prepare_request( headers=headers, ) - def map_model(self, model: Optional[str]) -> str: + def _map_model(self, model: Optional[str]) -> str: """Default implementation for mapping model HF model IDs to provider model IDs.""" if model is None: raise ValueError("Please provide a HF model ID supported by Together.")