Spaces:
Running
Running
Update app/providers.py
Browse files- app/providers.py +5 -4
app/providers.py
CHANGED
|
@@ -179,16 +179,17 @@ class OpenRouterProvider(BaseProvider):
|
|
| 179 |
|
| 180 |
|
| 181 |
class HuggingFaceProvider(BaseProvider):
|
| 182 |
-
"""HuggingFace Inference API —
|
| 183 |
|
| 184 |
-
|
| 185 |
-
|
|
|
|
| 186 |
"""
|
| 187 |
|
| 188 |
async def complete(self, prompt: str, model: str = None, max_tokens: int = 512) -> str:
|
| 189 |
m = model or self.model
|
| 190 |
data = await self._post(
|
| 191 |
-
f"{self.base_url}/
|
| 192 |
headers={
|
| 193 |
"Authorization": f"Bearer {self.key}",
|
| 194 |
"content-type": "application/json",
|
|
|
|
| 179 |
|
| 180 |
|
| 181 |
class HuggingFaceProvider(BaseProvider):
|
| 182 |
+
"""HuggingFace Inference API — OpenAI-compatible serverless endpoint.
|
| 183 |
|
| 184 |
+
base_url in .pyfun: https://api-inference.huggingface.co/v1
|
| 185 |
+
Model goes in payload, not in URL.
|
| 186 |
+
Free tier: max ~8B models. PRO required for 70B+.
|
| 187 |
"""
|
| 188 |
|
| 189 |
async def complete(self, prompt: str, model: str = None, max_tokens: int = 512) -> str:
|
| 190 |
m = model or self.model
|
| 191 |
data = await self._post(
|
| 192 |
+
f"{self.base_url}/chat/completions",
|
| 193 |
headers={
|
| 194 |
"Authorization": f"Bearer {self.key}",
|
| 195 |
"content-type": "application/json",
|