Skip to content

Conversation

@gabrielmbmb
Copy link
Contributor

Description

We were passing the base_url using the model_id argument to the huggingface_hub.AsyncInferenceClient. This worked if using Inference Endpoint solutions, but if using a local TGI deployment, it didn't causing the chat_completion endpoint to return a 422.

@gabrielmbmb gabrielmbmb requested a review from plaguss August 23, 2024 09:49
@gabrielmbmb gabrielmbmb self-assigned this Aug 23, 2024
@github-actions
Copy link

Documentation for this PR has been built. You can view it at: https://distilabel.argilla.io/pr-924/

@codspeed-hq
Copy link

codspeed-hq bot commented Aug 23, 2024

CodSpeed Performance Report

Merging #924 will not alter performance

Comparing fix-inference-endpoints (85b66cd) with main (c76d4a7)

Summary

✅ 1 untouched benchmarks

@gabrielmbmb gabrielmbmb merged commit 379c756 into main Aug 23, 2024
@gabrielmbmb gabrielmbmb deleted the fix-inference-endpoints branch August 23, 2024 12:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

3 participants