You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If you need the BFCL inference pipeline to spin up the vllm/sglang server to host the model, then you will need to implement a oss_model handler.
Don't worry about the naming of the proprietary_model handler vs oss_model handler; it's a bit confusing and will be addressed in #859. proprietary_model should mean API-based inference while oss_model means local-hosting-based inference.
If I deploy a proprietary model locally and use OSS model to access it, do I need to rewrite the file? Or can you reuse the proprietary_model Handler
The text was updated successfully, but these errors were encountered: