Open
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
This PR delivers Phase 11 Sprint 3 by adding the vLLM self-hosted adapter path on top of the shipped provider abstraction from
P11-S1and the local-adapter work fromP11-S2.What changed
vllmadapter and self-hosted helper wiring behind the existing provider registryPOST /v1/providers/vllm/registerandGET /v1/providers/{provider_id}/telemetryPOST /v1/providers/test,POST /v1/runtime/invoke,GET /v1/providers, andGET /v1/providers/{provider_id}working through the shipped normalized runtime seamP11-S3payloadUpgrade Overview
Protected Areas
Compatibility Impact
The
P11-S3changes are additive on top of the shipped provider abstraction. ExistingP11-S1OpenAI-compatible flows,P11-S2local-provider flows, andv0/responsesbehavior remain on the same normalized runtime seam; the new API surface is the additivePOST /v1/providers/vllm/registerandGET /v1/providers/{provider_id}/telemetrypath.Migration / Rollout
Apply the new
20260411_0054_phase11_vllm_telemetrymigration before using the vLLM telemetry path. Roll out by registering vLLM providers through the new endpoint and validating healthcheck, invoke, and telemetry behavior in one workspace before broader self-hosted adoption.Operator Action
Operators need to run the normal API migration flow, keep the vLLM service reachable at its configured base URL, and register the provider with only the bounded passthrough options supported by the adapter. No manual data backfill is required for existing providers.
Validation
Validation for this branch head is:
python3 scripts/check_control_doc_truth.pyPASS,./.venv/bin/python -m pytest tests/unit tests/integration -qPASS (1122 passed in 170.62s), andpnpm --dir apps/web testPASS (62 passedfiles,199 passedtests). The sprint also includes targeted runtime and telemetry coverage in the updated provider-runtime unit/integration tests and the new telemetry migration unit test.Rollback
Rollback is a standard application rollback plus reverting the
P11-S3commit and schema change if the vLLM path must be withdrawn. If rollout issues appear after registration, operators can stop using the registered vLLM provider path without affecting the shipped OpenAI-compatible, Ollama, or llama.cpp flows.Verification
python3 scripts/check_control_doc_truth.py./.venv/bin/python -m pytest tests/unit tests/integration -q1122 passed in 170.62s)pnpm --dir apps/web test62 passedfiles,199 passedtests)4.86sMerge Scope Notes
README.md,ARCHITECTURE.md, andPRODUCT_BRIEF.mdremain locally dirty and are explicitly excluded from this sprint merge scope.