You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: cagent-schema.json
+38-2Lines changed: 38 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -725,12 +725,13 @@
725
725
"description": "Retrieval strategy type",
726
726
"enum": [
727
727
"bm25",
728
-
"chunked-embeddings"
728
+
"chunked-embeddings",
729
+
"semantic-embeddings"
729
730
]
730
731
},
731
732
"embedding_model": {
732
733
"type": "string",
733
-
"description": "Embedding model reference for chunked-embeddings strategies (looked up in models map, or 'auto' for automatic selection)",
734
+
"description": "Embedding model reference for chunked-embeddings and semantic-embeddings strategies (looked up in models map, or 'auto' for automatic selection)",
734
735
"examples": [
735
736
"openai/text-embedding-3-small",
736
737
"dmr/embeddinggemma",
@@ -811,6 +812,41 @@
811
812
}
812
813
},
813
814
"additionalProperties": false
815
+
},
816
+
"embedding_batch_size": {
817
+
"type": "integer",
818
+
"description": "Number of text chunks to send to the embedding API in a single request (chunked-embeddings/semantic-embeddings only)",
819
+
"minimum": 1,
820
+
"default": 50
821
+
},
822
+
"max_embedding_concurrency": {
823
+
"type": "integer",
824
+
"description": "Maximum concurrent embedding batch API requests. For semantic-embeddings, also controls parallel LLM calls for generating chunk summaries.",
825
+
"minimum": 1,
826
+
"default": 3
827
+
},
828
+
"max_indexing_concurrency": {
829
+
"type": "integer",
830
+
"description": "Maximum number of files to index in parallel during initialization",
831
+
"minimum": 1,
832
+
"default": 3
833
+
},
834
+
"chat_model": {
835
+
"type": "string",
836
+
"description": "Chat model used to generate semantic representations for each chunk (semantic-embeddings only, required)",
837
+
"examples": [
838
+
"anthropic/claude-sonnet-4-5",
839
+
"openai/gpt-4o-mini"
840
+
]
841
+
},
842
+
"semantic_prompt": {
843
+
"type": "string",
844
+
"description": "Custom prompt template for semantic LLM. Uses JavaScript template literal syntax with the following placeholders: ${path} (full source file path), ${basename} (base name of file), ${chunk_index} (numeric chunk index), ${content} (raw chunk content), ${ast_context} (AST metadata when ast_context is enabled). Only applicable to semantic-embeddings strategy."
845
+
},
846
+
"ast_context": {
847
+
"type": "boolean",
848
+
"description": "Include TreeSitter-derived AST metadata in the semantic prompt (semantic-embeddings only, requires chunking.code_aware for best results)",
0 commit comments