inline::file-search
Description
File search tool runtime for document ingestion, chunking, and semantic search.
Configuration
| Field | Type | Required | Default | Description |
|---|---|---|---|---|
vector_stores_config | VectorStoresConfig | No | default_provider_id=None default_embedding_model=None default_reranker_model=None rewrite_query_params=None file_search_params=FileSearchParams(header_template='file_search tool found {num_chunks} chunks:\nBEGIN of file_search tool results.\n', footer_template='END of file_search tool results.\n') context_prompt_params=ContextPromptParams(chunk_annotation_template='Result {index}\nContent: {chunk.content}\nMetadata: {metadata}\n', context_template='The above results were retrieved to help answer the user's query: "{query}". Use them as supporting information only in answering this query. {annotation_instruction}\n') annotation_prompt_params=AnnotationPromptParams(enable_annotations=True, annotation_instruction_template="Cite sources immediately at the end of sentences before punctuation, using `< | file-id |
vector_stores_config.default_provider_id | str | None | No | ID of the vector_io provider to use as default when multiple providers are available and none is specified. | |
vector_stores_config.default_embedding_model | QualifiedModel | None | No | Default embedding model configuration for vector stores. | |
vector_stores_config.default_embedding_model.provider_id | str | No | ||
vector_stores_config.default_embedding_model.model_id | str | No | ||
vector_stores_config.default_embedding_model.embedding_dimensions | int | None | No | ||
vector_stores_config.default_reranker_model | RerankerModel | None | No | Default reranker model configuration for vector stores. | |
vector_stores_config.default_reranker_model.provider_id | str | No | ||
vector_stores_config.default_reranker_model.model_id | str | No | ||
vector_stores_config.rewrite_query_params | RewriteQueryParams | None | No | Parameters for query rewriting/expansion. None disables query rewriting. | |
vector_stores_config.rewrite_query_params.model | QualifiedModel | None | No | LLM model for query rewriting/expansion in vector search. | |
vector_stores_config.rewrite_query_params.prompt | str | No | Expand this query with relevant synonyms and related terms. Return only the improved query, no explanations: |
{query}
Improved query: | Prompt template for query rewriting. Use {query} as placeholder for the original query. |
| vector_stores_config.rewrite_query_params.max_tokens | int | No | 100 | Maximum number of tokens for query expansion responses. |
| vector_stores_config.rewrite_query_params.temperature | float | No | 0.3 | Temperature for query expansion model (0.0 = deterministic, 1.0 = creative). |
| vector_stores_config.file_search_params | FileSearchParams | No | header_template='file_search tool found {num_chunks} chunks:\nBEGIN of file_search tool results.\n' footer_template='END of file_search tool results.\n' | Configuration for file search tool output formatting. |
| vector_stores_config.file_search_params.header_template | str | No | file_search tool found {num_chunks} chunks:
BEGIN of file_search tool results.
| Template for the header text shown before search results. Available placeholders: {num_chunks} number of chunks found. |
| vector_stores_config.file_search_params.footer_template | str | No | END of file_search tool results.
| Template for the footer text shown after search results. |
| vector_stores_config.context_prompt_params | ContextPromptParams | No | chunk_annotation_template='Result {index}\nContent: {chunk.content}\nMetadata: {metadata}\n' context_template='The above results were retrieved to help answer the user's query: "{query}". Use them as supporting information only in answering this query. {annotation_instruction}\n' | Configuration for LLM prompt content and chunk formatting. |
| vector_stores_config.context_prompt_params.chunk_annotation_template | str | No | Result {index}
Content: {chunk.content}
Metadata: {metadata}
| Template for formatting individual chunks in search results. Available placeholders: {index} 1-based chunk index, {chunk.content} chunk content, {metadata} chunk metadata dict. |
| vector_stores_config.context_prompt_params.context_template | str | No | The above results were retrieved to help answer the user's query: "{query}". Use them as supporting information only in answering this query. {annotation_instruction}
| Template for explaining the search results to the model. Available placeholders: {query} user's query, {num_chunks} number of chunks. |
| vector_stores_config.annotation_prompt_params | AnnotationPromptParams | No | enable_annotations=True annotation_instruction_template="Cite sources immediately at the end of sentences before punctuation, using <|file-id|> format like 'This is a fact <|file-Cn3MSNn72ENTiiq11Qda4A|>.'. Do not add extra punctuation. Use only the file IDs provided, do not invent new ones." chunk_annotation_template='[{index}] {metadata_text} cite as <|{file_id}|>\n{chunk_text}\n' | Configuration for source annotation and attribution features. |
| vector_stores_config.annotation_prompt_params.enable_annotations | bool | No | True | Whether to include annotation information in results. |
| vector_stores_config.annotation_prompt_params.annotation_instruction_template | str | No | Cite sources immediately at the end of sentences before punctuation, using <|file-id|> format like 'This is a fact <|file-Cn3MSNn72ENTiiq11Qda4A|>.'. Do not add extra punctuation. Use only the file IDs provided, do not invent new ones. | Instructions for how the model should cite sources. Used when enable_annotations is True. |
| vector_stores_config.annotation_prompt_params.chunk_annotation_template | str | No | [{index}] {metadata_text} cite as <|{file_id}|>
{chunk_text}
| Template for chunks with annotation information. Available placeholders: {index} 1-based chunk index, {metadata_text} formatted metadata, {file_id} document identifier, {chunk_text} chunk content. |
| vector_stores_config.file_ingestion_params | FileIngestionParams | No | default_chunk_size_tokens=512 default_chunk_overlap_tokens=128 | Configuration for file processing during ingestion. |
| vector_stores_config.file_ingestion_params.default_chunk_size_tokens | int | No | 512 | Default chunk size for RAG tool operations when not specified |
| vector_stores_config.file_ingestion_params.default_chunk_overlap_tokens | int | No | 128 | Default overlap in tokens between chunks (original default: 512 // 4 = 128) |
| vector_stores_config.chunk_retrieval_params | ChunkRetrievalParams | No | chunk_multiplier=5 max_tokens_in_context=4000 default_reranker_strategy='rrf' rrf_impact_factor=60.0 weighted_search_alpha=0.5 | Configuration for chunk retrieval and ranking during search. |
| vector_stores_config.chunk_retrieval_params.chunk_multiplier | int | No | 5 | Multiplier for OpenAI API over-retrieval (affects all providers) |
| vector_stores_config.chunk_retrieval_params.max_tokens_in_context | int | No | 4000 | Maximum tokens allowed in RAG context before truncation |
| vector_stores_config.chunk_retrieval_params.default_reranker_strategy | str | No | rrf | Default reranker when not specified: 'rrf', 'weighted', or 'normalized' |
| vector_stores_config.chunk_retrieval_params.rrf_impact_factor | float | No | 60.0 | Impact factor for RRF (Reciprocal Rank Fusion) reranking |
| vector_stores_config.chunk_retrieval_params.weighted_search_alpha | float | No | 0.5 | Alpha weight for weighted search reranking (0.0-1.0) |
| vector_stores_config.file_batch_params | FileBatchParams | No | max_concurrent_files_per_batch=3 file_batch_chunk_size=10 cleanup_interval_seconds=86400 | Configuration for file batch processing. |
| vector_stores_config.file_batch_params.max_concurrent_files_per_batch | int | No | 3 | Maximum files processed concurrently in file batches |
| vector_stores_config.file_batch_params.file_batch_chunk_size | int | No | 10 | Number of files to process in each batch chunk |
| vector_stores_config.file_batch_params.cleanup_interval_seconds | int | No | 86400 | Interval for cleaning up expired file batches (seconds) |
| vector_stores_config.contextual_retrieval_params | ContextualRetrievalParams | No | model=None default_timeout_seconds=120 default_max_concurrency=3 max_document_tokens=100000 | Configuration for contextual retrieval during file ingestion. |
| vector_stores_config.contextual_retrieval_params.model | QualifiedModel \| None | No | | Default LLM model for contextual retrieval. Used when model_id is not specified in chunking strategy. |
| vector_stores_config.contextual_retrieval_params.default_timeout_seconds | int | No | 120 | Default timeout in seconds for each LLM contextualization call. |
| vector_stores_config.contextual_retrieval_params.default_max_concurrency | int | No | 3 | Default maximum concurrent LLM calls for contextualization. |
| vector_stores_config.contextual_retrieval_params.max_document_tokens | int | No | 100000 | Maximum document size in tokens. Documents exceeding this are rejected for contextual retrieval. |
Sample Configuration
{}