Skip to main content

inline::file-search

Description

File search tool runtime for document ingestion, chunking, and semantic search.

Configuration

FieldTypeRequiredDefaultDescription
vector_stores_configVectorStoresConfigNodefault_provider_id=None default_embedding_model=None default_reranker_model=None rewrite_query_params=None file_search_params=FileSearchParams(header_template='file_search tool found {num_chunks} chunks:\nBEGIN of file_search tool results.\n', footer_template='END of file_search tool results.\n') context_prompt_params=ContextPromptParams(chunk_annotation_template='Result {index}\nContent: {chunk.content}\nMetadata: {metadata}\n', context_template='The above results were retrieved to help answer the user's query: "{query}". Use them as supporting information only in answering this query. {annotation_instruction}\n') annotation_prompt_params=AnnotationPromptParams(enable_annotations=True, annotation_instruction_template="Cite sources immediately at the end of sentences before punctuation, using `<file-id
vector_stores_config.default_provider_idstr | NoneNoID of the vector_io provider to use as default when multiple providers are available and none is specified.
vector_stores_config.default_embedding_modelQualifiedModel | NoneNoDefault embedding model configuration for vector stores.
vector_stores_config.default_embedding_model.provider_idstrNo
vector_stores_config.default_embedding_model.model_idstrNo
vector_stores_config.default_embedding_model.embedding_dimensionsint | NoneNo
vector_stores_config.default_reranker_modelRerankerModel | NoneNoDefault reranker model configuration for vector stores.
vector_stores_config.default_reranker_model.provider_idstrNo
vector_stores_config.default_reranker_model.model_idstrNo
vector_stores_config.rewrite_query_paramsRewriteQueryParams | NoneNoParameters for query rewriting/expansion. None disables query rewriting.
vector_stores_config.rewrite_query_params.modelQualifiedModel | NoneNoLLM model for query rewriting/expansion in vector search.
vector_stores_config.rewrite_query_params.promptstrNoExpand this query with relevant synonyms and related terms. Return only the improved query, no explanations:

{query}

Improved query: | Prompt template for query rewriting. Use {query} as placeholder for the original query. | | vector_stores_config.rewrite_query_params.max_tokens | int | No | 100 | Maximum number of tokens for query expansion responses. | | vector_stores_config.rewrite_query_params.temperature | float | No | 0.3 | Temperature for query expansion model (0.0 = deterministic, 1.0 = creative). | | vector_stores_config.file_search_params | FileSearchParams | No | header_template='file_search tool found {num_chunks} chunks:\nBEGIN of file_search tool results.\n' footer_template='END of file_search tool results.\n' | Configuration for file search tool output formatting. | | vector_stores_config.file_search_params.header_template | str | No | file_search tool found {num_chunks} chunks: BEGIN of file_search tool results. | Template for the header text shown before search results. Available placeholders: {num_chunks} number of chunks found. | | vector_stores_config.file_search_params.footer_template | str | No | END of file_search tool results. | Template for the footer text shown after search results. | | vector_stores_config.context_prompt_params | ContextPromptParams | No | chunk_annotation_template='Result {index}\nContent: {chunk.content}\nMetadata: {metadata}\n' context_template='The above results were retrieved to help answer the user's query: "{query}". Use them as supporting information only in answering this query. {annotation_instruction}\n' | Configuration for LLM prompt content and chunk formatting. | | vector_stores_config.context_prompt_params.chunk_annotation_template | str | No | Result {index} Content: {chunk.content} Metadata: {metadata} | Template for formatting individual chunks in search results. Available placeholders: {index} 1-based chunk index, {chunk.content} chunk content, {metadata} chunk metadata dict. | | vector_stores_config.context_prompt_params.context_template | str | No | The above results were retrieved to help answer the user's query: "{query}". Use them as supporting information only in answering this query. {annotation_instruction} | Template for explaining the search results to the model. Available placeholders: {query} user's query, {num_chunks} number of chunks. | | vector_stores_config.annotation_prompt_params | AnnotationPromptParams | No | enable_annotations=True annotation_instruction_template="Cite sources immediately at the end of sentences before punctuation, using &lt;|file-id|&gt; format like 'This is a fact <|file-Cn3MSNn72ENTiiq11Qda4A|>.'. Do not add extra punctuation. Use only the file IDs provided, do not invent new ones." chunk_annotation_template='[{index}] {metadata_text} cite as <|{file_id}|>\n{chunk_text}\n' | Configuration for source annotation and attribution features. | | vector_stores_config.annotation_prompt_params.enable_annotations | bool | No | True | Whether to include annotation information in results. | | vector_stores_config.annotation_prompt_params.annotation_instruction_template | str | No | Cite sources immediately at the end of sentences before punctuation, using &lt;|file-id|&gt; format like 'This is a fact <|file-Cn3MSNn72ENTiiq11Qda4A|>.'. Do not add extra punctuation. Use only the file IDs provided, do not invent new ones. | Instructions for how the model should cite sources. Used when enable_annotations is True. | | vector_stores_config.annotation_prompt_params.chunk_annotation_template | str | No | [{index}] {metadata_text} cite as <|{file_id}|> {chunk_text} | Template for chunks with annotation information. Available placeholders: {index} 1-based chunk index, {metadata_text} formatted metadata, {file_id} document identifier, {chunk_text} chunk content. | | vector_stores_config.file_ingestion_params | FileIngestionParams | No | default_chunk_size_tokens=512 default_chunk_overlap_tokens=128 | Configuration for file processing during ingestion. | | vector_stores_config.file_ingestion_params.default_chunk_size_tokens | int | No | 512 | Default chunk size for RAG tool operations when not specified | | vector_stores_config.file_ingestion_params.default_chunk_overlap_tokens | int | No | 128 | Default overlap in tokens between chunks (original default: 512 // 4 = 128) | | vector_stores_config.chunk_retrieval_params | ChunkRetrievalParams | No | chunk_multiplier=5 max_tokens_in_context=4000 default_reranker_strategy='rrf' rrf_impact_factor=60.0 weighted_search_alpha=0.5 | Configuration for chunk retrieval and ranking during search. | | vector_stores_config.chunk_retrieval_params.chunk_multiplier | int | No | 5 | Multiplier for OpenAI API over-retrieval (affects all providers) | | vector_stores_config.chunk_retrieval_params.max_tokens_in_context | int | No | 4000 | Maximum tokens allowed in RAG context before truncation | | vector_stores_config.chunk_retrieval_params.default_reranker_strategy | str | No | rrf | Default reranker when not specified: 'rrf', 'weighted', or 'normalized' | | vector_stores_config.chunk_retrieval_params.rrf_impact_factor | float | No | 60.0 | Impact factor for RRF (Reciprocal Rank Fusion) reranking | | vector_stores_config.chunk_retrieval_params.weighted_search_alpha | float | No | 0.5 | Alpha weight for weighted search reranking (0.0-1.0) | | vector_stores_config.file_batch_params | FileBatchParams | No | max_concurrent_files_per_batch=3 file_batch_chunk_size=10 cleanup_interval_seconds=86400 | Configuration for file batch processing. | | vector_stores_config.file_batch_params.max_concurrent_files_per_batch | int | No | 3 | Maximum files processed concurrently in file batches | | vector_stores_config.file_batch_params.file_batch_chunk_size | int | No | 10 | Number of files to process in each batch chunk | | vector_stores_config.file_batch_params.cleanup_interval_seconds | int | No | 86400 | Interval for cleaning up expired file batches (seconds) | | vector_stores_config.contextual_retrieval_params | ContextualRetrievalParams | No | model=None default_timeout_seconds=120 default_max_concurrency=3 max_document_tokens=100000 | Configuration for contextual retrieval during file ingestion. | | vector_stores_config.contextual_retrieval_params.model | QualifiedModel \| None | No | | Default LLM model for contextual retrieval. Used when model_id is not specified in chunking strategy. | | vector_stores_config.contextual_retrieval_params.default_timeout_seconds | int | No | 120 | Default timeout in seconds for each LLM contextualization call. | | vector_stores_config.contextual_retrieval_params.default_max_concurrency | int | No | 3 | Default maximum concurrent LLM calls for contextualization. | | vector_stores_config.contextual_retrieval_params.max_document_tokens | int | No | 100000 | Maximum document size in tokens. Documents exceeding this are rejected for contextual retrieval. |

Sample Configuration

{}