123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475 |
- ---
- title: '🔍 Query configurations'
- ---
- ## AppConfig
- | option | description | type | default |
- |-----------|-----------------------|---------------------------------|------------------------|
- | log_level | log level | string | WARNING |
- | embedding_fn| embedding function | chromadb.utils.embedding_functions | \{text-embedding-ada-002\} |
- | db | vector database (experimental) | BaseVectorDB | ChromaDB |
- | collection_name | initial collection name for the database | string | embedchain_store |
- | collect_metrics | collect anonymous telemetry data to improve embedchain | boolean | true |
- ## AddConfig
- |option|description|type|default|
- |---|---|---|---|
- |chunker|chunker config|ChunkerConfig|Default values for chunker depends on the `data_type`. Please refer [ChunkerConfig](#chunker-config)|
- |loader|loader config|LoaderConfig|None|
- Yes, you are passing `ChunkerConfig` to `AddConfig`, like so:
- ```python
- chunker_config = ChunkerConfig(chunk_size=100)
- add_config = AddConfig(chunker=chunker_config)
- app.add_local("text", "lorem ipsum", config=add_config)
- ```
- ### ChunkerConfig
- |option|description|type|default|
- |---|---|---|---|
- |chunk_size|Maximum size of chunks to return|int|Default value for various `data_type` mentioned below|
- |chunk_overlap|Overlap in characters between chunks|int|Default value for various `data_type` mentioned below|
- |length_function|Function that measures the length of given chunks|typing.Callable|Default value for various `data_type` mentioned below|
- Default values of chunker config parameters for different `data_type`:
- |data_type|chunk_size|chunk_overlap|length_function|
- |---|---|---|---|
- |docx|1000|0|len|
- |text|300|0|len|
- |qna_pair|300|0|len|
- |web_page|500|0|len|
- |pdf_file|1000|0|len|
- |youtube_video|2000|0|len|
- |docs_site|500|50|len|
- |notion|300|0|len|
- ### LoaderConfig
- _coming soon_
- ## QueryConfig
- |option|description|type|default|
- |---|---|---|---|
- |number_documents|Absolute number of documents to pull from the database as context.|int|1
- |template|custom template for prompt. If history is used with query, $history has to be included as well.|Template|Template("Use the following pieces of context to answer the query at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer. \$context Query: \$query Helpful Answer:")|
- |model|name of the model used.|string|depends on app type|
- |temperature|Controls the randomness of the model's output. Higher values (closer to 1) make output more random, lower values make it more deterministic.|float|0|
- |max_tokens|Controls how many tokens are used. Exact implementation (whether it counts prompt and/or response) depends on the model.|int|1000|
- |top_p|Controls the diversity of words. Higher values (closer to 1) make word selection more diverse, lower values make words less diverse.|float|1|
- |history|include conversation history from your client or database.|any (recommendation: list[str])|None|
- |stream|control if response is streamed back to the user.|bool|False|
- ## ChatConfig
- All options for query and...
- _coming soon_
- `history` is not supported, as that is handled is handled automatically, the config option is not supported.
|