mirror of
https://github.com/microsoft/graphrag.git
synced 2026-01-14 09:07:20 +08:00
Fix INIT_YAML embeddings default settings (#1039)
Co-authored-by: Thanh Long Phan <long.phan@dida.do> Co-authored-by: Alonso Guevara <alonsog@microsoft.com>
This commit is contained in:
parent
22df2f80d0
commit
1b51827c66
@ -0,0 +1,4 @@
|
||||
{
|
||||
"type": "patch",
|
||||
"description": "Fix default settings for embedding"
|
||||
}
|
||||
@ -38,6 +38,8 @@ embeddings:
|
||||
## parallelization: override the global parallelization settings for embeddings
|
||||
async_mode: {defs.ASYNC_MODE.value} # or asyncio
|
||||
# target: {defs.EMBEDDING_TARGET.value} # or all
|
||||
# batch_size: {defs.EMBEDDING_BATCH_SIZE} # the number of documents to send in a single request
|
||||
# batch_max_tokens: {defs.EMBEDDING_BATCH_MAX_TOKENS} # the maximum number of tokens to send in a single request
|
||||
llm:
|
||||
api_key: ${{GRAPHRAG_API_KEY}}
|
||||
type: {defs.EMBEDDING_TYPE.value} # or azure_openai_embedding
|
||||
@ -52,8 +54,6 @@ embeddings:
|
||||
# max_retry_wait: {defs.LLM_MAX_RETRY_WAIT}
|
||||
# sleep_on_rate_limit_recommendation: true # whether to sleep when azure suggests wait-times
|
||||
# concurrent_requests: {defs.LLM_CONCURRENT_REQUESTS} # the number of parallel inflight requests that may be made
|
||||
# batch_size: {defs.EMBEDDING_BATCH_SIZE} # the number of documents to send in a single request
|
||||
# batch_max_tokens: {defs.EMBEDDING_BATCH_MAX_TOKENS} # the maximum number of tokens to send in a single request
|
||||
|
||||
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user