mirror of
https://github.com/microsoft/graphrag.git
synced 2026-01-14 09:07:20 +08:00
* Refactor config
- Add new ModelConfig to represent LLM settings
- Combines LLMParameters, ParallelizationParameters, encoding_model, and async_mode
- Add top level models config that is a list of available LLM ModelConfigs
- Remove LLMConfig inheritance and delete LLMConfig
- Replace the inheritance with a model_id reference to the ModelConfig listed in the top level models config
- Remove all fallbacks and hydration logic from create_graphrag_config
- This removes the automatic env variable overrides
- Support env variables within config files using Templating
- This requires "$" to be escaped with extra "$" so ".*\\.txt$" becomes ".*\\.txt$$"
- Update init content to initialize new config file with the ModelConfig structure
* Use dict of ModelConfig instead of list
* Add model validations and unit tests
* Fix ruff checks
* Add semversioner change
* Fix unit tests
* validate root_dir in pydantic model
* Rename ModelConfig to LanguageModelConfig
* Rename ModelConfigMissingError to LanguageModelConfigMissingError
* Add validationg for unexpected API keys
* Allow skipping pydantic validation for testing/mocking purposes.
* Add default lm configs to verb tests
* smoke test
* remove config from flows to fix llm arg mapping
* Fix embedding llm arg mapping
* Remove timestamp from smoke test outputs
* Remove unused "subworkflows" smoke test properties
* Add models to smoke test configs
* Update smoke test output path
* Send logs to logs folder
* Fix output path
* Fix csv test file pattern
* Update placeholder
* Format
* Instantiate default model configs
* Fix unit tests for config defaults
* Fix migration notebook
* Remove create_pipeline_config
* Remove several unused config models
* Remove indexing embedding and input configs
* Move embeddings function to config
* Remove skip_workflows
* Remove skip embeddings in favor of explicit naming
* fix unit test spelling mistake
* self.models[model_id] is already a language model. Remove redundant casting.
* update validation errors to instruct users to rerun graphrag init
* instantiate LanguageModelConfigs with validation
* skip validation in unit tests
* update verb tests to use default model settings instead of skipping validation
* test using llm settings
* cleanup verb tests
* remove unsafe default model config
* remove the ability to skip pydantic validation
* remove None union types when default values are set
* move vector_store from embeddings to top level of config and delete resolve_paths
* update vector store settings
* fix vector store and smoke tests
* fix serializing vector_store settings
* fix vector_store usage
* fix vector_store type
* support cli overrides for loading graphrag config
* rename storage to output
* Add --force flag to init
* Remove run_id and resume, fix Drift config assignment
* Ruff
---------
Co-authored-by: Nathan Evans <github@talkswithnumbers.com>
Co-authored-by: Alonso Guevara <alonsog@microsoft.com>
84 lines
2.5 KiB
Python
84 lines
2.5 KiB
Python
# Copyright (c) 2024 Microsoft Corporation.
|
|
# Licensed under the MIT License
|
|
|
|
|
|
from graphrag.callbacks.noop_workflow_callbacks import NoopWorkflowCallbacks
|
|
from graphrag.config.create_graphrag_config import create_graphrag_config
|
|
from graphrag.config.enums import LLMType
|
|
from graphrag.index.operations.summarize_communities.community_reports_extractor.community_reports_extractor import (
|
|
CommunityReportResponse,
|
|
FindingModel,
|
|
)
|
|
from graphrag.index.workflows.create_final_community_reports import (
|
|
run_workflow,
|
|
workflow_name,
|
|
)
|
|
from graphrag.utils.storage import load_table_from_storage
|
|
|
|
from .util import (
|
|
DEFAULT_MODEL_CONFIG,
|
|
compare_outputs,
|
|
create_test_context,
|
|
load_test_table,
|
|
)
|
|
|
|
MOCK_RESPONSES = [
|
|
CommunityReportResponse(
|
|
title="<report_title>",
|
|
summary="<executive_summary>",
|
|
rating=2,
|
|
rating_explanation="<rating_explanation>",
|
|
findings=[
|
|
FindingModel(
|
|
summary="<insight_1_summary>", explanation="<insight_1_explanation"
|
|
),
|
|
FindingModel(
|
|
summary="<insight_2_summary>", explanation="<insight_2_explanation"
|
|
),
|
|
],
|
|
)
|
|
]
|
|
|
|
|
|
async def test_create_final_community_reports():
|
|
expected = load_test_table(workflow_name)
|
|
|
|
context = await create_test_context(
|
|
storage=[
|
|
"create_final_nodes",
|
|
"create_final_covariates",
|
|
"create_final_relationships",
|
|
"create_final_entities",
|
|
"create_final_communities",
|
|
]
|
|
)
|
|
|
|
config = create_graphrag_config({"models": DEFAULT_MODEL_CONFIG})
|
|
llm_settings = config.get_language_model_config(
|
|
config.community_reports.model_id
|
|
).model_dump()
|
|
llm_settings["type"] = LLMType.StaticResponse
|
|
llm_settings["responses"] = MOCK_RESPONSES
|
|
llm_settings["parse_json"] = True
|
|
config.community_reports.strategy = {
|
|
"type": "graph_intelligence",
|
|
"llm": llm_settings,
|
|
}
|
|
|
|
await run_workflow(
|
|
config,
|
|
context,
|
|
NoopWorkflowCallbacks(),
|
|
)
|
|
|
|
actual = await load_table_from_storage(workflow_name, context.storage)
|
|
|
|
assert len(actual.columns) == len(expected.columns)
|
|
|
|
# only assert a couple of columns that are not mock - most of this table is LLM-generated
|
|
compare_outputs(actual, expected, columns=["community", "level"])
|
|
|
|
# assert a handful of mock data items to confirm they get put in the right spot
|
|
assert actual["rank"][:1][0] == 2
|
|
assert actual["rank_explanation"][:1][0] == "<rating_explanation>"
|