From 5650442b0bbfc9e928b8c95e3f5138d8ed6356ed Mon Sep 17 00:00:00 2001 From: Kevin Hu Date: Mon, 5 Aug 2024 18:22:01 +0800 Subject: [PATCH] refine context length (#1813) ### What problem does this PR solve? #1594 ### Type of change - [x] Performance Improvement --- graphrag/index.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/graphrag/index.py b/graphrag/index.py index 6c0a96e8f1c..9b1fa9785b6 100644 --- a/graphrag/index.py +++ b/graphrag/index.py @@ -68,7 +68,7 @@ def build_knowlege_graph_chunks(tenant_id: str, chunks: List[str], callback, ent llm_bdl = LLMBundle(tenant_id, LLMType.CHAT) ext = GraphExtractor(llm_bdl) left_token_count = llm_bdl.max_length - ext.prompt_token_count - 1024 - left_token_count = llm_bdl.max_length * 0.4 + left_token_count = max(llm_bdl.max_length * 0.8, left_token_count) assert left_token_count > 0, f"The LLM context length({llm_bdl.max_length}) is smaller than prompt({ext.prompt_token_count})"