-
Notifications
You must be signed in to change notification settings - Fork 3.4k
Open
Labels
bugSomething isn't workingSomething isn't workingtriageDefault label assignment, indicates new issue needs reviewed by a maintainerDefault label assignment, indicates new issue needs reviewed by a maintainer
Description
Do you need to file an issue?
- I have searched the existing issues and this bug is not already filed.
- My model is hosted on OpenAI or Azure. If not, please look at the "model providers" issue and don't file a new one here.
- I believe this is a legitimate bug, not just a question. If this is a question, please use the Discussions area.
Describe the bug
When indexing, i get the following error message:
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71525/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71526/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71527/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71528/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71529/106718
2026-03-21 06:03:32.0180 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71530/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71531/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71532/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71533/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71534/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71535/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71536/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71537/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71538/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71539/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71540/106718
2026-03-21 06:03:32.0181 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71541/106718
2026-03-21 06:03:32.0182 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71542/106718
2026-03-21 06:03:32.0303 - ERROR - graphrag_llm.middleware.with_logging - Async request failed with exception=litellm.BadRequestError: OpenAIException - We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 929, in acompletion
headers, response = await self.make_openai_chat_completion_request(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/logging_utils.py", line 297, in async_wrapper
result = await func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 460, in make_openai_chat_completion_request
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 437, in make_openai_chat_completion_request
await openai_aclient.chat.completions.with_raw_response.create(
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_legacy_response.py", line 384, in wrapped
return cast(LegacyAPIResponse[R], await func(*args, **kwargs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 2714, in create
return await self._post(
^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1884, in post
return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1669, in request
raise self._make_status_error_from_response(err.response) from None
openai.BadRequestError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 620, in acompletion
response = await init_response
^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 989, in acompletion
raise OpenAIError(
litellm.llms.openai.common_utils.OpenAIError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_logging.py", line 64, in _request_count_middleware_async
return await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_request_count.py", line 72, in _request_count_middleware_async
result = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_cache.py", line 145, in _cache_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_retries.py", line 55, in _retry_middleware_async
return await retrier.retry_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/retry/exponential_retry.py", line 102, in retry_async
return await func(**input_args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_metrics.py", line 80, in _metrics_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/completion/lite_llm_completion.py", line 290, in _base_completion_async
response = await litellm.acompletion(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 2072, in wrapper_async
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 1892, in wrapper_async
result = await original_function(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 639, in acompletion
raise exception_type(
^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 2456, in exception_type
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 478, in exception_type
raise BadRequestError(
litellm.exceptions.BadRequestError: litellm.BadRequestError: OpenAIException - We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)
2026-03-21 06:03:32.0324 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71543/106718
2026-03-21 06:03:32.0325 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71544/106718
2026-03-21 06:03:32.0715 - ERROR - graphrag.index.run.run_pipeline - error running workflow extract_graph
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 929, in acompletion
headers, response = await self.make_openai_chat_completion_request(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/logging_utils.py", line 297, in async_wrapper
result = await func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 460, in make_openai_chat_completion_request
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 437, in make_openai_chat_completion_request
await openai_aclient.chat.completions.with_raw_response.create(
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_legacy_response.py", line 384, in wrapped
return cast(LegacyAPIResponse[R], await func(*args, **kwargs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/resources/chat/completions/completions.py", line 2714, in create
return await self._post(
^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1884, in post
return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/openai/_base_client.py", line 1669, in request
raise self._make_status_error_from_response(err.response) from None
openai.BadRequestError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 620, in acompletion
response = await init_response
^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/llms/openai/openai.py", line 989, in acompletion
raise OpenAIError(
litellm.llms.openai.common_utils.OpenAIError: Error code: 400 - {'error': {'message': "We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)", 'type': 'invalid_request_error', 'param': None, 'code': None}}
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/run/run_pipeline.py", line 135, in _run_pipeline
result = await workflow_function(config, context)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/workflows/extract_graph.py", line 61, in run_workflow
entities, relationships, raw_entities, raw_relationships = await extract_graph(
^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/workflows/extract_graph.py", line 143, in extract_graph
entities, relationships = await get_summarized_entities_relationships(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/workflows/extract_graph.py", line 168, in get_summarized_entities_relationships
entity_summaries, relationship_summaries = await summarize_descriptions(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 115, in summarize_descriptions
return await get_summarized(entities_df, relationships_df, semaphore)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 80, in get_summarized
edge_results = await asyncio.gather(*edge_futures)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/.local/share/uv/python/cpython-3.12.13-linux-x86_64-gnu/lib/python3.12/asyncio/tasks.py", line 385, in __wakeup
future.result()
File "/home/dobin/.local/share/uv/python/cpython-3.12.13-linux-x86_64-gnu/lib/python3.12/asyncio/tasks.py", line 314, in __step_run_and_handle_result
result = coro.send(None)
^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 102, in do_summarize_descriptions
results = await run_summarize_descriptions(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/summarize_descriptions.py", line 139, in run_summarize_descriptions
result = await extractor(id=id, descriptions=descriptions)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/description_summary_extractor.py", line 68, in __call__
result = await self._summarize_descriptions(id, descriptions)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/description_summary_extractor.py", line 105, in _summarize_descriptions
result = await self._summarize_descriptions_with_llm(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag/index/operations/summarize_descriptions/description_summary_extractor.py", line 124, in _summarize_descriptions_with_llm
response: LLMCompletionResponse = await self._model.completion_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/completion/lite_llm_completion.py", line 193, in completion_async
response = await self._completion_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_logging.py", line 64, in _request_count_middleware_async
return await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_request_count.py", line 72, in _request_count_middleware_async
result = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_cache.py", line 145, in _cache_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_retries.py", line 55, in _retry_middleware_async
return await retrier.retry_async(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/retry/exponential_retry.py", line 102, in retry_async
return await func(**input_args)
^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/middleware/with_metrics.py", line 80, in _metrics_middleware_async
response = await async_middleware(**kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/graphrag_llm/completion/lite_llm_completion.py", line 290, in _base_completion_async
response = await litellm.acompletion(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 2072, in wrapper_async
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/utils.py", line 1892, in wrapper_async
result = await original_function(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/main.py", line 639, in acompletion
raise exception_type(
^^^^^^^^^^^^^^^
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 2456, in exception_type
raise e
File "/home/dobin/repos/bookmarkRag/.venv/lib/python3.12/site-packages/litellm/litellm_core_utils/exception_mapping_utils.py", line 478, in exception_type
raise BadRequestError(
litellm.exceptions.BadRequestError: litellm.BadRequestError: OpenAIException - We could not parse the JSON body of your request. (HINT: This likely means you aren't using your HTTP library correctly. The OpenAI API expects a JSON payload, but what was sent was not valid JSON. If you have trouble figuring out how to fix this, please contact us through our help center at help.openai.com.)
2026-03-21 06:03:32.0737 - ERROR - graphrag.api.index - Workflow extract_graph completed with errors
2026-03-21 06:03:32.0740 - INFO - graphrag.logger.progress - Summarize entity/relationship description progress: 71545/106718
2026-03-21 06:03:32.0744 - INFO - graphrag_llm.metrics.log_metrics_writer - Metrics for openai/text-embedding-3-small: {
"attempted_request_count": 1,
"successful_response_count": 1,
"failed_response_count": 0,
"failure_rate": 0.0,
"requests_with_retries": 0,
"retries": 0,
"retry_rate": 0.0,
"compute_duration_seconds": 2.2788126468658447,
"compute_duration_per_response_seconds": 2.2788126468658447,
"cache_hit_rate": 0.0,
"streaming_responses": 0,
"responses_with_tokens": 1,
"prompt_tokens": 9,
"total_tokens": 9,
"tokens_per_response": 9.0,
"cost_per_response": 0.0
}
2026-03-21 06:03:32.0745 - INFO - graphrag_llm.metrics.log_metrics_writer - Metrics for openai/gpt-5.4-mini: {
"attempted_request_count": 21831,
"successful_response_count": 21829,
"failed_response_count": 2,
"failure_rate": 9.161284412074573e-05,
"requests_with_retries": 0,
"retries": 0,
"retry_rate": 0.0,
"compute_duration_seconds": 48129.54413795471,
"compute_duration_per_response_seconds": 2.204844204404907,
"cached_responses": 6,
"cache_hit_rate": 0.0002748637134087682,
"streaming_responses": 0,
"responses_with_tokens": 21829,
"prompt_tokens": 22409866,
"completion_tokens": 8838469,
"total_tokens": 31248335,
"tokens_per_response": 1431.5055659901966,
"cost_per_response": 0.0
}
Steps to reproduce
graphrag init
git clone https://github.com/dobin/AwesomeMalDevLinks
mkdir input/
mv AwesomeMalDevLinks/out/maldev/*.md input/maldev
graphrag index
Expected Behavior
No error
GraphRAG Config Used
I changed model, and file pattern:
completion_models:
default_completion_model:
model_provider: openai
model: gpt-5.4-mini
auth_method: api_key # or azure_managed_identity
api_key: ${GRAPHRAG_API_KEY} # set this in the generated .env file, or remove if managed identity
retry:
type: exponential_backoff
embedding_models:
default_embedding_model:
model_provider: openai
model: text-embedding-3-small
auth_method: api_key
api_key: ${GRAPHRAG_API_KEY}
retry:
type: exponential_backoff
input:
type: text # [csv, text, json, jsonl]
file_pattern: ".*\\.md$$"Logs and screenshots
No response
Additional Information
- GraphRAG Version: 3.0.6
- Operating System: ubuntu wsl 22.04
- Python Version: 3.12.13
- Related Issues:
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't workingtriageDefault label assignment, indicates new issue needs reviewed by a maintainerDefault label assignment, indicates new issue needs reviewed by a maintainer