diff --git a/.semversioner/next-release/patch-20260421020145168929.json b/.semversioner/next-release/patch-20260421020145168929.json new file mode 100644 index 0000000000..8deda7c65b --- /dev/null +++ b/.semversioner/next-release/patch-20260421020145168929.json @@ -0,0 +1,4 @@ +{ + "type": "patch", + "description": "Add return type hints to generator functions in query/llm/text_utils.py" +} diff --git a/packages/graphrag/graphrag/query/llm/text_utils.py b/packages/graphrag/graphrag/query/llm/text_utils.py index d36aaed70f..bc39a984e7 100644 --- a/packages/graphrag/graphrag/query/llm/text_utils.py +++ b/packages/graphrag/graphrag/query/llm/text_utils.py @@ -17,7 +17,7 @@ logger = logging.getLogger(__name__) -def batched(iterable: Iterator, n: int): +def batched(iterable: Iterator, n: int) -> Iterator[tuple]: """ Batch data into tuples of length n. The last batch may be shorter. @@ -32,7 +32,7 @@ def batched(iterable: Iterator, n: int): yield batch -def chunk_text(text: str, max_tokens: int, tokenizer: Tokenizer | None = None): +def chunk_text(text: str, max_tokens: int, tokenizer: Tokenizer | None = None) -> Iterator[str]: """Chunk text by token length.""" if tokenizer is None: tokenizer = get_tokenizer()