From 72aa10023b04d4de1dab7de88938513b452003e6 Mon Sep 17 00:00:00 2001 From: James Braza Date: Mon, 4 Mar 2024 16:10:55 -0800 Subject: [PATCH] Expanded explanatory comment on Answer.token_counts (#240) --- paperqa/types.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/paperqa/types.py b/paperqa/types.py index 8d051db8..d1ab001a 100644 --- a/paperqa/types.py +++ b/paperqa/types.py @@ -177,7 +177,8 @@ class Answer(BaseModel): answer_length: str = "about 100 words" # just for convenience you can override this cost: float | None = None - # key is model name, value is (prompt, completion) token counts + # Map model name to a two-item list of LLM prompt token counts + # and LLM completion token counts token_counts: dict[str, list[int]] = Field(default_factory=dict) model_config = ConfigDict(extra="ignore")