From c6ac6d70011df17ed982105697469dbc19bb72c1 Mon Sep 17 00:00:00 2001 From: Paul Swingle Date: Mon, 22 Apr 2024 12:59:41 -0700 Subject: [PATCH] fix tests --- mentat/code_context.py | 1 - tests/conftest.py | 4 ++-- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/mentat/code_context.py b/mentat/code_context.py index 67c017659..7bfc58844 100644 --- a/mentat/code_context.py +++ b/mentat/code_context.py @@ -59,7 +59,6 @@ def __init__( self.include_files: Dict[Path, List[CodeFeature]] = {} self.ignore_files: Set[Path] = set() - @api_guard async def refresh_daemon(self): """Call before interacting with context to ensure daemon is up to date.""" diff --git a/tests/conftest.py b/tests/conftest.py index 089295956..234e661f7 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -98,7 +98,7 @@ def mock_call_llm_api(mocker): completion_mock = mocker.patch.object(LlmApiHandler, "call_llm_api") def wrap_unstreamed_string(value): - return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], False), value, 1, 0, 0, True, 1, None) + return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], False), value, 1, 0, 0, True, 1) def wrap_streamed_strings(values): class MockStreamingSpiceResponse: @@ -115,7 +115,7 @@ async def __anext__(self): return values[self.cur_value - 1] def current_response(self): - return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], True), "".join(values), 1, 0, 0, True, 1, None) + return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], True), "".join(values), 1, 0, 0, True, 1) mock_spice_response = MockStreamingSpiceResponse() return mock_spice_response