diff --git a/mentat/code_context.py b/mentat/code_context.py index 67c017659..7bfc58844 100644 --- a/mentat/code_context.py +++ b/mentat/code_context.py @@ -59,7 +59,6 @@ def __init__( self.include_files: Dict[Path, List[CodeFeature]] = {} self.ignore_files: Set[Path] = set() - @api_guard async def refresh_daemon(self): """Call before interacting with context to ensure daemon is up to date.""" diff --git a/tests/conftest.py b/tests/conftest.py index 089295956..234e661f7 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -98,7 +98,7 @@ def mock_call_llm_api(mocker): completion_mock = mocker.patch.object(LlmApiHandler, "call_llm_api") def wrap_unstreamed_string(value): - return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], False), value, 1, 0, 0, True, 1, None) + return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], False), value, 1, 0, 0, True, 1) def wrap_streamed_strings(values): class MockStreamingSpiceResponse: @@ -115,7 +115,7 @@ async def __anext__(self): return values[self.cur_value - 1] def current_response(self): - return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], True), "".join(values), 1, 0, 0, True, 1, None) + return SpiceResponse(SpiceCallArgs("gpt-3.5-turbo", [], True), "".join(values), 1, 0, 0, True, 1) mock_spice_response = MockStreamingSpiceResponse() return mock_spice_response