From 45b7f9577c0c3d6697fa21856ee77125c7fb3453 Mon Sep 17 00:00:00 2001 From: shamoon <4887959+shamoon@users.noreply.github.com> Date: Wed, 14 Jan 2026 13:13:05 -0800 Subject: [PATCH] Fix embedding mock --- src/paperless_ai/tests/test_chat.py | 10 ++++------ 1 file changed, 4 insertions(+), 6 deletions(-) diff --git a/src/paperless_ai/tests/test_chat.py b/src/paperless_ai/tests/test_chat.py index c91488cf1..9f5d5198f 100644 --- a/src/paperless_ai/tests/test_chat.py +++ b/src/paperless_ai/tests/test_chat.py @@ -11,14 +11,12 @@ from paperless_ai.chat import stream_chat_with_documents @pytest.fixture(autouse=True) def patch_embed_model(): from llama_index.core import settings as llama_settings + from llama_index.core.embeddings.utils import MockEmbedding - mock_embed_model = MagicMock() - mock_embed_model._get_text_embedding_batch.return_value = [ - [0.1] * 1536, - ] # 1 vector per input - llama_settings.Settings._embed_model = mock_embed_model + mock_embed_model = MockEmbedding(embed_dim=8) + llama_settings.Settings.embed_model = mock_embed_model yield - llama_settings.Settings._embed_model = None + llama_settings.Settings.embed_model = None @pytest.fixture(autouse=True)