mirror of
https://github.com/MODSetter/SurfSense.git
synced 2026-04-25 16:56:22 +02:00
chore: ran linting
This commit is contained in:
parent
74826b3714
commit
04691d572b
61 changed files with 1962 additions and 1516 deletions
|
|
@ -13,7 +13,9 @@ _EMBEDDING_DIM = app_config.embedding_model_instance.dimension
|
|||
pytestmark = pytest.mark.integration
|
||||
|
||||
|
||||
def _onedrive_doc(*, unique_id: str, search_space_id: int, connector_id: int, user_id: str) -> ConnectorDocument:
|
||||
def _onedrive_doc(
|
||||
*, unique_id: str, search_space_id: int, connector_id: int, user_id: str
|
||||
) -> ConnectorDocument:
|
||||
return ConnectorDocument(
|
||||
title=f"File {unique_id}.docx",
|
||||
source_markdown=f"## Document\n\nContent from {unique_id}",
|
||||
|
|
@ -32,7 +34,9 @@ def _onedrive_doc(*, unique_id: str, search_space_id: int, connector_id: int, us
|
|||
)
|
||||
|
||||
|
||||
@pytest.mark.usefixtures("patched_summarize", "patched_embed_texts", "patched_chunk_text")
|
||||
@pytest.mark.usefixtures(
|
||||
"patched_summarize", "patched_embed_texts", "patched_chunk_text"
|
||||
)
|
||||
async def test_onedrive_pipeline_creates_ready_document(
|
||||
db_session, db_search_space, db_connector, db_user, mocker
|
||||
):
|
||||
|
|
@ -61,7 +65,9 @@ async def test_onedrive_pipeline_creates_ready_document(
|
|||
assert DocumentStatus.is_state(row.status, DocumentStatus.READY)
|
||||
|
||||
|
||||
@pytest.mark.usefixtures("patched_summarize", "patched_embed_texts", "patched_chunk_text")
|
||||
@pytest.mark.usefixtures(
|
||||
"patched_summarize", "patched_embed_texts", "patched_chunk_text"
|
||||
)
|
||||
async def test_onedrive_duplicate_content_skipped(
|
||||
db_session, db_search_space, db_connector, db_user, mocker
|
||||
):
|
||||
|
|
@ -87,8 +93,6 @@ async def test_onedrive_duplicate_content_skipped(
|
|||
)
|
||||
first_doc = result.scalars().first()
|
||||
assert first_doc is not None
|
||||
first_id = first_doc.id
|
||||
|
||||
doc2 = _onedrive_doc(
|
||||
unique_id="od-dup-file",
|
||||
search_space_id=space_id,
|
||||
|
|
@ -97,4 +101,6 @@ async def test_onedrive_duplicate_content_skipped(
|
|||
)
|
||||
|
||||
prepared2 = await service.prepare_for_indexing([doc2])
|
||||
assert len(prepared2) == 0 or (len(prepared2) == 1 and prepared2[0].existing_document is not None)
|
||||
assert len(prepared2) == 0 or (
|
||||
len(prepared2) == 1 and prepared2[0].existing_document is not None
|
||||
)
|
||||
|
|
|
|||
|
|
@ -48,12 +48,14 @@ def patch_extract(monkeypatch):
|
|||
mock,
|
||||
)
|
||||
return mock
|
||||
|
||||
return _patch
|
||||
|
||||
|
||||
# Slice 1: Tracer bullet
|
||||
async def test_single_file_returns_one_connector_document(
|
||||
mock_onedrive_client, patch_extract,
|
||||
mock_onedrive_client,
|
||||
patch_extract,
|
||||
):
|
||||
patch_extract(return_value=_mock_extract_ok("f1", "test.txt"))
|
||||
|
||||
|
|
@ -75,7 +77,8 @@ async def test_single_file_returns_one_connector_document(
|
|||
|
||||
# Slice 2: Multiple files all produce documents
|
||||
async def test_multiple_files_all_produce_documents(
|
||||
mock_onedrive_client, patch_extract,
|
||||
mock_onedrive_client,
|
||||
patch_extract,
|
||||
):
|
||||
files = [_make_file_dict(f"f{i}", f"file{i}.txt") for i in range(3)]
|
||||
patch_extract(
|
||||
|
|
@ -98,7 +101,8 @@ async def test_multiple_files_all_produce_documents(
|
|||
|
||||
# Slice 3: Error isolation
|
||||
async def test_one_download_exception_does_not_block_others(
|
||||
mock_onedrive_client, patch_extract,
|
||||
mock_onedrive_client,
|
||||
patch_extract,
|
||||
):
|
||||
files = [_make_file_dict(f"f{i}", f"file{i}.txt") for i in range(3)]
|
||||
patch_extract(
|
||||
|
|
@ -125,7 +129,8 @@ async def test_one_download_exception_does_not_block_others(
|
|||
|
||||
# Slice 4: ETL error counts as download failure
|
||||
async def test_etl_error_counts_as_download_failure(
|
||||
mock_onedrive_client, patch_extract,
|
||||
mock_onedrive_client,
|
||||
patch_extract,
|
||||
):
|
||||
files = [_make_file_dict("f0", "good.txt"), _make_file_dict("f1", "bad.txt")]
|
||||
patch_extract(
|
||||
|
|
@ -150,7 +155,8 @@ async def test_etl_error_counts_as_download_failure(
|
|||
|
||||
# Slice 5: Semaphore bound
|
||||
async def test_concurrency_bounded_by_semaphore(
|
||||
mock_onedrive_client, monkeypatch,
|
||||
mock_onedrive_client,
|
||||
monkeypatch,
|
||||
):
|
||||
lock = asyncio.Lock()
|
||||
active = 0
|
||||
|
|
@ -190,7 +196,8 @@ async def test_concurrency_bounded_by_semaphore(
|
|||
|
||||
# Slice 6: Heartbeat fires
|
||||
async def test_heartbeat_fires_during_parallel_downloads(
|
||||
mock_onedrive_client, monkeypatch,
|
||||
mock_onedrive_client,
|
||||
monkeypatch,
|
||||
):
|
||||
import app.tasks.connector_indexers.onedrive_indexer as _mod
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue