feat: add integration and unit tests for local folder indexing and document versioning

This commit is contained in:
Anish Sarkar 2026-04-02 11:12:16 +05:30
parent 96a58d0d30
commit 775dea7894
4 changed files with 894 additions and 0 deletions

View file

@ -0,0 +1,609 @@
"""Integration tests for local folder indexer — Tier 3 (I1-I5), Tier 4 (F1-F5), Tier 5 (P1)."""
import os
from pathlib import Path
from unittest.mock import AsyncMock, MagicMock
import pytest
from sqlalchemy import func, select
from sqlalchemy.ext.asyncio import AsyncSession
from app.db import (
Document,
DocumentStatus,
DocumentType,
DocumentVersion,
Folder,
SearchSourceConnector,
SearchSpace,
User,
)
import app.tasks.connector_indexers.local_folder_indexer as _lfi_mod
pytestmark = pytest.mark.integration
@pytest.fixture
def patched_self_hosted(monkeypatch):
_cfg = type("_Cfg", (), {"is_self_hosted": staticmethod(lambda: True)})()
monkeypatch.setattr(_lfi_mod, "config", _cfg)
@pytest.fixture
def patched_embed_for_indexer(monkeypatch):
from app.config import config as app_config
dim = app_config.embedding_model_instance.dimension
mock = MagicMock(return_value=[0.1] * dim)
monkeypatch.setattr(_lfi_mod, "embed_text", mock)
return mock
@pytest.fixture
def patched_chunks_for_indexer(monkeypatch):
from app.db import Chunk
from app.config import config as app_config
dim = app_config.embedding_model_instance.dimension
async def mock_create_chunks(text):
return [Chunk(content="chunk", embedding=[0.1] * dim)]
monkeypatch.setattr(_lfi_mod, "create_document_chunks", mock_create_chunks)
@pytest.fixture
def patched_summary_for_indexer(monkeypatch):
monkeypatch.setattr(_lfi_mod, "get_user_long_context_llm", AsyncMock(return_value=None))
# ====================================================================
# Tier 3: Full Indexer Integration (I1-I5)
# ====================================================================
class TestFullIndexer:
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_i1_new_file_indexed(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""I1: Single new .md file is indexed with status READY."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
(tmp_path / "note.md").write_text("# Hello World\n\nContent here.")
count, skipped, err = await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
assert err is None
assert count == 1
docs = (
await db_session.execute(
select(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalars().all()
assert len(docs) == 1
assert docs[0].document_type == DocumentType.LOCAL_FOLDER_FILE
assert DocumentStatus.is_state(docs[0].status, DocumentStatus.READY)
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_i2_unchanged_skipped(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""I2: Second run on unchanged directory creates no new documents."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
(tmp_path / "note.md").write_text("# Hello\n\nSame content.")
count1, _, _ = await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
assert count1 == 1
# Second run — unchanged
count2, _, _ = await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
assert count2 == 0
total = (
await db_session.execute(
select(func.count()).select_from(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalar_one()
assert total == 1
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_i3_changed_reindexed(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""I3: Modified file content triggers re-index and creates a version."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
f = tmp_path / "note.md"
f.write_text("# Version 1\n\nOriginal.")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
# Modify
f.write_text("# Version 2\n\nUpdated.")
# Touch mtime to ensure it's detected as different
os.utime(f, (f.stat().st_atime + 10, f.stat().st_mtime + 10))
count, _, _ = await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
assert count == 1
# Should have a version snapshot
versions = (
await db_session.execute(
select(DocumentVersion).join(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalars().all()
assert len(versions) >= 1
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_i4_deleted_removed(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""I4: Deleted file is removed from DB on re-sync."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
f = tmp_path / "to_delete.md"
f.write_text("# Delete me")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
docs_before = (
await db_session.execute(
select(func.count()).select_from(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalar_one()
assert docs_before == 1
f.unlink()
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
docs_after = (
await db_session.execute(
select(func.count()).select_from(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalar_one()
assert docs_after == 0
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_i5_single_file_mode(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""I5: Single-file mode only processes the specified file."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
(tmp_path / "a.md").write_text("File A")
(tmp_path / "b.md").write_text("File B")
(tmp_path / "c.md").write_text("File C")
count, _, _ = await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
target_file_path=str(tmp_path / "b.md"),
)
assert count == 1
docs = (
await db_session.execute(
select(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalars().all()
assert len(docs) == 1
assert docs[0].title == "b"
# ====================================================================
# Tier 4: Folder Mirroring (F1-F5)
# ====================================================================
class TestFolderMirroring:
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_f1_root_folder_created(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""F1: First sync creates a root Folder and stores root_folder_id."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
(tmp_path / "root.md").write_text("Root file")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
# Refresh connector
await db_session.refresh(db_local_folder_connector)
root_id = db_local_folder_connector.config.get("root_folder_id")
assert root_id is not None
root_folder = (
await db_session.execute(select(Folder).where(Folder.id == root_id))
).scalar_one()
assert root_folder.name == "test-folder"
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_f2_nested_folder_rows(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""F2: Nested dirs create Folder rows with correct parent_id chain."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
daily = tmp_path / "notes" / "daily"
daily.mkdir(parents=True)
weekly = tmp_path / "notes" / "weekly"
weekly.mkdir(parents=True)
(daily / "today.md").write_text("today")
(weekly / "review.md").write_text("review")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
folders = (
await db_session.execute(
select(Folder).where(Folder.search_space_id == db_search_space.id)
)
).scalars().all()
folder_names = {f.name for f in folders}
assert "notes" in folder_names
assert "daily" in folder_names
assert "weekly" in folder_names
notes_folder = next(f for f in folders if f.name == "notes")
daily_folder = next(f for f in folders if f.name == "daily")
weekly_folder = next(f for f in folders if f.name == "weekly")
assert daily_folder.parent_id == notes_folder.id
assert weekly_folder.parent_id == notes_folder.id
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_f3_resync_reuses_folders(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""F3: Re-sync reuses existing Folder rows, no duplicates."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
sub = tmp_path / "docs"
sub.mkdir()
(sub / "file.md").write_text("content")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
folders_before = (
await db_session.execute(
select(Folder).where(Folder.search_space_id == db_search_space.id)
)
).scalars().all()
ids_before = {f.id for f in folders_before}
# Re-sync
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
folders_after = (
await db_session.execute(
select(Folder).where(Folder.search_space_id == db_search_space.id)
)
).scalars().all()
ids_after = {f.id for f in folders_after}
assert ids_before == ids_after
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_f4_folder_id_assigned(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""F4: Documents get correct folder_id based on their directory."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
daily = tmp_path / "notes" / "daily"
daily.mkdir(parents=True)
(daily / "today.md").write_text("today note")
(tmp_path / "root.md").write_text("root note")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
docs = (
await db_session.execute(
select(Document).where(
Document.connector_id == db_local_folder_connector.id
)
)
).scalars().all()
today_doc = next(d for d in docs if d.title == "today")
root_doc = next(d for d in docs if d.title == "root")
daily_folder = (
await db_session.execute(
select(Folder).where(Folder.name == "daily")
)
).scalar_one()
assert today_doc.folder_id == daily_folder.id
# Root doc should be in the root folder
await db_session.refresh(db_local_folder_connector)
root_fid = db_local_folder_connector.config.get("root_folder_id")
assert root_doc.folder_id == root_fid
@pytest.mark.usefixtures(
"patched_self_hosted",
"patched_embed_for_indexer",
"patched_chunks_for_indexer",
"patched_summary_for_indexer",
)
async def test_f5_empty_folder_cleanup(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
tmp_path: Path,
):
"""F5: Deleted dir's empty Folder row is cleaned up on re-sync."""
from app.tasks.connector_indexers.local_folder_indexer import index_local_folder
import shutil
daily = tmp_path / "notes" / "daily"
daily.mkdir(parents=True)
weekly = tmp_path / "notes" / "weekly"
weekly.mkdir(parents=True)
(daily / "today.md").write_text("today")
(weekly / "review.md").write_text("review")
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
# Verify weekly folder exists
weekly_folder = (
await db_session.execute(
select(Folder).where(Folder.name == "weekly")
)
).scalar_one_or_none()
assert weekly_folder is not None
# Delete weekly directory + its file
shutil.rmtree(weekly)
await index_local_folder(
session=db_session,
connector_id=db_local_folder_connector.id,
search_space_id=db_search_space.id,
user_id=str(db_user.id),
)
# weekly Folder should be gone (empty, dir removed)
weekly_after = (
await db_session.execute(
select(Folder).where(Folder.name == "weekly")
)
).scalar_one_or_none()
assert weekly_after is None
# daily should still exist
daily_after = (
await db_session.execute(
select(Folder).where(Folder.name == "daily")
)
).scalar_one_or_none()
assert daily_after is not None
# ====================================================================
# Tier 5: Pipeline Integration (P1)
# ====================================================================
class TestPipelineIntegration:
@pytest.mark.usefixtures(
"patched_summarize", "patched_embed_texts", "patched_chunk_text"
)
async def test_p1_local_folder_file_through_pipeline(
self,
db_session: AsyncSession,
db_local_folder_connector: SearchSourceConnector,
db_user: User,
db_search_space: SearchSpace,
mocker,
):
"""P1: LOCAL_FOLDER_FILE ConnectorDocument through prepare+index to READY."""
from app.indexing_pipeline.connector_document import ConnectorDocument
from app.indexing_pipeline.indexing_pipeline_service import IndexingPipelineService
doc = ConnectorDocument(
title="Test Local File",
source_markdown="## Local file\n\nContent from disk.",
unique_id="test-folder:test.md",
document_type=DocumentType.LOCAL_FOLDER_FILE,
search_space_id=db_search_space.id,
connector_id=db_local_folder_connector.id,
created_by_id=str(db_user.id),
)
service = IndexingPipelineService(session=db_session)
prepared = await service.prepare_for_indexing([doc])
assert len(prepared) == 1
db_doc = prepared[0]
result = await service.index(db_doc, doc, llm=mocker.Mock())
assert result is not None
docs = (
await db_session.execute(
select(Document).where(
Document.document_type == DocumentType.LOCAL_FOLDER_FILE,
Document.search_space_id == db_search_space.id,
)
)
).scalars().all()
assert len(docs) == 1
assert DocumentStatus.is_state(docs[0].status, DocumentStatus.READY)