mirror of
https://github.com/danswer-ai/danswer.git
synced 2025-05-08 10:50:26 +02:00
* fresh indexing feature branch * cherry pick test * Revert "cherry pick test" This reverts commit 2a624220687affdda3de347e30f2011136f64bda. * set multitenant so that vespa fields match when indexing * cleanup pass * mypy * pass through env var to control celery indexing concurrency * comments on task kickoff and some logging improvements * use get_session_with_tenant * comment out all of update.py * rename to RedisConnectorIndexingFenceData * first check num_indexing_workers * refactor RedisConnectorIndexingFenceData * comment out on_worker_process_init * fix where num_indexing_workers falls back * remove extra brace
173 lines
6.1 KiB
Python
173 lines
6.1 KiB
Python
import time
|
|
from uuid import uuid4
|
|
|
|
import requests
|
|
|
|
from tests.integration.common_utils.constants import API_SERVER_URL
|
|
from tests.integration.common_utils.constants import GENERAL_HEADERS
|
|
from tests.integration.common_utils.constants import MAX_DELAY
|
|
from tests.integration.common_utils.test_models import DATestDocumentSet
|
|
from tests.integration.common_utils.test_models import DATestUser
|
|
|
|
|
|
class DocumentSetManager:
|
|
@staticmethod
|
|
def create(
|
|
name: str | None = None,
|
|
description: str | None = None,
|
|
cc_pair_ids: list[int] | None = None,
|
|
is_public: bool = True,
|
|
users: list[str] | None = None,
|
|
groups: list[int] | None = None,
|
|
user_performing_action: DATestUser | None = None,
|
|
) -> DATestDocumentSet:
|
|
if name is None:
|
|
name = f"test_doc_set_{str(uuid4())}"
|
|
|
|
doc_set_creation_request = {
|
|
"name": name,
|
|
"description": description or name,
|
|
"cc_pair_ids": cc_pair_ids or [],
|
|
"is_public": is_public,
|
|
"users": users or [],
|
|
"groups": groups or [],
|
|
}
|
|
|
|
response = requests.post(
|
|
f"{API_SERVER_URL}/manage/admin/document-set",
|
|
json=doc_set_creation_request,
|
|
headers=user_performing_action.headers
|
|
if user_performing_action
|
|
else GENERAL_HEADERS,
|
|
)
|
|
response.raise_for_status()
|
|
|
|
return DATestDocumentSet(
|
|
id=int(response.json()),
|
|
name=name,
|
|
description=description or name,
|
|
cc_pair_ids=cc_pair_ids or [],
|
|
is_public=is_public,
|
|
is_up_to_date=True,
|
|
users=users or [],
|
|
groups=groups or [],
|
|
)
|
|
|
|
@staticmethod
|
|
def edit(
|
|
document_set: DATestDocumentSet,
|
|
user_performing_action: DATestUser | None = None,
|
|
) -> bool:
|
|
doc_set_update_request = {
|
|
"id": document_set.id,
|
|
"description": document_set.description,
|
|
"cc_pair_ids": document_set.cc_pair_ids,
|
|
"is_public": document_set.is_public,
|
|
"users": document_set.users,
|
|
"groups": document_set.groups,
|
|
}
|
|
response = requests.patch(
|
|
f"{API_SERVER_URL}/manage/admin/document-set",
|
|
json=doc_set_update_request,
|
|
headers=user_performing_action.headers
|
|
if user_performing_action
|
|
else GENERAL_HEADERS,
|
|
)
|
|
response.raise_for_status()
|
|
return True
|
|
|
|
@staticmethod
|
|
def delete(
|
|
document_set: DATestDocumentSet,
|
|
user_performing_action: DATestUser | None = None,
|
|
) -> bool:
|
|
response = requests.delete(
|
|
f"{API_SERVER_URL}/manage/admin/document-set/{document_set.id}",
|
|
headers=user_performing_action.headers
|
|
if user_performing_action
|
|
else GENERAL_HEADERS,
|
|
)
|
|
response.raise_for_status()
|
|
return True
|
|
|
|
@staticmethod
|
|
def get_all(
|
|
user_performing_action: DATestUser | None = None,
|
|
) -> list[DATestDocumentSet]:
|
|
response = requests.get(
|
|
f"{API_SERVER_URL}/manage/document-set",
|
|
headers=user_performing_action.headers
|
|
if user_performing_action
|
|
else GENERAL_HEADERS,
|
|
)
|
|
response.raise_for_status()
|
|
return [
|
|
DATestDocumentSet(
|
|
id=doc_set["id"],
|
|
name=doc_set["name"],
|
|
description=doc_set["description"],
|
|
cc_pair_ids=[
|
|
cc_pair["id"] for cc_pair in doc_set["cc_pair_descriptors"]
|
|
],
|
|
is_public=doc_set["is_public"],
|
|
is_up_to_date=doc_set["is_up_to_date"],
|
|
users=doc_set["users"],
|
|
groups=doc_set["groups"],
|
|
)
|
|
for doc_set in response.json()
|
|
]
|
|
|
|
@staticmethod
|
|
def wait_for_sync(
|
|
document_sets_to_check: list[DATestDocumentSet] | None = None,
|
|
user_performing_action: DATestUser | None = None,
|
|
) -> None:
|
|
# wait for document sets to be synced
|
|
start = time.time()
|
|
while True:
|
|
doc_sets = DocumentSetManager.get_all(user_performing_action)
|
|
if document_sets_to_check:
|
|
check_ids = {doc_set.id for doc_set in document_sets_to_check}
|
|
doc_set_ids = {doc_set.id for doc_set in doc_sets}
|
|
if not check_ids.issubset(doc_set_ids):
|
|
raise RuntimeError("Document set not found")
|
|
doc_sets = [doc_set for doc_set in doc_sets if doc_set.id in check_ids]
|
|
all_up_to_date = all(doc_set.is_up_to_date for doc_set in doc_sets)
|
|
|
|
if all_up_to_date:
|
|
print("Document sets synced successfully.")
|
|
break
|
|
|
|
if time.time() - start > MAX_DELAY:
|
|
raise TimeoutError(
|
|
f"Document sets were not synced within the {MAX_DELAY} seconds"
|
|
)
|
|
else:
|
|
print("Document sets were not synced yet, waiting...")
|
|
|
|
time.sleep(2)
|
|
|
|
@staticmethod
|
|
def verify(
|
|
document_set: DATestDocumentSet,
|
|
verify_deleted: bool = False,
|
|
user_performing_action: DATestUser | None = None,
|
|
) -> None:
|
|
doc_sets = DocumentSetManager.get_all(user_performing_action)
|
|
for doc_set in doc_sets:
|
|
if doc_set.id == document_set.id:
|
|
if verify_deleted:
|
|
raise ValueError(
|
|
f"Document set {document_set.id} found but should have been deleted"
|
|
)
|
|
if (
|
|
doc_set.name == document_set.name
|
|
and set(doc_set.cc_pair_ids) == set(document_set.cc_pair_ids)
|
|
and doc_set.is_public == document_set.is_public
|
|
and set(doc_set.users) == set(document_set.users)
|
|
and set(doc_set.groups) == set(document_set.groups)
|
|
):
|
|
return
|
|
if not verify_deleted:
|
|
raise ValueError(f"Document set {document_set.id} not found")
|