mirror of
https://github.com/danswer-ai/danswer.git
synced 2025-03-25 17:22:06 +01:00
* remove left over vim command * additional misconfigurations * ensure all regions updated
240 lines
7.6 KiB
Python
240 lines
7.6 KiB
Python
from typing import Any, Literal
|
|
from onyx.db.engine import get_iam_auth_token
|
|
from onyx.configs.app_configs import USE_IAM_AUTH
|
|
from onyx.configs.app_configs import POSTGRES_HOST
|
|
from onyx.configs.app_configs import POSTGRES_PORT
|
|
from onyx.configs.app_configs import POSTGRES_USER
|
|
from onyx.configs.app_configs import AWS_REGION_NAME
|
|
from onyx.db.engine import build_connection_string
|
|
from onyx.db.engine import get_all_tenant_ids
|
|
from sqlalchemy import event
|
|
from sqlalchemy import pool
|
|
from sqlalchemy import text
|
|
from sqlalchemy.engine.base import Connection
|
|
import os
|
|
import ssl
|
|
import asyncio
|
|
import logging
|
|
from logging.config import fileConfig
|
|
|
|
from alembic import context
|
|
from sqlalchemy.ext.asyncio import create_async_engine
|
|
from sqlalchemy.sql.schema import SchemaItem
|
|
from onyx.configs.constants import SSL_CERT_FILE
|
|
from shared_configs.configs import MULTI_TENANT, POSTGRES_DEFAULT_SCHEMA
|
|
from onyx.db.models import Base
|
|
from celery.backends.database.session import ResultModelBase # type: ignore
|
|
|
|
# Alembic Config object
|
|
config = context.config
|
|
|
|
if config.config_file_name is not None and config.attributes.get(
|
|
"configure_logger", True
|
|
):
|
|
fileConfig(config.config_file_name)
|
|
|
|
target_metadata = [Base.metadata, ResultModelBase.metadata]
|
|
|
|
EXCLUDE_TABLES = {"kombu_queue", "kombu_message"}
|
|
logger = logging.getLogger(__name__)
|
|
|
|
ssl_context: ssl.SSLContext | None = None
|
|
if USE_IAM_AUTH:
|
|
if not os.path.exists(SSL_CERT_FILE):
|
|
raise FileNotFoundError(f"Expected {SSL_CERT_FILE} when USE_IAM_AUTH is true.")
|
|
ssl_context = ssl.create_default_context(cafile=SSL_CERT_FILE)
|
|
|
|
|
|
def include_object(
|
|
object: SchemaItem,
|
|
name: str | None,
|
|
type_: Literal[
|
|
"schema",
|
|
"table",
|
|
"column",
|
|
"index",
|
|
"unique_constraint",
|
|
"foreign_key_constraint",
|
|
],
|
|
reflected: bool,
|
|
compare_to: SchemaItem | None,
|
|
) -> bool:
|
|
if type_ == "table" and name in EXCLUDE_TABLES:
|
|
return False
|
|
return True
|
|
|
|
|
|
def get_schema_options() -> tuple[str, bool, bool]:
|
|
x_args_raw = context.get_x_argument()
|
|
x_args = {}
|
|
for arg in x_args_raw:
|
|
for pair in arg.split(","):
|
|
if "=" in pair:
|
|
key, value = pair.split("=", 1)
|
|
x_args[key.strip()] = value.strip()
|
|
schema_name = x_args.get("schema", POSTGRES_DEFAULT_SCHEMA)
|
|
create_schema = x_args.get("create_schema", "true").lower() == "true"
|
|
upgrade_all_tenants = x_args.get("upgrade_all_tenants", "false").lower() == "true"
|
|
|
|
if (
|
|
MULTI_TENANT
|
|
and schema_name == POSTGRES_DEFAULT_SCHEMA
|
|
and not upgrade_all_tenants
|
|
):
|
|
raise ValueError(
|
|
"Cannot run default migrations in public schema when multi-tenancy is enabled. "
|
|
"Please specify a tenant-specific schema."
|
|
)
|
|
|
|
return schema_name, create_schema, upgrade_all_tenants
|
|
|
|
|
|
def do_run_migrations(
|
|
connection: Connection, schema_name: str, create_schema: bool
|
|
) -> None:
|
|
logger.info(f"About to migrate schema: {schema_name}")
|
|
|
|
if create_schema:
|
|
connection.execute(text(f'CREATE SCHEMA IF NOT EXISTS "{schema_name}"'))
|
|
connection.execute(text("COMMIT"))
|
|
|
|
connection.execute(text(f'SET search_path TO "{schema_name}"'))
|
|
|
|
context.configure(
|
|
connection=connection,
|
|
target_metadata=target_metadata, # type: ignore
|
|
include_object=include_object,
|
|
version_table_schema=schema_name,
|
|
include_schemas=True,
|
|
compare_type=True,
|
|
compare_server_default=True,
|
|
script_location=config.get_main_option("script_location"),
|
|
)
|
|
|
|
with context.begin_transaction():
|
|
context.run_migrations()
|
|
|
|
|
|
def provide_iam_token_for_alembic(
|
|
dialect: Any, conn_rec: Any, cargs: Any, cparams: Any
|
|
) -> None:
|
|
if USE_IAM_AUTH:
|
|
# Database connection settings
|
|
region = AWS_REGION_NAME
|
|
host = POSTGRES_HOST
|
|
port = POSTGRES_PORT
|
|
user = POSTGRES_USER
|
|
|
|
# Get IAM authentication token
|
|
token = get_iam_auth_token(host, port, user, region)
|
|
|
|
# For Alembic / SQLAlchemy in this context, set SSL and password
|
|
cparams["password"] = token
|
|
cparams["ssl"] = ssl_context
|
|
|
|
|
|
async def run_async_migrations() -> None:
|
|
schema_name, create_schema, upgrade_all_tenants = get_schema_options()
|
|
|
|
engine = create_async_engine(
|
|
build_connection_string(),
|
|
poolclass=pool.NullPool,
|
|
)
|
|
|
|
if USE_IAM_AUTH:
|
|
|
|
@event.listens_for(engine.sync_engine, "do_connect")
|
|
def event_provide_iam_token_for_alembic(
|
|
dialect: Any, conn_rec: Any, cargs: Any, cparams: Any
|
|
) -> None:
|
|
provide_iam_token_for_alembic(dialect, conn_rec, cargs, cparams)
|
|
|
|
if upgrade_all_tenants:
|
|
tenant_schemas = get_all_tenant_ids()
|
|
for schema in tenant_schemas:
|
|
try:
|
|
logger.info(f"Migrating schema: {schema}")
|
|
async with engine.connect() as connection:
|
|
await connection.run_sync(
|
|
do_run_migrations,
|
|
schema_name=schema,
|
|
create_schema=create_schema,
|
|
)
|
|
except Exception as e:
|
|
logger.error(f"Error migrating schema {schema}: {e}")
|
|
raise
|
|
else:
|
|
try:
|
|
logger.info(f"Migrating schema: {schema_name}")
|
|
async with engine.connect() as connection:
|
|
await connection.run_sync(
|
|
do_run_migrations,
|
|
schema_name=schema_name,
|
|
create_schema=create_schema,
|
|
)
|
|
except Exception as e:
|
|
logger.error(f"Error migrating schema {schema_name}: {e}")
|
|
raise
|
|
|
|
await engine.dispose()
|
|
|
|
|
|
def run_migrations_offline() -> None:
|
|
schema_name, _, upgrade_all_tenants = get_schema_options()
|
|
url = build_connection_string()
|
|
|
|
if upgrade_all_tenants:
|
|
engine = create_async_engine(url)
|
|
|
|
if USE_IAM_AUTH:
|
|
|
|
@event.listens_for(engine.sync_engine, "do_connect")
|
|
def event_provide_iam_token_for_alembic_offline(
|
|
dialect: Any, conn_rec: Any, cargs: Any, cparams: Any
|
|
) -> None:
|
|
provide_iam_token_for_alembic(dialect, conn_rec, cargs, cparams)
|
|
|
|
tenant_schemas = get_all_tenant_ids()
|
|
engine.sync_engine.dispose()
|
|
|
|
for schema in tenant_schemas:
|
|
logger.info(f"Migrating schema: {schema}")
|
|
context.configure(
|
|
url=url,
|
|
target_metadata=target_metadata, # type: ignore
|
|
literal_binds=True,
|
|
include_object=include_object,
|
|
version_table_schema=schema,
|
|
include_schemas=True,
|
|
script_location=config.get_main_option("script_location"),
|
|
dialect_opts={"paramstyle": "named"},
|
|
)
|
|
|
|
with context.begin_transaction():
|
|
context.run_migrations()
|
|
else:
|
|
logger.info(f"Migrating schema: {schema_name}")
|
|
context.configure(
|
|
url=url,
|
|
target_metadata=target_metadata, # type: ignore
|
|
literal_binds=True,
|
|
include_object=include_object,
|
|
version_table_schema=schema_name,
|
|
include_schemas=True,
|
|
script_location=config.get_main_option("script_location"),
|
|
dialect_opts={"paramstyle": "named"},
|
|
)
|
|
|
|
with context.begin_transaction():
|
|
context.run_migrations()
|
|
|
|
|
|
def run_migrations_online() -> None:
|
|
asyncio.run(run_async_migrations())
|
|
|
|
|
|
if context.is_offline_mode():
|
|
run_migrations_offline()
|
|
else:
|
|
run_migrations_online()
|