Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Use alembic for database table creation #298

Merged
merged 31 commits into from
Feb 16, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
31 commits
Select commit Hold shift + click to select a range
36bc011
Start adding alembic
jstutters Feb 12, 2024
8f4a322
Add alembic for creation tables in the pipeline schema
jstutters Feb 12, 2024
47a7412
Merge branch 'main' into jstutters/270-use-alembic-for-database-table…
jstutters Feb 12, 2024
e7be030
Add license
jstutters Feb 13, 2024
bf2c946
Customise ruff behaviour on alembic files
jstutters Feb 13, 2024
ba56779
Revoke change running alembic during postgres startup
jstutters Feb 14, 2024
942fa45
Apply alembic database migrations during startup of pixl_imaging
jstutters Feb 14, 2024
f19695a
Remove pixl_core from postgres container
jstutters Feb 14, 2024
2327e05
Add docstring
stefpiatek Feb 14, 2024
794d17d
Add extra debugging for migrate and run
stefpiatek Feb 14, 2024
ebdc5f0
Add temporary debug for CI
stefpiatek Feb 14, 2024
ae857c7
Empty commit, for some reason no GHA ran last time
stefpiatek Feb 14, 2024
6bf7641
Merge branch 'main' into jstutters/270-use-alembic-for-database-table…
stefpiatek Feb 14, 2024
40c1c8c
Remove unused test dockerfile
stefpiatek Feb 14, 2024
333390c
Skip alembic in imaging api tests
stefpiatek Feb 14, 2024
2729842
Remove debugging step
stefpiatek Feb 14, 2024
4bab792
Document alembic and move into imaging api
stefpiatek Feb 14, 2024
a4b71fe
Document alembic and move into imaging api
stefpiatek Feb 14, 2024
bbc7978
Document alembic and move into imaging api
stefpiatek Feb 14, 2024
c8dfb5b
Use underscore :facepalm:
stefpiatek Feb 14, 2024
4c484cf
Use alembic config within alembic dir
stefpiatek Feb 15, 2024
4323edc
Set healthcheck in docker compose
stefpiatek Feb 15, 2024
581565f
Add schema to alembic
stefpiatek Feb 15, 2024
8d40705
Add helper script to generate migrations
stefpiatek Feb 15, 2024
4b933aa
Document use of alembic
stefpiatek Feb 15, 2024
8f11e90
Document skip alembic env
stefpiatek Feb 15, 2024
385609f
Keep alembic table in same schema as pipeline
stefpiatek Feb 16, 2024
dfc0905
Name migrations
stefpiatek Feb 16, 2024
eb7a0a7
Merge branch 'main' into jstutters/270-use-alembic-for-database-table…
stefpiatek Feb 16, 2024
4d19dd8
Keep configuration for migrations in env file
stefpiatek Feb 16, 2024
cc10aed
Simplify testing out of migrations
stefpiatek Feb 16, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions cli/src/pixl_cli/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -341,6 +341,7 @@ class APIConfig:
-------
base_url()
Return the base url for the API

"""

def __init__(self, kwargs: dict) -> None:
Expand Down
7 changes: 7 additions & 0 deletions docker-compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -268,6 +268,7 @@ services:
orthanc-raw:
condition: service_healthy
healthcheck:
test: curl -f http://0.0.0.0:8000/heart-beat
interval: 10s
timeout: 30s
retries: 5
Expand All @@ -279,6 +280,12 @@ services:
ORTHANC_RAW_PASSWORD: ${ORTHANC_RAW_PASSWORD}
ORTHANC_RAW_AE_TITLE: ${ORTHANC_RAW_AE_TITLE}
VNAQR_MODALITY: ${VNAQR_MODALITY}
SKIP_ALEMBIC: ${SKIP_ALEMBIC}
PIXL_DB_HOST: ${PIXL_DB_HOST}
PIXL_DB_PORT: ${PIXL_DB_PORT}
PIXL_DB_NAME: ${PIXL_DB_NAME}
PIXL_DB_USER: ${PIXL_DB_USER}
PIXL_DB_PASSWORD: ${PIXL_DB_PASSWORD}
PIXL_DICOM_TRANSFER_TIMEOUT: ${PIXL_DICOM_TRANSFER_TIMEOUT}
PIXL_QUERY_TIMEOUT: ${PIXL_QUERY_TIMEOUT}
ports:
Expand Down
4 changes: 1 addition & 3 deletions docker/imaging-api/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,4 @@ RUN --mount=type=cache,target=/root/.cache \
pip install core/ . && \
if [ "$TEST" = "true" ]; then pip install core/ .[test]; fi

HEALTHCHECK CMD /usr/bin/curl -f http://0.0.0.0:8000/heart-beat || exit 1

ENTRYPOINT ["uvicorn", "pixl_imaging.main:app", "--host", "0.0.0.0", "--port", "8000"]
ENTRYPOINT ["/app/scripts/migrate_and_run.sh"]
18 changes: 1 addition & 17 deletions docker/postgres/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -23,20 +23,4 @@ RUN sed -i '/en_GB.UTF-8/s/^# //g' /etc/locale.gen && locale-gen

COPY --chmod=0755 ./postgres/postgres.conf /etc/postgresql/postgresql.conf

COPY --chmod=0777 ./postgres/pixl-db_init.sh /docker-entrypoint-initdb.d/pixl-db_init.sh

COPY --chmod=0777 ./postgres/create_pixl_tbls.py /pixl/create_pixl_tbls.py

# Install requirements before copying modules
COPY ./pixl_core/pyproject.toml /pixl/pixl_core/pyproject.toml
RUN --mount=type=cache,target=/root/.cache \
python3 -m venv /pixl/venv \
&& cd /pixl/venv/bin \
&& ./pip install /pixl/pixl_core/

WORKDIR /pixl/venv/bin

COPY ./pixl_core/ /pixl/pixl_core

RUN --mount=type=cache,target=/root/.cache \
./pip install /pixl/pixl_core/
COPY --chmod=0777 ./postgres/pixl-db_init.sh /docker-entrypoint-initdb.d/pixl-db_init.sh
11 changes: 11 additions & 0 deletions pixl_imaging/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,3 +24,14 @@ pip install -e ../pixl_core/ .
## Usage

Usage should be from the CLI driver, which interacts with the endpoint.


## Configuration and database interaction

The database tables are updated using alembic, see the [alembic](alembic) dir for more details.

The `SKIP_ALEMBIC` environmental variable is used to control whether migrations are applied to the database.

- Tests that don't use the database use `SKIP_ALEMBIC=true`, but otherwise you probably want to run this.
- If you wanted to test out new migrations from a test/dev deployment on the GAE with data in,
then you can redeploy just the `imaging-api` container while keeping the `postgres` container up.
31 changes: 31 additions & 0 deletions pixl_imaging/alembic/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,31 @@
# Alembic configuration

We've followed the [alembic tutorial](https://alembic.sqlalchemy.org/en/latest/tutorial.html),
there's a good explainer there of each of the files and directories.

## Running migrations in production

This is automatically done when the pixl-imaging container is started,
as the entry point is [migrate_and_run.sh](../scripts/migrate_and_run.sh).

## Creating a new migration after editing the database model

For convenience, the [autogenerate-migration.sh](autogenerate-migration.sh) has been made.

Which you can run giving a name for the migration like this:

```shell
cd alembic
./autogenerate-migration.sh "<Description of changes to models>"
```

- This creates a postgres container
- Runs the existing migrations
- Checks for differences between the SQLAlchemy [models](../../pixl_core/src/core/db/models.py)
and creates a new migration in [versions](versions)
- Takes down the postgres container

There's a couple of manual steps:

- Check the changes autogenerated and update to match your intention
- Commit the changes and review in a pull request as normal
116 changes: 116 additions & 0 deletions pixl_imaging/alembic/alembic.ini
Original file line number Diff line number Diff line change
@@ -0,0 +1,116 @@
# A generic, single database configuration.

[alembic]
# path to migration scripts
script_location = .

# template used to generate migration file names; The default value is %%(rev)s_%%(slug)s
# Uncomment the line below if you want the files to be prepended with date and time
# see https://alembic.sqlalchemy.org/en/latest/tutorial.html#editing-the-ini-file
# for all available tokens
# file_template = %%(year)d_%%(month).2d_%%(day).2d_%%(hour).2d%%(minute).2d-%%(rev)s_%%(slug)s

# sys.path path, will be prepended to sys.path if present.
# defaults to the current working directory.
prepend_sys_path = .

# timezone to use when rendering the date within the migration file
# as well as the filename.
# If specified, requires the python>=3.9 or backports.zoneinfo library.
# Any required deps can installed by adding `alembic[tz]` to the pip requirements
# string value is passed to ZoneInfo()
# leave blank for localtime
# timezone =

# max length of characters to apply to the
# "slug" field
# truncate_slug_length = 40

# set to 'true' to run the environment during
# the 'revision' command, regardless of autogenerate
# revision_environment = false

# set to 'true' to allow .pyc and .pyo files without
# a source .py file to be detected as revisions in the
# versions/ directory
# sourceless = false

# version location specification; This defaults
# to alembic/versions. When using multiple version
# directories, initial revisions must be specified with --version-path.
# The path separator used here should be the separator specified by "version_path_separator" below.
# version_locations = %(here)s/bar:%(here)s/bat:alembic/versions

# version path separator; As mentioned above, this is the character used to split
# version_locations. The default within new alembic.ini files is "os", which uses os.pathsep.
# If this key is omitted entirely, it falls back to the legacy behavior of splitting on spaces and/or commas.
# Valid values for version_path_separator are:
#
# version_path_separator = :
# version_path_separator = ;
# version_path_separator = space
version_path_separator = os # Use os.pathsep. Default configuration used for new projects.

# set to 'true' to search source files recursively
# in each "version_locations" directory
# new in Alembic version 1.10
# recursive_version_locations = false

# the output encoding used when revision files
# are written from script.py.mako
# output_encoding = utf-8

sqlalchemy.url = driver://user:pass@localhost/dbname


[post_write_hooks]
# post_write_hooks defines scripts or Python functions that are run
# on newly generated revision scripts. See the documentation for further
# detail and examples

# format using "black" - use the console_scripts runner, against the "black" entrypoint
# hooks = black
# black.type = console_scripts
# black.entrypoint = black
# black.options = -l 79 REVISION_SCRIPT_FILENAME

# lint with attempts to fix using "ruff" - use the exec runner, execute a binary
# hooks = ruff
# ruff.type = exec
# ruff.executable = %(here)s/.venv/bin/ruff
# ruff.options = --fix REVISION_SCRIPT_FILENAME

# Logging configuration
[loggers]
keys = root,sqlalchemy,alembic

[handlers]
keys = console

[formatters]
keys = generic

[logger_root]
level = WARN
handlers = console
qualname =

[logger_sqlalchemy]
level = WARN
handlers =
qualname = sqlalchemy.engine

[logger_alembic]
level = INFO
handlers =
qualname = alembic

[handler_console]
class = StreamHandler
args = (sys.stderr,)
level = NOTSET
formatter = generic

[formatter_generic]
format = %(levelname)-5.5s [%(name)s] %(message)s
datefmt = %H:%M:%S
34 changes: 18 additions & 16 deletions postgres/create_pixl_tbls.py → ...imaging/alembic/autogenerate-migration.sh
100644 → 100755
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
#!/bin/bash
# Copyright (c) University College London Hospitals NHS Foundation Trust
#
# Licensed under the Apache License, Version 2.0 (the "License");
Expand All @@ -11,25 +12,26 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
set -euxo pipefail
cd "$(dirname "${BASH_SOURCE[0]}")" && pwd

"""Create PIXL tables"""
import os
if [ $# -ne 1 ]
then
echo "Add a name for the migration in quotes"
exit 1
fi

from core.db.models import Base
from sqlalchemy import URL, create_engine
from sqlalchemy.sql.ddl import CreateSchema
export PIXL_DB_HOST=localhost

url = URL.create(
drivername="postgresql+psycopg2",
username=os.environ["POSTGRES_USER"],
password=os.environ["POSTGRES_PASSWORD"],
database=os.environ["POSTGRES_DB"],
)
# create postgres
(cd ../.. && \
docker compose --env-file test/.env -p migration-test up --wait -d --build postgres)

engine = create_engine(url, echo=True, echo_pool="debug")
# run current migrations
alembic upgrade head

with engine.connect() as connection:
connection.execute(CreateSchema("pipeline", if_not_exists=True))
connection.commit()
# generate new migrations
alembic revision --autogenerate -m "$1"

Base.metadata.create_all(engine)
# take containers down
(cd ../.. && docker compose --env-file test/.env -p migration-test down --volumes )
120 changes: 120 additions & 0 deletions pixl_imaging/alembic/env.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,120 @@
# Copyright (c) University College London Hospitals NHS Foundation Trust
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Alembic configuration"""
from logging.config import fileConfig

from decouple import Config, RepositoryEnv
from sqlalchemy import URL, create_engine

from alembic import context

env_config = Config(RepositoryEnv("migrations.env"))

# this is the Alembic Config object, which provides
# access to the values within the .ini file in use.
config = context.config

# Interpret the config file for Python logging.
# This line sets up loggers basically.
if config.config_file_name is not None:
fileConfig(config.config_file_name)

# add your model's MetaData object here
# for 'autogenerate' support
from core.db import models

target_metadata = models.Base.metadata


# other values from the config, defined by the needs of env.py,
# can be acquired:
# my_important_option = config.get_main_option("my_important_option")
# ... etc.


def include_name(name: str, type_: str, parent_names: str) -> bool: # noqa: ARG001
"""Filtering of allowed table names"""
if type_ == "schema":
return name in [target_metadata.schema]
return True


def get_pixl_db_url() -> URL:
"""Create PIXL database URL for connection."""
return URL.create(
drivername="postgresql+psycopg2",
host=env_config("PIXL_DB_HOST"),
port=env_config("PIXL_DB_PORT", int),
username=env_config("PIXL_DB_USER"),
password=env_config("PIXL_DB_PASSWORD"),
database=env_config("PIXL_DB_NAME"),
)


def run_migrations_offline() -> None:
"""
Run migrations in 'offline' mode.

This configures the context with just a URL
and not an Engine, though an Engine is acceptable
here as well. By skipping the Engine creation
we don't even need a DBAPI to be available.

Calls to context.execute() here emit the given string to the
script output.

"""
url = get_pixl_db_url()
context.configure(
url=url,
target_metadata=target_metadata,
literal_binds=True,
dialect_opts={"paramstyle": "named"},
include_schemas=True,
include_name=include_name,
version_table_schema=target_metadata.schema,
)

with context.begin_transaction():
context.run_migrations()


def run_migrations_online() -> None:
"""
Run migrations in 'online' mode.

In this scenario we need to create an Engine
and associate a connection with the context.

"""
url = get_pixl_db_url()
connectable = create_engine(url)

with connectable.connect() as connection:
context.configure(
connection=connection,
target_metadata=target_metadata,
include_schemas=True,
include_name=include_name,
version_table_schema=target_metadata.schema,
)

with context.begin_transaction():
context.run_migrations()


if context.is_offline_mode():
run_migrations_offline()
else:
run_migrations_online()
Loading
Loading