2015-10-13 04:38:48 -06:00
|
|
|
# -*- coding: utf-8 -*-
|
2016-01-05 11:01:18 -07:00
|
|
|
# Copyright 2014 - 2016 OpenMarket Ltd
|
2018-03-28 07:03:37 -06:00
|
|
|
# Copyright 2018 New Vector Ltd
|
2015-10-13 04:38:48 -06:00
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import imp
|
|
|
|
import logging
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
import attr
|
|
|
|
|
2019-06-03 06:02:47 -06:00
|
|
|
from synapse.storage.engines.postgres import PostgresEngine
|
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
# Remember to update this number every time a change is made to database
|
|
|
|
# schema files, so the users will be informed on server restarts.
|
2019-07-17 08:33:37 -06:00
|
|
|
SCHEMA_VERSION = 56
|
2015-10-13 04:38:48 -06:00
|
|
|
|
|
|
|
dir_path = os.path.abspath(os.path.dirname(__file__))
|
|
|
|
|
|
|
|
|
|
|
|
class PrepareDatabaseException(Exception):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class UpgradeDatabaseException(PrepareDatabaseException):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
2016-02-11 07:10:00 -07:00
|
|
|
def prepare_database(db_conn, database_engine, config):
|
2015-10-13 04:38:48 -06:00
|
|
|
"""Prepares a database for usage. Will either create all necessary tables
|
|
|
|
or upgrade from an older schema version.
|
2016-04-06 07:08:18 -06:00
|
|
|
|
|
|
|
If `config` is None then prepare_database will assert that no upgrade is
|
|
|
|
necessary, *or* will create a fresh database if the database is empty.
|
2017-10-30 09:16:21 -06:00
|
|
|
|
|
|
|
Args:
|
|
|
|
db_conn:
|
|
|
|
database_engine:
|
|
|
|
config (synapse.config.homeserver.HomeServerConfig|None):
|
|
|
|
application config, or None if we are connecting to an existing
|
|
|
|
database which we expect to be configured already
|
2015-10-13 04:38:48 -06:00
|
|
|
"""
|
2019-10-21 09:08:40 -06:00
|
|
|
|
|
|
|
# For now we only have the one datastore.
|
|
|
|
data_stores = ["main"]
|
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
try:
|
|
|
|
cur = db_conn.cursor()
|
|
|
|
version_info = _get_or_create_schema_state(cur, database_engine)
|
|
|
|
|
|
|
|
if version_info:
|
|
|
|
user_version, delta_files, upgraded = version_info
|
|
|
|
|
2016-04-06 07:08:18 -06:00
|
|
|
if config is None:
|
|
|
|
if user_version != SCHEMA_VERSION:
|
|
|
|
# If we don't pass in a config file then we are expecting to
|
|
|
|
# have already upgraded the DB.
|
|
|
|
raise UpgradeDatabaseException("Database needs to be upgraded")
|
|
|
|
else:
|
|
|
|
_upgrade_existing_database(
|
2019-10-21 09:08:40 -06:00
|
|
|
cur,
|
|
|
|
user_version,
|
|
|
|
delta_files,
|
|
|
|
upgraded,
|
|
|
|
database_engine,
|
|
|
|
config,
|
|
|
|
data_stores=data_stores,
|
2016-04-06 07:08:18 -06:00
|
|
|
)
|
|
|
|
else:
|
2019-10-21 09:08:40 -06:00
|
|
|
_setup_new_database(cur, database_engine, data_stores=data_stores)
|
2015-10-13 04:38:48 -06:00
|
|
|
|
2017-10-30 09:16:21 -06:00
|
|
|
# check if any of our configured dynamic modules want a database
|
|
|
|
if config is not None:
|
|
|
|
_apply_module_schemas(cur, database_engine, config)
|
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
cur.close()
|
|
|
|
db_conn.commit()
|
2017-10-23 08:52:32 -06:00
|
|
|
except Exception:
|
2015-10-13 04:38:48 -06:00
|
|
|
db_conn.rollback()
|
|
|
|
raise
|
|
|
|
|
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
def _setup_new_database(cur, database_engine, data_stores):
|
2015-10-13 04:38:48 -06:00
|
|
|
"""Sets up the database by finding a base set of "full schemas" and then
|
2019-10-22 11:43:31 -06:00
|
|
|
applying any necessary deltas, including schemas from the given data
|
|
|
|
stores.
|
2015-10-13 04:38:48 -06:00
|
|
|
|
|
|
|
The "full_schemas" directory has subdirectories named after versions. This
|
|
|
|
function searches for the highest version less than or equal to
|
|
|
|
`SCHEMA_VERSION` and executes all .sql files in that directory.
|
|
|
|
|
|
|
|
The function will then apply all deltas for all versions after the base
|
|
|
|
version.
|
|
|
|
|
|
|
|
Example directory structure:
|
|
|
|
|
|
|
|
schema/
|
|
|
|
delta/
|
|
|
|
...
|
|
|
|
full_schemas/
|
|
|
|
3/
|
|
|
|
test.sql
|
|
|
|
...
|
|
|
|
11/
|
|
|
|
foo.sql
|
|
|
|
bar.sql
|
|
|
|
...
|
|
|
|
|
|
|
|
In the example foo.sql and bar.sql would be run, and then any delta files
|
|
|
|
for versions strictly greater than 11.
|
2019-10-22 11:43:31 -06:00
|
|
|
|
|
|
|
Note: we apply the full schemas and deltas from the top level `schema/`
|
|
|
|
folder as well those in the data stores specified.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
cur (Cursor): a database cursor
|
|
|
|
database_engine (DatabaseEngine)
|
|
|
|
data_stores (list[str]): The names of the data stores to instantiate
|
|
|
|
on the given database.
|
2015-10-13 04:38:48 -06:00
|
|
|
"""
|
|
|
|
current_dir = os.path.join(dir_path, "schema", "full_schemas")
|
|
|
|
directory_entries = os.listdir(current_dir)
|
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
# First we find the highest full schema version we have
|
|
|
|
valid_versions = []
|
2019-06-03 06:02:47 -06:00
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
for filename in directory_entries:
|
2019-10-21 09:08:40 -06:00
|
|
|
try:
|
|
|
|
ver = int(filename)
|
|
|
|
except ValueError:
|
|
|
|
continue
|
2015-10-13 04:38:48 -06:00
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
if ver <= SCHEMA_VERSION:
|
|
|
|
valid_versions.append(ver)
|
|
|
|
|
|
|
|
if not valid_versions:
|
2015-10-13 04:38:48 -06:00
|
|
|
raise PrepareDatabaseException(
|
|
|
|
"Could not find a suitable base set of full schemas"
|
|
|
|
)
|
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
max_current_ver = max(valid_versions)
|
2015-10-13 04:38:48 -06:00
|
|
|
|
|
|
|
logger.debug("Initialising schema v%d", max_current_ver)
|
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
# Now lets find all the full schema files, both in the global schema and
|
|
|
|
# in data store schemas.
|
|
|
|
directories = [os.path.join(current_dir, str(max_current_ver))]
|
|
|
|
directories.extend(
|
|
|
|
os.path.join(
|
|
|
|
dir_path,
|
|
|
|
"data_stores",
|
|
|
|
data_store,
|
|
|
|
"schema",
|
|
|
|
"full_schemas",
|
|
|
|
str(max_current_ver),
|
|
|
|
)
|
|
|
|
for data_store in data_stores
|
|
|
|
)
|
|
|
|
|
|
|
|
directory_entries = []
|
|
|
|
for directory in directories:
|
|
|
|
directory_entries.extend(
|
|
|
|
_DirectoryListing(file_name, os.path.join(directory, file_name))
|
|
|
|
for file_name in os.listdir(directory)
|
|
|
|
)
|
|
|
|
|
|
|
|
if isinstance(database_engine, PostgresEngine):
|
|
|
|
specific = "postgres"
|
|
|
|
else:
|
|
|
|
specific = "sqlite"
|
2015-10-13 04:38:48 -06:00
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
directory_entries.sort()
|
|
|
|
for entry in directory_entries:
|
|
|
|
if entry.file_name.endswith(".sql") or entry.file_name.endswith(
|
|
|
|
".sql." + specific
|
|
|
|
):
|
|
|
|
logger.debug("Applying schema %s", entry.absolute_path)
|
|
|
|
executescript(cur, entry.absolute_path)
|
2015-10-13 04:38:48 -06:00
|
|
|
|
|
|
|
cur.execute(
|
|
|
|
database_engine.convert_param_style(
|
2019-10-21 09:08:40 -06:00
|
|
|
"INSERT INTO schema_version (version, upgraded) VALUES (?,?)"
|
2015-10-13 04:38:48 -06:00
|
|
|
),
|
2019-04-03 03:07:29 -06:00
|
|
|
(max_current_ver, False),
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
_upgrade_existing_database(
|
|
|
|
cur,
|
|
|
|
current_version=max_current_ver,
|
|
|
|
applied_delta_files=[],
|
|
|
|
upgraded=False,
|
|
|
|
database_engine=database_engine,
|
2016-04-06 07:08:18 -06:00
|
|
|
config=None,
|
2019-10-21 09:08:40 -06:00
|
|
|
data_stores=data_stores,
|
2016-04-06 07:08:18 -06:00
|
|
|
is_empty=True,
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2019-04-03 03:07:29 -06:00
|
|
|
def _upgrade_existing_database(
|
|
|
|
cur,
|
|
|
|
current_version,
|
|
|
|
applied_delta_files,
|
|
|
|
upgraded,
|
|
|
|
database_engine,
|
|
|
|
config,
|
2019-10-21 09:08:40 -06:00
|
|
|
data_stores,
|
2019-04-03 03:07:29 -06:00
|
|
|
is_empty=False,
|
|
|
|
):
|
2015-10-13 04:38:48 -06:00
|
|
|
"""Upgrades an existing database.
|
|
|
|
|
|
|
|
Delta files can either be SQL stored in *.sql files, or python modules
|
|
|
|
in *.py.
|
|
|
|
|
|
|
|
There can be multiple delta files per version. Synapse will keep track of
|
|
|
|
which delta files have been applied, and will apply any that haven't been
|
|
|
|
even if there has been no version bump. This is useful for development
|
|
|
|
where orthogonal schema changes may happen on separate branches.
|
|
|
|
|
|
|
|
Different delta files for the same version *must* be orthogonal and give
|
|
|
|
the same result when applied in any order. No guarantees are made on the
|
|
|
|
order of execution of these scripts.
|
|
|
|
|
|
|
|
This is a no-op of current_version == SCHEMA_VERSION.
|
|
|
|
|
|
|
|
Example directory structure:
|
|
|
|
|
|
|
|
schema/
|
|
|
|
delta/
|
|
|
|
11/
|
|
|
|
foo.sql
|
|
|
|
...
|
|
|
|
12/
|
|
|
|
foo.sql
|
|
|
|
bar.py
|
|
|
|
...
|
|
|
|
full_schemas/
|
|
|
|
...
|
|
|
|
|
|
|
|
In the example, if current_version is 11, then foo.sql will be run if and
|
|
|
|
only if `upgraded` is True. Then `foo.sql` and `bar.py` would be run in
|
|
|
|
some arbitrary order.
|
|
|
|
|
2019-10-22 11:43:31 -06:00
|
|
|
Note: we apply the delta files from the specified data stores as well as
|
|
|
|
those in the top-level schema. We apply all delta files across data stores
|
|
|
|
for a version before applying those in the next version.
|
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
Args:
|
|
|
|
cur (Cursor)
|
|
|
|
current_version (int): The current version of the schema.
|
|
|
|
applied_delta_files (list): A list of deltas that have already been
|
|
|
|
applied.
|
|
|
|
upgraded (bool): Whether the current version was generated by having
|
|
|
|
applied deltas or from full schema file. If `True` the function
|
|
|
|
will never apply delta files for the given `current_version`, since
|
|
|
|
the current_version wasn't generated by applying those delta files.
|
2019-10-22 11:43:31 -06:00
|
|
|
database_engine (DatabaseEngine)
|
|
|
|
config (synapse.config.homeserver.HomeServerConfig|None):
|
|
|
|
application config, or None if we are connecting to an existing
|
|
|
|
database which we expect to be configured already
|
|
|
|
data_stores (list[str]): The names of the data stores to instantiate
|
|
|
|
on the given database.
|
|
|
|
is_empty (bool): Is this a blank database? I.e. do we need to run the
|
|
|
|
upgrade portions of the delta scripts.
|
2015-10-13 04:38:48 -06:00
|
|
|
"""
|
|
|
|
|
|
|
|
if current_version > SCHEMA_VERSION:
|
|
|
|
raise ValueError(
|
2019-04-03 03:07:29 -06:00
|
|
|
"Cannot use this database as it is too "
|
|
|
|
+ "new for the server to understand"
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
start_ver = current_version
|
|
|
|
if not upgraded:
|
|
|
|
start_ver += 1
|
|
|
|
|
|
|
|
logger.debug("applied_delta_files: %s", applied_delta_files)
|
|
|
|
|
2019-08-27 01:52:20 -06:00
|
|
|
if isinstance(database_engine, PostgresEngine):
|
|
|
|
specific_engine_extension = ".postgres"
|
|
|
|
else:
|
|
|
|
specific_engine_extension = ".sqlite"
|
|
|
|
|
2019-08-27 02:39:11 -06:00
|
|
|
specific_engine_extensions = (".sqlite", ".postgres")
|
2019-08-27 01:52:20 -06:00
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
for v in range(start_ver, SCHEMA_VERSION + 1):
|
2016-02-09 07:44:12 -07:00
|
|
|
logger.info("Upgrading schema to v%d", v)
|
2015-10-13 04:38:48 -06:00
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
# We need to search both the global and per data store schema
|
|
|
|
# directories for schema updates.
|
2015-10-13 04:38:48 -06:00
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
# First we find the directories to search in
|
|
|
|
delta_dir = os.path.join(dir_path, "schema", "delta", str(v))
|
|
|
|
directories = [delta_dir]
|
|
|
|
for data_store in data_stores:
|
|
|
|
directories.append(
|
|
|
|
os.path.join(
|
|
|
|
dir_path, "data_stores", data_store, "schema", "delta", str(v)
|
|
|
|
)
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
|
|
|
|
2019-10-21 09:08:40 -06:00
|
|
|
# Now find which directories have anything of interest.
|
|
|
|
directory_entries = []
|
|
|
|
for directory in directories:
|
|
|
|
logger.debug("Looking for schema deltas in %s", directory)
|
|
|
|
try:
|
|
|
|
file_names = os.listdir(directory)
|
|
|
|
directory_entries.extend(
|
|
|
|
_DirectoryListing(file_name, os.path.join(directory, file_name))
|
|
|
|
for file_name in file_names
|
|
|
|
)
|
|
|
|
except FileNotFoundError:
|
|
|
|
# Data stores can have empty entries for a given version delta.
|
|
|
|
pass
|
|
|
|
except OSError:
|
|
|
|
raise UpgradeDatabaseException(
|
2019-10-22 11:43:31 -06:00
|
|
|
"Could not open delta dir for version %d: %s" % (v, directory)
|
2019-10-21 09:08:40 -06:00
|
|
|
)
|
|
|
|
|
2019-10-22 11:43:31 -06:00
|
|
|
# We sort to ensure that we apply the delta files in a consistent
|
|
|
|
# order (to avoid bugs caused by inconsistent directory listing order)
|
2015-10-13 04:38:48 -06:00
|
|
|
directory_entries.sort()
|
2019-10-21 09:08:40 -06:00
|
|
|
for entry in directory_entries:
|
|
|
|
file_name = entry.file_name
|
2015-10-13 04:38:48 -06:00
|
|
|
relative_path = os.path.join(str(v), file_name)
|
2019-10-21 09:08:40 -06:00
|
|
|
absolute_path = entry.absolute_path
|
|
|
|
|
2019-10-22 11:43:31 -06:00
|
|
|
logger.debug("Found file: %s (%s)", relative_path, absolute_path)
|
2015-10-13 04:38:48 -06:00
|
|
|
if relative_path in applied_delta_files:
|
|
|
|
continue
|
|
|
|
|
|
|
|
root_name, ext = os.path.splitext(file_name)
|
|
|
|
if ext == ".py":
|
|
|
|
# This is a python upgrade module. We need to import into some
|
|
|
|
# package and then execute its `run_upgrade` function.
|
2019-04-03 03:07:29 -06:00
|
|
|
module_name = "synapse.storage.v%d_%s" % (v, root_name)
|
2015-10-13 04:38:48 -06:00
|
|
|
with open(absolute_path) as python_file:
|
2019-04-03 03:07:29 -06:00
|
|
|
module = imp.load_source(module_name, absolute_path, python_file)
|
2016-09-08 06:40:46 -06:00
|
|
|
logger.info("Running script %s", relative_path)
|
2016-04-06 07:08:18 -06:00
|
|
|
module.run_create(cur, database_engine)
|
|
|
|
if not is_empty:
|
|
|
|
module.run_upgrade(cur, database_engine, config=config)
|
2018-11-20 15:46:51 -07:00
|
|
|
elif ext == ".pyc" or file_name == "__pycache__":
|
2015-10-13 04:38:48 -06:00
|
|
|
# Sometimes .pyc files turn up anyway even though we've
|
|
|
|
# disabled their generation; e.g. from distribution package
|
|
|
|
# installers. Silently skip it
|
2019-08-27 01:52:20 -06:00
|
|
|
continue
|
2015-10-13 04:38:48 -06:00
|
|
|
elif ext == ".sql":
|
|
|
|
# A plain old .sql file, just read and execute it
|
2016-09-08 06:40:46 -06:00
|
|
|
logger.info("Applying schema %s", relative_path)
|
2015-10-13 04:38:48 -06:00
|
|
|
executescript(cur, absolute_path)
|
2019-08-27 01:52:20 -06:00
|
|
|
elif ext == specific_engine_extension and root_name.endswith(".sql"):
|
|
|
|
# A .sql file specific to our engine; just read and execute it
|
|
|
|
logger.info("Applying engine-specific schema %s", relative_path)
|
|
|
|
executescript(cur, absolute_path)
|
|
|
|
elif ext in specific_engine_extensions and root_name.endswith(".sql"):
|
|
|
|
# A .sql file for a different engine; skip it.
|
|
|
|
continue
|
2015-10-13 04:38:48 -06:00
|
|
|
else:
|
|
|
|
# Not a valid delta file.
|
2019-08-27 01:52:20 -06:00
|
|
|
logger.warning(
|
|
|
|
"Found directory entry that did not end in .py or .sql: %s",
|
2015-10-13 04:38:48 -06:00
|
|
|
relative_path,
|
|
|
|
)
|
|
|
|
continue
|
|
|
|
|
|
|
|
# Mark as done.
|
|
|
|
cur.execute(
|
|
|
|
database_engine.convert_param_style(
|
2019-08-27 01:52:20 -06:00
|
|
|
"INSERT INTO applied_schema_deltas (version, file) VALUES (?,?)"
|
2015-10-13 04:38:48 -06:00
|
|
|
),
|
2019-04-03 03:07:29 -06:00
|
|
|
(v, relative_path),
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
cur.execute("DELETE FROM schema_version")
|
|
|
|
cur.execute(
|
|
|
|
database_engine.convert_param_style(
|
2019-08-27 01:52:20 -06:00
|
|
|
"INSERT INTO schema_version (version, upgraded) VALUES (?,?)"
|
2015-10-13 04:38:48 -06:00
|
|
|
),
|
2019-04-03 03:07:29 -06:00
|
|
|
(v, True),
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2017-10-30 09:16:21 -06:00
|
|
|
def _apply_module_schemas(txn, database_engine, config):
|
|
|
|
"""Apply the module schemas for the dynamic modules, if any
|
|
|
|
|
|
|
|
Args:
|
|
|
|
cur: database cursor
|
|
|
|
database_engine: synapse database engine class
|
|
|
|
config (synapse.config.homeserver.HomeServerConfig):
|
|
|
|
application config
|
|
|
|
"""
|
|
|
|
for (mod, _config) in config.password_providers:
|
2019-06-20 03:32:02 -06:00
|
|
|
if not hasattr(mod, "get_db_schema_files"):
|
2017-10-30 09:16:21 -06:00
|
|
|
continue
|
|
|
|
modname = ".".join((mod.__module__, mod.__name__))
|
|
|
|
_apply_module_schema_files(
|
2019-04-03 03:07:29 -06:00
|
|
|
txn, database_engine, modname, mod.get_db_schema_files()
|
2017-10-30 09:16:21 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def _apply_module_schema_files(cur, database_engine, modname, names_and_streams):
|
|
|
|
"""Apply the module schemas for a single module
|
|
|
|
|
|
|
|
Args:
|
|
|
|
cur: database cursor
|
|
|
|
database_engine: synapse database engine class
|
|
|
|
modname (str): fully qualified name of the module
|
|
|
|
names_and_streams (Iterable[(str, file)]): the names and streams of
|
|
|
|
schemas to be applied
|
|
|
|
"""
|
|
|
|
cur.execute(
|
|
|
|
database_engine.convert_param_style(
|
|
|
|
"SELECT file FROM applied_module_schemas WHERE module_name = ?"
|
|
|
|
),
|
2019-04-03 03:07:29 -06:00
|
|
|
(modname,),
|
2017-10-30 09:16:21 -06:00
|
|
|
)
|
|
|
|
applied_deltas = set(d for d, in cur)
|
|
|
|
for (name, stream) in names_and_streams:
|
|
|
|
if name in applied_deltas:
|
|
|
|
continue
|
|
|
|
|
|
|
|
root_name, ext = os.path.splitext(name)
|
2019-06-20 03:32:02 -06:00
|
|
|
if ext != ".sql":
|
2017-10-30 09:16:21 -06:00
|
|
|
raise PrepareDatabaseException(
|
2019-04-03 03:07:29 -06:00
|
|
|
"only .sql files are currently supported for module schemas"
|
2017-10-30 09:16:21 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
logger.info("applying schema %s for %s", name, modname)
|
|
|
|
for statement in get_statements(stream):
|
|
|
|
cur.execute(statement)
|
|
|
|
|
|
|
|
# Mark as done.
|
|
|
|
cur.execute(
|
|
|
|
database_engine.convert_param_style(
|
2019-11-21 05:00:14 -07:00
|
|
|
"INSERT INTO applied_module_schemas (module_name, file) VALUES (?,?)"
|
2017-10-30 09:16:21 -06:00
|
|
|
),
|
2019-04-03 03:07:29 -06:00
|
|
|
(modname, name),
|
2017-10-30 09:16:21 -06:00
|
|
|
)
|
|
|
|
|
|
|
|
|
2015-10-13 04:38:48 -06:00
|
|
|
def get_statements(f):
|
|
|
|
statement_buffer = ""
|
|
|
|
in_comment = False # If we're in a /* ... */ style comment
|
|
|
|
|
|
|
|
for line in f:
|
|
|
|
line = line.strip()
|
|
|
|
|
|
|
|
if in_comment:
|
|
|
|
# Check if this line contains an end to the comment
|
|
|
|
comments = line.split("*/", 1)
|
|
|
|
if len(comments) == 1:
|
|
|
|
continue
|
|
|
|
line = comments[1]
|
|
|
|
in_comment = False
|
|
|
|
|
|
|
|
# Remove inline block comments
|
|
|
|
line = re.sub(r"/\*.*\*/", " ", line)
|
|
|
|
|
|
|
|
# Does this line start a comment?
|
|
|
|
comments = line.split("/*", 1)
|
|
|
|
if len(comments) > 1:
|
|
|
|
line = comments[0]
|
|
|
|
in_comment = True
|
|
|
|
|
|
|
|
# Deal with line comments
|
|
|
|
line = line.split("--", 1)[0]
|
|
|
|
line = line.split("//", 1)[0]
|
|
|
|
|
|
|
|
# Find *all* semicolons. We need to treat first and last entry
|
|
|
|
# specially.
|
|
|
|
statements = line.split(";")
|
|
|
|
|
|
|
|
# We must prepend statement_buffer to the first statement
|
2019-04-03 03:07:29 -06:00
|
|
|
first_statement = "%s %s" % (statement_buffer.strip(), statements[0].strip())
|
2015-10-13 04:38:48 -06:00
|
|
|
statements[0] = first_statement
|
|
|
|
|
|
|
|
# Every entry, except the last, is a full statement
|
|
|
|
for statement in statements[:-1]:
|
|
|
|
yield statement.strip()
|
|
|
|
|
|
|
|
# The last entry did *not* end in a semicolon, so we store it for the
|
|
|
|
# next semicolon we find
|
|
|
|
statement_buffer = statements[-1].strip()
|
|
|
|
|
|
|
|
|
|
|
|
def executescript(txn, schema_path):
|
2019-06-20 03:32:02 -06:00
|
|
|
with open(schema_path, "r") as f:
|
2015-10-13 04:38:48 -06:00
|
|
|
for statement in get_statements(f):
|
|
|
|
txn.execute(statement)
|
|
|
|
|
|
|
|
|
|
|
|
def _get_or_create_schema_state(txn, database_engine):
|
|
|
|
# Bluntly try creating the schema_version tables.
|
2019-04-03 03:07:29 -06:00
|
|
|
schema_path = os.path.join(dir_path, "schema", "schema_version.sql")
|
2015-10-13 04:38:48 -06:00
|
|
|
executescript(txn, schema_path)
|
|
|
|
|
|
|
|
txn.execute("SELECT version, upgraded FROM schema_version")
|
|
|
|
row = txn.fetchone()
|
|
|
|
current_version = int(row[0]) if row else None
|
|
|
|
upgraded = bool(row[1]) if row else None
|
|
|
|
|
|
|
|
if current_version:
|
|
|
|
txn.execute(
|
|
|
|
database_engine.convert_param_style(
|
|
|
|
"SELECT file FROM applied_schema_deltas WHERE version >= ?"
|
|
|
|
),
|
2019-04-03 03:07:29 -06:00
|
|
|
(current_version,),
|
2015-10-13 04:38:48 -06:00
|
|
|
)
|
2017-03-23 11:53:49 -06:00
|
|
|
applied_deltas = [d for d, in txn]
|
2015-10-13 04:38:48 -06:00
|
|
|
return current_version, applied_deltas, upgraded
|
|
|
|
|
|
|
|
return None
|
2019-10-21 09:08:40 -06:00
|
|
|
|
|
|
|
|
|
|
|
@attr.s()
|
|
|
|
class _DirectoryListing(object):
|
|
|
|
"""Helper class to store schema file name and the
|
|
|
|
absolute path to it.
|
2019-10-22 11:43:31 -06:00
|
|
|
|
|
|
|
These entries get sorted, so for consistency we want to ensure that
|
|
|
|
`file_name` attr is kept first.
|
2019-10-21 09:08:40 -06:00
|
|
|
"""
|
|
|
|
|
|
|
|
file_name = attr.ib()
|
|
|
|
absolute_path = attr.ib()
|