mirror of https://github.com/morpheus65535/bazarr
1381 lines
48 KiB
Python
1381 lines
48 KiB
Python
from __future__ import annotations
|
|
|
|
from contextlib import contextmanager
|
|
from contextlib import nullcontext
|
|
import logging
|
|
import sys
|
|
from typing import Any
|
|
from typing import cast
|
|
from typing import Collection
|
|
from typing import ContextManager
|
|
from typing import Dict
|
|
from typing import Iterator
|
|
from typing import List
|
|
from typing import Optional
|
|
from typing import Set
|
|
from typing import Tuple
|
|
from typing import TYPE_CHECKING
|
|
from typing import Union
|
|
|
|
from sqlalchemy import Column
|
|
from sqlalchemy import literal_column
|
|
from sqlalchemy import MetaData
|
|
from sqlalchemy import PrimaryKeyConstraint
|
|
from sqlalchemy import String
|
|
from sqlalchemy import Table
|
|
from sqlalchemy.engine import Engine
|
|
from sqlalchemy.engine import url as sqla_url
|
|
from sqlalchemy.engine.strategies import MockEngineStrategy
|
|
|
|
from .. import ddl
|
|
from .. import util
|
|
from ..util import sqla_compat
|
|
from ..util.compat import EncodedIO
|
|
|
|
if TYPE_CHECKING:
|
|
from sqlalchemy.engine import Dialect
|
|
from sqlalchemy.engine import URL
|
|
from sqlalchemy.engine.base import Connection
|
|
from sqlalchemy.engine.base import Transaction
|
|
from sqlalchemy.engine.mock import MockConnection
|
|
from sqlalchemy.sql.elements import ClauseElement
|
|
|
|
from .environment import EnvironmentContext
|
|
from ..config import Config
|
|
from ..script.base import Script
|
|
from ..script.base import ScriptDirectory
|
|
from ..script.revision import _RevisionOrBase
|
|
from ..script.revision import Revision
|
|
from ..script.revision import RevisionMap
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
|
|
class _ProxyTransaction:
|
|
def __init__(self, migration_context: MigrationContext) -> None:
|
|
self.migration_context = migration_context
|
|
|
|
@property
|
|
def _proxied_transaction(self) -> Optional[Transaction]:
|
|
return self.migration_context._transaction
|
|
|
|
def rollback(self) -> None:
|
|
t = self._proxied_transaction
|
|
assert t is not None
|
|
t.rollback()
|
|
self.migration_context._transaction = None
|
|
|
|
def commit(self) -> None:
|
|
t = self._proxied_transaction
|
|
assert t is not None
|
|
t.commit()
|
|
self.migration_context._transaction = None
|
|
|
|
def __enter__(self) -> _ProxyTransaction:
|
|
return self
|
|
|
|
def __exit__(self, type_: None, value: None, traceback: None) -> None:
|
|
if self._proxied_transaction is not None:
|
|
self._proxied_transaction.__exit__(type_, value, traceback)
|
|
self.migration_context._transaction = None
|
|
|
|
|
|
class MigrationContext:
|
|
|
|
"""Represent the database state made available to a migration
|
|
script.
|
|
|
|
:class:`.MigrationContext` is the front end to an actual
|
|
database connection, or alternatively a string output
|
|
stream given a particular database dialect,
|
|
from an Alembic perspective.
|
|
|
|
When inside the ``env.py`` script, the :class:`.MigrationContext`
|
|
is available via the
|
|
:meth:`.EnvironmentContext.get_context` method,
|
|
which is available at ``alembic.context``::
|
|
|
|
# from within env.py script
|
|
from alembic import context
|
|
migration_context = context.get_context()
|
|
|
|
For usage outside of an ``env.py`` script, such as for
|
|
utility routines that want to check the current version
|
|
in the database, the :meth:`.MigrationContext.configure`
|
|
method to create new :class:`.MigrationContext` objects.
|
|
For example, to get at the current revision in the
|
|
database using :meth:`.MigrationContext.get_current_revision`::
|
|
|
|
# in any application, outside of an env.py script
|
|
from alembic.migration import MigrationContext
|
|
from sqlalchemy import create_engine
|
|
|
|
engine = create_engine("postgresql://mydatabase")
|
|
conn = engine.connect()
|
|
|
|
context = MigrationContext.configure(conn)
|
|
current_rev = context.get_current_revision()
|
|
|
|
The above context can also be used to produce
|
|
Alembic migration operations with an :class:`.Operations`
|
|
instance::
|
|
|
|
# in any application, outside of the normal Alembic environment
|
|
from alembic.operations import Operations
|
|
op = Operations(context)
|
|
op.alter_column("mytable", "somecolumn", nullable=True)
|
|
|
|
"""
|
|
|
|
def __init__(
|
|
self,
|
|
dialect: Dialect,
|
|
connection: Optional[Connection],
|
|
opts: Dict[str, Any],
|
|
environment_context: Optional[EnvironmentContext] = None,
|
|
) -> None:
|
|
self.environment_context = environment_context
|
|
self.opts = opts
|
|
self.dialect = dialect
|
|
self.script: Optional[ScriptDirectory] = opts.get("script")
|
|
as_sql: bool = opts.get("as_sql", False)
|
|
transactional_ddl = opts.get("transactional_ddl")
|
|
self._transaction_per_migration = opts.get(
|
|
"transaction_per_migration", False
|
|
)
|
|
self.on_version_apply_callbacks = opts.get("on_version_apply", ())
|
|
self._transaction: Optional[Transaction] = None
|
|
|
|
if as_sql:
|
|
self.connection = cast(
|
|
Optional["Connection"], self._stdout_connection(connection)
|
|
)
|
|
assert self.connection is not None
|
|
self._in_external_transaction = False
|
|
else:
|
|
self.connection = connection
|
|
self._in_external_transaction = (
|
|
sqla_compat._get_connection_in_transaction(connection)
|
|
)
|
|
|
|
self._migrations_fn = opts.get("fn")
|
|
self.as_sql = as_sql
|
|
|
|
self.purge = opts.get("purge", False)
|
|
|
|
if "output_encoding" in opts:
|
|
self.output_buffer = EncodedIO(
|
|
opts.get("output_buffer")
|
|
or sys.stdout, # type:ignore[arg-type]
|
|
opts["output_encoding"],
|
|
)
|
|
else:
|
|
self.output_buffer = opts.get("output_buffer", sys.stdout)
|
|
|
|
self._user_compare_type = opts.get("compare_type", False)
|
|
self._user_compare_server_default = opts.get(
|
|
"compare_server_default", False
|
|
)
|
|
self.version_table = version_table = opts.get(
|
|
"version_table", "alembic_version"
|
|
)
|
|
self.version_table_schema = version_table_schema = opts.get(
|
|
"version_table_schema", None
|
|
)
|
|
self._version = Table(
|
|
version_table,
|
|
MetaData(),
|
|
Column("version_num", String(32), nullable=False),
|
|
schema=version_table_schema,
|
|
)
|
|
if opts.get("version_table_pk", True):
|
|
self._version.append_constraint(
|
|
PrimaryKeyConstraint(
|
|
"version_num", name="%s_pkc" % version_table
|
|
)
|
|
)
|
|
|
|
self._start_from_rev: Optional[str] = opts.get("starting_rev")
|
|
self.impl = ddl.DefaultImpl.get_by_dialect(dialect)(
|
|
dialect,
|
|
self.connection,
|
|
self.as_sql,
|
|
transactional_ddl,
|
|
self.output_buffer,
|
|
opts,
|
|
)
|
|
log.info("Context impl %s.", self.impl.__class__.__name__)
|
|
if self.as_sql:
|
|
log.info("Generating static SQL")
|
|
log.info(
|
|
"Will assume %s DDL.",
|
|
"transactional"
|
|
if self.impl.transactional_ddl
|
|
else "non-transactional",
|
|
)
|
|
|
|
@classmethod
|
|
def configure(
|
|
cls,
|
|
connection: Optional[Connection] = None,
|
|
url: Optional[Union[str, URL]] = None,
|
|
dialect_name: Optional[str] = None,
|
|
dialect: Optional[Dialect] = None,
|
|
environment_context: Optional[EnvironmentContext] = None,
|
|
dialect_opts: Optional[Dict[str, str]] = None,
|
|
opts: Optional[Any] = None,
|
|
) -> MigrationContext:
|
|
"""Create a new :class:`.MigrationContext`.
|
|
|
|
This is a factory method usually called
|
|
by :meth:`.EnvironmentContext.configure`.
|
|
|
|
:param connection: a :class:`~sqlalchemy.engine.Connection`
|
|
to use for SQL execution in "online" mode. When present,
|
|
is also used to determine the type of dialect in use.
|
|
:param url: a string database url, or a
|
|
:class:`sqlalchemy.engine.url.URL` object.
|
|
The type of dialect to be used will be derived from this if
|
|
``connection`` is not passed.
|
|
:param dialect_name: string name of a dialect, such as
|
|
"postgresql", "mssql", etc. The type of dialect to be used will be
|
|
derived from this if ``connection`` and ``url`` are not passed.
|
|
:param opts: dictionary of options. Most other options
|
|
accepted by :meth:`.EnvironmentContext.configure` are passed via
|
|
this dictionary.
|
|
|
|
"""
|
|
if opts is None:
|
|
opts = {}
|
|
if dialect_opts is None:
|
|
dialect_opts = {}
|
|
|
|
if connection:
|
|
if isinstance(connection, Engine):
|
|
raise util.CommandError(
|
|
"'connection' argument to configure() is expected "
|
|
"to be a sqlalchemy.engine.Connection instance, "
|
|
"got %r" % connection,
|
|
)
|
|
|
|
dialect = connection.dialect
|
|
elif url:
|
|
url_obj = sqla_url.make_url(url)
|
|
dialect = url_obj.get_dialect()(**dialect_opts)
|
|
elif dialect_name:
|
|
url_obj = sqla_url.make_url("%s://" % dialect_name)
|
|
dialect = url_obj.get_dialect()(**dialect_opts)
|
|
elif not dialect:
|
|
raise Exception("Connection, url, or dialect_name is required.")
|
|
assert dialect is not None
|
|
return MigrationContext(dialect, connection, opts, environment_context)
|
|
|
|
@contextmanager
|
|
def autocommit_block(self) -> Iterator[None]:
|
|
"""Enter an "autocommit" block, for databases that support AUTOCOMMIT
|
|
isolation levels.
|
|
|
|
This special directive is intended to support the occasional database
|
|
DDL or system operation that specifically has to be run outside of
|
|
any kind of transaction block. The PostgreSQL database platform
|
|
is the most common target for this style of operation, as many
|
|
of its DDL operations must be run outside of transaction blocks, even
|
|
though the database overall supports transactional DDL.
|
|
|
|
The method is used as a context manager within a migration script, by
|
|
calling on :meth:`.Operations.get_context` to retrieve the
|
|
:class:`.MigrationContext`, then invoking
|
|
:meth:`.MigrationContext.autocommit_block` using the ``with:``
|
|
statement::
|
|
|
|
def upgrade():
|
|
with op.get_context().autocommit_block():
|
|
op.execute("ALTER TYPE mood ADD VALUE 'soso'")
|
|
|
|
Above, a PostgreSQL "ALTER TYPE..ADD VALUE" directive is emitted,
|
|
which must be run outside of a transaction block at the database level.
|
|
The :meth:`.MigrationContext.autocommit_block` method makes use of the
|
|
SQLAlchemy ``AUTOCOMMIT`` isolation level setting, which against the
|
|
psycogp2 DBAPI corresponds to the ``connection.autocommit`` setting,
|
|
to ensure that the database driver is not inside of a DBAPI level
|
|
transaction block.
|
|
|
|
.. warning::
|
|
|
|
As is necessary, **the database transaction preceding the block is
|
|
unconditionally committed**. This means that the run of migrations
|
|
preceding the operation will be committed, before the overall
|
|
migration operation is complete.
|
|
|
|
It is recommended that when an application includes migrations with
|
|
"autocommit" blocks, that
|
|
:paramref:`.EnvironmentContext.transaction_per_migration` be used
|
|
so that the calling environment is tuned to expect short per-file
|
|
migrations whether or not one of them has an autocommit block.
|
|
|
|
|
|
.. versionadded:: 1.2.0
|
|
|
|
"""
|
|
_in_connection_transaction = self._in_connection_transaction()
|
|
|
|
if self.impl.transactional_ddl and self.as_sql:
|
|
self.impl.emit_commit()
|
|
|
|
elif _in_connection_transaction:
|
|
assert self._transaction is not None
|
|
|
|
self._transaction.commit()
|
|
self._transaction = None
|
|
|
|
if not self.as_sql:
|
|
assert self.connection is not None
|
|
current_level = self.connection.get_isolation_level()
|
|
base_connection = self.connection
|
|
|
|
# in 1.3 and 1.4 non-future mode, the connection gets switched
|
|
# out. we can use the base connection with the new mode
|
|
# except that it will not know it's in "autocommit" and will
|
|
# emit deprecation warnings when an autocommit action takes
|
|
# place.
|
|
self.connection = (
|
|
self.impl.connection
|
|
) = base_connection.execution_options(isolation_level="AUTOCOMMIT")
|
|
|
|
# sqlalchemy future mode will "autobegin" in any case, so take
|
|
# control of that "transaction" here
|
|
fake_trans: Optional[Transaction] = self.connection.begin()
|
|
else:
|
|
fake_trans = None
|
|
try:
|
|
yield
|
|
finally:
|
|
if not self.as_sql:
|
|
assert self.connection is not None
|
|
if fake_trans is not None:
|
|
fake_trans.commit()
|
|
self.connection.execution_options(
|
|
isolation_level=current_level
|
|
)
|
|
self.connection = self.impl.connection = base_connection
|
|
|
|
if self.impl.transactional_ddl and self.as_sql:
|
|
self.impl.emit_begin()
|
|
|
|
elif _in_connection_transaction:
|
|
assert self.connection is not None
|
|
self._transaction = self.connection.begin()
|
|
|
|
def begin_transaction(
|
|
self, _per_migration: bool = False
|
|
) -> Union[_ProxyTransaction, ContextManager[None]]:
|
|
"""Begin a logical transaction for migration operations.
|
|
|
|
This method is used within an ``env.py`` script to demarcate where
|
|
the outer "transaction" for a series of migrations begins. Example::
|
|
|
|
def run_migrations_online():
|
|
connectable = create_engine(...)
|
|
|
|
with connectable.connect() as connection:
|
|
context.configure(
|
|
connection=connection, target_metadata=target_metadata
|
|
)
|
|
|
|
with context.begin_transaction():
|
|
context.run_migrations()
|
|
|
|
Above, :meth:`.MigrationContext.begin_transaction` is used to demarcate
|
|
where the outer logical transaction occurs around the
|
|
:meth:`.MigrationContext.run_migrations` operation.
|
|
|
|
A "Logical" transaction means that the operation may or may not
|
|
correspond to a real database transaction. If the target database
|
|
supports transactional DDL (or
|
|
:paramref:`.EnvironmentContext.configure.transactional_ddl` is true),
|
|
the :paramref:`.EnvironmentContext.configure.transaction_per_migration`
|
|
flag is not set, and the migration is against a real database
|
|
connection (as opposed to using "offline" ``--sql`` mode), a real
|
|
transaction will be started. If ``--sql`` mode is in effect, the
|
|
operation would instead correspond to a string such as "BEGIN" being
|
|
emitted to the string output.
|
|
|
|
The returned object is a Python context manager that should only be
|
|
used in the context of a ``with:`` statement as indicated above.
|
|
The object has no other guaranteed API features present.
|
|
|
|
.. seealso::
|
|
|
|
:meth:`.MigrationContext.autocommit_block`
|
|
|
|
"""
|
|
|
|
if self._in_external_transaction:
|
|
return nullcontext()
|
|
|
|
if self.impl.transactional_ddl:
|
|
transaction_now = _per_migration == self._transaction_per_migration
|
|
else:
|
|
transaction_now = _per_migration is True
|
|
|
|
if not transaction_now:
|
|
return nullcontext()
|
|
|
|
elif not self.impl.transactional_ddl:
|
|
assert _per_migration
|
|
|
|
if self.as_sql:
|
|
return nullcontext()
|
|
else:
|
|
# track our own notion of a "transaction block", which must be
|
|
# committed when complete. Don't rely upon whether or not the
|
|
# SQLAlchemy connection reports as "in transaction"; this
|
|
# because SQLAlchemy future connection features autobegin
|
|
# behavior, so it may already be in a transaction from our
|
|
# emitting of queries like "has_version_table", etc. While we
|
|
# could track these operations as well, that leaves open the
|
|
# possibility of new operations or other things happening in
|
|
# the user environment that still may be triggering
|
|
# "autobegin".
|
|
|
|
in_transaction = self._transaction is not None
|
|
|
|
if in_transaction:
|
|
return nullcontext()
|
|
else:
|
|
assert self.connection is not None
|
|
self._transaction = (
|
|
sqla_compat._safe_begin_connection_transaction(
|
|
self.connection
|
|
)
|
|
)
|
|
return _ProxyTransaction(self)
|
|
elif self.as_sql:
|
|
|
|
@contextmanager
|
|
def begin_commit():
|
|
self.impl.emit_begin()
|
|
yield
|
|
self.impl.emit_commit()
|
|
|
|
return begin_commit()
|
|
else:
|
|
assert self.connection is not None
|
|
self._transaction = sqla_compat._safe_begin_connection_transaction(
|
|
self.connection
|
|
)
|
|
return _ProxyTransaction(self)
|
|
|
|
def get_current_revision(self) -> Optional[str]:
|
|
"""Return the current revision, usually that which is present
|
|
in the ``alembic_version`` table in the database.
|
|
|
|
This method intends to be used only for a migration stream that
|
|
does not contain unmerged branches in the target database;
|
|
if there are multiple branches present, an exception is raised.
|
|
The :meth:`.MigrationContext.get_current_heads` should be preferred
|
|
over this method going forward in order to be compatible with
|
|
branch migration support.
|
|
|
|
If this :class:`.MigrationContext` was configured in "offline"
|
|
mode, that is with ``as_sql=True``, the ``starting_rev``
|
|
parameter is returned instead, if any.
|
|
|
|
"""
|
|
heads = self.get_current_heads()
|
|
if len(heads) == 0:
|
|
return None
|
|
elif len(heads) > 1:
|
|
raise util.CommandError(
|
|
"Version table '%s' has more than one head present; "
|
|
"please use get_current_heads()" % self.version_table
|
|
)
|
|
else:
|
|
return heads[0]
|
|
|
|
def get_current_heads(self) -> Tuple[str, ...]:
|
|
"""Return a tuple of the current 'head versions' that are represented
|
|
in the target database.
|
|
|
|
For a migration stream without branches, this will be a single
|
|
value, synonymous with that of
|
|
:meth:`.MigrationContext.get_current_revision`. However when multiple
|
|
unmerged branches exist within the target database, the returned tuple
|
|
will contain a value for each head.
|
|
|
|
If this :class:`.MigrationContext` was configured in "offline"
|
|
mode, that is with ``as_sql=True``, the ``starting_rev``
|
|
parameter is returned in a one-length tuple.
|
|
|
|
If no version table is present, or if there are no revisions
|
|
present, an empty tuple is returned.
|
|
|
|
"""
|
|
if self.as_sql:
|
|
start_from_rev: Any = self._start_from_rev
|
|
if start_from_rev == "base":
|
|
start_from_rev = None
|
|
elif start_from_rev is not None and self.script:
|
|
|
|
start_from_rev = [
|
|
cast("Script", self.script.get_revision(sfr)).revision
|
|
for sfr in util.to_list(start_from_rev)
|
|
if sfr not in (None, "base")
|
|
]
|
|
return util.to_tuple(start_from_rev, default=())
|
|
else:
|
|
if self._start_from_rev:
|
|
raise util.CommandError(
|
|
"Can't specify current_rev to context "
|
|
"when using a database connection"
|
|
)
|
|
if not self._has_version_table():
|
|
return ()
|
|
assert self.connection is not None
|
|
return tuple(
|
|
row[0] for row in self.connection.execute(self._version.select())
|
|
)
|
|
|
|
def _ensure_version_table(self, purge: bool = False) -> None:
|
|
with sqla_compat._ensure_scope_for_ddl(self.connection):
|
|
assert self.connection is not None
|
|
self._version.create(self.connection, checkfirst=True)
|
|
if purge:
|
|
assert self.connection is not None
|
|
self.connection.execute(self._version.delete())
|
|
|
|
def _has_version_table(self) -> bool:
|
|
assert self.connection is not None
|
|
return sqla_compat._connectable_has_table(
|
|
self.connection, self.version_table, self.version_table_schema
|
|
)
|
|
|
|
def stamp(self, script_directory: ScriptDirectory, revision: str) -> None:
|
|
"""Stamp the version table with a specific revision.
|
|
|
|
This method calculates those branches to which the given revision
|
|
can apply, and updates those branches as though they were migrated
|
|
towards that revision (either up or down). If no current branches
|
|
include the revision, it is added as a new branch head.
|
|
|
|
"""
|
|
heads = self.get_current_heads()
|
|
if not self.as_sql and not heads:
|
|
self._ensure_version_table()
|
|
head_maintainer = HeadMaintainer(self, heads)
|
|
for step in script_directory._stamp_revs(revision, heads):
|
|
head_maintainer.update_to_step(step)
|
|
|
|
def run_migrations(self, **kw: Any) -> None:
|
|
r"""Run the migration scripts established for this
|
|
:class:`.MigrationContext`, if any.
|
|
|
|
The commands in :mod:`alembic.command` will set up a function
|
|
that is ultimately passed to the :class:`.MigrationContext`
|
|
as the ``fn`` argument. This function represents the "work"
|
|
that will be done when :meth:`.MigrationContext.run_migrations`
|
|
is called, typically from within the ``env.py`` script of the
|
|
migration environment. The "work function" then provides an iterable
|
|
of version callables and other version information which
|
|
in the case of the ``upgrade`` or ``downgrade`` commands are the
|
|
list of version scripts to invoke. Other commands yield nothing,
|
|
in the case that a command wants to run some other operation
|
|
against the database such as the ``current`` or ``stamp`` commands.
|
|
|
|
:param \**kw: keyword arguments here will be passed to each
|
|
migration callable, that is the ``upgrade()`` or ``downgrade()``
|
|
method within revision scripts.
|
|
|
|
"""
|
|
self.impl.start_migrations()
|
|
|
|
heads: Tuple[str, ...]
|
|
if self.purge:
|
|
if self.as_sql:
|
|
raise util.CommandError("Can't use --purge with --sql mode")
|
|
self._ensure_version_table(purge=True)
|
|
heads = ()
|
|
else:
|
|
heads = self.get_current_heads()
|
|
|
|
dont_mutate = self.opts.get("dont_mutate", False)
|
|
|
|
if not self.as_sql and not heads and not dont_mutate:
|
|
self._ensure_version_table()
|
|
|
|
head_maintainer = HeadMaintainer(self, heads)
|
|
|
|
assert self._migrations_fn is not None
|
|
for step in self._migrations_fn(heads, self):
|
|
with self.begin_transaction(_per_migration=True):
|
|
|
|
if self.as_sql and not head_maintainer.heads:
|
|
# for offline mode, include a CREATE TABLE from
|
|
# the base
|
|
assert self.connection is not None
|
|
self._version.create(self.connection)
|
|
log.info("Running %s", step)
|
|
if self.as_sql:
|
|
self.impl.static_output(
|
|
"-- Running %s" % (step.short_log,)
|
|
)
|
|
step.migration_fn(**kw)
|
|
|
|
# previously, we wouldn't stamp per migration
|
|
# if we were in a transaction, however given the more
|
|
# complex model that involves any number of inserts
|
|
# and row-targeted updates and deletes, it's simpler for now
|
|
# just to run the operations on every version
|
|
head_maintainer.update_to_step(step)
|
|
for callback in self.on_version_apply_callbacks:
|
|
callback(
|
|
ctx=self,
|
|
step=step.info,
|
|
heads=set(head_maintainer.heads),
|
|
run_args=kw,
|
|
)
|
|
|
|
if self.as_sql and not head_maintainer.heads:
|
|
assert self.connection is not None
|
|
self._version.drop(self.connection)
|
|
|
|
def _in_connection_transaction(self) -> bool:
|
|
try:
|
|
meth = self.connection.in_transaction # type:ignore[union-attr]
|
|
except AttributeError:
|
|
return False
|
|
else:
|
|
return meth()
|
|
|
|
def execute(
|
|
self,
|
|
sql: Union[ClauseElement, str],
|
|
execution_options: Optional[dict] = None,
|
|
) -> None:
|
|
"""Execute a SQL construct or string statement.
|
|
|
|
The underlying execution mechanics are used, that is
|
|
if this is "offline mode" the SQL is written to the
|
|
output buffer, otherwise the SQL is emitted on
|
|
the current SQLAlchemy connection.
|
|
|
|
"""
|
|
self.impl._exec(sql, execution_options)
|
|
|
|
def _stdout_connection(
|
|
self, connection: Optional[Connection]
|
|
) -> MockConnection:
|
|
def dump(construct, *multiparams, **params):
|
|
self.impl._exec(construct)
|
|
|
|
return MockEngineStrategy.MockConnection(self.dialect, dump)
|
|
|
|
@property
|
|
def bind(self) -> Optional[Connection]:
|
|
"""Return the current "bind".
|
|
|
|
In online mode, this is an instance of
|
|
:class:`sqlalchemy.engine.Connection`, and is suitable
|
|
for ad-hoc execution of any kind of usage described
|
|
in :ref:`sqlexpression_toplevel` as well as
|
|
for usage with the :meth:`sqlalchemy.schema.Table.create`
|
|
and :meth:`sqlalchemy.schema.MetaData.create_all` methods
|
|
of :class:`~sqlalchemy.schema.Table`,
|
|
:class:`~sqlalchemy.schema.MetaData`.
|
|
|
|
Note that when "standard output" mode is enabled,
|
|
this bind will be a "mock" connection handler that cannot
|
|
return results and is only appropriate for a very limited
|
|
subset of commands.
|
|
|
|
"""
|
|
return self.connection
|
|
|
|
@property
|
|
def config(self) -> Optional[Config]:
|
|
"""Return the :class:`.Config` used by the current environment,
|
|
if any."""
|
|
|
|
if self.environment_context:
|
|
return self.environment_context.config
|
|
else:
|
|
return None
|
|
|
|
def _compare_type(
|
|
self, inspector_column: Column, metadata_column: Column
|
|
) -> bool:
|
|
if self._user_compare_type is False:
|
|
return False
|
|
|
|
if callable(self._user_compare_type):
|
|
user_value = self._user_compare_type(
|
|
self,
|
|
inspector_column,
|
|
metadata_column,
|
|
inspector_column.type,
|
|
metadata_column.type,
|
|
)
|
|
if user_value is not None:
|
|
return user_value
|
|
|
|
return self.impl.compare_type(inspector_column, metadata_column)
|
|
|
|
def _compare_server_default(
|
|
self,
|
|
inspector_column: Column,
|
|
metadata_column: Column,
|
|
rendered_metadata_default: Optional[str],
|
|
rendered_column_default: Optional[str],
|
|
) -> bool:
|
|
|
|
if self._user_compare_server_default is False:
|
|
return False
|
|
|
|
if callable(self._user_compare_server_default):
|
|
user_value = self._user_compare_server_default(
|
|
self,
|
|
inspector_column,
|
|
metadata_column,
|
|
rendered_column_default,
|
|
metadata_column.server_default,
|
|
rendered_metadata_default,
|
|
)
|
|
if user_value is not None:
|
|
return user_value
|
|
|
|
return self.impl.compare_server_default(
|
|
inspector_column,
|
|
metadata_column,
|
|
rendered_metadata_default,
|
|
rendered_column_default,
|
|
)
|
|
|
|
|
|
class HeadMaintainer:
|
|
def __init__(self, context: MigrationContext, heads: Any) -> None:
|
|
self.context = context
|
|
self.heads = set(heads)
|
|
|
|
def _insert_version(self, version: str) -> None:
|
|
assert version not in self.heads
|
|
self.heads.add(version)
|
|
|
|
self.context.impl._exec(
|
|
self.context._version.insert().values(
|
|
version_num=literal_column("'%s'" % version)
|
|
)
|
|
)
|
|
|
|
def _delete_version(self, version: str) -> None:
|
|
self.heads.remove(version)
|
|
|
|
ret = self.context.impl._exec(
|
|
self.context._version.delete().where(
|
|
self.context._version.c.version_num
|
|
== literal_column("'%s'" % version)
|
|
)
|
|
)
|
|
|
|
if (
|
|
not self.context.as_sql
|
|
and self.context.dialect.supports_sane_rowcount
|
|
and ret is not None
|
|
and ret.rowcount != 1
|
|
):
|
|
raise util.CommandError(
|
|
"Online migration expected to match one "
|
|
"row when deleting '%s' in '%s'; "
|
|
"%d found"
|
|
% (version, self.context.version_table, ret.rowcount)
|
|
)
|
|
|
|
def _update_version(self, from_: str, to_: str) -> None:
|
|
assert to_ not in self.heads
|
|
self.heads.remove(from_)
|
|
self.heads.add(to_)
|
|
|
|
ret = self.context.impl._exec(
|
|
self.context._version.update()
|
|
.values(version_num=literal_column("'%s'" % to_))
|
|
.where(
|
|
self.context._version.c.version_num
|
|
== literal_column("'%s'" % from_)
|
|
)
|
|
)
|
|
|
|
if (
|
|
not self.context.as_sql
|
|
and self.context.dialect.supports_sane_rowcount
|
|
and ret is not None
|
|
and ret.rowcount != 1
|
|
):
|
|
raise util.CommandError(
|
|
"Online migration expected to match one "
|
|
"row when updating '%s' to '%s' in '%s'; "
|
|
"%d found"
|
|
% (from_, to_, self.context.version_table, ret.rowcount)
|
|
)
|
|
|
|
def update_to_step(self, step: Union[RevisionStep, StampStep]) -> None:
|
|
if step.should_delete_branch(self.heads):
|
|
vers = step.delete_version_num
|
|
log.debug("branch delete %s", vers)
|
|
self._delete_version(vers)
|
|
elif step.should_create_branch(self.heads):
|
|
vers = step.insert_version_num
|
|
log.debug("new branch insert %s", vers)
|
|
self._insert_version(vers)
|
|
elif step.should_merge_branches(self.heads):
|
|
# delete revs, update from rev, update to rev
|
|
(
|
|
delete_revs,
|
|
update_from_rev,
|
|
update_to_rev,
|
|
) = step.merge_branch_idents(self.heads)
|
|
log.debug(
|
|
"merge, delete %s, update %s to %s",
|
|
delete_revs,
|
|
update_from_rev,
|
|
update_to_rev,
|
|
)
|
|
for delrev in delete_revs:
|
|
self._delete_version(delrev)
|
|
self._update_version(update_from_rev, update_to_rev)
|
|
elif step.should_unmerge_branches(self.heads):
|
|
(
|
|
update_from_rev,
|
|
update_to_rev,
|
|
insert_revs,
|
|
) = step.unmerge_branch_idents(self.heads)
|
|
log.debug(
|
|
"unmerge, insert %s, update %s to %s",
|
|
insert_revs,
|
|
update_from_rev,
|
|
update_to_rev,
|
|
)
|
|
for insrev in insert_revs:
|
|
self._insert_version(insrev)
|
|
self._update_version(update_from_rev, update_to_rev)
|
|
else:
|
|
from_, to_ = step.update_version_num(self.heads)
|
|
log.debug("update %s to %s", from_, to_)
|
|
self._update_version(from_, to_)
|
|
|
|
|
|
class MigrationInfo:
|
|
"""Exposes information about a migration step to a callback listener.
|
|
|
|
The :class:`.MigrationInfo` object is available exclusively for the
|
|
benefit of the :paramref:`.EnvironmentContext.on_version_apply`
|
|
callback hook.
|
|
|
|
"""
|
|
|
|
is_upgrade: bool
|
|
"""True/False: indicates whether this operation ascends or descends the
|
|
version tree."""
|
|
|
|
is_stamp: bool
|
|
"""True/False: indicates whether this operation is a stamp (i.e. whether
|
|
it results in any actual database operations)."""
|
|
|
|
up_revision_id: Optional[str]
|
|
"""Version string corresponding to :attr:`.Revision.revision`.
|
|
|
|
In the case of a stamp operation, it is advised to use the
|
|
:attr:`.MigrationInfo.up_revision_ids` tuple as a stamp operation can
|
|
make a single movement from one or more branches down to a single
|
|
branchpoint, in which case there will be multiple "up" revisions.
|
|
|
|
.. seealso::
|
|
|
|
:attr:`.MigrationInfo.up_revision_ids`
|
|
|
|
"""
|
|
|
|
up_revision_ids: Tuple[str, ...]
|
|
"""Tuple of version strings corresponding to :attr:`.Revision.revision`.
|
|
|
|
In the majority of cases, this tuple will be a single value, synonymous
|
|
with the scalar value of :attr:`.MigrationInfo.up_revision_id`.
|
|
It can be multiple revision identifiers only in the case of an
|
|
``alembic stamp`` operation which is moving downwards from multiple
|
|
branches down to their common branch point.
|
|
|
|
"""
|
|
|
|
down_revision_ids: Tuple[str, ...]
|
|
"""Tuple of strings representing the base revisions of this migration step.
|
|
|
|
If empty, this represents a root revision; otherwise, the first item
|
|
corresponds to :attr:`.Revision.down_revision`, and the rest are inferred
|
|
from dependencies.
|
|
"""
|
|
|
|
revision_map: RevisionMap
|
|
"""The revision map inside of which this operation occurs."""
|
|
|
|
def __init__(
|
|
self,
|
|
revision_map: RevisionMap,
|
|
is_upgrade: bool,
|
|
is_stamp: bool,
|
|
up_revisions: Union[str, Tuple[str, ...]],
|
|
down_revisions: Union[str, Tuple[str, ...]],
|
|
) -> None:
|
|
self.revision_map = revision_map
|
|
self.is_upgrade = is_upgrade
|
|
self.is_stamp = is_stamp
|
|
self.up_revision_ids = util.to_tuple(up_revisions, default=())
|
|
if self.up_revision_ids:
|
|
self.up_revision_id = self.up_revision_ids[0]
|
|
else:
|
|
# this should never be the case with
|
|
# "upgrade", "downgrade", or "stamp" as we are always
|
|
# measuring movement in terms of at least one upgrade version
|
|
self.up_revision_id = None
|
|
self.down_revision_ids = util.to_tuple(down_revisions, default=())
|
|
|
|
@property
|
|
def is_migration(self) -> bool:
|
|
"""True/False: indicates whether this operation is a migration.
|
|
|
|
At present this is true if and only the migration is not a stamp.
|
|
If other operation types are added in the future, both this attribute
|
|
and :attr:`~.MigrationInfo.is_stamp` will be false.
|
|
"""
|
|
return not self.is_stamp
|
|
|
|
@property
|
|
def source_revision_ids(self) -> Tuple[str, ...]:
|
|
"""Active revisions before this migration step is applied."""
|
|
return (
|
|
self.down_revision_ids if self.is_upgrade else self.up_revision_ids
|
|
)
|
|
|
|
@property
|
|
def destination_revision_ids(self) -> Tuple[str, ...]:
|
|
"""Active revisions after this migration step is applied."""
|
|
return (
|
|
self.up_revision_ids if self.is_upgrade else self.down_revision_ids
|
|
)
|
|
|
|
@property
|
|
def up_revision(self) -> Optional[Revision]:
|
|
"""Get :attr:`~.MigrationInfo.up_revision_id` as
|
|
a :class:`.Revision`.
|
|
|
|
"""
|
|
return self.revision_map.get_revision(self.up_revision_id)
|
|
|
|
@property
|
|
def up_revisions(self) -> Tuple[Optional[_RevisionOrBase], ...]:
|
|
"""Get :attr:`~.MigrationInfo.up_revision_ids` as a
|
|
:class:`.Revision`."""
|
|
return self.revision_map.get_revisions(self.up_revision_ids)
|
|
|
|
@property
|
|
def down_revisions(self) -> Tuple[Optional[_RevisionOrBase], ...]:
|
|
"""Get :attr:`~.MigrationInfo.down_revision_ids` as a tuple of
|
|
:class:`Revisions <.Revision>`."""
|
|
return self.revision_map.get_revisions(self.down_revision_ids)
|
|
|
|
@property
|
|
def source_revisions(self) -> Tuple[Optional[_RevisionOrBase], ...]:
|
|
"""Get :attr:`~MigrationInfo.source_revision_ids` as a tuple of
|
|
:class:`Revisions <.Revision>`."""
|
|
return self.revision_map.get_revisions(self.source_revision_ids)
|
|
|
|
@property
|
|
def destination_revisions(self) -> Tuple[Optional[_RevisionOrBase], ...]:
|
|
"""Get :attr:`~MigrationInfo.destination_revision_ids` as a tuple of
|
|
:class:`Revisions <.Revision>`."""
|
|
return self.revision_map.get_revisions(self.destination_revision_ids)
|
|
|
|
|
|
class MigrationStep:
|
|
|
|
from_revisions_no_deps: Tuple[str, ...]
|
|
to_revisions_no_deps: Tuple[str, ...]
|
|
is_upgrade: bool
|
|
migration_fn: Any
|
|
|
|
@property
|
|
def name(self) -> str:
|
|
return self.migration_fn.__name__
|
|
|
|
@classmethod
|
|
def upgrade_from_script(
|
|
cls, revision_map: RevisionMap, script: Script
|
|
) -> RevisionStep:
|
|
return RevisionStep(revision_map, script, True)
|
|
|
|
@classmethod
|
|
def downgrade_from_script(
|
|
cls, revision_map: RevisionMap, script: Script
|
|
) -> RevisionStep:
|
|
return RevisionStep(revision_map, script, False)
|
|
|
|
@property
|
|
def is_downgrade(self) -> bool:
|
|
return not self.is_upgrade
|
|
|
|
@property
|
|
def short_log(self) -> str:
|
|
return "%s %s -> %s" % (
|
|
self.name,
|
|
util.format_as_comma(self.from_revisions_no_deps),
|
|
util.format_as_comma(self.to_revisions_no_deps),
|
|
)
|
|
|
|
def __str__(self):
|
|
if self.doc:
|
|
return "%s %s -> %s, %s" % (
|
|
self.name,
|
|
util.format_as_comma(self.from_revisions_no_deps),
|
|
util.format_as_comma(self.to_revisions_no_deps),
|
|
self.doc,
|
|
)
|
|
else:
|
|
return self.short_log
|
|
|
|
|
|
class RevisionStep(MigrationStep):
|
|
def __init__(
|
|
self, revision_map: RevisionMap, revision: Script, is_upgrade: bool
|
|
) -> None:
|
|
self.revision_map = revision_map
|
|
self.revision = revision
|
|
self.is_upgrade = is_upgrade
|
|
if is_upgrade:
|
|
self.migration_fn = (
|
|
revision.module.upgrade # type:ignore[attr-defined]
|
|
)
|
|
else:
|
|
self.migration_fn = (
|
|
revision.module.downgrade # type:ignore[attr-defined]
|
|
)
|
|
|
|
def __repr__(self):
|
|
return "RevisionStep(%r, is_upgrade=%r)" % (
|
|
self.revision.revision,
|
|
self.is_upgrade,
|
|
)
|
|
|
|
def __eq__(self, other: object) -> bool:
|
|
return (
|
|
isinstance(other, RevisionStep)
|
|
and other.revision == self.revision
|
|
and self.is_upgrade == other.is_upgrade
|
|
)
|
|
|
|
@property
|
|
def doc(self) -> str:
|
|
return self.revision.doc
|
|
|
|
@property
|
|
def from_revisions(self) -> Tuple[str, ...]:
|
|
if self.is_upgrade:
|
|
return self.revision._normalized_down_revisions
|
|
else:
|
|
return (self.revision.revision,)
|
|
|
|
@property
|
|
def from_revisions_no_deps( # type:ignore[override]
|
|
self,
|
|
) -> Tuple[str, ...]:
|
|
if self.is_upgrade:
|
|
return self.revision._versioned_down_revisions
|
|
else:
|
|
return (self.revision.revision,)
|
|
|
|
@property
|
|
def to_revisions(self) -> Tuple[str, ...]:
|
|
if self.is_upgrade:
|
|
return (self.revision.revision,)
|
|
else:
|
|
return self.revision._normalized_down_revisions
|
|
|
|
@property
|
|
def to_revisions_no_deps( # type:ignore[override]
|
|
self,
|
|
) -> Tuple[str, ...]:
|
|
if self.is_upgrade:
|
|
return (self.revision.revision,)
|
|
else:
|
|
return self.revision._versioned_down_revisions
|
|
|
|
@property
|
|
def _has_scalar_down_revision(self) -> bool:
|
|
return len(self.revision._normalized_down_revisions) == 1
|
|
|
|
def should_delete_branch(self, heads: Set[str]) -> bool:
|
|
"""A delete is when we are a. in a downgrade and b.
|
|
we are going to the "base" or we are going to a version that
|
|
is implied as a dependency on another version that is remaining.
|
|
|
|
"""
|
|
if not self.is_downgrade:
|
|
return False
|
|
|
|
if self.revision.revision not in heads:
|
|
return False
|
|
|
|
downrevs = self.revision._normalized_down_revisions
|
|
|
|
if not downrevs:
|
|
# is a base
|
|
return True
|
|
else:
|
|
# determine what the ultimate "to_revisions" for an
|
|
# unmerge would be. If there are none, then we're a delete.
|
|
to_revisions = self._unmerge_to_revisions(heads)
|
|
return not to_revisions
|
|
|
|
def merge_branch_idents(
|
|
self, heads: Set[str]
|
|
) -> Tuple[List[str], str, str]:
|
|
other_heads = set(heads).difference(self.from_revisions)
|
|
|
|
if other_heads:
|
|
ancestors = {
|
|
r.revision
|
|
for r in self.revision_map._get_ancestor_nodes(
|
|
self.revision_map.get_revisions(other_heads), check=False
|
|
)
|
|
}
|
|
from_revisions = list(
|
|
set(self.from_revisions).difference(ancestors)
|
|
)
|
|
else:
|
|
from_revisions = list(self.from_revisions)
|
|
|
|
return (
|
|
# delete revs, update from rev, update to rev
|
|
list(from_revisions[0:-1]),
|
|
from_revisions[-1],
|
|
self.to_revisions[0],
|
|
)
|
|
|
|
def _unmerge_to_revisions(self, heads: Collection[str]) -> Tuple[str, ...]:
|
|
other_heads = set(heads).difference([self.revision.revision])
|
|
if other_heads:
|
|
ancestors = {
|
|
r.revision
|
|
for r in self.revision_map._get_ancestor_nodes(
|
|
self.revision_map.get_revisions(other_heads), check=False
|
|
)
|
|
}
|
|
return tuple(set(self.to_revisions).difference(ancestors))
|
|
else:
|
|
return self.to_revisions
|
|
|
|
def unmerge_branch_idents(
|
|
self, heads: Collection[str]
|
|
) -> Tuple[str, str, Tuple[str, ...]]:
|
|
to_revisions = self._unmerge_to_revisions(heads)
|
|
|
|
return (
|
|
# update from rev, update to rev, insert revs
|
|
self.from_revisions[0],
|
|
to_revisions[-1],
|
|
to_revisions[0:-1],
|
|
)
|
|
|
|
def should_create_branch(self, heads: Set[str]) -> bool:
|
|
if not self.is_upgrade:
|
|
return False
|
|
|
|
downrevs = self.revision._normalized_down_revisions
|
|
|
|
if not downrevs:
|
|
# is a base
|
|
return True
|
|
else:
|
|
# none of our downrevs are present, so...
|
|
# we have to insert our version. This is true whether
|
|
# or not there is only one downrev, or multiple (in the latter
|
|
# case, we're a merge point.)
|
|
if not heads.intersection(downrevs):
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
def should_merge_branches(self, heads: Set[str]) -> bool:
|
|
if not self.is_upgrade:
|
|
return False
|
|
|
|
downrevs = self.revision._normalized_down_revisions
|
|
|
|
if len(downrevs) > 1 and len(heads.intersection(downrevs)) > 1:
|
|
return True
|
|
|
|
return False
|
|
|
|
def should_unmerge_branches(self, heads: Set[str]) -> bool:
|
|
if not self.is_downgrade:
|
|
return False
|
|
|
|
downrevs = self.revision._normalized_down_revisions
|
|
|
|
if self.revision.revision in heads and len(downrevs) > 1:
|
|
return True
|
|
|
|
return False
|
|
|
|
def update_version_num(self, heads: Set[str]) -> Tuple[str, str]:
|
|
if not self._has_scalar_down_revision:
|
|
downrev = heads.intersection(
|
|
self.revision._normalized_down_revisions
|
|
)
|
|
assert (
|
|
len(downrev) == 1
|
|
), "Can't do an UPDATE because downrevision is ambiguous"
|
|
down_revision = list(downrev)[0]
|
|
else:
|
|
down_revision = self.revision._normalized_down_revisions[0]
|
|
|
|
if self.is_upgrade:
|
|
return down_revision, self.revision.revision
|
|
else:
|
|
return self.revision.revision, down_revision
|
|
|
|
@property
|
|
def delete_version_num(self) -> str:
|
|
return self.revision.revision
|
|
|
|
@property
|
|
def insert_version_num(self) -> str:
|
|
return self.revision.revision
|
|
|
|
@property
|
|
def info(self) -> MigrationInfo:
|
|
return MigrationInfo(
|
|
revision_map=self.revision_map,
|
|
up_revisions=self.revision.revision,
|
|
down_revisions=self.revision._normalized_down_revisions,
|
|
is_upgrade=self.is_upgrade,
|
|
is_stamp=False,
|
|
)
|
|
|
|
|
|
class StampStep(MigrationStep):
|
|
def __init__(
|
|
self,
|
|
from_: Optional[Union[str, Collection[str]]],
|
|
to_: Optional[Union[str, Collection[str]]],
|
|
is_upgrade: bool,
|
|
branch_move: bool,
|
|
revision_map: Optional[RevisionMap] = None,
|
|
) -> None:
|
|
self.from_: Tuple[str, ...] = util.to_tuple(from_, default=())
|
|
self.to_: Tuple[str, ...] = util.to_tuple(to_, default=())
|
|
self.is_upgrade = is_upgrade
|
|
self.branch_move = branch_move
|
|
self.migration_fn = self.stamp_revision
|
|
self.revision_map = revision_map
|
|
|
|
doc: None = None
|
|
|
|
def stamp_revision(self, **kw: Any) -> None:
|
|
return None
|
|
|
|
def __eq__(self, other):
|
|
return (
|
|
isinstance(other, StampStep)
|
|
and other.from_revisions == self.revisions
|
|
and other.to_revisions == self.to_revisions
|
|
and other.branch_move == self.branch_move
|
|
and self.is_upgrade == other.is_upgrade
|
|
)
|
|
|
|
@property
|
|
def from_revisions(self):
|
|
return self.from_
|
|
|
|
@property
|
|
def to_revisions(self) -> Tuple[str, ...]:
|
|
return self.to_
|
|
|
|
@property
|
|
def from_revisions_no_deps( # type:ignore[override]
|
|
self,
|
|
) -> Tuple[str, ...]:
|
|
return self.from_
|
|
|
|
@property
|
|
def to_revisions_no_deps( # type:ignore[override]
|
|
self,
|
|
) -> Tuple[str, ...]:
|
|
return self.to_
|
|
|
|
@property
|
|
def delete_version_num(self) -> str:
|
|
assert len(self.from_) == 1
|
|
return self.from_[0]
|
|
|
|
@property
|
|
def insert_version_num(self) -> str:
|
|
assert len(self.to_) == 1
|
|
return self.to_[0]
|
|
|
|
def update_version_num(self, heads: Set[str]) -> Tuple[str, str]:
|
|
assert len(self.from_) == 1
|
|
assert len(self.to_) == 1
|
|
return self.from_[0], self.to_[0]
|
|
|
|
def merge_branch_idents(
|
|
self, heads: Union[Set[str], List[str]]
|
|
) -> Union[Tuple[List[Any], str, str], Tuple[List[str], str, str]]:
|
|
return (
|
|
# delete revs, update from rev, update to rev
|
|
list(self.from_[0:-1]),
|
|
self.from_[-1],
|
|
self.to_[0],
|
|
)
|
|
|
|
def unmerge_branch_idents(
|
|
self, heads: Set[str]
|
|
) -> Tuple[str, str, List[str]]:
|
|
return (
|
|
# update from rev, update to rev, insert revs
|
|
self.from_[0],
|
|
self.to_[-1],
|
|
list(self.to_[0:-1]),
|
|
)
|
|
|
|
def should_delete_branch(self, heads: Set[str]) -> bool:
|
|
# TODO: we probably need to look for self.to_ inside of heads,
|
|
# in a similar manner as should_create_branch, however we have
|
|
# no tests for this yet (stamp downgrades w/ branches)
|
|
return self.is_downgrade and self.branch_move
|
|
|
|
def should_create_branch(self, heads: Set[str]) -> Union[Set[str], bool]:
|
|
return (
|
|
self.is_upgrade
|
|
and (self.branch_move or set(self.from_).difference(heads))
|
|
and set(self.to_).difference(heads)
|
|
)
|
|
|
|
def should_merge_branches(self, heads: Set[str]) -> bool:
|
|
return len(self.from_) > 1
|
|
|
|
def should_unmerge_branches(self, heads: Set[str]) -> bool:
|
|
return len(self.to_) > 1
|
|
|
|
@property
|
|
def info(self) -> MigrationInfo:
|
|
up, down = (
|
|
(self.to_, self.from_)
|
|
if self.is_upgrade
|
|
else (self.from_, self.to_)
|
|
)
|
|
assert self.revision_map is not None
|
|
return MigrationInfo(
|
|
revision_map=self.revision_map,
|
|
up_revisions=up,
|
|
down_revisions=down,
|
|
is_upgrade=self.is_upgrade,
|
|
is_stamp=True,
|
|
)
|