Compare commits
124 Commits
Author | SHA1 | Date | |
---|---|---|---|
|
d2e0a68351 | ||
|
ee6cc20c7d | ||
|
4e917495a0 | ||
|
bfa66f6dd4 | ||
|
f00715d4c4 | ||
|
6e3105690a | ||
|
c707f7ecb2 | ||
|
0bbc471e00 | ||
|
fb6cc62047 | ||
|
e9ceaf471f | ||
|
85fc3b2aa2 | ||
|
a677d506a9 | ||
|
9879004fee | ||
|
5760fe2040 | ||
|
b229c30558 | ||
|
5d2f1604c3 | ||
|
499c4e1c02 | ||
|
1463ee30bc | ||
|
3b801932f5 | ||
|
c2eb4dc9e3 | ||
|
5927febd0c | ||
|
a1c10ff330 | ||
|
f2013c931a | ||
|
b21b954d32 | ||
|
f5588a35c5 | ||
|
f5dff84476 | ||
|
e399821116 | ||
|
648f25a951 | ||
|
fa73e132e2 | ||
|
1bac33cd33 | ||
|
4e76f12ccf | ||
|
724379700e | ||
|
bb929f2b55 | ||
|
6339dc86a8 | ||
|
768747140a | ||
|
1fde3cd04e | ||
|
d0ce545ff5 | ||
|
09b89ed7d0 | ||
|
86c8382593 | ||
|
48e3ff48a3 | ||
|
1bf6d45bb0 | ||
|
342f4cdd3b | ||
|
8cace21fde | ||
|
9889d9492b | ||
|
823368aea8 | ||
|
6b1ad46cf1 | ||
|
ce8c0b1f06 | ||
|
43922d3734 | ||
|
48c5318737 | ||
|
002221e557 | ||
|
141d7205bf | ||
|
af4d4be19a | ||
|
3b4d9b47ce | ||
|
4b0c4ae7d0 | ||
|
dc821d8a02 | ||
|
d18a6b5be0 | ||
|
1e56a70f21 | ||
|
ab1e1aab75 | ||
|
00dd04f97d | ||
|
fc914acc80 | ||
|
ac03ecb002 | ||
|
235ef3f7ea | ||
|
e00eb7f3d9 | ||
|
d6c8941676 | ||
|
cf062c9310 | ||
|
309adec8c9 | ||
|
8674142ba8 | ||
|
cda9bd1c47 | ||
|
198e4e0032 | ||
|
1b440477a2 | ||
|
1263c6f735 | ||
|
6504384879 | ||
|
17ab0a1421 | ||
|
e1ffcb609b | ||
|
18cb75f555 | ||
|
dfe13ea250 | ||
|
b97ce0ff2f | ||
|
21001c0eda | ||
|
9bd96a9487 | ||
|
5bdaa32a9e | ||
|
d74e7b5630 | ||
|
119b15d597 | ||
|
f93ab6bbff | ||
|
bd8eb94a6e | ||
|
2fcb2626fd | ||
|
3728db4279 | ||
|
6ac1fb5332 | ||
|
87a17d443c | ||
|
55f18a69ed | ||
|
86a9c4cedd | ||
|
86e1d3defb | ||
|
01fa7fbbdb | ||
|
90196eb1bf | ||
|
3c111792a9 | ||
|
77e9d7bc91 | ||
|
fe2ddff88b | ||
|
0d23297f46 | ||
|
6be6d55e5b | ||
|
25674bc73a | ||
|
1715eda1a3 | ||
|
f5775049dd | ||
|
6fd0f8a42f | ||
|
f52dc009af | ||
|
9248d456f9 | ||
|
c24f2f6b09 | ||
|
19c9c2c30f | ||
|
73b75349ee | ||
|
7bc553221a | ||
|
7413a05e19 | ||
|
bf194ca8ce | ||
|
b06da0223a | ||
|
83554cdc5d | ||
|
6c76bfccad | ||
|
a1746e457c | ||
|
2a0435dea9 | ||
|
e87f67f1e1 | ||
|
7b4b7ac749 | ||
|
5b9b51db3f | ||
|
ffeee3c901 | ||
|
b4366d2427 | ||
|
ec1c80f3a9 | ||
|
d2083632eb | ||
|
dc8b4c2263 | ||
|
cb5dffeeb8 |
11
.github/workflows/pypi.yml
vendored
11
.github/workflows/pypi.yml
vendored
@@ -8,14 +8,17 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v2
|
- uses: actions/checkout@v2
|
||||||
- uses: actions/setup-python@v1
|
- uses: actions/setup-python@v2
|
||||||
with:
|
with:
|
||||||
python-version: '3.x'
|
python-version: '3.x'
|
||||||
|
- name: Install and configure Poetry
|
||||||
|
uses: snok/install-poetry@v1.1.1
|
||||||
|
with:
|
||||||
|
virtualenvs-create: false
|
||||||
- name: Build dists
|
- name: Build dists
|
||||||
run: |
|
run: make build
|
||||||
python3 setup.py sdist
|
|
||||||
- name: Pypi Publish
|
- name: Pypi Publish
|
||||||
uses: pypa/gh-action-pypi-publish@master
|
uses: pypa/gh-action-pypi-publish@master
|
||||||
with:
|
with:
|
||||||
user: __token__
|
user: __token__
|
||||||
password: ${{ secrets.pypi_password }}
|
password: ${{ secrets.pypi_password }}
|
||||||
|
12
.github/workflows/test.yml
vendored
12
.github/workflows/test.yml
vendored
@@ -1,5 +1,5 @@
|
|||||||
name: test
|
name: test
|
||||||
on: [push, pull_request]
|
on: [ push, pull_request ]
|
||||||
jobs:
|
jobs:
|
||||||
testall:
|
testall:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
@@ -19,10 +19,10 @@ jobs:
|
|||||||
- uses: actions/setup-python@v2
|
- uses: actions/setup-python@v2
|
||||||
with:
|
with:
|
||||||
python-version: '3.x'
|
python-version: '3.x'
|
||||||
- name: Install dependencies
|
- name: Install and configure Poetry
|
||||||
run: |
|
uses: snok/install-poetry@v1.1.1
|
||||||
python -m pip install --upgrade pip
|
with:
|
||||||
pip install -r requirements-dev.txt
|
virtualenvs-create: false
|
||||||
- name: CI
|
- name: CI
|
||||||
env:
|
env:
|
||||||
MYSQL_PASS: root
|
MYSQL_PASS: root
|
||||||
@@ -31,4 +31,4 @@ jobs:
|
|||||||
POSTGRES_PASS: 123456
|
POSTGRES_PASS: 123456
|
||||||
POSTGRES_HOST: 127.0.0.1
|
POSTGRES_HOST: 127.0.0.1
|
||||||
POSTGRES_PORT: 5432
|
POSTGRES_PORT: 5432
|
||||||
run: make testall
|
run: make ci
|
||||||
|
4
.gitignore
vendored
4
.gitignore
vendored
@@ -143,4 +143,6 @@ cython_debug/
|
|||||||
.idea
|
.idea
|
||||||
migrations
|
migrations
|
||||||
aerich.ini
|
aerich.ini
|
||||||
src
|
src
|
||||||
|
.vscode
|
||||||
|
.DS_Store
|
||||||
|
131
CHANGELOG.md
Normal file
131
CHANGELOG.md
Normal file
@@ -0,0 +1,131 @@
|
|||||||
|
# ChangeLog
|
||||||
|
|
||||||
|
## 0.4
|
||||||
|
|
||||||
|
### 0.4.2
|
||||||
|
|
||||||
|
- Use `pathlib` for path resolving. (#89)
|
||||||
|
- Fix upgrade in new db. (#96)
|
||||||
|
- Fix packaging error. (#92)
|
||||||
|
|
||||||
|
### 0.4.1
|
||||||
|
|
||||||
|
- Bug fix. (#91 #93)
|
||||||
|
|
||||||
|
### 0.4.0
|
||||||
|
|
||||||
|
- Use `.sql` instead of `.json` to store version file.
|
||||||
|
- Add `rename` column support MySQL5.
|
||||||
|
- Remove callable detection for defaults. (#87)
|
||||||
|
- Fix `sqlite` stuck. (#90)
|
||||||
|
|
||||||
|
## 0.3
|
||||||
|
|
||||||
|
### 0.3.3
|
||||||
|
|
||||||
|
- Fix encoding error. (#75)
|
||||||
|
- Support multiple databases. (#68)
|
||||||
|
- Compatible with models file in directory. (#70)
|
||||||
|
|
||||||
|
### 0.3.2
|
||||||
|
|
||||||
|
- Fix migrate to new database error. (#62)
|
||||||
|
|
||||||
|
### 0.3.1
|
||||||
|
|
||||||
|
- Fix first version error.
|
||||||
|
- Fix init error. (#61)
|
||||||
|
|
||||||
|
### 0.3.0
|
||||||
|
|
||||||
|
- Refactoring migrate logic, and this version is not compatible with previous version.
|
||||||
|
- Now there don't need `old_models.py` and it store in database.
|
||||||
|
- Upgrade steps:
|
||||||
|
1. Upgrade aerich version.
|
||||||
|
2. Drop aerich table in database.
|
||||||
|
3. Delete `migrations/{app}` folder and rerun `aerich init-db`.
|
||||||
|
4. Update model and `aerich migrate` normally.
|
||||||
|
|
||||||
|
## 0.2
|
||||||
|
|
||||||
|
### 0.2.5
|
||||||
|
|
||||||
|
- Fix windows support. (#46)
|
||||||
|
- Support `db_constraint` in fk, m2m should manual define table with fk. (#52)
|
||||||
|
|
||||||
|
### 0.2.4
|
||||||
|
|
||||||
|
- Raise error with SQLite unsupported features.
|
||||||
|
- Fix Postgres alter table. (#48)
|
||||||
|
- Add `Rename` support.
|
||||||
|
|
||||||
|
### 0.2.3
|
||||||
|
|
||||||
|
- Fix tortoise ssl config.
|
||||||
|
- PostgreSQL add/drop index/unique.
|
||||||
|
|
||||||
|
### 0.2.2
|
||||||
|
|
||||||
|
- Fix postgres drop fk.
|
||||||
|
- Fix version sort.
|
||||||
|
|
||||||
|
### 0.2.1
|
||||||
|
|
||||||
|
- Fix bug in windows.
|
||||||
|
- Enhance PostgreSQL support.
|
||||||
|
|
||||||
|
### 0.2.0
|
||||||
|
|
||||||
|
- Update model file find method.
|
||||||
|
- Set `--safe` bool.
|
||||||
|
|
||||||
|
## 0.1
|
||||||
|
|
||||||
|
### 0.1.9
|
||||||
|
|
||||||
|
- Fix default_connection when upgrade
|
||||||
|
- Find default app instead of default.
|
||||||
|
- Diff MySQL ddl.
|
||||||
|
- Check tortoise config.
|
||||||
|
|
||||||
|
### 0.1.8
|
||||||
|
|
||||||
|
- Fix upgrade error when migrate.
|
||||||
|
- Fix init db sql error.
|
||||||
|
- Support change column.
|
||||||
|
|
||||||
|
### 0.1.7
|
||||||
|
|
||||||
|
- Exclude models.Aerich.
|
||||||
|
- Add init record when init-db.
|
||||||
|
- Fix version num str.
|
||||||
|
|
||||||
|
### 0.1.6
|
||||||
|
|
||||||
|
- update dependency_links
|
||||||
|
|
||||||
|
### 0.1.5
|
||||||
|
|
||||||
|
- Add sqlite and postgres support.
|
||||||
|
- Fix dependency import.
|
||||||
|
- Store versions in db.
|
||||||
|
|
||||||
|
### 0.1.4
|
||||||
|
|
||||||
|
- Fix transaction and fields import.
|
||||||
|
- Make unique index worked.
|
||||||
|
- Add cli --version.
|
||||||
|
|
||||||
|
### 0.1.3
|
||||||
|
|
||||||
|
- Support indexes and unique_together.
|
||||||
|
|
||||||
|
### 0.1.2
|
||||||
|
|
||||||
|
- Now aerich support m2m.
|
||||||
|
- Add cli cmd init-db.
|
||||||
|
- Change cli options.
|
||||||
|
|
||||||
|
### 0.1.1
|
||||||
|
|
||||||
|
- Now aerich is basic worked.
|
@@ -1,54 +0,0 @@
|
|||||||
=========
|
|
||||||
ChangeLog
|
|
||||||
=========
|
|
||||||
|
|
||||||
0.1
|
|
||||||
===
|
|
||||||
0.1.9
|
|
||||||
-----
|
|
||||||
- Fix default_connection when upgrade
|
|
||||||
- Find default app instead of default.
|
|
||||||
- Diff MySQL ddl.
|
|
||||||
- Check tortoise config.
|
|
||||||
|
|
||||||
0.1.8
|
|
||||||
-----
|
|
||||||
- Fix upgrade error when migrate.
|
|
||||||
- Fix init db sql error.
|
|
||||||
- Support change column.
|
|
||||||
|
|
||||||
0.1.7
|
|
||||||
-----
|
|
||||||
- Exclude models.Aerich.
|
|
||||||
- Add init record when init-db.
|
|
||||||
- Fix version num str.
|
|
||||||
|
|
||||||
0.1.6
|
|
||||||
-----
|
|
||||||
- update dependency_links
|
|
||||||
|
|
||||||
0.1.5
|
|
||||||
-----
|
|
||||||
- Add sqlite and postgres support.
|
|
||||||
- Fix dependency import.
|
|
||||||
- Store versions in db.
|
|
||||||
|
|
||||||
0.1.4
|
|
||||||
-----
|
|
||||||
- Fix transaction and fields import.
|
|
||||||
- Make unique index worked.
|
|
||||||
- Add cli --version.
|
|
||||||
|
|
||||||
0.1.3
|
|
||||||
-----
|
|
||||||
- Support indexes and unique_together.
|
|
||||||
|
|
||||||
0.1.2
|
|
||||||
-----
|
|
||||||
- Now aerich support m2m.
|
|
||||||
- Add cli cmd init-db.
|
|
||||||
- Change cli options.
|
|
||||||
|
|
||||||
0.1.1
|
|
||||||
-----
|
|
||||||
- Now aerich is basic worked.
|
|
214
LICENSE
214
LICENSE
@@ -1,21 +1,201 @@
|
|||||||
The MIT License (MIT)
|
Apache License
|
||||||
|
Version 2.0, January 2004
|
||||||
|
http://www.apache.org/licenses/
|
||||||
|
|
||||||
Copyright (c) 2020 long2ice
|
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||||
|
|
||||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
1. Definitions.
|
||||||
of this software and associated documentation files (the "Software"), to deal
|
|
||||||
in the Software without restriction, including without limitation the rights
|
|
||||||
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
||||||
copies of the Software, and to permit persons to whom the Software is
|
|
||||||
furnished to do so, subject to the following conditions:
|
|
||||||
|
|
||||||
The above copyright notice and this permission notice shall be included in
|
"License" shall mean the terms and conditions for use, reproduction,
|
||||||
all copies or substantial portions of the Software.
|
and distribution as defined by Sections 1 through 9 of this document.
|
||||||
|
|
||||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
"Licensor" shall mean the copyright owner or entity authorized by
|
||||||
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
the copyright owner that is granting the License.
|
||||||
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
||||||
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
"Legal Entity" shall mean the union of the acting entity and all
|
||||||
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
other entities that control, are controlled by, or are under common
|
||||||
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
control with that entity. For the purposes of this definition,
|
||||||
THE SOFTWARE.
|
"control" means (i) the power, direct or indirect, to cause the
|
||||||
|
direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||||
|
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||||
|
|
||||||
|
"You" (or "Your") shall mean an individual or Legal Entity
|
||||||
|
exercising permissions granted by this License.
|
||||||
|
|
||||||
|
"Source" form shall mean the preferred form for making modifications,
|
||||||
|
including but not limited to software source code, documentation
|
||||||
|
source, and configuration files.
|
||||||
|
|
||||||
|
"Object" form shall mean any form resulting from mechanical
|
||||||
|
transformation or translation of a Source form, including but
|
||||||
|
not limited to compiled object code, generated documentation,
|
||||||
|
and conversions to other media types.
|
||||||
|
|
||||||
|
"Work" shall mean the work of authorship, whether in Source or
|
||||||
|
Object form, made available under the License, as indicated by a
|
||||||
|
copyright notice that is included in or attached to the work
|
||||||
|
(an example is provided in the Appendix below).
|
||||||
|
|
||||||
|
"Derivative Works" shall mean any work, whether in Source or Object
|
||||||
|
form, that is based on (or derived from) the Work and for which the
|
||||||
|
editorial revisions, annotations, elaborations, or other modifications
|
||||||
|
represent, as a whole, an original work of authorship. For the purposes
|
||||||
|
of this License, Derivative Works shall not include works that remain
|
||||||
|
separable from, or merely link (or bind by name) to the interfaces of,
|
||||||
|
the Work and Derivative Works thereof.
|
||||||
|
|
||||||
|
"Contribution" shall mean any work of authorship, including
|
||||||
|
the original version of the Work and any modifications or additions
|
||||||
|
to that Work or Derivative Works thereof, that is intentionally
|
||||||
|
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||||
|
or by an individual or Legal Entity authorized to submit on behalf of
|
||||||
|
the copyright owner. For the purposes of this definition, "submitted"
|
||||||
|
means any form of electronic, verbal, or written communication sent
|
||||||
|
to the Licensor or its representatives, including but not limited to
|
||||||
|
communication on electronic mailing lists, source code control systems,
|
||||||
|
and issue tracking systems that are managed by, or on behalf of, the
|
||||||
|
Licensor for the purpose of discussing and improving the Work, but
|
||||||
|
excluding communication that is conspicuously marked or otherwise
|
||||||
|
designated in writing by the copyright owner as "Not a Contribution."
|
||||||
|
|
||||||
|
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||||
|
on behalf of whom a Contribution has been received by Licensor and
|
||||||
|
subsequently incorporated within the Work.
|
||||||
|
|
||||||
|
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
copyright license to reproduce, prepare Derivative Works of,
|
||||||
|
publicly display, publicly perform, sublicense, and distribute the
|
||||||
|
Work and such Derivative Works in Source or Object form.
|
||||||
|
|
||||||
|
3. Grant of Patent License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
(except as stated in this section) patent license to make, have made,
|
||||||
|
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||||
|
where such license applies only to those patent claims licensable
|
||||||
|
by such Contributor that are necessarily infringed by their
|
||||||
|
Contribution(s) alone or by combination of their Contribution(s)
|
||||||
|
with the Work to which such Contribution(s) was submitted. If You
|
||||||
|
institute patent litigation against any entity (including a
|
||||||
|
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||||
|
or a Contribution incorporated within the Work constitutes direct
|
||||||
|
or contributory patent infringement, then any patent licenses
|
||||||
|
granted to You under this License for that Work shall terminate
|
||||||
|
as of the date such litigation is filed.
|
||||||
|
|
||||||
|
4. Redistribution. You may reproduce and distribute copies of the
|
||||||
|
Work or Derivative Works thereof in any medium, with or without
|
||||||
|
modifications, and in Source or Object form, provided that You
|
||||||
|
meet the following conditions:
|
||||||
|
|
||||||
|
(a) You must give any other recipients of the Work or
|
||||||
|
Derivative Works a copy of this License; and
|
||||||
|
|
||||||
|
(b) You must cause any modified files to carry prominent notices
|
||||||
|
stating that You changed the files; and
|
||||||
|
|
||||||
|
(c) You must retain, in the Source form of any Derivative Works
|
||||||
|
that You distribute, all copyright, patent, trademark, and
|
||||||
|
attribution notices from the Source form of the Work,
|
||||||
|
excluding those notices that do not pertain to any part of
|
||||||
|
the Derivative Works; and
|
||||||
|
|
||||||
|
(d) If the Work includes a "NOTICE" text file as part of its
|
||||||
|
distribution, then any Derivative Works that You distribute must
|
||||||
|
include a readable copy of the attribution notices contained
|
||||||
|
within such NOTICE file, excluding those notices that do not
|
||||||
|
pertain to any part of the Derivative Works, in at least one
|
||||||
|
of the following places: within a NOTICE text file distributed
|
||||||
|
as part of the Derivative Works; within the Source form or
|
||||||
|
documentation, if provided along with the Derivative Works; or,
|
||||||
|
within a display generated by the Derivative Works, if and
|
||||||
|
wherever such third-party notices normally appear. The contents
|
||||||
|
of the NOTICE file are for informational purposes only and
|
||||||
|
do not modify the License. You may add Your own attribution
|
||||||
|
notices within Derivative Works that You distribute, alongside
|
||||||
|
or as an addendum to the NOTICE text from the Work, provided
|
||||||
|
that such additional attribution notices cannot be construed
|
||||||
|
as modifying the License.
|
||||||
|
|
||||||
|
You may add Your own copyright statement to Your modifications and
|
||||||
|
may provide additional or different license terms and conditions
|
||||||
|
for use, reproduction, or distribution of Your modifications, or
|
||||||
|
for any such Derivative Works as a whole, provided Your use,
|
||||||
|
reproduction, and distribution of the Work otherwise complies with
|
||||||
|
the conditions stated in this License.
|
||||||
|
|
||||||
|
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||||
|
any Contribution intentionally submitted for inclusion in the Work
|
||||||
|
by You to the Licensor shall be under the terms and conditions of
|
||||||
|
this License, without any additional terms or conditions.
|
||||||
|
Notwithstanding the above, nothing herein shall supersede or modify
|
||||||
|
the terms of any separate license agreement you may have executed
|
||||||
|
with Licensor regarding such Contributions.
|
||||||
|
|
||||||
|
6. Trademarks. This License does not grant permission to use the trade
|
||||||
|
names, trademarks, service marks, or product names of the Licensor,
|
||||||
|
except as required for reasonable and customary use in describing the
|
||||||
|
origin of the Work and reproducing the content of the NOTICE file.
|
||||||
|
|
||||||
|
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||||
|
agreed to in writing, Licensor provides the Work (and each
|
||||||
|
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
implied, including, without limitation, any warranties or conditions
|
||||||
|
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||||
|
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||||
|
appropriateness of using or redistributing the Work and assume any
|
||||||
|
risks associated with Your exercise of permissions under this License.
|
||||||
|
|
||||||
|
8. Limitation of Liability. In no event and under no legal theory,
|
||||||
|
whether in tort (including negligence), contract, or otherwise,
|
||||||
|
unless required by applicable law (such as deliberate and grossly
|
||||||
|
negligent acts) or agreed to in writing, shall any Contributor be
|
||||||
|
liable to You for damages, including any direct, indirect, special,
|
||||||
|
incidental, or consequential damages of any character arising as a
|
||||||
|
result of this License or out of the use or inability to use the
|
||||||
|
Work (including but not limited to damages for loss of goodwill,
|
||||||
|
work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses), even if such Contributor
|
||||||
|
has been advised of the possibility of such damages.
|
||||||
|
|
||||||
|
9. Accepting Warranty or Additional Liability. While redistributing
|
||||||
|
the Work or Derivative Works thereof, You may choose to offer,
|
||||||
|
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||||
|
or other liability obligations and/or rights consistent with this
|
||||||
|
License. However, in accepting such obligations, You may act only
|
||||||
|
on Your own behalf and on Your sole responsibility, not on behalf
|
||||||
|
of any other Contributor, and only if You agree to indemnify,
|
||||||
|
defend, and hold each Contributor harmless for any liability
|
||||||
|
incurred by, or claims asserted against, such Contributor by reason
|
||||||
|
of your accepting any such warranty or additional liability.
|
||||||
|
|
||||||
|
END OF TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
APPENDIX: How to apply the Apache License to your work.
|
||||||
|
|
||||||
|
To apply the Apache License to your work, attach the following
|
||||||
|
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||||
|
replaced with your own identifying information. (Don't include
|
||||||
|
the brackets!) The text should be enclosed in the appropriate
|
||||||
|
comment syntax for the file format. We also recommend that a
|
||||||
|
file or class name and description of purpose be included on the
|
||||||
|
same "printed page" as the copyright notice for easier
|
||||||
|
identification within third-party archives.
|
||||||
|
|
||||||
|
Copyright 2020 long2ice
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
@@ -1,3 +0,0 @@
|
|||||||
include LICENSE
|
|
||||||
include README.rst
|
|
||||||
include requirements.txt
|
|
30
Makefile
30
Makefile
@@ -3,8 +3,10 @@ black_opts = -l 100 -t py38
|
|||||||
py_warn = PYTHONDEVMODE=1
|
py_warn = PYTHONDEVMODE=1
|
||||||
MYSQL_HOST ?= "127.0.0.1"
|
MYSQL_HOST ?= "127.0.0.1"
|
||||||
MYSQL_PORT ?= 3306
|
MYSQL_PORT ?= 3306
|
||||||
|
MYSQL_PASS ?= "123456"
|
||||||
POSTGRES_HOST ?= "127.0.0.1"
|
POSTGRES_HOST ?= "127.0.0.1"
|
||||||
POSTGRES_PORT ?= 5432
|
POSTGRES_PORT ?= 5432
|
||||||
|
POSTGRES_PASS ?= "123456"
|
||||||
|
|
||||||
help:
|
help:
|
||||||
@echo "Aerich development makefile"
|
@echo "Aerich development makefile"
|
||||||
@@ -18,23 +20,20 @@ help:
|
|||||||
@echo " test Runs all tests"
|
@echo " test Runs all tests"
|
||||||
@echo " style Auto-formats the code"
|
@echo " style Auto-formats the code"
|
||||||
|
|
||||||
|
up:
|
||||||
|
@poetry update
|
||||||
|
|
||||||
deps:
|
deps:
|
||||||
@which pip-sync > /dev/null || pip install -q pip-tools
|
@poetry install -E dbdrivers
|
||||||
@pip install -r requirements-dev.txt
|
|
||||||
|
|
||||||
style: deps
|
style: deps
|
||||||
isort -rc $(checkfiles)
|
isort -src $(checkfiles)
|
||||||
black $(black_opts) $(checkfiles)
|
black $(black_opts) $(checkfiles)
|
||||||
|
|
||||||
check: deps
|
check: deps
|
||||||
ifneq ($(shell which black),)
|
|
||||||
black --check $(black_opts) $(checkfiles) || (echo "Please run 'make style' to auto-fix style issues" && false)
|
black --check $(black_opts) $(checkfiles) || (echo "Please run 'make style' to auto-fix style issues" && false)
|
||||||
endif
|
|
||||||
flake8 $(checkfiles)
|
flake8 $(checkfiles)
|
||||||
mypy $(checkfiles)
|
bandit -x tests -r $(checkfiles)
|
||||||
pylint -d C,W,R $(checkfiles)
|
|
||||||
bandit -r $(checkfiles)
|
|
||||||
python setup.py check -mrs
|
|
||||||
|
|
||||||
test: deps
|
test: deps
|
||||||
$(py_warn) TEST_DB=sqlite://:memory: py.test
|
$(py_warn) TEST_DB=sqlite://:memory: py.test
|
||||||
@@ -43,17 +42,14 @@ test_sqlite:
|
|||||||
$(py_warn) TEST_DB=sqlite://:memory: py.test
|
$(py_warn) TEST_DB=sqlite://:memory: py.test
|
||||||
|
|
||||||
test_mysql:
|
test_mysql:
|
||||||
$(py_warn) TEST_DB="mysql://root:$(MYSQL_PASS)@$(MYSQL_HOST):$(MYSQL_PORT)/test_\{\}" py.test
|
$(py_warn) TEST_DB="mysql://root:$(MYSQL_PASS)@$(MYSQL_HOST):$(MYSQL_PORT)/test_\{\}" pytest -vv -s
|
||||||
|
|
||||||
test_postgres:
|
test_postgres:
|
||||||
$(py_warn) TEST_DB="postgres://postgres:$(POSTGRES_PASS)@$(POSTGRES_HOST):$(POSTGRES_PORT)/test_\{\}" py.test
|
$(py_warn) TEST_DB="postgres://postgres:$(POSTGRES_PASS)@$(POSTGRES_HOST):$(POSTGRES_PORT)/test_\{\}" pytest
|
||||||
|
|
||||||
testall: deps test_sqlite test_postgres test_mysql
|
testall: deps test_sqlite test_postgres test_mysql
|
||||||
|
|
||||||
publish: deps
|
build: deps
|
||||||
rm -fR dist/
|
@poetry build
|
||||||
python setup.py sdist
|
|
||||||
twine upload dist/*
|
|
||||||
|
|
||||||
ci:
|
ci: check testall
|
||||||
@act -P ubuntu-latest=nektos/act-environments-ubuntu:18.04 -b
|
|
||||||
|
192
README.md
Normal file
192
README.md
Normal file
@@ -0,0 +1,192 @@
|
|||||||
|
# Aerich
|
||||||
|
|
||||||
|
[](https://pypi.python.org/pypi/aerich)
|
||||||
|
[](https://github.com/long2ice/aerich)
|
||||||
|
[](https://github.com/long2ice/aerich/actions?query=workflow:pypi)
|
||||||
|
[](https://github.com/long2ice/aerich/actions?query=workflow:test)
|
||||||
|
|
||||||
|
## Introduction
|
||||||
|
|
||||||
|
Aerich is a database migrations tool for Tortoise-ORM, which like alembic for SQLAlchemy, or Django ORM with it\'s
|
||||||
|
own migrations solution.
|
||||||
|
|
||||||
|
**Important: You can only use absolutely import in your `models.py` to make `aerich` work.**
|
||||||
|
|
||||||
|
## Install
|
||||||
|
|
||||||
|
Just install from pypi:
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> pip install aerich
|
||||||
|
```
|
||||||
|
|
||||||
|
## Quick Start
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich -h
|
||||||
|
|
||||||
|
Usage: aerich [OPTIONS] COMMAND [ARGS]...
|
||||||
|
|
||||||
|
Options:
|
||||||
|
-c, --config TEXT Config file. [default: aerich.ini]
|
||||||
|
--app TEXT Tortoise-ORM app name. [default: models]
|
||||||
|
-n, --name TEXT Name of section in .ini file to use for aerich config.
|
||||||
|
[default: aerich]
|
||||||
|
-h, --help Show this message and exit.
|
||||||
|
|
||||||
|
Commands:
|
||||||
|
downgrade Downgrade to specified version.
|
||||||
|
heads Show current available heads in migrate location.
|
||||||
|
history List all migrate items.
|
||||||
|
init Init config file and generate root migrate location.
|
||||||
|
init-db Generate schema and generate app migrate location.
|
||||||
|
migrate Generate migrate changes file.
|
||||||
|
upgrade Upgrade to latest version.
|
||||||
|
```
|
||||||
|
|
||||||
|
## Usage
|
||||||
|
|
||||||
|
You need add `aerich.models` to your `Tortoise-ORM` config first,
|
||||||
|
example:
|
||||||
|
|
||||||
|
```python
|
||||||
|
TORTOISE_ORM = {
|
||||||
|
"connections": {"default": "mysql://root:123456@127.0.0.1:3306/test"},
|
||||||
|
"apps": {
|
||||||
|
"models": {
|
||||||
|
"models": ["tests.models", "aerich.models"],
|
||||||
|
"default_connection": "default",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Initialization
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich init -h
|
||||||
|
|
||||||
|
Usage: aerich init [OPTIONS]
|
||||||
|
|
||||||
|
Init config file and generate root migrate location.
|
||||||
|
|
||||||
|
Options:
|
||||||
|
-t, --tortoise-orm TEXT Tortoise-ORM config module dict variable, like settings.TORTOISE_ORM.
|
||||||
|
[required]
|
||||||
|
--location TEXT Migrate store location. [default: ./migrations]
|
||||||
|
-h, --help Show this message and exit.
|
||||||
|
```
|
||||||
|
|
||||||
|
Init config file and location:
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich init -t tests.backends.mysql.TORTOISE_ORM
|
||||||
|
|
||||||
|
Success create migrate location ./migrations
|
||||||
|
Success generate config file aerich.ini
|
||||||
|
```
|
||||||
|
|
||||||
|
### Init db
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich init-db
|
||||||
|
|
||||||
|
Success create app migrate location ./migrations/models
|
||||||
|
Success generate schema for app "models"
|
||||||
|
```
|
||||||
|
|
||||||
|
If your Tortoise-ORM app is not default `models`, you must specify
|
||||||
|
`--app` like `aerich --app other_models init-db`.
|
||||||
|
|
||||||
|
### Update models and make migrate
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich migrate --name drop_column
|
||||||
|
|
||||||
|
Success migrate 1_202029051520102929_drop_column.sql
|
||||||
|
```
|
||||||
|
|
||||||
|
Format of migrate filename is
|
||||||
|
`{version_num}_{datetime}_{name|update}.sql`.
|
||||||
|
|
||||||
|
And if `aerich` guess you are renaming a column, it will ask `Rename {old_column} to {new_column} [True]`, you can choice `True` to rename column without column drop, or choice `False` to drop column then create.
|
||||||
|
|
||||||
|
### Upgrade to latest version
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich upgrade
|
||||||
|
|
||||||
|
Success upgrade 1_202029051520102929_drop_column.sql
|
||||||
|
```
|
||||||
|
|
||||||
|
Now your db is migrated to latest.
|
||||||
|
|
||||||
|
### Downgrade to specified version
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich init -h
|
||||||
|
|
||||||
|
Usage: aerich downgrade [OPTIONS]
|
||||||
|
|
||||||
|
Downgrade to specified version.
|
||||||
|
|
||||||
|
Options:
|
||||||
|
-v, --version INTEGER Specified version, default to last. [default: -1]
|
||||||
|
-d, --delete Delete version files at the same time. [default:
|
||||||
|
False]
|
||||||
|
|
||||||
|
--yes Confirm the action without prompting.
|
||||||
|
-h, --help Show this message and exit.
|
||||||
|
```
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich downgrade
|
||||||
|
|
||||||
|
Success downgrade 1_202029051520102929_drop_column.sql
|
||||||
|
```
|
||||||
|
|
||||||
|
Now your db rollback to specified version.
|
||||||
|
|
||||||
|
### Show history
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich history
|
||||||
|
|
||||||
|
1_202029051520102929_drop_column.sql
|
||||||
|
```
|
||||||
|
|
||||||
|
### Show heads to be migrated
|
||||||
|
|
||||||
|
```shell
|
||||||
|
> aerich heads
|
||||||
|
|
||||||
|
1_202029051520102929_drop_column.sql
|
||||||
|
```
|
||||||
|
|
||||||
|
### Multiple databases
|
||||||
|
|
||||||
|
```python
|
||||||
|
tortoise_orm = {
|
||||||
|
"connections": {
|
||||||
|
"default": expand_db_url(db_url, True),
|
||||||
|
"second": expand_db_url(db_url_second, True),
|
||||||
|
},
|
||||||
|
"apps": {
|
||||||
|
"models": {"models": ["tests.models", "aerich.models"], "default_connection": "default"},
|
||||||
|
"models_second": {"models": ["tests.models_second"], "default_connection": "second",},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
You need only specify `aerich.models` in one app, and must specify `--app` when run `aerich migrate` and so on.
|
||||||
|
|
||||||
|
## Support this project
|
||||||
|
|
||||||
|
| AliPay | WeChatPay | PayPal |
|
||||||
|
| -------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------- | ---------------------------------------------------------------- |
|
||||||
|
| <img width="200" src="https://github.com/long2ice/aerich/raw/dev/images/alipay.jpeg"/> | <img width="200" src="https://github.com/long2ice/aerich/raw/dev/images/wechatpay.jpeg"/> | [PayPal](https://www.paypal.me/long2ice) to my account long2ice. |
|
||||||
|
|
||||||
|
## License
|
||||||
|
|
||||||
|
This project is licensed under the
|
||||||
|
[Apache-2.0](https://github.com/long2ice/aerich/blob/master/LICENSE) License.
|
169
README.rst
169
README.rst
@@ -1,169 +0,0 @@
|
|||||||
======
|
|
||||||
Aerich
|
|
||||||
======
|
|
||||||
|
|
||||||
.. image:: https://img.shields.io/pypi/v/aerich.svg?style=flat
|
|
||||||
:target: https://pypi.python.org/pypi/aerich
|
|
||||||
.. image:: https://img.shields.io/github/license/long2ice/aerich
|
|
||||||
:target: https://github.com/long2ice/aerich
|
|
||||||
.. image:: https://github.com/long2ice/aerich/workflows/pypi/badge.svg
|
|
||||||
:target: https://github.com/long2ice/aerich/actions?query=workflow:pypi
|
|
||||||
.. image:: https://github.com/long2ice/aerich/workflows/test/badge.svg
|
|
||||||
:target: https://github.com/long2ice/aerich/actions?query=workflow:test
|
|
||||||
|
|
||||||
Introduction
|
|
||||||
============
|
|
||||||
|
|
||||||
Tortoise-ORM is the best asyncio ORM now, but it lacks a database migrations tool like alembic for SQLAlchemy, or Django ORM with it's own migrations tool.
|
|
||||||
|
|
||||||
This project aim to be a best migrations tool for Tortoise-ORM and which written by one of contributors of Tortoise-ORM.
|
|
||||||
|
|
||||||
Install
|
|
||||||
=======
|
|
||||||
|
|
||||||
Just install from pypi:
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ pip install aerich
|
|
||||||
|
|
||||||
Quick Start
|
|
||||||
===========
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich -h
|
|
||||||
|
|
||||||
Usage: aerich [OPTIONS] COMMAND [ARGS]...
|
|
||||||
|
|
||||||
Options:
|
|
||||||
-c, --config TEXT Config file. [default: aerich.ini]
|
|
||||||
--app TEXT Tortoise-ORM app name. [default: models]
|
|
||||||
-n, --name TEXT Name of section in .ini file to use for aerich config.
|
|
||||||
[default: aerich]
|
|
||||||
-h, --help Show this message and exit.
|
|
||||||
|
|
||||||
Commands:
|
|
||||||
downgrade Downgrade to previous version.
|
|
||||||
heads Show current available heads in migrate location.
|
|
||||||
history List all migrate items.
|
|
||||||
init Init config file and generate root migrate location.
|
|
||||||
init-db Generate schema and generate app migrate location.
|
|
||||||
migrate Generate migrate changes file.
|
|
||||||
upgrade Upgrade to latest version.
|
|
||||||
|
|
||||||
Usage
|
|
||||||
=====
|
|
||||||
You need add ``aerich.models`` to your ``Tortoise-ORM`` config first, example:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
TORTOISE_ORM = {
|
|
||||||
"connections": {"default": "mysql://root:123456@127.0.0.1:3306/test"},
|
|
||||||
"apps": {
|
|
||||||
"models": {
|
|
||||||
"models": ["tests.models", "aerich.models"],
|
|
||||||
"default_connection": "default",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
Initialization
|
|
||||||
--------------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich init -h
|
|
||||||
|
|
||||||
Usage: aerich init [OPTIONS]
|
|
||||||
|
|
||||||
Init config file and generate root migrate location.
|
|
||||||
|
|
||||||
Options:
|
|
||||||
-t, --tortoise-orm TEXT Tortoise-ORM config module dict variable, like settings.TORTOISE_ORM.
|
|
||||||
[required]
|
|
||||||
--location TEXT Migrate store location. [default: ./migrations]
|
|
||||||
-h, --help Show this message and exit.
|
|
||||||
|
|
||||||
Init config file and location:
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich init -t tests.backends.mysql.TORTOISE_ORM
|
|
||||||
|
|
||||||
Success create migrate location ./migrations
|
|
||||||
Success generate config file aerich.ini
|
|
||||||
|
|
||||||
Init db
|
|
||||||
-------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich init-db
|
|
||||||
|
|
||||||
Success create app migrate location ./migrations/models
|
|
||||||
Success generate schema for app "models"
|
|
||||||
|
|
||||||
.. note::
|
|
||||||
|
|
||||||
If your Tortoise-ORM app is not default ``models``, you must specify ``--app`` like ``aerich --app other_models init-db``.
|
|
||||||
|
|
||||||
Update models and make migrate
|
|
||||||
------------------------------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich migrate --name drop_column
|
|
||||||
|
|
||||||
Success migrate 1_202029051520102929_drop_column.json
|
|
||||||
|
|
||||||
Format of migrate filename is ``{version_num}_{datetime}_{name|update}.json``
|
|
||||||
|
|
||||||
Upgrade to latest version
|
|
||||||
-------------------------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich upgrade
|
|
||||||
|
|
||||||
Success upgrade 1_202029051520102929_drop_column.json
|
|
||||||
|
|
||||||
Now your db is migrated to latest.
|
|
||||||
|
|
||||||
Downgrade to previous version
|
|
||||||
-----------------------------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich downgrade
|
|
||||||
|
|
||||||
Success downgrade 1_202029051520102929_drop_column.json
|
|
||||||
|
|
||||||
Now your db rollback to previous version.
|
|
||||||
|
|
||||||
Show history
|
|
||||||
------------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich history
|
|
||||||
|
|
||||||
1_202029051520102929_drop_column.json
|
|
||||||
|
|
||||||
Show heads to be migrated
|
|
||||||
-------------------------
|
|
||||||
|
|
||||||
.. code-block:: shell
|
|
||||||
|
|
||||||
$ aerich heads
|
|
||||||
|
|
||||||
1_202029051520102929_drop_column.json
|
|
||||||
|
|
||||||
Limitations
|
|
||||||
===========
|
|
||||||
* Not support ``rename column`` now.
|
|
||||||
* ``Sqlite`` and ``Postgres`` may not work as expected because I don't use those in my work.
|
|
||||||
|
|
||||||
License
|
|
||||||
=======
|
|
||||||
This project is licensed under the `MIT <https://github.com/long2ice/aerich/blob/master/LICENSE>`_ License.
|
|
@@ -1 +1 @@
|
|||||||
__version__ = "0.1.9"
|
__version__ = "0.4.2"
|
||||||
|
204
aerich/cli.py
204
aerich/cli.py
@@ -1,36 +1,54 @@
|
|||||||
import json
|
import asyncio
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
from configparser import ConfigParser
|
from configparser import ConfigParser
|
||||||
from enum import Enum
|
from functools import wraps
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
import asyncclick as click
|
import click
|
||||||
from asyncclick import Context, UsageError
|
from click import Context, UsageError
|
||||||
from tortoise import Tortoise, generate_schema_for_client
|
from tortoise import Tortoise, generate_schema_for_client
|
||||||
from tortoise.exceptions import OperationalError
|
from tortoise.exceptions import OperationalError
|
||||||
from tortoise.transactions import in_transaction
|
from tortoise.transactions import in_transaction
|
||||||
from tortoise.utils import get_schema_sql
|
from tortoise.utils import get_schema_sql
|
||||||
|
|
||||||
from aerich.migrate import Migrate
|
from aerich.migrate import Migrate
|
||||||
from aerich.utils import get_app_connection, get_app_connection_name, get_tortoise_config
|
from aerich.utils import (
|
||||||
|
get_app_connection,
|
||||||
|
get_app_connection_name,
|
||||||
|
get_tortoise_config,
|
||||||
|
get_version_content_from_file,
|
||||||
|
write_version_file,
|
||||||
|
)
|
||||||
|
|
||||||
from . import __version__
|
from . import __version__
|
||||||
|
from .enums import Color
|
||||||
from .models import Aerich
|
from .models import Aerich
|
||||||
|
|
||||||
|
|
||||||
class Color(str, Enum):
|
|
||||||
green = "green"
|
|
||||||
red = "red"
|
|
||||||
yellow = "yellow"
|
|
||||||
|
|
||||||
|
|
||||||
parser = ConfigParser()
|
parser = ConfigParser()
|
||||||
|
|
||||||
|
|
||||||
|
def coro(f):
|
||||||
|
@wraps(f)
|
||||||
|
def wrapper(*args, **kwargs):
|
||||||
|
loop = asyncio.get_event_loop()
|
||||||
|
ctx = args[0]
|
||||||
|
loop.run_until_complete(f(*args, **kwargs))
|
||||||
|
app = ctx.obj.get("app")
|
||||||
|
if app:
|
||||||
|
Migrate.remove_old_model_file(app, ctx.obj["location"])
|
||||||
|
|
||||||
|
return wrapper
|
||||||
|
|
||||||
|
|
||||||
@click.group(context_settings={"help_option_names": ["-h", "--help"]})
|
@click.group(context_settings={"help_option_names": ["-h", "--help"]})
|
||||||
@click.version_option(__version__)
|
@click.version_option(__version__, "-V", "--version")
|
||||||
@click.option(
|
@click.option(
|
||||||
"-c", "--config", default="aerich.ini", show_default=True, help="Config file.",
|
"-c",
|
||||||
|
"--config",
|
||||||
|
default="aerich.ini",
|
||||||
|
show_default=True,
|
||||||
|
help="Config file.",
|
||||||
)
|
)
|
||||||
@click.option("--app", required=False, help="Tortoise-ORM app name.")
|
@click.option("--app", required=False, help="Tortoise-ORM app name.")
|
||||||
@click.option(
|
@click.option(
|
||||||
@@ -41,6 +59,7 @@ parser = ConfigParser()
|
|||||||
help="Name of section in .ini file to use for aerich config.",
|
help="Name of section in .ini file to use for aerich config.",
|
||||||
)
|
)
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
|
@coro
|
||||||
async def cli(ctx: Context, config, app, name):
|
async def cli(ctx: Context, config, app, name):
|
||||||
ctx.ensure_object(dict)
|
ctx.ensure_object(dict)
|
||||||
ctx.obj["config_file"] = config
|
ctx.obj["config_file"] = config
|
||||||
@@ -48,7 +67,7 @@ async def cli(ctx: Context, config, app, name):
|
|||||||
|
|
||||||
invoked_subcommand = ctx.invoked_subcommand
|
invoked_subcommand = ctx.invoked_subcommand
|
||||||
if invoked_subcommand != "init":
|
if invoked_subcommand != "init":
|
||||||
if not os.path.exists(config):
|
if not Path(config).exists():
|
||||||
raise UsageError("You must exec init first", ctx=ctx)
|
raise UsageError("You must exec init first", ctx=ctx)
|
||||||
parser.read(config)
|
parser.read(config)
|
||||||
|
|
||||||
@@ -57,81 +76,113 @@ async def cli(ctx: Context, config, app, name):
|
|||||||
|
|
||||||
tortoise_config = get_tortoise_config(ctx, tortoise_orm)
|
tortoise_config = get_tortoise_config(ctx, tortoise_orm)
|
||||||
app = app or list(tortoise_config.get("apps").keys())[0]
|
app = app or list(tortoise_config.get("apps").keys())[0]
|
||||||
if "aerich.models" not in tortoise_config.get("apps").get(app).get("models"):
|
|
||||||
raise UsageError("Check your tortoise config and add aerich.models to it.", ctx=ctx)
|
|
||||||
ctx.obj["config"] = tortoise_config
|
ctx.obj["config"] = tortoise_config
|
||||||
ctx.obj["location"] = location
|
ctx.obj["location"] = location
|
||||||
ctx.obj["app"] = app
|
ctx.obj["app"] = app
|
||||||
|
Migrate.app = app
|
||||||
if invoked_subcommand != "init-db":
|
if invoked_subcommand != "init-db":
|
||||||
|
if not Path(location, app).exists():
|
||||||
|
raise UsageError("You must exec init-db first", ctx=ctx)
|
||||||
await Migrate.init_with_old_models(tortoise_config, app, location)
|
await Migrate.init_with_old_models(tortoise_config, app, location)
|
||||||
|
|
||||||
|
|
||||||
@cli.command(help="Generate migrate changes file.")
|
@cli.command(help="Generate migrate changes file.")
|
||||||
@click.option("--name", default="update", show_default=True, help="Migrate name.")
|
@click.option("--name", default="update", show_default=True, help="Migrate name.")
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
|
@coro
|
||||||
async def migrate(ctx: Context, name):
|
async def migrate(ctx: Context, name):
|
||||||
config = ctx.obj["config"]
|
|
||||||
location = ctx.obj["location"]
|
|
||||||
app = ctx.obj["app"]
|
|
||||||
|
|
||||||
ret = await Migrate.migrate(name)
|
ret = await Migrate.migrate(name)
|
||||||
if not ret:
|
if not ret:
|
||||||
return click.secho("No changes detected", fg=Color.yellow)
|
return click.secho("No changes detected", fg=Color.yellow)
|
||||||
Migrate.write_old_models(config, app, location)
|
|
||||||
click.secho(f"Success migrate {ret}", fg=Color.green)
|
click.secho(f"Success migrate {ret}", fg=Color.green)
|
||||||
|
|
||||||
|
|
||||||
@cli.command(help="Upgrade to latest version.")
|
@cli.command(help="Upgrade to specified version.")
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
|
@coro
|
||||||
async def upgrade(ctx: Context):
|
async def upgrade(ctx: Context):
|
||||||
config = ctx.obj["config"]
|
config = ctx.obj["config"]
|
||||||
app = ctx.obj["app"]
|
app = ctx.obj["app"]
|
||||||
|
location = ctx.obj["location"]
|
||||||
migrated = False
|
migrated = False
|
||||||
for version in Migrate.get_all_version_files():
|
for version_file in Migrate.get_all_version_files():
|
||||||
try:
|
try:
|
||||||
exists = await Aerich.exists(version=version, app=app)
|
exists = await Aerich.exists(version=version_file, app=app)
|
||||||
except OperationalError:
|
except OperationalError:
|
||||||
exists = False
|
exists = False
|
||||||
if not exists:
|
if not exists:
|
||||||
async with in_transaction(get_app_connection_name(config, app)) as conn:
|
async with in_transaction(get_app_connection_name(config, app)) as conn:
|
||||||
file_path = os.path.join(Migrate.migrate_location, version)
|
file_path = Path(Migrate.migrate_location, version_file)
|
||||||
with open(file_path, "r") as f:
|
content = get_version_content_from_file(file_path)
|
||||||
content = json.load(f)
|
upgrade_query_list = content.get("upgrade")
|
||||||
upgrade_query_list = content.get("upgrade")
|
for upgrade_query in upgrade_query_list:
|
||||||
for upgrade_query in upgrade_query_list:
|
await conn.execute_script(upgrade_query)
|
||||||
await conn.execute_query(upgrade_query)
|
await Aerich.create(
|
||||||
await Aerich.create(version=version, app=app)
|
version=version_file,
|
||||||
click.secho(f"Success upgrade {version}", fg=Color.green)
|
app=app,
|
||||||
|
content=Migrate.get_models_content(config, app, location),
|
||||||
|
)
|
||||||
|
click.secho(f"Success upgrade {version_file}", fg=Color.green)
|
||||||
migrated = True
|
migrated = True
|
||||||
if not migrated:
|
if not migrated:
|
||||||
click.secho("No migrate items", fg=Color.yellow)
|
click.secho("No upgrade items found", fg=Color.yellow)
|
||||||
|
|
||||||
|
|
||||||
@cli.command(help="Downgrade to previous version.")
|
@cli.command(help="Downgrade to specified version.")
|
||||||
|
@click.option(
|
||||||
|
"-v",
|
||||||
|
"--version",
|
||||||
|
default=-1,
|
||||||
|
type=int,
|
||||||
|
show_default=True,
|
||||||
|
help="Specified version, default to last.",
|
||||||
|
)
|
||||||
|
@click.option(
|
||||||
|
"-d",
|
||||||
|
"--delete",
|
||||||
|
is_flag=True,
|
||||||
|
default=False,
|
||||||
|
show_default=True,
|
||||||
|
help="Delete version files at the same time.",
|
||||||
|
)
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
async def downgrade(ctx: Context):
|
@click.confirmation_option(
|
||||||
|
prompt="Downgrade is dangerous, which maybe lose your data, are you sure?",
|
||||||
|
)
|
||||||
|
@coro
|
||||||
|
async def downgrade(ctx: Context, version: int, delete: bool):
|
||||||
app = ctx.obj["app"]
|
app = ctx.obj["app"]
|
||||||
config = ctx.obj["config"]
|
config = ctx.obj["config"]
|
||||||
last_version = await Migrate.get_last_version()
|
if version == -1:
|
||||||
if not last_version:
|
specified_version = await Migrate.get_last_version()
|
||||||
return click.secho("No last version found", fg=Color.yellow)
|
else:
|
||||||
file = last_version.version
|
specified_version = await Aerich.filter(app=app, version__startswith=f"{version}_").first()
|
||||||
async with in_transaction(get_app_connection_name(config, app)) as conn:
|
if not specified_version:
|
||||||
file_path = os.path.join(Migrate.migrate_location, file)
|
return click.secho("No specified version found", fg=Color.yellow)
|
||||||
with open(file_path, "r") as f:
|
if version == -1:
|
||||||
content = json.load(f)
|
versions = [specified_version]
|
||||||
|
else:
|
||||||
|
versions = await Aerich.filter(app=app, pk__gte=specified_version.pk)
|
||||||
|
for version in versions:
|
||||||
|
file = version.version
|
||||||
|
async with in_transaction(get_app_connection_name(config, app)) as conn:
|
||||||
|
file_path = Path(Migrate.migrate_location, file)
|
||||||
|
content = get_version_content_from_file(file_path)
|
||||||
downgrade_query_list = content.get("downgrade")
|
downgrade_query_list = content.get("downgrade")
|
||||||
if not downgrade_query_list:
|
if not downgrade_query_list:
|
||||||
return click.secho(f"No downgrade item dound", fg=Color.yellow)
|
click.secho("No downgrade items found", fg=Color.yellow)
|
||||||
|
return
|
||||||
for downgrade_query in downgrade_query_list:
|
for downgrade_query in downgrade_query_list:
|
||||||
await conn.execute_query(downgrade_query)
|
await conn.execute_query(downgrade_query)
|
||||||
await last_version.delete()
|
await version.delete()
|
||||||
return click.secho(f"Success downgrade {file}", fg=Color.green)
|
if delete:
|
||||||
|
os.unlink(file_path)
|
||||||
|
click.secho(f"Success downgrade {file}", fg=Color.green)
|
||||||
|
|
||||||
|
|
||||||
@cli.command(help="Show current available heads in migrate location.")
|
@cli.command(help="Show current available heads in migrate location.")
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
|
@coro
|
||||||
async def heads(ctx: Context):
|
async def heads(ctx: Context):
|
||||||
app = ctx.obj["app"]
|
app = ctx.obj["app"]
|
||||||
versions = Migrate.get_all_version_files()
|
versions = Migrate.get_all_version_files()
|
||||||
@@ -141,12 +192,13 @@ async def heads(ctx: Context):
|
|||||||
click.secho(version, fg=Color.green)
|
click.secho(version, fg=Color.green)
|
||||||
is_heads = True
|
is_heads = True
|
||||||
if not is_heads:
|
if not is_heads:
|
||||||
click.secho("No available heads,try migrate", fg=Color.green)
|
click.secho("No available heads,try migrate first", fg=Color.green)
|
||||||
|
|
||||||
|
|
||||||
@cli.command(help="List all migrate items.")
|
@cli.command(help="List all migrate items.")
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
def history(ctx):
|
@coro
|
||||||
|
async def history(ctx: Context):
|
||||||
versions = Migrate.get_all_version_files()
|
versions = Migrate.get_all_version_files()
|
||||||
for version in versions:
|
for version in versions:
|
||||||
click.secho(version, fg=Color.green)
|
click.secho(version, fg=Color.green)
|
||||||
@@ -162,25 +214,31 @@ def history(ctx):
|
|||||||
help="Tortoise-ORM config module dict variable, like settings.TORTOISE_ORM.",
|
help="Tortoise-ORM config module dict variable, like settings.TORTOISE_ORM.",
|
||||||
)
|
)
|
||||||
@click.option(
|
@click.option(
|
||||||
"--location", default="./migrations", show_default=True, help="Migrate store location."
|
"--location",
|
||||||
|
default="./migrations",
|
||||||
|
show_default=True,
|
||||||
|
help="Migrate store location.",
|
||||||
)
|
)
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
|
@coro
|
||||||
async def init(
|
async def init(
|
||||||
ctx: Context, tortoise_orm, location,
|
ctx: Context,
|
||||||
|
tortoise_orm,
|
||||||
|
location,
|
||||||
):
|
):
|
||||||
config_file = ctx.obj["config_file"]
|
config_file = ctx.obj["config_file"]
|
||||||
name = ctx.obj["name"]
|
name = ctx.obj["name"]
|
||||||
if os.path.exists(config_file):
|
if Path(config_file).exists():
|
||||||
return click.secho("You have inited", fg=Color.yellow)
|
return click.secho("You have inited", fg=Color.yellow)
|
||||||
|
|
||||||
parser.add_section(name)
|
parser.add_section(name)
|
||||||
parser.set(name, "tortoise_orm", tortoise_orm)
|
parser.set(name, "tortoise_orm", tortoise_orm)
|
||||||
parser.set(name, "location", location)
|
parser.set(name, "location", location)
|
||||||
|
|
||||||
with open(config_file, "w") as f:
|
with open(config_file, "w", encoding="utf-8") as f:
|
||||||
parser.write(f)
|
parser.write(f)
|
||||||
|
|
||||||
if not os.path.isdir(location):
|
if not Path(location).is_dir():
|
||||||
os.mkdir(location)
|
os.mkdir(location)
|
||||||
|
|
||||||
click.secho(f"Success create migrate location {location}", fg=Color.green)
|
click.secho(f"Success create migrate location {location}", fg=Color.green)
|
||||||
@@ -190,25 +248,26 @@ async def init(
|
|||||||
@cli.command(help="Generate schema and generate app migrate location.")
|
@cli.command(help="Generate schema and generate app migrate location.")
|
||||||
@click.option(
|
@click.option(
|
||||||
"--safe",
|
"--safe",
|
||||||
is_flag=True,
|
type=bool,
|
||||||
default=True,
|
default=True,
|
||||||
help="When set to true, creates the table only when it does not already exist.",
|
help="When set to true, creates the table only when it does not already exist.",
|
||||||
show_default=True,
|
show_default=True,
|
||||||
)
|
)
|
||||||
@click.pass_context
|
@click.pass_context
|
||||||
|
@coro
|
||||||
async def init_db(ctx: Context, safe):
|
async def init_db(ctx: Context, safe):
|
||||||
config = ctx.obj["config"]
|
config = ctx.obj["config"]
|
||||||
location = ctx.obj["location"]
|
location = ctx.obj["location"]
|
||||||
app = ctx.obj["app"]
|
app = ctx.obj["app"]
|
||||||
|
|
||||||
dirname = os.path.join(location, app)
|
dirname = Path(location, app)
|
||||||
if not os.path.isdir(dirname):
|
if not dirname.is_dir():
|
||||||
os.mkdir(dirname)
|
os.mkdir(dirname)
|
||||||
click.secho(f"Success create app migrate location {dirname}", fg=Color.green)
|
click.secho(f"Success create app migrate location {dirname}", fg=Color.green)
|
||||||
else:
|
else:
|
||||||
return click.secho(f"Inited {app} already", fg=Color.yellow)
|
return click.secho(
|
||||||
|
f"Inited {app} already, or delete {dirname} and try again.", fg=Color.yellow
|
||||||
Migrate.write_old_models(config, app, location)
|
)
|
||||||
|
|
||||||
await Tortoise.init(config=config)
|
await Tortoise.init(config=config)
|
||||||
connection = get_app_connection(config, app)
|
connection = get_app_connection(config, app)
|
||||||
@@ -217,15 +276,22 @@ async def init_db(ctx: Context, safe):
|
|||||||
schema = get_schema_sql(connection, safe)
|
schema = get_schema_sql(connection, safe)
|
||||||
|
|
||||||
version = await Migrate.generate_version()
|
version = await Migrate.generate_version()
|
||||||
await Aerich.create(version=version, app=app)
|
await Aerich.create(
|
||||||
with open(os.path.join(dirname, version), "w") as f:
|
version=version,
|
||||||
content = {
|
app=app,
|
||||||
"upgrade": [schema],
|
content=Migrate.get_models_content(config, app, location),
|
||||||
}
|
)
|
||||||
json.dump(content, f, ensure_ascii=False, indent=2)
|
content = {
|
||||||
return click.secho(f'Success generate schema for app "{app}"', fg=Color.green)
|
"upgrade": [schema],
|
||||||
|
}
|
||||||
|
write_version_file(Path(dirname, version), content)
|
||||||
|
click.secho(f'Success generate schema for app "{app}"', fg=Color.green)
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
sys.path.insert(0, ".")
|
sys.path.insert(0, ".")
|
||||||
cli(_anyio_backend="asyncio")
|
cli()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
|
@@ -2,7 +2,7 @@ from typing import List, Type
|
|||||||
|
|
||||||
from tortoise import BaseDBAsyncClient, ForeignKeyFieldInstance, ManyToManyFieldInstance, Model
|
from tortoise import BaseDBAsyncClient, ForeignKeyFieldInstance, ManyToManyFieldInstance, Model
|
||||||
from tortoise.backends.base.schema_generator import BaseSchemaGenerator
|
from tortoise.backends.base.schema_generator import BaseSchemaGenerator
|
||||||
from tortoise.fields import Field, JSONField, TextField, UUIDField
|
from tortoise.fields import CASCADE, Field, JSONField, TextField, UUIDField
|
||||||
|
|
||||||
|
|
||||||
class BaseDDL:
|
class BaseDDL:
|
||||||
@@ -11,14 +11,20 @@ class BaseDDL:
|
|||||||
_DROP_TABLE_TEMPLATE = 'DROP TABLE IF EXISTS "{table_name}"'
|
_DROP_TABLE_TEMPLATE = 'DROP TABLE IF EXISTS "{table_name}"'
|
||||||
_ADD_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" ADD {column}'
|
_ADD_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" ADD {column}'
|
||||||
_DROP_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" DROP COLUMN "{column_name}"'
|
_DROP_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" DROP COLUMN "{column_name}"'
|
||||||
|
_RENAME_COLUMN_TEMPLATE = (
|
||||||
|
'ALTER TABLE "{table_name}" RENAME COLUMN "{old_column_name}" TO "{new_column_name}"'
|
||||||
|
)
|
||||||
_ADD_INDEX_TEMPLATE = (
|
_ADD_INDEX_TEMPLATE = (
|
||||||
'ALTER TABLE "{table_name}" ADD {unique} INDEX "{index_name}" ({column_names})'
|
'ALTER TABLE "{table_name}" ADD {unique} INDEX "{index_name}" ({column_names})'
|
||||||
)
|
)
|
||||||
_DROP_INDEX_TEMPLATE = 'ALTER TABLE "{table_name}" DROP INDEX "{index_name}"'
|
_DROP_INDEX_TEMPLATE = 'ALTER TABLE "{table_name}" DROP INDEX "{index_name}"'
|
||||||
_ADD_FK_TEMPLATE = 'ALTER TABLE "{table_name}" ADD CONSTRAINT "{fk_name}" FOREIGN KEY ("{db_column}") REFERENCES "{table}" ("{field}") ON DELETE {on_delete}'
|
_ADD_FK_TEMPLATE = 'ALTER TABLE "{table_name}" ADD CONSTRAINT "{fk_name}" FOREIGN KEY ("{db_column}") REFERENCES "{table}" ("{field}") ON DELETE {on_delete}'
|
||||||
_DROP_FK_TEMPLATE = 'ALTER TABLE "{table_name}" DROP FOREIGN KEY "{fk_name}"'
|
_DROP_FK_TEMPLATE = 'ALTER TABLE "{table_name}" DROP FOREIGN KEY "{fk_name}"'
|
||||||
_M2M_TABLE_TEMPLATE = 'CREATE TABLE "{table_name}" ("{backward_key}" {backward_type} NOT NULL REFERENCES "{backward_table}" ("{backward_field}") ON DELETE CASCADE,"{forward_key}" {forward_type} NOT NULL REFERENCES "{forward_table}" ("{forward_field}") ON DELETE CASCADE){extra}{comment};'
|
_M2M_TABLE_TEMPLATE = 'CREATE TABLE "{table_name}" ("{backward_key}" {backward_type} NOT NULL REFERENCES "{backward_table}" ("{backward_field}") ON DELETE CASCADE,"{forward_key}" {forward_type} NOT NULL REFERENCES "{forward_table}" ("{forward_field}") ON DELETE {on_delete}){extra}{comment};'
|
||||||
_MODIFY_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" MODIFY COLUMN {column}'
|
_MODIFY_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" MODIFY COLUMN {column}'
|
||||||
|
_CHANGE_COLUMN_TEMPLATE = (
|
||||||
|
'ALTER TABLE "{table_name}" CHANGE {old_column_name} {new_column_name} {new_column_type}'
|
||||||
|
)
|
||||||
|
|
||||||
def __init__(self, client: "BaseDBAsyncClient"):
|
def __init__(self, client: "BaseDBAsyncClient"):
|
||||||
self.client = client
|
self.client = client
|
||||||
@@ -41,6 +47,7 @@ class BaseDDL:
|
|||||||
backward_type=model._meta.pk.get_for_dialect(self.DIALECT, "SQL_TYPE"),
|
backward_type=model._meta.pk.get_for_dialect(self.DIALECT, "SQL_TYPE"),
|
||||||
forward_key=field.forward_key,
|
forward_key=field.forward_key,
|
||||||
forward_type=field.related_model._meta.pk.get_for_dialect(self.DIALECT, "SQL_TYPE"),
|
forward_type=field.related_model._meta.pk.get_for_dialect(self.DIALECT, "SQL_TYPE"),
|
||||||
|
on_delete=CASCADE,
|
||||||
extra=self.schema_generator._table_generate_extra(table=field.through),
|
extra=self.schema_generator._table_generate_extra(table=field.through),
|
||||||
comment=self.schema_generator._table_comment_generator(
|
comment=self.schema_generator._table_comment_generator(
|
||||||
table=field.through, comment=field.description
|
table=field.through, comment=field.description
|
||||||
@@ -125,6 +132,23 @@ class BaseDDL:
|
|||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def rename_column(self, model: "Type[Model]", old_column_name: str, new_column_name: str):
|
||||||
|
return self._RENAME_COLUMN_TEMPLATE.format(
|
||||||
|
table_name=model._meta.db_table,
|
||||||
|
old_column_name=old_column_name,
|
||||||
|
new_column_name=new_column_name,
|
||||||
|
)
|
||||||
|
|
||||||
|
def change_column(
|
||||||
|
self, model: "Type[Model]", old_column_name: str, new_column_name: str, new_column_type: str
|
||||||
|
):
|
||||||
|
return self._CHANGE_COLUMN_TEMPLATE.format(
|
||||||
|
table_name=model._meta.db_table,
|
||||||
|
old_column_name=old_column_name,
|
||||||
|
new_column_name=new_column_name,
|
||||||
|
new_column_type=new_column_type,
|
||||||
|
)
|
||||||
|
|
||||||
def add_index(self, model: "Type[Model]", field_names: List[str], unique=False):
|
def add_index(self, model: "Type[Model]", field_names: List[str], unique=False):
|
||||||
return self._ADD_INDEX_TEMPLATE.format(
|
return self._ADD_INDEX_TEMPLATE.format(
|
||||||
unique="UNIQUE" if unique else "",
|
unique="UNIQUE" if unique else "",
|
||||||
@@ -179,3 +203,12 @@ class BaseDDL:
|
|||||||
to_field=to_field_name,
|
to_field=to_field_name,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def alter_column_default(self, model: "Type[Model]", field_object: Field):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def alter_column_null(self, model: "Type[Model]", field_object: Field):
|
||||||
|
pass
|
||||||
|
|
||||||
|
def set_comment(self, model: "Type[Model]", field_object: Field):
|
||||||
|
pass
|
||||||
|
@@ -9,6 +9,9 @@ class MysqlDDL(BaseDDL):
|
|||||||
_DROP_TABLE_TEMPLATE = "DROP TABLE IF EXISTS `{table_name}`"
|
_DROP_TABLE_TEMPLATE = "DROP TABLE IF EXISTS `{table_name}`"
|
||||||
_ADD_COLUMN_TEMPLATE = "ALTER TABLE `{table_name}` ADD {column}"
|
_ADD_COLUMN_TEMPLATE = "ALTER TABLE `{table_name}` ADD {column}"
|
||||||
_DROP_COLUMN_TEMPLATE = "ALTER TABLE `{table_name}` DROP COLUMN `{column_name}`"
|
_DROP_COLUMN_TEMPLATE = "ALTER TABLE `{table_name}` DROP COLUMN `{column_name}`"
|
||||||
|
_RENAME_COLUMN_TEMPLATE = (
|
||||||
|
"ALTER TABLE `{table_name}` RENAME COLUMN `{old_column_name}` TO `{new_column_name}`"
|
||||||
|
)
|
||||||
_ADD_INDEX_TEMPLATE = (
|
_ADD_INDEX_TEMPLATE = (
|
||||||
"ALTER TABLE `{table_name}` ADD {unique} INDEX `{index_name}` ({column_names})"
|
"ALTER TABLE `{table_name}` ADD {unique} INDEX `{index_name}` ({column_names})"
|
||||||
)
|
)
|
||||||
|
@@ -1,4 +1,8 @@
|
|||||||
|
from typing import List, Type
|
||||||
|
|
||||||
|
from tortoise import Model
|
||||||
from tortoise.backends.asyncpg.schema_generator import AsyncpgSchemaGenerator
|
from tortoise.backends.asyncpg.schema_generator import AsyncpgSchemaGenerator
|
||||||
|
from tortoise.fields import Field
|
||||||
|
|
||||||
from aerich.ddl import BaseDDL
|
from aerich.ddl import BaseDDL
|
||||||
|
|
||||||
@@ -6,3 +10,66 @@ from aerich.ddl import BaseDDL
|
|||||||
class PostgresDDL(BaseDDL):
|
class PostgresDDL(BaseDDL):
|
||||||
schema_generator_cls = AsyncpgSchemaGenerator
|
schema_generator_cls = AsyncpgSchemaGenerator
|
||||||
DIALECT = AsyncpgSchemaGenerator.DIALECT
|
DIALECT = AsyncpgSchemaGenerator.DIALECT
|
||||||
|
_ADD_INDEX_TEMPLATE = 'CREATE INDEX "{index_name}" ON "{table_name}" ({column_names})'
|
||||||
|
_ADD_UNIQUE_TEMPLATE = (
|
||||||
|
'ALTER TABLE "{table_name}" ADD CONSTRAINT "{index_name}" UNIQUE ({column_names})'
|
||||||
|
)
|
||||||
|
_DROP_INDEX_TEMPLATE = 'DROP INDEX "{index_name}"'
|
||||||
|
_DROP_UNIQUE_TEMPLATE = 'ALTER TABLE "{table_name}" DROP CONSTRAINT "{index_name}"'
|
||||||
|
_ALTER_DEFAULT_TEMPLATE = 'ALTER TABLE "{table_name}" ALTER COLUMN "{column}" {default}'
|
||||||
|
_ALTER_NULL_TEMPLATE = 'ALTER TABLE "{table_name}" ALTER COLUMN "{column}" {set_drop} NOT NULL'
|
||||||
|
_MODIFY_COLUMN_TEMPLATE = 'ALTER TABLE "{table_name}" ALTER COLUMN "{column}" TYPE {datatype}'
|
||||||
|
_SET_COMMENT_TEMPLATE = 'COMMENT ON COLUMN "{table_name}"."{column}" IS {comment}'
|
||||||
|
_DROP_FK_TEMPLATE = 'ALTER TABLE "{table_name}" DROP CONSTRAINT "{fk_name}"'
|
||||||
|
|
||||||
|
def alter_column_default(self, model: "Type[Model]", field_object: Field):
|
||||||
|
db_table = model._meta.db_table
|
||||||
|
default = self._get_default(model, field_object)
|
||||||
|
return self._ALTER_DEFAULT_TEMPLATE.format(
|
||||||
|
table_name=db_table,
|
||||||
|
column=field_object.model_field_name,
|
||||||
|
default="SET" + default if default else "DROP DEFAULT",
|
||||||
|
)
|
||||||
|
|
||||||
|
def alter_column_null(self, model: "Type[Model]", field_object: Field):
|
||||||
|
db_table = model._meta.db_table
|
||||||
|
return self._ALTER_NULL_TEMPLATE.format(
|
||||||
|
table_name=db_table,
|
||||||
|
column=field_object.model_field_name,
|
||||||
|
set_drop="DROP" if field_object.null else "SET",
|
||||||
|
)
|
||||||
|
|
||||||
|
def modify_column(self, model: "Type[Model]", field_object: Field):
|
||||||
|
db_table = model._meta.db_table
|
||||||
|
return self._MODIFY_COLUMN_TEMPLATE.format(
|
||||||
|
table_name=db_table,
|
||||||
|
column=field_object.model_field_name,
|
||||||
|
datatype=field_object.get_for_dialect(self.DIALECT, "SQL_TYPE"),
|
||||||
|
)
|
||||||
|
|
||||||
|
def add_index(self, model: "Type[Model]", field_names: List[str], unique=False):
|
||||||
|
template = self._ADD_UNIQUE_TEMPLATE if unique else self._ADD_INDEX_TEMPLATE
|
||||||
|
return template.format(
|
||||||
|
index_name=self.schema_generator._generate_index_name(
|
||||||
|
"uid" if unique else "idx", model, field_names
|
||||||
|
),
|
||||||
|
table_name=model._meta.db_table,
|
||||||
|
column_names=", ".join([self.schema_generator.quote(f) for f in field_names]),
|
||||||
|
)
|
||||||
|
|
||||||
|
def drop_index(self, model: "Type[Model]", field_names: List[str], unique=False):
|
||||||
|
template = self._DROP_UNIQUE_TEMPLATE if unique else self._DROP_INDEX_TEMPLATE
|
||||||
|
return template.format(
|
||||||
|
index_name=self.schema_generator._generate_index_name(
|
||||||
|
"uid" if unique else "idx", model, field_names
|
||||||
|
),
|
||||||
|
table_name=model._meta.db_table,
|
||||||
|
)
|
||||||
|
|
||||||
|
def set_comment(self, model: "Type[Model]", field_object: Field):
|
||||||
|
db_table = model._meta.db_table
|
||||||
|
return self._SET_COMMENT_TEMPLATE.format(
|
||||||
|
table_name=db_table,
|
||||||
|
column=field_object.model_field_name,
|
||||||
|
comment="'{}'".format(field_object.description) if field_object.description else "NULL",
|
||||||
|
)
|
||||||
|
@@ -1,8 +1,19 @@
|
|||||||
|
from typing import Type
|
||||||
|
|
||||||
|
from tortoise import Model
|
||||||
from tortoise.backends.sqlite.schema_generator import SqliteSchemaGenerator
|
from tortoise.backends.sqlite.schema_generator import SqliteSchemaGenerator
|
||||||
|
from tortoise.fields import Field
|
||||||
|
|
||||||
from aerich.ddl import BaseDDL
|
from aerich.ddl import BaseDDL
|
||||||
|
from aerich.exceptions import NotSupportError
|
||||||
|
|
||||||
|
|
||||||
class SqliteDDL(BaseDDL):
|
class SqliteDDL(BaseDDL):
|
||||||
schema_generator_cls = SqliteSchemaGenerator
|
schema_generator_cls = SqliteSchemaGenerator
|
||||||
DIALECT = SqliteSchemaGenerator.DIALECT
|
DIALECT = SqliteSchemaGenerator.DIALECT
|
||||||
|
|
||||||
|
def drop_column(self, model: "Type[Model]", column_name: str):
|
||||||
|
raise NotSupportError("Drop column is unsupported in SQLite.")
|
||||||
|
|
||||||
|
def modify_column(self, model: "Type[Model]", field_object: Field):
|
||||||
|
raise NotSupportError("Modify column is unsupported in SQLite.")
|
||||||
|
7
aerich/enums.py
Normal file
7
aerich/enums.py
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
from enum import Enum
|
||||||
|
|
||||||
|
|
||||||
|
class Color(str, Enum):
|
||||||
|
green = "green"
|
||||||
|
red = "red"
|
||||||
|
yellow = "yellow"
|
@@ -1,6 +1,4 @@
|
|||||||
class ConfigurationError(Exception):
|
class NotSupportError(Exception):
|
||||||
"""
|
"""
|
||||||
config error
|
raise when features not support
|
||||||
"""
|
"""
|
||||||
|
|
||||||
pass
|
|
||||||
|
@@ -1,23 +1,28 @@
|
|||||||
import json
|
import inspect
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
from copy import deepcopy
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from typing import Dict, List, Tuple, Type
|
from importlib import import_module
|
||||||
|
from io import StringIO
|
||||||
|
from pathlib import Path
|
||||||
|
from typing import Dict, List, Optional, Tuple, Type
|
||||||
|
|
||||||
|
import click
|
||||||
from tortoise import (
|
from tortoise import (
|
||||||
BackwardFKRelation,
|
BackwardFKRelation,
|
||||||
BackwardOneToOneRelation,
|
BackwardOneToOneRelation,
|
||||||
|
BaseDBAsyncClient,
|
||||||
ForeignKeyFieldInstance,
|
ForeignKeyFieldInstance,
|
||||||
ManyToManyFieldInstance,
|
ManyToManyFieldInstance,
|
||||||
Model,
|
Model,
|
||||||
Tortoise,
|
Tortoise,
|
||||||
)
|
)
|
||||||
|
from tortoise.exceptions import OperationalError
|
||||||
from tortoise.fields import Field
|
from tortoise.fields import Field
|
||||||
|
|
||||||
from aerich.ddl import BaseDDL
|
from aerich.ddl import BaseDDL
|
||||||
from aerich.models import Aerich
|
from aerich.models import MAX_VERSION_LENGTH, Aerich
|
||||||
from aerich.utils import get_app_connection
|
from aerich.utils import get_app_connection, write_version_file
|
||||||
|
|
||||||
|
|
||||||
class Migrate:
|
class Migrate:
|
||||||
@@ -28,6 +33,8 @@ class Migrate:
|
|||||||
_upgrade_m2m: List[str] = []
|
_upgrade_m2m: List[str] = []
|
||||||
_downgrade_m2m: List[str] = []
|
_downgrade_m2m: List[str] = []
|
||||||
_aerich = Aerich.__name__
|
_aerich = Aerich.__name__
|
||||||
|
_rename_old = []
|
||||||
|
_rename_new = []
|
||||||
|
|
||||||
ddl: BaseDDL
|
ddl: BaseDDL
|
||||||
migrate_config: dict
|
migrate_config: dict
|
||||||
@@ -35,44 +42,71 @@ class Migrate:
|
|||||||
diff_app = "diff_models"
|
diff_app = "diff_models"
|
||||||
app: str
|
app: str
|
||||||
migrate_location: str
|
migrate_location: str
|
||||||
|
dialect: str
|
||||||
|
_db_version: Optional[str] = None
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def get_old_model_file(cls):
|
def get_old_model_file(cls, app: str, location: str):
|
||||||
return cls.old_models + ".py"
|
return Path(location, app, cls.old_models + ".py")
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def get_all_version_files(cls) -> List[str]:
|
def get_all_version_files(cls) -> List[str]:
|
||||||
return sorted(filter(lambda x: x.endswith("json"), os.listdir(cls.migrate_location)))
|
return sorted(
|
||||||
|
filter(lambda x: x.endswith("sql"), os.listdir(cls.migrate_location)),
|
||||||
|
key=lambda x: int(x.split("_")[0]),
|
||||||
|
)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def get_last_version(cls) -> Aerich:
|
async def get_last_version(cls) -> Optional[Aerich]:
|
||||||
return await Aerich.filter(app=cls.app).first()
|
try:
|
||||||
|
return await Aerich.filter(app=cls.app).first()
|
||||||
|
except OperationalError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def remove_old_model_file(cls, app: str, location: str):
|
||||||
|
try:
|
||||||
|
os.unlink(cls.get_old_model_file(app, location))
|
||||||
|
except (OSError, FileNotFoundError):
|
||||||
|
pass
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
async def _get_db_version(cls, connection: BaseDBAsyncClient):
|
||||||
|
if cls.dialect == "mysql":
|
||||||
|
sql = "select version() as version"
|
||||||
|
ret = await connection.execute_query(sql)
|
||||||
|
cls._db_version = ret[1][0].get("version")
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def init_with_old_models(cls, config: dict, app: str, location: str):
|
async def init_with_old_models(cls, config: dict, app: str, location: str):
|
||||||
migrate_config = cls._get_migrate_config(config, app, location)
|
await Tortoise.init(config=config)
|
||||||
|
last_version = await cls.get_last_version()
|
||||||
cls.app = app
|
cls.app = app
|
||||||
cls.migrate_config = migrate_config
|
cls.migrate_location = Path(location, app)
|
||||||
cls.migrate_location = os.path.join(location, app)
|
if last_version:
|
||||||
|
content = last_version.content
|
||||||
|
with open(cls.get_old_model_file(app, location), "w", encoding="utf-8") as f:
|
||||||
|
f.write(content)
|
||||||
|
|
||||||
await Tortoise.init(config=migrate_config)
|
migrate_config = cls._get_migrate_config(config, app, location)
|
||||||
|
cls.migrate_config = migrate_config
|
||||||
|
await Tortoise.init(config=migrate_config)
|
||||||
|
|
||||||
connection = get_app_connection(config, app)
|
connection = get_app_connection(config, app)
|
||||||
if connection.schema_generator.DIALECT == "mysql":
|
cls.dialect = connection.schema_generator.DIALECT
|
||||||
|
if cls.dialect == "mysql":
|
||||||
from aerich.ddl.mysql import MysqlDDL
|
from aerich.ddl.mysql import MysqlDDL
|
||||||
|
|
||||||
cls.ddl = MysqlDDL(connection)
|
cls.ddl = MysqlDDL(connection)
|
||||||
elif connection.schema_generator.DIALECT == "sqlite":
|
elif cls.dialect == "sqlite":
|
||||||
from aerich.ddl.sqlite import SqliteDDL
|
from aerich.ddl.sqlite import SqliteDDL
|
||||||
|
|
||||||
cls.ddl = SqliteDDL(connection)
|
cls.ddl = SqliteDDL(connection)
|
||||||
elif connection.schema_generator.DIALECT == "postgres":
|
elif cls.dialect == "postgres":
|
||||||
from aerich.ddl.postgres import PostgresDDL
|
from aerich.ddl.postgres import PostgresDDL
|
||||||
|
|
||||||
cls.ddl = PostgresDDL(connection)
|
cls.ddl = PostgresDDL(connection)
|
||||||
else:
|
await cls._get_db_version(connection)
|
||||||
raise NotImplementedError("Current only support MySQL")
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def _get_last_version_num(cls):
|
async def _get_last_version_num(cls):
|
||||||
@@ -80,25 +114,31 @@ class Migrate:
|
|||||||
if not last_version:
|
if not last_version:
|
||||||
return None
|
return None
|
||||||
version = last_version.version
|
version = last_version.version
|
||||||
return int(version.split("_")[0])
|
return int(version.split("_", 1)[0])
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def generate_version(cls, name=None):
|
async def generate_version(cls, name=None):
|
||||||
now = datetime.now().strftime("%Y%M%D%H%M%S").replace("/", "")
|
now = datetime.now().strftime("%Y%m%d%H%M%S").replace("/", "")
|
||||||
last_version_num = await cls._get_last_version_num()
|
last_version_num = await cls._get_last_version_num()
|
||||||
if last_version_num is None:
|
if last_version_num is None:
|
||||||
return f"0_{now}_init.json"
|
return f"0_{now}_init.sql"
|
||||||
return f"{last_version_num + 1}_{now}_{name}.json"
|
version = f"{last_version_num + 1}_{now}_{name}.sql"
|
||||||
|
if len(version) > MAX_VERSION_LENGTH:
|
||||||
|
raise ValueError(f"Version name exceeds maximum length ({MAX_VERSION_LENGTH})")
|
||||||
|
return version
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def _generate_diff_sql(cls, name):
|
async def _generate_diff_sql(cls, name):
|
||||||
version = await cls.generate_version(name)
|
version = await cls.generate_version(name)
|
||||||
|
# delete if same version exists
|
||||||
|
for version_file in cls.get_all_version_files():
|
||||||
|
if version_file.startswith(version.split("_")[0]):
|
||||||
|
os.unlink(Path(cls.migrate_location, version_file))
|
||||||
content = {
|
content = {
|
||||||
"upgrade": cls.upgrade_operators,
|
"upgrade": cls.upgrade_operators,
|
||||||
"downgrade": cls.downgrade_operators,
|
"downgrade": cls.downgrade_operators,
|
||||||
}
|
}
|
||||||
with open(os.path.join(cls.migrate_location, version), "w") as f:
|
write_version_file(Path(cls.migrate_location, version), content)
|
||||||
json.dump(content, f, indent=2, ensure_ascii=False)
|
|
||||||
return version
|
return version
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
@@ -123,7 +163,7 @@ class Migrate:
|
|||||||
return await cls._generate_diff_sql(name)
|
return await cls._generate_diff_sql(name)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _add_operator(cls, operator: str, upgrade=True, fk=False):
|
def _add_operator(cls, operator: str, upgrade=True, fk_m2m=False):
|
||||||
"""
|
"""
|
||||||
add operator,differentiate fk because fk is order limit
|
add operator,differentiate fk because fk is order limit
|
||||||
:param operator:
|
:param operator:
|
||||||
@@ -132,36 +172,16 @@ class Migrate:
|
|||||||
:return:
|
:return:
|
||||||
"""
|
"""
|
||||||
if upgrade:
|
if upgrade:
|
||||||
if fk:
|
if fk_m2m:
|
||||||
cls._upgrade_fk_m2m_index_operators.append(operator)
|
cls._upgrade_fk_m2m_index_operators.append(operator)
|
||||||
else:
|
else:
|
||||||
cls.upgrade_operators.append(operator)
|
cls.upgrade_operators.append(operator)
|
||||||
else:
|
else:
|
||||||
if fk:
|
if fk_m2m:
|
||||||
cls._downgrade_fk_m2m_index_operators.append(operator)
|
cls._downgrade_fk_m2m_index_operators.append(operator)
|
||||||
else:
|
else:
|
||||||
cls.downgrade_operators.append(operator)
|
cls.downgrade_operators.append(operator)
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def cp_models(
|
|
||||||
cls, app: str, model_files: List[str], old_model_file,
|
|
||||||
):
|
|
||||||
"""
|
|
||||||
cp currents models to old_model_files
|
|
||||||
:param app:
|
|
||||||
:param model_files:
|
|
||||||
:param old_model_file:
|
|
||||||
:return:
|
|
||||||
"""
|
|
||||||
pattern = rf"(\n)?('|\")({app})(.\w+)('|\")"
|
|
||||||
for i, model_file in enumerate(model_files):
|
|
||||||
with open(model_file, "r") as f:
|
|
||||||
content = f.read()
|
|
||||||
ret = re.sub(pattern, rf"\2{cls.diff_app}\4\5", content)
|
|
||||||
mode = "w" if i == 0 else "a"
|
|
||||||
with open(old_model_file, mode) as f:
|
|
||||||
f.write(ret)
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _get_migrate_config(cls, config: dict, app: str, location: str):
|
def _get_migrate_config(cls, config: dict, app: str, location: str):
|
||||||
"""
|
"""
|
||||||
@@ -171,17 +191,15 @@ class Migrate:
|
|||||||
:param location:
|
:param location:
|
||||||
:return:
|
:return:
|
||||||
"""
|
"""
|
||||||
temp_config = deepcopy(config)
|
path = Path(location, app, cls.old_models).as_posix().replace("/", ".")
|
||||||
path = os.path.join(location, app, cls.old_models)
|
config["apps"][cls.diff_app] = {
|
||||||
path = path.replace("/", ".").lstrip(".")
|
|
||||||
temp_config["apps"][cls.diff_app] = {
|
|
||||||
"models": [path],
|
"models": [path],
|
||||||
"default_connection": config.get("apps").get(app).get("default_connection", "default"),
|
"default_connection": config.get("apps").get(app).get("default_connection", "default"),
|
||||||
}
|
}
|
||||||
return temp_config
|
return config
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def write_old_models(cls, config: dict, app: str, location: str):
|
def get_models_content(cls, config: dict, app: str, location: str):
|
||||||
"""
|
"""
|
||||||
write new models to old models
|
write new models to old models
|
||||||
:param config:
|
:param config:
|
||||||
@@ -189,15 +207,26 @@ class Migrate:
|
|||||||
:param location:
|
:param location:
|
||||||
:return:
|
:return:
|
||||||
"""
|
"""
|
||||||
cls.app = app
|
|
||||||
|
|
||||||
old_model_files = []
|
old_model_files = []
|
||||||
models = config.get("apps").get(app).get("models")
|
models = config.get("apps").get(app).get("models")
|
||||||
for model in models:
|
for model in models:
|
||||||
if model != "aerich.models":
|
module = import_module(model)
|
||||||
old_model_files.append(model.replace(".", "/") + ".py")
|
possible_models = [getattr(module, attr_name) for attr_name in dir(module)]
|
||||||
|
for attr in filter(
|
||||||
cls.cp_models(app, old_model_files, os.path.join(location, app, cls.get_old_model_file()))
|
lambda x: inspect.isclass(x) and issubclass(x, Model) and x is not Model,
|
||||||
|
possible_models,
|
||||||
|
):
|
||||||
|
file = inspect.getfile(attr)
|
||||||
|
if file not in old_model_files:
|
||||||
|
old_model_files.append(file)
|
||||||
|
pattern = rf"(\n)?('|\")({app})(.\w+)('|\")"
|
||||||
|
str_io = StringIO()
|
||||||
|
for i, model_file in enumerate(old_model_files):
|
||||||
|
with open(model_file, "r", encoding="utf-8") as f:
|
||||||
|
content = f.read()
|
||||||
|
ret = re.sub(pattern, rf"\2{cls.diff_app}\4\5", content)
|
||||||
|
str_io.write(f"{ret}\n")
|
||||||
|
return str_io.getvalue()
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def diff_models(
|
def diff_models(
|
||||||
@@ -260,11 +289,48 @@ class Migrate:
|
|||||||
if cls._exclude_field(new_field, upgrade):
|
if cls._exclude_field(new_field, upgrade):
|
||||||
continue
|
continue
|
||||||
if new_key not in old_keys:
|
if new_key not in old_keys:
|
||||||
cls._add_operator(
|
new_field_dict = new_field.describe(serializable=True)
|
||||||
cls._add_field(new_model, new_field),
|
new_field_dict.pop("name", None)
|
||||||
upgrade,
|
new_field_dict.pop("db_column", None)
|
||||||
isinstance(new_field, (ForeignKeyFieldInstance, ManyToManyFieldInstance)),
|
for diff_key in old_keys - new_keys:
|
||||||
)
|
old_field = old_fields_map.get(diff_key)
|
||||||
|
old_field_dict = old_field.describe(serializable=True)
|
||||||
|
old_field_dict.pop("name", None)
|
||||||
|
old_field_dict.pop("db_column", None)
|
||||||
|
if old_field_dict == new_field_dict:
|
||||||
|
if upgrade:
|
||||||
|
is_rename = click.prompt(
|
||||||
|
f"Rename {diff_key} to {new_key}?",
|
||||||
|
default=True,
|
||||||
|
type=bool,
|
||||||
|
show_choices=True,
|
||||||
|
)
|
||||||
|
cls._rename_new.append(new_key)
|
||||||
|
cls._rename_old.append(diff_key)
|
||||||
|
else:
|
||||||
|
is_rename = diff_key in cls._rename_new
|
||||||
|
if is_rename:
|
||||||
|
if (
|
||||||
|
cls.dialect == "mysql"
|
||||||
|
and cls._db_version
|
||||||
|
and cls._db_version.startswith("5.")
|
||||||
|
):
|
||||||
|
cls._add_operator(
|
||||||
|
cls._change_field(new_model, old_field, new_field),
|
||||||
|
upgrade,
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._rename_field(new_model, old_field, new_field),
|
||||||
|
upgrade,
|
||||||
|
)
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._add_field(new_model, new_field),
|
||||||
|
upgrade,
|
||||||
|
cls._is_fk_m2m(new_field),
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
old_field = old_fields_map.get(new_key)
|
old_field = old_fields_map.get(new_key)
|
||||||
new_field_dict = new_field.describe(serializable=True)
|
new_field_dict = new_field.describe(serializable=True)
|
||||||
@@ -274,7 +340,27 @@ class Migrate:
|
|||||||
old_field_dict.pop("unique")
|
old_field_dict.pop("unique")
|
||||||
old_field_dict.pop("indexed")
|
old_field_dict.pop("indexed")
|
||||||
if not cls._is_fk_m2m(new_field) and new_field_dict != old_field_dict:
|
if not cls._is_fk_m2m(new_field) and new_field_dict != old_field_dict:
|
||||||
cls._add_operator(cls._modify_field(new_model, new_field), upgrade=upgrade)
|
if cls.dialect == "postgres":
|
||||||
|
if new_field.null != old_field.null:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._alter_null(new_model, new_field), upgrade=upgrade
|
||||||
|
)
|
||||||
|
if new_field.default != old_field.default and not callable(
|
||||||
|
new_field.default
|
||||||
|
):
|
||||||
|
cls._add_operator(
|
||||||
|
cls._alter_default(new_model, new_field), upgrade=upgrade
|
||||||
|
)
|
||||||
|
if new_field.description != old_field.description:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._set_comment(new_model, new_field), upgrade=upgrade
|
||||||
|
)
|
||||||
|
if new_field.field_type != old_field.field_type:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._modify_field(new_model, new_field), upgrade=upgrade
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
cls._add_operator(cls._modify_field(new_model, new_field), upgrade=upgrade)
|
||||||
if (old_field.index and not new_field.index) or (
|
if (old_field.index and not new_field.index) or (
|
||||||
old_field.unique and not new_field.unique
|
old_field.unique and not new_field.unique
|
||||||
):
|
):
|
||||||
@@ -293,17 +379,41 @@ class Migrate:
|
|||||||
upgrade,
|
upgrade,
|
||||||
cls._is_fk_m2m(new_field),
|
cls._is_fk_m2m(new_field),
|
||||||
)
|
)
|
||||||
|
if isinstance(new_field, ForeignKeyFieldInstance):
|
||||||
|
if old_field.db_constraint and not new_field.db_constraint:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._drop_fk(new_model, new_field),
|
||||||
|
upgrade,
|
||||||
|
True,
|
||||||
|
)
|
||||||
|
if new_field.db_constraint and not old_field.db_constraint:
|
||||||
|
cls._add_operator(
|
||||||
|
cls._add_fk(new_model, new_field),
|
||||||
|
upgrade,
|
||||||
|
True,
|
||||||
|
)
|
||||||
|
|
||||||
for old_key in old_keys:
|
for old_key in old_keys:
|
||||||
field = old_fields_map.get(old_key)
|
field = old_fields_map.get(old_key)
|
||||||
if old_key not in new_keys and not cls._exclude_field(field, upgrade):
|
if old_key not in new_keys and not cls._exclude_field(field, upgrade):
|
||||||
cls._add_operator(
|
if (upgrade and old_key not in cls._rename_old) or (
|
||||||
cls._remove_field(old_model, field), upgrade, cls._is_fk_m2m(field),
|
not upgrade and old_key not in cls._rename_new
|
||||||
)
|
):
|
||||||
|
cls._add_operator(
|
||||||
|
cls._remove_field(old_model, field),
|
||||||
|
upgrade,
|
||||||
|
cls._is_fk_m2m(field),
|
||||||
|
)
|
||||||
|
|
||||||
for new_index in new_indexes:
|
for new_index in new_indexes:
|
||||||
if new_index not in old_indexes:
|
if new_index not in old_indexes:
|
||||||
cls._add_operator(cls._add_index(new_model, new_index,), upgrade)
|
cls._add_operator(
|
||||||
|
cls._add_index(
|
||||||
|
new_model,
|
||||||
|
new_index,
|
||||||
|
),
|
||||||
|
upgrade,
|
||||||
|
)
|
||||||
for old_index in old_indexes:
|
for old_index in old_indexes:
|
||||||
if old_index not in new_indexes:
|
if old_index not in new_indexes:
|
||||||
cls._add_operator(cls._remove_index(old_model, old_index), upgrade)
|
cls._add_operator(cls._remove_index(old_model, old_index), upgrade)
|
||||||
@@ -367,10 +477,26 @@ class Migrate:
|
|||||||
return cls.ddl.create_m2m_table(model, field)
|
return cls.ddl.create_m2m_table(model, field)
|
||||||
return cls.ddl.add_column(model, field)
|
return cls.ddl.add_column(model, field)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def _alter_default(cls, model: Type[Model], field: Field):
|
||||||
|
return cls.ddl.alter_column_default(model, field)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def _alter_null(cls, model: Type[Model], field: Field):
|
||||||
|
return cls.ddl.alter_column_null(model, field)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def _set_comment(cls, model: Type[Model], field: Field):
|
||||||
|
return cls.ddl.set_comment(model, field)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _modify_field(cls, model: Type[Model], field: Field):
|
def _modify_field(cls, model: Type[Model], field: Field):
|
||||||
return cls.ddl.modify_column(model, field)
|
return cls.ddl.modify_column(model, field)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def _drop_fk(cls, model: Type[Model], field: ForeignKeyFieldInstance):
|
||||||
|
return cls.ddl.drop_fk(model, field)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _remove_field(cls, model: Type[Model], field: Field):
|
def _remove_field(cls, model: Type[Model], field: Field):
|
||||||
if isinstance(field, ForeignKeyFieldInstance):
|
if isinstance(field, ForeignKeyFieldInstance):
|
||||||
@@ -379,6 +505,19 @@ class Migrate:
|
|||||||
return cls.ddl.drop_m2m(field)
|
return cls.ddl.drop_m2m(field)
|
||||||
return cls.ddl.drop_column(model, field.model_field_name)
|
return cls.ddl.drop_column(model, field.model_field_name)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def _rename_field(cls, model: Type[Model], old_field: Field, new_field: Field):
|
||||||
|
return cls.ddl.rename_column(model, old_field.model_field_name, new_field.model_field_name)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def _change_field(cls, model: Type[Model], old_field: Field, new_field: Field):
|
||||||
|
return cls.ddl.change_column(
|
||||||
|
model,
|
||||||
|
old_field.model_field_name,
|
||||||
|
new_field.model_field_name,
|
||||||
|
new_field.get_for_dialect(cls.dialect, "SQL_TYPE"),
|
||||||
|
)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _add_fk(cls, model: Type[Model], field: ForeignKeyFieldInstance):
|
def _add_fk(cls, model: Type[Model], field: ForeignKeyFieldInstance):
|
||||||
"""
|
"""
|
||||||
|
@@ -1,9 +1,12 @@
|
|||||||
from tortoise import Model, fields
|
from tortoise import Model, fields
|
||||||
|
|
||||||
|
MAX_VERSION_LENGTH = 255
|
||||||
|
|
||||||
|
|
||||||
class Aerich(Model):
|
class Aerich(Model):
|
||||||
version = fields.CharField(max_length=50)
|
version = fields.CharField(max_length=MAX_VERSION_LENGTH)
|
||||||
app = fields.CharField(max_length=20)
|
app = fields.CharField(max_length=20)
|
||||||
|
content = fields.TextField()
|
||||||
|
|
||||||
class Meta:
|
class Meta:
|
||||||
ordering = ["-id"]
|
ordering = ["-id"]
|
||||||
|
@@ -1,17 +1,24 @@
|
|||||||
import importlib
|
import importlib
|
||||||
|
from typing import Dict
|
||||||
|
|
||||||
from asyncclick import BadOptionUsage, Context
|
from click import BadOptionUsage, Context
|
||||||
from tortoise import BaseDBAsyncClient, Tortoise
|
from tortoise import BaseDBAsyncClient, Tortoise
|
||||||
|
|
||||||
|
|
||||||
def get_app_connection_name(config, app) -> str:
|
def get_app_connection_name(config, app_name: str) -> str:
|
||||||
"""
|
"""
|
||||||
get connection name
|
get connection name
|
||||||
:param config:
|
:param config:
|
||||||
:param app:
|
:param app_name:
|
||||||
:return:
|
:return:
|
||||||
"""
|
"""
|
||||||
return config.get("apps").get(app).get("default_connection")
|
app = config.get("apps").get(app_name)
|
||||||
|
if app:
|
||||||
|
return app.get("default_connection", "default")
|
||||||
|
raise BadOptionUsage(
|
||||||
|
option_name="--app",
|
||||||
|
message=f'Can\'t get app named "{app_name}"',
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def get_app_connection(config, app) -> BaseDBAsyncClient:
|
def get_app_connection(config, app) -> BaseDBAsyncClient:
|
||||||
@@ -49,3 +56,55 @@ def get_tortoise_config(ctx: Context, tortoise_orm: str) -> dict:
|
|||||||
ctx=ctx,
|
ctx=ctx,
|
||||||
)
|
)
|
||||||
return config
|
return config
|
||||||
|
|
||||||
|
|
||||||
|
_UPGRADE = "##### upgrade #####\n"
|
||||||
|
_DOWNGRADE = "##### downgrade #####\n"
|
||||||
|
|
||||||
|
|
||||||
|
def get_version_content_from_file(version_file: str) -> Dict:
|
||||||
|
"""
|
||||||
|
get version content
|
||||||
|
:param version_file:
|
||||||
|
:return:
|
||||||
|
"""
|
||||||
|
with open(version_file, "r", encoding="utf-8") as f:
|
||||||
|
content = f.read()
|
||||||
|
first = content.index(_UPGRADE)
|
||||||
|
try:
|
||||||
|
second = content.index(_DOWNGRADE)
|
||||||
|
except ValueError:
|
||||||
|
second = len(content) - 1
|
||||||
|
upgrade_content = content[first + len(_UPGRADE) : second].strip() # noqa:E203
|
||||||
|
downgrade_content = content[second + len(_DOWNGRADE) :].strip() # noqa:E203
|
||||||
|
ret = {
|
||||||
|
"upgrade": list(filter(lambda x: x or False, upgrade_content.split(";\n"))),
|
||||||
|
"downgrade": list(filter(lambda x: x or False, downgrade_content.split(";\n"))),
|
||||||
|
}
|
||||||
|
return ret
|
||||||
|
|
||||||
|
|
||||||
|
def write_version_file(version_file: str, content: Dict):
|
||||||
|
"""
|
||||||
|
write version file
|
||||||
|
:param version_file:
|
||||||
|
:param content:
|
||||||
|
:return:
|
||||||
|
"""
|
||||||
|
with open(version_file, "w", encoding="utf-8") as f:
|
||||||
|
f.write(_UPGRADE)
|
||||||
|
upgrade = content.get("upgrade")
|
||||||
|
if len(upgrade) > 1:
|
||||||
|
f.write(";\n".join(upgrade) + ";\n")
|
||||||
|
else:
|
||||||
|
f.write(f"{upgrade[0]}")
|
||||||
|
if not upgrade[0].endswith(";"):
|
||||||
|
f.write(";")
|
||||||
|
f.write("\n")
|
||||||
|
downgrade = content.get("downgrade")
|
||||||
|
if downgrade:
|
||||||
|
f.write(_DOWNGRADE)
|
||||||
|
if len(downgrade) > 1:
|
||||||
|
f.write(";\n".join(downgrade) + ";\n")
|
||||||
|
else:
|
||||||
|
f.write(f"{downgrade[0]};\n")
|
||||||
|
36
conftest.py
36
conftest.py
@@ -13,10 +13,15 @@ from aerich.ddl.sqlite import SqliteDDL
|
|||||||
from aerich.migrate import Migrate
|
from aerich.migrate import Migrate
|
||||||
|
|
||||||
db_url = os.getenv("TEST_DB", "sqlite://:memory:")
|
db_url = os.getenv("TEST_DB", "sqlite://:memory:")
|
||||||
|
db_url_second = os.getenv("TEST_DB_SECOND", "sqlite://:memory:")
|
||||||
tortoise_orm = {
|
tortoise_orm = {
|
||||||
"connections": {"default": expand_db_url(db_url, True)},
|
"connections": {
|
||||||
|
"default": expand_db_url(db_url, True),
|
||||||
|
"second": expand_db_url(db_url_second, True),
|
||||||
|
},
|
||||||
"apps": {
|
"apps": {
|
||||||
"models": {"models": ["tests.models", "aerich.models"], "default_connection": "default",},
|
"models": {"models": ["tests.models", "aerich.models"], "default_connection": "default"},
|
||||||
|
"models_second": {"models": ["tests.models_second"], "default_connection": "second"},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -31,24 +36,29 @@ def reset_migrate():
|
|||||||
Migrate._downgrade_m2m = []
|
Migrate._downgrade_m2m = []
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.yield_fixture(scope="session")
|
||||||
def loop():
|
def event_loop():
|
||||||
loop = asyncio.get_event_loop()
|
policy = asyncio.get_event_loop_policy()
|
||||||
return loop
|
res = policy.new_event_loop()
|
||||||
|
asyncio.set_event_loop(res)
|
||||||
|
res._close = res.close
|
||||||
|
res.close = lambda: None
|
||||||
|
|
||||||
|
yield res
|
||||||
|
|
||||||
|
res._close()
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session", autouse=True)
|
@pytest.fixture(scope="session", autouse=True)
|
||||||
def initialize_tests(loop, request):
|
async def initialize_tests(event_loop, request):
|
||||||
tortoise_orm["connections"]["diff_models"] = "sqlite://:memory:"
|
tortoise_orm["connections"]["diff_models"] = "sqlite://:memory:"
|
||||||
tortoise_orm["apps"]["diff_models"] = {
|
tortoise_orm["apps"]["diff_models"] = {
|
||||||
"models": ["tests.diff_models"],
|
"models": ["tests.diff_models"],
|
||||||
"default_connection": "diff_models",
|
"default_connection": "diff_models",
|
||||||
}
|
}
|
||||||
|
|
||||||
loop.run_until_complete(Tortoise.init(config=tortoise_orm, _create_db=True))
|
await Tortoise.init(config=tortoise_orm, _create_db=True)
|
||||||
loop.run_until_complete(
|
await generate_schema_for_client(Tortoise.get_connection("default"), safe=True)
|
||||||
generate_schema_for_client(Tortoise.get_connection("default"), safe=True)
|
|
||||||
)
|
|
||||||
|
|
||||||
client = Tortoise.get_connection("default")
|
client = Tortoise.get_connection("default")
|
||||||
if client.schema_generator is MySQLSchemaGenerator:
|
if client.schema_generator is MySQLSchemaGenerator:
|
||||||
@@ -57,5 +67,5 @@ def initialize_tests(loop, request):
|
|||||||
Migrate.ddl = SqliteDDL(client)
|
Migrate.ddl = SqliteDDL(client)
|
||||||
elif client.schema_generator is AsyncpgSchemaGenerator:
|
elif client.schema_generator is AsyncpgSchemaGenerator:
|
||||||
Migrate.ddl = PostgresDDL(client)
|
Migrate.ddl = PostgresDDL(client)
|
||||||
|
Migrate.dialect = Migrate.ddl.DIALECT
|
||||||
request.addfinalizer(lambda: loop.run_until_complete(Tortoise._drop_databases()))
|
request.addfinalizer(lambda: event_loop.run_until_complete(Tortoise._drop_databases()))
|
||||||
|
BIN
images/alipay.jpeg
Normal file
BIN
images/alipay.jpeg
Normal file
Binary file not shown.
After Width: | Height: | Size: 75 KiB |
BIN
images/wechatpay.jpeg
Normal file
BIN
images/wechatpay.jpeg
Normal file
Binary file not shown.
After Width: | Height: | Size: 76 KiB |
1007
poetry.lock
generated
1007
poetry.lock
generated
File diff suppressed because it is too large
Load Diff
@@ -1,32 +1,43 @@
|
|||||||
[tool.poetry]
|
[tool.poetry]
|
||||||
name = "aerich"
|
name = "aerich"
|
||||||
version = "0.1.9"
|
version = "0.4.2"
|
||||||
description = "A database migrations tool for Tortoise ORM."
|
description = "A database migrations tool for Tortoise ORM."
|
||||||
authors = ["long2ice <long2ice@gmail.com>"]
|
authors = ["long2ice <long2ice@gmail.com>"]
|
||||||
|
license = "Apache-2.0"
|
||||||
|
readme = "README.md"
|
||||||
|
homepage = "https://github.com/long2ice/aerich"
|
||||||
|
repository = "https://github.com/long2ice/aerich.git"
|
||||||
|
documentation = "https://github.com/long2ice/aerich"
|
||||||
|
keywords = ["migrate", "Tortoise-ORM", "mysql"]
|
||||||
|
packages = [
|
||||||
|
{ include = "aerich" }
|
||||||
|
]
|
||||||
|
include = ["CHANGELOG.md", "LICENSE", "README.md"]
|
||||||
|
|
||||||
[tool.poetry.dependencies]
|
[tool.poetry.dependencies]
|
||||||
python = "^3.8"
|
python = "^3.7"
|
||||||
tortoise-orm = "*"
|
tortoise-orm = "*"
|
||||||
asyncclick = "*"
|
click = "*"
|
||||||
pydantic = "*"
|
pydantic = "*"
|
||||||
|
aiomysql = {version = "*", optional = true}
|
||||||
|
asyncpg = {version = "*", optional = true}
|
||||||
|
|
||||||
[tool.poetry.dev-dependencies]
|
[tool.poetry.dev-dependencies]
|
||||||
taskipy = "*"
|
|
||||||
flake8 = "*"
|
flake8 = "*"
|
||||||
isort = "*"
|
isort = "*"
|
||||||
black = "^19.10b0"
|
black = "^20.8b1"
|
||||||
pytest = "*"
|
pytest = "*"
|
||||||
aiomysql = "*"
|
|
||||||
asyncpg = "*"
|
|
||||||
pytest-xdist = "*"
|
pytest-xdist = "*"
|
||||||
mypy = "*"
|
|
||||||
pytest-asyncio = "*"
|
pytest-asyncio = "*"
|
||||||
|
bandit = "*"
|
||||||
|
pytest-mock = "*"
|
||||||
|
|
||||||
[tool.taskipy.tasks]
|
[tool.poetry.extras]
|
||||||
export = "poetry export -f requirements.txt --without-hashes > requirements.txt"
|
dbdrivers = ["aiomysql", "asyncpg"]
|
||||||
export-dev = "poetry export -f requirements.txt --dev --without-hashes > requirements-dev.txt"
|
|
||||||
|
|
||||||
[build-system]
|
[build-system]
|
||||||
requires = ["poetry>=0.12"]
|
requires = ["poetry>=0.12"]
|
||||||
build-backend = "poetry.masonry.api"
|
build-backend = "poetry.masonry.api"
|
||||||
|
|
||||||
|
[tool.poetry.scripts]
|
||||||
|
aerich = "aerich.cli:main"
|
||||||
|
@@ -1,2 +0,0 @@
|
|||||||
[pytest]
|
|
||||||
addopts = -p no:warnings --ignore=src
|
|
@@ -1,48 +0,0 @@
|
|||||||
aiomysql==0.0.20
|
|
||||||
aiosqlite==0.13.0
|
|
||||||
anyio==1.3.0
|
|
||||||
apipkg==1.5
|
|
||||||
appdirs==1.4.4
|
|
||||||
async-generator==1.10
|
|
||||||
asyncclick==7.0.9
|
|
||||||
asyncpg==0.20.1
|
|
||||||
atomicwrites==1.4.0; sys_platform == "win32"
|
|
||||||
attrs==19.3.0
|
|
||||||
black==19.10b0
|
|
||||||
cffi==1.14.0
|
|
||||||
ciso8601==2.1.3; sys_platform != "win32" and implementation_name == "cpython"
|
|
||||||
click==7.1.2
|
|
||||||
colorama==0.4.3; sys_platform == "win32"
|
|
||||||
cryptography==2.9.2
|
|
||||||
execnet==1.7.1
|
|
||||||
flake8==3.8.2
|
|
||||||
iso8601==0.1.12; sys_platform == "win32" or implementation_name != "cpython"
|
|
||||||
isort==4.3.21
|
|
||||||
mccabe==0.6.1
|
|
||||||
more-itertools==8.3.0
|
|
||||||
mypy==0.770
|
|
||||||
mypy-extensions==0.4.3
|
|
||||||
packaging==20.4
|
|
||||||
pathspec==0.8.0
|
|
||||||
pluggy==0.13.1
|
|
||||||
py==1.8.1
|
|
||||||
pycodestyle==2.6.0
|
|
||||||
pycparser==2.20
|
|
||||||
pydantic==1.5.1
|
|
||||||
pyflakes==2.2.0
|
|
||||||
pymysql==0.9.2
|
|
||||||
pyparsing==2.4.7
|
|
||||||
pypika==0.37.6
|
|
||||||
pytest==5.4.2
|
|
||||||
pytest-asyncio==0.12.0
|
|
||||||
pytest-forked==1.1.3
|
|
||||||
pytest-xdist==1.32.0
|
|
||||||
regex==2020.5.14
|
|
||||||
six==1.15.0
|
|
||||||
sniffio==1.1.0
|
|
||||||
taskipy==1.2.1
|
|
||||||
toml==0.10.1
|
|
||||||
tortoise-orm==0.16.12
|
|
||||||
typed-ast==1.4.1
|
|
||||||
typing-extensions==3.7.4.2
|
|
||||||
wcwidth==0.1.9
|
|
@@ -1,11 +0,0 @@
|
|||||||
aiosqlite==0.13.0
|
|
||||||
anyio==1.3.0
|
|
||||||
async-generator==1.10
|
|
||||||
asyncclick==7.0.9
|
|
||||||
ciso8601==2.1.3; sys_platform != "win32" and implementation_name == "cpython"
|
|
||||||
iso8601==0.1.12; sys_platform == "win32" or implementation_name != "cpython"
|
|
||||||
pydantic==1.5.1
|
|
||||||
pypika==0.37.6
|
|
||||||
sniffio==1.1.0
|
|
||||||
tortoise-orm==0.16.12
|
|
||||||
typing-extensions==3.7.4.2
|
|
47
setup.cfg
47
setup.cfg
@@ -1,47 +1,2 @@
|
|||||||
[flake8]
|
[flake8]
|
||||||
max-line-length = 100
|
ignore = E501,W503
|
||||||
exclude =
|
|
||||||
ignore = E501,W503,DAR101,DAR201,DAR402
|
|
||||||
|
|
||||||
[darglint]
|
|
||||||
docstring_style=sphinx
|
|
||||||
|
|
||||||
[isort]
|
|
||||||
not_skip=__init__.py
|
|
||||||
multi_line_output=3
|
|
||||||
include_trailing_comma=True
|
|
||||||
force_grid_wrap=0
|
|
||||||
use_parentheses=True
|
|
||||||
line_length=100
|
|
||||||
|
|
||||||
[tool:pytest]
|
|
||||||
addopts = -n auto --tb=native -q
|
|
||||||
|
|
||||||
[mypy]
|
|
||||||
pretty = True
|
|
||||||
ignore_missing_imports = True
|
|
||||||
check_untyped_defs = True
|
|
||||||
disallow_subclassing_any = True
|
|
||||||
disallow_untyped_calls = True
|
|
||||||
disallow_untyped_defs = False
|
|
||||||
disallow_incomplete_defs = False
|
|
||||||
disallow_untyped_decorators = True
|
|
||||||
no_implicit_optional = True
|
|
||||||
warn_redundant_casts = True
|
|
||||||
warn_unused_ignores = True
|
|
||||||
warn_no_return = True
|
|
||||||
warn_return_any = False
|
|
||||||
warn_unused_configs = True
|
|
||||||
warn_unreachable = True
|
|
||||||
allow_redefinition = True
|
|
||||||
strict_equality = True
|
|
||||||
show_error_context = True
|
|
||||||
|
|
||||||
[mypy-tests.*]
|
|
||||||
check_untyped_defs = False
|
|
||||||
disallow_untyped_defs = False
|
|
||||||
disallow_incomplete_defs = False
|
|
||||||
warn_unreachable = False
|
|
||||||
|
|
||||||
[mypy-conftest]
|
|
||||||
disallow_untyped_defs = False
|
|
44
setup.py
44
setup.py
@@ -1,44 +0,0 @@
|
|||||||
import os
|
|
||||||
import re
|
|
||||||
from setuptools import find_packages, setup
|
|
||||||
|
|
||||||
|
|
||||||
def version():
|
|
||||||
ver_str_line = open('aerich/__init__.py', 'rt').read()
|
|
||||||
mob = re.search(r"^__version__ = ['\"]([^'\"]*)['\"]", ver_str_line, re.M)
|
|
||||||
if not mob:
|
|
||||||
raise RuntimeError("Unable to find version string")
|
|
||||||
return mob.group(1)
|
|
||||||
|
|
||||||
|
|
||||||
with open(os.path.join(os.path.dirname(__file__), 'README.rst')) as f:
|
|
||||||
long_description = f.read()
|
|
||||||
|
|
||||||
|
|
||||||
def requirements():
|
|
||||||
return open('requirements.txt', 'rt').read().splitlines()
|
|
||||||
|
|
||||||
|
|
||||||
setup(
|
|
||||||
name='aerich',
|
|
||||||
version=version(),
|
|
||||||
description='A database migrations tool for Tortoise-ORM.',
|
|
||||||
author='long2ice',
|
|
||||||
long_description_content_type='text/x-rst',
|
|
||||||
long_description=long_description,
|
|
||||||
author_email='long2ice@gmail.com',
|
|
||||||
url='https://github.com/long2ice/aerich',
|
|
||||||
license='MIT License',
|
|
||||||
packages=find_packages(include=['aerich*']),
|
|
||||||
include_package_data=True,
|
|
||||||
zip_safe=True,
|
|
||||||
entry_points={
|
|
||||||
'console_scripts': ['aerich = aerich.cli:main'],
|
|
||||||
},
|
|
||||||
platforms='any',
|
|
||||||
keywords=(
|
|
||||||
'migrate Tortoise-ORM mysql'
|
|
||||||
),
|
|
||||||
dependency_links=['https://github.com/tortoise-orm/tortoise-orm.git@develop#egg=tortoise-orm'],
|
|
||||||
install_requires=requirements(),
|
|
||||||
)
|
|
@@ -22,15 +22,21 @@ class Status(IntEnum):
|
|||||||
|
|
||||||
|
|
||||||
class User(Model):
|
class User(Model):
|
||||||
username = fields.CharField(max_length=20,)
|
username = fields.CharField(max_length=20)
|
||||||
password = fields.CharField(max_length=200)
|
password = fields.CharField(max_length=200)
|
||||||
last_login = fields.DatetimeField(description="Last Login", default=datetime.datetime.now)
|
last_login_at = fields.DatetimeField(description="Last Login", default=datetime.datetime.now)
|
||||||
is_active = fields.BooleanField(default=True, description="Is Active")
|
is_active = fields.BooleanField(default=True, description="Is Active")
|
||||||
is_superuser = fields.BooleanField(default=False, description="Is SuperUser")
|
is_superuser = fields.BooleanField(default=False, description="Is SuperUser")
|
||||||
avatar = fields.CharField(max_length=200, default="")
|
avatar = fields.CharField(max_length=200, default="")
|
||||||
intro = fields.TextField(default="")
|
intro = fields.TextField(default="")
|
||||||
|
|
||||||
|
|
||||||
|
class Email(Model):
|
||||||
|
email = fields.CharField(max_length=200)
|
||||||
|
is_primary = fields.BooleanField(default=False)
|
||||||
|
user = fields.ForeignKeyField("diff_models.User", db_constraint=True)
|
||||||
|
|
||||||
|
|
||||||
class Category(Model):
|
class Category(Model):
|
||||||
slug = fields.CharField(max_length=200)
|
slug = fields.CharField(max_length=200)
|
||||||
user = fields.ForeignKeyField("diff_models.User", description="User")
|
user = fields.ForeignKeyField("diff_models.User", description="User")
|
||||||
|
@@ -31,6 +31,12 @@ class User(Model):
|
|||||||
intro = fields.TextField(default="")
|
intro = fields.TextField(default="")
|
||||||
|
|
||||||
|
|
||||||
|
class Email(Model):
|
||||||
|
email = fields.CharField(max_length=200)
|
||||||
|
is_primary = fields.BooleanField(default=False)
|
||||||
|
user = fields.ForeignKeyField("models.User", db_constraint=False)
|
||||||
|
|
||||||
|
|
||||||
class Category(Model):
|
class Category(Model):
|
||||||
slug = fields.CharField(max_length=200)
|
slug = fields.CharField(max_length=200)
|
||||||
name = fields.CharField(max_length=200)
|
name = fields.CharField(max_length=200)
|
||||||
|
63
tests/models_second.py
Normal file
63
tests/models_second.py
Normal file
@@ -0,0 +1,63 @@
|
|||||||
|
import datetime
|
||||||
|
from enum import IntEnum
|
||||||
|
|
||||||
|
from tortoise import Model, fields
|
||||||
|
|
||||||
|
|
||||||
|
class ProductType(IntEnum):
|
||||||
|
article = 1
|
||||||
|
page = 2
|
||||||
|
|
||||||
|
|
||||||
|
class PermissionAction(IntEnum):
|
||||||
|
create = 1
|
||||||
|
delete = 2
|
||||||
|
update = 3
|
||||||
|
read = 4
|
||||||
|
|
||||||
|
|
||||||
|
class Status(IntEnum):
|
||||||
|
on = 1
|
||||||
|
off = 0
|
||||||
|
|
||||||
|
|
||||||
|
class User(Model):
|
||||||
|
username = fields.CharField(max_length=20, unique=True)
|
||||||
|
password = fields.CharField(max_length=200)
|
||||||
|
last_login = fields.DatetimeField(description="Last Login", default=datetime.datetime.now)
|
||||||
|
is_active = fields.BooleanField(default=True, description="Is Active")
|
||||||
|
is_superuser = fields.BooleanField(default=False, description="Is SuperUser")
|
||||||
|
avatar = fields.CharField(max_length=200, default="")
|
||||||
|
intro = fields.TextField(default="")
|
||||||
|
|
||||||
|
|
||||||
|
class Email(Model):
|
||||||
|
email = fields.CharField(max_length=200)
|
||||||
|
is_primary = fields.BooleanField(default=False)
|
||||||
|
user = fields.ForeignKeyField("models_second.User", db_constraint=False)
|
||||||
|
|
||||||
|
|
||||||
|
class Category(Model):
|
||||||
|
slug = fields.CharField(max_length=200)
|
||||||
|
name = fields.CharField(max_length=200)
|
||||||
|
user = fields.ForeignKeyField("models_second.User", description="User")
|
||||||
|
created_at = fields.DatetimeField(auto_now_add=True)
|
||||||
|
|
||||||
|
|
||||||
|
class Product(Model):
|
||||||
|
categories = fields.ManyToManyField("models_second.Category")
|
||||||
|
name = fields.CharField(max_length=50)
|
||||||
|
view_num = fields.IntField(description="View Num")
|
||||||
|
sort = fields.IntField()
|
||||||
|
is_reviewed = fields.BooleanField(description="Is Reviewed")
|
||||||
|
type = fields.IntEnumField(ProductType, description="Product Type")
|
||||||
|
image = fields.CharField(max_length=200)
|
||||||
|
body = fields.TextField()
|
||||||
|
created_at = fields.DatetimeField(auto_now_add=True)
|
||||||
|
|
||||||
|
|
||||||
|
class Config(Model):
|
||||||
|
label = fields.CharField(max_length=200)
|
||||||
|
key = fields.CharField(max_length=20)
|
||||||
|
value = fields.JSONField()
|
||||||
|
status: Status = fields.IntEnumField(Status, default=Status.on)
|
@@ -1,8 +1,11 @@
|
|||||||
|
import pytest
|
||||||
|
|
||||||
from aerich.ddl.mysql import MysqlDDL
|
from aerich.ddl.mysql import MysqlDDL
|
||||||
from aerich.ddl.postgres import PostgresDDL
|
from aerich.ddl.postgres import PostgresDDL
|
||||||
from aerich.ddl.sqlite import SqliteDDL
|
from aerich.ddl.sqlite import SqliteDDL
|
||||||
|
from aerich.exceptions import NotSupportError
|
||||||
from aerich.migrate import Migrate
|
from aerich.migrate import Migrate
|
||||||
from tests.models import Category
|
from tests.models import Category, User
|
||||||
|
|
||||||
|
|
||||||
def test_create_table():
|
def test_create_table():
|
||||||
@@ -39,7 +42,7 @@ def test_create_table():
|
|||||||
"id" SERIAL NOT NULL PRIMARY KEY,
|
"id" SERIAL NOT NULL PRIMARY KEY,
|
||||||
"slug" VARCHAR(200) NOT NULL,
|
"slug" VARCHAR(200) NOT NULL,
|
||||||
"name" VARCHAR(200) NOT NULL,
|
"name" VARCHAR(200) NOT NULL,
|
||||||
"created_at" TIMESTAMP NOT NULL DEFAULT CURRENT_TIMESTAMP,
|
"created_at" TIMESTAMPTZ NOT NULL DEFAULT CURRENT_TIMESTAMP,
|
||||||
"user_id" INT NOT NULL REFERENCES "user" ("id") ON DELETE CASCADE
|
"user_id" INT NOT NULL REFERENCES "user" ("id") ON DELETE CASCADE
|
||||||
);
|
);
|
||||||
COMMENT ON COLUMN "category"."user_id" IS 'User';"""
|
COMMENT ON COLUMN "category"."user_id" IS 'User';"""
|
||||||
@@ -63,18 +66,81 @@ def test_add_column():
|
|||||||
|
|
||||||
|
|
||||||
def test_modify_column():
|
def test_modify_column():
|
||||||
ret = Migrate.ddl.modify_column(Category, Category._meta.fields_map.get("name"))
|
if isinstance(Migrate.ddl, SqliteDDL):
|
||||||
if isinstance(Migrate.ddl, MysqlDDL):
|
with pytest.raises(NotSupportError):
|
||||||
assert ret == "ALTER TABLE `category` MODIFY COLUMN `name` VARCHAR(200) NOT NULL"
|
ret0 = Migrate.ddl.modify_column(Category, Category._meta.fields_map.get("name"))
|
||||||
|
ret1 = Migrate.ddl.modify_column(User, User._meta.fields_map.get("is_active"))
|
||||||
|
|
||||||
else:
|
else:
|
||||||
assert ret == 'ALTER TABLE "category" MODIFY COLUMN "name" VARCHAR(200) NOT NULL'
|
ret0 = Migrate.ddl.modify_column(Category, Category._meta.fields_map.get("name"))
|
||||||
|
ret1 = Migrate.ddl.modify_column(User, User._meta.fields_map.get("is_active"))
|
||||||
|
if isinstance(Migrate.ddl, MysqlDDL):
|
||||||
|
assert ret0 == "ALTER TABLE `category` MODIFY COLUMN `name` VARCHAR(200) NOT NULL"
|
||||||
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret0 == 'ALTER TABLE "category" ALTER COLUMN "name" TYPE VARCHAR(200)'
|
||||||
|
|
||||||
|
if isinstance(Migrate.ddl, MysqlDDL):
|
||||||
|
assert (
|
||||||
|
ret1
|
||||||
|
== "ALTER TABLE `user` MODIFY COLUMN `is_active` BOOL NOT NULL COMMENT 'Is Active' DEFAULT 1"
|
||||||
|
)
|
||||||
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret1 == 'ALTER TABLE "user" ALTER COLUMN "is_active" TYPE BOOL'
|
||||||
|
|
||||||
|
|
||||||
|
def test_alter_column_default():
|
||||||
|
ret = Migrate.ddl.alter_column_default(Category, Category._meta.fields_map.get("name"))
|
||||||
|
if isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'ALTER TABLE "category" ALTER COLUMN "name" DROP DEFAULT'
|
||||||
|
else:
|
||||||
|
assert ret is None
|
||||||
|
|
||||||
|
ret = Migrate.ddl.alter_column_default(Category, Category._meta.fields_map.get("created_at"))
|
||||||
|
if isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert (
|
||||||
|
ret == 'ALTER TABLE "category" ALTER COLUMN "created_at" SET DEFAULT CURRENT_TIMESTAMP'
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
assert ret is None
|
||||||
|
|
||||||
|
ret = Migrate.ddl.alter_column_default(User, User._meta.fields_map.get("avatar"))
|
||||||
|
if isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'ALTER TABLE "user" ALTER COLUMN "avatar" SET DEFAULT \'\''
|
||||||
|
else:
|
||||||
|
assert ret is None
|
||||||
|
|
||||||
|
|
||||||
|
def test_alter_column_null():
|
||||||
|
ret = Migrate.ddl.alter_column_null(Category, Category._meta.fields_map.get("name"))
|
||||||
|
if isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'ALTER TABLE "category" ALTER COLUMN "name" SET NOT NULL'
|
||||||
|
else:
|
||||||
|
assert ret is None
|
||||||
|
|
||||||
|
|
||||||
|
def test_set_comment():
|
||||||
|
ret = Migrate.ddl.set_comment(Category, Category._meta.fields_map.get("name"))
|
||||||
|
if isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'COMMENT ON COLUMN "category"."name" IS NULL'
|
||||||
|
else:
|
||||||
|
assert ret is None
|
||||||
|
|
||||||
|
ret = Migrate.ddl.set_comment(Category, Category._meta.fields_map.get("user"))
|
||||||
|
if isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'COMMENT ON COLUMN "category"."user" IS \'User\''
|
||||||
|
else:
|
||||||
|
assert ret is None
|
||||||
|
|
||||||
|
|
||||||
def test_drop_column():
|
def test_drop_column():
|
||||||
ret = Migrate.ddl.drop_column(Category, "name")
|
if isinstance(Migrate.ddl, SqliteDDL):
|
||||||
|
with pytest.raises(NotSupportError):
|
||||||
|
ret = Migrate.ddl.drop_column(Category, "name")
|
||||||
|
else:
|
||||||
|
ret = Migrate.ddl.drop_column(Category, "name")
|
||||||
if isinstance(Migrate.ddl, MysqlDDL):
|
if isinstance(Migrate.ddl, MysqlDDL):
|
||||||
assert ret == "ALTER TABLE `category` DROP COLUMN `name`"
|
assert ret == "ALTER TABLE `category` DROP COLUMN `name`"
|
||||||
else:
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
assert ret == 'ALTER TABLE "category" DROP COLUMN "name"'
|
assert ret == 'ALTER TABLE "category" DROP COLUMN "name"'
|
||||||
|
|
||||||
|
|
||||||
@@ -86,6 +152,12 @@ def test_add_index():
|
|||||||
assert (
|
assert (
|
||||||
index_u == "ALTER TABLE `category` ADD UNIQUE INDEX `uid_category_name_8b0cb9` (`name`)"
|
index_u == "ALTER TABLE `category` ADD UNIQUE INDEX `uid_category_name_8b0cb9` (`name`)"
|
||||||
)
|
)
|
||||||
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert index == 'CREATE INDEX "idx_category_name_8b0cb9" ON "category" ("name")'
|
||||||
|
assert (
|
||||||
|
index_u
|
||||||
|
== 'ALTER TABLE "category" ADD CONSTRAINT "uid_category_name_8b0cb9" UNIQUE ("name")'
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
assert index == 'ALTER TABLE "category" ADD INDEX "idx_category_name_8b0cb9" ("name")'
|
assert index == 'ALTER TABLE "category" ADD INDEX "idx_category_name_8b0cb9" ("name")'
|
||||||
assert (
|
assert (
|
||||||
@@ -95,10 +167,16 @@ def test_add_index():
|
|||||||
|
|
||||||
def test_drop_index():
|
def test_drop_index():
|
||||||
ret = Migrate.ddl.drop_index(Category, ["name"])
|
ret = Migrate.ddl.drop_index(Category, ["name"])
|
||||||
|
ret_u = Migrate.ddl.drop_index(Category, ["name"], True)
|
||||||
if isinstance(Migrate.ddl, MysqlDDL):
|
if isinstance(Migrate.ddl, MysqlDDL):
|
||||||
assert ret == "ALTER TABLE `category` DROP INDEX `idx_category_name_8b0cb9`"
|
assert ret == "ALTER TABLE `category` DROP INDEX `idx_category_name_8b0cb9`"
|
||||||
|
assert ret_u == "ALTER TABLE `category` DROP INDEX `uid_category_name_8b0cb9`"
|
||||||
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'DROP INDEX "idx_category_name_8b0cb9"'
|
||||||
|
assert ret_u == 'ALTER TABLE "category" DROP CONSTRAINT "uid_category_name_8b0cb9"'
|
||||||
else:
|
else:
|
||||||
assert ret == 'ALTER TABLE "category" DROP INDEX "idx_category_name_8b0cb9"'
|
assert ret == 'ALTER TABLE "category" DROP INDEX "idx_category_name_8b0cb9"'
|
||||||
|
assert ret_u == 'ALTER TABLE "category" DROP INDEX "uid_category_name_8b0cb9"'
|
||||||
|
|
||||||
|
|
||||||
def test_add_fk():
|
def test_add_fk():
|
||||||
@@ -119,5 +197,7 @@ def test_drop_fk():
|
|||||||
ret = Migrate.ddl.drop_fk(Category, Category._meta.fields_map.get("user"))
|
ret = Migrate.ddl.drop_fk(Category, Category._meta.fields_map.get("user"))
|
||||||
if isinstance(Migrate.ddl, MysqlDDL):
|
if isinstance(Migrate.ddl, MysqlDDL):
|
||||||
assert ret == "ALTER TABLE `category` DROP FOREIGN KEY `fk_category_user_e2e3874c`"
|
assert ret == "ALTER TABLE `category` DROP FOREIGN KEY `fk_category_user_e2e3874c`"
|
||||||
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
|
assert ret == 'ALTER TABLE "category" DROP CONSTRAINT "fk_category_user_e2e3874c"'
|
||||||
else:
|
else:
|
||||||
assert ret == 'ALTER TABLE "category" DROP FOREIGN KEY "fk_category_user_e2e3874c"'
|
assert ret == 'ALTER TABLE "category" DROP FOREIGN KEY "fk_category_user_e2e3874c"'
|
||||||
|
@@ -1,30 +1,79 @@
|
|||||||
|
import pytest
|
||||||
|
from pytest_mock import MockerFixture
|
||||||
from tortoise import Tortoise
|
from tortoise import Tortoise
|
||||||
|
|
||||||
from aerich.ddl.mysql import MysqlDDL
|
from aerich.ddl.mysql import MysqlDDL
|
||||||
|
from aerich.ddl.postgres import PostgresDDL
|
||||||
|
from aerich.ddl.sqlite import SqliteDDL
|
||||||
|
from aerich.exceptions import NotSupportError
|
||||||
from aerich.migrate import Migrate
|
from aerich.migrate import Migrate
|
||||||
|
|
||||||
|
|
||||||
def test_migrate():
|
def test_migrate(mocker: MockerFixture):
|
||||||
|
mocker.patch("click.prompt", return_value=True)
|
||||||
apps = Tortoise.apps
|
apps = Tortoise.apps
|
||||||
models = apps.get("models")
|
models = apps.get("models")
|
||||||
diff_models = apps.get("diff_models")
|
diff_models = apps.get("diff_models")
|
||||||
Migrate.diff_models(diff_models, models)
|
Migrate.diff_models(diff_models, models)
|
||||||
Migrate.diff_models(models, diff_models, False)
|
if isinstance(Migrate.ddl, SqliteDDL):
|
||||||
|
with pytest.raises(NotSupportError):
|
||||||
|
Migrate.diff_models(models, diff_models, False)
|
||||||
|
else:
|
||||||
|
Migrate.diff_models(models, diff_models, False)
|
||||||
|
Migrate._merge_operators()
|
||||||
if isinstance(Migrate.ddl, MysqlDDL):
|
if isinstance(Migrate.ddl, MysqlDDL):
|
||||||
assert Migrate.upgrade_operators == [
|
assert Migrate.upgrade_operators == [
|
||||||
|
"ALTER TABLE `email` DROP FOREIGN KEY `fk_email_user_5b58673d`",
|
||||||
"ALTER TABLE `category` ADD `name` VARCHAR(200) NOT NULL",
|
"ALTER TABLE `category` ADD `name` VARCHAR(200) NOT NULL",
|
||||||
"ALTER TABLE `user` ADD UNIQUE INDEX `uid_user_usernam_9987ab` (`username`)",
|
"ALTER TABLE `user` ADD UNIQUE INDEX `uid_user_usernam_9987ab` (`username`)",
|
||||||
|
"ALTER TABLE `user` RENAME COLUMN `last_login_at` TO `last_login`",
|
||||||
]
|
]
|
||||||
assert Migrate.downgrade_operators == [
|
assert Migrate.downgrade_operators == [
|
||||||
"ALTER TABLE `category` DROP COLUMN `name`",
|
"ALTER TABLE `category` DROP COLUMN `name`",
|
||||||
"ALTER TABLE `user` DROP INDEX `uid_user_usernam_9987ab`",
|
"ALTER TABLE `user` DROP INDEX `uid_user_usernam_9987ab`",
|
||||||
|
"ALTER TABLE `user` RENAME COLUMN `last_login` TO `last_login_at`",
|
||||||
|
"ALTER TABLE `email` ADD CONSTRAINT `fk_email_user_5b58673d` FOREIGN KEY "
|
||||||
|
"(`user_id`) REFERENCES `user` (`id`) ON DELETE CASCADE",
|
||||||
]
|
]
|
||||||
else:
|
elif isinstance(Migrate.ddl, PostgresDDL):
|
||||||
assert Migrate.upgrade_operators == [
|
assert Migrate.upgrade_operators == [
|
||||||
|
'ALTER TABLE "email" DROP CONSTRAINT "fk_email_user_5b58673d"',
|
||||||
'ALTER TABLE "category" ADD "name" VARCHAR(200) NOT NULL',
|
'ALTER TABLE "category" ADD "name" VARCHAR(200) NOT NULL',
|
||||||
'ALTER TABLE "user" ADD UNIQUE INDEX "uid_user_usernam_9987ab" ("username")',
|
'ALTER TABLE "user" ADD CONSTRAINT "uid_user_usernam_9987ab" UNIQUE ("username")',
|
||||||
|
'ALTER TABLE "user" RENAME COLUMN "last_login_at" TO "last_login"',
|
||||||
]
|
]
|
||||||
assert Migrate.downgrade_operators == [
|
assert Migrate.downgrade_operators == [
|
||||||
'ALTER TABLE "category" DROP COLUMN "name"',
|
'ALTER TABLE "category" DROP COLUMN "name"',
|
||||||
'ALTER TABLE "user" DROP INDEX "uid_user_usernam_9987ab"',
|
'ALTER TABLE "user" DROP CONSTRAINT "uid_user_usernam_9987ab"',
|
||||||
|
'ALTER TABLE "user" RENAME COLUMN "last_login" TO "last_login_at"',
|
||||||
|
'ALTER TABLE "email" ADD CONSTRAINT "fk_email_user_5b58673d" FOREIGN KEY ("user_id") REFERENCES "user" ("id") ON DELETE CASCADE',
|
||||||
]
|
]
|
||||||
|
elif isinstance(Migrate.ddl, SqliteDDL):
|
||||||
|
assert Migrate.upgrade_operators == [
|
||||||
|
'ALTER TABLE "email" DROP FOREIGN KEY "fk_email_user_5b58673d"',
|
||||||
|
'ALTER TABLE "category" ADD "name" VARCHAR(200) NOT NULL',
|
||||||
|
'ALTER TABLE "user" ADD UNIQUE INDEX "uid_user_usernam_9987ab" ("username")',
|
||||||
|
'ALTER TABLE "user" RENAME COLUMN "last_login_at" TO "last_login"',
|
||||||
|
]
|
||||||
|
assert Migrate.downgrade_operators == []
|
||||||
|
|
||||||
|
|
||||||
|
def test_sort_all_version_files(mocker):
|
||||||
|
mocker.patch(
|
||||||
|
"os.listdir",
|
||||||
|
return_value=[
|
||||||
|
"1_datetime_update.sql",
|
||||||
|
"11_datetime_update.sql",
|
||||||
|
"10_datetime_update.sql",
|
||||||
|
"2_datetime_update.sql",
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
|
Migrate.migrate_location = "."
|
||||||
|
|
||||||
|
assert Migrate.get_all_version_files() == [
|
||||||
|
"1_datetime_update.sql",
|
||||||
|
"2_datetime_update.sql",
|
||||||
|
"10_datetime_update.sql",
|
||||||
|
"11_datetime_update.sql",
|
||||||
|
]
|
||||||
|
Reference in New Issue
Block a user