Merge branch 'master' of github.com:MongoEngine/mongoengine into fix_count_documents_deprecation
This commit is contained in:
commit
aca4de728e
11
.gitignore
vendored
11
.gitignore
vendored
@ -1,8 +1,15 @@
|
|||||||
.*
|
|
||||||
!.gitignore
|
!.gitignore
|
||||||
*~
|
*~
|
||||||
*.py[co]
|
*.py[co]
|
||||||
.*.sw[po]
|
.*.sw[po]
|
||||||
|
.cache/
|
||||||
|
.coverage
|
||||||
|
.coveragerc
|
||||||
|
.env
|
||||||
|
.idea/
|
||||||
|
.pytest_cache/
|
||||||
|
.tox/
|
||||||
|
.eggs/
|
||||||
*.egg
|
*.egg
|
||||||
docs/.build
|
docs/.build
|
||||||
docs/_build
|
docs/_build
|
||||||
@ -13,8 +20,6 @@ env/
|
|||||||
.settings
|
.settings
|
||||||
.project
|
.project
|
||||||
.pydevproject
|
.pydevproject
|
||||||
tests/test_bugfix.py
|
|
||||||
htmlcov/
|
htmlcov/
|
||||||
venv
|
venv
|
||||||
venv3
|
venv3
|
||||||
scratchpad
|
|
||||||
|
@ -5,17 +5,12 @@ pylint:
|
|||||||
|
|
||||||
options:
|
options:
|
||||||
additional-builtins:
|
additional-builtins:
|
||||||
# add xrange and long as valid built-ins. In Python 3, xrange is
|
# add long as valid built-ins.
|
||||||
# translated into range and long is translated into int via 2to3 (see
|
|
||||||
# "use_2to3" in setup.py). This should be removed when we drop Python
|
|
||||||
# 2 support (which probably won't happen any time soon).
|
|
||||||
- xrange
|
|
||||||
- long
|
- long
|
||||||
|
|
||||||
pyflakes:
|
pyflakes:
|
||||||
disable:
|
disable:
|
||||||
# undefined variables are already covered by pylint (and exclude
|
# undefined variables are already covered by pylint (and exclude long)
|
||||||
# xrange & long)
|
|
||||||
- F821
|
- F821
|
||||||
|
|
||||||
ignore-paths:
|
ignore-paths:
|
||||||
|
12
.pre-commit-config.yaml
Normal file
12
.pre-commit-config.yaml
Normal file
@ -0,0 +1,12 @@
|
|||||||
|
fail_fast: false
|
||||||
|
repos:
|
||||||
|
- repo: https://github.com/ambv/black
|
||||||
|
rev: 19.10b0
|
||||||
|
hooks:
|
||||||
|
- id: black
|
||||||
|
- repo: https://gitlab.com/pycqa/flake8
|
||||||
|
rev: 3.8.0a2
|
||||||
|
hooks:
|
||||||
|
- id: flake8
|
||||||
|
additional_dependencies:
|
||||||
|
- flake8-import-order
|
52
.travis.yml
52
.travis.yml
@ -1,13 +1,10 @@
|
|||||||
# For full coverage, we'd have to test all supported Python, MongoDB, and
|
# For full coverage, we'd have to test all supported Python, MongoDB, and
|
||||||
# PyMongo combinations. However, that would result in an overly long build
|
# PyMongo combinations. However, that would result in an overly long build
|
||||||
# with a very large number of jobs, hence we only test a subset of all the
|
# with a very large number of jobs, hence we only test a subset of all the
|
||||||
# combinations:
|
# combinations.
|
||||||
# * MongoDB v3.4 & the latest PyMongo v3.x is currently the "main" setup,
|
# * Python3.7, MongoDB v3.4 & the latest PyMongo v3.x is currently the "main" setup,
|
||||||
# tested against Python v2.7, v3.5, v3.6, v3.7, v3.8 and PyPy.
|
# Other combinations are tested. See below for the details or check the travis jobs
|
||||||
# * Besides that, we test the lowest actively supported Python/MongoDB/PyMongo
|
|
||||||
# combination: MongoDB v3.4, PyMongo v3.4, Python v2.7.
|
|
||||||
# * MongoDB v3.6 is tested against Python v3.6, and PyMongo v3.6, v3.7, v3.8.
|
|
||||||
#
|
|
||||||
# We should periodically check MongoDB Server versions supported by MongoDB
|
# We should periodically check MongoDB Server versions supported by MongoDB
|
||||||
# Inc., add newly released versions to the test matrix, and remove versions
|
# Inc., add newly released versions to the test matrix, and remove versions
|
||||||
# which have reached their End of Life. See:
|
# which have reached their End of Life. See:
|
||||||
@ -16,44 +13,43 @@
|
|||||||
#
|
#
|
||||||
# Reminder: Update README.rst if you change MongoDB versions we test.
|
# Reminder: Update README.rst if you change MongoDB versions we test.
|
||||||
|
|
||||||
|
|
||||||
language: python
|
language: python
|
||||||
|
dist: xenial
|
||||||
python:
|
python:
|
||||||
- 2.7
|
|
||||||
- 3.5
|
- 3.5
|
||||||
- 3.6
|
- 3.6
|
||||||
- 3.7
|
- 3.7
|
||||||
- 3.8
|
- 3.8
|
||||||
- pypy
|
|
||||||
- pypy3
|
- pypy3
|
||||||
|
|
||||||
dist: xenial
|
|
||||||
|
|
||||||
env:
|
env:
|
||||||
global:
|
global:
|
||||||
- MONGODB_3_4=3.4.17
|
- MONGODB_3_4=3.4.17
|
||||||
- MONGODB_3_6=3.6.12
|
- MONGODB_3_6=3.6.12
|
||||||
- PYMONGO_3_10=3.10
|
- MONGODB_4_0=4.0.13
|
||||||
- PYMONGO_3_9=3.9
|
|
||||||
- PYMONGO_3_6=3.6
|
|
||||||
- PYMONGO_3_4=3.4
|
- PYMONGO_3_4=3.4
|
||||||
|
- PYMONGO_3_6=3.6
|
||||||
|
- PYMONGO_3_9=3.9
|
||||||
|
- PYMONGO_3_10=3.10
|
||||||
|
|
||||||
|
- MAIN_PYTHON_VERSION=3.7
|
||||||
matrix:
|
matrix:
|
||||||
- MONGODB=${MONGODB_3_4} PYMONGO=${PYMONGO_3_10}
|
- MONGODB=${MONGODB_3_4} PYMONGO=${PYMONGO_3_10}
|
||||||
|
|
||||||
matrix:
|
matrix:
|
||||||
|
|
||||||
# Finish the build as soon as one job fails
|
# Finish the build as soon as one job fails
|
||||||
fast_finish: true
|
fast_finish: true
|
||||||
|
|
||||||
include:
|
include:
|
||||||
- python: 2.7
|
|
||||||
env: MONGODB=${MONGODB_3_4} PYMONGO=${PYMONGO_3_4}
|
|
||||||
- python: 3.7
|
- python: 3.7
|
||||||
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_6}
|
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_6}
|
||||||
- python: 3.7
|
- python: 3.7
|
||||||
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_9}
|
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_9}
|
||||||
- python: 3.7
|
- python: 3.7
|
||||||
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_10}
|
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_10}
|
||||||
|
- python: 3.8
|
||||||
|
env: MONGODB=${MONGODB_4_0} PYMONGO=${PYMONGO_3_10}
|
||||||
|
|
||||||
install:
|
install:
|
||||||
# Install Mongo
|
# Install Mongo
|
||||||
@ -63,29 +59,23 @@ install:
|
|||||||
# Install Python dependencies.
|
# Install Python dependencies.
|
||||||
- pip install --upgrade pip
|
- pip install --upgrade pip
|
||||||
- pip install coveralls
|
- pip install coveralls
|
||||||
- pip install flake8 flake8-import-order
|
- pip install pre-commit
|
||||||
- pip install tox # tox 3.11.0 has requirement virtualenv>=14.0.0
|
- pip install tox
|
||||||
- pip install virtualenv # virtualenv>=14.0.0 has dropped Python 3.2 support (and pypy3 is based on py32)
|
|
||||||
# tox dryrun to setup the tox venv (we run a mock test).
|
# tox dryrun to setup the tox venv (we run a mock test).
|
||||||
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "-k=test_ci_placeholder"
|
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "-k=test_ci_placeholder"
|
||||||
# Install black for Python v3.7 only.
|
|
||||||
- if [[ $TRAVIS_PYTHON_VERSION == '3.7' ]]; then pip install black; fi
|
|
||||||
|
|
||||||
before_script:
|
before_script:
|
||||||
- mkdir ${PWD}/mongodb-linux-x86_64-${MONGODB}/data
|
- mkdir ${PWD}/mongodb-linux-x86_64-${MONGODB}/data
|
||||||
- ${PWD}/mongodb-linux-x86_64-${MONGODB}/bin/mongod --dbpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/data --logpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/mongodb.log --fork
|
- ${PWD}/mongodb-linux-x86_64-${MONGODB}/bin/mongod --dbpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/data --logpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/mongodb.log --fork
|
||||||
- if [[ $TRAVIS_PYTHON_VERSION == '2.7' ]]; then flake8 .; else echo "flake8 only runs on py27"; fi # Run flake8 for Python 2.7 only
|
# Run pre-commit hooks (black, flake8, etc) on entire codebase
|
||||||
- if [[ $TRAVIS_PYTHON_VERSION == '3.7' ]]; then black --check .; else echo "black only runs on py37"; fi # Run black for Python 3.7 only
|
- if [[ $TRAVIS_PYTHON_VERSION == $MAIN_PYTHON_VERSION ]]; then pre-commit run -a; else echo "pre-commit checks only runs on py37"; fi
|
||||||
- mongo --eval 'db.version();' # Make sure mongo is awake
|
- mongo --eval 'db.version();' # Make sure mongo is awake
|
||||||
|
|
||||||
script:
|
script:
|
||||||
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "--cov=mongoengine"
|
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "--cov=mongoengine"
|
||||||
|
|
||||||
# For now only submit coveralls for Python v2.7. Python v3.x currently shows
|
|
||||||
# 0% coverage. That's caused by 'use_2to3', which builds the py3-compatible
|
|
||||||
# code in a separate dir and runs tests on that.
|
|
||||||
after_success:
|
after_success:
|
||||||
- if [[ $TRAVIS_PYTHON_VERSION == '2.7' ]]; then coveralls --verbose; else echo "coveralls only sent for py27"; fi
|
- - if [[ $TRAVIS_PYTHON_VERSION == $MAIN_PYTHON_VERSION ]]; then coveralls --verbose; else echo "coveralls only sent for py37"; fi
|
||||||
|
|
||||||
notifications:
|
notifications:
|
||||||
irc: irc.freenode.org#mongoengine
|
irc: irc.freenode.org#mongoengine
|
||||||
@ -107,11 +97,11 @@ deploy:
|
|||||||
distributions: "sdist bdist_wheel"
|
distributions: "sdist bdist_wheel"
|
||||||
|
|
||||||
# Only deploy on tagged commits (aka GitHub releases) and only for the parent
|
# Only deploy on tagged commits (aka GitHub releases) and only for the parent
|
||||||
# repo's builds running Python v2.7 along with PyMongo v3.x and MongoDB v3.4.
|
# repo's builds running Python v3.7 along with PyMongo v3.x and MongoDB v3.4.
|
||||||
# We run Travis against many different Python, PyMongo, and MongoDB versions
|
# We run Travis against many different Python, PyMongo, and MongoDB versions
|
||||||
# and we don't want the deploy to occur multiple times).
|
# and we don't want the deploy to occur multiple times).
|
||||||
on:
|
on:
|
||||||
tags: true
|
tags: true
|
||||||
repo: MongoEngine/mongoengine
|
repo: MongoEngine/mongoengine
|
||||||
condition: ($PYMONGO = ${PYMONGO_3_10}) && ($MONGODB = ${MONGODB_3_4})
|
condition: ($PYMONGO = ${PYMONGO_3_10}) && ($MONGODB = ${MONGODB_3_4})
|
||||||
python: 2.7
|
python: 3.7
|
||||||
|
2
AUTHORS
2
AUTHORS
@ -255,3 +255,5 @@ that much better:
|
|||||||
* Filip Kucharczyk (https://github.com/Pacu2)
|
* Filip Kucharczyk (https://github.com/Pacu2)
|
||||||
* Eric Timmons (https://github.com/daewok)
|
* Eric Timmons (https://github.com/daewok)
|
||||||
* Matthew Simpson (https://github.com/mcsimps2)
|
* Matthew Simpson (https://github.com/mcsimps2)
|
||||||
|
* Leonardo Domingues (https://github.com/leodmgs)
|
||||||
|
* Agustin Barto (https://github.com/abarto)
|
||||||
|
@ -20,19 +20,43 @@ post to the `user group <http://groups.google.com/group/mongoengine-users>`
|
|||||||
Supported Interpreters
|
Supported Interpreters
|
||||||
----------------------
|
----------------------
|
||||||
|
|
||||||
MongoEngine supports CPython 2.7 and newer. Language
|
MongoEngine supports CPython 3.5 and newer as well as Pypy3.
|
||||||
features not supported by all interpreters can not be used.
|
Language features not supported by all interpreters can not be used.
|
||||||
The codebase is written in python 2 so you must be using python 2
|
|
||||||
when developing new features. Compatibility of the library with Python 3
|
Python3 codebase
|
||||||
relies on the 2to3 package that gets executed as part of the installation
|
----------------------
|
||||||
build. You should ensure that your code is properly converted by
|
|
||||||
`2to3 <http://docs.python.org/library/2to3.html>`_.
|
Since 0.20, the codebase is exclusively Python 3.
|
||||||
|
|
||||||
|
Earlier versions were exclusively Python2, and were relying on 2to3 to support Python3 installs.
|
||||||
|
Travis runs the tests against the main Python 3.x versions.
|
||||||
|
|
||||||
|
|
||||||
Style Guide
|
Style Guide
|
||||||
-----------
|
-----------
|
||||||
|
|
||||||
MongoEngine uses `black <https://github.com/python/black>`_ for code
|
MongoEngine's codebase is formatted with `black <https://github.com/python/black>`_, other tools like
|
||||||
formatting.
|
flake8 are also used. Those tools will run as part of the CI and will fail in case the code is not formatted properly.
|
||||||
|
|
||||||
|
To install all development tools, simply run the following commands:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ python -m pip install -r requirements-dev.txt
|
||||||
|
|
||||||
|
|
||||||
|
You can install `pre-commit <https://pre-commit.com/>`_ into your git hooks,
|
||||||
|
to automatically check and fix any formatting issue before creating a
|
||||||
|
git commit.
|
||||||
|
|
||||||
|
To enable ``pre-commit`` simply run:
|
||||||
|
|
||||||
|
.. code-block:: console
|
||||||
|
|
||||||
|
$ pre-commit install
|
||||||
|
|
||||||
|
See the ``.pre-commit-config.yaml`` configuration file for more information
|
||||||
|
on how it works.
|
||||||
|
|
||||||
Testing
|
Testing
|
||||||
-------
|
-------
|
||||||
|
17
README.rst
17
README.rst
@ -26,15 +26,15 @@ an `API reference <https://mongoengine-odm.readthedocs.io/apireference.html>`_.
|
|||||||
|
|
||||||
Supported MongoDB Versions
|
Supported MongoDB Versions
|
||||||
==========================
|
==========================
|
||||||
MongoEngine is currently tested against MongoDB v3.4 and v3.6. Future versions
|
MongoEngine is currently tested against MongoDB v3.4, v3.6 and v4.0. Future versions
|
||||||
should be supported as well, but aren't actively tested at the moment. Make
|
should be supported as well, but aren't actively tested at the moment. Make
|
||||||
sure to open an issue or submit a pull request if you experience any problems
|
sure to open an issue or submit a pull request if you experience any problems
|
||||||
with MongoDB version > 3.6.
|
with MongoDB version > 4.0.
|
||||||
|
|
||||||
Installation
|
Installation
|
||||||
============
|
============
|
||||||
We recommend the use of `virtualenv <https://virtualenv.pypa.io/>`_ and of
|
We recommend the use of `virtualenv <https://virtualenv.pypa.io/>`_ and of
|
||||||
`pip <https://pip.pypa.io/>`_. You can then use ``pip install -U mongoengine``.
|
`pip <https://pip.pypa.io/>`_. You can then use ``python -m pip install -U mongoengine``.
|
||||||
You may also have `setuptools <http://peak.telecommunity.com/DevCenter/setuptools>`_
|
You may also have `setuptools <http://peak.telecommunity.com/DevCenter/setuptools>`_
|
||||||
and thus you can use ``easy_install -U mongoengine``. Another option is
|
and thus you can use ``easy_install -U mongoengine``. Another option is
|
||||||
`pipenv <https://docs.pipenv.org/>`_. You can then use ``pipenv install mongoengine``
|
`pipenv <https://docs.pipenv.org/>`_. You can then use ``pipenv install mongoengine``
|
||||||
@ -42,13 +42,14 @@ to both create the virtual environment and install the package. Otherwise, you c
|
|||||||
download the source from `GitHub <http://github.com/MongoEngine/mongoengine>`_ and
|
download the source from `GitHub <http://github.com/MongoEngine/mongoengine>`_ and
|
||||||
run ``python setup.py install``.
|
run ``python setup.py install``.
|
||||||
|
|
||||||
|
The support for Python2 was dropped with MongoEngine 0.20.0
|
||||||
|
|
||||||
Dependencies
|
Dependencies
|
||||||
============
|
============
|
||||||
All of the dependencies can easily be installed via `pip <https://pip.pypa.io/>`_.
|
All of the dependencies can easily be installed via `python -m pip <https://pip.pypa.io/>`_.
|
||||||
At the very least, you'll need these two packages to use MongoEngine:
|
At the very least, you'll need these two packages to use MongoEngine:
|
||||||
|
|
||||||
- pymongo>=3.4
|
- pymongo>=3.4
|
||||||
- six>=1.10.0
|
|
||||||
|
|
||||||
If you utilize a ``DateTimeField``, you might also use a more flexible date parser:
|
If you utilize a ``DateTimeField``, you might also use a more flexible date parser:
|
||||||
|
|
||||||
@ -58,6 +59,10 @@ If you need to use an ``ImageField`` or ``ImageGridFsProxy``:
|
|||||||
|
|
||||||
- Pillow>=2.0.0
|
- Pillow>=2.0.0
|
||||||
|
|
||||||
|
If you need to use signals:
|
||||||
|
|
||||||
|
- blinker>=1.3
|
||||||
|
|
||||||
Examples
|
Examples
|
||||||
========
|
========
|
||||||
Some simple examples of what MongoEngine code looks like:
|
Some simple examples of what MongoEngine code looks like:
|
||||||
@ -125,7 +130,7 @@ installed in your environment and then:
|
|||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
# Install tox
|
# Install tox
|
||||||
$ pip install tox
|
$ python -m pip install tox
|
||||||
# Run the test suites
|
# Run the test suites
|
||||||
$ tox
|
$ tox
|
||||||
|
|
||||||
|
@ -4,12 +4,14 @@ import timeit
|
|||||||
def main():
|
def main():
|
||||||
setup = """
|
setup = """
|
||||||
from pymongo import MongoClient
|
from pymongo import MongoClient
|
||||||
|
|
||||||
connection = MongoClient()
|
connection = MongoClient()
|
||||||
connection.drop_database('mongoengine_benchmark_test')
|
connection.drop_database('mongoengine_benchmark_test')
|
||||||
"""
|
"""
|
||||||
|
|
||||||
stmt = """
|
stmt = """
|
||||||
from pymongo import MongoClient
|
from pymongo import MongoClient
|
||||||
|
|
||||||
connection = MongoClient()
|
connection = MongoClient()
|
||||||
|
|
||||||
db = connection.mongoengine_benchmark_test
|
db = connection.mongoengine_benchmark_test
|
||||||
@ -56,6 +58,7 @@ myNoddys = noddy.find()
|
|||||||
|
|
||||||
setup = """
|
setup = """
|
||||||
from pymongo import MongoClient
|
from pymongo import MongoClient
|
||||||
|
|
||||||
connection = MongoClient()
|
connection = MongoClient()
|
||||||
connection.drop_database('mongoengine_benchmark_test')
|
connection.drop_database('mongoengine_benchmark_test')
|
||||||
connection.close()
|
connection.close()
|
||||||
|
@ -9,10 +9,31 @@ Development
|
|||||||
- When using pymongo >= 3.7, make use of Collection.count_documents instead of Collection.count
|
- When using pymongo >= 3.7, make use of Collection.count_documents instead of Collection.count
|
||||||
and Cursor.count that got deprecated in pymongo >= 3.7.
|
and Cursor.count that got deprecated in pymongo >= 3.7.
|
||||||
This should have a negative impact on performance of count see Issue #2219
|
This should have a negative impact on performance of count see Issue #2219
|
||||||
|
- Fix a bug that made the queryset drop the read_preference after clone().
|
||||||
|
- Fix the behavior of Doc.objects.limit(0) which should return all documents (similar to mongodb) #2311
|
||||||
|
|
||||||
|
Changes in 0.20.0
|
||||||
|
=================
|
||||||
|
- ATTENTION: Drop support for Python2
|
||||||
|
- Add Mongo 4.0 to Travis
|
||||||
|
- Fix error when setting a string as a ComplexDateTimeField #2253
|
||||||
|
- Bump development Status classifier to Production/Stable #2232
|
||||||
|
- Improve Queryset.get to avoid confusing MultipleObjectsReturned message in case multiple match are found #630
|
||||||
|
- Fixed a bug causing inaccurate query results, while combining ``__raw__`` and regular filters for the same field #2264
|
||||||
|
- Add support for the `elemMatch` projection operator in .fields() (e.g BlogPost.objects.fields(elemMatch__comments="test")) #2267
|
||||||
|
- DictField validate failed without default connection (bug introduced in 0.19.0) #2239
|
||||||
|
- Remove methods that were deprecated years ago:
|
||||||
|
- name parameter in Field constructor e.g `StringField(name="...")`, was replaced by db_field
|
||||||
|
- Queryset.slave_okay() was deprecated since pymongo3
|
||||||
|
- dropDups was dropped with MongoDB3
|
||||||
|
- ``Queryset._ensure_indexes`` and ``Queryset.ensure_indexes``, the right method to use is ``Document.ensure_indexes``
|
||||||
|
- Added pre-commit for development/CI #2212
|
||||||
|
- Renamed requirements-lint.txt to requirements-dev.txt #2212
|
||||||
|
- Support for setting ReadConcern #2255
|
||||||
|
|
||||||
Changes in 0.19.1
|
Changes in 0.19.1
|
||||||
=================
|
=================
|
||||||
- Requires Pillow < 7.0.0 as it dropped Python2 support
|
- Tests require Pillow < 7.0.0 as it dropped Python2 support
|
||||||
- DEPRECATION: The interface of ``QuerySet.aggregate`` method was changed, it no longer takes an unpacked list of
|
- DEPRECATION: The interface of ``QuerySet.aggregate`` method was changed, it no longer takes an unpacked list of
|
||||||
pipeline steps (*pipeline) but simply takes the pipeline list just like ``pymongo.Collection.aggregate`` does. #2079
|
pipeline steps (*pipeline) but simply takes the pipeline list just like ``pymongo.Collection.aggregate`` does. #2079
|
||||||
|
|
||||||
|
@ -10,4 +10,3 @@ If this is a requirement for your project, check the alternative: `uMongo`_ and
|
|||||||
|
|
||||||
.. _uMongo: https://umongo.readthedocs.io/
|
.. _uMongo: https://umongo.readthedocs.io/
|
||||||
.. _MotorEngine: https://motorengine.readthedocs.io/
|
.. _MotorEngine: https://motorengine.readthedocs.io/
|
||||||
|
|
||||||
|
@ -555,7 +555,6 @@ There are a few top level defaults for all indexes that can be set::
|
|||||||
'index_background': True,
|
'index_background': True,
|
||||||
'index_cls': False,
|
'index_cls': False,
|
||||||
'auto_create_index': True,
|
'auto_create_index': True,
|
||||||
'index_drop_dups': True,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
@ -574,11 +573,6 @@ There are a few top level defaults for all indexes that can be set::
|
|||||||
in systems where indexes are managed separately. Disabling this will improve
|
in systems where indexes are managed separately. Disabling this will improve
|
||||||
performance.
|
performance.
|
||||||
|
|
||||||
:attr:`index_drop_dups` (Optional)
|
|
||||||
Set the default value for if an index should drop duplicates
|
|
||||||
Since MongoDB 3.0 drop_dups is not supported anymore. Raises a Warning
|
|
||||||
and has no effect
|
|
||||||
|
|
||||||
|
|
||||||
Compound Indexes and Indexing sub documents
|
Compound Indexes and Indexing sub documents
|
||||||
-------------------------------------------
|
-------------------------------------------
|
||||||
|
@ -12,7 +12,7 @@ MongoEngine is available on PyPI, so you can use :program:`pip`:
|
|||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
$ pip install mongoengine
|
$ python -m pip install mongoengine
|
||||||
|
|
||||||
Alternatively, if you don't have setuptools installed, `download it from PyPi
|
Alternatively, if you don't have setuptools installed, `download it from PyPi
|
||||||
<http://pypi.python.org/pypi/mongoengine/>`_ and run
|
<http://pypi.python.org/pypi/mongoengine/>`_ and run
|
||||||
|
@ -7,7 +7,7 @@ MongoDB. To install it, simply run
|
|||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
$ pip install -U mongoengine
|
$ python -m pip install -U mongoengine
|
||||||
|
|
||||||
:doc:`tutorial`
|
:doc:`tutorial`
|
||||||
A quick tutorial building a tumblelog to get you up and running with
|
A quick tutorial building a tumblelog to get you up and running with
|
||||||
@ -91,4 +91,3 @@ Indices and tables
|
|||||||
* :ref:`genindex`
|
* :ref:`genindex`
|
||||||
* :ref:`modindex`
|
* :ref:`modindex`
|
||||||
* :ref:`search`
|
* :ref:`search`
|
||||||
|
|
||||||
|
@ -18,7 +18,7 @@ location --- running it locally will be easier, but if that is not an option
|
|||||||
then it may be run on a remote server. If you haven't installed MongoEngine,
|
then it may be run on a remote server. If you haven't installed MongoEngine,
|
||||||
simply use pip to install it like so::
|
simply use pip to install it like so::
|
||||||
|
|
||||||
$ pip install mongoengine
|
$ python -m pip install mongoengine
|
||||||
|
|
||||||
Before we can start using MongoEngine, we need to tell it how to connect to our
|
Before we can start using MongoEngine, we need to tell it how to connect to our
|
||||||
instance of :program:`mongod`. For this we use the :func:`~mongoengine.connect`
|
instance of :program:`mongod`. For this we use the :func:`~mongoengine.connect`
|
||||||
|
@ -85,10 +85,10 @@ by default from now on.
|
|||||||
|
|
||||||
The 0.8.7 package on pypi was corrupted. If upgrading from 0.8.7 to 0.9.0 please follow: ::
|
The 0.8.7 package on pypi was corrupted. If upgrading from 0.8.7 to 0.9.0 please follow: ::
|
||||||
|
|
||||||
pip uninstall pymongo
|
python -m pip uninstall pymongo
|
||||||
pip uninstall mongoengine
|
python -m pip uninstall mongoengine
|
||||||
pip install pymongo==2.8
|
python -m pip install pymongo==2.8
|
||||||
pip install mongoengine
|
python -m pip install mongoengine
|
||||||
|
|
||||||
0.8.7
|
0.8.7
|
||||||
*****
|
*****
|
||||||
@ -153,7 +153,7 @@ inherited classes like so: ::
|
|||||||
|
|
||||||
# 4. Remove indexes
|
# 4. Remove indexes
|
||||||
info = collection.index_information()
|
info = collection.index_information()
|
||||||
indexes_to_drop = [key for key, value in info.iteritems()
|
indexes_to_drop = [key for key, value in info.items()
|
||||||
if '_types' in dict(value['key'])]
|
if '_types' in dict(value['key'])]
|
||||||
for index in indexes_to_drop:
|
for index in indexes_to_drop:
|
||||||
collection.drop_index(index)
|
collection.drop_index(index)
|
||||||
|
@ -28,7 +28,7 @@ __all__ = (
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
VERSION = (0, 19, 1)
|
VERSION = (0, 20, 0)
|
||||||
|
|
||||||
|
|
||||||
def get_version():
|
def get_version():
|
||||||
|
@ -1,8 +1,6 @@
|
|||||||
import weakref
|
import weakref
|
||||||
|
|
||||||
from bson import DBRef
|
from bson import DBRef
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.errors import DoesNotExist, MultipleObjectsReturned
|
from mongoengine.errors import DoesNotExist, MultipleObjectsReturned
|
||||||
@ -53,7 +51,7 @@ class BaseDict(dict):
|
|||||||
if isinstance(instance, BaseDocument):
|
if isinstance(instance, BaseDocument):
|
||||||
self._instance = weakref.proxy(instance)
|
self._instance = weakref.proxy(instance)
|
||||||
self._name = name
|
self._name = name
|
||||||
super(BaseDict, self).__init__(dict_items)
|
super().__init__(dict_items)
|
||||||
|
|
||||||
def get(self, key, default=None):
|
def get(self, key, default=None):
|
||||||
# get does not use __getitem__ by default so we must override it as well
|
# get does not use __getitem__ by default so we must override it as well
|
||||||
@ -63,18 +61,18 @@ class BaseDict(dict):
|
|||||||
return default
|
return default
|
||||||
|
|
||||||
def __getitem__(self, key):
|
def __getitem__(self, key):
|
||||||
value = super(BaseDict, self).__getitem__(key)
|
value = super().__getitem__(key)
|
||||||
|
|
||||||
EmbeddedDocument = _import_class("EmbeddedDocument")
|
EmbeddedDocument = _import_class("EmbeddedDocument")
|
||||||
if isinstance(value, EmbeddedDocument) and value._instance is None:
|
if isinstance(value, EmbeddedDocument) and value._instance is None:
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
||||||
value = BaseDict(value, None, "%s.%s" % (self._name, key))
|
value = BaseDict(value, None, "{}.{}".format(self._name, key))
|
||||||
super(BaseDict, self).__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, list) and not isinstance(value, BaseList):
|
elif isinstance(value, list) and not isinstance(value, BaseList):
|
||||||
value = BaseList(value, None, "%s.%s" % (self._name, key))
|
value = BaseList(value, None, "{}.{}".format(self._name, key))
|
||||||
super(BaseDict, self).__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
return value
|
return value
|
||||||
|
|
||||||
@ -99,7 +97,7 @@ class BaseDict(dict):
|
|||||||
def _mark_as_changed(self, key=None):
|
def _mark_as_changed(self, key=None):
|
||||||
if hasattr(self._instance, "_mark_as_changed"):
|
if hasattr(self._instance, "_mark_as_changed"):
|
||||||
if key:
|
if key:
|
||||||
self._instance._mark_as_changed("%s.%s" % (self._name, key))
|
self._instance._mark_as_changed("{}.{}".format(self._name, key))
|
||||||
else:
|
else:
|
||||||
self._instance._mark_as_changed(self._name)
|
self._instance._mark_as_changed(self._name)
|
||||||
|
|
||||||
@ -117,13 +115,13 @@ class BaseList(list):
|
|||||||
if isinstance(instance, BaseDocument):
|
if isinstance(instance, BaseDocument):
|
||||||
self._instance = weakref.proxy(instance)
|
self._instance = weakref.proxy(instance)
|
||||||
self._name = name
|
self._name = name
|
||||||
super(BaseList, self).__init__(list_items)
|
super().__init__(list_items)
|
||||||
|
|
||||||
def __getitem__(self, key):
|
def __getitem__(self, key):
|
||||||
# change index to positive value because MongoDB does not support negative one
|
# change index to positive value because MongoDB does not support negative one
|
||||||
if isinstance(key, int) and key < 0:
|
if isinstance(key, int) and key < 0:
|
||||||
key = len(self) + key
|
key = len(self) + key
|
||||||
value = super(BaseList, self).__getitem__(key)
|
value = super().__getitem__(key)
|
||||||
|
|
||||||
if isinstance(key, slice):
|
if isinstance(key, slice):
|
||||||
# When receiving a slice operator, we don't convert the structure and bind
|
# When receiving a slice operator, we don't convert the structure and bind
|
||||||
@ -135,19 +133,18 @@ class BaseList(list):
|
|||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
||||||
# Replace dict by BaseDict
|
# Replace dict by BaseDict
|
||||||
value = BaseDict(value, None, "%s.%s" % (self._name, key))
|
value = BaseDict(value, None, "{}.{}".format(self._name, key))
|
||||||
super(BaseList, self).__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, list) and not isinstance(value, BaseList):
|
elif isinstance(value, list) and not isinstance(value, BaseList):
|
||||||
# Replace list by BaseList
|
# Replace list by BaseList
|
||||||
value = BaseList(value, None, "%s.%s" % (self._name, key))
|
value = BaseList(value, None, "{}.{}".format(self._name, key))
|
||||||
super(BaseList, self).__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
return value
|
return value
|
||||||
|
|
||||||
def __iter__(self):
|
def __iter__(self):
|
||||||
for v in super(BaseList, self).__iter__():
|
yield from super().__iter__()
|
||||||
yield v
|
|
||||||
|
|
||||||
def __getstate__(self):
|
def __getstate__(self):
|
||||||
self.instance = None
|
self.instance = None
|
||||||
@ -165,7 +162,7 @@ class BaseList(list):
|
|||||||
# instead, we simply marks the whole list as changed
|
# instead, we simply marks the whole list as changed
|
||||||
changed_key = None
|
changed_key = None
|
||||||
|
|
||||||
result = super(BaseList, self).__setitem__(key, value)
|
result = super().__setitem__(key, value)
|
||||||
self._mark_as_changed(changed_key)
|
self._mark_as_changed(changed_key)
|
||||||
return result
|
return result
|
||||||
|
|
||||||
@ -180,30 +177,19 @@ class BaseList(list):
|
|||||||
__iadd__ = mark_as_changed_wrapper(list.__iadd__)
|
__iadd__ = mark_as_changed_wrapper(list.__iadd__)
|
||||||
__imul__ = mark_as_changed_wrapper(list.__imul__)
|
__imul__ = mark_as_changed_wrapper(list.__imul__)
|
||||||
|
|
||||||
if six.PY2:
|
|
||||||
# Under py3 __setslice__, __delslice__ and __getslice__
|
|
||||||
# are replaced by __setitem__, __delitem__ and __getitem__ with a slice as parameter
|
|
||||||
# so we mimic this under python 2
|
|
||||||
def __setslice__(self, i, j, sequence):
|
|
||||||
return self.__setitem__(slice(i, j), sequence)
|
|
||||||
|
|
||||||
def __delslice__(self, i, j):
|
|
||||||
return self.__delitem__(slice(i, j))
|
|
||||||
|
|
||||||
def __getslice__(self, i, j):
|
|
||||||
return self.__getitem__(slice(i, j))
|
|
||||||
|
|
||||||
def _mark_as_changed(self, key=None):
|
def _mark_as_changed(self, key=None):
|
||||||
if hasattr(self._instance, "_mark_as_changed"):
|
if hasattr(self._instance, "_mark_as_changed"):
|
||||||
if key:
|
if key:
|
||||||
self._instance._mark_as_changed("%s.%s" % (self._name, key % len(self)))
|
self._instance._mark_as_changed(
|
||||||
|
"{}.{}".format(self._name, key % len(self))
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
self._instance._mark_as_changed(self._name)
|
self._instance._mark_as_changed(self._name)
|
||||||
|
|
||||||
|
|
||||||
class EmbeddedDocumentList(BaseList):
|
class EmbeddedDocumentList(BaseList):
|
||||||
def __init__(self, list_items, instance, name):
|
def __init__(self, list_items, instance, name):
|
||||||
super(EmbeddedDocumentList, self).__init__(list_items, instance, name)
|
super().__init__(list_items, instance, name)
|
||||||
self._instance = instance
|
self._instance = instance
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
@ -213,7 +199,7 @@ class EmbeddedDocumentList(BaseList):
|
|||||||
"""
|
"""
|
||||||
for key, expected_value in kwargs.items():
|
for key, expected_value in kwargs.items():
|
||||||
doc_val = getattr(embedded_doc, key)
|
doc_val = getattr(embedded_doc, key)
|
||||||
if doc_val != expected_value and six.text_type(doc_val) != expected_value:
|
if doc_val != expected_value and str(doc_val) != expected_value:
|
||||||
return False
|
return False
|
||||||
return True
|
return True
|
||||||
|
|
||||||
@ -303,11 +289,11 @@ class EmbeddedDocumentList(BaseList):
|
|||||||
|
|
||||||
def create(self, **values):
|
def create(self, **values):
|
||||||
"""
|
"""
|
||||||
Creates a new embedded document and saves it to the database.
|
Creates a new instance of the EmbeddedDocument and appends it to this EmbeddedDocumentList.
|
||||||
|
|
||||||
.. note::
|
.. note::
|
||||||
The embedded document changes are not automatically saved
|
the instance of the EmbeddedDocument is not automatically saved to the database.
|
||||||
to the database after calling this method.
|
You still need to call .save() on the parent Document.
|
||||||
|
|
||||||
:param values: A dictionary of values for the embedded document.
|
:param values: A dictionary of values for the embedded document.
|
||||||
:return: The new embedded document instance.
|
:return: The new embedded document instance.
|
||||||
@ -368,13 +354,13 @@ class EmbeddedDocumentList(BaseList):
|
|||||||
return len(values)
|
return len(values)
|
||||||
|
|
||||||
|
|
||||||
class StrictDict(object):
|
class StrictDict:
|
||||||
__slots__ = ()
|
__slots__ = ()
|
||||||
_special_fields = {"get", "pop", "iteritems", "items", "keys", "create"}
|
_special_fields = {"get", "pop", "iteritems", "items", "keys", "create"}
|
||||||
_classes = {}
|
_classes = {}
|
||||||
|
|
||||||
def __init__(self, **kwargs):
|
def __init__(self, **kwargs):
|
||||||
for k, v in iteritems(kwargs):
|
for k, v in kwargs.items():
|
||||||
setattr(self, k, v)
|
setattr(self, k, v)
|
||||||
|
|
||||||
def __getitem__(self, key):
|
def __getitem__(self, key):
|
||||||
@ -422,13 +408,13 @@ class StrictDict(object):
|
|||||||
return (key for key in self.__slots__ if hasattr(self, key))
|
return (key for key in self.__slots__ if hasattr(self, key))
|
||||||
|
|
||||||
def __len__(self):
|
def __len__(self):
|
||||||
return len(list(iteritems(self)))
|
return len(list(self.items()))
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
return self.items() == other.items()
|
return list(self.items()) == list(other.items())
|
||||||
|
|
||||||
def __ne__(self, other):
|
def __ne__(self, other):
|
||||||
return self.items() != other.items()
|
return not (self == other)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def create(cls, allowed_keys):
|
def create(cls, allowed_keys):
|
||||||
@ -443,7 +429,7 @@ class StrictDict(object):
|
|||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "{%s}" % ", ".join(
|
return "{%s}" % ", ".join(
|
||||||
'"{0!s}": {1!r}'.format(k, v) for k, v in self.items()
|
'"{!s}": {!r}'.format(k, v) for k, v in self.items()
|
||||||
)
|
)
|
||||||
|
|
||||||
cls._classes[allowed_keys] = SpecificStrictDict
|
cls._classes[allowed_keys] = SpecificStrictDict
|
||||||
@ -468,9 +454,7 @@ class LazyReference(DBRef):
|
|||||||
self.document_type = document_type
|
self.document_type = document_type
|
||||||
self._cached_doc = cached_doc
|
self._cached_doc = cached_doc
|
||||||
self.passthrough = passthrough
|
self.passthrough = passthrough
|
||||||
super(LazyReference, self).__init__(
|
super().__init__(self.document_type._get_collection_name(), pk)
|
||||||
self.document_type._get_collection_name(), pk
|
|
||||||
)
|
|
||||||
|
|
||||||
def __getitem__(self, name):
|
def __getitem__(self, name):
|
||||||
if not self.passthrough:
|
if not self.passthrough:
|
||||||
@ -488,4 +472,4 @@ class LazyReference(DBRef):
|
|||||||
raise AttributeError()
|
raise AttributeError()
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "<LazyReference(%s, %r)>" % (self.document_type, self.pk)
|
return "<LazyReference({}, {!r})>".format(self.document_type, self.pk)
|
||||||
|
@ -1,11 +1,10 @@
|
|||||||
import copy
|
import copy
|
||||||
|
|
||||||
import numbers
|
import numbers
|
||||||
from functools import partial
|
from functools import partial
|
||||||
|
|
||||||
from bson import DBRef, ObjectId, SON, json_util
|
from bson import DBRef, ObjectId, SON, json_util
|
||||||
import pymongo
|
import pymongo
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
from mongoengine.base.common import get_document
|
from mongoengine.base.common import get_document
|
||||||
@ -25,14 +24,13 @@ from mongoengine.errors import (
|
|||||||
OperationError,
|
OperationError,
|
||||||
ValidationError,
|
ValidationError,
|
||||||
)
|
)
|
||||||
from mongoengine.python_support import Hashable
|
|
||||||
|
|
||||||
__all__ = ("BaseDocument", "NON_FIELD_ERRORS")
|
__all__ = ("BaseDocument", "NON_FIELD_ERRORS")
|
||||||
|
|
||||||
NON_FIELD_ERRORS = "__all__"
|
NON_FIELD_ERRORS = "__all__"
|
||||||
|
|
||||||
|
|
||||||
class BaseDocument(object):
|
class BaseDocument:
|
||||||
# TODO simplify how `_changed_fields` is used.
|
# TODO simplify how `_changed_fields` is used.
|
||||||
# Currently, handling of `_changed_fields` seems unnecessarily convoluted:
|
# Currently, handling of `_changed_fields` seems unnecessarily convoluted:
|
||||||
# 1. `BaseDocument` defines `_changed_fields` in its `__slots__`, yet it's
|
# 1. `BaseDocument` defines `_changed_fields` in its `__slots__`, yet it's
|
||||||
@ -92,10 +90,10 @@ class BaseDocument(object):
|
|||||||
# if so raise an Exception.
|
# if so raise an Exception.
|
||||||
if not self._dynamic and (self._meta.get("strict", True) or _created):
|
if not self._dynamic and (self._meta.get("strict", True) or _created):
|
||||||
_undefined_fields = set(values.keys()) - set(
|
_undefined_fields = set(values.keys()) - set(
|
||||||
self._fields.keys() + ["id", "pk", "_cls", "_text_score"]
|
list(self._fields.keys()) + ["id", "pk", "_cls", "_text_score"]
|
||||||
)
|
)
|
||||||
if _undefined_fields:
|
if _undefined_fields:
|
||||||
msg = ('The fields "{0}" do not exist on the document "{1}"').format(
|
msg = ('The fields "{}" do not exist on the document "{}"').format(
|
||||||
_undefined_fields, self._class_name
|
_undefined_fields, self._class_name
|
||||||
)
|
)
|
||||||
raise FieldDoesNotExist(msg)
|
raise FieldDoesNotExist(msg)
|
||||||
@ -110,7 +108,7 @@ class BaseDocument(object):
|
|||||||
# Assign default values to the instance.
|
# Assign default values to the instance.
|
||||||
# We set default values only for fields loaded from DB. See
|
# We set default values only for fields loaded from DB. See
|
||||||
# https://github.com/mongoengine/mongoengine/issues/399 for more info.
|
# https://github.com/mongoengine/mongoengine/issues/399 for more info.
|
||||||
for key, field in iteritems(self._fields):
|
for key, field in self._fields.items():
|
||||||
if self._db_field_map.get(key, key) in __only_fields:
|
if self._db_field_map.get(key, key) in __only_fields:
|
||||||
continue
|
continue
|
||||||
value = getattr(self, key, None)
|
value = getattr(self, key, None)
|
||||||
@ -122,14 +120,14 @@ class BaseDocument(object):
|
|||||||
# Set passed values after initialisation
|
# Set passed values after initialisation
|
||||||
if self._dynamic:
|
if self._dynamic:
|
||||||
dynamic_data = {}
|
dynamic_data = {}
|
||||||
for key, value in iteritems(values):
|
for key, value in values.items():
|
||||||
if key in self._fields or key == "_id":
|
if key in self._fields or key == "_id":
|
||||||
setattr(self, key, value)
|
setattr(self, key, value)
|
||||||
else:
|
else:
|
||||||
dynamic_data[key] = value
|
dynamic_data[key] = value
|
||||||
else:
|
else:
|
||||||
FileField = _import_class("FileField")
|
FileField = _import_class("FileField")
|
||||||
for key, value in iteritems(values):
|
for key, value in values.items():
|
||||||
key = self._reverse_db_field_map.get(key, key)
|
key = self._reverse_db_field_map.get(key, key)
|
||||||
if key in self._fields or key in ("id", "pk", "_cls"):
|
if key in self._fields or key in ("id", "pk", "_cls"):
|
||||||
if __auto_convert and value is not None:
|
if __auto_convert and value is not None:
|
||||||
@ -145,7 +143,7 @@ class BaseDocument(object):
|
|||||||
|
|
||||||
if self._dynamic:
|
if self._dynamic:
|
||||||
self._dynamic_lock = False
|
self._dynamic_lock = False
|
||||||
for key, value in iteritems(dynamic_data):
|
for key, value in dynamic_data.items():
|
||||||
setattr(self, key, value)
|
setattr(self, key, value)
|
||||||
|
|
||||||
# Flag initialised
|
# Flag initialised
|
||||||
@ -163,7 +161,7 @@ class BaseDocument(object):
|
|||||||
default = default()
|
default = default()
|
||||||
setattr(self, field_name, default)
|
setattr(self, field_name, default)
|
||||||
else:
|
else:
|
||||||
super(BaseDocument, self).__delattr__(*args, **kwargs)
|
super().__delattr__(*args, **kwargs)
|
||||||
|
|
||||||
def __setattr__(self, name, value):
|
def __setattr__(self, name, value):
|
||||||
# Handle dynamic data only if an initialised dynamic document
|
# Handle dynamic data only if an initialised dynamic document
|
||||||
@ -210,9 +208,9 @@ class BaseDocument(object):
|
|||||||
and self__created
|
and self__created
|
||||||
and name == self._meta.get("id_field")
|
and name == self._meta.get("id_field")
|
||||||
):
|
):
|
||||||
super(BaseDocument, self).__setattr__("_created", False)
|
super().__setattr__("_created", False)
|
||||||
|
|
||||||
super(BaseDocument, self).__setattr__(name, value)
|
super().__setattr__(name, value)
|
||||||
|
|
||||||
def __getstate__(self):
|
def __getstate__(self):
|
||||||
data = {}
|
data = {}
|
||||||
@ -288,16 +286,13 @@ class BaseDocument(object):
|
|||||||
except (UnicodeEncodeError, UnicodeDecodeError):
|
except (UnicodeEncodeError, UnicodeDecodeError):
|
||||||
u = "[Bad Unicode data]"
|
u = "[Bad Unicode data]"
|
||||||
repr_type = str if u is None else type(u)
|
repr_type = str if u is None else type(u)
|
||||||
return repr_type("<%s: %s>" % (self.__class__.__name__, u))
|
return repr_type("<{}: {}>".format(self.__class__.__name__, u))
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
# TODO this could be simpler?
|
# TODO this could be simpler?
|
||||||
if hasattr(self, "__unicode__"):
|
if hasattr(self, "__unicode__"):
|
||||||
if six.PY3:
|
return self.__unicode__()
|
||||||
return self.__unicode__()
|
return "%s object" % self.__class__.__name__
|
||||||
else:
|
|
||||||
return six.text_type(self).encode("utf-8")
|
|
||||||
return six.text_type("%s object" % self.__class__.__name__)
|
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
if (
|
if (
|
||||||
@ -446,7 +441,7 @@ class BaseDocument(object):
|
|||||||
pk = self.pk
|
pk = self.pk
|
||||||
elif self._instance and hasattr(self._instance, "pk"):
|
elif self._instance and hasattr(self._instance, "pk"):
|
||||||
pk = self._instance.pk
|
pk = self._instance.pk
|
||||||
message = "ValidationError (%s:%s) " % (self._class_name, pk)
|
message = "ValidationError ({}:{}) ".format(self._class_name, pk)
|
||||||
raise ValidationError(message, errors=errors)
|
raise ValidationError(message, errors=errors)
|
||||||
|
|
||||||
def to_json(self, *args, **kwargs):
|
def to_json(self, *args, **kwargs):
|
||||||
@ -519,7 +514,7 @@ class BaseDocument(object):
|
|||||||
if "." in key:
|
if "." in key:
|
||||||
key, rest = key.split(".", 1)
|
key, rest = key.split(".", 1)
|
||||||
key = self._db_field_map.get(key, key)
|
key = self._db_field_map.get(key, key)
|
||||||
key = "%s.%s" % (key, rest)
|
key = "{}.{}".format(key, rest)
|
||||||
else:
|
else:
|
||||||
key = self._db_field_map.get(key, key)
|
key = self._db_field_map.get(key, key)
|
||||||
|
|
||||||
@ -578,10 +573,10 @@ class BaseDocument(object):
|
|||||||
if not hasattr(data, "items"):
|
if not hasattr(data, "items"):
|
||||||
iterator = enumerate(data)
|
iterator = enumerate(data)
|
||||||
else:
|
else:
|
||||||
iterator = iteritems(data)
|
iterator = data.items()
|
||||||
|
|
||||||
for index_or_key, value in iterator:
|
for index_or_key, value in iterator:
|
||||||
item_key = "%s%s." % (base_key, index_or_key)
|
item_key = "{}{}.".format(base_key, index_or_key)
|
||||||
# don't check anything lower if this key is already marked
|
# don't check anything lower if this key is already marked
|
||||||
# as changed.
|
# as changed.
|
||||||
if item_key[:-1] in changed_fields:
|
if item_key[:-1] in changed_fields:
|
||||||
@ -589,7 +584,7 @@ class BaseDocument(object):
|
|||||||
|
|
||||||
if hasattr(value, "_get_changed_fields"):
|
if hasattr(value, "_get_changed_fields"):
|
||||||
changed = value._get_changed_fields()
|
changed = value._get_changed_fields()
|
||||||
changed_fields += ["%s%s" % (item_key, k) for k in changed if k]
|
changed_fields += ["{}{}".format(item_key, k) for k in changed if k]
|
||||||
elif isinstance(value, (list, tuple, dict)):
|
elif isinstance(value, (list, tuple, dict)):
|
||||||
self._nestable_types_changed_fields(changed_fields, item_key, value)
|
self._nestable_types_changed_fields(changed_fields, item_key, value)
|
||||||
|
|
||||||
@ -620,7 +615,7 @@ class BaseDocument(object):
|
|||||||
if isinstance(data, EmbeddedDocument):
|
if isinstance(data, EmbeddedDocument):
|
||||||
# Find all embedded fields that have been changed
|
# Find all embedded fields that have been changed
|
||||||
changed = data._get_changed_fields()
|
changed = data._get_changed_fields()
|
||||||
changed_fields += ["%s%s" % (key, k) for k in changed if k]
|
changed_fields += ["{}{}".format(key, k) for k in changed if k]
|
||||||
elif isinstance(data, (list, tuple, dict)):
|
elif isinstance(data, (list, tuple, dict)):
|
||||||
if hasattr(field, "field") and isinstance(
|
if hasattr(field, "field") and isinstance(
|
||||||
field.field, (ReferenceField, GenericReferenceField)
|
field.field, (ReferenceField, GenericReferenceField)
|
||||||
@ -670,7 +665,7 @@ class BaseDocument(object):
|
|||||||
del set_data["_id"]
|
del set_data["_id"]
|
||||||
|
|
||||||
# Determine if any changed items were actually unset.
|
# Determine if any changed items were actually unset.
|
||||||
for path, value in set_data.items():
|
for path, value in list(set_data.items()):
|
||||||
if value or isinstance(
|
if value or isinstance(
|
||||||
value, (numbers.Number, bool)
|
value, (numbers.Number, bool)
|
||||||
): # Account for 0 and True that are truthy
|
): # Account for 0 and True that are truthy
|
||||||
@ -744,7 +739,7 @@ class BaseDocument(object):
|
|||||||
# Convert SON to a data dict, making sure each key is a string and
|
# Convert SON to a data dict, making sure each key is a string and
|
||||||
# corresponds to the right db field.
|
# corresponds to the right db field.
|
||||||
data = {}
|
data = {}
|
||||||
for key, value in iteritems(son):
|
for key, value in son.items():
|
||||||
key = str(key)
|
key = str(key)
|
||||||
key = cls._db_field_map.get(key, key)
|
key = cls._db_field_map.get(key, key)
|
||||||
data[key] = value
|
data[key] = value
|
||||||
@ -759,7 +754,7 @@ class BaseDocument(object):
|
|||||||
if not _auto_dereference:
|
if not _auto_dereference:
|
||||||
fields = copy.deepcopy(fields)
|
fields = copy.deepcopy(fields)
|
||||||
|
|
||||||
for field_name, field in iteritems(fields):
|
for field_name, field in fields.items():
|
||||||
field._auto_dereference = _auto_dereference
|
field._auto_dereference = _auto_dereference
|
||||||
if field.db_field in data:
|
if field.db_field in data:
|
||||||
value = data[field.db_field]
|
value = data[field.db_field]
|
||||||
@ -774,17 +769,16 @@ class BaseDocument(object):
|
|||||||
|
|
||||||
if errors_dict:
|
if errors_dict:
|
||||||
errors = "\n".join(
|
errors = "\n".join(
|
||||||
["Field '%s' - %s" % (k, v) for k, v in errors_dict.items()]
|
["Field '{}' - {}".format(k, v) for k, v in errors_dict.items()]
|
||||||
)
|
)
|
||||||
msg = "Invalid data to create a `%s` instance.\n%s" % (
|
msg = "Invalid data to create a `{}` instance.\n{}".format(
|
||||||
cls._class_name,
|
cls._class_name, errors,
|
||||||
errors,
|
|
||||||
)
|
)
|
||||||
raise InvalidDocumentError(msg)
|
raise InvalidDocumentError(msg)
|
||||||
|
|
||||||
# In STRICT documents, remove any keys that aren't in cls._fields
|
# In STRICT documents, remove any keys that aren't in cls._fields
|
||||||
if cls.STRICT:
|
if cls.STRICT:
|
||||||
data = {k: v for k, v in iteritems(data) if k in cls._fields}
|
data = {k: v for k, v in data.items() if k in cls._fields}
|
||||||
|
|
||||||
obj = cls(
|
obj = cls(
|
||||||
__auto_convert=False, _created=created, __only_fields=only_fields, **data
|
__auto_convert=False, _created=created, __only_fields=only_fields, **data
|
||||||
@ -831,7 +825,7 @@ class BaseDocument(object):
|
|||||||
@classmethod
|
@classmethod
|
||||||
def _build_index_spec(cls, spec):
|
def _build_index_spec(cls, spec):
|
||||||
"""Build a PyMongo index spec from a MongoEngine index spec."""
|
"""Build a PyMongo index spec from a MongoEngine index spec."""
|
||||||
if isinstance(spec, six.string_types):
|
if isinstance(spec, str):
|
||||||
spec = {"fields": [spec]}
|
spec = {"fields": [spec]}
|
||||||
elif isinstance(spec, (list, tuple)):
|
elif isinstance(spec, (list, tuple)):
|
||||||
spec = {"fields": list(spec)}
|
spec = {"fields": list(spec)}
|
||||||
@ -928,7 +922,7 @@ class BaseDocument(object):
|
|||||||
|
|
||||||
# Add any unique_with fields to the back of the index spec
|
# Add any unique_with fields to the back of the index spec
|
||||||
if field.unique_with:
|
if field.unique_with:
|
||||||
if isinstance(field.unique_with, six.string_types):
|
if isinstance(field.unique_with, str):
|
||||||
field.unique_with = [field.unique_with]
|
field.unique_with = [field.unique_with]
|
||||||
|
|
||||||
# Convert unique_with field names to real field names
|
# Convert unique_with field names to real field names
|
||||||
@ -949,7 +943,8 @@ class BaseDocument(object):
|
|||||||
|
|
||||||
# Add the new index to the list
|
# Add the new index to the list
|
||||||
fields = [
|
fields = [
|
||||||
("%s%s" % (namespace, f), pymongo.ASCENDING) for f in unique_fields
|
("{}{}".format(namespace, f), pymongo.ASCENDING)
|
||||||
|
for f in unique_fields
|
||||||
]
|
]
|
||||||
index = {"fields": fields, "unique": True, "sparse": sparse}
|
index = {"fields": fields, "unique": True, "sparse": sparse}
|
||||||
unique_indexes.append(index)
|
unique_indexes.append(index)
|
||||||
@ -1006,7 +1001,7 @@ class BaseDocument(object):
|
|||||||
elif field._geo_index:
|
elif field._geo_index:
|
||||||
field_name = field.db_field
|
field_name = field.db_field
|
||||||
if parent_field:
|
if parent_field:
|
||||||
field_name = "%s.%s" % (parent_field, field_name)
|
field_name = "{}.{}".format(parent_field, field_name)
|
||||||
geo_indices.append({"fields": [(field_name, field._geo_index)]})
|
geo_indices.append({"fields": [(field_name, field._geo_index)]})
|
||||||
|
|
||||||
return geo_indices
|
return geo_indices
|
||||||
@ -1175,9 +1170,6 @@ class BaseDocument(object):
|
|||||||
else [value]
|
else [value]
|
||||||
)
|
)
|
||||||
return sep.join(
|
return sep.join(
|
||||||
[
|
[str(dict(field.choices).get(val, val)) for val in values or []]
|
||||||
six.text_type(dict(field.choices).get(val, val))
|
|
||||||
for val in values or []
|
|
||||||
]
|
|
||||||
)
|
)
|
||||||
return value
|
return value
|
||||||
|
@ -4,8 +4,6 @@ import weakref
|
|||||||
|
|
||||||
from bson import DBRef, ObjectId, SON
|
from bson import DBRef, ObjectId, SON
|
||||||
import pymongo
|
import pymongo
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine.base.common import UPDATE_OPERATORS
|
from mongoengine.base.common import UPDATE_OPERATORS
|
||||||
from mongoengine.base.datastructures import BaseDict, BaseList, EmbeddedDocumentList
|
from mongoengine.base.datastructures import BaseDict, BaseList, EmbeddedDocumentList
|
||||||
@ -15,7 +13,7 @@ from mongoengine.errors import DeprecatedError, ValidationError
|
|||||||
__all__ = ("BaseField", "ComplexBaseField", "ObjectIdField", "GeoJsonBaseField")
|
__all__ = ("BaseField", "ComplexBaseField", "ObjectIdField", "GeoJsonBaseField")
|
||||||
|
|
||||||
|
|
||||||
class BaseField(object):
|
class BaseField:
|
||||||
"""A base class for fields in a MongoDB document. Instances of this class
|
"""A base class for fields in a MongoDB document. Instances of this class
|
||||||
may be added to subclasses of `Document` to define a document's schema.
|
may be added to subclasses of `Document` to define a document's schema.
|
||||||
|
|
||||||
@ -36,7 +34,6 @@ class BaseField(object):
|
|||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
db_field=None,
|
db_field=None,
|
||||||
name=None,
|
|
||||||
required=False,
|
required=False,
|
||||||
default=None,
|
default=None,
|
||||||
unique=False,
|
unique=False,
|
||||||
@ -51,7 +48,6 @@ class BaseField(object):
|
|||||||
"""
|
"""
|
||||||
:param db_field: The database field to store this field in
|
:param db_field: The database field to store this field in
|
||||||
(defaults to the name of the field)
|
(defaults to the name of the field)
|
||||||
:param name: Deprecated - use db_field
|
|
||||||
:param required: If the field is required. Whether it has to have a
|
:param required: If the field is required. Whether it has to have a
|
||||||
value or not. Defaults to False.
|
value or not. Defaults to False.
|
||||||
:param default: (optional) The default value for this field if no value
|
:param default: (optional) The default value for this field if no value
|
||||||
@ -75,11 +71,8 @@ class BaseField(object):
|
|||||||
existing attributes. Common metadata includes `verbose_name` and
|
existing attributes. Common metadata includes `verbose_name` and
|
||||||
`help_text`.
|
`help_text`.
|
||||||
"""
|
"""
|
||||||
self.db_field = (db_field or name) if not primary_key else "_id"
|
self.db_field = db_field if not primary_key else "_id"
|
||||||
|
|
||||||
if name:
|
|
||||||
msg = 'Field\'s "name" attribute deprecated in favour of "db_field"'
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
self.required = required or primary_key
|
self.required = required or primary_key
|
||||||
self.default = default
|
self.default = default
|
||||||
self.unique = bool(unique or unique_with)
|
self.unique = bool(unique or unique_with)
|
||||||
@ -92,13 +85,11 @@ class BaseField(object):
|
|||||||
self._owner_document = None
|
self._owner_document = None
|
||||||
|
|
||||||
# Make sure db_field is a string (if it's explicitly defined).
|
# Make sure db_field is a string (if it's explicitly defined).
|
||||||
if self.db_field is not None and not isinstance(
|
if self.db_field is not None and not isinstance(self.db_field, str):
|
||||||
self.db_field, six.string_types
|
|
||||||
):
|
|
||||||
raise TypeError("db_field should be a string.")
|
raise TypeError("db_field should be a string.")
|
||||||
|
|
||||||
# Make sure db_field doesn't contain any forbidden characters.
|
# Make sure db_field doesn't contain any forbidden characters.
|
||||||
if isinstance(self.db_field, six.string_types) and (
|
if isinstance(self.db_field, str) and (
|
||||||
"." in self.db_field
|
"." in self.db_field
|
||||||
or "\0" in self.db_field
|
or "\0" in self.db_field
|
||||||
or self.db_field.startswith("$")
|
or self.db_field.startswith("$")
|
||||||
@ -221,14 +212,12 @@ class BaseField(object):
|
|||||||
# Choices which are other types of Documents
|
# Choices which are other types of Documents
|
||||||
if isinstance(value, (Document, EmbeddedDocument)):
|
if isinstance(value, (Document, EmbeddedDocument)):
|
||||||
if not any(isinstance(value, c) for c in choice_list):
|
if not any(isinstance(value, c) for c in choice_list):
|
||||||
self.error(
|
self.error("Value must be an instance of %s" % (choice_list))
|
||||||
"Value must be an instance of %s" % (six.text_type(choice_list))
|
|
||||||
)
|
|
||||||
# Choices which are types other than Documents
|
# Choices which are types other than Documents
|
||||||
else:
|
else:
|
||||||
values = value if isinstance(value, (list, tuple)) else [value]
|
values = value if isinstance(value, (list, tuple)) else [value]
|
||||||
if len(set(values) - set(choice_list)):
|
if len(set(values) - set(choice_list)):
|
||||||
self.error("Value must be one of %s" % six.text_type(choice_list))
|
self.error("Value must be one of %s" % str(choice_list))
|
||||||
|
|
||||||
def _validate(self, value, **kwargs):
|
def _validate(self, value, **kwargs):
|
||||||
# Check the Choices Constraint
|
# Check the Choices Constraint
|
||||||
@ -316,7 +305,7 @@ class ComplexBaseField(BaseField):
|
|||||||
if hasattr(instance._data[self.name], "_dereferenced"):
|
if hasattr(instance._data[self.name], "_dereferenced"):
|
||||||
instance._data[self.name]._dereferenced = True
|
instance._data[self.name]._dereferenced = True
|
||||||
|
|
||||||
value = super(ComplexBaseField, self).__get__(instance, owner)
|
value = super().__get__(instance, owner)
|
||||||
|
|
||||||
# Convert lists / values so we can watch for any changes on them
|
# Convert lists / values so we can watch for any changes on them
|
||||||
if isinstance(value, (list, tuple)):
|
if isinstance(value, (list, tuple)):
|
||||||
@ -345,7 +334,7 @@ class ComplexBaseField(BaseField):
|
|||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
"""Convert a MongoDB-compatible type to a Python type."""
|
"""Convert a MongoDB-compatible type to a Python type."""
|
||||||
if isinstance(value, six.string_types):
|
if isinstance(value, str):
|
||||||
return value
|
return value
|
||||||
|
|
||||||
if hasattr(value, "to_python"):
|
if hasattr(value, "to_python"):
|
||||||
@ -399,7 +388,7 @@ class ComplexBaseField(BaseField):
|
|||||||
EmbeddedDocument = _import_class("EmbeddedDocument")
|
EmbeddedDocument = _import_class("EmbeddedDocument")
|
||||||
GenericReferenceField = _import_class("GenericReferenceField")
|
GenericReferenceField = _import_class("GenericReferenceField")
|
||||||
|
|
||||||
if isinstance(value, six.string_types):
|
if isinstance(value, str):
|
||||||
return value
|
return value
|
||||||
|
|
||||||
if hasattr(value, "to_mongo"):
|
if hasattr(value, "to_mongo"):
|
||||||
@ -423,11 +412,11 @@ class ComplexBaseField(BaseField):
|
|||||||
if self.field:
|
if self.field:
|
||||||
value_dict = {
|
value_dict = {
|
||||||
key: self.field._to_mongo_safe_call(item, use_db_field, fields)
|
key: self.field._to_mongo_safe_call(item, use_db_field, fields)
|
||||||
for key, item in iteritems(value)
|
for key, item in value.items()
|
||||||
}
|
}
|
||||||
else:
|
else:
|
||||||
value_dict = {}
|
value_dict = {}
|
||||||
for k, v in iteritems(value):
|
for k, v in value.items():
|
||||||
if isinstance(v, Document):
|
if isinstance(v, Document):
|
||||||
# We need the id from the saved object to create the DBRef
|
# We need the id from the saved object to create the DBRef
|
||||||
if v.pk is None:
|
if v.pk is None:
|
||||||
@ -466,8 +455,8 @@ class ComplexBaseField(BaseField):
|
|||||||
"""If field is provided ensure the value is valid."""
|
"""If field is provided ensure the value is valid."""
|
||||||
errors = {}
|
errors = {}
|
||||||
if self.field:
|
if self.field:
|
||||||
if hasattr(value, "iteritems") or hasattr(value, "items"):
|
if hasattr(value, "items"):
|
||||||
sequence = iteritems(value)
|
sequence = value.items()
|
||||||
else:
|
else:
|
||||||
sequence = enumerate(value)
|
sequence = enumerate(value)
|
||||||
for k, v in sequence:
|
for k, v in sequence:
|
||||||
@ -480,7 +469,9 @@ class ComplexBaseField(BaseField):
|
|||||||
|
|
||||||
if errors:
|
if errors:
|
||||||
field_class = self.field.__class__.__name__
|
field_class = self.field.__class__.__name__
|
||||||
self.error("Invalid %s item (%s)" % (field_class, value), errors=errors)
|
self.error(
|
||||||
|
"Invalid {} item ({})".format(field_class, value), errors=errors
|
||||||
|
)
|
||||||
# Don't allow empty values if required
|
# Don't allow empty values if required
|
||||||
if self.required and not value:
|
if self.required and not value:
|
||||||
self.error("Field is required and cannot be empty")
|
self.error("Field is required and cannot be empty")
|
||||||
@ -513,10 +504,9 @@ class ObjectIdField(BaseField):
|
|||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
if not isinstance(value, ObjectId):
|
if not isinstance(value, ObjectId):
|
||||||
try:
|
try:
|
||||||
return ObjectId(six.text_type(value))
|
return ObjectId(str(value))
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
# e.message attribute has been deprecated since Python 2.6
|
self.error(str(e))
|
||||||
self.error(six.text_type(e))
|
|
||||||
return value
|
return value
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
@ -524,9 +514,9 @@ class ObjectIdField(BaseField):
|
|||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
try:
|
try:
|
||||||
ObjectId(six.text_type(value))
|
ObjectId(str(value))
|
||||||
except Exception:
|
except Exception:
|
||||||
self.error("Invalid Object ID")
|
self.error("Invalid ObjectID")
|
||||||
|
|
||||||
|
|
||||||
class GeoJsonBaseField(BaseField):
|
class GeoJsonBaseField(BaseField):
|
||||||
@ -546,14 +536,14 @@ class GeoJsonBaseField(BaseField):
|
|||||||
self._name = "%sField" % self._type
|
self._name = "%sField" % self._type
|
||||||
if not auto_index:
|
if not auto_index:
|
||||||
self._geo_index = False
|
self._geo_index = False
|
||||||
super(GeoJsonBaseField, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
"""Validate the GeoJson object based on its type."""
|
"""Validate the GeoJson object based on its type."""
|
||||||
if isinstance(value, dict):
|
if isinstance(value, dict):
|
||||||
if set(value.keys()) == {"type", "coordinates"}:
|
if set(value.keys()) == {"type", "coordinates"}:
|
||||||
if value["type"] != self._type:
|
if value["type"] != self._type:
|
||||||
self.error('%s type must be "%s"' % (self._name, self._type))
|
self.error('{} type must be "{}"'.format(self._name, self._type))
|
||||||
return self.validate(value["coordinates"])
|
return self.validate(value["coordinates"])
|
||||||
else:
|
else:
|
||||||
self.error(
|
self.error(
|
||||||
|
@ -1,9 +1,6 @@
|
|||||||
import itertools
|
import itertools
|
||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
import six
|
|
||||||
from six import iteritems, itervalues
|
|
||||||
|
|
||||||
from mongoengine.base.common import _document_registry
|
from mongoengine.base.common import _document_registry
|
||||||
from mongoengine.base.fields import BaseField, ComplexBaseField, ObjectIdField
|
from mongoengine.base.fields import BaseField, ComplexBaseField, ObjectIdField
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
@ -25,7 +22,7 @@ class DocumentMetaclass(type):
|
|||||||
# TODO lower complexity of this method
|
# TODO lower complexity of this method
|
||||||
def __new__(mcs, name, bases, attrs):
|
def __new__(mcs, name, bases, attrs):
|
||||||
flattened_bases = mcs._get_bases(bases)
|
flattened_bases = mcs._get_bases(bases)
|
||||||
super_new = super(DocumentMetaclass, mcs).__new__
|
super_new = super().__new__
|
||||||
|
|
||||||
# If a base class just call super
|
# If a base class just call super
|
||||||
metaclass = attrs.get("my_metaclass")
|
metaclass = attrs.get("my_metaclass")
|
||||||
@ -69,7 +66,7 @@ class DocumentMetaclass(type):
|
|||||||
# Standard object mixin - merge in any Fields
|
# Standard object mixin - merge in any Fields
|
||||||
if not hasattr(base, "_meta"):
|
if not hasattr(base, "_meta"):
|
||||||
base_fields = {}
|
base_fields = {}
|
||||||
for attr_name, attr_value in iteritems(base.__dict__):
|
for attr_name, attr_value in base.__dict__.items():
|
||||||
if not isinstance(attr_value, BaseField):
|
if not isinstance(attr_value, BaseField):
|
||||||
continue
|
continue
|
||||||
attr_value.name = attr_name
|
attr_value.name = attr_name
|
||||||
@ -81,7 +78,7 @@ class DocumentMetaclass(type):
|
|||||||
|
|
||||||
# Discover any document fields
|
# Discover any document fields
|
||||||
field_names = {}
|
field_names = {}
|
||||||
for attr_name, attr_value in iteritems(attrs):
|
for attr_name, attr_value in attrs.items():
|
||||||
if not isinstance(attr_value, BaseField):
|
if not isinstance(attr_value, BaseField):
|
||||||
continue
|
continue
|
||||||
attr_value.name = attr_name
|
attr_value.name = attr_name
|
||||||
@ -111,9 +108,7 @@ class DocumentMetaclass(type):
|
|||||||
|
|
||||||
attrs["_fields_ordered"] = tuple(
|
attrs["_fields_ordered"] = tuple(
|
||||||
i[1]
|
i[1]
|
||||||
for i in sorted(
|
for i in sorted((v.creation_counter, v.name) for v in doc_fields.values())
|
||||||
(v.creation_counter, v.name) for v in itervalues(doc_fields)
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
#
|
#
|
||||||
@ -173,24 +168,8 @@ class DocumentMetaclass(type):
|
|||||||
# Add class to the _document_registry
|
# Add class to the _document_registry
|
||||||
_document_registry[new_class._class_name] = new_class
|
_document_registry[new_class._class_name] = new_class
|
||||||
|
|
||||||
# In Python 2, User-defined methods objects have special read-only
|
|
||||||
# attributes 'im_func' and 'im_self' which contain the function obj
|
|
||||||
# and class instance object respectively. With Python 3 these special
|
|
||||||
# attributes have been replaced by __func__ and __self__. The Blinker
|
|
||||||
# module continues to use im_func and im_self, so the code below
|
|
||||||
# copies __func__ into im_func and __self__ into im_self for
|
|
||||||
# classmethod objects in Document derived classes.
|
|
||||||
if six.PY3:
|
|
||||||
for val in new_class.__dict__.values():
|
|
||||||
if isinstance(val, classmethod):
|
|
||||||
f = val.__get__(new_class)
|
|
||||||
if hasattr(f, "__func__") and not hasattr(f, "im_func"):
|
|
||||||
f.__dict__.update({"im_func": getattr(f, "__func__")})
|
|
||||||
if hasattr(f, "__self__") and not hasattr(f, "im_self"):
|
|
||||||
f.__dict__.update({"im_self": getattr(f, "__self__")})
|
|
||||||
|
|
||||||
# Handle delete rules
|
# Handle delete rules
|
||||||
for field in itervalues(new_class._fields):
|
for field in new_class._fields.values():
|
||||||
f = field
|
f = field
|
||||||
if f.owner_document is None:
|
if f.owner_document is None:
|
||||||
f.owner_document = new_class
|
f.owner_document = new_class
|
||||||
@ -252,8 +231,7 @@ class DocumentMetaclass(type):
|
|||||||
if base is object:
|
if base is object:
|
||||||
continue
|
continue
|
||||||
yield base
|
yield base
|
||||||
for child_base in mcs.__get_bases(base.__bases__):
|
yield from mcs.__get_bases(base.__bases__)
|
||||||
yield child_base
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _import_classes(mcs):
|
def _import_classes(mcs):
|
||||||
@ -271,7 +249,7 @@ class TopLevelDocumentMetaclass(DocumentMetaclass):
|
|||||||
|
|
||||||
def __new__(mcs, name, bases, attrs):
|
def __new__(mcs, name, bases, attrs):
|
||||||
flattened_bases = mcs._get_bases(bases)
|
flattened_bases = mcs._get_bases(bases)
|
||||||
super_new = super(TopLevelDocumentMetaclass, mcs).__new__
|
super_new = super().__new__
|
||||||
|
|
||||||
# Set default _meta data if base class, otherwise get user defined meta
|
# Set default _meta data if base class, otherwise get user defined meta
|
||||||
if attrs.get("my_metaclass") == TopLevelDocumentMetaclass:
|
if attrs.get("my_metaclass") == TopLevelDocumentMetaclass:
|
||||||
@ -284,7 +262,6 @@ class TopLevelDocumentMetaclass(DocumentMetaclass):
|
|||||||
"indexes": [], # indexes to be ensured at runtime
|
"indexes": [], # indexes to be ensured at runtime
|
||||||
"id_field": None,
|
"id_field": None,
|
||||||
"index_background": False,
|
"index_background": False,
|
||||||
"index_drop_dups": False,
|
|
||||||
"index_opts": None,
|
"index_opts": None,
|
||||||
"delete_rules": None,
|
"delete_rules": None,
|
||||||
# allow_inheritance can be True, False, and None. True means
|
# allow_inheritance can be True, False, and None. True means
|
||||||
@ -399,7 +376,7 @@ class TopLevelDocumentMetaclass(DocumentMetaclass):
|
|||||||
new_class.objects = QuerySetManager()
|
new_class.objects = QuerySetManager()
|
||||||
|
|
||||||
# Validate the fields and set primary key if needed
|
# Validate the fields and set primary key if needed
|
||||||
for field_name, field in iteritems(new_class._fields):
|
for field_name, field in new_class._fields.items():
|
||||||
if field.primary_key:
|
if field.primary_key:
|
||||||
# Ensure only one primary key is set
|
# Ensure only one primary key is set
|
||||||
current_pk = new_class._meta.get("id_field")
|
current_pk = new_class._meta.get("id_field")
|
||||||
@ -462,8 +439,8 @@ class TopLevelDocumentMetaclass(DocumentMetaclass):
|
|||||||
|
|
||||||
id_basename, id_db_basename, i = ("auto_id", "_auto_id", 0)
|
id_basename, id_db_basename, i = ("auto_id", "_auto_id", 0)
|
||||||
for i in itertools.count():
|
for i in itertools.count():
|
||||||
id_name = "{0}_{1}".format(id_basename, i)
|
id_name = "{}_{}".format(id_basename, i)
|
||||||
id_db_name = "{0}_{1}".format(id_db_basename, i)
|
id_db_name = "{}_{}".format(id_db_basename, i)
|
||||||
if id_name not in existing_fields and id_db_name not in existing_db_fields:
|
if id_name not in existing_fields and id_db_name not in existing_db_fields:
|
||||||
return id_name, id_db_name
|
return id_name, id_db_name
|
||||||
|
|
||||||
@ -476,7 +453,7 @@ class MetaDict(dict):
|
|||||||
_merge_options = ("indexes",)
|
_merge_options = ("indexes",)
|
||||||
|
|
||||||
def merge(self, new_options):
|
def merge(self, new_options):
|
||||||
for k, v in iteritems(new_options):
|
for k, v in new_options.items():
|
||||||
if k in self._merge_options:
|
if k in self._merge_options:
|
||||||
self[k] = self.get(k, []) + v
|
self[k] = self.get(k, []) + v
|
||||||
else:
|
else:
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
import re
|
import re
|
||||||
|
|
||||||
|
|
||||||
class LazyRegexCompiler(object):
|
class LazyRegexCompiler:
|
||||||
"""Descriptor to allow lazy compilation of regex"""
|
"""Descriptor to allow lazy compilation of regex"""
|
||||||
|
|
||||||
def __init__(self, pattern, flags=0):
|
def __init__(self, pattern, flags=0):
|
||||||
|
@ -1,6 +1,5 @@
|
|||||||
from pymongo import MongoClient, ReadPreference, uri_parser
|
from pymongo import MongoClient, ReadPreference, uri_parser
|
||||||
from pymongo.database import _check_name
|
from pymongo.database import _check_name
|
||||||
import six
|
|
||||||
|
|
||||||
__all__ = [
|
__all__ = [
|
||||||
"DEFAULT_CONNECTION_NAME",
|
"DEFAULT_CONNECTION_NAME",
|
||||||
@ -39,8 +38,8 @@ def _check_db_name(name):
|
|||||||
"""Check if a database name is valid.
|
"""Check if a database name is valid.
|
||||||
This functionality is copied from pymongo Database class constructor.
|
This functionality is copied from pymongo Database class constructor.
|
||||||
"""
|
"""
|
||||||
if not isinstance(name, six.string_types):
|
if not isinstance(name, str):
|
||||||
raise TypeError("name must be an instance of %s" % six.string_types)
|
raise TypeError("name must be an instance of %s" % str)
|
||||||
elif name != "$external":
|
elif name != "$external":
|
||||||
_check_name(name)
|
_check_name(name)
|
||||||
|
|
||||||
@ -93,7 +92,7 @@ def _get_connection_settings(
|
|||||||
conn_host = conn_settings["host"]
|
conn_host = conn_settings["host"]
|
||||||
|
|
||||||
# Host can be a list or a string, so if string, force to a list.
|
# Host can be a list or a string, so if string, force to a list.
|
||||||
if isinstance(conn_host, six.string_types):
|
if isinstance(conn_host, str):
|
||||||
conn_host = [conn_host]
|
conn_host = [conn_host]
|
||||||
|
|
||||||
resolved_hosts = []
|
resolved_hosts = []
|
||||||
@ -148,7 +147,7 @@ def _get_connection_settings(
|
|||||||
# TODO simplify the code below once we drop support for
|
# TODO simplify the code below once we drop support for
|
||||||
# PyMongo v3.4.
|
# PyMongo v3.4.
|
||||||
read_pf_mode = uri_options["readpreference"]
|
read_pf_mode = uri_options["readpreference"]
|
||||||
if isinstance(read_pf_mode, six.string_types):
|
if isinstance(read_pf_mode, str):
|
||||||
read_pf_mode = read_pf_mode.lower()
|
read_pf_mode = read_pf_mode.lower()
|
||||||
for preference in read_preferences:
|
for preference in read_preferences:
|
||||||
if (
|
if (
|
||||||
@ -318,7 +317,7 @@ def _create_connection(alias, connection_class, **connection_settings):
|
|||||||
try:
|
try:
|
||||||
return connection_class(**connection_settings)
|
return connection_class(**connection_settings)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise ConnectionFailure("Cannot connect to database %s :\n%s" % (alias, e))
|
raise ConnectionFailure("Cannot connect to database {} :\n{}".format(alias, e))
|
||||||
|
|
||||||
|
|
||||||
def _find_existing_connection(connection_settings):
|
def _find_existing_connection(connection_settings):
|
||||||
@ -396,8 +395,8 @@ def connect(db=None, alias=DEFAULT_CONNECTION_NAME, **kwargs):
|
|||||||
|
|
||||||
if new_conn_settings != prev_conn_setting:
|
if new_conn_settings != prev_conn_setting:
|
||||||
err_msg = (
|
err_msg = (
|
||||||
u"A different connection with alias `{}` was already "
|
"A different connection with alias `{}` was already "
|
||||||
u"registered. Use disconnect() first"
|
"registered. Use disconnect() first"
|
||||||
).format(alias)
|
).format(alias)
|
||||||
raise ConnectionFailure(err_msg)
|
raise ConnectionFailure(err_msg)
|
||||||
else:
|
else:
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
from contextlib import contextmanager
|
from contextlib import contextmanager
|
||||||
|
|
||||||
|
from pymongo.read_concern import ReadConcern
|
||||||
from pymongo.write_concern import WriteConcern
|
from pymongo.write_concern import WriteConcern
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
||||||
@ -14,10 +14,11 @@ __all__ = (
|
|||||||
"no_sub_classes",
|
"no_sub_classes",
|
||||||
"query_counter",
|
"query_counter",
|
||||||
"set_write_concern",
|
"set_write_concern",
|
||||||
|
"set_read_write_concern",
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class switch_db(object):
|
class switch_db:
|
||||||
"""switch_db alias context manager.
|
"""switch_db alias context manager.
|
||||||
|
|
||||||
Example ::
|
Example ::
|
||||||
@ -58,7 +59,7 @@ class switch_db(object):
|
|||||||
self.cls._collection = self.collection
|
self.cls._collection = self.collection
|
||||||
|
|
||||||
|
|
||||||
class switch_collection(object):
|
class switch_collection:
|
||||||
"""switch_collection alias context manager.
|
"""switch_collection alias context manager.
|
||||||
|
|
||||||
Example ::
|
Example ::
|
||||||
@ -100,7 +101,7 @@ class switch_collection(object):
|
|||||||
self.cls._get_collection_name = self.ori_get_collection_name
|
self.cls._get_collection_name = self.ori_get_collection_name
|
||||||
|
|
||||||
|
|
||||||
class no_dereference(object):
|
class no_dereference:
|
||||||
"""no_dereference context manager.
|
"""no_dereference context manager.
|
||||||
|
|
||||||
Turns off all dereferencing in Documents for the duration of the context
|
Turns off all dereferencing in Documents for the duration of the context
|
||||||
@ -123,7 +124,7 @@ class no_dereference(object):
|
|||||||
|
|
||||||
self.deref_fields = [
|
self.deref_fields = [
|
||||||
k
|
k
|
||||||
for k, v in iteritems(self.cls._fields)
|
for k, v in self.cls._fields.items()
|
||||||
if isinstance(v, (ReferenceField, GenericReferenceField, ComplexBaseField))
|
if isinstance(v, (ReferenceField, GenericReferenceField, ComplexBaseField))
|
||||||
]
|
]
|
||||||
|
|
||||||
@ -140,7 +141,7 @@ class no_dereference(object):
|
|||||||
return self.cls
|
return self.cls
|
||||||
|
|
||||||
|
|
||||||
class no_sub_classes(object):
|
class no_sub_classes:
|
||||||
"""no_sub_classes context manager.
|
"""no_sub_classes context manager.
|
||||||
|
|
||||||
Only returns instances of this class and no sub (inherited) classes::
|
Only returns instances of this class and no sub (inherited) classes::
|
||||||
@ -168,7 +169,7 @@ class no_sub_classes(object):
|
|||||||
self.cls._subclasses = self.cls_initial_subclasses
|
self.cls._subclasses = self.cls_initial_subclasses
|
||||||
|
|
||||||
|
|
||||||
class query_counter(object):
|
class query_counter:
|
||||||
"""Query_counter context manager to get the number of queries.
|
"""Query_counter context manager to get the number of queries.
|
||||||
This works by updating the `profiling_level` of the database so that all queries get logged,
|
This works by updating the `profiling_level` of the database so that all queries get logged,
|
||||||
resetting the db.system.profile collection at the beginning of the context and counting the new entries.
|
resetting the db.system.profile collection at the beginning of the context and counting the new entries.
|
||||||
@ -235,7 +236,7 @@ class query_counter(object):
|
|||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
"""repr query_counter as the number of queries."""
|
"""repr query_counter as the number of queries."""
|
||||||
return u"%s" % self._get_count()
|
return "%s" % self._get_count()
|
||||||
|
|
||||||
def _get_count(self):
|
def _get_count(self):
|
||||||
"""Get the number of queries by counting the current number of entries in db.system.profile
|
"""Get the number of queries by counting the current number of entries in db.system.profile
|
||||||
@ -257,3 +258,21 @@ def set_write_concern(collection, write_concerns):
|
|||||||
combined_concerns = dict(collection.write_concern.document.items())
|
combined_concerns = dict(collection.write_concern.document.items())
|
||||||
combined_concerns.update(write_concerns)
|
combined_concerns.update(write_concerns)
|
||||||
yield collection.with_options(write_concern=WriteConcern(**combined_concerns))
|
yield collection.with_options(write_concern=WriteConcern(**combined_concerns))
|
||||||
|
|
||||||
|
|
||||||
|
@contextmanager
|
||||||
|
def set_read_write_concern(collection, write_concerns, read_concerns):
|
||||||
|
combined_write_concerns = dict(collection.write_concern.document.items())
|
||||||
|
|
||||||
|
if write_concerns is not None:
|
||||||
|
combined_write_concerns.update(write_concerns)
|
||||||
|
|
||||||
|
combined_read_concerns = dict(collection.read_concern.document.items())
|
||||||
|
|
||||||
|
if read_concerns is not None:
|
||||||
|
combined_read_concerns.update(read_concerns)
|
||||||
|
|
||||||
|
yield collection.with_options(
|
||||||
|
write_concern=WriteConcern(**combined_write_concerns),
|
||||||
|
read_concern=ReadConcern(**combined_read_concerns),
|
||||||
|
)
|
||||||
|
@ -1,6 +1,4 @@
|
|||||||
from bson import DBRef, SON
|
from bson import DBRef, SON
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine.base import (
|
from mongoengine.base import (
|
||||||
BaseDict,
|
BaseDict,
|
||||||
@ -16,7 +14,7 @@ from mongoengine.fields import DictField, ListField, MapField, ReferenceField
|
|||||||
from mongoengine.queryset import QuerySet
|
from mongoengine.queryset import QuerySet
|
||||||
|
|
||||||
|
|
||||||
class DeReference(object):
|
class DeReference:
|
||||||
def __call__(self, items, max_depth=1, instance=None, name=None):
|
def __call__(self, items, max_depth=1, instance=None, name=None):
|
||||||
"""
|
"""
|
||||||
Cheaply dereferences the items to a set depth.
|
Cheaply dereferences the items to a set depth.
|
||||||
@ -30,7 +28,7 @@ class DeReference(object):
|
|||||||
:class:`~mongoengine.base.ComplexBaseField`
|
:class:`~mongoengine.base.ComplexBaseField`
|
||||||
:param get: A boolean determining if being called by __get__
|
:param get: A boolean determining if being called by __get__
|
||||||
"""
|
"""
|
||||||
if items is None or isinstance(items, six.string_types):
|
if items is None or isinstance(items, str):
|
||||||
return items
|
return items
|
||||||
|
|
||||||
# cheapest way to convert a queryset to a list
|
# cheapest way to convert a queryset to a list
|
||||||
@ -79,7 +77,7 @@ class DeReference(object):
|
|||||||
|
|
||||||
def _get_items_from_dict(items):
|
def _get_items_from_dict(items):
|
||||||
new_items = {}
|
new_items = {}
|
||||||
for k, v in iteritems(items):
|
for k, v in items.items():
|
||||||
value = v
|
value = v
|
||||||
if isinstance(v, list):
|
if isinstance(v, list):
|
||||||
value = _get_items_from_list(v)
|
value = _get_items_from_list(v)
|
||||||
@ -120,7 +118,7 @@ class DeReference(object):
|
|||||||
depth += 1
|
depth += 1
|
||||||
for item in iterator:
|
for item in iterator:
|
||||||
if isinstance(item, (Document, EmbeddedDocument)):
|
if isinstance(item, (Document, EmbeddedDocument)):
|
||||||
for field_name, field in iteritems(item._fields):
|
for field_name, field in item._fields.items():
|
||||||
v = item._data.get(field_name, None)
|
v = item._data.get(field_name, None)
|
||||||
if isinstance(v, LazyReference):
|
if isinstance(v, LazyReference):
|
||||||
# LazyReference inherits DBRef but should not be dereferenced here !
|
# LazyReference inherits DBRef but should not be dereferenced here !
|
||||||
@ -136,7 +134,7 @@ class DeReference(object):
|
|||||||
getattr(field, "field", None), "document_type", None
|
getattr(field, "field", None), "document_type", None
|
||||||
)
|
)
|
||||||
references = self._find_references(v, depth)
|
references = self._find_references(v, depth)
|
||||||
for key, refs in iteritems(references):
|
for key, refs in references.items():
|
||||||
if isinstance(
|
if isinstance(
|
||||||
field_cls, (Document, TopLevelDocumentMetaclass)
|
field_cls, (Document, TopLevelDocumentMetaclass)
|
||||||
):
|
):
|
||||||
@ -153,7 +151,7 @@ class DeReference(object):
|
|||||||
)
|
)
|
||||||
elif isinstance(item, (dict, list, tuple)) and depth - 1 <= self.max_depth:
|
elif isinstance(item, (dict, list, tuple)) and depth - 1 <= self.max_depth:
|
||||||
references = self._find_references(item, depth - 1)
|
references = self._find_references(item, depth - 1)
|
||||||
for key, refs in iteritems(references):
|
for key, refs in references.items():
|
||||||
reference_map.setdefault(key, set()).update(refs)
|
reference_map.setdefault(key, set()).update(refs)
|
||||||
|
|
||||||
return reference_map
|
return reference_map
|
||||||
@ -162,7 +160,7 @@ class DeReference(object):
|
|||||||
"""Fetch all references and convert to their document objects
|
"""Fetch all references and convert to their document objects
|
||||||
"""
|
"""
|
||||||
object_map = {}
|
object_map = {}
|
||||||
for collection, dbrefs in iteritems(self.reference_map):
|
for collection, dbrefs in self.reference_map.items():
|
||||||
|
|
||||||
# we use getattr instead of hasattr because hasattr swallows any exception under python2
|
# we use getattr instead of hasattr because hasattr swallows any exception under python2
|
||||||
# so it could hide nasty things without raising exceptions (cfr bug #1688))
|
# so it could hide nasty things without raising exceptions (cfr bug #1688))
|
||||||
@ -174,7 +172,7 @@ class DeReference(object):
|
|||||||
dbref for dbref in dbrefs if (col_name, dbref) not in object_map
|
dbref for dbref in dbrefs if (col_name, dbref) not in object_map
|
||||||
]
|
]
|
||||||
references = collection.objects.in_bulk(refs)
|
references = collection.objects.in_bulk(refs)
|
||||||
for key, doc in iteritems(references):
|
for key, doc in references.items():
|
||||||
object_map[(col_name, key)] = doc
|
object_map[(col_name, key)] = doc
|
||||||
else: # Generic reference: use the refs data to convert to document
|
else: # Generic reference: use the refs data to convert to document
|
||||||
if isinstance(doc_type, (ListField, DictField, MapField)):
|
if isinstance(doc_type, (ListField, DictField, MapField)):
|
||||||
@ -250,7 +248,7 @@ class DeReference(object):
|
|||||||
data = []
|
data = []
|
||||||
else:
|
else:
|
||||||
is_list = False
|
is_list = False
|
||||||
iterator = iteritems(items)
|
iterator = items.items()
|
||||||
data = {}
|
data = {}
|
||||||
|
|
||||||
depth += 1
|
depth += 1
|
||||||
@ -274,14 +272,12 @@ class DeReference(object):
|
|||||||
(v["_ref"].collection, v["_ref"].id), v
|
(v["_ref"].collection, v["_ref"].id), v
|
||||||
)
|
)
|
||||||
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
||||||
item_name = six.text_type("{0}.{1}.{2}").format(
|
item_name = "{}.{}.{}".format(name, k, field_name)
|
||||||
name, k, field_name
|
|
||||||
)
|
|
||||||
data[k]._data[field_name] = self._attach_objects(
|
data[k]._data[field_name] = self._attach_objects(
|
||||||
v, depth, instance=instance, name=item_name
|
v, depth, instance=instance, name=item_name
|
||||||
)
|
)
|
||||||
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
||||||
item_name = "%s.%s" % (name, k) if name else name
|
item_name = "{}.{}".format(name, k) if name else name
|
||||||
data[k] = self._attach_objects(
|
data[k] = self._attach_objects(
|
||||||
v, depth - 1, instance=instance, name=item_name
|
v, depth - 1, instance=instance, name=item_name
|
||||||
)
|
)
|
||||||
|
@ -4,8 +4,6 @@ import warnings
|
|||||||
from bson.dbref import DBRef
|
from bson.dbref import DBRef
|
||||||
import pymongo
|
import pymongo
|
||||||
from pymongo.read_preferences import ReadPreference
|
from pymongo.read_preferences import ReadPreference
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
from mongoengine.base import (
|
from mongoengine.base import (
|
||||||
@ -44,7 +42,7 @@ def includes_cls(fields):
|
|||||||
"""Helper function used for ensuring and comparing indexes."""
|
"""Helper function used for ensuring and comparing indexes."""
|
||||||
first_field = None
|
first_field = None
|
||||||
if len(fields):
|
if len(fields):
|
||||||
if isinstance(fields[0], six.string_types):
|
if isinstance(fields[0], str):
|
||||||
first_field = fields[0]
|
first_field = fields[0]
|
||||||
elif isinstance(fields[0], (list, tuple)) and len(fields[0]):
|
elif isinstance(fields[0], (list, tuple)) and len(fields[0]):
|
||||||
first_field = fields[0][0]
|
first_field = fields[0][0]
|
||||||
@ -55,8 +53,8 @@ class InvalidCollectionError(Exception):
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class EmbeddedDocument(six.with_metaclass(DocumentMetaclass, BaseDocument)):
|
class EmbeddedDocument(BaseDocument, metaclass=DocumentMetaclass):
|
||||||
"""A :class:`~mongoengine.Document` that isn't stored in its own
|
r"""A :class:`~mongoengine.Document` that isn't stored in its own
|
||||||
collection. :class:`~mongoengine.EmbeddedDocument`\ s should be used as
|
collection. :class:`~mongoengine.EmbeddedDocument`\ s should be used as
|
||||||
fields on :class:`~mongoengine.Document`\ s through the
|
fields on :class:`~mongoengine.Document`\ s through the
|
||||||
:class:`~mongoengine.EmbeddedDocumentField` field type.
|
:class:`~mongoengine.EmbeddedDocumentField` field type.
|
||||||
@ -71,7 +69,6 @@ class EmbeddedDocument(six.with_metaclass(DocumentMetaclass, BaseDocument)):
|
|||||||
|
|
||||||
__slots__ = ("_instance",)
|
__slots__ = ("_instance",)
|
||||||
|
|
||||||
# The __metaclass__ attribute is removed by 2to3 when running with Python3
|
|
||||||
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
||||||
my_metaclass = DocumentMetaclass
|
my_metaclass = DocumentMetaclass
|
||||||
|
|
||||||
@ -82,7 +79,7 @@ class EmbeddedDocument(six.with_metaclass(DocumentMetaclass, BaseDocument)):
|
|||||||
__hash__ = None
|
__hash__ = None
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
super(EmbeddedDocument, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
self._instance = None
|
self._instance = None
|
||||||
self._changed_fields = []
|
self._changed_fields = []
|
||||||
|
|
||||||
@ -95,7 +92,7 @@ class EmbeddedDocument(six.with_metaclass(DocumentMetaclass, BaseDocument)):
|
|||||||
return not self.__eq__(other)
|
return not self.__eq__(other)
|
||||||
|
|
||||||
def to_mongo(self, *args, **kwargs):
|
def to_mongo(self, *args, **kwargs):
|
||||||
data = super(EmbeddedDocument, self).to_mongo(*args, **kwargs)
|
data = super().to_mongo(*args, **kwargs)
|
||||||
|
|
||||||
# remove _id from the SON if it's in it and it's None
|
# remove _id from the SON if it's in it and it's None
|
||||||
if "_id" in data and data["_id"] is None:
|
if "_id" in data and data["_id"] is None:
|
||||||
@ -104,7 +101,7 @@ class EmbeddedDocument(six.with_metaclass(DocumentMetaclass, BaseDocument)):
|
|||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
||||||
"""The base class used for defining the structure and properties of
|
"""The base class used for defining the structure and properties of
|
||||||
collections of documents stored in MongoDB. Inherit from this class, and
|
collections of documents stored in MongoDB. Inherit from this class, and
|
||||||
add fields as class attributes to define a document's structure.
|
add fields as class attributes to define a document's structure.
|
||||||
@ -156,7 +153,6 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
in the :attr:`meta` dictionary.
|
in the :attr:`meta` dictionary.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# The __metaclass__ attribute is removed by 2to3 when running with Python3
|
|
||||||
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
||||||
my_metaclass = TopLevelDocumentMetaclass
|
my_metaclass = TopLevelDocumentMetaclass
|
||||||
|
|
||||||
@ -260,7 +256,7 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
return db.create_collection(collection_name, **opts)
|
return db.create_collection(collection_name, **opts)
|
||||||
|
|
||||||
def to_mongo(self, *args, **kwargs):
|
def to_mongo(self, *args, **kwargs):
|
||||||
data = super(Document, self).to_mongo(*args, **kwargs)
|
data = super().to_mongo(*args, **kwargs)
|
||||||
|
|
||||||
# If '_id' is None, try and set it from self._data. If that
|
# If '_id' is None, try and set it from self._data. If that
|
||||||
# doesn't exist either, remove '_id' from the SON completely.
|
# doesn't exist either, remove '_id' from the SON completely.
|
||||||
@ -332,7 +328,7 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
):
|
):
|
||||||
"""Save the :class:`~mongoengine.Document` to the database. If the
|
"""Save the :class:`~mongoengine.Document` to the database. If the
|
||||||
document already exists, it will be updated, otherwise it will be
|
document already exists, it will be updated, otherwise it will be
|
||||||
created.
|
created. Returns the saved object instance.
|
||||||
|
|
||||||
:param force_insert: only try to create a new document, don't allow
|
:param force_insert: only try to create a new document, don't allow
|
||||||
updates of existing documents.
|
updates of existing documents.
|
||||||
@ -431,16 +427,16 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
self.cascade_save(**kwargs)
|
self.cascade_save(**kwargs)
|
||||||
|
|
||||||
except pymongo.errors.DuplicateKeyError as err:
|
except pymongo.errors.DuplicateKeyError as err:
|
||||||
message = u"Tried to save duplicate unique keys (%s)"
|
message = "Tried to save duplicate unique keys (%s)"
|
||||||
raise NotUniqueError(message % six.text_type(err))
|
raise NotUniqueError(message % err)
|
||||||
except pymongo.errors.OperationFailure as err:
|
except pymongo.errors.OperationFailure as err:
|
||||||
message = "Could not save document (%s)"
|
message = "Could not save document (%s)"
|
||||||
if re.match("^E1100[01] duplicate key", six.text_type(err)):
|
if re.match("^E1100[01] duplicate key", str(err)):
|
||||||
# E11000 - duplicate key error index
|
# E11000 - duplicate key error index
|
||||||
# E11001 - duplicate key on update
|
# E11001 - duplicate key on update
|
||||||
message = u"Tried to save duplicate unique keys (%s)"
|
message = "Tried to save duplicate unique keys (%s)"
|
||||||
raise NotUniqueError(message % six.text_type(err))
|
raise NotUniqueError(message % err)
|
||||||
raise OperationError(message % six.text_type(err))
|
raise OperationError(message % err)
|
||||||
|
|
||||||
# Make sure we store the PK on this document now that it's saved
|
# Make sure we store the PK on this document now that it's saved
|
||||||
id_field = self._meta["id_field"]
|
id_field = self._meta["id_field"]
|
||||||
@ -559,7 +555,7 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
if not getattr(ref, "_changed_fields", True):
|
if not getattr(ref, "_changed_fields", True):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
ref_id = "%s,%s" % (ref.__class__.__name__, str(ref._data))
|
ref_id = "{},{}".format(ref.__class__.__name__, str(ref._data))
|
||||||
if ref and ref_id not in _refs:
|
if ref and ref_id not in _refs:
|
||||||
_refs.append(ref_id)
|
_refs.append(ref_id)
|
||||||
kwargs["_refs"] = _refs
|
kwargs["_refs"] = _refs
|
||||||
@ -634,7 +630,7 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
|
|
||||||
# Delete FileFields separately
|
# Delete FileFields separately
|
||||||
FileField = _import_class("FileField")
|
FileField = _import_class("FileField")
|
||||||
for name, field in iteritems(self._fields):
|
for name, field in self._fields.items():
|
||||||
if isinstance(field, FileField):
|
if isinstance(field, FileField):
|
||||||
getattr(self, name).delete()
|
getattr(self, name).delete()
|
||||||
|
|
||||||
@ -643,7 +639,7 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
write_concern=write_concern, _from_doc_delete=True
|
write_concern=write_concern, _from_doc_delete=True
|
||||||
)
|
)
|
||||||
except pymongo.errors.OperationFailure as err:
|
except pymongo.errors.OperationFailure as err:
|
||||||
message = u"Could not delete document (%s)" % err.message
|
message = "Could not delete document (%s)" % err.args
|
||||||
raise OperationError(message)
|
raise OperationError(message)
|
||||||
signals.post_delete.send(self.__class__, document=self, **signal_kwargs)
|
signals.post_delete.send(self.__class__, document=self, **signal_kwargs)
|
||||||
|
|
||||||
@ -851,17 +847,13 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
index_spec = cls._build_index_spec(keys)
|
index_spec = cls._build_index_spec(keys)
|
||||||
index_spec = index_spec.copy()
|
index_spec = index_spec.copy()
|
||||||
fields = index_spec.pop("fields")
|
fields = index_spec.pop("fields")
|
||||||
drop_dups = kwargs.get("drop_dups", False)
|
|
||||||
if drop_dups:
|
|
||||||
msg = "drop_dups is deprecated and is removed when using PyMongo 3+."
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
index_spec["background"] = background
|
index_spec["background"] = background
|
||||||
index_spec.update(kwargs)
|
index_spec.update(kwargs)
|
||||||
|
|
||||||
return cls._get_collection().create_index(fields, **index_spec)
|
return cls._get_collection().create_index(fields, **index_spec)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def ensure_index(cls, key_or_list, drop_dups=False, background=False, **kwargs):
|
def ensure_index(cls, key_or_list, background=False, **kwargs):
|
||||||
"""Ensure that the given indexes are in place. Deprecated in favour
|
"""Ensure that the given indexes are in place. Deprecated in favour
|
||||||
of create_index.
|
of create_index.
|
||||||
|
|
||||||
@ -869,12 +861,7 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
construct a multi-field index); keys may be prefixed with a **+**
|
construct a multi-field index); keys may be prefixed with a **+**
|
||||||
or a **-** to determine the index ordering
|
or a **-** to determine the index ordering
|
||||||
:param background: Allows index creation in the background
|
:param background: Allows index creation in the background
|
||||||
:param drop_dups: Was removed/ignored with MongoDB >2.7.5. The value
|
|
||||||
will be removed if PyMongo3+ is used
|
|
||||||
"""
|
"""
|
||||||
if drop_dups:
|
|
||||||
msg = "drop_dups is deprecated and is removed when using PyMongo 3+."
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
return cls.create_index(key_or_list, background=background, **kwargs)
|
return cls.create_index(key_or_list, background=background, **kwargs)
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
@ -887,12 +874,8 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
`auto_create_index` to False in the documents meta data
|
`auto_create_index` to False in the documents meta data
|
||||||
"""
|
"""
|
||||||
background = cls._meta.get("index_background", False)
|
background = cls._meta.get("index_background", False)
|
||||||
drop_dups = cls._meta.get("index_drop_dups", False)
|
|
||||||
index_opts = cls._meta.get("index_opts") or {}
|
index_opts = cls._meta.get("index_opts") or {}
|
||||||
index_cls = cls._meta.get("index_cls", True)
|
index_cls = cls._meta.get("index_cls", True)
|
||||||
if drop_dups:
|
|
||||||
msg = "drop_dups is deprecated and is removed when using PyMongo 3+."
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
|
|
||||||
collection = cls._get_collection()
|
collection = cls._get_collection()
|
||||||
# 746: when connection is via mongos, the read preference is not necessarily an indication that
|
# 746: when connection is via mongos, the read preference is not necessarily an indication that
|
||||||
@ -992,10 +975,10 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
indexes.append(index)
|
indexes.append(index)
|
||||||
|
|
||||||
# finish up by appending { '_id': 1 } and { '_cls': 1 }, if needed
|
# finish up by appending { '_id': 1 } and { '_cls': 1 }, if needed
|
||||||
if [(u"_id", 1)] not in indexes:
|
if [("_id", 1)] not in indexes:
|
||||||
indexes.append([(u"_id", 1)])
|
indexes.append([("_id", 1)])
|
||||||
if cls._meta.get("index_cls", True) and cls._meta.get("allow_inheritance"):
|
if cls._meta.get("index_cls", True) and cls._meta.get("allow_inheritance"):
|
||||||
indexes.append([(u"_cls", 1)])
|
indexes.append([("_cls", 1)])
|
||||||
|
|
||||||
return indexes
|
return indexes
|
||||||
|
|
||||||
@ -1019,19 +1002,19 @@ class Document(six.with_metaclass(TopLevelDocumentMetaclass, BaseDocument)):
|
|||||||
extra = [index for index in existing if index not in required]
|
extra = [index for index in existing if index not in required]
|
||||||
|
|
||||||
# if { _cls: 1 } is missing, make sure it's *really* necessary
|
# if { _cls: 1 } is missing, make sure it's *really* necessary
|
||||||
if [(u"_cls", 1)] in missing:
|
if [("_cls", 1)] in missing:
|
||||||
cls_obsolete = False
|
cls_obsolete = False
|
||||||
for index in existing:
|
for index in existing:
|
||||||
if includes_cls(index) and index not in extra:
|
if includes_cls(index) and index not in extra:
|
||||||
cls_obsolete = True
|
cls_obsolete = True
|
||||||
break
|
break
|
||||||
if cls_obsolete:
|
if cls_obsolete:
|
||||||
missing.remove([(u"_cls", 1)])
|
missing.remove([("_cls", 1)])
|
||||||
|
|
||||||
return {"missing": missing, "extra": extra}
|
return {"missing": missing, "extra": extra}
|
||||||
|
|
||||||
|
|
||||||
class DynamicDocument(six.with_metaclass(TopLevelDocumentMetaclass, Document)):
|
class DynamicDocument(Document, metaclass=TopLevelDocumentMetaclass):
|
||||||
"""A Dynamic Document class allowing flexible, expandable and uncontrolled
|
"""A Dynamic Document class allowing flexible, expandable and uncontrolled
|
||||||
schemas. As a :class:`~mongoengine.Document` subclass, acts in the same
|
schemas. As a :class:`~mongoengine.Document` subclass, acts in the same
|
||||||
way as an ordinary document but has expanded style properties. Any data
|
way as an ordinary document but has expanded style properties. Any data
|
||||||
@ -1045,7 +1028,6 @@ class DynamicDocument(six.with_metaclass(TopLevelDocumentMetaclass, Document)):
|
|||||||
There is one caveat on Dynamic Documents: undeclared fields cannot start with `_`
|
There is one caveat on Dynamic Documents: undeclared fields cannot start with `_`
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# The __metaclass__ attribute is removed by 2to3 when running with Python3
|
|
||||||
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
||||||
my_metaclass = TopLevelDocumentMetaclass
|
my_metaclass = TopLevelDocumentMetaclass
|
||||||
|
|
||||||
@ -1060,16 +1042,15 @@ class DynamicDocument(six.with_metaclass(TopLevelDocumentMetaclass, Document)):
|
|||||||
setattr(self, field_name, None)
|
setattr(self, field_name, None)
|
||||||
self._dynamic_fields[field_name].null = False
|
self._dynamic_fields[field_name].null = False
|
||||||
else:
|
else:
|
||||||
super(DynamicDocument, self).__delattr__(*args, **kwargs)
|
super().__delattr__(*args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
class DynamicEmbeddedDocument(six.with_metaclass(DocumentMetaclass, EmbeddedDocument)):
|
class DynamicEmbeddedDocument(EmbeddedDocument, metaclass=DocumentMetaclass):
|
||||||
"""A Dynamic Embedded Document class allowing flexible, expandable and
|
"""A Dynamic Embedded Document class allowing flexible, expandable and
|
||||||
uncontrolled schemas. See :class:`~mongoengine.DynamicDocument` for more
|
uncontrolled schemas. See :class:`~mongoengine.DynamicDocument` for more
|
||||||
information about dynamic documents.
|
information about dynamic documents.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# The __metaclass__ attribute is removed by 2to3 when running with Python3
|
|
||||||
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
# my_metaclass is defined so that metaclass can be queried in Python 2 & 3
|
||||||
my_metaclass = DocumentMetaclass
|
my_metaclass = DocumentMetaclass
|
||||||
|
|
||||||
@ -1089,7 +1070,7 @@ class DynamicEmbeddedDocument(six.with_metaclass(DocumentMetaclass, EmbeddedDocu
|
|||||||
setattr(self, field_name, None)
|
setattr(self, field_name, None)
|
||||||
|
|
||||||
|
|
||||||
class MapReduceDocument(object):
|
class MapReduceDocument:
|
||||||
"""A document returned from a map/reduce query.
|
"""A document returned from a map/reduce query.
|
||||||
|
|
||||||
:param collection: An instance of :class:`~pymongo.Collection`
|
:param collection: An instance of :class:`~pymongo.Collection`
|
||||||
|
@ -1,7 +1,5 @@
|
|||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
|
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
"NotRegistered",
|
"NotRegistered",
|
||||||
@ -87,24 +85,24 @@ class ValidationError(AssertionError):
|
|||||||
_message = None
|
_message = None
|
||||||
|
|
||||||
def __init__(self, message="", **kwargs):
|
def __init__(self, message="", **kwargs):
|
||||||
super(ValidationError, self).__init__(message)
|
super().__init__(message)
|
||||||
self.errors = kwargs.get("errors", {})
|
self.errors = kwargs.get("errors", {})
|
||||||
self.field_name = kwargs.get("field_name")
|
self.field_name = kwargs.get("field_name")
|
||||||
self.message = message
|
self.message = message
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
return six.text_type(self.message)
|
return str(self.message)
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "%s(%s,)" % (self.__class__.__name__, self.message)
|
return "{}({},)".format(self.__class__.__name__, self.message)
|
||||||
|
|
||||||
def __getattribute__(self, name):
|
def __getattribute__(self, name):
|
||||||
message = super(ValidationError, self).__getattribute__(name)
|
message = super().__getattribute__(name)
|
||||||
if name == "message":
|
if name == "message":
|
||||||
if self.field_name:
|
if self.field_name:
|
||||||
message = "%s" % message
|
message = "%s" % message
|
||||||
if self.errors:
|
if self.errors:
|
||||||
message = "%s(%s)" % (message, self._format_errors())
|
message = "{}({})".format(message, self._format_errors())
|
||||||
return message
|
return message
|
||||||
|
|
||||||
def _get_message(self):
|
def _get_message(self):
|
||||||
@ -126,12 +124,12 @@ class ValidationError(AssertionError):
|
|||||||
def build_dict(source):
|
def build_dict(source):
|
||||||
errors_dict = {}
|
errors_dict = {}
|
||||||
if isinstance(source, dict):
|
if isinstance(source, dict):
|
||||||
for field_name, error in iteritems(source):
|
for field_name, error in source.items():
|
||||||
errors_dict[field_name] = build_dict(error)
|
errors_dict[field_name] = build_dict(error)
|
||||||
elif isinstance(source, ValidationError) and source.errors:
|
elif isinstance(source, ValidationError) and source.errors:
|
||||||
return build_dict(source.errors)
|
return build_dict(source.errors)
|
||||||
else:
|
else:
|
||||||
return six.text_type(source)
|
return str(source)
|
||||||
|
|
||||||
return errors_dict
|
return errors_dict
|
||||||
|
|
||||||
@ -147,15 +145,15 @@ class ValidationError(AssertionError):
|
|||||||
if isinstance(value, list):
|
if isinstance(value, list):
|
||||||
value = " ".join([generate_key(k) for k in value])
|
value = " ".join([generate_key(k) for k in value])
|
||||||
elif isinstance(value, dict):
|
elif isinstance(value, dict):
|
||||||
value = " ".join([generate_key(v, k) for k, v in iteritems(value)])
|
value = " ".join([generate_key(v, k) for k, v in value.items()])
|
||||||
|
|
||||||
results = "%s.%s" % (prefix, value) if prefix else value
|
results = "{}.{}".format(prefix, value) if prefix else value
|
||||||
return results
|
return results
|
||||||
|
|
||||||
error_dict = defaultdict(list)
|
error_dict = defaultdict(list)
|
||||||
for k, v in iteritems(self.to_dict()):
|
for k, v in self.to_dict().items():
|
||||||
error_dict[generate_key(v)].append(k)
|
error_dict[generate_key(v)].append(k)
|
||||||
return " ".join(["%s: %s" % (k, v) for k, v in iteritems(error_dict)])
|
return " ".join(["{}: {}".format(k, v) for k, v in error_dict.items()])
|
||||||
|
|
||||||
|
|
||||||
class DeprecatedError(Exception):
|
class DeprecatedError(Exception):
|
||||||
|
@ -5,14 +5,14 @@ import re
|
|||||||
import socket
|
import socket
|
||||||
import time
|
import time
|
||||||
import uuid
|
import uuid
|
||||||
|
from io import BytesIO
|
||||||
from operator import itemgetter
|
from operator import itemgetter
|
||||||
|
|
||||||
from bson import Binary, DBRef, ObjectId, SON
|
from bson import Binary, DBRef, ObjectId, SON
|
||||||
|
from bson.int64 import Int64
|
||||||
import gridfs
|
import gridfs
|
||||||
import pymongo
|
import pymongo
|
||||||
from pymongo import ReturnDocument
|
from pymongo import ReturnDocument
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import dateutil
|
import dateutil
|
||||||
@ -21,11 +21,6 @@ except ImportError:
|
|||||||
else:
|
else:
|
||||||
import dateutil.parser
|
import dateutil.parser
|
||||||
|
|
||||||
try:
|
|
||||||
from bson.int64 import Int64
|
|
||||||
except ImportError:
|
|
||||||
Int64 = long
|
|
||||||
|
|
||||||
|
|
||||||
from mongoengine.base import (
|
from mongoengine.base import (
|
||||||
BaseDocument,
|
BaseDocument,
|
||||||
@ -42,7 +37,6 @@ from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
|||||||
from mongoengine.document import Document, EmbeddedDocument
|
from mongoengine.document import Document, EmbeddedDocument
|
||||||
from mongoengine.errors import DoesNotExist, InvalidQueryError, ValidationError
|
from mongoengine.errors import DoesNotExist, InvalidQueryError, ValidationError
|
||||||
from mongoengine.mongodb_support import MONGODB_36, get_mongodb_version
|
from mongoengine.mongodb_support import MONGODB_36, get_mongodb_version
|
||||||
from mongoengine.python_support import StringIO
|
|
||||||
from mongoengine.queryset import DO_NOTHING
|
from mongoengine.queryset import DO_NOTHING
|
||||||
from mongoengine.queryset.base import BaseQuerySet
|
from mongoengine.queryset.base import BaseQuerySet
|
||||||
from mongoengine.queryset.transform import STRING_OPERATORS
|
from mongoengine.queryset.transform import STRING_OPERATORS
|
||||||
@ -53,11 +47,6 @@ except ImportError:
|
|||||||
Image = None
|
Image = None
|
||||||
ImageOps = None
|
ImageOps = None
|
||||||
|
|
||||||
if six.PY3:
|
|
||||||
# Useless as long as 2to3 gets executed
|
|
||||||
# as it turns `long` into `int` blindly
|
|
||||||
long = int
|
|
||||||
|
|
||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
"StringField",
|
"StringField",
|
||||||
@ -114,10 +103,10 @@ class StringField(BaseField):
|
|||||||
self.regex = re.compile(regex) if regex else None
|
self.regex = re.compile(regex) if regex else None
|
||||||
self.max_length = max_length
|
self.max_length = max_length
|
||||||
self.min_length = min_length
|
self.min_length = min_length
|
||||||
super(StringField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
if isinstance(value, six.text_type):
|
if isinstance(value, str):
|
||||||
return value
|
return value
|
||||||
try:
|
try:
|
||||||
value = value.decode("utf-8")
|
value = value.decode("utf-8")
|
||||||
@ -126,7 +115,7 @@ class StringField(BaseField):
|
|||||||
return value
|
return value
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if not isinstance(value, six.string_types):
|
if not isinstance(value, str):
|
||||||
self.error("StringField only accepts string values")
|
self.error("StringField only accepts string values")
|
||||||
|
|
||||||
if self.max_length is not None and len(value) > self.max_length:
|
if self.max_length is not None and len(value) > self.max_length:
|
||||||
@ -142,7 +131,7 @@ class StringField(BaseField):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if not isinstance(op, six.string_types):
|
if not isinstance(op, str):
|
||||||
return value
|
return value
|
||||||
|
|
||||||
if op in STRING_OPERATORS:
|
if op in STRING_OPERATORS:
|
||||||
@ -162,7 +151,7 @@ class StringField(BaseField):
|
|||||||
# escape unsafe characters which could lead to a re.error
|
# escape unsafe characters which could lead to a re.error
|
||||||
value = re.escape(value)
|
value = re.escape(value)
|
||||||
value = re.compile(regex % value, flags)
|
value = re.compile(regex % value, flags)
|
||||||
return super(StringField, self).prepare_query_value(op, value)
|
return super().prepare_query_value(op, value)
|
||||||
|
|
||||||
|
|
||||||
class URLField(StringField):
|
class URLField(StringField):
|
||||||
@ -186,17 +175,17 @@ class URLField(StringField):
|
|||||||
def __init__(self, url_regex=None, schemes=None, **kwargs):
|
def __init__(self, url_regex=None, schemes=None, **kwargs):
|
||||||
self.url_regex = url_regex or self._URL_REGEX
|
self.url_regex = url_regex or self._URL_REGEX
|
||||||
self.schemes = schemes or self._URL_SCHEMES
|
self.schemes = schemes or self._URL_SCHEMES
|
||||||
super(URLField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
# Check first if the scheme is valid
|
# Check first if the scheme is valid
|
||||||
scheme = value.split("://")[0].lower()
|
scheme = value.split("://")[0].lower()
|
||||||
if scheme not in self.schemes:
|
if scheme not in self.schemes:
|
||||||
self.error(u"Invalid scheme {} in URL: {}".format(scheme, value))
|
self.error("Invalid scheme {} in URL: {}".format(scheme, value))
|
||||||
|
|
||||||
# Then check full URL
|
# Then check full URL
|
||||||
if not self.url_regex.match(value):
|
if not self.url_regex.match(value):
|
||||||
self.error(u"Invalid URL: {}".format(value))
|
self.error("Invalid URL: {}".format(value))
|
||||||
|
|
||||||
|
|
||||||
class EmailField(StringField):
|
class EmailField(StringField):
|
||||||
@ -214,7 +203,7 @@ class EmailField(StringField):
|
|||||||
)
|
)
|
||||||
|
|
||||||
UTF8_USER_REGEX = LazyRegexCompiler(
|
UTF8_USER_REGEX = LazyRegexCompiler(
|
||||||
six.u(
|
(
|
||||||
# RFC 6531 Section 3.3 extends `atext` (used by dot-atom) to
|
# RFC 6531 Section 3.3 extends `atext` (used by dot-atom) to
|
||||||
# include `UTF8-non-ascii`.
|
# include `UTF8-non-ascii`.
|
||||||
r"(^[-!#$%&'*+/=?^_`{}|~0-9A-Z\u0080-\U0010FFFF]+(\.[-!#$%&'*+/=?^_`{}|~0-9A-Z\u0080-\U0010FFFF]+)*\Z"
|
r"(^[-!#$%&'*+/=?^_`{}|~0-9A-Z\u0080-\U0010FFFF]+(\.[-!#$%&'*+/=?^_`{}|~0-9A-Z\u0080-\U0010FFFF]+)*\Z"
|
||||||
@ -229,7 +218,7 @@ class EmailField(StringField):
|
|||||||
re.IGNORECASE,
|
re.IGNORECASE,
|
||||||
)
|
)
|
||||||
|
|
||||||
error_msg = u"Invalid email address: %s"
|
error_msg = "Invalid email address: %s"
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
@ -253,7 +242,7 @@ class EmailField(StringField):
|
|||||||
self.domain_whitelist = domain_whitelist or []
|
self.domain_whitelist = domain_whitelist or []
|
||||||
self.allow_utf8_user = allow_utf8_user
|
self.allow_utf8_user = allow_utf8_user
|
||||||
self.allow_ip_domain = allow_ip_domain
|
self.allow_ip_domain = allow_ip_domain
|
||||||
super(EmailField, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def validate_user_part(self, user_part):
|
def validate_user_part(self, user_part):
|
||||||
"""Validate the user part of the email address. Return True if
|
"""Validate the user part of the email address. Return True if
|
||||||
@ -280,13 +269,13 @@ class EmailField(StringField):
|
|||||||
try:
|
try:
|
||||||
socket.inet_pton(addr_family, domain_part[1:-1])
|
socket.inet_pton(addr_family, domain_part[1:-1])
|
||||||
return True
|
return True
|
||||||
except (socket.error, UnicodeEncodeError):
|
except (OSError, UnicodeEncodeError):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
super(EmailField, self).validate(value)
|
super().validate(value)
|
||||||
|
|
||||||
if "@" not in value:
|
if "@" not in value:
|
||||||
self.error(self.error_msg % value)
|
self.error(self.error_msg % value)
|
||||||
@ -303,12 +292,16 @@ class EmailField(StringField):
|
|||||||
domain_part = domain_part.encode("idna").decode("ascii")
|
domain_part = domain_part.encode("idna").decode("ascii")
|
||||||
except UnicodeError:
|
except UnicodeError:
|
||||||
self.error(
|
self.error(
|
||||||
"%s %s" % (self.error_msg % value, "(domain failed IDN encoding)")
|
"{} {}".format(
|
||||||
|
self.error_msg % value, "(domain failed IDN encoding)"
|
||||||
|
)
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
if not self.validate_domain_part(domain_part):
|
if not self.validate_domain_part(domain_part):
|
||||||
self.error(
|
self.error(
|
||||||
"%s %s" % (self.error_msg % value, "(domain validation failed)")
|
"{} {}".format(
|
||||||
|
self.error_msg % value, "(domain validation failed)"
|
||||||
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -317,7 +310,7 @@ class IntField(BaseField):
|
|||||||
|
|
||||||
def __init__(self, min_value=None, max_value=None, **kwargs):
|
def __init__(self, min_value=None, max_value=None, **kwargs):
|
||||||
self.min_value, self.max_value = min_value, max_value
|
self.min_value, self.max_value = min_value, max_value
|
||||||
super(IntField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
try:
|
try:
|
||||||
@ -342,19 +335,19 @@ class IntField(BaseField):
|
|||||||
if value is None:
|
if value is None:
|
||||||
return value
|
return value
|
||||||
|
|
||||||
return super(IntField, self).prepare_query_value(op, int(value))
|
return super().prepare_query_value(op, int(value))
|
||||||
|
|
||||||
|
|
||||||
class LongField(BaseField):
|
class LongField(BaseField):
|
||||||
"""64-bit integer field."""
|
"""64-bit integer field. (Equivalent to IntField since the support to Python2 was dropped)"""
|
||||||
|
|
||||||
def __init__(self, min_value=None, max_value=None, **kwargs):
|
def __init__(self, min_value=None, max_value=None, **kwargs):
|
||||||
self.min_value, self.max_value = min_value, max_value
|
self.min_value, self.max_value = min_value, max_value
|
||||||
super(LongField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
try:
|
try:
|
||||||
value = long(value)
|
value = int(value)
|
||||||
except (TypeError, ValueError):
|
except (TypeError, ValueError):
|
||||||
pass
|
pass
|
||||||
return value
|
return value
|
||||||
@ -364,7 +357,7 @@ class LongField(BaseField):
|
|||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
try:
|
try:
|
||||||
value = long(value)
|
value = int(value)
|
||||||
except (TypeError, ValueError):
|
except (TypeError, ValueError):
|
||||||
self.error("%s could not be converted to long" % value)
|
self.error("%s could not be converted to long" % value)
|
||||||
|
|
||||||
@ -378,7 +371,7 @@ class LongField(BaseField):
|
|||||||
if value is None:
|
if value is None:
|
||||||
return value
|
return value
|
||||||
|
|
||||||
return super(LongField, self).prepare_query_value(op, long(value))
|
return super().prepare_query_value(op, int(value))
|
||||||
|
|
||||||
|
|
||||||
class FloatField(BaseField):
|
class FloatField(BaseField):
|
||||||
@ -386,7 +379,7 @@ class FloatField(BaseField):
|
|||||||
|
|
||||||
def __init__(self, min_value=None, max_value=None, **kwargs):
|
def __init__(self, min_value=None, max_value=None, **kwargs):
|
||||||
self.min_value, self.max_value = min_value, max_value
|
self.min_value, self.max_value = min_value, max_value
|
||||||
super(FloatField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
try:
|
try:
|
||||||
@ -396,7 +389,7 @@ class FloatField(BaseField):
|
|||||||
return value
|
return value
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if isinstance(value, six.integer_types):
|
if isinstance(value, int):
|
||||||
try:
|
try:
|
||||||
value = float(value)
|
value = float(value)
|
||||||
except OverflowError:
|
except OverflowError:
|
||||||
@ -415,7 +408,7 @@ class FloatField(BaseField):
|
|||||||
if value is None:
|
if value is None:
|
||||||
return value
|
return value
|
||||||
|
|
||||||
return super(FloatField, self).prepare_query_value(op, float(value))
|
return super().prepare_query_value(op, float(value))
|
||||||
|
|
||||||
|
|
||||||
class DecimalField(BaseField):
|
class DecimalField(BaseField):
|
||||||
@ -462,7 +455,7 @@ class DecimalField(BaseField):
|
|||||||
self.precision = precision
|
self.precision = precision
|
||||||
self.rounding = rounding
|
self.rounding = rounding
|
||||||
|
|
||||||
super(DecimalField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
if value is None:
|
if value is None:
|
||||||
@ -481,13 +474,13 @@ class DecimalField(BaseField):
|
|||||||
if value is None:
|
if value is None:
|
||||||
return value
|
return value
|
||||||
if self.force_string:
|
if self.force_string:
|
||||||
return six.text_type(self.to_python(value))
|
return str(self.to_python(value))
|
||||||
return float(self.to_python(value))
|
return float(self.to_python(value))
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if not isinstance(value, decimal.Decimal):
|
if not isinstance(value, decimal.Decimal):
|
||||||
if not isinstance(value, six.string_types):
|
if not isinstance(value, str):
|
||||||
value = six.text_type(value)
|
value = str(value)
|
||||||
try:
|
try:
|
||||||
value = decimal.Decimal(value)
|
value = decimal.Decimal(value)
|
||||||
except (TypeError, ValueError, decimal.InvalidOperation) as exc:
|
except (TypeError, ValueError, decimal.InvalidOperation) as exc:
|
||||||
@ -500,7 +493,7 @@ class DecimalField(BaseField):
|
|||||||
self.error("Decimal value is too large")
|
self.error("Decimal value is too large")
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
return super(DecimalField, self).prepare_query_value(op, self.to_mongo(value))
|
return super().prepare_query_value(op, self.to_mongo(value))
|
||||||
|
|
||||||
|
|
||||||
class BooleanField(BaseField):
|
class BooleanField(BaseField):
|
||||||
@ -540,7 +533,7 @@ class DateTimeField(BaseField):
|
|||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
new_value = self.to_mongo(value)
|
new_value = self.to_mongo(value)
|
||||||
if not isinstance(new_value, (datetime.datetime, datetime.date)):
|
if not isinstance(new_value, (datetime.datetime, datetime.date)):
|
||||||
self.error(u'cannot parse date "%s"' % value)
|
self.error('cannot parse date "%s"' % value)
|
||||||
|
|
||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
if value is None:
|
if value is None:
|
||||||
@ -552,7 +545,7 @@ class DateTimeField(BaseField):
|
|||||||
if callable(value):
|
if callable(value):
|
||||||
return value()
|
return value()
|
||||||
|
|
||||||
if not isinstance(value, six.string_types):
|
if not isinstance(value, str):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
return self._parse_datetime(value)
|
return self._parse_datetime(value)
|
||||||
@ -597,19 +590,19 @@ class DateTimeField(BaseField):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
return super(DateTimeField, self).prepare_query_value(op, self.to_mongo(value))
|
return super().prepare_query_value(op, self.to_mongo(value))
|
||||||
|
|
||||||
|
|
||||||
class DateField(DateTimeField):
|
class DateField(DateTimeField):
|
||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
value = super(DateField, self).to_mongo(value)
|
value = super().to_mongo(value)
|
||||||
# drop hours, minutes, seconds
|
# drop hours, minutes, seconds
|
||||||
if isinstance(value, datetime.datetime):
|
if isinstance(value, datetime.datetime):
|
||||||
value = datetime.datetime(value.year, value.month, value.day)
|
value = datetime.datetime(value.year, value.month, value.day)
|
||||||
return value
|
return value
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
value = super(DateField, self).to_python(value)
|
value = super().to_python(value)
|
||||||
# convert datetime to date
|
# convert datetime to date
|
||||||
if isinstance(value, datetime.datetime):
|
if isinstance(value, datetime.datetime):
|
||||||
value = datetime.date(value.year, value.month, value.day)
|
value = datetime.date(value.year, value.month, value.day)
|
||||||
@ -643,7 +636,7 @@ class ComplexDateTimeField(StringField):
|
|||||||
"""
|
"""
|
||||||
self.separator = separator
|
self.separator = separator
|
||||||
self.format = separator.join(["%Y", "%m", "%d", "%H", "%M", "%S", "%f"])
|
self.format = separator.join(["%Y", "%m", "%d", "%H", "%M", "%S", "%f"])
|
||||||
super(ComplexDateTimeField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def _convert_from_datetime(self, val):
|
def _convert_from_datetime(self, val):
|
||||||
"""
|
"""
|
||||||
@ -674,17 +667,20 @@ class ComplexDateTimeField(StringField):
|
|||||||
if instance is None:
|
if instance is None:
|
||||||
return self
|
return self
|
||||||
|
|
||||||
data = super(ComplexDateTimeField, self).__get__(instance, owner)
|
data = super().__get__(instance, owner)
|
||||||
|
|
||||||
if isinstance(data, datetime.datetime) or data is None:
|
if isinstance(data, datetime.datetime) or data is None:
|
||||||
return data
|
return data
|
||||||
return self._convert_from_string(data)
|
return self._convert_from_string(data)
|
||||||
|
|
||||||
def __set__(self, instance, value):
|
def __set__(self, instance, value):
|
||||||
super(ComplexDateTimeField, self).__set__(instance, value)
|
super().__set__(instance, value)
|
||||||
value = instance._data[self.name]
|
value = instance._data[self.name]
|
||||||
if value is not None:
|
if value is not None:
|
||||||
instance._data[self.name] = self._convert_from_datetime(value)
|
if isinstance(value, datetime.datetime):
|
||||||
|
instance._data[self.name] = self._convert_from_datetime(value)
|
||||||
|
else:
|
||||||
|
instance._data[self.name] = value
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
value = self.to_python(value)
|
value = self.to_python(value)
|
||||||
@ -703,9 +699,7 @@ class ComplexDateTimeField(StringField):
|
|||||||
return self._convert_from_datetime(value)
|
return self._convert_from_datetime(value)
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
return super(ComplexDateTimeField, self).prepare_query_value(
|
return super().prepare_query_value(op, self._convert_from_datetime(value))
|
||||||
op, self._convert_from_datetime(value)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class EmbeddedDocumentField(BaseField):
|
class EmbeddedDocumentField(BaseField):
|
||||||
@ -716,7 +710,7 @@ class EmbeddedDocumentField(BaseField):
|
|||||||
def __init__(self, document_type, **kwargs):
|
def __init__(self, document_type, **kwargs):
|
||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
if not (
|
if not (
|
||||||
isinstance(document_type, six.string_types)
|
isinstance(document_type, str)
|
||||||
or issubclass(document_type, EmbeddedDocument)
|
or issubclass(document_type, EmbeddedDocument)
|
||||||
):
|
):
|
||||||
self.error(
|
self.error(
|
||||||
@ -725,11 +719,11 @@ class EmbeddedDocumentField(BaseField):
|
|||||||
)
|
)
|
||||||
|
|
||||||
self.document_type_obj = document_type
|
self.document_type_obj = document_type
|
||||||
super(EmbeddedDocumentField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def document_type(self):
|
def document_type(self):
|
||||||
if isinstance(self.document_type_obj, six.string_types):
|
if isinstance(self.document_type_obj, str):
|
||||||
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
||||||
resolved_document_type = self.owner_document
|
resolved_document_type = self.owner_document
|
||||||
else:
|
else:
|
||||||
@ -779,6 +773,9 @@ class EmbeddedDocumentField(BaseField):
|
|||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if value is not None and not isinstance(value, self.document_type):
|
if value is not None and not isinstance(value, self.document_type):
|
||||||
|
# Short circuit for special operators, returning them as is
|
||||||
|
if isinstance(value, dict) and all(k.startswith("$") for k in value.keys()):
|
||||||
|
return value
|
||||||
try:
|
try:
|
||||||
value = self.document_type._from_son(value)
|
value = self.document_type._from_son(value)
|
||||||
except ValueError:
|
except ValueError:
|
||||||
@ -786,7 +783,7 @@ class EmbeddedDocumentField(BaseField):
|
|||||||
"Querying the embedded document '%s' failed, due to an invalid query value"
|
"Querying the embedded document '%s' failed, due to an invalid query value"
|
||||||
% (self.document_type._class_name,)
|
% (self.document_type._class_name,)
|
||||||
)
|
)
|
||||||
super(EmbeddedDocumentField, self).prepare_query_value(op, value)
|
super().prepare_query_value(op, value)
|
||||||
return self.to_mongo(value)
|
return self.to_mongo(value)
|
||||||
|
|
||||||
|
|
||||||
@ -802,9 +799,7 @@ class GenericEmbeddedDocumentField(BaseField):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
return super(GenericEmbeddedDocumentField, self).prepare_query_value(
|
return super().prepare_query_value(op, self.to_mongo(value))
|
||||||
op, self.to_mongo(value)
|
|
||||||
)
|
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
if isinstance(value, dict):
|
if isinstance(value, dict):
|
||||||
@ -855,7 +850,7 @@ class DynamicField(BaseField):
|
|||||||
"""Convert a Python type to a MongoDB compatible type.
|
"""Convert a Python type to a MongoDB compatible type.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if isinstance(value, six.string_types):
|
if isinstance(value, str):
|
||||||
return value
|
return value
|
||||||
|
|
||||||
if hasattr(value, "to_mongo"):
|
if hasattr(value, "to_mongo"):
|
||||||
@ -877,12 +872,12 @@ class DynamicField(BaseField):
|
|||||||
value = {k: v for k, v in enumerate(value)}
|
value = {k: v for k, v in enumerate(value)}
|
||||||
|
|
||||||
data = {}
|
data = {}
|
||||||
for k, v in iteritems(value):
|
for k, v in value.items():
|
||||||
data[k] = self.to_mongo(v, use_db_field, fields)
|
data[k] = self.to_mongo(v, use_db_field, fields)
|
||||||
|
|
||||||
value = data
|
value = data
|
||||||
if is_list: # Convert back to a list
|
if is_list: # Convert back to a list
|
||||||
value = [v for k, v in sorted(iteritems(data), key=itemgetter(0))]
|
value = [v for k, v in sorted(data.items(), key=itemgetter(0))]
|
||||||
return value
|
return value
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
@ -892,15 +887,15 @@ class DynamicField(BaseField):
|
|||||||
value = doc_cls._get_db().dereference(value["_ref"])
|
value = doc_cls._get_db().dereference(value["_ref"])
|
||||||
return doc_cls._from_son(value)
|
return doc_cls._from_son(value)
|
||||||
|
|
||||||
return super(DynamicField, self).to_python(value)
|
return super().to_python(value)
|
||||||
|
|
||||||
def lookup_member(self, member_name):
|
def lookup_member(self, member_name):
|
||||||
return member_name
|
return member_name
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if isinstance(value, six.string_types):
|
if isinstance(value, str):
|
||||||
return StringField().prepare_query_value(op, value)
|
return StringField().prepare_query_value(op, value)
|
||||||
return super(DynamicField, self).prepare_query_value(op, self.to_mongo(value))
|
return super().prepare_query_value(op, self.to_mongo(value))
|
||||||
|
|
||||||
def validate(self, value, clean=True):
|
def validate(self, value, clean=True):
|
||||||
if hasattr(value, "validate"):
|
if hasattr(value, "validate"):
|
||||||
@ -921,7 +916,7 @@ class ListField(ComplexBaseField):
|
|||||||
self.field = field
|
self.field = field
|
||||||
self.max_length = max_length
|
self.max_length = max_length
|
||||||
kwargs.setdefault("default", lambda: [])
|
kwargs.setdefault("default", lambda: [])
|
||||||
super(ListField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
if instance is None:
|
if instance is None:
|
||||||
@ -935,7 +930,7 @@ class ListField(ComplexBaseField):
|
|||||||
and value
|
and value
|
||||||
):
|
):
|
||||||
instance._data[self.name] = [self.field.build_lazyref(x) for x in value]
|
instance._data[self.name] = [self.field.build_lazyref(x) for x in value]
|
||||||
return super(ListField, self).__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
"""Make sure that a list of valid fields is being used."""
|
"""Make sure that a list of valid fields is being used."""
|
||||||
@ -949,7 +944,7 @@ class ListField(ComplexBaseField):
|
|||||||
if self.max_length is not None and len(value) > self.max_length:
|
if self.max_length is not None and len(value) > self.max_length:
|
||||||
self.error("List is too long")
|
self.error("List is too long")
|
||||||
|
|
||||||
super(ListField, self).validate(value)
|
super().validate(value)
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
# Validate that the `set` operator doesn't contain more items than `max_length`.
|
# Validate that the `set` operator doesn't contain more items than `max_length`.
|
||||||
@ -963,14 +958,14 @@ class ListField(ComplexBaseField):
|
|||||||
if (
|
if (
|
||||||
op in ("set", "unset", None)
|
op in ("set", "unset", None)
|
||||||
and hasattr(value, "__iter__")
|
and hasattr(value, "__iter__")
|
||||||
and not isinstance(value, six.string_types)
|
and not isinstance(value, str)
|
||||||
and not isinstance(value, BaseDocument)
|
and not isinstance(value, BaseDocument)
|
||||||
):
|
):
|
||||||
return [self.field.prepare_query_value(op, v) for v in value]
|
return [self.field.prepare_query_value(op, v) for v in value]
|
||||||
|
|
||||||
return self.field.prepare_query_value(op, value)
|
return self.field.prepare_query_value(op, value)
|
||||||
|
|
||||||
return super(ListField, self).prepare_query_value(op, value)
|
return super().prepare_query_value(op, value)
|
||||||
|
|
||||||
|
|
||||||
class EmbeddedDocumentListField(ListField):
|
class EmbeddedDocumentListField(ListField):
|
||||||
@ -991,9 +986,7 @@ class EmbeddedDocumentListField(ListField):
|
|||||||
:param kwargs: Keyword arguments passed directly into the parent
|
:param kwargs: Keyword arguments passed directly into the parent
|
||||||
:class:`~mongoengine.ListField`.
|
:class:`~mongoengine.ListField`.
|
||||||
"""
|
"""
|
||||||
super(EmbeddedDocumentListField, self).__init__(
|
super().__init__(field=EmbeddedDocumentField(document_type), **kwargs)
|
||||||
field=EmbeddedDocumentField(document_type), **kwargs
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class SortedListField(ListField):
|
class SortedListField(ListField):
|
||||||
@ -1019,10 +1012,10 @@ class SortedListField(ListField):
|
|||||||
self._ordering = kwargs.pop("ordering")
|
self._ordering = kwargs.pop("ordering")
|
||||||
if "reverse" in kwargs.keys():
|
if "reverse" in kwargs.keys():
|
||||||
self._order_reverse = kwargs.pop("reverse")
|
self._order_reverse = kwargs.pop("reverse")
|
||||||
super(SortedListField, self).__init__(field, **kwargs)
|
super().__init__(field, **kwargs)
|
||||||
|
|
||||||
def to_mongo(self, value, use_db_field=True, fields=None):
|
def to_mongo(self, value, use_db_field=True, fields=None):
|
||||||
value = super(SortedListField, self).to_mongo(value, use_db_field, fields)
|
value = super().to_mongo(value, use_db_field, fields)
|
||||||
if self._ordering is not None:
|
if self._ordering is not None:
|
||||||
return sorted(
|
return sorted(
|
||||||
value, key=itemgetter(self._ordering), reverse=self._order_reverse
|
value, key=itemgetter(self._ordering), reverse=self._order_reverse
|
||||||
@ -1035,9 +1028,7 @@ def key_not_string(d):
|
|||||||
dictionary is not a string.
|
dictionary is not a string.
|
||||||
"""
|
"""
|
||||||
for k, v in d.items():
|
for k, v in d.items():
|
||||||
if not isinstance(k, six.string_types) or (
|
if not isinstance(k, str) or (isinstance(v, dict) and key_not_string(v)):
|
||||||
isinstance(v, dict) and key_not_string(v)
|
|
||||||
):
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
@ -1077,7 +1068,7 @@ class DictField(ComplexBaseField):
|
|||||||
self._auto_dereference = False
|
self._auto_dereference = False
|
||||||
|
|
||||||
kwargs.setdefault("default", lambda: {})
|
kwargs.setdefault("default", lambda: {})
|
||||||
super(DictField, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
"""Make sure that a list of valid fields is being used."""
|
"""Make sure that a list of valid fields is being used."""
|
||||||
@ -1088,18 +1079,16 @@ class DictField(ComplexBaseField):
|
|||||||
msg = "Invalid dictionary key - documents must have only string keys"
|
msg = "Invalid dictionary key - documents must have only string keys"
|
||||||
self.error(msg)
|
self.error(msg)
|
||||||
|
|
||||||
curr_mongo_ver = get_mongodb_version()
|
# Following condition applies to MongoDB >= 3.6
|
||||||
|
# older Mongo has stricter constraints but
|
||||||
if curr_mongo_ver < MONGODB_36 and key_has_dot_or_dollar(value):
|
# it will be rejected upon insertion anyway
|
||||||
self.error(
|
# Having a validation that depends on the MongoDB version
|
||||||
'Invalid dictionary key name - keys may not contain "."'
|
# is not straightforward as the field isn't aware of the connected Mongo
|
||||||
' or startswith "$" characters'
|
if key_starts_with_dollar(value):
|
||||||
)
|
|
||||||
elif curr_mongo_ver >= MONGODB_36 and key_starts_with_dollar(value):
|
|
||||||
self.error(
|
self.error(
|
||||||
'Invalid dictionary key name - keys may not startswith "$" characters'
|
'Invalid dictionary key name - keys may not startswith "$" characters'
|
||||||
)
|
)
|
||||||
super(DictField, self).validate(value)
|
super().validate(value)
|
||||||
|
|
||||||
def lookup_member(self, member_name):
|
def lookup_member(self, member_name):
|
||||||
return DictField(db_field=member_name)
|
return DictField(db_field=member_name)
|
||||||
@ -1116,7 +1105,7 @@ class DictField(ComplexBaseField):
|
|||||||
"iexact",
|
"iexact",
|
||||||
]
|
]
|
||||||
|
|
||||||
if op in match_operators and isinstance(value, six.string_types):
|
if op in match_operators and isinstance(value, str):
|
||||||
return StringField().prepare_query_value(op, value)
|
return StringField().prepare_query_value(op, value)
|
||||||
|
|
||||||
if hasattr(
|
if hasattr(
|
||||||
@ -1128,7 +1117,7 @@ class DictField(ComplexBaseField):
|
|||||||
}
|
}
|
||||||
return self.field.prepare_query_value(op, value)
|
return self.field.prepare_query_value(op, value)
|
||||||
|
|
||||||
return super(DictField, self).prepare_query_value(op, value)
|
return super().prepare_query_value(op, value)
|
||||||
|
|
||||||
|
|
||||||
class MapField(DictField):
|
class MapField(DictField):
|
||||||
@ -1143,7 +1132,7 @@ class MapField(DictField):
|
|||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
if not isinstance(field, BaseField):
|
if not isinstance(field, BaseField):
|
||||||
self.error("Argument to MapField constructor must be a valid field")
|
self.error("Argument to MapField constructor must be a valid field")
|
||||||
super(MapField, self).__init__(field=field, *args, **kwargs)
|
super().__init__(field=field, *args, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
class ReferenceField(BaseField):
|
class ReferenceField(BaseField):
|
||||||
@ -1203,7 +1192,7 @@ class ReferenceField(BaseField):
|
|||||||
:class:`~pymongo.dbref.DBRef`, regardless of the value of `dbref`.
|
:class:`~pymongo.dbref.DBRef`, regardless of the value of `dbref`.
|
||||||
"""
|
"""
|
||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
if not isinstance(document_type, six.string_types) and not issubclass(
|
if not isinstance(document_type, str) and not issubclass(
|
||||||
document_type, Document
|
document_type, Document
|
||||||
):
|
):
|
||||||
self.error(
|
self.error(
|
||||||
@ -1214,11 +1203,11 @@ class ReferenceField(BaseField):
|
|||||||
self.dbref = dbref
|
self.dbref = dbref
|
||||||
self.document_type_obj = document_type
|
self.document_type_obj = document_type
|
||||||
self.reverse_delete_rule = reverse_delete_rule
|
self.reverse_delete_rule = reverse_delete_rule
|
||||||
super(ReferenceField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def document_type(self):
|
def document_type(self):
|
||||||
if isinstance(self.document_type_obj, six.string_types):
|
if isinstance(self.document_type_obj, str):
|
||||||
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
||||||
self.document_type_obj = self.owner_document
|
self.document_type_obj = self.owner_document
|
||||||
else:
|
else:
|
||||||
@ -1247,7 +1236,7 @@ class ReferenceField(BaseField):
|
|||||||
else:
|
else:
|
||||||
instance._data[self.name] = cls._from_son(dereferenced)
|
instance._data[self.name] = cls._from_son(dereferenced)
|
||||||
|
|
||||||
return super(ReferenceField, self).__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
def to_mongo(self, document):
|
def to_mongo(self, document):
|
||||||
if isinstance(document, DBRef):
|
if isinstance(document, DBRef):
|
||||||
@ -1298,7 +1287,7 @@ class ReferenceField(BaseField):
|
|||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if value is None:
|
if value is None:
|
||||||
return None
|
return None
|
||||||
super(ReferenceField, self).prepare_query_value(op, value)
|
super().prepare_query_value(op, value)
|
||||||
return self.to_mongo(value)
|
return self.to_mongo(value)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
@ -1334,7 +1323,7 @@ class CachedReferenceField(BaseField):
|
|||||||
fields = []
|
fields = []
|
||||||
|
|
||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
if not isinstance(document_type, six.string_types) and not issubclass(
|
if not isinstance(document_type, str) and not issubclass(
|
||||||
document_type, Document
|
document_type, Document
|
||||||
):
|
):
|
||||||
self.error(
|
self.error(
|
||||||
@ -1345,7 +1334,7 @@ class CachedReferenceField(BaseField):
|
|||||||
self.auto_sync = auto_sync
|
self.auto_sync = auto_sync
|
||||||
self.document_type_obj = document_type
|
self.document_type_obj = document_type
|
||||||
self.fields = fields
|
self.fields = fields
|
||||||
super(CachedReferenceField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def start_listener(self):
|
def start_listener(self):
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
@ -1357,7 +1346,7 @@ class CachedReferenceField(BaseField):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
update_kwargs = {
|
update_kwargs = {
|
||||||
"set__%s__%s" % (self.name, key): val
|
"set__{}__{}".format(self.name, key): val
|
||||||
for key, val in document._delta()[0].items()
|
for key, val in document._delta()[0].items()
|
||||||
if key in self.fields
|
if key in self.fields
|
||||||
}
|
}
|
||||||
@ -1379,7 +1368,7 @@ class CachedReferenceField(BaseField):
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def document_type(self):
|
def document_type(self):
|
||||||
if isinstance(self.document_type_obj, six.string_types):
|
if isinstance(self.document_type_obj, str):
|
||||||
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
||||||
self.document_type_obj = self.owner_document
|
self.document_type_obj = self.owner_document
|
||||||
else:
|
else:
|
||||||
@ -1403,7 +1392,7 @@ class CachedReferenceField(BaseField):
|
|||||||
else:
|
else:
|
||||||
instance._data[self.name] = self.document_type._from_son(dereferenced)
|
instance._data[self.name] = self.document_type._from_son(dereferenced)
|
||||||
|
|
||||||
return super(CachedReferenceField, self).__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
def to_mongo(self, document, use_db_field=True, fields=None):
|
def to_mongo(self, document, use_db_field=True, fields=None):
|
||||||
id_field_name = self.document_type._meta["id_field"]
|
id_field_name = self.document_type._meta["id_field"]
|
||||||
@ -1502,12 +1491,12 @@ class GenericReferenceField(BaseField):
|
|||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
choices = kwargs.pop("choices", None)
|
choices = kwargs.pop("choices", None)
|
||||||
super(GenericReferenceField, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
self.choices = []
|
self.choices = []
|
||||||
# Keep the choices as a list of allowed Document class names
|
# Keep the choices as a list of allowed Document class names
|
||||||
if choices:
|
if choices:
|
||||||
for choice in choices:
|
for choice in choices:
|
||||||
if isinstance(choice, six.string_types):
|
if isinstance(choice, str):
|
||||||
self.choices.append(choice)
|
self.choices.append(choice)
|
||||||
elif isinstance(choice, type) and issubclass(choice, Document):
|
elif isinstance(choice, type) and issubclass(choice, Document):
|
||||||
self.choices.append(choice._class_name)
|
self.choices.append(choice._class_name)
|
||||||
@ -1516,7 +1505,7 @@ class GenericReferenceField(BaseField):
|
|||||||
# method.
|
# method.
|
||||||
self.error(
|
self.error(
|
||||||
"Invalid choices provided: must be a list of"
|
"Invalid choices provided: must be a list of"
|
||||||
"Document subclasses and/or six.string_typess"
|
"Document subclasses and/or str"
|
||||||
)
|
)
|
||||||
|
|
||||||
def _validate_choices(self, value):
|
def _validate_choices(self, value):
|
||||||
@ -1526,7 +1515,7 @@ class GenericReferenceField(BaseField):
|
|||||||
value = value.get("_cls")
|
value = value.get("_cls")
|
||||||
elif isinstance(value, Document):
|
elif isinstance(value, Document):
|
||||||
value = value._class_name
|
value = value._class_name
|
||||||
super(GenericReferenceField, self)._validate_choices(value)
|
super()._validate_choices(value)
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
if instance is None:
|
if instance is None:
|
||||||
@ -1542,7 +1531,7 @@ class GenericReferenceField(BaseField):
|
|||||||
else:
|
else:
|
||||||
instance._data[self.name] = dereferenced
|
instance._data[self.name] = dereferenced
|
||||||
|
|
||||||
return super(GenericReferenceField, self).__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if not isinstance(value, (Document, DBRef, dict, SON)):
|
if not isinstance(value, (Document, DBRef, dict, SON)):
|
||||||
@ -1606,22 +1595,22 @@ class BinaryField(BaseField):
|
|||||||
|
|
||||||
def __init__(self, max_bytes=None, **kwargs):
|
def __init__(self, max_bytes=None, **kwargs):
|
||||||
self.max_bytes = max_bytes
|
self.max_bytes = max_bytes
|
||||||
super(BinaryField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def __set__(self, instance, value):
|
def __set__(self, instance, value):
|
||||||
"""Handle bytearrays in python 3.1"""
|
"""Handle bytearrays in python 3.1"""
|
||||||
if six.PY3 and isinstance(value, bytearray):
|
if isinstance(value, bytearray):
|
||||||
value = six.binary_type(value)
|
value = bytes(value)
|
||||||
return super(BinaryField, self).__set__(instance, value)
|
return super().__set__(instance, value)
|
||||||
|
|
||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
return Binary(value)
|
return Binary(value)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if not isinstance(value, (six.binary_type, Binary)):
|
if not isinstance(value, (bytes, Binary)):
|
||||||
self.error(
|
self.error(
|
||||||
"BinaryField only accepts instances of "
|
"BinaryField only accepts instances of "
|
||||||
"(%s, %s, Binary)" % (six.binary_type.__name__, Binary.__name__)
|
"(%s, %s, Binary)" % (bytes.__name__, Binary.__name__)
|
||||||
)
|
)
|
||||||
|
|
||||||
if self.max_bytes is not None and len(value) > self.max_bytes:
|
if self.max_bytes is not None and len(value) > self.max_bytes:
|
||||||
@ -1630,14 +1619,14 @@ class BinaryField(BaseField):
|
|||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if value is None:
|
if value is None:
|
||||||
return value
|
return value
|
||||||
return super(BinaryField, self).prepare_query_value(op, self.to_mongo(value))
|
return super().prepare_query_value(op, self.to_mongo(value))
|
||||||
|
|
||||||
|
|
||||||
class GridFSError(Exception):
|
class GridFSError(Exception):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
class GridFSProxy(object):
|
class GridFSProxy:
|
||||||
"""Proxy object to handle writing and reading of files to and from GridFS
|
"""Proxy object to handle writing and reading of files to and from GridFS
|
||||||
|
|
||||||
.. versionadded:: 0.4
|
.. versionadded:: 0.4
|
||||||
@ -1687,8 +1676,6 @@ class GridFSProxy(object):
|
|||||||
def __bool__(self):
|
def __bool__(self):
|
||||||
return bool(self.grid_id)
|
return bool(self.grid_id)
|
||||||
|
|
||||||
__nonzero__ = __bool__ # For Py2 support
|
|
||||||
|
|
||||||
def __getstate__(self):
|
def __getstate__(self):
|
||||||
self_dict = self.__dict__
|
self_dict = self.__dict__
|
||||||
self_dict["_fs"] = None
|
self_dict["_fs"] = None
|
||||||
@ -1703,12 +1690,12 @@ class GridFSProxy(object):
|
|||||||
return self.__copy__()
|
return self.__copy__()
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "<%s: %s>" % (self.__class__.__name__, self.grid_id)
|
return "<{}: {}>".format(self.__class__.__name__, self.grid_id)
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
gridout = self.get()
|
gridout = self.get()
|
||||||
filename = getattr(gridout, "filename") if gridout else "<no file>"
|
filename = getattr(gridout, "filename") if gridout else "<no file>"
|
||||||
return "<%s: %s (%s)>" % (self.__class__.__name__, filename, self.grid_id)
|
return "<{}: {} ({})>".format(self.__class__.__name__, filename, self.grid_id)
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
if isinstance(other, GridFSProxy):
|
if isinstance(other, GridFSProxy):
|
||||||
@ -1819,7 +1806,7 @@ class FileField(BaseField):
|
|||||||
def __init__(
|
def __init__(
|
||||||
self, db_alias=DEFAULT_CONNECTION_NAME, collection_name="fs", **kwargs
|
self, db_alias=DEFAULT_CONNECTION_NAME, collection_name="fs", **kwargs
|
||||||
):
|
):
|
||||||
super(FileField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
self.collection_name = collection_name
|
self.collection_name = collection_name
|
||||||
self.db_alias = db_alias
|
self.db_alias = db_alias
|
||||||
|
|
||||||
@ -1842,7 +1829,7 @@ class FileField(BaseField):
|
|||||||
key = self.name
|
key = self.name
|
||||||
if (
|
if (
|
||||||
hasattr(value, "read") and not isinstance(value, GridFSProxy)
|
hasattr(value, "read") and not isinstance(value, GridFSProxy)
|
||||||
) or isinstance(value, (six.binary_type, six.string_types)):
|
) or isinstance(value, (bytes, str)):
|
||||||
# using "FileField() = file/string" notation
|
# using "FileField() = file/string" notation
|
||||||
grid_file = instance._data.get(self.name)
|
grid_file = instance._data.get(self.name)
|
||||||
# If a file already exists, delete it
|
# If a file already exists, delete it
|
||||||
@ -1960,11 +1947,11 @@ class ImageGridFsProxy(GridFSProxy):
|
|||||||
|
|
||||||
w, h = img.size
|
w, h = img.size
|
||||||
|
|
||||||
io = StringIO()
|
io = BytesIO()
|
||||||
img.save(io, img_format, progressive=progressive)
|
img.save(io, img_format, progressive=progressive)
|
||||||
io.seek(0)
|
io.seek(0)
|
||||||
|
|
||||||
return super(ImageGridFsProxy, self).put(
|
return super().put(
|
||||||
io, width=w, height=h, format=img_format, thumbnail_id=thumb_id, **kwargs
|
io, width=w, height=h, format=img_format, thumbnail_id=thumb_id, **kwargs
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -1974,12 +1961,12 @@ class ImageGridFsProxy(GridFSProxy):
|
|||||||
if out and out.thumbnail_id:
|
if out and out.thumbnail_id:
|
||||||
self.fs.delete(out.thumbnail_id)
|
self.fs.delete(out.thumbnail_id)
|
||||||
|
|
||||||
return super(ImageGridFsProxy, self).delete()
|
return super().delete()
|
||||||
|
|
||||||
def _put_thumbnail(self, thumbnail, format, progressive, **kwargs):
|
def _put_thumbnail(self, thumbnail, format, progressive, **kwargs):
|
||||||
w, h = thumbnail.size
|
w, h = thumbnail.size
|
||||||
|
|
||||||
io = StringIO()
|
io = BytesIO()
|
||||||
thumbnail.save(io, format, progressive=progressive)
|
thumbnail.save(io, format, progressive=progressive)
|
||||||
io.seek(0)
|
io.seek(0)
|
||||||
|
|
||||||
@ -2049,16 +2036,11 @@ class ImageField(FileField):
|
|||||||
for att_name, att in extra_args.items():
|
for att_name, att in extra_args.items():
|
||||||
value = None
|
value = None
|
||||||
if isinstance(att, (tuple, list)):
|
if isinstance(att, (tuple, list)):
|
||||||
if six.PY3:
|
value = dict(itertools.zip_longest(params_size, att, fillvalue=None))
|
||||||
value = dict(
|
|
||||||
itertools.zip_longest(params_size, att, fillvalue=None)
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
value = dict(map(None, params_size, att))
|
|
||||||
|
|
||||||
setattr(self, att_name, value)
|
setattr(self, att_name, value)
|
||||||
|
|
||||||
super(ImageField, self).__init__(collection_name=collection_name, **kwargs)
|
super().__init__(collection_name=collection_name, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
class SequenceField(BaseField):
|
class SequenceField(BaseField):
|
||||||
@ -2110,14 +2092,14 @@ class SequenceField(BaseField):
|
|||||||
self.value_decorator = (
|
self.value_decorator = (
|
||||||
value_decorator if callable(value_decorator) else self.VALUE_DECORATOR
|
value_decorator if callable(value_decorator) else self.VALUE_DECORATOR
|
||||||
)
|
)
|
||||||
super(SequenceField, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def generate(self):
|
def generate(self):
|
||||||
"""
|
"""
|
||||||
Generate and Increment the counter
|
Generate and Increment the counter
|
||||||
"""
|
"""
|
||||||
sequence_name = self.get_sequence_name()
|
sequence_name = self.get_sequence_name()
|
||||||
sequence_id = "%s.%s" % (sequence_name, self.name)
|
sequence_id = "{}.{}".format(sequence_name, self.name)
|
||||||
collection = get_db(alias=self.db_alias)[self.collection_name]
|
collection = get_db(alias=self.db_alias)[self.collection_name]
|
||||||
|
|
||||||
counter = collection.find_one_and_update(
|
counter = collection.find_one_and_update(
|
||||||
@ -2131,7 +2113,7 @@ class SequenceField(BaseField):
|
|||||||
def set_next_value(self, value):
|
def set_next_value(self, value):
|
||||||
"""Helper method to set the next sequence value"""
|
"""Helper method to set the next sequence value"""
|
||||||
sequence_name = self.get_sequence_name()
|
sequence_name = self.get_sequence_name()
|
||||||
sequence_id = "%s.%s" % (sequence_name, self.name)
|
sequence_id = "{}.{}".format(sequence_name, self.name)
|
||||||
collection = get_db(alias=self.db_alias)[self.collection_name]
|
collection = get_db(alias=self.db_alias)[self.collection_name]
|
||||||
counter = collection.find_one_and_update(
|
counter = collection.find_one_and_update(
|
||||||
filter={"_id": sequence_id},
|
filter={"_id": sequence_id},
|
||||||
@ -2148,7 +2130,7 @@ class SequenceField(BaseField):
|
|||||||
as it is only fixed on set.
|
as it is only fixed on set.
|
||||||
"""
|
"""
|
||||||
sequence_name = self.get_sequence_name()
|
sequence_name = self.get_sequence_name()
|
||||||
sequence_id = "%s.%s" % (sequence_name, self.name)
|
sequence_id = "{}.{}".format(sequence_name, self.name)
|
||||||
collection = get_db(alias=self.db_alias)[self.collection_name]
|
collection = get_db(alias=self.db_alias)[self.collection_name]
|
||||||
data = collection.find_one({"_id": sequence_id})
|
data = collection.find_one({"_id": sequence_id})
|
||||||
|
|
||||||
@ -2171,7 +2153,7 @@ class SequenceField(BaseField):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
value = super(SequenceField, self).__get__(instance, owner)
|
value = super().__get__(instance, owner)
|
||||||
if value is None and instance._initialised:
|
if value is None and instance._initialised:
|
||||||
value = self.generate()
|
value = self.generate()
|
||||||
instance._data[self.name] = value
|
instance._data[self.name] = value
|
||||||
@ -2184,7 +2166,7 @@ class SequenceField(BaseField):
|
|||||||
if value is None and instance._initialised:
|
if value is None and instance._initialised:
|
||||||
value = self.generate()
|
value = self.generate()
|
||||||
|
|
||||||
return super(SequenceField, self).__set__(instance, value)
|
return super().__set__(instance, value)
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
"""
|
"""
|
||||||
@ -2218,14 +2200,14 @@ class UUIDField(BaseField):
|
|||||||
.. versionchanged:: 0.6.19
|
.. versionchanged:: 0.6.19
|
||||||
"""
|
"""
|
||||||
self._binary = binary
|
self._binary = binary
|
||||||
super(UUIDField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
if not self._binary:
|
if not self._binary:
|
||||||
original_value = value
|
original_value = value
|
||||||
try:
|
try:
|
||||||
if not isinstance(value, six.string_types):
|
if not isinstance(value, str):
|
||||||
value = six.text_type(value)
|
value = str(value)
|
||||||
return uuid.UUID(value)
|
return uuid.UUID(value)
|
||||||
except (ValueError, TypeError, AttributeError):
|
except (ValueError, TypeError, AttributeError):
|
||||||
return original_value
|
return original_value
|
||||||
@ -2233,8 +2215,8 @@ class UUIDField(BaseField):
|
|||||||
|
|
||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
if not self._binary:
|
if not self._binary:
|
||||||
return six.text_type(value)
|
return str(value)
|
||||||
elif isinstance(value, six.string_types):
|
elif isinstance(value, str):
|
||||||
return uuid.UUID(value)
|
return uuid.UUID(value)
|
||||||
return value
|
return value
|
||||||
|
|
||||||
@ -2245,7 +2227,7 @@ class UUIDField(BaseField):
|
|||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if not isinstance(value, uuid.UUID):
|
if not isinstance(value, uuid.UUID):
|
||||||
if not isinstance(value, six.string_types):
|
if not isinstance(value, str):
|
||||||
value = str(value)
|
value = str(value)
|
||||||
try:
|
try:
|
||||||
uuid.UUID(value)
|
uuid.UUID(value)
|
||||||
@ -2444,7 +2426,7 @@ class LazyReferenceField(BaseField):
|
|||||||
document. Note this only work getting field (not setting or deleting).
|
document. Note this only work getting field (not setting or deleting).
|
||||||
"""
|
"""
|
||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
if not isinstance(document_type, six.string_types) and not issubclass(
|
if not isinstance(document_type, str) and not issubclass(
|
||||||
document_type, Document
|
document_type, Document
|
||||||
):
|
):
|
||||||
self.error(
|
self.error(
|
||||||
@ -2456,11 +2438,11 @@ class LazyReferenceField(BaseField):
|
|||||||
self.passthrough = passthrough
|
self.passthrough = passthrough
|
||||||
self.document_type_obj = document_type
|
self.document_type_obj = document_type
|
||||||
self.reverse_delete_rule = reverse_delete_rule
|
self.reverse_delete_rule = reverse_delete_rule
|
||||||
super(LazyReferenceField, self).__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def document_type(self):
|
def document_type(self):
|
||||||
if isinstance(self.document_type_obj, six.string_types):
|
if isinstance(self.document_type_obj, str):
|
||||||
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
if self.document_type_obj == RECURSIVE_REFERENCE_CONSTANT:
|
||||||
self.document_type_obj = self.owner_document
|
self.document_type_obj = self.owner_document
|
||||||
else:
|
else:
|
||||||
@ -2499,7 +2481,7 @@ class LazyReferenceField(BaseField):
|
|||||||
if value:
|
if value:
|
||||||
instance._data[self.name] = value
|
instance._data[self.name] = value
|
||||||
|
|
||||||
return super(LazyReferenceField, self).__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
if isinstance(value, LazyReference):
|
if isinstance(value, LazyReference):
|
||||||
@ -2563,7 +2545,7 @@ class LazyReferenceField(BaseField):
|
|||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if value is None:
|
if value is None:
|
||||||
return None
|
return None
|
||||||
super(LazyReferenceField, self).prepare_query_value(op, value)
|
super().prepare_query_value(op, value)
|
||||||
return self.to_mongo(value)
|
return self.to_mongo(value)
|
||||||
|
|
||||||
def lookup_member(self, member_name):
|
def lookup_member(self, member_name):
|
||||||
@ -2590,12 +2572,12 @@ class GenericLazyReferenceField(GenericReferenceField):
|
|||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
self.passthrough = kwargs.pop("passthrough", False)
|
self.passthrough = kwargs.pop("passthrough", False)
|
||||||
super(GenericLazyReferenceField, self).__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def _validate_choices(self, value):
|
def _validate_choices(self, value):
|
||||||
if isinstance(value, LazyReference):
|
if isinstance(value, LazyReference):
|
||||||
value = value.document_type._class_name
|
value = value.document_type._class_name
|
||||||
super(GenericLazyReferenceField, self)._validate_choices(value)
|
super()._validate_choices(value)
|
||||||
|
|
||||||
def build_lazyref(self, value):
|
def build_lazyref(self, value):
|
||||||
if isinstance(value, LazyReference):
|
if isinstance(value, LazyReference):
|
||||||
@ -2624,7 +2606,7 @@ class GenericLazyReferenceField(GenericReferenceField):
|
|||||||
if value:
|
if value:
|
||||||
instance._data[self.name] = value
|
instance._data[self.name] = value
|
||||||
|
|
||||||
return super(GenericLazyReferenceField, self).__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
if isinstance(value, LazyReference) and value.pk is None:
|
if isinstance(value, LazyReference) and value.pk is None:
|
||||||
@ -2632,7 +2614,7 @@ class GenericLazyReferenceField(GenericReferenceField):
|
|||||||
"You can only reference documents once they have been"
|
"You can only reference documents once they have been"
|
||||||
" saved to the database"
|
" saved to the database"
|
||||||
)
|
)
|
||||||
return super(GenericLazyReferenceField, self).validate(value)
|
return super().validate(value)
|
||||||
|
|
||||||
def to_mongo(self, document):
|
def to_mongo(self, document):
|
||||||
if document is None:
|
if document is None:
|
||||||
@ -2651,4 +2633,4 @@ class GenericLazyReferenceField(GenericReferenceField):
|
|||||||
)
|
)
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
return super(GenericLazyReferenceField, self).to_mongo(document)
|
return super().to_mongo(document)
|
||||||
|
@ -11,7 +11,7 @@ MONGODB_36 = (3, 6)
|
|||||||
|
|
||||||
|
|
||||||
def get_mongodb_version():
|
def get_mongodb_version():
|
||||||
"""Return the version of the connected mongoDB (first 2 digits)
|
"""Return the version of the default connected mongoDB (first 2 digits)
|
||||||
|
|
||||||
:return: tuple(int, int)
|
:return: tuple(int, int)
|
||||||
"""
|
"""
|
||||||
|
@ -1,23 +0,0 @@
|
|||||||
"""
|
|
||||||
Helper functions, constants, and types to aid with Python v2.7 - v3.x support
|
|
||||||
"""
|
|
||||||
import six
|
|
||||||
|
|
||||||
# six.BytesIO resolves to StringIO.StringIO in Py2 and io.BytesIO in Py3.
|
|
||||||
StringIO = six.BytesIO
|
|
||||||
|
|
||||||
# Additionally for Py2, try to use the faster cStringIO, if available
|
|
||||||
if not six.PY3:
|
|
||||||
try:
|
|
||||||
import cStringIO
|
|
||||||
except ImportError:
|
|
||||||
pass
|
|
||||||
else:
|
|
||||||
StringIO = cStringIO.StringIO
|
|
||||||
|
|
||||||
|
|
||||||
if six.PY3:
|
|
||||||
from collections.abc import Hashable
|
|
||||||
else:
|
|
||||||
# raises DeprecationWarnings in Python >=3.7
|
|
||||||
from collections import Hashable
|
|
@ -1,24 +1,27 @@
|
|||||||
from __future__ import absolute_import
|
|
||||||
|
|
||||||
import copy
|
import copy
|
||||||
import itertools
|
import itertools
|
||||||
import re
|
import re
|
||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
|
from collections.abc import Mapping
|
||||||
|
|
||||||
from bson import SON, json_util
|
from bson import SON, json_util
|
||||||
from bson.code import Code
|
from bson.code import Code
|
||||||
import pymongo
|
import pymongo
|
||||||
import pymongo.errors
|
import pymongo.errors
|
||||||
from pymongo.collection import ReturnDocument
|
from pymongo.collection import ReturnDocument
|
||||||
from pymongo.common import validate_read_preference
|
from pymongo.common import validate_read_preference
|
||||||
import six
|
from pymongo.read_concern import ReadConcern
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
from mongoengine.base import get_document
|
from mongoengine.base import get_document
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.context_managers import set_write_concern, switch_db
|
from mongoengine.context_managers import (
|
||||||
|
set_read_write_concern,
|
||||||
|
set_write_concern,
|
||||||
|
switch_db,
|
||||||
|
)
|
||||||
from mongoengine.errors import (
|
from mongoengine.errors import (
|
||||||
BulkWriteError,
|
BulkWriteError,
|
||||||
InvalidQueryError,
|
InvalidQueryError,
|
||||||
@ -42,7 +45,7 @@ DENY = 3
|
|||||||
PULL = 4
|
PULL = 4
|
||||||
|
|
||||||
|
|
||||||
class BaseQuerySet(object):
|
class BaseQuerySet:
|
||||||
"""A set of results returned from a query. Wraps a MongoDB cursor,
|
"""A set of results returned from a query. Wraps a MongoDB cursor,
|
||||||
providing :class:`~mongoengine.Document` objects as the results.
|
providing :class:`~mongoengine.Document` objects as the results.
|
||||||
"""
|
"""
|
||||||
@ -61,8 +64,8 @@ class BaseQuerySet(object):
|
|||||||
self._ordering = None
|
self._ordering = None
|
||||||
self._snapshot = False
|
self._snapshot = False
|
||||||
self._timeout = True
|
self._timeout = True
|
||||||
self._slave_okay = False
|
|
||||||
self._read_preference = None
|
self._read_preference = None
|
||||||
|
self._read_concern = None
|
||||||
self._iter = False
|
self._iter = False
|
||||||
self._scalar = []
|
self._scalar = []
|
||||||
self._none = False
|
self._none = False
|
||||||
@ -81,6 +84,7 @@ class BaseQuerySet(object):
|
|||||||
self._cursor_obj = None
|
self._cursor_obj = None
|
||||||
self._limit = None
|
self._limit = None
|
||||||
self._skip = None
|
self._skip = None
|
||||||
|
|
||||||
self._hint = -1 # Using -1 as None is a valid value for hint
|
self._hint = -1 # Using -1 as None is a valid value for hint
|
||||||
self._collation = None
|
self._collation = None
|
||||||
self._batch_size = None
|
self._batch_size = None
|
||||||
@ -88,6 +92,13 @@ class BaseQuerySet(object):
|
|||||||
self._max_time_ms = None
|
self._max_time_ms = None
|
||||||
self._comment = None
|
self._comment = None
|
||||||
|
|
||||||
|
# Hack - As people expect cursor[5:5] to return
|
||||||
|
# an empty result set. It's hard to do that right, though, because the
|
||||||
|
# server uses limit(0) to mean 'no limit'. So we set _empty
|
||||||
|
# in that case and check for it when iterating. We also unset
|
||||||
|
# it anytime we change _limit. Inspired by how it is done in pymongo.Cursor
|
||||||
|
self._empty = False
|
||||||
|
|
||||||
def __call__(self, q_obj=None, **query):
|
def __call__(self, q_obj=None, **query):
|
||||||
"""Filter the selected documents by calling the
|
"""Filter the selected documents by calling the
|
||||||
:class:`~mongoengine.queryset.QuerySet` with a query.
|
:class:`~mongoengine.queryset.QuerySet` with a query.
|
||||||
@ -160,6 +171,7 @@ class BaseQuerySet(object):
|
|||||||
[<User: User object>, <User: User object>]
|
[<User: User object>, <User: User object>]
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
|
queryset._empty = False
|
||||||
|
|
||||||
# Handle a slice
|
# Handle a slice
|
||||||
if isinstance(key, slice):
|
if isinstance(key, slice):
|
||||||
@ -167,6 +179,8 @@ class BaseQuerySet(object):
|
|||||||
queryset._skip, queryset._limit = key.start, key.stop
|
queryset._skip, queryset._limit = key.start, key.stop
|
||||||
if key.start and key.stop:
|
if key.start and key.stop:
|
||||||
queryset._limit = key.stop - key.start
|
queryset._limit = key.stop - key.start
|
||||||
|
if queryset._limit == 0:
|
||||||
|
queryset._empty = True
|
||||||
|
|
||||||
# Allow further QuerySet modifications to be performed
|
# Allow further QuerySet modifications to be performed
|
||||||
return queryset
|
return queryset
|
||||||
@ -205,8 +219,6 @@ class BaseQuerySet(object):
|
|||||||
"""Avoid to open all records in an if stmt in Py3."""
|
"""Avoid to open all records in an if stmt in Py3."""
|
||||||
return self._has_data()
|
return self._has_data()
|
||||||
|
|
||||||
__nonzero__ = __bool__ # For Py2 support
|
|
||||||
|
|
||||||
# Core functions
|
# Core functions
|
||||||
|
|
||||||
def all(self):
|
def all(self):
|
||||||
@ -257,20 +269,21 @@ class BaseQuerySet(object):
|
|||||||
queryset = queryset.filter(*q_objs, **query)
|
queryset = queryset.filter(*q_objs, **query)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
result = six.next(queryset)
|
result = next(queryset)
|
||||||
except StopIteration:
|
except StopIteration:
|
||||||
msg = "%s matching query does not exist." % queryset._document._class_name
|
msg = "%s matching query does not exist." % queryset._document._class_name
|
||||||
raise queryset._document.DoesNotExist(msg)
|
raise queryset._document.DoesNotExist(msg)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
six.next(queryset)
|
# Check if there is another match
|
||||||
|
next(queryset)
|
||||||
except StopIteration:
|
except StopIteration:
|
||||||
return result
|
return result
|
||||||
|
|
||||||
# If we were able to retrieve the 2nd doc, rewind the cursor and
|
# If we were able to retrieve the 2nd doc, raise the MultipleObjectsReturned exception.
|
||||||
# raise the MultipleObjectsReturned exception.
|
raise queryset._document.MultipleObjectsReturned(
|
||||||
queryset.rewind()
|
"2 or more items returned, instead of 1"
|
||||||
message = u"%d items returned, instead of 1" % queryset.count()
|
)
|
||||||
raise queryset._document.MultipleObjectsReturned(message)
|
|
||||||
|
|
||||||
def create(self, **kwargs):
|
def create(self, **kwargs):
|
||||||
"""Create new object. Returns the saved object instance.
|
"""Create new object. Returns the saved object instance.
|
||||||
@ -354,20 +367,20 @@ class BaseQuerySet(object):
|
|||||||
)
|
)
|
||||||
except pymongo.errors.DuplicateKeyError as err:
|
except pymongo.errors.DuplicateKeyError as err:
|
||||||
message = "Could not save document (%s)"
|
message = "Could not save document (%s)"
|
||||||
raise NotUniqueError(message % six.text_type(err))
|
raise NotUniqueError(message % err)
|
||||||
except pymongo.errors.BulkWriteError as err:
|
except pymongo.errors.BulkWriteError as err:
|
||||||
# inserting documents that already have an _id field will
|
# inserting documents that already have an _id field will
|
||||||
# give huge performance debt or raise
|
# give huge performance debt or raise
|
||||||
message = u"Bulk write error: (%s)"
|
message = "Bulk write error: (%s)"
|
||||||
raise BulkWriteError(message % six.text_type(err.details))
|
raise BulkWriteError(message % err.details)
|
||||||
except pymongo.errors.OperationFailure as err:
|
except pymongo.errors.OperationFailure as err:
|
||||||
message = "Could not save document (%s)"
|
message = "Could not save document (%s)"
|
||||||
if re.match("^E1100[01] duplicate key", six.text_type(err)):
|
if re.match("^E1100[01] duplicate key", str(err)):
|
||||||
# E11000 - duplicate key error index
|
# E11000 - duplicate key error index
|
||||||
# E11001 - duplicate key on update
|
# E11001 - duplicate key on update
|
||||||
message = u"Tried to save duplicate unique keys (%s)"
|
message = "Tried to save duplicate unique keys (%s)"
|
||||||
raise NotUniqueError(message % six.text_type(err))
|
raise NotUniqueError(message % err)
|
||||||
raise OperationError(message % six.text_type(err))
|
raise OperationError(message % err)
|
||||||
|
|
||||||
# Apply inserted_ids to documents
|
# Apply inserted_ids to documents
|
||||||
for doc, doc_id in zip(docs, ids):
|
for doc, doc_id in zip(docs, ids):
|
||||||
@ -395,7 +408,12 @@ class BaseQuerySet(object):
|
|||||||
"""
|
"""
|
||||||
# mimic the fact that setting .limit(0) in pymongo sets no limit
|
# mimic the fact that setting .limit(0) in pymongo sets no limit
|
||||||
# https://docs.mongodb.com/manual/reference/method/cursor.limit/#zero-value
|
# https://docs.mongodb.com/manual/reference/method/cursor.limit/#zero-value
|
||||||
if self._limit == 0 and with_limit_and_skip is False or self._none:
|
if (
|
||||||
|
self._limit == 0
|
||||||
|
and with_limit_and_skip is False
|
||||||
|
or self._none
|
||||||
|
or self._empty
|
||||||
|
):
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
kwargs = (
|
kwargs = (
|
||||||
@ -513,7 +531,13 @@ class BaseQuerySet(object):
|
|||||||
return result.deleted_count
|
return result.deleted_count
|
||||||
|
|
||||||
def update(
|
def update(
|
||||||
self, upsert=False, multi=True, write_concern=None, full_result=False, **update
|
self,
|
||||||
|
upsert=False,
|
||||||
|
multi=True,
|
||||||
|
write_concern=None,
|
||||||
|
read_concern=None,
|
||||||
|
full_result=False,
|
||||||
|
**update
|
||||||
):
|
):
|
||||||
"""Perform an atomic update on the fields matched by the query.
|
"""Perform an atomic update on the fields matched by the query.
|
||||||
|
|
||||||
@ -525,6 +549,7 @@ class BaseQuerySet(object):
|
|||||||
``save(..., write_concern={w: 2, fsync: True}, ...)`` will
|
``save(..., write_concern={w: 2, fsync: True}, ...)`` will
|
||||||
wait until at least two servers have recorded the write and
|
wait until at least two servers have recorded the write and
|
||||||
will force an fsync on the primary server.
|
will force an fsync on the primary server.
|
||||||
|
:param read_concern: Override the read concern for the operation
|
||||||
:param full_result: Return the associated ``pymongo.UpdateResult`` rather than just the number
|
:param full_result: Return the associated ``pymongo.UpdateResult`` rather than just the number
|
||||||
updated items
|
updated items
|
||||||
:param update: Django-style update keyword arguments
|
:param update: Django-style update keyword arguments
|
||||||
@ -551,7 +576,9 @@ class BaseQuerySet(object):
|
|||||||
else:
|
else:
|
||||||
update["$set"] = {"_cls": queryset._document._class_name}
|
update["$set"] = {"_cls": queryset._document._class_name}
|
||||||
try:
|
try:
|
||||||
with set_write_concern(queryset._collection, write_concern) as collection:
|
with set_read_write_concern(
|
||||||
|
queryset._collection, write_concern, read_concern
|
||||||
|
) as collection:
|
||||||
update_func = collection.update_one
|
update_func = collection.update_one
|
||||||
if multi:
|
if multi:
|
||||||
update_func = collection.update_many
|
update_func = collection.update_many
|
||||||
@ -561,14 +588,14 @@ class BaseQuerySet(object):
|
|||||||
elif result.raw_result:
|
elif result.raw_result:
|
||||||
return result.raw_result["n"]
|
return result.raw_result["n"]
|
||||||
except pymongo.errors.DuplicateKeyError as err:
|
except pymongo.errors.DuplicateKeyError as err:
|
||||||
raise NotUniqueError(u"Update failed (%s)" % six.text_type(err))
|
raise NotUniqueError("Update failed (%s)" % err)
|
||||||
except pymongo.errors.OperationFailure as err:
|
except pymongo.errors.OperationFailure as err:
|
||||||
if six.text_type(err) == u"multi not coded yet":
|
if str(err) == "multi not coded yet":
|
||||||
message = u"update() method requires MongoDB 1.1.3+"
|
message = "update() method requires MongoDB 1.1.3+"
|
||||||
raise OperationError(message)
|
raise OperationError(message)
|
||||||
raise OperationError(u"Update failed (%s)" % six.text_type(err))
|
raise OperationError("Update failed (%s)" % err)
|
||||||
|
|
||||||
def upsert_one(self, write_concern=None, **update):
|
def upsert_one(self, write_concern=None, read_concern=None, **update):
|
||||||
"""Overwrite or add the first document matched by the query.
|
"""Overwrite or add the first document matched by the query.
|
||||||
|
|
||||||
:param write_concern: Extra keyword arguments are passed down which
|
:param write_concern: Extra keyword arguments are passed down which
|
||||||
@ -577,6 +604,7 @@ class BaseQuerySet(object):
|
|||||||
``save(..., write_concern={w: 2, fsync: True}, ...)`` will
|
``save(..., write_concern={w: 2, fsync: True}, ...)`` will
|
||||||
wait until at least two servers have recorded the write and
|
wait until at least two servers have recorded the write and
|
||||||
will force an fsync on the primary server.
|
will force an fsync on the primary server.
|
||||||
|
:param read_concern: Override the read concern for the operation
|
||||||
:param update: Django-style update keyword arguments
|
:param update: Django-style update keyword arguments
|
||||||
|
|
||||||
:returns the new or overwritten document
|
:returns the new or overwritten document
|
||||||
@ -588,6 +616,7 @@ class BaseQuerySet(object):
|
|||||||
multi=False,
|
multi=False,
|
||||||
upsert=True,
|
upsert=True,
|
||||||
write_concern=write_concern,
|
write_concern=write_concern,
|
||||||
|
read_concern=read_concern,
|
||||||
full_result=True,
|
full_result=True,
|
||||||
**update
|
**update
|
||||||
)
|
)
|
||||||
@ -684,9 +713,9 @@ class BaseQuerySet(object):
|
|||||||
**self._cursor_args
|
**self._cursor_args
|
||||||
)
|
)
|
||||||
except pymongo.errors.DuplicateKeyError as err:
|
except pymongo.errors.DuplicateKeyError as err:
|
||||||
raise NotUniqueError(u"Update failed (%s)" % err)
|
raise NotUniqueError("Update failed (%s)" % err)
|
||||||
except pymongo.errors.OperationFailure as err:
|
except pymongo.errors.OperationFailure as err:
|
||||||
raise OperationError(u"Update failed (%s)" % err)
|
raise OperationError("Update failed (%s)" % err)
|
||||||
|
|
||||||
if full_response:
|
if full_response:
|
||||||
if result["value"] is not None:
|
if result["value"] is not None:
|
||||||
@ -715,10 +744,10 @@ class BaseQuerySet(object):
|
|||||||
return queryset.filter(pk=object_id).first()
|
return queryset.filter(pk=object_id).first()
|
||||||
|
|
||||||
def in_bulk(self, object_ids):
|
def in_bulk(self, object_ids):
|
||||||
"""Retrieve a set of documents by their ids.
|
""""Retrieve a set of documents by their ids.
|
||||||
|
|
||||||
:param object_ids: a list or tuple of ``ObjectId``\ s
|
:param object_ids: a list or tuple of ObjectId's
|
||||||
:rtype: dict of ObjectIds as keys and collection-specific
|
:rtype: dict of ObjectId's as keys and collection-specific
|
||||||
Document subclasses as values.
|
Document subclasses as values.
|
||||||
|
|
||||||
.. versionadded:: 0.3
|
.. versionadded:: 0.3
|
||||||
@ -745,7 +774,9 @@ class BaseQuerySet(object):
|
|||||||
return doc_map
|
return doc_map
|
||||||
|
|
||||||
def none(self):
|
def none(self):
|
||||||
"""Helper that just returns a list"""
|
"""Returns a queryset that never returns any objects and no query will be executed when accessing the results
|
||||||
|
inspired by django none() https://docs.djangoproject.com/en/dev/ref/models/querysets/#none
|
||||||
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset._none = True
|
queryset._none = True
|
||||||
return queryset
|
return queryset
|
||||||
@ -798,13 +829,14 @@ class BaseQuerySet(object):
|
|||||||
"_ordering",
|
"_ordering",
|
||||||
"_snapshot",
|
"_snapshot",
|
||||||
"_timeout",
|
"_timeout",
|
||||||
"_slave_okay",
|
|
||||||
"_read_preference",
|
"_read_preference",
|
||||||
|
"_read_concern",
|
||||||
"_iter",
|
"_iter",
|
||||||
"_scalar",
|
"_scalar",
|
||||||
"_as_pymongo",
|
"_as_pymongo",
|
||||||
"_limit",
|
"_limit",
|
||||||
"_skip",
|
"_skip",
|
||||||
|
"_empty",
|
||||||
"_hint",
|
"_hint",
|
||||||
"_collation",
|
"_collation",
|
||||||
"_auto_dereference",
|
"_auto_dereference",
|
||||||
@ -845,6 +877,7 @@ class BaseQuerySet(object):
|
|||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset._limit = n
|
queryset._limit = n
|
||||||
|
queryset._empty = False # cancels the effect of empty
|
||||||
|
|
||||||
# If a cursor object has already been created, apply the limit to it.
|
# If a cursor object has already been created, apply the limit to it.
|
||||||
if queryset._cursor_obj:
|
if queryset._cursor_obj:
|
||||||
@ -1012,7 +1045,7 @@ class BaseQuerySet(object):
|
|||||||
.. versionchanged:: 0.5 - Added subfield support
|
.. versionchanged:: 0.5 - Added subfield support
|
||||||
"""
|
"""
|
||||||
fields = {f: QueryFieldList.ONLY for f in fields}
|
fields = {f: QueryFieldList.ONLY for f in fields}
|
||||||
self.only_fields = fields.keys()
|
self.only_fields = list(fields.keys())
|
||||||
return self.fields(True, **fields)
|
return self.fields(True, **fields)
|
||||||
|
|
||||||
def exclude(self, *fields):
|
def exclude(self, *fields):
|
||||||
@ -1049,9 +1082,11 @@ class BaseQuerySet(object):
|
|||||||
|
|
||||||
posts = BlogPost.objects(...).fields(comments=0)
|
posts = BlogPost.objects(...).fields(comments=0)
|
||||||
|
|
||||||
To retrieve a subrange of array elements:
|
To retrieve a subrange or sublist of array elements,
|
||||||
|
support exist for both the `slice` and `elemMatch` projection operator:
|
||||||
|
|
||||||
posts = BlogPost.objects(...).fields(slice__comments=5)
|
posts = BlogPost.objects(...).fields(slice__comments=5)
|
||||||
|
posts = BlogPost.objects(...).fields(elemMatch__comments="test")
|
||||||
|
|
||||||
:param kwargs: A set of keyword arguments identifying what to
|
:param kwargs: A set of keyword arguments identifying what to
|
||||||
include, exclude, or slice.
|
include, exclude, or slice.
|
||||||
@ -1060,7 +1095,7 @@ class BaseQuerySet(object):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
# Check for an operator and transform to mongo-style if there is
|
# Check for an operator and transform to mongo-style if there is
|
||||||
operators = ["slice"]
|
operators = ["slice", "elemMatch"]
|
||||||
cleaned_fields = []
|
cleaned_fields = []
|
||||||
for key, value in kwargs.items():
|
for key, value in kwargs.items():
|
||||||
parts = key.split("__")
|
parts = key.split("__")
|
||||||
@ -1163,7 +1198,7 @@ class BaseQuerySet(object):
|
|||||||
|
|
||||||
def explain(self):
|
def explain(self):
|
||||||
"""Return an explain plan record for the
|
"""Return an explain plan record for the
|
||||||
:class:`~mongoengine.queryset.QuerySet`\ 's cursor.
|
:class:`~mongoengine.queryset.QuerySet` cursor.
|
||||||
"""
|
"""
|
||||||
return self._cursor.explain()
|
return self._cursor.explain()
|
||||||
|
|
||||||
@ -1193,20 +1228,6 @@ class BaseQuerySet(object):
|
|||||||
queryset._timeout = enabled
|
queryset._timeout = enabled
|
||||||
return queryset
|
return queryset
|
||||||
|
|
||||||
# DEPRECATED. Has no more impact on PyMongo 3+
|
|
||||||
def slave_okay(self, enabled):
|
|
||||||
"""Enable or disable the slave_okay when querying.
|
|
||||||
|
|
||||||
:param enabled: whether or not the slave_okay is enabled
|
|
||||||
|
|
||||||
.. deprecated:: Ignored with PyMongo 3+
|
|
||||||
"""
|
|
||||||
msg = "slave_okay is deprecated as it has no impact when using PyMongo 3+."
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
queryset = self.clone()
|
|
||||||
queryset._slave_okay = enabled
|
|
||||||
return queryset
|
|
||||||
|
|
||||||
def read_preference(self, read_preference):
|
def read_preference(self, read_preference):
|
||||||
"""Change the read_preference when querying.
|
"""Change the read_preference when querying.
|
||||||
|
|
||||||
@ -1219,6 +1240,22 @@ class BaseQuerySet(object):
|
|||||||
queryset._cursor_obj = None # we need to re-create the cursor object whenever we apply read_preference
|
queryset._cursor_obj = None # we need to re-create the cursor object whenever we apply read_preference
|
||||||
return queryset
|
return queryset
|
||||||
|
|
||||||
|
def read_concern(self, read_concern):
|
||||||
|
"""Change the read_concern when querying.
|
||||||
|
|
||||||
|
:param read_concern: override ReplicaSetConnection-level
|
||||||
|
preference.
|
||||||
|
"""
|
||||||
|
if read_concern is not None and not isinstance(read_concern, Mapping):
|
||||||
|
raise TypeError("%r is not a valid read concern." % (read_concern,))
|
||||||
|
|
||||||
|
queryset = self.clone()
|
||||||
|
queryset._read_concern = (
|
||||||
|
ReadConcern(**read_concern) if read_concern is not None else None
|
||||||
|
)
|
||||||
|
queryset._cursor_obj = None # we need to re-create the cursor object whenever we apply read_concern
|
||||||
|
return queryset
|
||||||
|
|
||||||
def scalar(self, *fields):
|
def scalar(self, *fields):
|
||||||
"""Instead of returning Document instances, return either a specific
|
"""Instead of returning Document instances, return either a specific
|
||||||
value or a tuple of values in order.
|
value or a tuple of values in order.
|
||||||
@ -1318,10 +1355,11 @@ class BaseQuerySet(object):
|
|||||||
final_pipeline = initial_pipeline + user_pipeline
|
final_pipeline = initial_pipeline + user_pipeline
|
||||||
|
|
||||||
collection = self._collection
|
collection = self._collection
|
||||||
if self._read_preference is not None:
|
if self._read_preference is not None or self._read_concern is not None:
|
||||||
collection = self._collection.with_options(
|
collection = self._collection.with_options(
|
||||||
read_preference=self._read_preference
|
read_preference=self._read_preference, read_concern=self._read_concern
|
||||||
)
|
)
|
||||||
|
|
||||||
return collection.aggregate(final_pipeline, cursor={}, **kwargs)
|
return collection.aggregate(final_pipeline, cursor={}, **kwargs)
|
||||||
|
|
||||||
# JS functionality
|
# JS functionality
|
||||||
@ -1375,13 +1413,13 @@ class BaseQuerySet(object):
|
|||||||
map_f_scope = {}
|
map_f_scope = {}
|
||||||
if isinstance(map_f, Code):
|
if isinstance(map_f, Code):
|
||||||
map_f_scope = map_f.scope
|
map_f_scope = map_f.scope
|
||||||
map_f = six.text_type(map_f)
|
map_f = str(map_f)
|
||||||
map_f = Code(queryset._sub_js_fields(map_f), map_f_scope)
|
map_f = Code(queryset._sub_js_fields(map_f), map_f_scope)
|
||||||
|
|
||||||
reduce_f_scope = {}
|
reduce_f_scope = {}
|
||||||
if isinstance(reduce_f, Code):
|
if isinstance(reduce_f, Code):
|
||||||
reduce_f_scope = reduce_f.scope
|
reduce_f_scope = reduce_f.scope
|
||||||
reduce_f = six.text_type(reduce_f)
|
reduce_f = str(reduce_f)
|
||||||
reduce_f_code = queryset._sub_js_fields(reduce_f)
|
reduce_f_code = queryset._sub_js_fields(reduce_f)
|
||||||
reduce_f = Code(reduce_f_code, reduce_f_scope)
|
reduce_f = Code(reduce_f_code, reduce_f_scope)
|
||||||
|
|
||||||
@ -1391,7 +1429,7 @@ class BaseQuerySet(object):
|
|||||||
finalize_f_scope = {}
|
finalize_f_scope = {}
|
||||||
if isinstance(finalize_f, Code):
|
if isinstance(finalize_f, Code):
|
||||||
finalize_f_scope = finalize_f.scope
|
finalize_f_scope = finalize_f.scope
|
||||||
finalize_f = six.text_type(finalize_f)
|
finalize_f = str(finalize_f)
|
||||||
finalize_f_code = queryset._sub_js_fields(finalize_f)
|
finalize_f_code = queryset._sub_js_fields(finalize_f)
|
||||||
finalize_f = Code(finalize_f_code, finalize_f_scope)
|
finalize_f = Code(finalize_f_code, finalize_f_scope)
|
||||||
mr_args["finalize"] = finalize_f
|
mr_args["finalize"] = finalize_f
|
||||||
@ -1407,7 +1445,7 @@ class BaseQuerySet(object):
|
|||||||
else:
|
else:
|
||||||
map_reduce_function = "map_reduce"
|
map_reduce_function = "map_reduce"
|
||||||
|
|
||||||
if isinstance(output, six.string_types):
|
if isinstance(output, str):
|
||||||
mr_args["out"] = output
|
mr_args["out"] = output
|
||||||
|
|
||||||
elif isinstance(output, dict):
|
elif isinstance(output, dict):
|
||||||
@ -1591,10 +1629,10 @@ class BaseQuerySet(object):
|
|||||||
def __next__(self):
|
def __next__(self):
|
||||||
"""Wrap the result in a :class:`~mongoengine.Document` object.
|
"""Wrap the result in a :class:`~mongoengine.Document` object.
|
||||||
"""
|
"""
|
||||||
if self._limit == 0 or self._none:
|
if self._none or self._empty:
|
||||||
raise StopIteration
|
raise StopIteration
|
||||||
|
|
||||||
raw_doc = six.next(self._cursor)
|
raw_doc = next(self._cursor)
|
||||||
|
|
||||||
if self._as_pymongo:
|
if self._as_pymongo:
|
||||||
return raw_doc
|
return raw_doc
|
||||||
@ -1610,8 +1648,6 @@ class BaseQuerySet(object):
|
|||||||
|
|
||||||
return doc
|
return doc
|
||||||
|
|
||||||
next = __next__ # For Python2 support
|
|
||||||
|
|
||||||
def rewind(self):
|
def rewind(self):
|
||||||
"""Rewind the cursor to its unevaluated state.
|
"""Rewind the cursor to its unevaluated state.
|
||||||
|
|
||||||
@ -1665,9 +1701,9 @@ class BaseQuerySet(object):
|
|||||||
# XXX In PyMongo 3+, we define the read preference on a collection
|
# XXX In PyMongo 3+, we define the read preference on a collection
|
||||||
# level, not a cursor level. Thus, we need to get a cloned collection
|
# level, not a cursor level. Thus, we need to get a cloned collection
|
||||||
# object using `with_options` first.
|
# object using `with_options` first.
|
||||||
if self._read_preference is not None:
|
if self._read_preference is not None or self._read_concern is not None:
|
||||||
self._cursor_obj = self._collection.with_options(
|
self._cursor_obj = self._collection.with_options(
|
||||||
read_preference=self._read_preference
|
read_preference=self._read_preference, read_concern=self._read_concern
|
||||||
).find(self._query, **self._cursor_args)
|
).find(self._query, **self._cursor_args)
|
||||||
else:
|
else:
|
||||||
self._cursor_obj = self._collection.find(self._query, **self._cursor_args)
|
self._cursor_obj = self._collection.find(self._query, **self._cursor_args)
|
||||||
@ -1839,13 +1875,13 @@ class BaseQuerySet(object):
|
|||||||
}
|
}
|
||||||
"""
|
"""
|
||||||
total, data, types = self.exec_js(freq_func, field)
|
total, data, types = self.exec_js(freq_func, field)
|
||||||
values = {types.get(k): int(v) for k, v in iteritems(data)}
|
values = {types.get(k): int(v) for k, v in data.items()}
|
||||||
|
|
||||||
if normalize:
|
if normalize:
|
||||||
values = {k: float(v) / total for k, v in values.items()}
|
values = {k: float(v) / total for k, v in values.items()}
|
||||||
|
|
||||||
frequencies = {}
|
frequencies = {}
|
||||||
for k, v in iteritems(values):
|
for k, v in values.items():
|
||||||
if isinstance(k, float):
|
if isinstance(k, float):
|
||||||
if int(k) == k:
|
if int(k) == k:
|
||||||
k = int(k)
|
k = int(k)
|
||||||
@ -1865,7 +1901,7 @@ class BaseQuerySet(object):
|
|||||||
field_parts = field.split(".")
|
field_parts = field.split(".")
|
||||||
try:
|
try:
|
||||||
field = ".".join(
|
field = ".".join(
|
||||||
f if isinstance(f, six.string_types) else f.db_field
|
f if isinstance(f, str) else f.db_field
|
||||||
for f in self._document._lookup_field(field_parts)
|
for f in self._document._lookup_field(field_parts)
|
||||||
)
|
)
|
||||||
db_field_paths.append(field)
|
db_field_paths.append(field)
|
||||||
@ -1877,7 +1913,7 @@ class BaseQuerySet(object):
|
|||||||
for subdoc in subclasses:
|
for subdoc in subclasses:
|
||||||
try:
|
try:
|
||||||
subfield = ".".join(
|
subfield = ".".join(
|
||||||
f if isinstance(f, six.string_types) else f.db_field
|
f if isinstance(f, str) else f.db_field
|
||||||
for f in subdoc._lookup_field(field_parts)
|
for f in subdoc._lookup_field(field_parts)
|
||||||
)
|
)
|
||||||
db_field_paths.append(subfield)
|
db_field_paths.append(subfield)
|
||||||
@ -1951,7 +1987,7 @@ class BaseQuerySet(object):
|
|||||||
field_name = match.group(1).split(".")
|
field_name = match.group(1).split(".")
|
||||||
fields = self._document._lookup_field(field_name)
|
fields = self._document._lookup_field(field_name)
|
||||||
# Substitute the correct name for the field into the javascript
|
# Substitute the correct name for the field into the javascript
|
||||||
return u'["%s"]' % fields[-1].db_field
|
return '["%s"]' % fields[-1].db_field
|
||||||
|
|
||||||
def field_path_sub(match):
|
def field_path_sub(match):
|
||||||
# Extract just the field name, and look up the field objects
|
# Extract just the field name, and look up the field objects
|
||||||
@ -1981,23 +2017,3 @@ class BaseQuerySet(object):
|
|||||||
setattr(queryset, "_" + method_name, val)
|
setattr(queryset, "_" + method_name, val)
|
||||||
|
|
||||||
return queryset
|
return queryset
|
||||||
|
|
||||||
# Deprecated
|
|
||||||
def ensure_index(self, **kwargs):
|
|
||||||
"""Deprecated use :func:`Document.ensure_index`"""
|
|
||||||
msg = (
|
|
||||||
"Doc.objects()._ensure_index() is deprecated. "
|
|
||||||
"Use Doc.ensure_index() instead."
|
|
||||||
)
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
self._document.__class__.ensure_index(**kwargs)
|
|
||||||
return self
|
|
||||||
|
|
||||||
def _ensure_indexes(self):
|
|
||||||
"""Deprecated use :func:`~Document.ensure_indexes`"""
|
|
||||||
msg = (
|
|
||||||
"Doc.objects()._ensure_indexes() is deprecated. "
|
|
||||||
"Use Doc.ensure_indexes() instead."
|
|
||||||
)
|
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
self._document.__class__.ensure_indexes()
|
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
__all__ = ("QueryFieldList",)
|
__all__ = ("QueryFieldList",)
|
||||||
|
|
||||||
|
|
||||||
class QueryFieldList(object):
|
class QueryFieldList:
|
||||||
"""Object that handles combinations of .only() and .exclude() calls"""
|
"""Object that handles combinations of .only() and .exclude() calls"""
|
||||||
|
|
||||||
ONLY = 1
|
ONLY = 1
|
||||||
@ -69,8 +69,6 @@ class QueryFieldList(object):
|
|||||||
def __bool__(self):
|
def __bool__(self):
|
||||||
return bool(self.fields)
|
return bool(self.fields)
|
||||||
|
|
||||||
__nonzero__ = __bool__ # For Py2 support
|
|
||||||
|
|
||||||
def as_dict(self):
|
def as_dict(self):
|
||||||
field_list = {field: self.value for field in self.fields}
|
field_list = {field: self.value for field in self.fields}
|
||||||
if self.slice:
|
if self.slice:
|
||||||
@ -80,7 +78,7 @@ class QueryFieldList(object):
|
|||||||
return field_list
|
return field_list
|
||||||
|
|
||||||
def reset(self):
|
def reset(self):
|
||||||
self.fields = set([])
|
self.fields = set()
|
||||||
self.slice = {}
|
self.slice = {}
|
||||||
self.value = self.ONLY
|
self.value = self.ONLY
|
||||||
|
|
||||||
|
@ -4,7 +4,7 @@ from mongoengine.queryset.queryset import QuerySet
|
|||||||
__all__ = ("queryset_manager", "QuerySetManager")
|
__all__ = ("queryset_manager", "QuerySetManager")
|
||||||
|
|
||||||
|
|
||||||
class QuerySetManager(object):
|
class QuerySetManager:
|
||||||
"""
|
"""
|
||||||
The default QuerySet Manager.
|
The default QuerySet Manager.
|
||||||
|
|
||||||
|
@ -1,5 +1,3 @@
|
|||||||
import six
|
|
||||||
|
|
||||||
from mongoengine.errors import OperationError
|
from mongoengine.errors import OperationError
|
||||||
from mongoengine.queryset.base import (
|
from mongoengine.queryset.base import (
|
||||||
BaseQuerySet,
|
BaseQuerySet,
|
||||||
@ -127,8 +125,8 @@ class QuerySet(BaseQuerySet):
|
|||||||
# Pull in ITER_CHUNK_SIZE docs from the database and store them in
|
# Pull in ITER_CHUNK_SIZE docs from the database and store them in
|
||||||
# the result cache.
|
# the result cache.
|
||||||
try:
|
try:
|
||||||
for _ in six.moves.range(ITER_CHUNK_SIZE):
|
for _ in range(ITER_CHUNK_SIZE):
|
||||||
self._result_cache.append(six.next(self))
|
self._result_cache.append(next(self))
|
||||||
except StopIteration:
|
except StopIteration:
|
||||||
# Getting this exception means there are no more docs in the
|
# Getting this exception means there are no more docs in the
|
||||||
# db cursor. Set _has_more to False so that we can use that
|
# db cursor. Set _has_more to False so that we can use that
|
||||||
@ -143,11 +141,11 @@ class QuerySet(BaseQuerySet):
|
|||||||
getting the count
|
getting the count
|
||||||
"""
|
"""
|
||||||
if with_limit_and_skip is False:
|
if with_limit_and_skip is False:
|
||||||
return super(QuerySet, self).count(with_limit_and_skip)
|
return super().count(with_limit_and_skip)
|
||||||
|
|
||||||
if self._len is None:
|
if self._len is None:
|
||||||
# cache the length
|
# cache the length
|
||||||
self._len = super(QuerySet, self).count(with_limit_and_skip)
|
self._len = super().count(with_limit_and_skip)
|
||||||
|
|
||||||
return self._len
|
return self._len
|
||||||
|
|
||||||
@ -181,9 +179,9 @@ class QuerySetNoCache(BaseQuerySet):
|
|||||||
return ".. queryset mid-iteration .."
|
return ".. queryset mid-iteration .."
|
||||||
|
|
||||||
data = []
|
data = []
|
||||||
for _ in six.moves.range(REPR_OUTPUT_SIZE + 1):
|
for _ in range(REPR_OUTPUT_SIZE + 1):
|
||||||
try:
|
try:
|
||||||
data.append(six.next(self))
|
data.append(next(self))
|
||||||
except StopIteration:
|
except StopIteration:
|
||||||
break
|
break
|
||||||
|
|
||||||
|
@ -3,14 +3,12 @@ from collections import defaultdict
|
|||||||
from bson import ObjectId, SON
|
from bson import ObjectId, SON
|
||||||
from bson.dbref import DBRef
|
from bson.dbref import DBRef
|
||||||
import pymongo
|
import pymongo
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine.base import UPDATE_OPERATORS
|
from mongoengine.base import UPDATE_OPERATORS
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.errors import InvalidQueryError
|
from mongoengine.errors import InvalidQueryError
|
||||||
|
|
||||||
__all__ = ("query", "update")
|
__all__ = ("query", "update", "STRING_OPERATORS")
|
||||||
|
|
||||||
COMPARISON_OPERATORS = (
|
COMPARISON_OPERATORS = (
|
||||||
"ne",
|
"ne",
|
||||||
@ -101,7 +99,7 @@ def query(_doc_cls=None, **kwargs):
|
|||||||
cleaned_fields = []
|
cleaned_fields = []
|
||||||
for field in fields:
|
for field in fields:
|
||||||
append_field = True
|
append_field = True
|
||||||
if isinstance(field, six.string_types):
|
if isinstance(field, str):
|
||||||
parts.append(field)
|
parts.append(field)
|
||||||
append_field = False
|
append_field = False
|
||||||
# is last and CachedReferenceField
|
# is last and CachedReferenceField
|
||||||
@ -169,9 +167,9 @@ def query(_doc_cls=None, **kwargs):
|
|||||||
|
|
||||||
key = ".".join(parts)
|
key = ".".join(parts)
|
||||||
|
|
||||||
if op is None or key not in mongo_query:
|
if key not in mongo_query:
|
||||||
mongo_query[key] = value
|
mongo_query[key] = value
|
||||||
elif key in mongo_query:
|
else:
|
||||||
if isinstance(mongo_query[key], dict) and isinstance(value, dict):
|
if isinstance(mongo_query[key], dict) and isinstance(value, dict):
|
||||||
mongo_query[key].update(value)
|
mongo_query[key].update(value)
|
||||||
# $max/minDistance needs to come last - convert to SON
|
# $max/minDistance needs to come last - convert to SON
|
||||||
@ -180,7 +178,7 @@ def query(_doc_cls=None, **kwargs):
|
|||||||
"$near" in value_dict or "$nearSphere" in value_dict
|
"$near" in value_dict or "$nearSphere" in value_dict
|
||||||
):
|
):
|
||||||
value_son = SON()
|
value_son = SON()
|
||||||
for k, v in iteritems(value_dict):
|
for k, v in value_dict.items():
|
||||||
if k == "$maxDistance" or k == "$minDistance":
|
if k == "$maxDistance" or k == "$minDistance":
|
||||||
continue
|
continue
|
||||||
value_son[k] = v
|
value_son[k] = v
|
||||||
@ -281,7 +279,7 @@ def update(_doc_cls=None, **update):
|
|||||||
appended_sub_field = False
|
appended_sub_field = False
|
||||||
for field in fields:
|
for field in fields:
|
||||||
append_field = True
|
append_field = True
|
||||||
if isinstance(field, six.string_types):
|
if isinstance(field, str):
|
||||||
# Convert the S operator to $
|
# Convert the S operator to $
|
||||||
if field == "S":
|
if field == "S":
|
||||||
field = "$"
|
field = "$"
|
||||||
@ -435,7 +433,9 @@ def _geo_operator(field, op, value):
|
|||||||
value = {"$near": _infer_geometry(value)}
|
value = {"$near": _infer_geometry(value)}
|
||||||
else:
|
else:
|
||||||
raise NotImplementedError(
|
raise NotImplementedError(
|
||||||
'Geo method "%s" has not been implemented for a %s ' % (op, field._name)
|
'Geo method "{}" has not been implemented for a {} '.format(
|
||||||
|
op, field._name
|
||||||
|
)
|
||||||
)
|
)
|
||||||
return value
|
return value
|
||||||
|
|
||||||
|
@ -7,7 +7,12 @@ from mongoengine.queryset import transform
|
|||||||
__all__ = ("Q", "QNode")
|
__all__ = ("Q", "QNode")
|
||||||
|
|
||||||
|
|
||||||
class QNodeVisitor(object):
|
def warn_empty_is_deprecated():
|
||||||
|
msg = "'empty' property is deprecated in favour of using 'not bool(filter)'"
|
||||||
|
warnings.warn(msg, DeprecationWarning, stacklevel=2)
|
||||||
|
|
||||||
|
|
||||||
|
class QNodeVisitor:
|
||||||
"""Base visitor class for visiting Q-object nodes in a query tree.
|
"""Base visitor class for visiting Q-object nodes in a query tree.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@ -79,7 +84,7 @@ class QueryCompilerVisitor(QNodeVisitor):
|
|||||||
return transform.query(self.document, **query.query)
|
return transform.query(self.document, **query.query)
|
||||||
|
|
||||||
|
|
||||||
class QNode(object):
|
class QNode:
|
||||||
"""Base class for nodes in query trees."""
|
"""Base class for nodes in query trees."""
|
||||||
|
|
||||||
AND = 0
|
AND = 0
|
||||||
@ -98,19 +103,18 @@ class QNode(object):
|
|||||||
object.
|
object.
|
||||||
"""
|
"""
|
||||||
# If the other Q() is empty, ignore it and just use `self`.
|
# If the other Q() is empty, ignore it and just use `self`.
|
||||||
if getattr(other, "empty", True):
|
if not bool(other):
|
||||||
return self
|
return self
|
||||||
|
|
||||||
# Or if this Q is empty, ignore it and just use `other`.
|
# Or if this Q is empty, ignore it and just use `other`.
|
||||||
if self.empty:
|
if not bool(self):
|
||||||
return other
|
return other
|
||||||
|
|
||||||
return QCombination(operation, [self, other])
|
return QCombination(operation, [self, other])
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def empty(self):
|
def empty(self):
|
||||||
msg = "'empty' property is deprecated in favour of using 'not bool(filter)'"
|
warn_empty_is_deprecated()
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
def __or__(self, other):
|
def __or__(self, other):
|
||||||
@ -143,8 +147,6 @@ class QCombination(QNode):
|
|||||||
def __bool__(self):
|
def __bool__(self):
|
||||||
return bool(self.children)
|
return bool(self.children)
|
||||||
|
|
||||||
__nonzero__ = __bool__ # For Py2 support
|
|
||||||
|
|
||||||
def accept(self, visitor):
|
def accept(self, visitor):
|
||||||
for i in range(len(self.children)):
|
for i in range(len(self.children)):
|
||||||
if isinstance(self.children[i], QNode):
|
if isinstance(self.children[i], QNode):
|
||||||
@ -154,8 +156,7 @@ class QCombination(QNode):
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def empty(self):
|
def empty(self):
|
||||||
msg = "'empty' property is deprecated in favour of using 'not bool(filter)'"
|
warn_empty_is_deprecated()
|
||||||
warnings.warn(msg, DeprecationWarning)
|
|
||||||
return not bool(self.children)
|
return not bool(self.children)
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
@ -180,8 +181,6 @@ class Q(QNode):
|
|||||||
def __bool__(self):
|
def __bool__(self):
|
||||||
return bool(self.query)
|
return bool(self.query)
|
||||||
|
|
||||||
__nonzero__ = __bool__ # For Py2 support
|
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
return self.__class__ == other.__class__ and self.query == other.query
|
return self.__class__ == other.__class__ and self.query == other.query
|
||||||
|
|
||||||
@ -190,4 +189,5 @@ class Q(QNode):
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def empty(self):
|
def empty(self):
|
||||||
|
warn_empty_is_deprecated()
|
||||||
return not bool(self.query)
|
return not bool(self.query)
|
||||||
|
@ -15,11 +15,11 @@ try:
|
|||||||
signals_available = True
|
signals_available = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
|
|
||||||
class Namespace(object):
|
class Namespace:
|
||||||
def signal(self, name, doc=None):
|
def signal(self, name, doc=None):
|
||||||
return _FakeSignal(name, doc)
|
return _FakeSignal(name, doc)
|
||||||
|
|
||||||
class _FakeSignal(object):
|
class _FakeSignal:
|
||||||
"""If blinker is unavailable, create a fake class with the same
|
"""If blinker is unavailable, create a fake class with the same
|
||||||
interface that allows sending of signals but will fail with an
|
interface that allows sending of signals but will fail with an
|
||||||
error on anything else. Instead of doing anything on send, it
|
error on anything else. Instead of doing anything on send, it
|
||||||
|
8
requirements-dev.txt
Normal file
8
requirements-dev.txt
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
black
|
||||||
|
flake8
|
||||||
|
flake8-import-order
|
||||||
|
pre-commit
|
||||||
|
pytest
|
||||||
|
ipdb
|
||||||
|
ipython
|
||||||
|
tox
|
@ -1,3 +0,0 @@
|
|||||||
black
|
|
||||||
flake8
|
|
||||||
flake8-import-order
|
|
@ -1,4 +1,3 @@
|
|||||||
pymongo>=3.4
|
pymongo>=3.4
|
||||||
six==1.10.0
|
|
||||||
Sphinx==1.5.5
|
Sphinx==1.5.5
|
||||||
sphinx-rtd-theme==0.2.4
|
sphinx-rtd-theme==0.2.4
|
||||||
|
24
setup.py
24
setup.py
@ -92,16 +92,16 @@ version_line = list(filter(lambda l: l.startswith("VERSION"), open(init)))[0]
|
|||||||
VERSION = get_version(eval(version_line.split("=")[-1]))
|
VERSION = get_version(eval(version_line.split("=")[-1]))
|
||||||
|
|
||||||
CLASSIFIERS = [
|
CLASSIFIERS = [
|
||||||
"Development Status :: 4 - Beta",
|
"Development Status :: 5 - Production/Stable",
|
||||||
"Intended Audience :: Developers",
|
"Intended Audience :: Developers",
|
||||||
"License :: OSI Approved :: MIT License",
|
"License :: OSI Approved :: MIT License",
|
||||||
"Operating System :: OS Independent",
|
"Operating System :: OS Independent",
|
||||||
"Programming Language :: Python",
|
"Programming Language :: Python",
|
||||||
"Programming Language :: Python :: 2",
|
|
||||||
"Programming Language :: Python :: 2.7",
|
|
||||||
"Programming Language :: Python :: 3",
|
"Programming Language :: Python :: 3",
|
||||||
"Programming Language :: Python :: 3.5",
|
"Programming Language :: Python :: 3.5",
|
||||||
"Programming Language :: Python :: 3.6",
|
"Programming Language :: Python :: 3.6",
|
||||||
|
"Programming Language :: Python :: 3.7",
|
||||||
|
"Programming Language :: Python :: 3.8",
|
||||||
"Programming Language :: Python :: Implementation :: CPython",
|
"Programming Language :: Python :: Implementation :: CPython",
|
||||||
"Programming Language :: Python :: Implementation :: PyPy",
|
"Programming Language :: Python :: Implementation :: PyPy",
|
||||||
"Topic :: Database",
|
"Topic :: Database",
|
||||||
@ -118,15 +118,12 @@ extra_opts = {
|
|||||||
"Pillow>=2.0.0, <7.0.0", # 7.0.0 dropped Python2 support
|
"Pillow>=2.0.0, <7.0.0", # 7.0.0 dropped Python2 support
|
||||||
],
|
],
|
||||||
}
|
}
|
||||||
if sys.version_info[0] == 3:
|
|
||||||
extra_opts["use_2to3"] = True
|
if "test" in sys.argv:
|
||||||
if "test" in sys.argv:
|
extra_opts["packages"] = find_packages()
|
||||||
extra_opts["packages"] = find_packages()
|
extra_opts["package_data"] = {
|
||||||
extra_opts["package_data"] = {
|
"tests": ["fields/mongoengine.png", "fields/mongodb_leaf.png"]
|
||||||
"tests": ["fields/mongoengine.png", "fields/mongodb_leaf.png"]
|
}
|
||||||
}
|
|
||||||
else:
|
|
||||||
extra_opts["tests_require"] += ["python-dateutil"]
|
|
||||||
|
|
||||||
setup(
|
setup(
|
||||||
name="mongoengine",
|
name="mongoengine",
|
||||||
@ -143,7 +140,8 @@ setup(
|
|||||||
long_description=LONG_DESCRIPTION,
|
long_description=LONG_DESCRIPTION,
|
||||||
platforms=["any"],
|
platforms=["any"],
|
||||||
classifiers=CLASSIFIERS,
|
classifiers=CLASSIFIERS,
|
||||||
install_requires=["pymongo>=3.4, <4.0", "six>=1.10.0"],
|
python_requires=">=3.5",
|
||||||
|
install_requires=["pymongo>=3.4, <4.0"],
|
||||||
cmdclass={"test": PyTest},
|
cmdclass={"test": PyTest},
|
||||||
**extra_opts
|
**extra_opts
|
||||||
)
|
)
|
||||||
|
@ -5,7 +5,6 @@ from datetime import datetime
|
|||||||
from pymongo.collation import Collation
|
from pymongo.collation import Collation
|
||||||
from pymongo.errors import OperationFailure
|
from pymongo.errors import OperationFailure
|
||||||
import pytest
|
import pytest
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
@ -59,7 +58,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
info = BlogPost.objects._collection.index_information()
|
info = BlogPost.objects._collection.index_information()
|
||||||
# _id, '-date', 'tags', ('cat', 'date')
|
# _id, '-date', 'tags', ('cat', 'date')
|
||||||
assert len(info) == 4
|
assert len(info) == 4
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
for expected in expected_specs:
|
for expected in expected_specs:
|
||||||
assert expected["fields"] in info
|
assert expected["fields"] in info
|
||||||
|
|
||||||
@ -87,7 +86,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
# the indices on -date and tags will both contain
|
# the indices on -date and tags will both contain
|
||||||
# _cls as first element in the key
|
# _cls as first element in the key
|
||||||
assert len(info) == 4
|
assert len(info) == 4
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
for expected in expected_specs:
|
for expected in expected_specs:
|
||||||
assert expected["fields"] in info
|
assert expected["fields"] in info
|
||||||
|
|
||||||
@ -102,7 +101,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
|
|
||||||
ExtendedBlogPost.ensure_indexes()
|
ExtendedBlogPost.ensure_indexes()
|
||||||
info = ExtendedBlogPost.objects._collection.index_information()
|
info = ExtendedBlogPost.objects._collection.index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
for expected in expected_specs:
|
for expected in expected_specs:
|
||||||
assert expected["fields"] in info
|
assert expected["fields"] in info
|
||||||
|
|
||||||
@ -192,7 +191,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
# Indexes are lazy so use list() to perform query
|
# Indexes are lazy so use list() to perform query
|
||||||
list(Person.objects)
|
list(Person.objects)
|
||||||
info = Person.objects._collection.index_information()
|
info = Person.objects._collection.index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("rank.title", 1)] in info
|
assert [("rank.title", 1)] in info
|
||||||
|
|
||||||
def test_explicit_geo2d_index(self):
|
def test_explicit_geo2d_index(self):
|
||||||
@ -207,7 +206,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
|
|
||||||
Place.ensure_indexes()
|
Place.ensure_indexes()
|
||||||
info = Place._get_collection().index_information()
|
info = Place._get_collection().index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("location.point", "2d")] in info
|
assert [("location.point", "2d")] in info
|
||||||
|
|
||||||
def test_explicit_geo2d_index_embedded(self):
|
def test_explicit_geo2d_index_embedded(self):
|
||||||
@ -227,7 +226,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
|
|
||||||
Place.ensure_indexes()
|
Place.ensure_indexes()
|
||||||
info = Place._get_collection().index_information()
|
info = Place._get_collection().index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("current.location.point", "2d")] in info
|
assert [("current.location.point", "2d")] in info
|
||||||
|
|
||||||
def test_explicit_geosphere_index(self):
|
def test_explicit_geosphere_index(self):
|
||||||
@ -244,7 +243,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
|
|
||||||
Place.ensure_indexes()
|
Place.ensure_indexes()
|
||||||
info = Place._get_collection().index_information()
|
info = Place._get_collection().index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("location.point", "2dsphere")] in info
|
assert [("location.point", "2dsphere")] in info
|
||||||
|
|
||||||
def test_explicit_geohaystack_index(self):
|
def test_explicit_geohaystack_index(self):
|
||||||
@ -266,7 +265,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
|
|
||||||
Place.ensure_indexes()
|
Place.ensure_indexes()
|
||||||
info = Place._get_collection().index_information()
|
info = Place._get_collection().index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("location.point", "geoHaystack")] in info
|
assert [("location.point", "geoHaystack")] in info
|
||||||
|
|
||||||
def test_create_geohaystack_index(self):
|
def test_create_geohaystack_index(self):
|
||||||
@ -279,7 +278,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
|
|
||||||
Place.create_index({"fields": (")location.point", "name")}, bucketSize=10)
|
Place.create_index({"fields": (")location.point", "name")}, bucketSize=10)
|
||||||
info = Place._get_collection().index_information()
|
info = Place._get_collection().index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("location.point", "geoHaystack"), ("name", 1)] in info
|
assert [("location.point", "geoHaystack"), ("name", 1)] in info
|
||||||
|
|
||||||
def test_dictionary_indexes(self):
|
def test_dictionary_indexes(self):
|
||||||
@ -308,7 +307,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
info = BlogPost.objects._collection.index_information()
|
info = BlogPost.objects._collection.index_information()
|
||||||
info = [
|
info = [
|
||||||
(value["key"], value.get("unique", False), value.get("sparse", False))
|
(value["key"], value.get("unique", False), value.get("sparse", False))
|
||||||
for key, value in iteritems(info)
|
for key, value in info.items()
|
||||||
]
|
]
|
||||||
assert ([("addDate", -1)], True, True) in info
|
assert ([("addDate", -1)], True, True) in info
|
||||||
|
|
||||||
@ -807,18 +806,6 @@ class TestIndexes(unittest.TestCase):
|
|||||||
info = Log.objects._collection.index_information()
|
info = Log.objects._collection.index_information()
|
||||||
assert 3600 == info["created_1"]["expireAfterSeconds"]
|
assert 3600 == info["created_1"]["expireAfterSeconds"]
|
||||||
|
|
||||||
def test_index_drop_dups_silently_ignored(self):
|
|
||||||
class Customer(Document):
|
|
||||||
cust_id = IntField(unique=True, required=True)
|
|
||||||
meta = {
|
|
||||||
"indexes": ["cust_id"],
|
|
||||||
"index_drop_dups": True,
|
|
||||||
"allow_inheritance": False,
|
|
||||||
}
|
|
||||||
|
|
||||||
Customer.drop_collection()
|
|
||||||
Customer.objects.first()
|
|
||||||
|
|
||||||
def test_unique_and_indexes(self):
|
def test_unique_and_indexes(self):
|
||||||
"""Ensure that 'unique' constraints aren't overridden by
|
"""Ensure that 'unique' constraints aren't overridden by
|
||||||
meta.indexes.
|
meta.indexes.
|
||||||
@ -902,7 +889,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
self.fail("Unbound local error at index + pk definition")
|
self.fail("Unbound local error at index + pk definition")
|
||||||
|
|
||||||
info = BlogPost.objects._collection.index_information()
|
info = BlogPost.objects._collection.index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
index_item = [("_id", 1), ("comments.comment_id", 1)]
|
index_item = [("_id", 1), ("comments.comment_id", 1)]
|
||||||
assert index_item in info
|
assert index_item in info
|
||||||
|
|
||||||
@ -943,7 +930,7 @@ class TestIndexes(unittest.TestCase):
|
|||||||
meta = {"indexes": ["provider_ids.foo", "provider_ids.bar"]}
|
meta = {"indexes": ["provider_ids.foo", "provider_ids.bar"]}
|
||||||
|
|
||||||
info = MyDoc.objects._collection.index_information()
|
info = MyDoc.objects._collection.index_information()
|
||||||
info = [value["key"] for key, value in iteritems(info)]
|
info = [value["key"] for key, value in info.items()]
|
||||||
assert [("provider_ids.foo", 1)] in info
|
assert [("provider_ids.foo", 1)] in info
|
||||||
assert [("provider_ids.bar", 1)] in info
|
assert [("provider_ids.bar", 1)] in info
|
||||||
|
|
||||||
@ -1059,10 +1046,6 @@ class TestIndexes(unittest.TestCase):
|
|||||||
del index_info[key][
|
del index_info[key][
|
||||||
"ns"
|
"ns"
|
||||||
] # drop the index namespace - we don't care about that here, MongoDB 3+
|
] # drop the index namespace - we don't care about that here, MongoDB 3+
|
||||||
if "dropDups" in index_info[key]:
|
|
||||||
del index_info[key][
|
|
||||||
"dropDups"
|
|
||||||
] # drop the index dropDups - it is deprecated in MongoDB 3+
|
|
||||||
|
|
||||||
assert index_info == {
|
assert index_info == {
|
||||||
"txt_1": {"key": [("txt", 1)], "background": False},
|
"txt_1": {"key": [("txt", 1)], "background": False},
|
||||||
|
@ -3,7 +3,6 @@ import unittest
|
|||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import (
|
from mongoengine import (
|
||||||
BooleanField,
|
BooleanField,
|
||||||
@ -523,7 +522,6 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
|
|
||||||
defaults = {
|
defaults = {
|
||||||
"index_background": True,
|
"index_background": True,
|
||||||
"index_drop_dups": True,
|
|
||||||
"index_opts": {"hello": "world"},
|
"index_opts": {"hello": "world"},
|
||||||
"allow_inheritance": True,
|
"allow_inheritance": True,
|
||||||
"queryset_class": "QuerySet",
|
"queryset_class": "QuerySet",
|
||||||
@ -550,7 +548,7 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
class Human(Mammal):
|
class Human(Mammal):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
for k, v in iteritems(defaults):
|
for k, v in defaults.items():
|
||||||
for cls in [Animal, Fish, Guppy]:
|
for cls in [Animal, Fish, Guppy]:
|
||||||
assert cls._meta[k] == v
|
assert cls._meta[k] == v
|
||||||
|
|
||||||
|
@ -10,7 +10,6 @@ import bson
|
|||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
from pymongo.errors import DuplicateKeyError
|
from pymongo.errors import DuplicateKeyError
|
||||||
import pytest
|
import pytest
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
@ -1415,7 +1414,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
assert raw_doc["first_name"] == "John"
|
assert raw_doc["first_name"] == "John"
|
||||||
|
|
||||||
def test_inserts_if_you_set_the_pk(self):
|
def test_inserts_if_you_set_the_pk(self):
|
||||||
p1 = self.Person(name="p1", id=bson.ObjectId()).save()
|
_ = self.Person(name="p1", id=bson.ObjectId()).save()
|
||||||
p2 = self.Person(name="p2")
|
p2 = self.Person(name="p2")
|
||||||
p2.id = bson.ObjectId()
|
p2.id = bson.ObjectId()
|
||||||
p2.save()
|
p2.save()
|
||||||
@ -2196,7 +2195,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
|
|
||||||
user = User(name="Mike").save()
|
user = User(name="Mike").save()
|
||||||
reviewer = User(name="John").save()
|
reviewer = User(name="John").save()
|
||||||
book = Book(author=user, reviewer=reviewer).save()
|
_ = Book(author=user, reviewer=reviewer).save()
|
||||||
|
|
||||||
reviewer.delete()
|
reviewer.delete()
|
||||||
assert Book.objects.count() == 1
|
assert Book.objects.count() == 1
|
||||||
@ -2222,7 +2221,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
|
|
||||||
user_1 = User(id=1).save()
|
user_1 = User(id=1).save()
|
||||||
user_2 = User(id=2).save()
|
user_2 = User(id=2).save()
|
||||||
book_1 = Book(id=1, author=user_2).save()
|
_ = Book(id=1, author=user_2).save()
|
||||||
book_2 = Book(id=2, author=user_1).save()
|
book_2 = Book(id=2, author=user_1).save()
|
||||||
|
|
||||||
user_2.delete()
|
user_2.delete()
|
||||||
@ -2231,7 +2230,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
assert Book.objects.get() == book_2
|
assert Book.objects.get() == book_2
|
||||||
|
|
||||||
user_3 = User(id=3).save()
|
user_3 = User(id=3).save()
|
||||||
book_3 = Book(id=3, author=user_3).save()
|
_ = Book(id=3, author=user_3).save()
|
||||||
|
|
||||||
user_3.delete()
|
user_3.delete()
|
||||||
# Deleting user_3 should also delete book_3
|
# Deleting user_3 should also delete book_3
|
||||||
@ -3205,7 +3204,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
def test_positional_creation(self):
|
def test_positional_creation(self):
|
||||||
"""Document cannot be instantiated using positional arguments."""
|
"""Document cannot be instantiated using positional arguments."""
|
||||||
with pytest.raises(TypeError) as exc_info:
|
with pytest.raises(TypeError) as exc_info:
|
||||||
person = self.Person("Test User", 42)
|
self.Person("Test User", 42)
|
||||||
|
|
||||||
expected_msg = (
|
expected_msg = (
|
||||||
"Instantiating a document with positional arguments is not "
|
"Instantiating a document with positional arguments is not "
|
||||||
@ -3274,7 +3273,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
|
|
||||||
def expand(self):
|
def expand(self):
|
||||||
self.flattened_parameter = {}
|
self.flattened_parameter = {}
|
||||||
for parameter_name, parameter in iteritems(self.parameters):
|
for parameter_name, parameter in self.parameters.items():
|
||||||
parameter.expand()
|
parameter.expand()
|
||||||
|
|
||||||
class NodesSystem(Document):
|
class NodesSystem(Document):
|
||||||
@ -3282,7 +3281,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
nodes = MapField(ReferenceField(Node, dbref=False))
|
nodes = MapField(ReferenceField(Node, dbref=False))
|
||||||
|
|
||||||
def save(self, *args, **kwargs):
|
def save(self, *args, **kwargs):
|
||||||
for node_name, node in iteritems(self.nodes):
|
for node_name, node in self.nodes.items():
|
||||||
node.expand()
|
node.expand()
|
||||||
node.save(*args, **kwargs)
|
node.save(*args, **kwargs)
|
||||||
super(NodesSystem, self).save(*args, **kwargs)
|
super(NodesSystem, self).save(*args, **kwargs)
|
||||||
@ -3607,13 +3606,13 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
v = StringField()
|
v = StringField()
|
||||||
|
|
||||||
class A(Document):
|
class A(Document):
|
||||||
l = ListField(EmbeddedDocumentField(B))
|
array = ListField(EmbeddedDocumentField(B))
|
||||||
|
|
||||||
A.objects.delete()
|
A.objects.delete()
|
||||||
A(l=[B(v="1"), B(v="2"), B(v="3")]).save()
|
A(array=[B(v="1"), B(v="2"), B(v="3")]).save()
|
||||||
a = A.objects.get()
|
a = A.objects.get()
|
||||||
assert a.l._instance == a
|
assert a.array._instance == a
|
||||||
for idx, b in enumerate(a.l):
|
for idx, b in enumerate(a.array):
|
||||||
assert b._instance == a
|
assert b._instance == a
|
||||||
assert idx == 2
|
assert idx == 2
|
||||||
|
|
||||||
|
@ -3,13 +3,12 @@ import uuid
|
|||||||
|
|
||||||
from bson import Binary
|
from bson import Binary
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
BIN_VALUE = six.b(
|
BIN_VALUE = "\xa9\xf3\x8d(\xd7\x03\x84\xb4k[\x0f\xe3\xa2\x19\x85p[J\xa3\xd2>\xde\xe6\x87\xb1\x7f\xc6\xe6\xd9r\x18\xf5".encode(
|
||||||
"\xa9\xf3\x8d(\xd7\x03\x84\xb4k[\x0f\xe3\xa2\x19\x85p[J\xa3\xd2>\xde\xe6\x87\xb1\x7f\xc6\xe6\xd9r\x18\xf5"
|
"latin-1"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -22,7 +21,7 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
content_type = StringField()
|
content_type = StringField()
|
||||||
blob = BinaryField()
|
blob = BinaryField()
|
||||||
|
|
||||||
BLOB = six.b("\xe6\x00\xc4\xff\x07")
|
BLOB = "\xe6\x00\xc4\xff\x07".encode("latin-1")
|
||||||
MIME_TYPE = "application/octet-stream"
|
MIME_TYPE = "application/octet-stream"
|
||||||
|
|
||||||
Attachment.drop_collection()
|
Attachment.drop_collection()
|
||||||
@ -32,7 +31,7 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
|
|
||||||
attachment_1 = Attachment.objects().first()
|
attachment_1 = Attachment.objects().first()
|
||||||
assert MIME_TYPE == attachment_1.content_type
|
assert MIME_TYPE == attachment_1.content_type
|
||||||
assert BLOB == six.binary_type(attachment_1.blob)
|
assert BLOB == bytes(attachment_1.blob)
|
||||||
|
|
||||||
def test_validation_succeeds(self):
|
def test_validation_succeeds(self):
|
||||||
"""Ensure that valid values can be assigned to binary fields.
|
"""Ensure that valid values can be assigned to binary fields.
|
||||||
@ -47,11 +46,11 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
attachment_required = AttachmentRequired()
|
attachment_required = AttachmentRequired()
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
attachment_required.validate()
|
attachment_required.validate()
|
||||||
attachment_required.blob = Binary(six.b("\xe6\x00\xc4\xff\x07"))
|
attachment_required.blob = Binary("\xe6\x00\xc4\xff\x07".encode("latin-1"))
|
||||||
attachment_required.validate()
|
attachment_required.validate()
|
||||||
|
|
||||||
_5_BYTES = six.b("\xe6\x00\xc4\xff\x07")
|
_5_BYTES = "\xe6\x00\xc4\xff\x07".encode("latin-1")
|
||||||
_4_BYTES = six.b("\xe6\x00\xc4\xff")
|
_4_BYTES = "\xe6\x00\xc4\xff".encode("latin-1")
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
AttachmentSizeLimit(blob=_5_BYTES).validate()
|
AttachmentSizeLimit(blob=_5_BYTES).validate()
|
||||||
AttachmentSizeLimit(blob=_4_BYTES).validate()
|
AttachmentSizeLimit(blob=_4_BYTES).validate()
|
||||||
@ -123,10 +122,7 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
upsert=True, new=True, set__bin_field=BIN_VALUE
|
upsert=True, new=True, set__bin_field=BIN_VALUE
|
||||||
)
|
)
|
||||||
assert doc.some_field == "test"
|
assert doc.some_field == "test"
|
||||||
if six.PY3:
|
assert doc.bin_field == BIN_VALUE
|
||||||
assert doc.bin_field == BIN_VALUE
|
|
||||||
else:
|
|
||||||
assert doc.bin_field == Binary(BIN_VALUE)
|
|
||||||
|
|
||||||
def test_update_one(self):
|
def test_update_one(self):
|
||||||
"""Ensures no regression of bug #1127"""
|
"""Ensures no regression of bug #1127"""
|
||||||
@ -136,7 +132,7 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
|
|
||||||
MyDocument.drop_collection()
|
MyDocument.drop_collection()
|
||||||
|
|
||||||
bin_data = six.b("\xe6\x00\xc4\xff\x07")
|
bin_data = "\xe6\x00\xc4\xff\x07".encode("latin-1")
|
||||||
doc = MyDocument(bin_field=bin_data).save()
|
doc = MyDocument(bin_field=bin_data).save()
|
||||||
|
|
||||||
n_updated = MyDocument.objects(bin_field=bin_data).update_one(
|
n_updated = MyDocument.objects(bin_field=bin_data).update_one(
|
||||||
@ -144,7 +140,4 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
)
|
)
|
||||||
assert n_updated == 1
|
assert n_updated == 1
|
||||||
fetched = MyDocument.objects.with_id(doc.id)
|
fetched = MyDocument.objects.with_id(doc.id)
|
||||||
if six.PY3:
|
assert fetched.bin_field == BIN_VALUE
|
||||||
assert fetched.bin_field == BIN_VALUE
|
|
||||||
else:
|
|
||||||
assert fetched.bin_field == Binary(BIN_VALUE)
|
|
||||||
|
@ -4,6 +4,8 @@ import itertools
|
|||||||
import math
|
import math
|
||||||
import re
|
import re
|
||||||
|
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
@ -65,7 +67,7 @@ class ComplexDateTimeFieldTest(MongoDBTestCase):
|
|||||||
for values in itertools.product([2014], mm, dd, hh, ii, ss, microsecond):
|
for values in itertools.product([2014], mm, dd, hh, ii, ss, microsecond):
|
||||||
stored = LogEntry(date=datetime.datetime(*values)).to_mongo()["date"]
|
stored = LogEntry(date=datetime.datetime(*values)).to_mongo()["date"]
|
||||||
assert (
|
assert (
|
||||||
re.match("^\d{4},\d{2},\d{2},\d{2},\d{2},\d{2},\d{6}$", stored)
|
re.match(r"^\d{4},\d{2},\d{2},\d{2},\d{2},\d{2},\d{6}$", stored)
|
||||||
is not None
|
is not None
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -74,7 +76,7 @@ class ComplexDateTimeFieldTest(MongoDBTestCase):
|
|||||||
"date_with_dots"
|
"date_with_dots"
|
||||||
]
|
]
|
||||||
assert (
|
assert (
|
||||||
re.match("^\d{4}.\d{2}.\d{2}.\d{2}.\d{2}.\d{2}.\d{6}$", stored) is not None
|
re.match(r"^\d{4}.\d{2}.\d{2}.\d{2}.\d{2}.\d{2}.\d{6}$", stored) is not None
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_complexdatetime_usage(self):
|
def test_complexdatetime_usage(self):
|
||||||
@ -191,3 +193,18 @@ class ComplexDateTimeFieldTest(MongoDBTestCase):
|
|||||||
|
|
||||||
fetched_log = Log.objects.with_id(log.id)
|
fetched_log = Log.objects.with_id(log.id)
|
||||||
assert fetched_log.timestamp >= NOW
|
assert fetched_log.timestamp >= NOW
|
||||||
|
|
||||||
|
def test_setting_bad_value_does_not_raise_unless_validate_is_called(self):
|
||||||
|
# test regression of #2253
|
||||||
|
|
||||||
|
class Log(Document):
|
||||||
|
timestamp = ComplexDateTimeField()
|
||||||
|
|
||||||
|
Log.drop_collection()
|
||||||
|
|
||||||
|
log = Log(timestamp="garbage")
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
log.validate()
|
||||||
|
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
log.save()
|
||||||
|
@ -2,7 +2,6 @@
|
|||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import dateutil
|
import dateutil
|
||||||
@ -89,17 +88,6 @@ class TestDateField(MongoDBTestCase):
|
|||||||
assert log.date == d1.date()
|
assert log.date == d1.date()
|
||||||
assert log.date == d2.date()
|
assert log.date == d2.date()
|
||||||
|
|
||||||
if not six.PY3:
|
|
||||||
# Pre UTC dates microseconds below 1000 are dropped
|
|
||||||
# This does not seem to be true in PY3
|
|
||||||
d1 = datetime.datetime(1969, 12, 31, 23, 59, 59, 999)
|
|
||||||
d2 = datetime.datetime(1969, 12, 31, 23, 59, 59)
|
|
||||||
log.date = d1
|
|
||||||
log.save()
|
|
||||||
log.reload()
|
|
||||||
assert log.date == d1.date()
|
|
||||||
assert log.date == d2.date()
|
|
||||||
|
|
||||||
def test_regular_usage(self):
|
def test_regular_usage(self):
|
||||||
"""Tests for regular datetime fields"""
|
"""Tests for regular datetime fields"""
|
||||||
|
|
||||||
|
@ -2,7 +2,6 @@
|
|||||||
import datetime as dt
|
import datetime as dt
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import dateutil
|
import dateutil
|
||||||
@ -98,17 +97,6 @@ class TestDateTimeField(MongoDBTestCase):
|
|||||||
assert log.date != d1
|
assert log.date != d1
|
||||||
assert log.date == d2
|
assert log.date == d2
|
||||||
|
|
||||||
if not six.PY3:
|
|
||||||
# Pre UTC dates microseconds below 1000 are dropped
|
|
||||||
# This does not seem to be true in PY3
|
|
||||||
d1 = dt.datetime(1969, 12, 31, 23, 59, 59, 999)
|
|
||||||
d2 = dt.datetime(1969, 12, 31, 23, 59, 59)
|
|
||||||
log.date = d1
|
|
||||||
log.save()
|
|
||||||
log.reload()
|
|
||||||
assert log.date != d1
|
|
||||||
assert log.date == d2
|
|
||||||
|
|
||||||
def test_regular_usage(self):
|
def test_regular_usage(self):
|
||||||
"""Tests for regular datetime fields"""
|
"""Tests for regular datetime fields"""
|
||||||
|
|
||||||
@ -213,7 +201,7 @@ class TestDateTimeField(MongoDBTestCase):
|
|||||||
# make sure that passing a parsable datetime works
|
# make sure that passing a parsable datetime works
|
||||||
dtd = DTDoc()
|
dtd = DTDoc()
|
||||||
dtd.date = date_str
|
dtd.date = date_str
|
||||||
assert isinstance(dtd.date, six.string_types)
|
assert isinstance(dtd.date, str)
|
||||||
dtd.save()
|
dtd.save()
|
||||||
dtd.reload()
|
dtd.reload()
|
||||||
|
|
||||||
|
@ -1,4 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
|
from bson import InvalidDocument
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
@ -19,22 +20,24 @@ class TestDictField(MongoDBTestCase):
|
|||||||
post = BlogPost(info=info).save()
|
post = BlogPost(info=info).save()
|
||||||
assert get_as_pymongo(post) == {"_id": post.id, "info": info}
|
assert get_as_pymongo(post) == {"_id": post.id, "info": info}
|
||||||
|
|
||||||
def test_general_things(self):
|
def test_validate_invalid_type(self):
|
||||||
"""Ensure that dict types work as expected."""
|
class BlogPost(Document):
|
||||||
|
info = DictField()
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
|
invalid_infos = ["my post", ["test", "test"], {1: "test"}]
|
||||||
|
for invalid_info in invalid_infos:
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
BlogPost(info=invalid_info).validate()
|
||||||
|
|
||||||
|
def test_keys_with_dots_or_dollars(self):
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
info = DictField()
|
info = DictField()
|
||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
post = BlogPost()
|
post = BlogPost()
|
||||||
post.info = "my post"
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
post.validate()
|
|
||||||
|
|
||||||
post.info = ["test", "test"]
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
post.validate()
|
|
||||||
|
|
||||||
post.info = {"$title": "test"}
|
post.info = {"$title": "test"}
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
@ -48,25 +51,34 @@ class TestDictField(MongoDBTestCase):
|
|||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
post.validate()
|
post.validate()
|
||||||
|
|
||||||
post.info = {1: "test"}
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
post.validate()
|
|
||||||
|
|
||||||
post.info = {"nested": {"the.title": "test"}}
|
post.info = {"nested": {"the.title": "test"}}
|
||||||
if get_mongodb_version() < MONGODB_36:
|
if get_mongodb_version() < MONGODB_36:
|
||||||
with pytest.raises(ValidationError):
|
# MongoDB < 3.6 rejects dots
|
||||||
post.validate()
|
# To avoid checking the mongodb version from the DictField class
|
||||||
|
# we rely on MongoDB to reject the data during the save
|
||||||
|
post.validate()
|
||||||
|
with pytest.raises(InvalidDocument):
|
||||||
|
post.save()
|
||||||
else:
|
else:
|
||||||
post.validate()
|
post.validate()
|
||||||
|
|
||||||
post.info = {"dollar_and_dot": {"te$st.test": "test"}}
|
post.info = {"dollar_and_dot": {"te$st.test": "test"}}
|
||||||
if get_mongodb_version() < MONGODB_36:
|
if get_mongodb_version() < MONGODB_36:
|
||||||
with pytest.raises(ValidationError):
|
post.validate()
|
||||||
post.validate()
|
with pytest.raises(InvalidDocument):
|
||||||
|
post.save()
|
||||||
else:
|
else:
|
||||||
post.validate()
|
post.validate()
|
||||||
|
|
||||||
post.info = {"title": "test"}
|
def test_general_things(self):
|
||||||
|
"""Ensure that dict types work as expected."""
|
||||||
|
|
||||||
|
class BlogPost(Document):
|
||||||
|
info = DictField()
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
|
post = BlogPost(info={"title": "test"})
|
||||||
post.save()
|
post.save()
|
||||||
|
|
||||||
post = BlogPost()
|
post = BlogPost()
|
||||||
|
@ -75,7 +75,7 @@ class TestEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
Person.objects(settings__notexist="bar").first()
|
Person.objects(settings__notexist="bar").first()
|
||||||
assert unicode(exc_info.value) == u'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
with pytest.raises(LookUpError):
|
with pytest.raises(LookUpError):
|
||||||
Person.objects.only("settings.notexist")
|
Person.objects.only("settings.notexist")
|
||||||
@ -111,7 +111,7 @@ class TestEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
assert Person.objects(settings__notexist="bar").first().id == p.id
|
assert Person.objects(settings__notexist="bar").first().id == p.id
|
||||||
assert unicode(exc_info.value) == u'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
# Test existing attribute
|
# Test existing attribute
|
||||||
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
||||||
@ -319,7 +319,7 @@ class TestGenericEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
Person.objects(settings__notexist="bar").first()
|
Person.objects(settings__notexist="bar").first()
|
||||||
assert unicode(exc_info.value) == u'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
with pytest.raises(LookUpError):
|
with pytest.raises(LookUpError):
|
||||||
Person.objects.only("settings.notexist")
|
Person.objects.only("settings.notexist")
|
||||||
@ -347,7 +347,7 @@ class TestGenericEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
assert Person.objects(settings__notexist="bar").first().id == p.id
|
assert Person.objects(settings__notexist="bar").first().id == p.id
|
||||||
assert unicode(exc_info.value) == u'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
# Test existing attribute
|
# Test existing attribute
|
||||||
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
||||||
|
@ -3,14 +3,13 @@ import copy
|
|||||||
import os
|
import os
|
||||||
import tempfile
|
import tempfile
|
||||||
import unittest
|
import unittest
|
||||||
|
from io import BytesIO
|
||||||
|
|
||||||
import gridfs
|
import gridfs
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.python_support import StringIO
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from PIL import Image
|
from PIL import Image
|
||||||
@ -30,7 +29,7 @@ TEST_IMAGE2_PATH = os.path.join(os.path.dirname(__file__), "mongodb_leaf.png")
|
|||||||
def get_file(path):
|
def get_file(path):
|
||||||
"""Use a BytesIO instead of a file to allow
|
"""Use a BytesIO instead of a file to allow
|
||||||
to have a one-liner and avoid that the file remains opened"""
|
to have a one-liner and avoid that the file remains opened"""
|
||||||
bytes_io = StringIO()
|
bytes_io = BytesIO()
|
||||||
with open(path, "rb") as f:
|
with open(path, "rb") as f:
|
||||||
bytes_io.write(f.read())
|
bytes_io.write(f.read())
|
||||||
bytes_io.seek(0)
|
bytes_io.seek(0)
|
||||||
@ -58,7 +57,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
PutFile.drop_collection()
|
PutFile.drop_collection()
|
||||||
|
|
||||||
text = six.b("Hello, World!")
|
text = "Hello, World!".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
putfile = PutFile()
|
putfile = PutFile()
|
||||||
@ -80,7 +79,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
PutFile.drop_collection()
|
PutFile.drop_collection()
|
||||||
|
|
||||||
putfile = PutFile()
|
putfile = PutFile()
|
||||||
putstring = StringIO()
|
putstring = BytesIO()
|
||||||
putstring.write(text)
|
putstring.write(text)
|
||||||
putstring.seek(0)
|
putstring.seek(0)
|
||||||
putfile.the_file.put(putstring, content_type=content_type)
|
putfile.the_file.put(putstring, content_type=content_type)
|
||||||
@ -101,8 +100,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
StreamFile.drop_collection()
|
StreamFile.drop_collection()
|
||||||
|
|
||||||
text = six.b("Hello, World!")
|
text = "Hello, World!".encode("latin-1")
|
||||||
more_text = six.b("Foo Bar")
|
more_text = "Foo Bar".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
streamfile = StreamFile()
|
streamfile = StreamFile()
|
||||||
@ -137,8 +136,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
StreamFile.drop_collection()
|
StreamFile.drop_collection()
|
||||||
|
|
||||||
text = six.b("Hello, World!")
|
text = "Hello, World!".encode("latin-1")
|
||||||
more_text = six.b("Foo Bar")
|
more_text = "Foo Bar".encode("latin-1")
|
||||||
|
|
||||||
streamfile = StreamFile()
|
streamfile = StreamFile()
|
||||||
streamfile.save()
|
streamfile.save()
|
||||||
@ -167,8 +166,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
class SetFile(Document):
|
class SetFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
|
|
||||||
text = six.b("Hello, World!")
|
text = "Hello, World!".encode("latin-1")
|
||||||
more_text = six.b("Foo Bar")
|
more_text = "Foo Bar".encode("latin-1")
|
||||||
|
|
||||||
SetFile.drop_collection()
|
SetFile.drop_collection()
|
||||||
|
|
||||||
@ -196,7 +195,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
GridDocument.drop_collection()
|
GridDocument.drop_collection()
|
||||||
|
|
||||||
with tempfile.TemporaryFile() as f:
|
with tempfile.TemporaryFile() as f:
|
||||||
f.write(six.b("Hello World!"))
|
f.write("Hello World!".encode("latin-1"))
|
||||||
f.flush()
|
f.flush()
|
||||||
|
|
||||||
# Test without default
|
# Test without default
|
||||||
@ -213,7 +212,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
assert doc_b.the_file.grid_id == doc_c.the_file.grid_id
|
assert doc_b.the_file.grid_id == doc_c.the_file.grid_id
|
||||||
|
|
||||||
# Test with default
|
# Test with default
|
||||||
doc_d = GridDocument(the_file=six.b(""))
|
doc_d = GridDocument(the_file="".encode("latin-1"))
|
||||||
doc_d.save()
|
doc_d.save()
|
||||||
|
|
||||||
doc_e = GridDocument.objects.with_id(doc_d.id)
|
doc_e = GridDocument.objects.with_id(doc_d.id)
|
||||||
@ -240,7 +239,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
# First instance
|
# First instance
|
||||||
test_file = TestFile()
|
test_file = TestFile()
|
||||||
test_file.name = "Hello, World!"
|
test_file.name = "Hello, World!"
|
||||||
test_file.the_file.put(six.b("Hello, World!"))
|
test_file.the_file.put("Hello, World!".encode("latin-1"))
|
||||||
test_file.save()
|
test_file.save()
|
||||||
|
|
||||||
# Second instance
|
# Second instance
|
||||||
@ -297,7 +296,9 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
test_file = TestFile()
|
test_file = TestFile()
|
||||||
assert not bool(test_file.the_file)
|
assert not bool(test_file.the_file)
|
||||||
test_file.the_file.put(six.b("Hello, World!"), content_type="text/plain")
|
test_file.the_file.put(
|
||||||
|
"Hello, World!".encode("latin-1"), content_type="text/plain"
|
||||||
|
)
|
||||||
test_file.save()
|
test_file.save()
|
||||||
assert bool(test_file.the_file)
|
assert bool(test_file.the_file)
|
||||||
|
|
||||||
@ -319,7 +320,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
class TestFile(Document):
|
class TestFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
|
|
||||||
text = six.b("Hello, World!")
|
text = "Hello, World!".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
testfile = TestFile()
|
testfile = TestFile()
|
||||||
@ -363,7 +364,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
testfile.the_file.put(text, content_type=content_type, filename="hello")
|
testfile.the_file.put(text, content_type=content_type, filename="hello")
|
||||||
testfile.save()
|
testfile.save()
|
||||||
|
|
||||||
text = six.b("Bonjour, World!")
|
text = "Bonjour, World!".encode("latin-1")
|
||||||
testfile.the_file.replace(text, content_type=content_type, filename="hello")
|
testfile.the_file.replace(text, content_type=content_type, filename="hello")
|
||||||
testfile.save()
|
testfile.save()
|
||||||
|
|
||||||
@ -387,7 +388,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
TestImage.drop_collection()
|
TestImage.drop_collection()
|
||||||
|
|
||||||
with tempfile.TemporaryFile() as f:
|
with tempfile.TemporaryFile() as f:
|
||||||
f.write(six.b("Hello World!"))
|
f.write("Hello World!".encode("latin-1"))
|
||||||
f.flush()
|
f.flush()
|
||||||
|
|
||||||
t = TestImage()
|
t = TestImage()
|
||||||
@ -503,21 +504,21 @@ class TestFileField(MongoDBTestCase):
|
|||||||
# First instance
|
# First instance
|
||||||
test_file = TestFile()
|
test_file = TestFile()
|
||||||
test_file.name = "Hello, World!"
|
test_file.name = "Hello, World!"
|
||||||
test_file.the_file.put(six.b("Hello, World!"), name="hello.txt")
|
test_file.the_file.put("Hello, World!".encode("latin-1"), name="hello.txt")
|
||||||
test_file.save()
|
test_file.save()
|
||||||
|
|
||||||
data = get_db("test_files").macumba.files.find_one()
|
data = get_db("test_files").macumba.files.find_one()
|
||||||
assert data.get("name") == "hello.txt"
|
assert data.get("name") == "hello.txt"
|
||||||
|
|
||||||
test_file = TestFile.objects.first()
|
test_file = TestFile.objects.first()
|
||||||
assert test_file.the_file.read() == six.b("Hello, World!")
|
assert test_file.the_file.read() == "Hello, World!".encode("latin-1")
|
||||||
|
|
||||||
test_file = TestFile.objects.first()
|
test_file = TestFile.objects.first()
|
||||||
test_file.the_file = six.b("HELLO, WORLD!")
|
test_file.the_file = "Hello, World!".encode("latin-1")
|
||||||
test_file.save()
|
test_file.save()
|
||||||
|
|
||||||
test_file = TestFile.objects.first()
|
test_file = TestFile.objects.first()
|
||||||
assert test_file.the_file.read() == six.b("HELLO, WORLD!")
|
assert test_file.the_file.read() == "Hello, World!".encode("latin-1")
|
||||||
|
|
||||||
def test_copyable(self):
|
def test_copyable(self):
|
||||||
class PutFile(Document):
|
class PutFile(Document):
|
||||||
@ -525,7 +526,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
PutFile.drop_collection()
|
PutFile.drop_collection()
|
||||||
|
|
||||||
text = six.b("Hello, World!")
|
text = "Hello, World!".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
putfile = PutFile()
|
putfile = PutFile()
|
||||||
|
@ -1,6 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
@ -52,9 +51,8 @@ class TestFloatField(MongoDBTestCase):
|
|||||||
|
|
||||||
big_person = BigPerson()
|
big_person = BigPerson()
|
||||||
|
|
||||||
for value, value_type in enumerate(six.integer_types):
|
big_person.height = int(0)
|
||||||
big_person.height = value_type(value)
|
big_person.validate()
|
||||||
big_person.validate()
|
|
||||||
|
|
||||||
big_person.height = 2 ** 500
|
big_person.height = 2 ** 500
|
||||||
big_person.validate()
|
big_person.validate()
|
||||||
|
@ -152,7 +152,7 @@ class TestLazyReferenceField(MongoDBTestCase):
|
|||||||
LazyReference(BadDoc, animal.pk),
|
LazyReference(BadDoc, animal.pk),
|
||||||
):
|
):
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
p = Ocurrence(person="test", animal=bad).save()
|
Ocurrence(person="test", animal=bad).save()
|
||||||
|
|
||||||
def test_lazy_reference_query_conversion(self):
|
def test_lazy_reference_query_conversion(self):
|
||||||
"""Ensure that LazyReferenceFields can be queried using objects and values
|
"""Ensure that LazyReferenceFields can be queried using objects and values
|
||||||
@ -386,7 +386,7 @@ class TestGenericLazyReferenceField(MongoDBTestCase):
|
|||||||
mineral = Mineral(name="Granite").save()
|
mineral = Mineral(name="Granite").save()
|
||||||
|
|
||||||
occ_animal = Ocurrence(living_thing=animal, thing=animal).save()
|
occ_animal = Ocurrence(living_thing=animal, thing=animal).save()
|
||||||
occ_vegetal = Ocurrence(living_thing=vegetal, thing=vegetal).save()
|
_ = Ocurrence(living_thing=vegetal, thing=vegetal).save()
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
Ocurrence(living_thing=mineral).save()
|
Ocurrence(living_thing=mineral).save()
|
||||||
|
|
||||||
@ -458,7 +458,7 @@ class TestGenericLazyReferenceField(MongoDBTestCase):
|
|||||||
baddoc = BadDoc().save()
|
baddoc = BadDoc().save()
|
||||||
for bad in (42, "foo", baddoc, LazyReference(BadDoc, animal.pk)):
|
for bad in (42, "foo", baddoc, LazyReference(BadDoc, animal.pk)):
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
p = Ocurrence(person="test", animal=bad).save()
|
Ocurrence(person="test", animal=bad).save()
|
||||||
|
|
||||||
def test_generic_lazy_reference_query_conversion(self):
|
def test_generic_lazy_reference_query_conversion(self):
|
||||||
class Member(Document):
|
class Member(Document):
|
||||||
|
@ -1,11 +1,5 @@
|
|||||||
# -*- coding: utf-8 -*-
|
from bson.int64 import Int64
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
|
|
||||||
try:
|
|
||||||
from bson.int64 import Int64
|
|
||||||
except ImportError:
|
|
||||||
Int64 = long
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
@ -28,7 +22,7 @@ class TestLongField(MongoDBTestCase):
|
|||||||
assert isinstance(
|
assert isinstance(
|
||||||
db.test_long_field_considered_as_int64.find()[0]["some_long"], Int64
|
db.test_long_field_considered_as_int64.find()[0]["some_long"], Int64
|
||||||
)
|
)
|
||||||
assert isinstance(doc.some_long, six.integer_types)
|
assert isinstance(doc.some_long, int)
|
||||||
|
|
||||||
def test_long_validation(self):
|
def test_long_validation(self):
|
||||||
"""Ensure that invalid values cannot be assigned to long fields.
|
"""Ensure that invalid values cannot be assigned to long fields.
|
||||||
|
@ -21,7 +21,7 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
|
|
||||||
ids = [i.id for i in Person.objects]
|
ids = [i.id for i in Person.objects]
|
||||||
assert ids == range(1, 11)
|
assert ids == list(range(1, 11))
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
@ -76,7 +76,7 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
|
|
||||||
ids = [i.id for i in Person.objects]
|
ids = [i.id for i in Person.objects]
|
||||||
assert ids == range(1, 11)
|
assert ids == list(range(1, 11))
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "jelly.id"})
|
c = self.db["mongoengine.counters"].find_one({"_id": "jelly.id"})
|
||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
@ -101,10 +101,10 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
|
|
||||||
ids = [i.id for i in Person.objects]
|
ids = [i.id for i in Person.objects]
|
||||||
assert ids == range(1, 11)
|
assert ids == list(range(1, 11))
|
||||||
|
|
||||||
counters = [i.counter for i in Person.objects]
|
counters = [i.counter for i in Person.objects]
|
||||||
assert counters == range(1, 11)
|
assert counters == list(range(1, 11))
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
@ -166,10 +166,10 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
|
|
||||||
ids = [i.id for i in Person.objects]
|
ids = [i.id for i in Person.objects]
|
||||||
assert ids == range(1, 11)
|
assert ids == list(range(1, 11))
|
||||||
|
|
||||||
id = [i.id for i in Animal.objects]
|
id = [i.id for i in Animal.objects]
|
||||||
assert id == range(1, 11)
|
assert id == list(range(1, 11))
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
@ -193,7 +193,7 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
|
|
||||||
ids = [i.id for i in Person.objects]
|
ids = [i.id for i in Person.objects]
|
||||||
assert ids == map(str, range(1, 11))
|
assert ids == [str(i) for i in range(1, 11)]
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
@ -267,12 +267,12 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
foo = Foo(name="Foo")
|
foo = Foo(name="Foo")
|
||||||
foo.save()
|
foo.save()
|
||||||
|
|
||||||
assert not (
|
assert "base.counter" not in self.db["mongoengine.counters"].find().distinct(
|
||||||
"base.counter" in self.db["mongoengine.counters"].find().distinct("_id")
|
"_id"
|
||||||
)
|
)
|
||||||
assert ("foo.counter" and "bar.counter") in self.db[
|
existing_counters = self.db["mongoengine.counters"].find().distinct("_id")
|
||||||
"mongoengine.counters"
|
assert "foo.counter" in existing_counters
|
||||||
].find().distinct("_id")
|
assert "bar.counter" in existing_counters
|
||||||
assert foo.counter == bar.counter
|
assert foo.counter == bar.counter
|
||||||
assert foo._fields["counter"].owner_document == Foo
|
assert foo._fields["counter"].owner_document == Foo
|
||||||
assert bar._fields["counter"].owner_document == Bar
|
assert bar._fields["counter"].owner_document == Bar
|
||||||
|
@ -2,7 +2,6 @@
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@ -35,7 +34,7 @@ class TestURLField(MongoDBTestCase):
|
|||||||
with pytest.raises(ValidationError) as exc_info:
|
with pytest.raises(ValidationError) as exc_info:
|
||||||
link.validate()
|
link.validate()
|
||||||
assert (
|
assert (
|
||||||
unicode(exc_info.value)
|
str(exc_info.value)
|
||||||
== u"ValidationError (Link:None) (Invalid URL: http://\u043f\u0440\u0438\u0432\u0435\u0442.com: ['url'])"
|
== u"ValidationError (Link:None) (Invalid URL: http://\u043f\u0440\u0438\u0432\u0435\u0442.com: ['url'])"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -7,11 +7,10 @@ from decimal import Decimal
|
|||||||
|
|
||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
import pymongo
|
import pymongo
|
||||||
|
from pymongo.read_concern import ReadConcern
|
||||||
from pymongo.read_preferences import ReadPreference
|
from pymongo.read_preferences import ReadPreference
|
||||||
from pymongo.results import UpdateResult
|
from pymongo.results import UpdateResult
|
||||||
import pytest
|
import pytest
|
||||||
import six
|
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
@ -110,15 +109,47 @@ class TestQueryset(unittest.TestCase):
|
|||||||
# Filter people by age
|
# Filter people by age
|
||||||
people = self.Person.objects(age=20)
|
people = self.Person.objects(age=20)
|
||||||
assert people.count() == 1
|
assert people.count() == 1
|
||||||
person = people.next()
|
person = next(people)
|
||||||
assert person == user_a
|
assert person == user_a
|
||||||
assert person.name == "User A"
|
assert person.name == "User A"
|
||||||
assert person.age == 20
|
assert person.age == 20
|
||||||
|
|
||||||
|
def test_slicing_sets_empty_limit_skip(self):
|
||||||
|
self.Person.objects.insert(
|
||||||
|
[self.Person(name="User {}".format(i), age=i) for i in range(5)],
|
||||||
|
load_bulk=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.Person.objects.create(name="User B", age=30)
|
||||||
|
self.Person.objects.create(name="User C", age=40)
|
||||||
|
|
||||||
|
qs = self.Person.objects()[1:2]
|
||||||
|
assert (qs._empty, qs._skip, qs._limit) == (False, 1, 1)
|
||||||
|
assert len(list(qs)) == 1
|
||||||
|
|
||||||
|
# Test edge case of [1:1] which should return nothing
|
||||||
|
# and require a hack so that it doesn't clash with limit(0)
|
||||||
|
qs = self.Person.objects()[1:1]
|
||||||
|
assert (qs._empty, qs._skip, qs._limit) == (True, 1, 0)
|
||||||
|
assert len(list(qs)) == 0
|
||||||
|
|
||||||
|
qs2 = qs[1:5] # Make sure that further slicing resets _empty
|
||||||
|
assert (qs2._empty, qs2._skip, qs2._limit) == (False, 1, 4)
|
||||||
|
assert len(list(qs2)) == 4
|
||||||
|
|
||||||
|
def test_limit_0_returns_all_documents(self):
|
||||||
|
self.Person.objects.create(name="User A", age=20)
|
||||||
|
self.Person.objects.create(name="User B", age=30)
|
||||||
|
|
||||||
|
n_docs = self.Person.objects().count()
|
||||||
|
|
||||||
|
persons = list(self.Person.objects().limit(0))
|
||||||
|
assert len(persons) == 2 == n_docs
|
||||||
|
|
||||||
def test_limit(self):
|
def test_limit(self):
|
||||||
"""Ensure that QuerySet.limit works as expected."""
|
"""Ensure that QuerySet.limit works as expected."""
|
||||||
user_a = self.Person.objects.create(name="User A", age=20)
|
user_a = self.Person.objects.create(name="User A", age=20)
|
||||||
user_b = self.Person.objects.create(name="User B", age=30)
|
_ = self.Person.objects.create(name="User B", age=30)
|
||||||
|
|
||||||
# Test limit on a new queryset
|
# Test limit on a new queryset
|
||||||
people = list(self.Person.objects.limit(1))
|
people = list(self.Person.objects.limit(1))
|
||||||
@ -150,6 +181,11 @@ class TestQueryset(unittest.TestCase):
|
|||||||
user_b = self.Person.objects.create(name="User B", age=30)
|
user_b = self.Person.objects.create(name="User B", age=30)
|
||||||
|
|
||||||
# Test skip on a new queryset
|
# Test skip on a new queryset
|
||||||
|
people = list(self.Person.objects.skip(0))
|
||||||
|
assert len(people) == 2
|
||||||
|
assert people[0] == user_a
|
||||||
|
assert people[1] == user_b
|
||||||
|
|
||||||
people = list(self.Person.objects.skip(1))
|
people = list(self.Person.objects.skip(1))
|
||||||
assert len(people) == 1
|
assert len(people) == 1
|
||||||
assert people[0] == user_b
|
assert people[0] == user_b
|
||||||
@ -274,32 +310,47 @@ class TestQueryset(unittest.TestCase):
|
|||||||
with pytest.raises(InvalidQueryError):
|
with pytest.raises(InvalidQueryError):
|
||||||
self.Person.objects(name="User A").with_id(person1.id)
|
self.Person.objects(name="User A").with_id(person1.id)
|
||||||
|
|
||||||
def test_find_only_one(self):
|
def test_get_no_document_exists_raises_doesnotexist(self):
|
||||||
"""Ensure that a query using ``get`` returns at most one result.
|
assert self.Person.objects.count() == 0
|
||||||
"""
|
|
||||||
# Try retrieving when no objects exists
|
# Try retrieving when no objects exists
|
||||||
with pytest.raises(DoesNotExist):
|
with pytest.raises(DoesNotExist):
|
||||||
self.Person.objects.get()
|
self.Person.objects.get()
|
||||||
with pytest.raises(self.Person.DoesNotExist):
|
with pytest.raises(self.Person.DoesNotExist):
|
||||||
self.Person.objects.get()
|
self.Person.objects.get()
|
||||||
|
|
||||||
|
def test_get_multiple_match_raises_multipleobjectsreturned(self):
|
||||||
|
"""Ensure that a query using ``get`` returns at most one result.
|
||||||
|
"""
|
||||||
|
assert self.Person.objects().count() == 0
|
||||||
|
|
||||||
person1 = self.Person(name="User A", age=20)
|
person1 = self.Person(name="User A", age=20)
|
||||||
person1.save()
|
person1.save()
|
||||||
person2 = self.Person(name="User B", age=30)
|
|
||||||
|
p = self.Person.objects.get()
|
||||||
|
assert p == person1
|
||||||
|
|
||||||
|
person2 = self.Person(name="User B", age=20)
|
||||||
person2.save()
|
person2.save()
|
||||||
|
|
||||||
# Retrieve the first person from the database
|
person3 = self.Person(name="User C", age=30)
|
||||||
|
person3.save()
|
||||||
|
|
||||||
|
# .get called without argument
|
||||||
with pytest.raises(MultipleObjectsReturned):
|
with pytest.raises(MultipleObjectsReturned):
|
||||||
self.Person.objects.get()
|
self.Person.objects.get()
|
||||||
with pytest.raises(self.Person.MultipleObjectsReturned):
|
with pytest.raises(self.Person.MultipleObjectsReturned):
|
||||||
self.Person.objects.get()
|
self.Person.objects.get()
|
||||||
|
|
||||||
|
# check filtering
|
||||||
|
with pytest.raises(MultipleObjectsReturned):
|
||||||
|
self.Person.objects.get(age__lt=30)
|
||||||
|
with pytest.raises(MultipleObjectsReturned) as exc_info:
|
||||||
|
self.Person.objects(age__lt=30).get()
|
||||||
|
assert "2 or more items returned, instead of 1" == str(exc_info.value)
|
||||||
|
|
||||||
# Use a query to filter the people found to just person2
|
# Use a query to filter the people found to just person2
|
||||||
person = self.Person.objects.get(age=30)
|
person = self.Person.objects.get(age=30)
|
||||||
assert person.name == "User B"
|
assert person == person3
|
||||||
|
|
||||||
person = self.Person.objects.get(age__lt=30)
|
|
||||||
assert person.name == "User A"
|
|
||||||
|
|
||||||
def test_find_array_position(self):
|
def test_find_array_position(self):
|
||||||
"""Ensure that query by array position works.
|
"""Ensure that query by array position works.
|
||||||
@ -358,6 +409,9 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
assert list(A.objects.none()) == []
|
assert list(A.objects.none()) == []
|
||||||
assert list(A.objects.none().all()) == []
|
assert list(A.objects.none().all()) == []
|
||||||
|
assert list(A.objects.none().limit(1)) == []
|
||||||
|
assert list(A.objects.none().skip(1)) == []
|
||||||
|
assert list(A.objects.none()[:5]) == []
|
||||||
|
|
||||||
def test_chaining(self):
|
def test_chaining(self):
|
||||||
class A(Document):
|
class A(Document):
|
||||||
@ -2573,13 +2627,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
age = IntField()
|
age = IntField()
|
||||||
|
|
||||||
with db_ops_tracker() as q:
|
with db_ops_tracker() as q:
|
||||||
adult1 = (
|
User.objects.filter(age__gte=18).comment("looking for an adult").first()
|
||||||
User.objects.filter(age__gte=18).comment("looking for an adult").first()
|
User.objects.comment("looking for an adult").filter(age__gte=18).first()
|
||||||
)
|
|
||||||
|
|
||||||
adult2 = (
|
|
||||||
User.objects.comment("looking for an adult").filter(age__gte=18).first()
|
|
||||||
)
|
|
||||||
|
|
||||||
ops = q.get_ops()
|
ops = q.get_ops()
|
||||||
assert len(ops) == 2
|
assert len(ops) == 2
|
||||||
@ -2768,7 +2817,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
# start a map/reduce
|
# start a map/reduce
|
||||||
cursor.next()
|
next(cursor)
|
||||||
|
|
||||||
results = Person.objects.map_reduce(
|
results = Person.objects.map_reduce(
|
||||||
map_f=map_person,
|
map_f=map_person,
|
||||||
@ -4007,6 +4056,32 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
Number.drop_collection()
|
Number.drop_collection()
|
||||||
|
|
||||||
|
def test_clone_retains_settings(self):
|
||||||
|
"""Ensure that cloning retains the read_preference and read_concern
|
||||||
|
"""
|
||||||
|
|
||||||
|
class Number(Document):
|
||||||
|
n = IntField()
|
||||||
|
|
||||||
|
Number.drop_collection()
|
||||||
|
|
||||||
|
qs = Number.objects
|
||||||
|
qs_clone = qs.clone()
|
||||||
|
assert qs._read_preference == qs_clone._read_preference
|
||||||
|
assert qs._read_concern == qs_clone._read_concern
|
||||||
|
|
||||||
|
qs = Number.objects.read_preference(ReadPreference.PRIMARY_PREFERRED)
|
||||||
|
qs_clone = qs.clone()
|
||||||
|
assert qs._read_preference == ReadPreference.PRIMARY_PREFERRED
|
||||||
|
assert qs._read_preference == qs_clone._read_preference
|
||||||
|
|
||||||
|
qs = Number.objects.read_concern({"level": "majority"})
|
||||||
|
qs_clone = qs.clone()
|
||||||
|
assert qs._read_concern.document == {"level": "majority"}
|
||||||
|
assert qs._read_concern == qs_clone._read_concern
|
||||||
|
|
||||||
|
Number.drop_collection()
|
||||||
|
|
||||||
def test_using(self):
|
def test_using(self):
|
||||||
"""Ensure that switching databases for a queryset is possible
|
"""Ensure that switching databases for a queryset is possible
|
||||||
"""
|
"""
|
||||||
@ -4093,7 +4168,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
info = Comment.objects._collection.index_information()
|
info = Comment.objects._collection.index_information()
|
||||||
info = [
|
info = [
|
||||||
(value["key"], value.get("unique", False), value.get("sparse", False))
|
(value["key"], value.get("unique", False), value.get("sparse", False))
|
||||||
for key, value in iteritems(info)
|
for key, value in info.items()
|
||||||
]
|
]
|
||||||
assert ([("_cls", 1), ("message", 1)], False, False) in info
|
assert ([("_cls", 1), ("message", 1)], False, False) in info
|
||||||
|
|
||||||
@ -4395,7 +4470,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
# Use a query to filter the people found to just person1
|
# Use a query to filter the people found to just person1
|
||||||
people = self.Person.objects(age=20).scalar("name")
|
people = self.Person.objects(age=20).scalar("name")
|
||||||
assert people.count() == 1
|
assert people.count() == 1
|
||||||
person = people.next()
|
person = next(people)
|
||||||
assert person == "User A"
|
assert person == "User A"
|
||||||
|
|
||||||
# Test limit
|
# Test limit
|
||||||
@ -4428,7 +4503,9 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert len(people) == 1
|
assert len(people) == 1
|
||||||
assert people[0] == "User B"
|
assert people[0] == "User B"
|
||||||
|
|
||||||
people = list(self.Person.objects[1:1].scalar("name"))
|
# people = list(self.Person.objects[1:1].scalar("name"))
|
||||||
|
people = self.Person.objects[1:1]
|
||||||
|
people = people.scalar("name")
|
||||||
assert len(people) == 0
|
assert len(people) == 0
|
||||||
|
|
||||||
# Test slice out of range
|
# Test slice out of range
|
||||||
@ -4445,24 +4522,14 @@ class TestQueryset(unittest.TestCase):
|
|||||||
"A0" == "%s" % self.Person.objects.order_by("name").scalar("name").first()
|
"A0" == "%s" % self.Person.objects.order_by("name").scalar("name").first()
|
||||||
)
|
)
|
||||||
assert "A0" == "%s" % self.Person.objects.scalar("name").order_by("name")[0]
|
assert "A0" == "%s" % self.Person.objects.scalar("name").order_by("name")[0]
|
||||||
if six.PY3:
|
assert (
|
||||||
assert (
|
"['A1', 'A2']"
|
||||||
"['A1', 'A2']"
|
== "%s" % self.Person.objects.order_by("age").scalar("name")[1:3]
|
||||||
== "%s" % self.Person.objects.order_by("age").scalar("name")[1:3]
|
)
|
||||||
)
|
assert (
|
||||||
assert (
|
"['A51', 'A52']"
|
||||||
"['A51', 'A52']"
|
== "%s" % self.Person.objects.order_by("age").scalar("name")[51:53]
|
||||||
== "%s" % self.Person.objects.order_by("age").scalar("name")[51:53]
|
)
|
||||||
)
|
|
||||||
else:
|
|
||||||
assert (
|
|
||||||
"[u'A1', u'A2']"
|
|
||||||
== "%s" % self.Person.objects.order_by("age").scalar("name")[1:3]
|
|
||||||
)
|
|
||||||
assert (
|
|
||||||
"[u'A51', u'A52']"
|
|
||||||
== "%s" % self.Person.objects.order_by("age").scalar("name")[51:53]
|
|
||||||
)
|
|
||||||
|
|
||||||
# with_id and in_bulk
|
# with_id and in_bulk
|
||||||
person = self.Person.objects.order_by("name").first()
|
person = self.Person.objects.order_by("name").first()
|
||||||
@ -4470,12 +4537,77 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
pks = self.Person.objects.order_by("age").scalar("pk")[1:3]
|
pks = self.Person.objects.order_by("age").scalar("pk")[1:3]
|
||||||
names = self.Person.objects.scalar("name").in_bulk(list(pks)).values()
|
names = self.Person.objects.scalar("name").in_bulk(list(pks)).values()
|
||||||
if six.PY3:
|
expected = "['A1', 'A2']"
|
||||||
expected = "['A1', 'A2']"
|
|
||||||
else:
|
|
||||||
expected = "[u'A1', u'A2']"
|
|
||||||
assert expected == "%s" % sorted(names)
|
assert expected == "%s" % sorted(names)
|
||||||
|
|
||||||
|
def test_fields(self):
|
||||||
|
class Bar(EmbeddedDocument):
|
||||||
|
v = StringField()
|
||||||
|
z = StringField()
|
||||||
|
|
||||||
|
class Foo(Document):
|
||||||
|
x = StringField()
|
||||||
|
y = IntField()
|
||||||
|
items = EmbeddedDocumentListField(Bar)
|
||||||
|
|
||||||
|
Foo.drop_collection()
|
||||||
|
|
||||||
|
Foo(x="foo1", y=1).save()
|
||||||
|
Foo(x="foo2", y=2, items=[]).save()
|
||||||
|
Foo(x="foo3", y=3, items=[Bar(z="a", v="V")]).save()
|
||||||
|
Foo(
|
||||||
|
x="foo4",
|
||||||
|
y=4,
|
||||||
|
items=[
|
||||||
|
Bar(z="a", v="V"),
|
||||||
|
Bar(z="b", v="W"),
|
||||||
|
Bar(z="b", v="X"),
|
||||||
|
Bar(z="c", v="V"),
|
||||||
|
],
|
||||||
|
).save()
|
||||||
|
Foo(
|
||||||
|
x="foo5",
|
||||||
|
y=5,
|
||||||
|
items=[
|
||||||
|
Bar(z="b", v="X"),
|
||||||
|
Bar(z="c", v="V"),
|
||||||
|
Bar(z="d", v="V"),
|
||||||
|
Bar(z="e", v="V"),
|
||||||
|
],
|
||||||
|
).save()
|
||||||
|
|
||||||
|
foos_with_x = list(Foo.objects.order_by("y").fields(x=1))
|
||||||
|
|
||||||
|
assert all(o.x is not None for o in foos_with_x)
|
||||||
|
|
||||||
|
foos_without_y = list(Foo.objects.order_by("y").fields(y=0))
|
||||||
|
|
||||||
|
assert all(o.y is None for o in foos_without_y)
|
||||||
|
|
||||||
|
foos_with_sliced_items = list(Foo.objects.order_by("y").fields(slice__items=1))
|
||||||
|
|
||||||
|
assert foos_with_sliced_items[0].items == []
|
||||||
|
assert foos_with_sliced_items[1].items == []
|
||||||
|
assert len(foos_with_sliced_items[2].items) == 1
|
||||||
|
assert foos_with_sliced_items[2].items[0].z == "a"
|
||||||
|
assert len(foos_with_sliced_items[3].items) == 1
|
||||||
|
assert foos_with_sliced_items[3].items[0].z == "a"
|
||||||
|
assert len(foos_with_sliced_items[4].items) == 1
|
||||||
|
assert foos_with_sliced_items[4].items[0].z == "b"
|
||||||
|
|
||||||
|
foos_with_elem_match_items = list(
|
||||||
|
Foo.objects.order_by("y").fields(elemMatch__items={"z": "b"})
|
||||||
|
)
|
||||||
|
|
||||||
|
assert foos_with_elem_match_items[0].items == []
|
||||||
|
assert foos_with_elem_match_items[1].items == []
|
||||||
|
assert foos_with_elem_match_items[2].items == []
|
||||||
|
assert len(foos_with_elem_match_items[3].items) == 1
|
||||||
|
assert foos_with_elem_match_items[3].items[0].z == "b"
|
||||||
|
assert foos_with_elem_match_items[3].items[0].v == "W"
|
||||||
|
assert len(foos_with_elem_match_items[4].items) == 1
|
||||||
|
assert foos_with_elem_match_items[4].items[0].z == "b"
|
||||||
|
|
||||||
def test_elem_match(self):
|
def test_elem_match(self):
|
||||||
class Foo(EmbeddedDocument):
|
class Foo(EmbeddedDocument):
|
||||||
shape = StringField()
|
shape = StringField()
|
||||||
@ -4658,6 +4790,46 @@ class TestQueryset(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
assert_read_pref(bars, ReadPreference.SECONDARY_PREFERRED)
|
assert_read_pref(bars, ReadPreference.SECONDARY_PREFERRED)
|
||||||
|
|
||||||
|
def test_read_concern(self):
|
||||||
|
class Bar(Document):
|
||||||
|
txt = StringField()
|
||||||
|
|
||||||
|
meta = {"indexes": ["txt"]}
|
||||||
|
|
||||||
|
Bar.drop_collection()
|
||||||
|
bar = Bar.objects.create(txt="xyz")
|
||||||
|
|
||||||
|
bars = list(Bar.objects.read_concern(None))
|
||||||
|
assert bars == [bar]
|
||||||
|
|
||||||
|
bars = Bar.objects.read_concern({"level": "local"})
|
||||||
|
assert bars._read_concern.document == {"level": "local"}
|
||||||
|
assert bars._cursor.collection.read_concern.document == {"level": "local"}
|
||||||
|
|
||||||
|
# Make sure that `.read_concern(...)` does not accept string values.
|
||||||
|
with pytest.raises(TypeError):
|
||||||
|
Bar.objects.read_concern("local")
|
||||||
|
|
||||||
|
def assert_read_concern(qs, expected_read_concern):
|
||||||
|
assert qs._read_concern.document == expected_read_concern
|
||||||
|
assert qs._cursor.collection.read_concern.document == expected_read_concern
|
||||||
|
|
||||||
|
# Make sure read concern is respected after a `.skip(...)`.
|
||||||
|
bars = Bar.objects.skip(1).read_concern({"level": "local"})
|
||||||
|
assert_read_concern(bars, {"level": "local"})
|
||||||
|
|
||||||
|
# Make sure read concern is respected after a `.limit(...)`.
|
||||||
|
bars = Bar.objects.limit(1).read_concern({"level": "local"})
|
||||||
|
assert_read_concern(bars, {"level": "local"})
|
||||||
|
|
||||||
|
# Make sure read concern is respected after an `.order_by(...)`.
|
||||||
|
bars = Bar.objects.order_by("txt").read_concern({"level": "local"})
|
||||||
|
assert_read_concern(bars, {"level": "local"})
|
||||||
|
|
||||||
|
# Make sure read concern is respected after a `.hint(...)`.
|
||||||
|
bars = Bar.objects.hint([("txt", 1)]).read_concern({"level": "majority"})
|
||||||
|
assert_read_concern(bars, {"level": "majority"})
|
||||||
|
|
||||||
def test_json_simple(self):
|
def test_json_simple(self):
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
string = StringField()
|
string = StringField()
|
||||||
@ -5294,7 +5466,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
if not test:
|
if not test:
|
||||||
raise AssertionError("Cursor has data and returned False")
|
raise AssertionError("Cursor has data and returned False")
|
||||||
|
|
||||||
queryset.next()
|
next(queryset)
|
||||||
if not queryset:
|
if not queryset:
|
||||||
raise AssertionError(
|
raise AssertionError(
|
||||||
"Cursor has data and it must returns True, even in the last item."
|
"Cursor has data and it must returns True, even in the last item."
|
||||||
@ -5527,7 +5699,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
self.Person.objects.create(name="Baz")
|
self.Person.objects.create(name="Baz")
|
||||||
assert self.Person.objects.count(with_limit_and_skip=True) == 3
|
assert self.Person.objects.count(with_limit_and_skip=True) == 3
|
||||||
|
|
||||||
newPerson = self.Person.objects.create(name="Foo_1")
|
self.Person.objects.create(name="Foo_1")
|
||||||
assert self.Person.objects.count(with_limit_and_skip=True) == 4
|
assert self.Person.objects.count(with_limit_and_skip=True) == 4
|
||||||
|
|
||||||
def test_no_cursor_timeout(self):
|
def test_no_cursor_timeout(self):
|
||||||
|
@ -24,6 +24,12 @@ class TestTransform(unittest.TestCase):
|
|||||||
}
|
}
|
||||||
assert transform.query(friend__age__gte=30) == {"friend.age": {"$gte": 30}}
|
assert transform.query(friend__age__gte=30) == {"friend.age": {"$gte": 30}}
|
||||||
assert transform.query(name__exists=True) == {"name": {"$exists": True}}
|
assert transform.query(name__exists=True) == {"name": {"$exists": True}}
|
||||||
|
assert transform.query(name=["Mark"], __raw__={"name": {"$in": "Tom"}}) == {
|
||||||
|
"$and": [{"name": ["Mark"]}, {"name": {"$in": "Tom"}}]
|
||||||
|
}
|
||||||
|
assert transform.query(name__in=["Tom"], __raw__={"name": "Mark"}) == {
|
||||||
|
"$and": [{"name": {"$in": ["Tom"]}}, {"name": "Mark"}]
|
||||||
|
}
|
||||||
|
|
||||||
def test_transform_update(self):
|
def test_transform_update(self):
|
||||||
class LisDoc(Document):
|
class LisDoc(Document):
|
||||||
@ -338,6 +344,31 @@ class TestTransform(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
assert update == {"$pull": {"content.text": {"word": {"$nin": ["foo", "bar"]}}}}
|
assert update == {"$pull": {"content.text": {"word": {"$nin": ["foo", "bar"]}}}}
|
||||||
|
|
||||||
|
def test_transform_embedded_document_list_fields(self):
|
||||||
|
"""
|
||||||
|
Test added to check filtering
|
||||||
|
EmbeddedDocumentListField which is inside a EmbeddedDocumentField
|
||||||
|
"""
|
||||||
|
|
||||||
|
class Drink(EmbeddedDocument):
|
||||||
|
id = StringField()
|
||||||
|
meta = {"strict": False}
|
||||||
|
|
||||||
|
class Shop(Document):
|
||||||
|
drinks = EmbeddedDocumentListField(Drink)
|
||||||
|
|
||||||
|
Shop.drop_collection()
|
||||||
|
drinks = [Drink(id="drink_1"), Drink(id="drink_2")]
|
||||||
|
Shop.objects.create(drinks=drinks)
|
||||||
|
q_obj = transform.query(
|
||||||
|
Shop, drinks__all=[{"$elemMatch": {"_id": x.id}} for x in drinks]
|
||||||
|
)
|
||||||
|
assert q_obj == {
|
||||||
|
"drinks": {"$all": [{"$elemMatch": {"_id": x.id}} for x in drinks]}
|
||||||
|
}
|
||||||
|
|
||||||
|
Shop.drop_collection()
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@ -348,7 +348,7 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
|
|
||||||
def test_disconnect_cleans_cached_collection_attribute_in_document(self):
|
def test_disconnect_cleans_cached_collection_attribute_in_document(self):
|
||||||
"""Ensure that the disconnect() method works properly"""
|
"""Ensure that the disconnect() method works properly"""
|
||||||
conn1 = connect("mongoenginetest")
|
connect("mongoenginetest")
|
||||||
|
|
||||||
class History(Document):
|
class History(Document):
|
||||||
pass
|
pass
|
||||||
@ -518,7 +518,7 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
"""Ensure connect() uses the username & password params if the URI
|
"""Ensure connect() uses the username & password params if the URI
|
||||||
doesn't explicitly specify them.
|
doesn't explicitly specify them.
|
||||||
"""
|
"""
|
||||||
c = connect(
|
connect(
|
||||||
host="mongodb://localhost/mongoenginetest", username="user", password="pass"
|
host="mongodb://localhost/mongoenginetest", username="user", password="pass"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -632,7 +632,7 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
"""Ensure connect() works when specifying a replicaSet via the
|
"""Ensure connect() works when specifying a replicaSet via the
|
||||||
MongoDB URI.
|
MongoDB URI.
|
||||||
"""
|
"""
|
||||||
c = connect(host="mongodb://localhost/test?replicaSet=local-rs")
|
connect(host="mongodb://localhost/test?replicaSet=local-rs")
|
||||||
db = get_db()
|
db = get_db()
|
||||||
assert isinstance(db, pymongo.database.Database)
|
assert isinstance(db, pymongo.database.Database)
|
||||||
assert db.name == "test"
|
assert db.name == "test"
|
||||||
|
@ -8,6 +8,8 @@ from mongoengine.context_managers import (
|
|||||||
no_dereference,
|
no_dereference,
|
||||||
no_sub_classes,
|
no_sub_classes,
|
||||||
query_counter,
|
query_counter,
|
||||||
|
set_read_write_concern,
|
||||||
|
set_write_concern,
|
||||||
switch_collection,
|
switch_collection,
|
||||||
switch_db,
|
switch_db,
|
||||||
)
|
)
|
||||||
@ -15,6 +17,52 @@ from mongoengine.pymongo_support import count_documents
|
|||||||
|
|
||||||
|
|
||||||
class TestContextManagers:
|
class TestContextManagers:
|
||||||
|
def test_set_write_concern(self):
|
||||||
|
connect("mongoenginetest")
|
||||||
|
|
||||||
|
class User(Document):
|
||||||
|
name = StringField()
|
||||||
|
|
||||||
|
collection = User._get_collection()
|
||||||
|
original_write_concern = collection.write_concern
|
||||||
|
|
||||||
|
with set_write_concern(
|
||||||
|
collection, {"w": "majority", "j": True, "wtimeout": 1234}
|
||||||
|
) as updated_collection:
|
||||||
|
assert updated_collection.write_concern.document == {
|
||||||
|
"w": "majority",
|
||||||
|
"j": True,
|
||||||
|
"wtimeout": 1234,
|
||||||
|
}
|
||||||
|
|
||||||
|
assert original_write_concern.document == collection.write_concern.document
|
||||||
|
|
||||||
|
def test_set_read_write_concern(self):
|
||||||
|
connect("mongoenginetest")
|
||||||
|
|
||||||
|
class User(Document):
|
||||||
|
name = StringField()
|
||||||
|
|
||||||
|
collection = User._get_collection()
|
||||||
|
|
||||||
|
original_read_concern = collection.read_concern
|
||||||
|
original_write_concern = collection.write_concern
|
||||||
|
|
||||||
|
with set_read_write_concern(
|
||||||
|
collection,
|
||||||
|
{"w": "majority", "j": True, "wtimeout": 1234},
|
||||||
|
{"level": "local"},
|
||||||
|
) as update_collection:
|
||||||
|
assert update_collection.read_concern.document == {"level": "local"}
|
||||||
|
assert update_collection.write_concern.document == {
|
||||||
|
"w": "majority",
|
||||||
|
"j": True,
|
||||||
|
"wtimeout": 1234,
|
||||||
|
}
|
||||||
|
|
||||||
|
assert original_read_concern.document == collection.read_concern.document
|
||||||
|
assert original_write_concern.document == collection.write_concern.document
|
||||||
|
|
||||||
def test_switch_db_context_manager(self):
|
def test_switch_db_context_manager(self):
|
||||||
connect("mongoenginetest")
|
connect("mongoenginetest")
|
||||||
register_connection("testdb-1", "mongoenginetest2")
|
register_connection("testdb-1", "mongoenginetest2")
|
||||||
@ -216,7 +264,7 @@ class TestContextManagers:
|
|||||||
|
|
||||||
def test_query_counter_does_not_swallow_exception(self):
|
def test_query_counter_does_not_swallow_exception(self):
|
||||||
with pytest.raises(TypeError):
|
with pytest.raises(TypeError):
|
||||||
with query_counter() as q:
|
with query_counter():
|
||||||
raise TypeError()
|
raise TypeError()
|
||||||
|
|
||||||
def test_query_counter_temporarily_modifies_profiling_level(self):
|
def test_query_counter_temporarily_modifies_profiling_level(self):
|
||||||
@ -226,12 +274,12 @@ class TestContextManagers:
|
|||||||
initial_profiling_level = db.profiling_level()
|
initial_profiling_level = db.profiling_level()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
NEW_LEVEL = 1
|
new_level = 1
|
||||||
db.set_profiling_level(NEW_LEVEL)
|
db.set_profiling_level(new_level)
|
||||||
assert db.profiling_level() == NEW_LEVEL
|
assert db.profiling_level() == new_level
|
||||||
with query_counter() as q:
|
with query_counter():
|
||||||
assert db.profiling_level() == 2
|
assert db.profiling_level() == 2
|
||||||
assert db.profiling_level() == NEW_LEVEL
|
assert db.profiling_level() == new_level
|
||||||
except Exception:
|
except Exception:
|
||||||
db.set_profiling_level(
|
db.set_profiling_level(
|
||||||
initial_profiling_level
|
initial_profiling_level
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
from six import iterkeys
|
|
||||||
|
|
||||||
from mongoengine import Document
|
from mongoengine import Document
|
||||||
from mongoengine.base.datastructures import BaseDict, BaseList, StrictDict
|
from mongoengine.base.datastructures import BaseDict, BaseList, StrictDict
|
||||||
@ -287,7 +286,7 @@ class TestBaseList:
|
|||||||
base_list[:] = [
|
base_list[:] = [
|
||||||
0,
|
0,
|
||||||
1,
|
1,
|
||||||
] # Will use __setslice__ under py2 and __setitem__ under py3
|
]
|
||||||
assert base_list._instance._changed_fields == ["my_name"]
|
assert base_list._instance._changed_fields == ["my_name"]
|
||||||
assert base_list == [0, 1]
|
assert base_list == [0, 1]
|
||||||
|
|
||||||
@ -296,13 +295,13 @@ class TestBaseList:
|
|||||||
base_list[0:2] = [
|
base_list[0:2] = [
|
||||||
1,
|
1,
|
||||||
0,
|
0,
|
||||||
] # Will use __setslice__ under py2 and __setitem__ under py3
|
]
|
||||||
assert base_list._instance._changed_fields == ["my_name"]
|
assert base_list._instance._changed_fields == ["my_name"]
|
||||||
assert base_list == [1, 0, 2]
|
assert base_list == [1, 0, 2]
|
||||||
|
|
||||||
def test___setitem___calls_with_step_slice_mark_as_changed(self):
|
def test___setitem___calls_with_step_slice_mark_as_changed(self):
|
||||||
base_list = self._get_baselist([0, 1, 2])
|
base_list = self._get_baselist([0, 1, 2])
|
||||||
base_list[0:3:2] = [-1, -2] # uses __setitem__ in both py2 & 3
|
base_list[0:3:2] = [-1, -2] # uses __setitem__
|
||||||
assert base_list._instance._changed_fields == ["my_name"]
|
assert base_list._instance._changed_fields == ["my_name"]
|
||||||
assert base_list == [-1, 1, -2]
|
assert base_list == [-1, 1, -2]
|
||||||
|
|
||||||
@ -372,7 +371,7 @@ class TestStrictDict(unittest.TestCase):
|
|||||||
|
|
||||||
def test_iterkeys(self):
|
def test_iterkeys(self):
|
||||||
d = self.dtype(a=1)
|
d = self.dtype(a=1)
|
||||||
assert list(iterkeys(d)) == ["a"]
|
assert list(d.keys()) == ["a"]
|
||||||
|
|
||||||
def test_len(self):
|
def test_len(self):
|
||||||
d = self.dtype(a=1)
|
d = self.dtype(a=1)
|
||||||
|
@ -2,10 +2,8 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
from six import iteritems
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
|
||||||
from mongoengine.context_managers import query_counter
|
from mongoengine.context_managers import query_counter
|
||||||
|
|
||||||
|
|
||||||
@ -739,7 +737,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, User)
|
assert isinstance(m, User)
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@ -752,7 +750,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, User)
|
assert isinstance(m, User)
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@ -766,7 +764,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, User)
|
assert isinstance(m, User)
|
||||||
|
|
||||||
User.drop_collection()
|
User.drop_collection()
|
||||||
@ -820,7 +818,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@ -836,7 +834,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@ -853,7 +851,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
Group.objects.delete()
|
Group.objects.delete()
|
||||||
@ -910,7 +908,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, UserA)
|
assert isinstance(m, UserA)
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@ -926,7 +924,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, UserA)
|
assert isinstance(m, UserA)
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@ -943,7 +941,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, UserA)
|
assert isinstance(m, UserA)
|
||||||
|
|
||||||
UserA.drop_collection()
|
UserA.drop_collection()
|
||||||
@ -997,7 +995,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@ -1013,7 +1011,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@ -1030,7 +1028,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
[m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in iteritems(group_obj.members):
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
Group.objects.delete()
|
Group.objects.delete()
|
||||||
|
@ -58,7 +58,9 @@ class TestSignal(unittest.TestCase):
|
|||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def post_save(cls, sender, document, **kwargs):
|
def post_save(cls, sender, document, **kwargs):
|
||||||
dirty_keys = document._delta()[0].keys() + document._delta()[1].keys()
|
dirty_keys = list(document._delta()[0].keys()) + list(
|
||||||
|
document._delta()[1].keys()
|
||||||
|
)
|
||||||
signal_output.append("post_save signal, %s" % document)
|
signal_output.append("post_save signal, %s" % document)
|
||||||
signal_output.append("post_save dirty keys, %s" % dirty_keys)
|
signal_output.append("post_save dirty keys, %s" % dirty_keys)
|
||||||
if kwargs.pop("created", False):
|
if kwargs.pop("created", False):
|
||||||
@ -265,7 +267,7 @@ class TestSignal(unittest.TestCase):
|
|||||||
a = self.Author(name="Bill Shakespeare")
|
a = self.Author(name="Bill Shakespeare")
|
||||||
a.save()
|
a.save()
|
||||||
self.get_signal_output(lambda: None) # eliminate signal output
|
self.get_signal_output(lambda: None) # eliminate signal output
|
||||||
a1 = self.Author.objects(name="Bill Shakespeare")[0]
|
_ = self.Author.objects(name="Bill Shakespeare")[0]
|
||||||
|
|
||||||
assert self.get_signal_output(create_author) == [
|
assert self.get_signal_output(create_author) == [
|
||||||
"pre_init signal, Author",
|
"pre_init signal, Author",
|
||||||
|
Loading…
x
Reference in New Issue
Block a user