Compare commits
4 Commits
fix-hash
...
test-conne
Author | SHA1 | Date | |
---|---|---|---|
|
4e8bb14131 | ||
|
9cc4fad614 | ||
|
2a486ee537 | ||
|
2579ed754f |
1
.gitignore
vendored
1
.gitignore
vendored
@@ -17,4 +17,3 @@ tests/test_bugfix.py
|
|||||||
htmlcov/
|
htmlcov/
|
||||||
venv
|
venv
|
||||||
venv3
|
venv3
|
||||||
scratchpad
|
|
||||||
|
@@ -1,23 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
sudo apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 7F0CEB10
|
|
||||||
|
|
||||||
if [ "$MONGODB" = "2.4" ]; then
|
|
||||||
echo "deb http://downloads-distro.mongodb.org/repo/ubuntu-upstart dist 10gen" | sudo tee /etc/apt/sources.list.d/mongodb.list
|
|
||||||
sudo apt-get update
|
|
||||||
sudo apt-get install mongodb-10gen=2.4.14
|
|
||||||
sudo service mongodb start
|
|
||||||
elif [ "$MONGODB" = "2.6" ]; then
|
|
||||||
echo "deb http://downloads-distro.mongodb.org/repo/ubuntu-upstart dist 10gen" | sudo tee /etc/apt/sources.list.d/mongodb.list
|
|
||||||
sudo apt-get update
|
|
||||||
sudo apt-get install mongodb-org-server=2.6.12
|
|
||||||
# service should be started automatically
|
|
||||||
elif [ "$MONGODB" = "3.0" ]; then
|
|
||||||
echo "deb http://repo.mongodb.org/apt/ubuntu precise/mongodb-org/3.0 multiverse" | sudo tee /etc/apt/sources.list.d/mongodb.list
|
|
||||||
sudo apt-get update
|
|
||||||
sudo apt-get install mongodb-org-server=3.0.14
|
|
||||||
# service should be started automatically
|
|
||||||
else
|
|
||||||
echo "Invalid MongoDB version, expected 2.4, 2.6, or 3.0."
|
|
||||||
exit 1
|
|
||||||
fi;
|
|
67
.travis.yml
67
.travis.yml
@@ -1,48 +1,28 @@
|
|||||||
# For full coverage, we'd have to test all supported Python, MongoDB, and
|
|
||||||
# PyMongo combinations. However, that would result in an overly long build
|
|
||||||
# with a very large number of jobs, hence we only test a subset of all the
|
|
||||||
# combinations:
|
|
||||||
# * MongoDB v2.4 & v3.0 are only tested against Python v2.7 & v3.5.
|
|
||||||
# * MongoDB v2.4 is tested against PyMongo v2.7 & v3.x.
|
|
||||||
# * MongoDB v3.0 is tested against PyMongo v3.x.
|
|
||||||
# * MongoDB v2.6 is currently the "main" version tested against Python v2.7,
|
|
||||||
# v3.5, PyPy & PyPy3, and PyMongo v2.7, v2.8 & v3.x.
|
|
||||||
#
|
|
||||||
# Reminder: Update README.rst if you change MongoDB versions we test.
|
|
||||||
|
|
||||||
language: python
|
language: python
|
||||||
|
|
||||||
python:
|
python:
|
||||||
- 2.7
|
- '2.7'
|
||||||
- 3.5
|
- '3.3'
|
||||||
|
- '3.4'
|
||||||
|
- '3.5'
|
||||||
- pypy
|
- pypy
|
||||||
- pypy3
|
- pypy3
|
||||||
|
|
||||||
env:
|
env:
|
||||||
- MONGODB=2.6 PYMONGO=2.7
|
- PYMONGO=2.7
|
||||||
- MONGODB=2.6 PYMONGO=2.8
|
- PYMONGO=2.8
|
||||||
- MONGODB=2.6 PYMONGO=3.0
|
- PYMONGO=3.0
|
||||||
|
- PYMONGO=dev
|
||||||
|
|
||||||
matrix:
|
matrix:
|
||||||
# Finish the build as soon as one job fails
|
|
||||||
fast_finish: true
|
fast_finish: true
|
||||||
|
|
||||||
include:
|
|
||||||
- python: 2.7
|
|
||||||
env: MONGODB=2.4 PYMONGO=2.7
|
|
||||||
- python: 2.7
|
|
||||||
env: MONGODB=2.4 PYMONGO=3.0
|
|
||||||
- python: 2.7
|
|
||||||
env: MONGODB=3.0 PYMONGO=3.0
|
|
||||||
- python: 3.5
|
|
||||||
env: MONGODB=2.4 PYMONGO=2.7
|
|
||||||
- python: 3.5
|
|
||||||
env: MONGODB=2.4 PYMONGO=3.0
|
|
||||||
- python: 3.5
|
|
||||||
env: MONGODB=3.0 PYMONGO=3.0
|
|
||||||
|
|
||||||
before_install:
|
before_install:
|
||||||
- bash .install_mongodb_on_travis.sh
|
- travis_retry sudo apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 7F0CEB10
|
||||||
|
- echo 'deb http://downloads-distro.mongodb.org/repo/ubuntu-upstart dist 10gen' |
|
||||||
|
sudo tee /etc/apt/sources.list.d/mongodb.list
|
||||||
|
- travis_retry sudo apt-get update
|
||||||
|
- travis_retry sudo apt-get install mongodb-org-server
|
||||||
|
|
||||||
install:
|
install:
|
||||||
- sudo apt-get install python-dev python3-dev libopenjpeg-dev zlib1g-dev libjpeg-turbo8-dev
|
- sudo apt-get install python-dev python3-dev libopenjpeg-dev zlib1g-dev libjpeg-turbo8-dev
|
||||||
@@ -50,17 +30,14 @@ install:
|
|||||||
python-tk
|
python-tk
|
||||||
- travis_retry pip install --upgrade pip
|
- travis_retry pip install --upgrade pip
|
||||||
- travis_retry pip install coveralls
|
- travis_retry pip install coveralls
|
||||||
- travis_retry pip install flake8 flake8-import-order
|
- travis_retry pip install flake8
|
||||||
- travis_retry pip install tox>=1.9
|
- travis_retry pip install tox>=1.9
|
||||||
- travis_retry pip install "virtualenv<14.0.0" # virtualenv>=14.0.0 has dropped Python 3.2 support (and pypy3 is based on py32)
|
- travis_retry pip install "virtualenv<14.0.0" # virtualenv>=14.0.0 has dropped Python 3.2 support (and pypy3 is based on py32)
|
||||||
- travis_retry tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -e test
|
- travis_retry tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -e test
|
||||||
|
|
||||||
# Cache dependencies installed via pip
|
|
||||||
cache: pip
|
|
||||||
|
|
||||||
# Run flake8 for py27
|
# Run flake8 for py27
|
||||||
before_script:
|
before_script:
|
||||||
- if [[ $TRAVIS_PYTHON_VERSION == '2.7' ]]; then flake8 .; else echo "flake8 only runs on py27"; fi
|
- if [[ $TRAVIS_PYTHON_VERSION == '2.7' ]]; then tox -e flake8; fi
|
||||||
|
|
||||||
script:
|
script:
|
||||||
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- --with-coverage
|
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- --with-coverage
|
||||||
@@ -68,34 +45,22 @@ script:
|
|||||||
# For now only submit coveralls for Python v2.7. Python v3.x currently shows
|
# For now only submit coveralls for Python v2.7. Python v3.x currently shows
|
||||||
# 0% coverage. That's caused by 'use_2to3', which builds the py3-compatible
|
# 0% coverage. That's caused by 'use_2to3', which builds the py3-compatible
|
||||||
# code in a separate dir and runs tests on that.
|
# code in a separate dir and runs tests on that.
|
||||||
after_success:
|
after_script:
|
||||||
- if [[ $TRAVIS_PYTHON_VERSION == '2.7' ]]; then coveralls --verbose; fi
|
- if [[ $TRAVIS_PYTHON_VERSION == '2.7' ]]; then coveralls --verbose; fi
|
||||||
|
|
||||||
notifications:
|
notifications:
|
||||||
irc: irc.freenode.org#mongoengine
|
irc: irc.freenode.org#mongoengine
|
||||||
|
|
||||||
# Only run builds on the master branch and GitHub releases (tagged as vX.Y.Z)
|
|
||||||
branches:
|
branches:
|
||||||
only:
|
only:
|
||||||
- master
|
- master
|
||||||
- /^v.*$/
|
- /^v.*$/
|
||||||
|
|
||||||
# Whenever a new release is created via GitHub, publish it on PyPI.
|
|
||||||
deploy:
|
deploy:
|
||||||
provider: pypi
|
provider: pypi
|
||||||
user: the_drow
|
user: the_drow
|
||||||
password:
|
password:
|
||||||
secure: QMyatmWBnC6ZN3XLW2+fTBDU4LQcp1m/LjR2/0uamyeUzWKdlOoh/Wx5elOgLwt/8N9ppdPeG83ose1jOz69l5G0MUMjv8n/RIcMFSpCT59tGYqn3kh55b0cIZXFT9ar+5cxlif6a5rS72IHm5li7QQyxexJIII6Uxp0kpvUmek=
|
secure: QMyatmWBnC6ZN3XLW2+fTBDU4LQcp1m/LjR2/0uamyeUzWKdlOoh/Wx5elOgLwt/8N9ppdPeG83ose1jOz69l5G0MUMjv8n/RIcMFSpCT59tGYqn3kh55b0cIZXFT9ar+5cxlif6a5rS72IHm5li7QQyxexJIII6Uxp0kpvUmek=
|
||||||
|
|
||||||
# create a source distribution and a pure python wheel for faster installs
|
|
||||||
distributions: "sdist bdist_wheel"
|
|
||||||
|
|
||||||
# only deploy on tagged commits (aka GitHub releases) and only for the
|
|
||||||
# parent repo's builds running Python 2.7 along with dev PyMongo (we run
|
|
||||||
# Travis against many different Python and PyMongo versions and we don't
|
|
||||||
# want the deploy to occur multiple times).
|
|
||||||
on:
|
on:
|
||||||
tags: true
|
tags: true
|
||||||
repo: MongoEngine/mongoengine
|
repo: MongoEngine/mongoengine
|
||||||
condition: "$PYMONGO = 3.0"
|
|
||||||
python: 2.7
|
|
||||||
|
@@ -29,20 +29,19 @@ Style Guide
|
|||||||
-----------
|
-----------
|
||||||
|
|
||||||
MongoEngine aims to follow `PEP8 <http://www.python.org/dev/peps/pep-0008/>`_
|
MongoEngine aims to follow `PEP8 <http://www.python.org/dev/peps/pep-0008/>`_
|
||||||
including 4 space indents. When possible we try to stick to 79 character line
|
including 4 space indents. When possible we try to stick to 79 character line limits.
|
||||||
limits. However, screens got bigger and an ORM has a strong focus on
|
However, screens got bigger and an ORM has a strong focus on readability and
|
||||||
readability and if it can help, we accept 119 as maximum line length, in a
|
if it can help, we accept 119 as maximum line length, in a similar way as
|
||||||
similar way as `django does
|
`django does <https://docs.djangoproject.com/en/dev/internals/contributing/writing-code/coding-style/#python-style>`_
|
||||||
<https://docs.djangoproject.com/en/dev/internals/contributing/writing-code/coding-style/#python-style>`_
|
|
||||||
|
|
||||||
Testing
|
Testing
|
||||||
-------
|
-------
|
||||||
|
|
||||||
All tests are run on `Travis <http://travis-ci.org/MongoEngine/mongoengine>`_
|
All tests are run on `Travis <http://travis-ci.org/MongoEngine/mongoengine>`_
|
||||||
and any pull requests are automatically tested. Any pull requests without
|
and any pull requests are automatically tested by Travis. Any pull requests
|
||||||
tests will take longer to be integrated and might be refused.
|
without tests will take longer to be integrated and might be refused.
|
||||||
|
|
||||||
You may also submit a simple failing test as a pull request if you don't know
|
You may also submit a simple failing test as a PullRequest if you don't know
|
||||||
how to fix it, it will be easier for other people to work on it and it may get
|
how to fix it, it will be easier for other people to work on it and it may get
|
||||||
fixed faster.
|
fixed faster.
|
||||||
|
|
||||||
@@ -50,18 +49,13 @@ General Guidelines
|
|||||||
------------------
|
------------------
|
||||||
|
|
||||||
- Avoid backward breaking changes if at all possible.
|
- Avoid backward breaking changes if at all possible.
|
||||||
- If you *have* to introduce a breaking change, make it very clear in your
|
|
||||||
pull request's description. Also, describe how users of this package
|
|
||||||
should adapt to the breaking change in docs/upgrade.rst.
|
|
||||||
- Write inline documentation for new classes and methods.
|
- Write inline documentation for new classes and methods.
|
||||||
- Write tests and make sure they pass (make sure you have a mongod
|
- Write tests and make sure they pass (make sure you have a mongod
|
||||||
running on the default port, then execute ``python setup.py nosetests``
|
running on the default port, then execute ``python setup.py nosetests``
|
||||||
from the cmd line to run the test suite).
|
from the cmd line to run the test suite).
|
||||||
- Ensure tests pass on all supported Python, PyMongo, and MongoDB versions.
|
- Ensure tests pass on every Python and PyMongo versions.
|
||||||
You can test various Python and PyMongo versions locally by executing
|
You can test on these versions locally by executing ``tox``
|
||||||
``tox``. For different MongoDB versions, you can rely on our automated
|
- Add enhancements or problematic bug fixes to docs/changelog.rst
|
||||||
Travis tests.
|
|
||||||
- Add enhancements or problematic bug fixes to docs/changelog.rst.
|
|
||||||
- Add yourself to AUTHORS :)
|
- Add yourself to AUTHORS :)
|
||||||
|
|
||||||
Documentation
|
Documentation
|
||||||
@@ -75,6 +69,3 @@ just make your changes to the inline documentation of the appropriate
|
|||||||
branch and submit a `pull request <https://help.github.com/articles/using-pull-requests>`_.
|
branch and submit a `pull request <https://help.github.com/articles/using-pull-requests>`_.
|
||||||
You might also use the github `Edit <https://github.com/blog/844-forking-with-the-edit-button>`_
|
You might also use the github `Edit <https://github.com/blog/844-forking-with-the-edit-button>`_
|
||||||
button.
|
button.
|
||||||
|
|
||||||
If you want to test your documentation changes locally, you need to install
|
|
||||||
the ``sphinx`` package.
|
|
||||||
|
45
README.rst
45
README.rst
@@ -19,31 +19,23 @@ MongoEngine
|
|||||||
About
|
About
|
||||||
=====
|
=====
|
||||||
MongoEngine is a Python Object-Document Mapper for working with MongoDB.
|
MongoEngine is a Python Object-Document Mapper for working with MongoDB.
|
||||||
Documentation is available at https://mongoengine-odm.readthedocs.io - there
|
Documentation available at https://mongoengine-odm.readthedocs.io - there is currently
|
||||||
is currently a `tutorial <https://mongoengine-odm.readthedocs.io/tutorial.html>`_,
|
a `tutorial <https://mongoengine-odm.readthedocs.io/tutorial.html>`_, a `user guide
|
||||||
a `user guide <https://mongoengine-odm.readthedocs.io/guide/index.html>`_, and
|
<https://mongoengine-odm.readthedocs.io/guide/index.html>`_ and an `API reference
|
||||||
an `API reference <https://mongoengine-odm.readthedocs.io/apireference.html>`_.
|
<https://mongoengine-odm.readthedocs.io/apireference.html>`_.
|
||||||
|
|
||||||
Supported MongoDB Versions
|
|
||||||
==========================
|
|
||||||
MongoEngine is currently tested against MongoDB v2.4, v2.6, and v3.0. Future
|
|
||||||
versions should be supported as well, but aren't actively tested at the moment.
|
|
||||||
Make sure to open an issue or submit a pull request if you experience any
|
|
||||||
problems with MongoDB v3.2+.
|
|
||||||
|
|
||||||
Installation
|
Installation
|
||||||
============
|
============
|
||||||
We recommend the use of `virtualenv <https://virtualenv.pypa.io/>`_ and of
|
We recommend the use of `virtualenv <https://virtualenv.pypa.io/>`_ and of
|
||||||
`pip <https://pip.pypa.io/>`_. You can then use ``pip install -U mongoengine``.
|
`pip <https://pip.pypa.io/>`_. You can then use ``pip install -U mongoengine``.
|
||||||
You may also have `setuptools <http://peak.telecommunity.com/DevCenter/setuptools>`_
|
You may also have `setuptools <http://peak.telecommunity.com/DevCenter/setuptools>`_ and thus
|
||||||
and thus you can use ``easy_install -U mongoengine``. Otherwise, you can download the
|
you can use ``easy_install -U mongoengine``. Otherwise, you can download the
|
||||||
source from `GitHub <http://github.com/MongoEngine/mongoengine>`_ and run ``python
|
source from `GitHub <http://github.com/MongoEngine/mongoengine>`_ and run ``python
|
||||||
setup.py install``.
|
setup.py install``.
|
||||||
|
|
||||||
Dependencies
|
Dependencies
|
||||||
============
|
============
|
||||||
All of the dependencies can easily be installed via `pip <https://pip.pypa.io/>`_.
|
All of the dependencies can easily be installed via `pip <https://pip.pypa.io/>`_. At the very least, you'll need these two packages to use MongoEngine:
|
||||||
At the very least, you'll need these two packages to use MongoEngine:
|
|
||||||
|
|
||||||
- pymongo>=2.7.1
|
- pymongo>=2.7.1
|
||||||
- six>=1.10.0
|
- six>=1.10.0
|
||||||
@@ -56,6 +48,10 @@ If you need to use an ``ImageField`` or ``ImageGridFsProxy``:
|
|||||||
|
|
||||||
- Pillow>=2.0.0
|
- Pillow>=2.0.0
|
||||||
|
|
||||||
|
If you want to generate the documentation (e.g. to contribute to it):
|
||||||
|
|
||||||
|
- sphinx
|
||||||
|
|
||||||
Examples
|
Examples
|
||||||
========
|
========
|
||||||
Some simple examples of what MongoEngine code looks like:
|
Some simple examples of what MongoEngine code looks like:
|
||||||
@@ -114,11 +110,11 @@ Some simple examples of what MongoEngine code looks like:
|
|||||||
Tests
|
Tests
|
||||||
=====
|
=====
|
||||||
To run the test suite, ensure you are running a local instance of MongoDB on
|
To run the test suite, ensure you are running a local instance of MongoDB on
|
||||||
the standard port and have ``nose`` installed. Then, run ``python setup.py nosetests``.
|
the standard port and have ``nose`` installed. Then, run: ``python setup.py nosetests``.
|
||||||
|
|
||||||
To run the test suite on every supported Python and PyMongo version, you can
|
To run the test suite on every supported Python version and every supported PyMongo version,
|
||||||
use ``tox``. You'll need to make sure you have each supported Python version
|
you can use ``tox``.
|
||||||
installed in your environment and then:
|
tox and each supported Python version should be installed in your environment:
|
||||||
|
|
||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
@@ -127,16 +123,13 @@ installed in your environment and then:
|
|||||||
# Run the test suites
|
# Run the test suites
|
||||||
$ tox
|
$ tox
|
||||||
|
|
||||||
If you wish to run a subset of tests, use the nosetests convention:
|
If you wish to run one single or selected tests, use the nosetest convention. It will find the folder,
|
||||||
|
eventually the file, go to the TestClass specified after the colon and eventually right to the single test.
|
||||||
|
Also use the -s argument if you want to print out whatever or access pdb while testing.
|
||||||
|
|
||||||
.. code-block:: shell
|
.. code-block:: shell
|
||||||
|
|
||||||
# Run all the tests in a particular test file
|
$ python setup.py nosetests --tests tests/fields/fields.py:FieldTest.test_cls_field -s
|
||||||
$ python setup.py nosetests --tests tests/fields/fields.py
|
|
||||||
# Run only particular test class in that file
|
|
||||||
$ python setup.py nosetests --tests tests/fields/fields.py:FieldTest
|
|
||||||
# Use the -s option if you want to print some debug statements or use pdb
|
|
||||||
$ python setup.py nosetests --tests tests/fields/fields.py:FieldTest -s
|
|
||||||
|
|
||||||
Community
|
Community
|
||||||
=========
|
=========
|
||||||
|
@@ -5,27 +5,11 @@ Changelog
|
|||||||
Development
|
Development
|
||||||
===========
|
===========
|
||||||
- (Fill this out as you fix issues and develop your features).
|
- (Fill this out as you fix issues and develop your features).
|
||||||
|
- Fixed using sets in field choices #1481
|
||||||
Changes in 0.13.0
|
|
||||||
=================
|
|
||||||
- POTENTIAL BREAKING CHANGE: Added Unicode support to the `EmailField`, see
|
|
||||||
docs/upgrade.rst for details.
|
|
||||||
|
|
||||||
Changes in 0.12.0
|
|
||||||
=================
|
|
||||||
- POTENTIAL BREAKING CHANGE: Fixed limit/skip/hint/batch_size chaining #1476
|
- POTENTIAL BREAKING CHANGE: Fixed limit/skip/hint/batch_size chaining #1476
|
||||||
- POTENTIAL BREAKING CHANGE: Changed a public `QuerySet.clone_into` method to a private `QuerySet._clone_into` #1476
|
- POTENTIAL BREAKING CHANGE: Changed a public `QuerySet.clone_into` method to a private `QuerySet._clone_into` #1476
|
||||||
- Fixed the way `Document.objects.create` works with duplicate IDs #1485
|
|
||||||
- Fixed connecting to a replica set with PyMongo 2.x #1436
|
- Fixed connecting to a replica set with PyMongo 2.x #1436
|
||||||
- Fixed using sets in field choices #1481
|
|
||||||
- Fixed deleting items from a `ListField` #1318
|
|
||||||
- Fixed an obscure error message when filtering by `field__in=non_iterable`. #1237
|
- Fixed an obscure error message when filtering by `field__in=non_iterable`. #1237
|
||||||
- Fixed behavior of a `dec` update operator #1450
|
|
||||||
- Added a `rename` update operator #1454
|
|
||||||
- Added validation for the `db_field` parameter #1448
|
|
||||||
- Fixed the error message displayed when querying an `EmbeddedDocumentField` by an invalid value #1440
|
|
||||||
- Fixed the error message displayed when validating unicode URLs #1486
|
|
||||||
- Raise an error when trying to save an abstract document #1449
|
|
||||||
|
|
||||||
Changes in 0.11.0
|
Changes in 0.11.0
|
||||||
=================
|
=================
|
||||||
|
@@ -42,18 +42,13 @@ the :attr:`host` to
|
|||||||
will establish connection to ``production`` database using
|
will establish connection to ``production`` database using
|
||||||
``admin`` username and ``qwerty`` password.
|
``admin`` username and ``qwerty`` password.
|
||||||
|
|
||||||
Replica Sets
|
ReplicaSets
|
||||||
============
|
===========
|
||||||
|
|
||||||
MongoEngine supports connecting to replica sets::
|
MongoEngine supports
|
||||||
|
:class:`~pymongo.mongo_replica_set_client.MongoReplicaSetClient`. To use them,
|
||||||
from mongoengine import connect
|
please use an URI style connection and provide the ``replicaSet`` name
|
||||||
|
in the connection kwargs.
|
||||||
# Regular connect
|
|
||||||
connect('dbname', replicaset='rs-name')
|
|
||||||
|
|
||||||
# MongoDB URI-style connect
|
|
||||||
connect(host='mongodb://localhost/dbname?replicaSet=rs-name')
|
|
||||||
|
|
||||||
Read preferences are supported through the connection or via individual
|
Read preferences are supported through the connection or via individual
|
||||||
queries by passing the read_preference ::
|
queries by passing the read_preference ::
|
||||||
@@ -64,74 +59,76 @@ queries by passing the read_preference ::
|
|||||||
Multiple Databases
|
Multiple Databases
|
||||||
==================
|
==================
|
||||||
|
|
||||||
To use multiple databases you can use :func:`~mongoengine.connect` and provide
|
Multiple database support was added in MongoEngine 0.6. To use multiple
|
||||||
an `alias` name for the connection - if no `alias` is provided then "default"
|
databases you can use :func:`~mongoengine.connect` and provide an `alias` name
|
||||||
is used.
|
for the connection - if no `alias` is provided then "default" is used.
|
||||||
|
|
||||||
In the background this uses :func:`~mongoengine.register_connection` to
|
In the background this uses :func:`~mongoengine.register_connection` to
|
||||||
store the data and you can register all aliases up front if required.
|
store the data and you can register all aliases up front if required.
|
||||||
|
|
||||||
Individual documents can also support multiple databases by providing a
|
Individual documents can also support multiple databases by providing a
|
||||||
`db_alias` in their meta data. This allows :class:`~pymongo.dbref.DBRef`
|
`db_alias` in their meta data. This allows :class:`~pymongo.dbref.DBRef` objects
|
||||||
objects to point across databases and collections. Below is an example schema,
|
to point across databases and collections. Below is an example schema, using
|
||||||
using 3 different databases to store data::
|
3 different databases to store data::
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
meta = {'db_alias': 'user-db'}
|
meta = {"db_alias": "user-db"}
|
||||||
|
|
||||||
class Book(Document):
|
class Book(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
meta = {'db_alias': 'book-db'}
|
meta = {"db_alias": "book-db"}
|
||||||
|
|
||||||
class AuthorBooks(Document):
|
class AuthorBooks(Document):
|
||||||
author = ReferenceField(User)
|
author = ReferenceField(User)
|
||||||
book = ReferenceField(Book)
|
book = ReferenceField(Book)
|
||||||
|
|
||||||
meta = {'db_alias': 'users-books-db'}
|
meta = {"db_alias": "users-books-db"}
|
||||||
|
|
||||||
|
|
||||||
Context Managers
|
Context Managers
|
||||||
================
|
================
|
||||||
Sometimes you may want to switch the database or collection to query against.
|
Sometimes you may want to switch the database or collection to query against
|
||||||
|
for a class.
|
||||||
For example, archiving older data into a separate database for performance
|
For example, archiving older data into a separate database for performance
|
||||||
reasons or writing functions that dynamically choose collections to write
|
reasons or writing functions that dynamically choose collections to write
|
||||||
a document to.
|
document to.
|
||||||
|
|
||||||
Switch Database
|
Switch Database
|
||||||
---------------
|
---------------
|
||||||
The :class:`~mongoengine.context_managers.switch_db` context manager allows
|
The :class:`~mongoengine.context_managers.switch_db` context manager allows
|
||||||
you to change the database alias for a given class allowing quick and easy
|
you to change the database alias for a given class allowing quick and easy
|
||||||
access to the same User document across databases::
|
access the same User document across databases::
|
||||||
|
|
||||||
from mongoengine.context_managers import switch_db
|
from mongoengine.context_managers import switch_db
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
meta = {'db_alias': 'user-db'}
|
meta = {"db_alias": "user-db"}
|
||||||
|
|
||||||
with switch_db(User, 'archive-user-db') as User:
|
with switch_db(User, 'archive-user-db') as User:
|
||||||
User(name='Ross').save() # Saves the 'archive-user-db'
|
User(name="Ross").save() # Saves the 'archive-user-db'
|
||||||
|
|
||||||
|
|
||||||
Switch Collection
|
Switch Collection
|
||||||
-----------------
|
-----------------
|
||||||
The :class:`~mongoengine.context_managers.switch_collection` context manager
|
The :class:`~mongoengine.context_managers.switch_collection` context manager
|
||||||
allows you to change the collection for a given class allowing quick and easy
|
allows you to change the collection for a given class allowing quick and easy
|
||||||
access to the same Group document across collection::
|
access the same Group document across collection::
|
||||||
|
|
||||||
from mongoengine.context_managers import switch_collection
|
from mongoengine.context_managers import switch_collection
|
||||||
|
|
||||||
class Group(Document):
|
class Group(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
Group(name='test').save() # Saves in the default db
|
Group(name="test").save() # Saves in the default db
|
||||||
|
|
||||||
with switch_collection(Group, 'group2000') as Group:
|
with switch_collection(Group, 'group2000') as Group:
|
||||||
Group(name='hello Group 2000 collection!').save() # Saves in group2000 collection
|
Group(name="hello Group 2000 collection!").save() # Saves in group2000 collection
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
.. note:: Make sure any aliases have been registered with
|
.. note:: Make sure any aliases have been registered with
|
||||||
|
@@ -2,13 +2,13 @@
|
|||||||
Installing MongoEngine
|
Installing MongoEngine
|
||||||
======================
|
======================
|
||||||
|
|
||||||
To use MongoEngine, you will need to download `MongoDB <http://mongodb.com/>`_
|
To use MongoEngine, you will need to download `MongoDB <http://mongodb.org/>`_
|
||||||
and ensure it is running in an accessible location. You will also need
|
and ensure it is running in an accessible location. You will also need
|
||||||
`PyMongo <http://api.mongodb.org/python>`_ to use MongoEngine, but if you
|
`PyMongo <http://api.mongodb.org/python>`_ to use MongoEngine, but if you
|
||||||
install MongoEngine using setuptools, then the dependencies will be handled for
|
install MongoEngine using setuptools, then the dependencies will be handled for
|
||||||
you.
|
you.
|
||||||
|
|
||||||
MongoEngine is available on PyPI, so you can use :program:`pip`:
|
MongoEngine is available on PyPI, so to use it you can use :program:`pip`:
|
||||||
|
|
||||||
.. code-block:: console
|
.. code-block:: console
|
||||||
|
|
||||||
|
@@ -340,19 +340,14 @@ Javascript code that is executed on the database server.
|
|||||||
|
|
||||||
Counting results
|
Counting results
|
||||||
----------------
|
----------------
|
||||||
Just as with limiting and skipping results, there is a method on a
|
Just as with limiting and skipping results, there is a method on
|
||||||
:class:`~mongoengine.queryset.QuerySet` object --
|
:class:`~mongoengine.queryset.QuerySet` objects --
|
||||||
:meth:`~mongoengine.queryset.QuerySet.count`::
|
:meth:`~mongoengine.queryset.QuerySet.count`, but there is also a more Pythonic
|
||||||
|
way of achieving this::
|
||||||
|
|
||||||
num_users = User.objects.count()
|
num_users = len(User.objects)
|
||||||
|
|
||||||
You could technically use ``len(User.objects)`` to get the same result, but it
|
Even if len() is the Pythonic way of counting results, keep in mind that if you concerned about performance, :meth:`~mongoengine.queryset.QuerySet.count` is the way to go since it only execute a server side count query, while len() retrieves the results, places them in cache, and finally counts them. If we compare the performance of the two operations, len() is much slower than :meth:`~mongoengine.queryset.QuerySet.count`.
|
||||||
would be significantly slower than :meth:`~mongoengine.queryset.QuerySet.count`.
|
|
||||||
When you execute a server-side count query, you let MongoDB do the heavy
|
|
||||||
lifting and you receive a single integer over the wire. Meanwhile, len()
|
|
||||||
retrieves all the results, places them in a local cache, and finally counts
|
|
||||||
them. If we compare the performance of the two operations, len() is much slower
|
|
||||||
than :meth:`~mongoengine.queryset.QuerySet.count`.
|
|
||||||
|
|
||||||
Further aggregation
|
Further aggregation
|
||||||
-------------------
|
-------------------
|
||||||
|
@@ -3,10 +3,11 @@ Tutorial
|
|||||||
========
|
========
|
||||||
|
|
||||||
This tutorial introduces **MongoEngine** by means of example --- we will walk
|
This tutorial introduces **MongoEngine** by means of example --- we will walk
|
||||||
through how to create a simple **Tumblelog** application. A tumblelog is a
|
through how to create a simple **Tumblelog** application. A Tumblelog is a type
|
||||||
blog that supports mixed media content, including text, images, links, video,
|
of blog where posts are not constrained to being conventional text-based posts.
|
||||||
audio, etc. For simplicity's sake, we'll stick to text, image, and link
|
As well as text-based entries, users may post images, links, videos, etc. For
|
||||||
entries. As the purpose of this tutorial is to introduce MongoEngine, we'll
|
simplicity's sake, we'll stick to text, image and link entries in our
|
||||||
|
application. As the purpose of this tutorial is to introduce MongoEngine, we'll
|
||||||
focus on the data-modelling side of the application, leaving out a user
|
focus on the data-modelling side of the application, leaving out a user
|
||||||
interface.
|
interface.
|
||||||
|
|
||||||
@@ -15,14 +16,14 @@ Getting started
|
|||||||
|
|
||||||
Before we start, make sure that a copy of MongoDB is running in an accessible
|
Before we start, make sure that a copy of MongoDB is running in an accessible
|
||||||
location --- running it locally will be easier, but if that is not an option
|
location --- running it locally will be easier, but if that is not an option
|
||||||
then it may be run on a remote server. If you haven't installed MongoEngine,
|
then it may be run on a remote server. If you haven't installed mongoengine,
|
||||||
simply use pip to install it like so::
|
simply use pip to install it like so::
|
||||||
|
|
||||||
$ pip install mongoengine
|
$ pip install mongoengine
|
||||||
|
|
||||||
Before we can start using MongoEngine, we need to tell it how to connect to our
|
Before we can start using MongoEngine, we need to tell it how to connect to our
|
||||||
instance of :program:`mongod`. For this we use the :func:`~mongoengine.connect`
|
instance of :program:`mongod`. For this we use the :func:`~mongoengine.connect`
|
||||||
function. If running locally, the only argument we need to provide is the name
|
function. If running locally the only argument we need to provide is the name
|
||||||
of the MongoDB database to use::
|
of the MongoDB database to use::
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
@@ -38,18 +39,18 @@ Defining our documents
|
|||||||
MongoDB is *schemaless*, which means that no schema is enforced by the database
|
MongoDB is *schemaless*, which means that no schema is enforced by the database
|
||||||
--- we may add and remove fields however we want and MongoDB won't complain.
|
--- we may add and remove fields however we want and MongoDB won't complain.
|
||||||
This makes life a lot easier in many regards, especially when there is a change
|
This makes life a lot easier in many regards, especially when there is a change
|
||||||
to the data model. However, defining schemas for our documents can help to iron
|
to the data model. However, defining schemata for our documents can help to
|
||||||
out bugs involving incorrect types or missing fields, and also allow us to
|
iron out bugs involving incorrect types or missing fields, and also allow us to
|
||||||
define utility methods on our documents in the same way that traditional
|
define utility methods on our documents in the same way that traditional
|
||||||
:abbr:`ORMs (Object-Relational Mappers)` do.
|
:abbr:`ORMs (Object-Relational Mappers)` do.
|
||||||
|
|
||||||
In our Tumblelog application we need to store several different types of
|
In our Tumblelog application we need to store several different types of
|
||||||
information. We will need to have a collection of **users**, so that we may
|
information. We will need to have a collection of **users**, so that we may
|
||||||
link posts to an individual. We also need to store our different types of
|
link posts to an individual. We also need to store our different types of
|
||||||
**posts** (eg: text, image and link) in the database. To aid navigation of our
|
**posts** (eg: text, image and link) in the database. To aid navigation of our
|
||||||
Tumblelog, posts may have **tags** associated with them, so that the list of
|
Tumblelog, posts may have **tags** associated with them, so that the list of
|
||||||
posts shown to the user may be limited to posts that have been assigned a
|
posts shown to the user may be limited to posts that have been assigned a
|
||||||
specific tag. Finally, it would be nice if **comments** could be added to
|
specific tag. Finally, it would be nice if **comments** could be added to
|
||||||
posts. We'll start with **users**, as the other document models are slightly
|
posts. We'll start with **users**, as the other document models are slightly
|
||||||
more involved.
|
more involved.
|
||||||
|
|
||||||
@@ -77,7 +78,7 @@ Now we'll think about how to store the rest of the information. If we were
|
|||||||
using a relational database, we would most likely have a table of **posts**, a
|
using a relational database, we would most likely have a table of **posts**, a
|
||||||
table of **comments** and a table of **tags**. To associate the comments with
|
table of **comments** and a table of **tags**. To associate the comments with
|
||||||
individual posts, we would put a column in the comments table that contained a
|
individual posts, we would put a column in the comments table that contained a
|
||||||
foreign key to the posts table. We'd also need a link table to provide the
|
foreign key to the posts table. We'd also need a link table to provide the
|
||||||
many-to-many relationship between posts and tags. Then we'd need to address the
|
many-to-many relationship between posts and tags. Then we'd need to address the
|
||||||
problem of storing the specialised post-types (text, image and link). There are
|
problem of storing the specialised post-types (text, image and link). There are
|
||||||
several ways we can achieve this, but each of them have their problems --- none
|
several ways we can achieve this, but each of them have their problems --- none
|
||||||
@@ -95,7 +96,7 @@ using* the new fields we need to support video posts. This fits with the
|
|||||||
Object-Oriented principle of *inheritance* nicely. We can think of
|
Object-Oriented principle of *inheritance* nicely. We can think of
|
||||||
:class:`Post` as a base class, and :class:`TextPost`, :class:`ImagePost` and
|
:class:`Post` as a base class, and :class:`TextPost`, :class:`ImagePost` and
|
||||||
:class:`LinkPost` as subclasses of :class:`Post`. In fact, MongoEngine supports
|
:class:`LinkPost` as subclasses of :class:`Post`. In fact, MongoEngine supports
|
||||||
this kind of modeling out of the box --- all you need do is turn on inheritance
|
this kind of modelling out of the box --- all you need do is turn on inheritance
|
||||||
by setting :attr:`allow_inheritance` to True in the :attr:`meta`::
|
by setting :attr:`allow_inheritance` to True in the :attr:`meta`::
|
||||||
|
|
||||||
class Post(Document):
|
class Post(Document):
|
||||||
@@ -127,8 +128,8 @@ link table, we can just store a list of tags in each post. So, for both
|
|||||||
efficiency and simplicity's sake, we'll store the tags as strings directly
|
efficiency and simplicity's sake, we'll store the tags as strings directly
|
||||||
within the post, rather than storing references to tags in a separate
|
within the post, rather than storing references to tags in a separate
|
||||||
collection. Especially as tags are generally very short (often even shorter
|
collection. Especially as tags are generally very short (often even shorter
|
||||||
than a document's id), this denormalization won't impact the size of the
|
than a document's id), this denormalisation won't impact very strongly on the
|
||||||
database very strongly. Let's take a look at the code of our modified
|
size of our database. So let's take a look that the code our modified
|
||||||
:class:`Post` class::
|
:class:`Post` class::
|
||||||
|
|
||||||
class Post(Document):
|
class Post(Document):
|
||||||
@@ -140,7 +141,7 @@ The :class:`~mongoengine.fields.ListField` object that is used to define a Post'
|
|||||||
takes a field object as its first argument --- this means that you can have
|
takes a field object as its first argument --- this means that you can have
|
||||||
lists of any type of field (including lists).
|
lists of any type of field (including lists).
|
||||||
|
|
||||||
.. note:: We don't need to modify the specialized post types as they all
|
.. note:: We don't need to modify the specialised post types as they all
|
||||||
inherit from :class:`Post`.
|
inherit from :class:`Post`.
|
||||||
|
|
||||||
Comments
|
Comments
|
||||||
@@ -148,7 +149,7 @@ Comments
|
|||||||
|
|
||||||
A comment is typically associated with *one* post. In a relational database, to
|
A comment is typically associated with *one* post. In a relational database, to
|
||||||
display a post with its comments, we would have to retrieve the post from the
|
display a post with its comments, we would have to retrieve the post from the
|
||||||
database and then query the database again for the comments associated with the
|
database, then query the database again for the comments associated with the
|
||||||
post. This works, but there is no real reason to be storing the comments
|
post. This works, but there is no real reason to be storing the comments
|
||||||
separately from their associated posts, other than to work around the
|
separately from their associated posts, other than to work around the
|
||||||
relational model. Using MongoDB we can store the comments as a list of
|
relational model. Using MongoDB we can store the comments as a list of
|
||||||
@@ -218,8 +219,8 @@ Now that we've got our user in the database, let's add a couple of posts::
|
|||||||
post2.tags = ['mongoengine']
|
post2.tags = ['mongoengine']
|
||||||
post2.save()
|
post2.save()
|
||||||
|
|
||||||
.. note:: If you change a field on an object that has already been saved and
|
.. note:: If you change a field on a object that has already been saved, then
|
||||||
then call :meth:`save` again, the document will be updated.
|
call :meth:`save` again, the document will be updated.
|
||||||
|
|
||||||
Accessing our data
|
Accessing our data
|
||||||
==================
|
==================
|
||||||
@@ -231,17 +232,17 @@ used to access the documents in the database collection associated with that
|
|||||||
class. So let's see how we can get our posts' titles::
|
class. So let's see how we can get our posts' titles::
|
||||||
|
|
||||||
for post in Post.objects:
|
for post in Post.objects:
|
||||||
print(post.title)
|
print post.title
|
||||||
|
|
||||||
Retrieving type-specific information
|
Retrieving type-specific information
|
||||||
------------------------------------
|
------------------------------------
|
||||||
|
|
||||||
This will print the titles of our posts, one on each line. But what if we want
|
This will print the titles of our posts, one on each line. But What if we want
|
||||||
to access the type-specific data (link_url, content, etc.)? One way is simply
|
to access the type-specific data (link_url, content, etc.)? One way is simply
|
||||||
to use the :attr:`objects` attribute of a subclass of :class:`Post`::
|
to use the :attr:`objects` attribute of a subclass of :class:`Post`::
|
||||||
|
|
||||||
for post in TextPost.objects:
|
for post in TextPost.objects:
|
||||||
print(post.content)
|
print post.content
|
||||||
|
|
||||||
Using TextPost's :attr:`objects` attribute only returns documents that were
|
Using TextPost's :attr:`objects` attribute only returns documents that were
|
||||||
created using :class:`TextPost`. Actually, there is a more general rule here:
|
created using :class:`TextPost`. Actually, there is a more general rule here:
|
||||||
@@ -258,14 +259,16 @@ instances of :class:`Post` --- they were instances of the subclass of
|
|||||||
practice::
|
practice::
|
||||||
|
|
||||||
for post in Post.objects:
|
for post in Post.objects:
|
||||||
print(post.title)
|
print post.title
|
||||||
print('=' * len(post.title))
|
print '=' * len(post.title)
|
||||||
|
|
||||||
if isinstance(post, TextPost):
|
if isinstance(post, TextPost):
|
||||||
print(post.content)
|
print post.content
|
||||||
|
|
||||||
if isinstance(post, LinkPost):
|
if isinstance(post, LinkPost):
|
||||||
print('Link: {}'.format(post.link_url))
|
print 'Link:', post.link_url
|
||||||
|
|
||||||
|
print
|
||||||
|
|
||||||
This would print the title of each post, followed by the content if it was a
|
This would print the title of each post, followed by the content if it was a
|
||||||
text post, and "Link: <url>" if it was a link post.
|
text post, and "Link: <url>" if it was a link post.
|
||||||
@@ -280,7 +283,7 @@ your query. Let's adjust our query so that only posts with the tag "mongodb"
|
|||||||
are returned::
|
are returned::
|
||||||
|
|
||||||
for post in Post.objects(tags='mongodb'):
|
for post in Post.objects(tags='mongodb'):
|
||||||
print(post.title)
|
print post.title
|
||||||
|
|
||||||
There are also methods available on :class:`~mongoengine.queryset.QuerySet`
|
There are also methods available on :class:`~mongoengine.queryset.QuerySet`
|
||||||
objects that allow different results to be returned, for example, calling
|
objects that allow different results to be returned, for example, calling
|
||||||
@@ -289,11 +292,11 @@ the first matched by the query you provide. Aggregation functions may also be
|
|||||||
used on :class:`~mongoengine.queryset.QuerySet` objects::
|
used on :class:`~mongoengine.queryset.QuerySet` objects::
|
||||||
|
|
||||||
num_posts = Post.objects(tags='mongodb').count()
|
num_posts = Post.objects(tags='mongodb').count()
|
||||||
print('Found {} posts with tag "mongodb"'.format(num_posts))
|
print 'Found %d posts with tag "mongodb"' % num_posts
|
||||||
|
|
||||||
Learning more about MongoEngine
|
Learning more about mongoengine
|
||||||
-------------------------------
|
-------------------------------
|
||||||
|
|
||||||
If you got this far you've made a great start, so well done! The next step on
|
If you got this far you've made a great start, so well done! The next step on
|
||||||
your MongoEngine journey is the `full user guide <guide/index.html>`_, where
|
your mongoengine journey is the `full user guide <guide/index.html>`_, where you
|
||||||
you can learn in-depth about how to use MongoEngine and MongoDB.
|
can learn indepth about how to use mongoengine and mongodb.
|
||||||
|
@@ -6,20 +6,6 @@ Development
|
|||||||
***********
|
***********
|
||||||
(Fill this out whenever you introduce breaking changes to MongoEngine)
|
(Fill this out whenever you introduce breaking changes to MongoEngine)
|
||||||
|
|
||||||
0.13.0
|
|
||||||
******
|
|
||||||
This release adds Unicode support to the `EmailField` and changes its
|
|
||||||
structure significantly. Previously, email addresses containing Unicode
|
|
||||||
characters didn't work at all. Starting with v0.13.0, domains with Unicode
|
|
||||||
characters are supported out of the box, meaning some emails that previously
|
|
||||||
didn't pass validation now do. Make sure the rest of your application can
|
|
||||||
accept such email addresses. Additionally, if you subclassed the `EmailField`
|
|
||||||
in your application and overrode `EmailField.EMAIL_REGEX`, you will have to
|
|
||||||
adjust your code to override `EmailField.USER_REGEX`, `EmailField.DOMAIN_REGEX`,
|
|
||||||
and potentially `EmailField.UTF8_USER_REGEX`.
|
|
||||||
|
|
||||||
0.12.0
|
|
||||||
******
|
|
||||||
This release includes various fixes for the `BaseQuerySet` methods and how they
|
This release includes various fixes for the `BaseQuerySet` methods and how they
|
||||||
are chained together. Since version 0.10.1 applying limit/skip/hint/batch_size
|
are chained together. Since version 0.10.1 applying limit/skip/hint/batch_size
|
||||||
to an already-existing queryset wouldn't modify the underlying PyMongo cursor.
|
to an already-existing queryset wouldn't modify the underlying PyMongo cursor.
|
||||||
|
@@ -23,7 +23,7 @@ __all__ = (list(document.__all__) + list(fields.__all__) +
|
|||||||
list(signals.__all__) + list(errors.__all__))
|
list(signals.__all__) + list(errors.__all__))
|
||||||
|
|
||||||
|
|
||||||
VERSION = (0, 13, 0)
|
VERSION = (0, 11, 0)
|
||||||
|
|
||||||
|
|
||||||
def get_version():
|
def get_version():
|
||||||
|
@@ -272,6 +272,13 @@ class BaseDocument(object):
|
|||||||
def __ne__(self, other):
|
def __ne__(self, other):
|
||||||
return not self.__eq__(other)
|
return not self.__eq__(other)
|
||||||
|
|
||||||
|
def __hash__(self):
|
||||||
|
if getattr(self, 'pk', None) is None:
|
||||||
|
# For new object
|
||||||
|
return super(BaseDocument, self).__hash__()
|
||||||
|
else:
|
||||||
|
return hash(self.pk)
|
||||||
|
|
||||||
def clean(self):
|
def clean(self):
|
||||||
"""
|
"""
|
||||||
Hook for doing document level data cleaning before validation is run.
|
Hook for doing document level data cleaning before validation is run.
|
||||||
@@ -677,13 +684,8 @@ class BaseDocument(object):
|
|||||||
# class if unavailable
|
# class if unavailable
|
||||||
class_name = son.get('_cls', cls._class_name)
|
class_name = son.get('_cls', cls._class_name)
|
||||||
|
|
||||||
# Convert SON to a data dict, making sure each key is a string and
|
# Convert SON to a dict, making sure each key is a string
|
||||||
# corresponds to the right db field.
|
data = {str(key): value for key, value in son.iteritems()}
|
||||||
data = {}
|
|
||||||
for key, value in son.iteritems():
|
|
||||||
key = str(key)
|
|
||||||
key = cls._db_field_map.get(key, key)
|
|
||||||
data[key] = value
|
|
||||||
|
|
||||||
# Return correct subclass for document type
|
# Return correct subclass for document type
|
||||||
if class_name != cls._class_name:
|
if class_name != cls._class_name:
|
||||||
|
@@ -1,4 +1,3 @@
|
|||||||
from collections import OrderedDict
|
|
||||||
from bson import DBRef, SON
|
from bson import DBRef, SON
|
||||||
import six
|
import six
|
||||||
|
|
||||||
@@ -202,10 +201,6 @@ class DeReference(object):
|
|||||||
as_tuple = isinstance(items, tuple)
|
as_tuple = isinstance(items, tuple)
|
||||||
iterator = enumerate(items)
|
iterator = enumerate(items)
|
||||||
data = []
|
data = []
|
||||||
elif isinstance(items, OrderedDict):
|
|
||||||
is_list = False
|
|
||||||
iterator = items.iteritems()
|
|
||||||
data = OrderedDict()
|
|
||||||
else:
|
else:
|
||||||
is_list = False
|
is_list = False
|
||||||
iterator = items.iteritems()
|
iterator = items.iteritems()
|
||||||
|
@@ -60,12 +60,6 @@ class EmbeddedDocument(BaseDocument):
|
|||||||
my_metaclass = DocumentMetaclass
|
my_metaclass = DocumentMetaclass
|
||||||
__metaclass__ = DocumentMetaclass
|
__metaclass__ = DocumentMetaclass
|
||||||
|
|
||||||
# A generic embedded document doesn't have any immutable properties
|
|
||||||
# that describe it uniquely, hence it shouldn't be hashable. You can
|
|
||||||
# define your own __hash__ method on a subclass if you need your
|
|
||||||
# embedded documents to be hashable.
|
|
||||||
__hash__ = None
|
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
super(EmbeddedDocument, self).__init__(*args, **kwargs)
|
super(EmbeddedDocument, self).__init__(*args, **kwargs)
|
||||||
self._instance = None
|
self._instance = None
|
||||||
@@ -166,15 +160,6 @@ class Document(BaseDocument):
|
|||||||
"""Set the primary key."""
|
"""Set the primary key."""
|
||||||
return setattr(self, self._meta['id_field'], value)
|
return setattr(self, self._meta['id_field'], value)
|
||||||
|
|
||||||
def __hash__(self):
|
|
||||||
"""Return the hash based on the PK of this document. If it's new
|
|
||||||
and doesn't have a PK yet, return the default object hash instead.
|
|
||||||
"""
|
|
||||||
if self.pk is None:
|
|
||||||
return super(BaseDocument, self).__hash__()
|
|
||||||
else:
|
|
||||||
return hash(self.pk)
|
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _get_db(cls):
|
def _get_db(cls):
|
||||||
"""Some Model using other db_alias"""
|
"""Some Model using other db_alias"""
|
||||||
|
@@ -2,11 +2,9 @@ import datetime
|
|||||||
import decimal
|
import decimal
|
||||||
import itertools
|
import itertools
|
||||||
import re
|
import re
|
||||||
import socket
|
|
||||||
import time
|
import time
|
||||||
import uuid
|
import uuid
|
||||||
import warnings
|
import warnings
|
||||||
from collections import Mapping
|
|
||||||
from operator import itemgetter
|
from operator import itemgetter
|
||||||
|
|
||||||
from bson import Binary, DBRef, ObjectId, SON
|
from bson import Binary, DBRef, ObjectId, SON
|
||||||
@@ -155,105 +153,21 @@ class EmailField(StringField):
|
|||||||
|
|
||||||
.. versionadded:: 0.4
|
.. versionadded:: 0.4
|
||||||
"""
|
"""
|
||||||
USER_REGEX = re.compile(
|
|
||||||
# `dot-atom` defined in RFC 5322 Section 3.2.3.
|
EMAIL_REGEX = re.compile(
|
||||||
r"(^[-!#$%&'*+/=?^_`{}|~0-9A-Z]+(\.[-!#$%&'*+/=?^_`{}|~0-9A-Z]+)*\Z"
|
# dot-atom
|
||||||
# `quoted-string` defined in RFC 5322 Section 3.2.4.
|
r"(^[-!#$%&'*+/=?^_`{}|~0-9A-Z]+(\.[-!#$%&'*+/=?^_`{}|~0-9A-Z]+)*"
|
||||||
r'|^"([\001-\010\013\014\016-\037!#-\[\]-\177]|\\[\001-\011\013\014\016-\177])*"\Z)',
|
# quoted-string
|
||||||
re.IGNORECASE
|
r'|^"([\001-\010\013\014\016-\037!#-\[\]-\177]|\\[\001-011\013\014\016-\177])*"'
|
||||||
|
# domain (max length of an ICAAN TLD is 22 characters)
|
||||||
|
r')@(?:[A-Z0-9](?:[A-Z0-9-]{0,61}[A-Z0-9])?\.)+(?:[A-Z]{2,6}|[A-Z0-9-]{2,}(?<!-))$', re.IGNORECASE
|
||||||
)
|
)
|
||||||
|
|
||||||
UTF8_USER_REGEX = re.compile(
|
|
||||||
six.u(
|
|
||||||
# RFC 6531 Section 3.3 extends `atext` (used by dot-atom) to
|
|
||||||
# include `UTF8-non-ascii`.
|
|
||||||
r"(^[-!#$%&'*+/=?^_`{}|~0-9A-Z\u0080-\U0010FFFF]+(\.[-!#$%&'*+/=?^_`{}|~0-9A-Z\u0080-\U0010FFFF]+)*\Z"
|
|
||||||
# `quoted-string`
|
|
||||||
r'|^"([\001-\010\013\014\016-\037!#-\[\]-\177]|\\[\001-\011\013\014\016-\177])*"\Z)'
|
|
||||||
), re.IGNORECASE | re.UNICODE
|
|
||||||
)
|
|
||||||
|
|
||||||
DOMAIN_REGEX = re.compile(
|
|
||||||
r'((?:[A-Z0-9](?:[A-Z0-9-]{0,61}[A-Z0-9])?\.)+)(?:[A-Z0-9-]{2,63}(?<!-))\Z',
|
|
||||||
re.IGNORECASE
|
|
||||||
)
|
|
||||||
|
|
||||||
error_msg = u'Invalid email address: %s'
|
|
||||||
|
|
||||||
def __init__(self, domain_whitelist=None, allow_utf8_user=False,
|
|
||||||
allow_ip_domain=False, *args, **kwargs):
|
|
||||||
"""Initialize the EmailField.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
domain_whitelist (list) - list of otherwise invalid domain
|
|
||||||
names which you'd like to support.
|
|
||||||
allow_utf8_user (bool) - if True, the user part of the email
|
|
||||||
address can contain UTF8 characters.
|
|
||||||
False by default.
|
|
||||||
allow_ip_domain (bool) - if True, the domain part of the email
|
|
||||||
can be a valid IPv4 or IPv6 address.
|
|
||||||
"""
|
|
||||||
self.domain_whitelist = domain_whitelist or []
|
|
||||||
self.allow_utf8_user = allow_utf8_user
|
|
||||||
self.allow_ip_domain = allow_ip_domain
|
|
||||||
super(EmailField, self).__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
def validate_user_part(self, user_part):
|
|
||||||
"""Validate the user part of the email address. Return True if
|
|
||||||
valid and False otherwise.
|
|
||||||
"""
|
|
||||||
if self.allow_utf8_user:
|
|
||||||
return self.UTF8_USER_REGEX.match(user_part)
|
|
||||||
return self.USER_REGEX.match(user_part)
|
|
||||||
|
|
||||||
def validate_domain_part(self, domain_part):
|
|
||||||
"""Validate the domain part of the email address. Return True if
|
|
||||||
valid and False otherwise.
|
|
||||||
"""
|
|
||||||
# Skip domain validation if it's in the whitelist.
|
|
||||||
if domain_part in self.domain_whitelist:
|
|
||||||
return True
|
|
||||||
|
|
||||||
if self.DOMAIN_REGEX.match(domain_part):
|
|
||||||
return True
|
|
||||||
|
|
||||||
# Validate IPv4/IPv6, e.g. user@[192.168.0.1]
|
|
||||||
if (
|
|
||||||
self.allow_ip_domain and
|
|
||||||
domain_part[0] == '[' and
|
|
||||||
domain_part[-1] == ']'
|
|
||||||
):
|
|
||||||
for addr_family in (socket.AF_INET, socket.AF_INET6):
|
|
||||||
try:
|
|
||||||
socket.inet_pton(addr_family, domain_part[1:-1])
|
|
||||||
return True
|
|
||||||
except (socket.error, UnicodeEncodeError):
|
|
||||||
pass
|
|
||||||
|
|
||||||
return False
|
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
|
if not EmailField.EMAIL_REGEX.match(value):
|
||||||
|
self.error('Invalid email address: %s' % value)
|
||||||
super(EmailField, self).validate(value)
|
super(EmailField, self).validate(value)
|
||||||
|
|
||||||
if '@' not in value:
|
|
||||||
self.error(self.error_msg % value)
|
|
||||||
|
|
||||||
user_part, domain_part = value.rsplit('@', 1)
|
|
||||||
|
|
||||||
# Validate the user part.
|
|
||||||
if not self.validate_user_part(user_part):
|
|
||||||
self.error(self.error_msg % value)
|
|
||||||
|
|
||||||
# Validate the domain and, if invalid, see if it's IDN-encoded.
|
|
||||||
if not self.validate_domain_part(domain_part):
|
|
||||||
try:
|
|
||||||
domain_part = domain_part.encode('idna').decode('ascii')
|
|
||||||
except UnicodeError:
|
|
||||||
self.error(self.error_msg % value)
|
|
||||||
else:
|
|
||||||
if not self.validate_domain_part(domain_part):
|
|
||||||
self.error(self.error_msg % value)
|
|
||||||
|
|
||||||
|
|
||||||
class IntField(BaseField):
|
class IntField(BaseField):
|
||||||
"""32-bit integer field."""
|
"""32-bit integer field."""
|
||||||
@@ -705,14 +619,6 @@ class DynamicField(BaseField):
|
|||||||
|
|
||||||
Used by :class:`~mongoengine.DynamicDocument` to handle dynamic data"""
|
Used by :class:`~mongoengine.DynamicDocument` to handle dynamic data"""
|
||||||
|
|
||||||
def __init__(self, container_class=dict, *args, **kwargs):
|
|
||||||
self._container_cls = container_class
|
|
||||||
if not issubclass(self._container_cls, Mapping):
|
|
||||||
self.error('The class that is specified in `container_class` parameter '
|
|
||||||
'must be a subclass of `dict`.')
|
|
||||||
|
|
||||||
super(DynamicField, self).__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
def to_mongo(self, value, use_db_field=True, fields=None):
|
def to_mongo(self, value, use_db_field=True, fields=None):
|
||||||
"""Convert a Python type to a MongoDB compatible type.
|
"""Convert a Python type to a MongoDB compatible type.
|
||||||
"""
|
"""
|
||||||
@@ -738,7 +644,7 @@ class DynamicField(BaseField):
|
|||||||
is_list = True
|
is_list = True
|
||||||
value = {k: v for k, v in enumerate(value)}
|
value = {k: v for k, v in enumerate(value)}
|
||||||
|
|
||||||
data = self._container_cls()
|
data = {}
|
||||||
for k, v in value.iteritems():
|
for k, v in value.iteritems():
|
||||||
data[k] = self.to_mongo(v, use_db_field, fields)
|
data[k] = self.to_mongo(v, use_db_field, fields)
|
||||||
|
|
||||||
@@ -1092,8 +998,8 @@ class ReferenceField(BaseField):
|
|||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
|
|
||||||
if not isinstance(value, (self.document_type, DBRef, ObjectId)):
|
if not isinstance(value, (self.document_type, DBRef)):
|
||||||
self.error('A ReferenceField only accepts DBRef, ObjectId or documents')
|
self.error('A ReferenceField only accepts DBRef or documents')
|
||||||
|
|
||||||
if isinstance(value, Document) and value.id is None:
|
if isinstance(value, Document) and value.id is None:
|
||||||
self.error('You can only reference documents once they have been '
|
self.error('You can only reference documents once they have been '
|
||||||
|
@@ -158,49 +158,44 @@ class BaseQuerySet(object):
|
|||||||
# self._cursor
|
# self._cursor
|
||||||
|
|
||||||
def __getitem__(self, key):
|
def __getitem__(self, key):
|
||||||
"""Return a document instance corresponding to a given index if
|
"""Support skip and limit using getitem and slicing syntax."""
|
||||||
the key is an integer. If the key is a slice, translate its
|
|
||||||
bounds into a skip and a limit, and return a cloned queryset
|
|
||||||
with that skip/limit applied. For example:
|
|
||||||
|
|
||||||
>>> User.objects[0]
|
|
||||||
<User: User object>
|
|
||||||
>>> User.objects[1:3]
|
|
||||||
[<User: User object>, <User: User object>]
|
|
||||||
"""
|
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
|
|
||||||
# Handle a slice
|
# Slice provided
|
||||||
if isinstance(key, slice):
|
if isinstance(key, slice):
|
||||||
queryset._cursor_obj = queryset._cursor[key]
|
try:
|
||||||
queryset._skip, queryset._limit = key.start, key.stop
|
queryset._cursor_obj = queryset._cursor[key]
|
||||||
if key.start and key.stop:
|
queryset._skip, queryset._limit = key.start, key.stop
|
||||||
queryset._limit = key.stop - key.start
|
if key.start and key.stop:
|
||||||
|
queryset._limit = key.stop - key.start
|
||||||
|
except IndexError as err:
|
||||||
|
# PyMongo raises an error if key.start == key.stop, catch it,
|
||||||
|
# bin it, kill it.
|
||||||
|
start = key.start or 0
|
||||||
|
if start >= 0 and key.stop >= 0 and key.step is None:
|
||||||
|
if start == key.stop:
|
||||||
|
queryset.limit(0)
|
||||||
|
queryset._skip = key.start
|
||||||
|
queryset._limit = key.stop - start
|
||||||
|
return queryset
|
||||||
|
raise err
|
||||||
# Allow further QuerySet modifications to be performed
|
# Allow further QuerySet modifications to be performed
|
||||||
return queryset
|
return queryset
|
||||||
|
# Integer index provided
|
||||||
# Handle an index
|
|
||||||
elif isinstance(key, int):
|
elif isinstance(key, int):
|
||||||
if queryset._scalar:
|
if queryset._scalar:
|
||||||
return queryset._get_scalar(
|
return queryset._get_scalar(
|
||||||
queryset._document._from_son(
|
queryset._document._from_son(queryset._cursor[key],
|
||||||
queryset._cursor[key],
|
_auto_dereference=self._auto_dereference,
|
||||||
_auto_dereference=self._auto_dereference,
|
only_fields=self.only_fields))
|
||||||
only_fields=self.only_fields
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
if queryset._as_pymongo:
|
if queryset._as_pymongo:
|
||||||
return queryset._get_as_pymongo(queryset._cursor[key])
|
return queryset._get_as_pymongo(queryset._cursor[key])
|
||||||
|
return queryset._document._from_son(queryset._cursor[key],
|
||||||
|
_auto_dereference=self._auto_dereference,
|
||||||
|
only_fields=self.only_fields)
|
||||||
|
|
||||||
return queryset._document._from_son(
|
raise AttributeError
|
||||||
queryset._cursor[key],
|
|
||||||
_auto_dereference=self._auto_dereference,
|
|
||||||
only_fields=self.only_fields
|
|
||||||
)
|
|
||||||
|
|
||||||
raise AttributeError('Provide a slice or an integer index')
|
|
||||||
|
|
||||||
def __iter__(self):
|
def __iter__(self):
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
@@ -291,7 +286,7 @@ class BaseQuerySet(object):
|
|||||||
|
|
||||||
.. versionadded:: 0.4
|
.. versionadded:: 0.4
|
||||||
"""
|
"""
|
||||||
return self._document(**kwargs).save(force_insert=True)
|
return self._document(**kwargs).save()
|
||||||
|
|
||||||
def first(self):
|
def first(self):
|
||||||
"""Retrieve the first object matching the query."""
|
"""Retrieve the first object matching the query."""
|
||||||
|
@@ -2,14 +2,14 @@
|
|||||||
import unittest
|
import unittest
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
from nose.plugins.skip import SkipTest
|
|
||||||
from datetime import datetime
|
|
||||||
import pymongo
|
import pymongo
|
||||||
|
|
||||||
from mongoengine import *
|
from nose.plugins.skip import SkipTest
|
||||||
from mongoengine.connection import get_db
|
from datetime import datetime
|
||||||
|
|
||||||
from tests.utils import get_mongodb_version, needs_mongodb_v26
|
from mongoengine import *
|
||||||
|
from mongoengine.connection import get_db, get_connection
|
||||||
|
|
||||||
__all__ = ("IndexesTest", )
|
__all__ = ("IndexesTest", )
|
||||||
|
|
||||||
@@ -412,6 +412,7 @@ class IndexesTest(unittest.TestCase):
|
|||||||
User.ensure_indexes()
|
User.ensure_indexes()
|
||||||
info = User.objects._collection.index_information()
|
info = User.objects._collection.index_information()
|
||||||
self.assertEqual(sorted(info.keys()), ['_cls_1_user_guid_1', '_id_'])
|
self.assertEqual(sorted(info.keys()), ['_cls_1_user_guid_1', '_id_'])
|
||||||
|
User.drop_collection()
|
||||||
|
|
||||||
def test_embedded_document_index(self):
|
def test_embedded_document_index(self):
|
||||||
"""Tests settings an index on an embedded document
|
"""Tests settings an index on an embedded document
|
||||||
@@ -433,6 +434,7 @@ class IndexesTest(unittest.TestCase):
|
|||||||
|
|
||||||
info = BlogPost.objects._collection.index_information()
|
info = BlogPost.objects._collection.index_information()
|
||||||
self.assertEqual(sorted(info.keys()), ['_id_', 'date.yr_-1'])
|
self.assertEqual(sorted(info.keys()), ['_id_', 'date.yr_-1'])
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_list_embedded_document_index(self):
|
def test_list_embedded_document_index(self):
|
||||||
"""Ensure list embedded documents can be indexed
|
"""Ensure list embedded documents can be indexed
|
||||||
@@ -459,6 +461,7 @@ class IndexesTest(unittest.TestCase):
|
|||||||
post1 = BlogPost(title="Embedded Indexes tests in place",
|
post1 = BlogPost(title="Embedded Indexes tests in place",
|
||||||
tags=[Tag(name="about"), Tag(name="time")])
|
tags=[Tag(name="about"), Tag(name="time")])
|
||||||
post1.save()
|
post1.save()
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_recursive_embedded_objects_dont_break_indexes(self):
|
def test_recursive_embedded_objects_dont_break_indexes(self):
|
||||||
|
|
||||||
@@ -491,7 +494,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
obj = Test(a=1)
|
obj = Test(a=1)
|
||||||
obj.save()
|
obj.save()
|
||||||
|
|
||||||
IS_MONGODB_3 = get_mongodb_version()[0] >= 3
|
connection = get_connection()
|
||||||
|
IS_MONGODB_3 = connection.server_info()['versionArray'][0] >= 3
|
||||||
|
|
||||||
# Need to be explicit about covered indexes as mongoDB doesn't know if
|
# Need to be explicit about covered indexes as mongoDB doesn't know if
|
||||||
# the documents returned might have more keys in that here.
|
# the documents returned might have more keys in that here.
|
||||||
@@ -619,6 +623,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
post3 = BlogPost(title='test3', date=Date(year=2010), slug='test')
|
post3 = BlogPost(title='test3', date=Date(year=2010), slug='test')
|
||||||
self.assertRaises(OperationError, post3.save)
|
self.assertRaises(OperationError, post3.save)
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_unique_embedded_document(self):
|
def test_unique_embedded_document(self):
|
||||||
"""Ensure that uniqueness constraints are applied to fields on embedded documents.
|
"""Ensure that uniqueness constraints are applied to fields on embedded documents.
|
||||||
"""
|
"""
|
||||||
@@ -646,6 +652,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
sub=SubDocument(year=2010, slug='test'))
|
sub=SubDocument(year=2010, slug='test'))
|
||||||
self.assertRaises(NotUniqueError, post3.save)
|
self.assertRaises(NotUniqueError, post3.save)
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_unique_embedded_document_in_list(self):
|
def test_unique_embedded_document_in_list(self):
|
||||||
"""
|
"""
|
||||||
Ensure that the uniqueness constraints are applied to fields in
|
Ensure that the uniqueness constraints are applied to fields in
|
||||||
@@ -676,6 +684,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
|
|
||||||
self.assertRaises(NotUniqueError, post2.save)
|
self.assertRaises(NotUniqueError, post2.save)
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_unique_with_embedded_document_and_embedded_unique(self):
|
def test_unique_with_embedded_document_and_embedded_unique(self):
|
||||||
"""Ensure that uniqueness constraints are applied to fields on
|
"""Ensure that uniqueness constraints are applied to fields on
|
||||||
embedded documents. And work with unique_with as well.
|
embedded documents. And work with unique_with as well.
|
||||||
@@ -709,6 +719,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
sub=SubDocument(year=2009, slug='test-1'))
|
sub=SubDocument(year=2009, slug='test-1'))
|
||||||
self.assertRaises(NotUniqueError, post3.save)
|
self.assertRaises(NotUniqueError, post3.save)
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_ttl_indexes(self):
|
def test_ttl_indexes(self):
|
||||||
|
|
||||||
class Log(Document):
|
class Log(Document):
|
||||||
@@ -721,6 +733,14 @@ class IndexesTest(unittest.TestCase):
|
|||||||
|
|
||||||
Log.drop_collection()
|
Log.drop_collection()
|
||||||
|
|
||||||
|
if pymongo.version_tuple[0] < 2 and pymongo.version_tuple[1] < 3:
|
||||||
|
raise SkipTest('pymongo needs to be 2.3 or higher for this test')
|
||||||
|
|
||||||
|
connection = get_connection()
|
||||||
|
version_array = connection.server_info()['versionArray']
|
||||||
|
if version_array[0] < 2 and version_array[1] < 2:
|
||||||
|
raise SkipTest('MongoDB needs to be 2.2 or higher for this test')
|
||||||
|
|
||||||
# Indexes are lazy so use list() to perform query
|
# Indexes are lazy so use list() to perform query
|
||||||
list(Log.objects)
|
list(Log.objects)
|
||||||
info = Log.objects._collection.index_information()
|
info = Log.objects._collection.index_information()
|
||||||
@@ -748,11 +768,13 @@ class IndexesTest(unittest.TestCase):
|
|||||||
raise AssertionError("We saved a dupe!")
|
raise AssertionError("We saved a dupe!")
|
||||||
except NotUniqueError:
|
except NotUniqueError:
|
||||||
pass
|
pass
|
||||||
|
Customer.drop_collection()
|
||||||
|
|
||||||
def test_unique_and_primary(self):
|
def test_unique_and_primary(self):
|
||||||
"""If you set a field as primary, then unexpected behaviour can occur.
|
"""If you set a field as primary, then unexpected behaviour can occur.
|
||||||
You won't create a duplicate but you will update an existing document.
|
You won't create a duplicate but you will update an existing document.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField(primary_key=True, unique=True)
|
name = StringField(primary_key=True, unique=True)
|
||||||
password = StringField()
|
password = StringField()
|
||||||
@@ -768,23 +790,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
self.assertEqual(User.objects.count(), 1)
|
self.assertEqual(User.objects.count(), 1)
|
||||||
self.assertEqual(User.objects.get().password, 'secret2')
|
self.assertEqual(User.objects.get().password, 'secret2')
|
||||||
|
|
||||||
def test_unique_and_primary_create(self):
|
|
||||||
"""Create a new record with a duplicate primary key
|
|
||||||
throws an exception
|
|
||||||
"""
|
|
||||||
class User(Document):
|
|
||||||
name = StringField(primary_key=True)
|
|
||||||
password = StringField()
|
|
||||||
|
|
||||||
User.drop_collection()
|
User.drop_collection()
|
||||||
|
|
||||||
User.objects.create(name='huangz', password='secret')
|
|
||||||
with self.assertRaises(NotUniqueError):
|
|
||||||
User.objects.create(name='huangz', password='secret2')
|
|
||||||
|
|
||||||
self.assertEqual(User.objects.count(), 1)
|
|
||||||
self.assertEqual(User.objects.get().password, 'secret')
|
|
||||||
|
|
||||||
def test_index_with_pk(self):
|
def test_index_with_pk(self):
|
||||||
"""Ensure you can use `pk` as part of a query"""
|
"""Ensure you can use `pk` as part of a query"""
|
||||||
|
|
||||||
@@ -867,8 +874,8 @@ class IndexesTest(unittest.TestCase):
|
|||||||
info['provider_ids.foo_1_provider_ids.bar_1']['key'])
|
info['provider_ids.foo_1_provider_ids.bar_1']['key'])
|
||||||
self.assertTrue(info['provider_ids.foo_1_provider_ids.bar_1']['sparse'])
|
self.assertTrue(info['provider_ids.foo_1_provider_ids.bar_1']['sparse'])
|
||||||
|
|
||||||
@needs_mongodb_v26
|
|
||||||
def test_text_indexes(self):
|
def test_text_indexes(self):
|
||||||
|
|
||||||
class Book(Document):
|
class Book(Document):
|
||||||
title = DictField()
|
title = DictField()
|
||||||
meta = {
|
meta = {
|
||||||
|
@@ -28,6 +28,8 @@ TEST_IMAGE_PATH = os.path.join(os.path.dirname(__file__),
|
|||||||
__all__ = ("InstanceTest",)
|
__all__ = ("InstanceTest",)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class InstanceTest(unittest.TestCase):
|
class InstanceTest(unittest.TestCase):
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
@@ -70,7 +72,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(field._instance, instance)
|
self.assertEqual(field._instance, instance)
|
||||||
|
|
||||||
def test_capped_collection(self):
|
def test_capped_collection(self):
|
||||||
"""Ensure that capped collections work properly."""
|
"""Ensure that capped collections work properly.
|
||||||
|
"""
|
||||||
class Log(Document):
|
class Log(Document):
|
||||||
date = DateTimeField(default=datetime.now)
|
date = DateTimeField(default=datetime.now)
|
||||||
meta = {
|
meta = {
|
||||||
@@ -178,7 +181,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual('<Article: привет мир>', repr(doc))
|
self.assertEqual('<Article: привет мир>', repr(doc))
|
||||||
|
|
||||||
def test_repr_none(self):
|
def test_repr_none(self):
|
||||||
"""Ensure None values are handled correctly."""
|
"""Ensure None values handled correctly
|
||||||
|
"""
|
||||||
class Article(Document):
|
class Article(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
|
|
||||||
@@ -186,23 +190,25 @@ class InstanceTest(unittest.TestCase):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
doc = Article(title=u'привет мир')
|
doc = Article(title=u'привет мир')
|
||||||
|
|
||||||
self.assertEqual('<Article: None>', repr(doc))
|
self.assertEqual('<Article: None>', repr(doc))
|
||||||
|
|
||||||
def test_queryset_resurrects_dropped_collection(self):
|
def test_queryset_resurrects_dropped_collection(self):
|
||||||
self.Person.drop_collection()
|
self.Person.drop_collection()
|
||||||
|
|
||||||
self.assertEqual([], list(self.Person.objects()))
|
self.assertEqual([], list(self.Person.objects()))
|
||||||
|
|
||||||
# Ensure works correctly with inhertited classes
|
|
||||||
class Actor(self.Person):
|
class Actor(self.Person):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
# Ensure works correctly with inhertited classes
|
||||||
Actor.objects()
|
Actor.objects()
|
||||||
self.Person.drop_collection()
|
self.Person.drop_collection()
|
||||||
self.assertEqual([], list(Actor.objects()))
|
self.assertEqual([], list(Actor.objects()))
|
||||||
|
|
||||||
def test_polymorphic_references(self):
|
def test_polymorphic_references(self):
|
||||||
"""Ensure that the correct subclasses are returned from a query
|
"""Ensure that the correct subclasses are returned from a query when
|
||||||
when using references / generic references
|
using references / generic references
|
||||||
"""
|
"""
|
||||||
class Animal(Document):
|
class Animal(Document):
|
||||||
meta = {'allow_inheritance': True}
|
meta = {'allow_inheritance': True}
|
||||||
@@ -252,6 +258,9 @@ class InstanceTest(unittest.TestCase):
|
|||||||
classes = [a.__class__ for a in Zoo.objects.first().animals]
|
classes = [a.__class__ for a in Zoo.objects.first().animals]
|
||||||
self.assertEqual(classes, [Animal, Fish, Mammal, Dog, Human])
|
self.assertEqual(classes, [Animal, Fish, Mammal, Dog, Human])
|
||||||
|
|
||||||
|
Zoo.drop_collection()
|
||||||
|
Animal.drop_collection()
|
||||||
|
|
||||||
def test_reference_inheritance(self):
|
def test_reference_inheritance(self):
|
||||||
class Stats(Document):
|
class Stats(Document):
|
||||||
created = DateTimeField(default=datetime.now)
|
created = DateTimeField(default=datetime.now)
|
||||||
@@ -278,7 +287,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(list_stats, CompareStats.objects.first().stats)
|
self.assertEqual(list_stats, CompareStats.objects.first().stats)
|
||||||
|
|
||||||
def test_db_field_load(self):
|
def test_db_field_load(self):
|
||||||
"""Ensure we load data correctly from the right db field."""
|
"""Ensure we load data correctly
|
||||||
|
"""
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField(required=True)
|
name = StringField(required=True)
|
||||||
_rank = StringField(required=False, db_field="rank")
|
_rank = StringField(required=False, db_field="rank")
|
||||||
@@ -297,7 +307,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(Person.objects.get(name="Fred").rank, "Private")
|
self.assertEqual(Person.objects.get(name="Fred").rank, "Private")
|
||||||
|
|
||||||
def test_db_embedded_doc_field_load(self):
|
def test_db_embedded_doc_field_load(self):
|
||||||
"""Ensure we load embedded document data correctly."""
|
"""Ensure we load embedded document data correctly
|
||||||
|
"""
|
||||||
class Rank(EmbeddedDocument):
|
class Rank(EmbeddedDocument):
|
||||||
title = StringField(required=True)
|
title = StringField(required=True)
|
||||||
|
|
||||||
@@ -322,7 +333,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(Person.objects.get(name="Fred").rank, "Private")
|
self.assertEqual(Person.objects.get(name="Fred").rank, "Private")
|
||||||
|
|
||||||
def test_custom_id_field(self):
|
def test_custom_id_field(self):
|
||||||
"""Ensure that documents may be created with custom primary keys."""
|
"""Ensure that documents may be created with custom primary keys.
|
||||||
|
"""
|
||||||
class User(Document):
|
class User(Document):
|
||||||
username = StringField(primary_key=True)
|
username = StringField(primary_key=True)
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -370,7 +382,10 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(user_son['_id'], 'mongo')
|
self.assertEqual(user_son['_id'], 'mongo')
|
||||||
self.assertTrue('username' not in user_son['_id'])
|
self.assertTrue('username' not in user_son['_id'])
|
||||||
|
|
||||||
|
User.drop_collection()
|
||||||
|
|
||||||
def test_document_not_registered(self):
|
def test_document_not_registered(self):
|
||||||
|
|
||||||
class Place(Document):
|
class Place(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -392,6 +407,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
list(Place.objects.all())
|
list(Place.objects.all())
|
||||||
|
|
||||||
def test_document_registry_regressions(self):
|
def test_document_registry_regressions(self):
|
||||||
|
|
||||||
class Location(Document):
|
class Location(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
meta = {'allow_inheritance': True}
|
meta = {'allow_inheritance': True}
|
||||||
@@ -405,16 +421,18 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(Area, get_document("Location.Area"))
|
self.assertEqual(Area, get_document("Location.Area"))
|
||||||
|
|
||||||
def test_creation(self):
|
def test_creation(self):
|
||||||
"""Ensure that document may be created using keyword arguments."""
|
"""Ensure that document may be created using keyword arguments.
|
||||||
|
"""
|
||||||
person = self.Person(name="Test User", age=30)
|
person = self.Person(name="Test User", age=30)
|
||||||
self.assertEqual(person.name, "Test User")
|
self.assertEqual(person.name, "Test User")
|
||||||
self.assertEqual(person.age, 30)
|
self.assertEqual(person.age, 30)
|
||||||
|
|
||||||
def test_to_dbref(self):
|
def test_to_dbref(self):
|
||||||
"""Ensure that you can get a dbref of a document."""
|
"""Ensure that you can get a dbref of a document"""
|
||||||
person = self.Person(name="Test User", age=30)
|
person = self.Person(name="Test User", age=30)
|
||||||
self.assertRaises(OperationError, person.to_dbref)
|
self.assertRaises(OperationError, person.to_dbref)
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
person.to_dbref()
|
person.to_dbref()
|
||||||
|
|
||||||
def test_save_abstract_document(self):
|
def test_save_abstract_document(self):
|
||||||
@@ -427,7 +445,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
Doc(name='aaa').save()
|
Doc(name='aaa').save()
|
||||||
|
|
||||||
def test_reload(self):
|
def test_reload(self):
|
||||||
"""Ensure that attributes may be reloaded."""
|
"""Ensure that attributes may be reloaded.
|
||||||
|
"""
|
||||||
person = self.Person(name="Test User", age=20)
|
person = self.Person(name="Test User", age=20)
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
@@ -460,6 +479,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
doc = Animal(superphylum='Deuterostomia')
|
doc = Animal(superphylum='Deuterostomia')
|
||||||
doc.save()
|
doc.save()
|
||||||
doc.reload()
|
doc.reload()
|
||||||
|
Animal.drop_collection()
|
||||||
|
|
||||||
def test_reload_sharded_nested(self):
|
def test_reload_sharded_nested(self):
|
||||||
class SuperPhylum(EmbeddedDocument):
|
class SuperPhylum(EmbeddedDocument):
|
||||||
@@ -473,9 +493,11 @@ class InstanceTest(unittest.TestCase):
|
|||||||
doc = Animal(superphylum=SuperPhylum(name='Deuterostomia'))
|
doc = Animal(superphylum=SuperPhylum(name='Deuterostomia'))
|
||||||
doc.save()
|
doc.save()
|
||||||
doc.reload()
|
doc.reload()
|
||||||
|
Animal.drop_collection()
|
||||||
|
|
||||||
def test_reload_referencing(self):
|
def test_reload_referencing(self):
|
||||||
"""Ensures reloading updates weakrefs correctly."""
|
"""Ensures reloading updates weakrefs correctly
|
||||||
|
"""
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
dict_field = DictField()
|
dict_field = DictField()
|
||||||
list_field = ListField()
|
list_field = ListField()
|
||||||
@@ -547,7 +569,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertFalse("Threw wrong exception")
|
self.assertFalse("Threw wrong exception")
|
||||||
|
|
||||||
def test_reload_of_non_strict_with_special_field_name(self):
|
def test_reload_of_non_strict_with_special_field_name(self):
|
||||||
"""Ensures reloading works for documents with meta strict == False."""
|
"""Ensures reloading works for documents with meta strict == False
|
||||||
|
"""
|
||||||
class Post(Document):
|
class Post(Document):
|
||||||
meta = {
|
meta = {
|
||||||
'strict': False
|
'strict': False
|
||||||
@@ -568,7 +591,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(post.items, ["more lorem", "even more ipsum"])
|
self.assertEqual(post.items, ["more lorem", "even more ipsum"])
|
||||||
|
|
||||||
def test_dictionary_access(self):
|
def test_dictionary_access(self):
|
||||||
"""Ensure that dictionary-style field access works properly."""
|
"""Ensure that dictionary-style field access works properly.
|
||||||
|
"""
|
||||||
person = self.Person(name='Test User', age=30, job=self.Job())
|
person = self.Person(name='Test User', age=30, job=self.Job())
|
||||||
self.assertEqual(person['name'], 'Test User')
|
self.assertEqual(person['name'], 'Test User')
|
||||||
|
|
||||||
@@ -610,7 +634,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(sub_doc.to_mongo().keys(), ['id'])
|
self.assertEqual(sub_doc.to_mongo().keys(), ['id'])
|
||||||
|
|
||||||
def test_embedded_document(self):
|
def test_embedded_document(self):
|
||||||
"""Ensure that embedded documents are set up correctly."""
|
"""Ensure that embedded documents are set up correctly.
|
||||||
|
"""
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
|
|
||||||
@@ -618,7 +643,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertFalse('id' in Comment._fields)
|
self.assertFalse('id' in Comment._fields)
|
||||||
|
|
||||||
def test_embedded_document_instance(self):
|
def test_embedded_document_instance(self):
|
||||||
"""Ensure that embedded documents can reference parent instance."""
|
"""Ensure that embedded documents can reference parent instance
|
||||||
|
"""
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
string = StringField()
|
string = StringField()
|
||||||
|
|
||||||
@@ -626,7 +652,6 @@ class InstanceTest(unittest.TestCase):
|
|||||||
embedded_field = EmbeddedDocumentField(Embedded)
|
embedded_field = EmbeddedDocumentField(Embedded)
|
||||||
|
|
||||||
Doc.drop_collection()
|
Doc.drop_collection()
|
||||||
|
|
||||||
doc = Doc(embedded_field=Embedded(string="Hi"))
|
doc = Doc(embedded_field=Embedded(string="Hi"))
|
||||||
self.assertHasInstance(doc.embedded_field, doc)
|
self.assertHasInstance(doc.embedded_field, doc)
|
||||||
|
|
||||||
@@ -636,8 +661,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
|
|
||||||
def test_embedded_document_complex_instance(self):
|
def test_embedded_document_complex_instance(self):
|
||||||
"""Ensure that embedded documents in complex fields can reference
|
"""Ensure that embedded documents in complex fields can reference
|
||||||
parent instance.
|
parent instance"""
|
||||||
"""
|
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
string = StringField()
|
string = StringField()
|
||||||
|
|
||||||
@@ -653,7 +677,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertHasInstance(doc.embedded_field[0], doc)
|
self.assertHasInstance(doc.embedded_field[0], doc)
|
||||||
|
|
||||||
def test_embedded_document_complex_instance_no_use_db_field(self):
|
def test_embedded_document_complex_instance_no_use_db_field(self):
|
||||||
"""Ensure that use_db_field is propagated to list of Emb Docs."""
|
"""Ensure that use_db_field is propagated to list of Emb Docs
|
||||||
|
"""
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
string = StringField(db_field='s')
|
string = StringField(db_field='s')
|
||||||
|
|
||||||
@@ -665,6 +690,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(d['embedded_field'], [{'string': 'Hi'}])
|
self.assertEqual(d['embedded_field'], [{'string': 'Hi'}])
|
||||||
|
|
||||||
def test_instance_is_set_on_setattr(self):
|
def test_instance_is_set_on_setattr(self):
|
||||||
|
|
||||||
class Email(EmbeddedDocument):
|
class Email(EmbeddedDocument):
|
||||||
email = EmailField()
|
email = EmailField()
|
||||||
|
|
||||||
@@ -672,7 +698,6 @@ class InstanceTest(unittest.TestCase):
|
|||||||
email = EmbeddedDocumentField(Email)
|
email = EmbeddedDocumentField(Email)
|
||||||
|
|
||||||
Account.drop_collection()
|
Account.drop_collection()
|
||||||
|
|
||||||
acc = Account()
|
acc = Account()
|
||||||
acc.email = Email(email='test@example.com')
|
acc.email = Email(email='test@example.com')
|
||||||
self.assertHasInstance(acc._data["email"], acc)
|
self.assertHasInstance(acc._data["email"], acc)
|
||||||
@@ -682,6 +707,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertHasInstance(acc1._data["email"], acc1)
|
self.assertHasInstance(acc1._data["email"], acc1)
|
||||||
|
|
||||||
def test_instance_is_set_on_setattr_on_embedded_document_list(self):
|
def test_instance_is_set_on_setattr_on_embedded_document_list(self):
|
||||||
|
|
||||||
class Email(EmbeddedDocument):
|
class Email(EmbeddedDocument):
|
||||||
email = EmailField()
|
email = EmailField()
|
||||||
|
|
||||||
@@ -827,28 +853,32 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertDbEqual([dict(other_doc.to_mongo()), dict(doc.to_mongo())])
|
self.assertDbEqual([dict(other_doc.to_mongo()), dict(doc.to_mongo())])
|
||||||
|
|
||||||
def test_save(self):
|
def test_save(self):
|
||||||
"""Ensure that a document may be saved in the database."""
|
"""Ensure that a document may be saved in the database.
|
||||||
|
"""
|
||||||
# Create person object and save it to the database
|
# Create person object and save it to the database
|
||||||
person = self.Person(name='Test User', age=30)
|
person = self.Person(name='Test User', age=30)
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
# Ensure that the object is in the database
|
# Ensure that the object is in the database
|
||||||
collection = self.db[self.Person._get_collection_name()]
|
collection = self.db[self.Person._get_collection_name()]
|
||||||
person_obj = collection.find_one({'name': 'Test User'})
|
person_obj = collection.find_one({'name': 'Test User'})
|
||||||
self.assertEqual(person_obj['name'], 'Test User')
|
self.assertEqual(person_obj['name'], 'Test User')
|
||||||
self.assertEqual(person_obj['age'], 30)
|
self.assertEqual(person_obj['age'], 30)
|
||||||
self.assertEqual(person_obj['_id'], person.id)
|
self.assertEqual(person_obj['_id'], person.id)
|
||||||
|
|
||||||
# Test skipping validation on save
|
# Test skipping validation on save
|
||||||
|
|
||||||
class Recipient(Document):
|
class Recipient(Document):
|
||||||
email = EmailField(required=True)
|
email = EmailField(required=True)
|
||||||
|
|
||||||
recipient = Recipient(email='not-an-email')
|
recipient = Recipient(email='root@localhost')
|
||||||
self.assertRaises(ValidationError, recipient.save)
|
self.assertRaises(ValidationError, recipient.save)
|
||||||
recipient.save(validate=False)
|
|
||||||
|
try:
|
||||||
|
recipient.save(validate=False)
|
||||||
|
except ValidationError:
|
||||||
|
self.fail()
|
||||||
|
|
||||||
def test_save_to_a_value_that_equates_to_false(self):
|
def test_save_to_a_value_that_equates_to_false(self):
|
||||||
|
|
||||||
class Thing(EmbeddedDocument):
|
class Thing(EmbeddedDocument):
|
||||||
count = IntField()
|
count = IntField()
|
||||||
|
|
||||||
@@ -868,6 +898,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(user.thing.count, 0)
|
self.assertEqual(user.thing.count, 0)
|
||||||
|
|
||||||
def test_save_max_recursion_not_hit(self):
|
def test_save_max_recursion_not_hit(self):
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
parent = ReferenceField('self')
|
parent = ReferenceField('self')
|
||||||
@@ -893,6 +924,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
p0.save()
|
p0.save()
|
||||||
|
|
||||||
def test_save_max_recursion_not_hit_with_file_field(self):
|
def test_save_max_recursion_not_hit_with_file_field(self):
|
||||||
|
|
||||||
class Foo(Document):
|
class Foo(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
picture = FileField()
|
picture = FileField()
|
||||||
@@ -916,6 +948,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(b.picture, b.bar.picture, b.bar.bar.picture)
|
self.assertEqual(b.picture, b.bar.picture, b.bar.bar.picture)
|
||||||
|
|
||||||
def test_save_cascades(self):
|
def test_save_cascades(self):
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
parent = ReferenceField('self')
|
parent = ReferenceField('self')
|
||||||
@@ -938,6 +971,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(p1.name, p.parent.name)
|
self.assertEqual(p1.name, p.parent.name)
|
||||||
|
|
||||||
def test_save_cascade_kwargs(self):
|
def test_save_cascade_kwargs(self):
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
parent = ReferenceField('self')
|
parent = ReferenceField('self')
|
||||||
@@ -958,6 +992,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(p1.name, p2.parent.name)
|
self.assertEqual(p1.name, p2.parent.name)
|
||||||
|
|
||||||
def test_save_cascade_meta_false(self):
|
def test_save_cascade_meta_false(self):
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
parent = ReferenceField('self')
|
parent = ReferenceField('self')
|
||||||
@@ -986,6 +1021,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(p1.name, p.parent.name)
|
self.assertEqual(p1.name, p.parent.name)
|
||||||
|
|
||||||
def test_save_cascade_meta_true(self):
|
def test_save_cascade_meta_true(self):
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
parent = ReferenceField('self')
|
parent = ReferenceField('self')
|
||||||
@@ -1010,6 +1046,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertNotEqual(p1.name, p.parent.name)
|
self.assertNotEqual(p1.name, p.parent.name)
|
||||||
|
|
||||||
def test_save_cascades_generically(self):
|
def test_save_cascades_generically(self):
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
parent = GenericReferenceField()
|
parent = GenericReferenceField()
|
||||||
@@ -1035,6 +1072,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(p1.name, p.parent.name)
|
self.assertEqual(p1.name, p.parent.name)
|
||||||
|
|
||||||
def test_save_atomicity_condition(self):
|
def test_save_atomicity_condition(self):
|
||||||
|
|
||||||
class Widget(Document):
|
class Widget(Document):
|
||||||
toggle = BooleanField(default=False)
|
toggle = BooleanField(default=False)
|
||||||
count = IntField(default=0)
|
count = IntField(default=0)
|
||||||
@@ -1112,8 +1150,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
|
|
||||||
def test_update(self):
|
def test_update(self):
|
||||||
"""Ensure that an existing document is updated instead of be
|
"""Ensure that an existing document is updated instead of be
|
||||||
overwritten.
|
overwritten."""
|
||||||
"""
|
|
||||||
# Create person object and save it to the database
|
# Create person object and save it to the database
|
||||||
person = self.Person(name='Test User', age=30)
|
person = self.Person(name='Test User', age=30)
|
||||||
person.save()
|
person.save()
|
||||||
@@ -1217,6 +1254,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(2, self.Person.objects.count())
|
self.assertEqual(2, self.Person.objects.count())
|
||||||
|
|
||||||
def test_can_save_if_not_included(self):
|
def test_can_save_if_not_included(self):
|
||||||
|
|
||||||
class EmbeddedDoc(EmbeddedDocument):
|
class EmbeddedDoc(EmbeddedDocument):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -1303,7 +1341,10 @@ class InstanceTest(unittest.TestCase):
|
|||||||
doc2.update(set__name=doc1.name)
|
doc2.update(set__name=doc1.name)
|
||||||
|
|
||||||
def test_embedded_update(self):
|
def test_embedded_update(self):
|
||||||
"""Test update on `EmbeddedDocumentField` fields."""
|
"""
|
||||||
|
Test update on `EmbeddedDocumentField` fields
|
||||||
|
"""
|
||||||
|
|
||||||
class Page(EmbeddedDocument):
|
class Page(EmbeddedDocument):
|
||||||
log_message = StringField(verbose_name="Log message",
|
log_message = StringField(verbose_name="Log message",
|
||||||
required=True)
|
required=True)
|
||||||
@@ -1324,9 +1365,11 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(site.page.log_message, "Error: Dummy message")
|
self.assertEqual(site.page.log_message, "Error: Dummy message")
|
||||||
|
|
||||||
def test_embedded_update_db_field(self):
|
def test_embedded_update_db_field(self):
|
||||||
"""Test update on `EmbeddedDocumentField` fields when db_field
|
|
||||||
is other than default.
|
|
||||||
"""
|
"""
|
||||||
|
Test update on `EmbeddedDocumentField` fields when db_field is other
|
||||||
|
than default.
|
||||||
|
"""
|
||||||
|
|
||||||
class Page(EmbeddedDocument):
|
class Page(EmbeddedDocument):
|
||||||
log_message = StringField(verbose_name="Log message",
|
log_message = StringField(verbose_name="Log message",
|
||||||
db_field="page_log_message",
|
db_field="page_log_message",
|
||||||
@@ -1349,7 +1392,9 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(site.page.log_message, "Error: Dummy message")
|
self.assertEqual(site.page.log_message, "Error: Dummy message")
|
||||||
|
|
||||||
def test_save_only_changed_fields(self):
|
def test_save_only_changed_fields(self):
|
||||||
"""Ensure save only sets / unsets changed fields."""
|
"""Ensure save only sets / unsets changed fields
|
||||||
|
"""
|
||||||
|
|
||||||
class User(self.Person):
|
class User(self.Person):
|
||||||
active = BooleanField(default=True)
|
active = BooleanField(default=True)
|
||||||
|
|
||||||
@@ -1469,8 +1514,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(q, 3)
|
self.assertEqual(q, 3)
|
||||||
|
|
||||||
def test_set_unset_one_operation(self):
|
def test_set_unset_one_operation(self):
|
||||||
"""Ensure that $set and $unset actions are performed in the
|
"""Ensure that $set and $unset actions are performed in the same
|
||||||
same operation.
|
operation.
|
||||||
"""
|
"""
|
||||||
class FooBar(Document):
|
class FooBar(Document):
|
||||||
foo = StringField(default=None)
|
foo = StringField(default=None)
|
||||||
@@ -1491,7 +1536,9 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(1, q)
|
self.assertEqual(1, q)
|
||||||
|
|
||||||
def test_save_only_changed_fields_recursive(self):
|
def test_save_only_changed_fields_recursive(self):
|
||||||
"""Ensure save only sets / unsets changed fields."""
|
"""Ensure save only sets / unsets changed fields
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
published = BooleanField(default=True)
|
published = BooleanField(default=True)
|
||||||
|
|
||||||
@@ -1531,7 +1578,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertFalse(person.comments_dict['first_post'].published)
|
self.assertFalse(person.comments_dict['first_post'].published)
|
||||||
|
|
||||||
def test_delete(self):
|
def test_delete(self):
|
||||||
"""Ensure that document may be deleted using the delete method."""
|
"""Ensure that document may be deleted using the delete method.
|
||||||
|
"""
|
||||||
person = self.Person(name="Test User", age=30)
|
person = self.Person(name="Test User", age=30)
|
||||||
person.save()
|
person.save()
|
||||||
self.assertEqual(self.Person.objects.count(), 1)
|
self.assertEqual(self.Person.objects.count(), 1)
|
||||||
@@ -1539,34 +1587,33 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(self.Person.objects.count(), 0)
|
self.assertEqual(self.Person.objects.count(), 0)
|
||||||
|
|
||||||
def test_save_custom_id(self):
|
def test_save_custom_id(self):
|
||||||
"""Ensure that a document may be saved with a custom _id."""
|
"""Ensure that a document may be saved with a custom _id.
|
||||||
|
"""
|
||||||
# Create person object and save it to the database
|
# Create person object and save it to the database
|
||||||
person = self.Person(name='Test User', age=30,
|
person = self.Person(name='Test User', age=30,
|
||||||
id='497ce96f395f2f052a494fd4')
|
id='497ce96f395f2f052a494fd4')
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
# Ensure that the object is in the database with the correct _id
|
# Ensure that the object is in the database with the correct _id
|
||||||
collection = self.db[self.Person._get_collection_name()]
|
collection = self.db[self.Person._get_collection_name()]
|
||||||
person_obj = collection.find_one({'name': 'Test User'})
|
person_obj = collection.find_one({'name': 'Test User'})
|
||||||
self.assertEqual(str(person_obj['_id']), '497ce96f395f2f052a494fd4')
|
self.assertEqual(str(person_obj['_id']), '497ce96f395f2f052a494fd4')
|
||||||
|
|
||||||
def test_save_custom_pk(self):
|
def test_save_custom_pk(self):
|
||||||
"""Ensure that a document may be saved with a custom _id using
|
"""
|
||||||
pk alias.
|
Ensure that a document may be saved with a custom _id using pk alias.
|
||||||
"""
|
"""
|
||||||
# Create person object and save it to the database
|
# Create person object and save it to the database
|
||||||
person = self.Person(name='Test User', age=30,
|
person = self.Person(name='Test User', age=30,
|
||||||
pk='497ce96f395f2f052a494fd4')
|
pk='497ce96f395f2f052a494fd4')
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
# Ensure that the object is in the database with the correct _id
|
# Ensure that the object is in the database with the correct _id
|
||||||
collection = self.db[self.Person._get_collection_name()]
|
collection = self.db[self.Person._get_collection_name()]
|
||||||
person_obj = collection.find_one({'name': 'Test User'})
|
person_obj = collection.find_one({'name': 'Test User'})
|
||||||
self.assertEqual(str(person_obj['_id']), '497ce96f395f2f052a494fd4')
|
self.assertEqual(str(person_obj['_id']), '497ce96f395f2f052a494fd4')
|
||||||
|
|
||||||
def test_save_list(self):
|
def test_save_list(self):
|
||||||
"""Ensure that a list field may be properly saved."""
|
"""Ensure that a list field may be properly saved.
|
||||||
|
"""
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
|
|
||||||
@@ -1589,6 +1636,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
for comment_obj, comment in zip(post_obj['comments'], comments):
|
for comment_obj, comment in zip(post_obj['comments'], comments):
|
||||||
self.assertEqual(comment_obj['content'], comment['content'])
|
self.assertEqual(comment_obj['content'], comment['content'])
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_list_search_by_embedded(self):
|
def test_list_search_by_embedded(self):
|
||||||
class User(Document):
|
class User(Document):
|
||||||
username = StringField(required=True)
|
username = StringField(required=True)
|
||||||
@@ -1648,8 +1697,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
list(Page.objects.filter(comments__user=u3)))
|
list(Page.objects.filter(comments__user=u3)))
|
||||||
|
|
||||||
def test_save_embedded_document(self):
|
def test_save_embedded_document(self):
|
||||||
"""Ensure that a document with an embedded document field may
|
"""Ensure that a document with an embedded document field may be
|
||||||
be saved in the database.
|
saved in the database.
|
||||||
"""
|
"""
|
||||||
class EmployeeDetails(EmbeddedDocument):
|
class EmployeeDetails(EmbeddedDocument):
|
||||||
position = StringField()
|
position = StringField()
|
||||||
@@ -1668,13 +1717,13 @@ class InstanceTest(unittest.TestCase):
|
|||||||
employee_obj = collection.find_one({'name': 'Test Employee'})
|
employee_obj = collection.find_one({'name': 'Test Employee'})
|
||||||
self.assertEqual(employee_obj['name'], 'Test Employee')
|
self.assertEqual(employee_obj['name'], 'Test Employee')
|
||||||
self.assertEqual(employee_obj['age'], 50)
|
self.assertEqual(employee_obj['age'], 50)
|
||||||
|
|
||||||
# Ensure that the 'details' embedded object saved correctly
|
# Ensure that the 'details' embedded object saved correctly
|
||||||
self.assertEqual(employee_obj['details']['position'], 'Developer')
|
self.assertEqual(employee_obj['details']['position'], 'Developer')
|
||||||
|
|
||||||
def test_embedded_update_after_save(self):
|
def test_embedded_update_after_save(self):
|
||||||
"""Test update of `EmbeddedDocumentField` attached to a newly
|
"""
|
||||||
saved document.
|
Test update of `EmbeddedDocumentField` attached to a newly saved
|
||||||
|
document.
|
||||||
"""
|
"""
|
||||||
class Page(EmbeddedDocument):
|
class Page(EmbeddedDocument):
|
||||||
log_message = StringField(verbose_name="Log message",
|
log_message = StringField(verbose_name="Log message",
|
||||||
@@ -1695,8 +1744,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(site.page.log_message, "Error: Dummy message")
|
self.assertEqual(site.page.log_message, "Error: Dummy message")
|
||||||
|
|
||||||
def test_updating_an_embedded_document(self):
|
def test_updating_an_embedded_document(self):
|
||||||
"""Ensure that a document with an embedded document field may
|
"""Ensure that a document with an embedded document field may be
|
||||||
be saved in the database.
|
saved in the database.
|
||||||
"""
|
"""
|
||||||
class EmployeeDetails(EmbeddedDocument):
|
class EmployeeDetails(EmbeddedDocument):
|
||||||
position = StringField()
|
position = StringField()
|
||||||
@@ -1731,6 +1780,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(promoted_employee.details, None)
|
self.assertEqual(promoted_employee.details, None)
|
||||||
|
|
||||||
def test_object_mixins(self):
|
def test_object_mixins(self):
|
||||||
|
|
||||||
class NameMixin(object):
|
class NameMixin(object):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -1769,9 +1819,9 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(t.count, 12)
|
self.assertEqual(t.count, 12)
|
||||||
|
|
||||||
def test_save_reference(self):
|
def test_save_reference(self):
|
||||||
"""Ensure that a document reference field may be saved in the
|
"""Ensure that a document reference field may be saved in the database.
|
||||||
database.
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
meta = {'collection': 'blogpost_1'}
|
meta = {'collection': 'blogpost_1'}
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -1802,6 +1852,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
author = list(self.Person.objects(name='Test User'))[-1]
|
author = list(self.Person.objects(name='Test User'))[-1]
|
||||||
self.assertEqual(author.age, 25)
|
self.assertEqual(author.age, 25)
|
||||||
|
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_duplicate_db_fields_raise_invalid_document_error(self):
|
def test_duplicate_db_fields_raise_invalid_document_error(self):
|
||||||
"""Ensure a InvalidDocumentError is thrown if duplicate fields
|
"""Ensure a InvalidDocumentError is thrown if duplicate fields
|
||||||
declare the same db_field.
|
declare the same db_field.
|
||||||
@@ -1812,7 +1864,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
name2 = StringField(db_field='name')
|
name2 = StringField(db_field='name')
|
||||||
|
|
||||||
def test_invalid_son(self):
|
def test_invalid_son(self):
|
||||||
"""Raise an error if loading invalid data."""
|
"""Raise an error if loading invalid data"""
|
||||||
class Occurrence(EmbeddedDocument):
|
class Occurrence(EmbeddedDocument):
|
||||||
number = IntField()
|
number = IntField()
|
||||||
|
|
||||||
@@ -1835,9 +1887,9 @@ class InstanceTest(unittest.TestCase):
|
|||||||
Word._from_son('this is not a valid SON dict')
|
Word._from_son('this is not a valid SON dict')
|
||||||
|
|
||||||
def test_reverse_delete_rule_cascade_and_nullify(self):
|
def test_reverse_delete_rule_cascade_and_nullify(self):
|
||||||
"""Ensure that a referenced document is also deleted upon
|
"""Ensure that a referenced document is also deleted upon deletion.
|
||||||
deletion.
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
author = ReferenceField(self.Person, reverse_delete_rule=CASCADE)
|
author = ReferenceField(self.Person, reverse_delete_rule=CASCADE)
|
||||||
@@ -1892,8 +1944,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(Book.objects.count(), 0)
|
self.assertEqual(Book.objects.count(), 0)
|
||||||
|
|
||||||
def test_reverse_delete_rule_with_shared_id_among_collections(self):
|
def test_reverse_delete_rule_with_shared_id_among_collections(self):
|
||||||
"""Ensure that cascade delete rule doesn't mix id among
|
"""Ensure that cascade delete rule doesn't mix id among collections.
|
||||||
collections.
|
|
||||||
"""
|
"""
|
||||||
class User(Document):
|
class User(Document):
|
||||||
id = IntField(primary_key=True)
|
id = IntField(primary_key=True)
|
||||||
@@ -1924,9 +1975,10 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(Book.objects.get(), book_2)
|
self.assertEqual(Book.objects.get(), book_2)
|
||||||
|
|
||||||
def test_reverse_delete_rule_with_document_inheritance(self):
|
def test_reverse_delete_rule_with_document_inheritance(self):
|
||||||
"""Ensure that a referenced document is also deleted upon
|
"""Ensure that a referenced document is also deleted upon deletion
|
||||||
deletion of a child document.
|
of a child document.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class Writer(self.Person):
|
class Writer(self.Person):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -1958,9 +2010,10 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(BlogPost.objects.count(), 0)
|
self.assertEqual(BlogPost.objects.count(), 0)
|
||||||
|
|
||||||
def test_reverse_delete_rule_cascade_and_nullify_complex_field(self):
|
def test_reverse_delete_rule_cascade_and_nullify_complex_field(self):
|
||||||
"""Ensure that a referenced document is also deleted upon
|
"""Ensure that a referenced document is also deleted upon deletion for
|
||||||
deletion for complex fields.
|
complex fields.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
authors = ListField(ReferenceField(
|
authors = ListField(ReferenceField(
|
||||||
@@ -1969,6 +2022,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.Person, reverse_delete_rule=NULLIFY))
|
self.Person, reverse_delete_rule=NULLIFY))
|
||||||
|
|
||||||
self.Person.drop_collection()
|
self.Person.drop_collection()
|
||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
author = self.Person(name='Test User')
|
author = self.Person(name='Test User')
|
||||||
@@ -1992,10 +2046,10 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(BlogPost.objects.count(), 0)
|
self.assertEqual(BlogPost.objects.count(), 0)
|
||||||
|
|
||||||
def test_reverse_delete_rule_cascade_triggers_pre_delete_signal(self):
|
def test_reverse_delete_rule_cascade_triggers_pre_delete_signal(self):
|
||||||
"""Ensure the pre_delete signal is triggered upon a cascading
|
""" ensure the pre_delete signal is triggered upon a cascading deletion
|
||||||
deletion setup a blog post with content, an author and editor
|
setup a blog post with content, an author and editor
|
||||||
delete the author which triggers deletion of blogpost via
|
delete the author which triggers deletion of blogpost via cascade
|
||||||
cascade blog post's pre_delete signal alters an editor attribute.
|
blog post's pre_delete signal alters an editor attribute
|
||||||
"""
|
"""
|
||||||
class Editor(self.Person):
|
class Editor(self.Person):
|
||||||
review_queue = IntField(default=0)
|
review_queue = IntField(default=0)
|
||||||
@@ -2023,7 +2077,6 @@ class InstanceTest(unittest.TestCase):
|
|||||||
|
|
||||||
# delete the author, the post is also deleted due to the CASCADE rule
|
# delete the author, the post is also deleted due to the CASCADE rule
|
||||||
author.delete()
|
author.delete()
|
||||||
|
|
||||||
# the pre-delete signal should have decremented the editor's queue
|
# the pre-delete signal should have decremented the editor's queue
|
||||||
editor = Editor.objects(name='Max P.').get()
|
editor = Editor.objects(name='Max P.').get()
|
||||||
self.assertEqual(editor.review_queue, 0)
|
self.assertEqual(editor.review_queue, 0)
|
||||||
@@ -2032,6 +2085,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
"""Ensure that Bi-Directional relationships work with
|
"""Ensure that Bi-Directional relationships work with
|
||||||
reverse_delete_rule
|
reverse_delete_rule
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class Bar(Document):
|
class Bar(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
foo = ReferenceField('Foo')
|
foo = ReferenceField('Foo')
|
||||||
@@ -2077,8 +2131,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
mother = ReferenceField('Person', reverse_delete_rule=DENY)
|
mother = ReferenceField('Person', reverse_delete_rule=DENY)
|
||||||
|
|
||||||
def test_reverse_delete_rule_cascade_recurs(self):
|
def test_reverse_delete_rule_cascade_recurs(self):
|
||||||
"""Ensure that a chain of documents is also deleted upon
|
"""Ensure that a chain of documents is also deleted upon cascaded
|
||||||
cascaded deletion.
|
deletion.
|
||||||
"""
|
"""
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -2108,10 +2162,15 @@ class InstanceTest(unittest.TestCase):
|
|||||||
author.delete()
|
author.delete()
|
||||||
self.assertEqual(Comment.objects.count(), 0)
|
self.assertEqual(Comment.objects.count(), 0)
|
||||||
|
|
||||||
|
self.Person.drop_collection()
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
Comment.drop_collection()
|
||||||
|
|
||||||
def test_reverse_delete_rule_deny(self):
|
def test_reverse_delete_rule_deny(self):
|
||||||
"""Ensure that a document cannot be referenced if there are
|
"""Ensure that a document cannot be referenced if there are still
|
||||||
still documents referring to it.
|
documents referring to it.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
author = ReferenceField(self.Person, reverse_delete_rule=DENY)
|
author = ReferenceField(self.Person, reverse_delete_rule=DENY)
|
||||||
@@ -2139,7 +2198,11 @@ class InstanceTest(unittest.TestCase):
|
|||||||
author.delete()
|
author.delete()
|
||||||
self.assertEqual(self.Person.objects.count(), 1)
|
self.assertEqual(self.Person.objects.count(), 1)
|
||||||
|
|
||||||
|
self.Person.drop_collection()
|
||||||
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def subclasses_and_unique_keys_works(self):
|
def subclasses_and_unique_keys_works(self):
|
||||||
|
|
||||||
class A(Document):
|
class A(Document):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -2155,16 +2218,19 @@ class InstanceTest(unittest.TestCase):
|
|||||||
|
|
||||||
self.assertEqual(A.objects.count(), 2)
|
self.assertEqual(A.objects.count(), 2)
|
||||||
self.assertEqual(B.objects.count(), 1)
|
self.assertEqual(B.objects.count(), 1)
|
||||||
|
A.drop_collection()
|
||||||
|
B.drop_collection()
|
||||||
|
|
||||||
def test_document_hash(self):
|
def test_document_hash(self):
|
||||||
"""Test document in list, dict, set."""
|
"""Test document in list, dict, set
|
||||||
|
"""
|
||||||
class User(Document):
|
class User(Document):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
# Clear old data
|
# Clear old datas
|
||||||
User.drop_collection()
|
User.drop_collection()
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
@@ -2176,18 +2242,17 @@ class InstanceTest(unittest.TestCase):
|
|||||||
b1 = BlogPost.objects.create()
|
b1 = BlogPost.objects.create()
|
||||||
b2 = BlogPost.objects.create()
|
b2 = BlogPost.objects.create()
|
||||||
|
|
||||||
# Make sure docs are properly identified in a list (__eq__ is used
|
# in List
|
||||||
# for the comparison).
|
|
||||||
all_user_list = list(User.objects.all())
|
all_user_list = list(User.objects.all())
|
||||||
|
|
||||||
self.assertTrue(u1 in all_user_list)
|
self.assertTrue(u1 in all_user_list)
|
||||||
self.assertTrue(u2 in all_user_list)
|
self.assertTrue(u2 in all_user_list)
|
||||||
self.assertTrue(u3 in all_user_list)
|
self.assertTrue(u3 in all_user_list)
|
||||||
self.assertTrue(u4 not in all_user_list) # New object
|
self.assertFalse(u4 in all_user_list) # New object
|
||||||
self.assertTrue(b1 not in all_user_list) # Other object
|
self.assertFalse(b1 in all_user_list) # Other object
|
||||||
self.assertTrue(b2 not in all_user_list) # Other object
|
self.assertFalse(b2 in all_user_list) # Other object
|
||||||
|
|
||||||
# Make sure docs can be used as keys in a dict (__hash__ is used
|
# in Dict
|
||||||
# for hashing the docs).
|
|
||||||
all_user_dic = {}
|
all_user_dic = {}
|
||||||
for u in User.objects.all():
|
for u in User.objects.all():
|
||||||
all_user_dic[u] = "OK"
|
all_user_dic[u] = "OK"
|
||||||
@@ -2199,22 +2264,13 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(all_user_dic.get(b1, False), False) # Other object
|
self.assertEqual(all_user_dic.get(b1, False), False) # Other object
|
||||||
self.assertEqual(all_user_dic.get(b2, False), False) # Other object
|
self.assertEqual(all_user_dic.get(b2, False), False) # Other object
|
||||||
|
|
||||||
# Make sure docs are properly identified in a set (__hash__ is used
|
# in Set
|
||||||
# for hashing the docs).
|
|
||||||
all_user_set = set(User.objects.all())
|
all_user_set = set(User.objects.all())
|
||||||
self.assertTrue(u1 in all_user_set)
|
|
||||||
self.assertTrue(u4 not in all_user_set)
|
|
||||||
self.assertTrue(b1 not in all_user_list)
|
|
||||||
self.assertTrue(b2 not in all_user_list)
|
|
||||||
|
|
||||||
# Make sure duplicate docs aren't accepted in the set
|
self.assertTrue(u1 in all_user_set)
|
||||||
self.assertEqual(len(all_user_set), 3)
|
|
||||||
all_user_set.add(u1)
|
|
||||||
all_user_set.add(u2)
|
|
||||||
all_user_set.add(u3)
|
|
||||||
self.assertEqual(len(all_user_set), 3)
|
|
||||||
|
|
||||||
def test_picklable(self):
|
def test_picklable(self):
|
||||||
|
|
||||||
pickle_doc = PickleTest(number=1, string="One", lists=['1', '2'])
|
pickle_doc = PickleTest(number=1, string="One", lists=['1', '2'])
|
||||||
pickle_doc.embedded = PickleEmbedded()
|
pickle_doc.embedded = PickleEmbedded()
|
||||||
pickled_doc = pickle.dumps(pickle_doc) # make sure pickling works even before the doc is saved
|
pickled_doc = pickle.dumps(pickle_doc) # make sure pickling works even before the doc is saved
|
||||||
@@ -2240,6 +2296,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(pickle_doc.lists, ["1", "2", "3"])
|
self.assertEqual(pickle_doc.lists, ["1", "2", "3"])
|
||||||
|
|
||||||
def test_regular_document_pickle(self):
|
def test_regular_document_pickle(self):
|
||||||
|
|
||||||
pickle_doc = PickleTest(number=1, string="One", lists=['1', '2'])
|
pickle_doc = PickleTest(number=1, string="One", lists=['1', '2'])
|
||||||
pickled_doc = pickle.dumps(pickle_doc) # make sure pickling works even before the doc is saved
|
pickled_doc = pickle.dumps(pickle_doc) # make sure pickling works even before the doc is saved
|
||||||
pickle_doc.save()
|
pickle_doc.save()
|
||||||
@@ -2262,6 +2319,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
fixtures.PickleTest = PickleTest
|
fixtures.PickleTest = PickleTest
|
||||||
|
|
||||||
def test_dynamic_document_pickle(self):
|
def test_dynamic_document_pickle(self):
|
||||||
|
|
||||||
pickle_doc = PickleDynamicTest(
|
pickle_doc = PickleDynamicTest(
|
||||||
name="test", number=1, string="One", lists=['1', '2'])
|
name="test", number=1, string="One", lists=['1', '2'])
|
||||||
pickle_doc.embedded = PickleDynamicEmbedded(foo="Bar")
|
pickle_doc.embedded = PickleDynamicEmbedded(foo="Bar")
|
||||||
@@ -2300,6 +2358,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
validate = DictField()
|
validate = DictField()
|
||||||
|
|
||||||
def test_mutating_documents(self):
|
def test_mutating_documents(self):
|
||||||
|
|
||||||
class B(EmbeddedDocument):
|
class B(EmbeddedDocument):
|
||||||
field1 = StringField(default='field1')
|
field1 = StringField(default='field1')
|
||||||
|
|
||||||
@@ -2307,7 +2366,6 @@ class InstanceTest(unittest.TestCase):
|
|||||||
b = EmbeddedDocumentField(B, default=lambda: B())
|
b = EmbeddedDocumentField(B, default=lambda: B())
|
||||||
|
|
||||||
A.drop_collection()
|
A.drop_collection()
|
||||||
|
|
||||||
a = A()
|
a = A()
|
||||||
a.save()
|
a.save()
|
||||||
a.reload()
|
a.reload()
|
||||||
@@ -2331,13 +2389,12 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(a.b.field2.c_field, 'new value')
|
self.assertEqual(a.b.field2.c_field, 'new value')
|
||||||
|
|
||||||
def test_can_save_false_values(self):
|
def test_can_save_false_values(self):
|
||||||
"""Ensures you can save False values on save."""
|
"""Ensures you can save False values on save"""
|
||||||
class Doc(Document):
|
class Doc(Document):
|
||||||
foo = StringField()
|
foo = StringField()
|
||||||
archived = BooleanField(default=False, required=True)
|
archived = BooleanField(default=False, required=True)
|
||||||
|
|
||||||
Doc.drop_collection()
|
Doc.drop_collection()
|
||||||
|
|
||||||
d = Doc()
|
d = Doc()
|
||||||
d.save()
|
d.save()
|
||||||
d.archived = False
|
d.archived = False
|
||||||
@@ -2346,12 +2403,11 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(Doc.objects(archived=False).count(), 1)
|
self.assertEqual(Doc.objects(archived=False).count(), 1)
|
||||||
|
|
||||||
def test_can_save_false_values_dynamic(self):
|
def test_can_save_false_values_dynamic(self):
|
||||||
"""Ensures you can save False values on dynamic docs."""
|
"""Ensures you can save False values on dynamic docs"""
|
||||||
class Doc(DynamicDocument):
|
class Doc(DynamicDocument):
|
||||||
foo = StringField()
|
foo = StringField()
|
||||||
|
|
||||||
Doc.drop_collection()
|
Doc.drop_collection()
|
||||||
|
|
||||||
d = Doc()
|
d = Doc()
|
||||||
d.save()
|
d.save()
|
||||||
d.archived = False
|
d.archived = False
|
||||||
@@ -2391,7 +2447,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
Collection.update = orig_update
|
Collection.update = orig_update
|
||||||
|
|
||||||
def test_db_alias_tests(self):
|
def test_db_alias_tests(self):
|
||||||
"""DB Alias tests."""
|
""" DB Alias tests """
|
||||||
# mongoenginetest - Is default connection alias from setUp()
|
# mongoenginetest - Is default connection alias from setUp()
|
||||||
# Register Aliases
|
# Register Aliases
|
||||||
register_connection('testdb-1', 'mongoenginetest2')
|
register_connection('testdb-1', 'mongoenginetest2')
|
||||||
@@ -2453,7 +2509,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
get_db("testdb-3")[AuthorBooks._get_collection_name()])
|
get_db("testdb-3")[AuthorBooks._get_collection_name()])
|
||||||
|
|
||||||
def test_db_alias_overrides(self):
|
def test_db_alias_overrides(self):
|
||||||
"""Test db_alias can be overriden."""
|
"""db_alias can be overriden
|
||||||
|
"""
|
||||||
# Register a connection with db_alias testdb-2
|
# Register a connection with db_alias testdb-2
|
||||||
register_connection('testdb-2', 'mongoenginetest2')
|
register_connection('testdb-2', 'mongoenginetest2')
|
||||||
|
|
||||||
@@ -2477,7 +2534,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
B._get_collection().database.name)
|
B._get_collection().database.name)
|
||||||
|
|
||||||
def test_db_alias_propagates(self):
|
def test_db_alias_propagates(self):
|
||||||
"""db_alias propagates?"""
|
"""db_alias propagates?
|
||||||
|
"""
|
||||||
register_connection('testdb-1', 'mongoenginetest2')
|
register_connection('testdb-1', 'mongoenginetest2')
|
||||||
|
|
||||||
class A(Document):
|
class A(Document):
|
||||||
@@ -2490,7 +2548,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual('testdb-1', B._meta.get('db_alias'))
|
self.assertEqual('testdb-1', B._meta.get('db_alias'))
|
||||||
|
|
||||||
def test_db_ref_usage(self):
|
def test_db_ref_usage(self):
|
||||||
"""DB Ref usage in dict_fields."""
|
""" DB Ref usage in dict_fields"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -2725,6 +2784,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(user.thing._data['data'], [1, 2, 3])
|
self.assertEqual(user.thing._data['data'], [1, 2, 3])
|
||||||
|
|
||||||
def test_spaces_in_keys(self):
|
def test_spaces_in_keys(self):
|
||||||
|
|
||||||
class Embedded(DynamicEmbeddedDocument):
|
class Embedded(DynamicEmbeddedDocument):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -2813,6 +2873,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
log.machine = "127.0.0.1"
|
log.machine = "127.0.0.1"
|
||||||
|
|
||||||
def test_kwargs_simple(self):
|
def test_kwargs_simple(self):
|
||||||
|
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -2832,6 +2893,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(classic_doc._data, dict_doc._data)
|
self.assertEqual(classic_doc._data, dict_doc._data)
|
||||||
|
|
||||||
def test_kwargs_complex(self):
|
def test_kwargs_complex(self):
|
||||||
|
|
||||||
class Embedded(EmbeddedDocument):
|
class Embedded(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -2854,35 +2916,36 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(classic_doc._data, dict_doc._data)
|
self.assertEqual(classic_doc._data, dict_doc._data)
|
||||||
|
|
||||||
def test_positional_creation(self):
|
def test_positional_creation(self):
|
||||||
"""Ensure that document may be created using positional arguments."""
|
"""Ensure that document may be created using positional arguments.
|
||||||
|
"""
|
||||||
person = self.Person("Test User", 42)
|
person = self.Person("Test User", 42)
|
||||||
self.assertEqual(person.name, "Test User")
|
self.assertEqual(person.name, "Test User")
|
||||||
self.assertEqual(person.age, 42)
|
self.assertEqual(person.age, 42)
|
||||||
|
|
||||||
def test_mixed_creation(self):
|
def test_mixed_creation(self):
|
||||||
"""Ensure that document may be created using mixed arguments."""
|
"""Ensure that document may be created using mixed arguments.
|
||||||
|
"""
|
||||||
person = self.Person("Test User", age=42)
|
person = self.Person("Test User", age=42)
|
||||||
self.assertEqual(person.name, "Test User")
|
self.assertEqual(person.name, "Test User")
|
||||||
self.assertEqual(person.age, 42)
|
self.assertEqual(person.age, 42)
|
||||||
|
|
||||||
def test_positional_creation_embedded(self):
|
def test_positional_creation_embedded(self):
|
||||||
"""Ensure that embedded document may be created using positional
|
"""Ensure that embedded document may be created using positional arguments.
|
||||||
arguments.
|
|
||||||
"""
|
"""
|
||||||
job = self.Job("Test Job", 4)
|
job = self.Job("Test Job", 4)
|
||||||
self.assertEqual(job.name, "Test Job")
|
self.assertEqual(job.name, "Test Job")
|
||||||
self.assertEqual(job.years, 4)
|
self.assertEqual(job.years, 4)
|
||||||
|
|
||||||
def test_mixed_creation_embedded(self):
|
def test_mixed_creation_embedded(self):
|
||||||
"""Ensure that embedded document may be created using mixed
|
"""Ensure that embedded document may be created using mixed arguments.
|
||||||
arguments.
|
|
||||||
"""
|
"""
|
||||||
job = self.Job("Test Job", years=4)
|
job = self.Job("Test Job", years=4)
|
||||||
self.assertEqual(job.name, "Test Job")
|
self.assertEqual(job.name, "Test Job")
|
||||||
self.assertEqual(job.years, 4)
|
self.assertEqual(job.years, 4)
|
||||||
|
|
||||||
def test_mixed_creation_dynamic(self):
|
def test_mixed_creation_dynamic(self):
|
||||||
"""Ensure that document may be created using mixed arguments."""
|
"""Ensure that document may be created using mixed arguments.
|
||||||
|
"""
|
||||||
class Person(DynamicDocument):
|
class Person(DynamicDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -2891,14 +2954,14 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(person.age, 42)
|
self.assertEqual(person.age, 42)
|
||||||
|
|
||||||
def test_bad_mixed_creation(self):
|
def test_bad_mixed_creation(self):
|
||||||
"""Ensure that document gives correct error when duplicating
|
"""Ensure that document gives correct error when duplicating arguments
|
||||||
arguments.
|
|
||||||
"""
|
"""
|
||||||
with self.assertRaises(TypeError):
|
with self.assertRaises(TypeError):
|
||||||
return self.Person("Test User", 42, name="Bad User")
|
return self.Person("Test User", 42, name="Bad User")
|
||||||
|
|
||||||
def test_data_contains_id_field(self):
|
def test_data_contains_id_field(self):
|
||||||
"""Ensure that asking for _data returns 'id'."""
|
"""Ensure that asking for _data returns 'id'
|
||||||
|
"""
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
@@ -2910,6 +2973,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(person._data.get('id'), person.id)
|
self.assertEqual(person._data.get('id'), person.id)
|
||||||
|
|
||||||
def test_complex_nesting_document_and_embedded_document(self):
|
def test_complex_nesting_document_and_embedded_document(self):
|
||||||
|
|
||||||
class Macro(EmbeddedDocument):
|
class Macro(EmbeddedDocument):
|
||||||
value = DynamicField(default="UNDEFINED")
|
value = DynamicField(default="UNDEFINED")
|
||||||
|
|
||||||
@@ -2952,6 +3016,7 @@ class InstanceTest(unittest.TestCase):
|
|||||||
system.nodes["node"].parameters["param"].macros["test"].value)
|
system.nodes["node"].parameters["param"].macros["test"].value)
|
||||||
|
|
||||||
def test_embedded_document_equality(self):
|
def test_embedded_document_equality(self):
|
||||||
|
|
||||||
class Test(Document):
|
class Test(Document):
|
||||||
field = StringField(required=True)
|
field = StringField(required=True)
|
||||||
|
|
||||||
@@ -3137,7 +3202,8 @@ class InstanceTest(unittest.TestCase):
|
|||||||
self.assertEqual(idx, 2)
|
self.assertEqual(idx, 2)
|
||||||
|
|
||||||
def test_falsey_pk(self):
|
def test_falsey_pk(self):
|
||||||
"""Ensure that we can create and update a document with Falsey PK."""
|
"""Ensure that we can create and update a document with Falsey PK.
|
||||||
|
"""
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
age = IntField(primary_key=True)
|
age = IntField(primary_key=True)
|
||||||
height = FloatField()
|
height = FloatField()
|
||||||
|
File diff suppressed because it is too large
Load Diff
@@ -1,139 +1,105 @@
|
|||||||
import datetime
|
from datetime import datetime, timedelta
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
|
from pymongo.errors import OperationFailure
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
from mongoengine.connection import get_connection
|
||||||
from tests.utils import MongoDBTestCase, needs_mongodb_v3
|
from nose.plugins.skip import SkipTest
|
||||||
|
|
||||||
|
|
||||||
__all__ = ("GeoQueriesTest",)
|
__all__ = ("GeoQueriesTest",)
|
||||||
|
|
||||||
|
|
||||||
class GeoQueriesTest(MongoDBTestCase):
|
class GeoQueriesTest(unittest.TestCase):
|
||||||
|
|
||||||
def _create_event_data(self, point_field_class=GeoPointField):
|
def setUp(self):
|
||||||
"""Create some sample data re-used in many of the tests below."""
|
connect(db='mongoenginetest')
|
||||||
|
|
||||||
|
def test_geospatial_operators(self):
|
||||||
|
"""Ensure that geospatial queries are working.
|
||||||
|
"""
|
||||||
class Event(Document):
|
class Event(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
date = DateTimeField()
|
date = DateTimeField()
|
||||||
location = point_field_class()
|
location = GeoPointField()
|
||||||
|
|
||||||
def __unicode__(self):
|
def __unicode__(self):
|
||||||
return self.title
|
return self.title
|
||||||
|
|
||||||
self.Event = Event
|
|
||||||
|
|
||||||
Event.drop_collection()
|
Event.drop_collection()
|
||||||
|
|
||||||
event1 = Event.objects.create(
|
event1 = Event(title="Coltrane Motion @ Double Door",
|
||||||
title="Coltrane Motion @ Double Door",
|
date=datetime.now() - timedelta(days=1),
|
||||||
date=datetime.datetime.now() - datetime.timedelta(days=1),
|
location=[-87.677137, 41.909889]).save()
|
||||||
location=[-87.677137, 41.909889])
|
event2 = Event(title="Coltrane Motion @ Bottom of the Hill",
|
||||||
event2 = Event.objects.create(
|
date=datetime.now() - timedelta(days=10),
|
||||||
title="Coltrane Motion @ Bottom of the Hill",
|
location=[-122.4194155, 37.7749295]).save()
|
||||||
date=datetime.datetime.now() - datetime.timedelta(days=10),
|
event3 = Event(title="Coltrane Motion @ Empty Bottle",
|
||||||
location=[-122.4194155, 37.7749295])
|
date=datetime.now(),
|
||||||
event3 = Event.objects.create(
|
location=[-87.686638, 41.900474]).save()
|
||||||
title="Coltrane Motion @ Empty Bottle",
|
|
||||||
date=datetime.datetime.now(),
|
|
||||||
location=[-87.686638, 41.900474])
|
|
||||||
|
|
||||||
return event1, event2, event3
|
|
||||||
|
|
||||||
def test_near(self):
|
|
||||||
"""Make sure the "near" operator works."""
|
|
||||||
event1, event2, event3 = self._create_event_data()
|
|
||||||
|
|
||||||
# find all events "near" pitchfork office, chicago.
|
# find all events "near" pitchfork office, chicago.
|
||||||
# note that "near" will show the san francisco event, too,
|
# note that "near" will show the san francisco event, too,
|
||||||
# although it sorts to last.
|
# although it sorts to last.
|
||||||
events = self.Event.objects(location__near=[-87.67892, 41.9120459])
|
events = Event.objects(location__near=[-87.67892, 41.9120459])
|
||||||
self.assertEqual(events.count(), 3)
|
self.assertEqual(events.count(), 3)
|
||||||
self.assertEqual(list(events), [event1, event3, event2])
|
self.assertEqual(list(events), [event1, event3, event2])
|
||||||
|
|
||||||
# ensure ordering is respected by "near"
|
|
||||||
events = self.Event.objects(location__near=[-87.67892, 41.9120459])
|
|
||||||
events = events.order_by("-date")
|
|
||||||
self.assertEqual(events.count(), 3)
|
|
||||||
self.assertEqual(list(events), [event3, event1, event2])
|
|
||||||
|
|
||||||
def test_near_and_max_distance(self):
|
|
||||||
"""Ensure the "max_distance" operator works alongside the "near"
|
|
||||||
operator.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data()
|
|
||||||
|
|
||||||
# find events within 10 degrees of san francisco
|
|
||||||
point = [-122.415579, 37.7566023]
|
|
||||||
events = self.Event.objects(location__near=point,
|
|
||||||
location__max_distance=10)
|
|
||||||
self.assertEqual(events.count(), 1)
|
|
||||||
self.assertEqual(events[0], event2)
|
|
||||||
|
|
||||||
# $minDistance was added in MongoDB v2.6, but continued being buggy
|
|
||||||
# until v3.0; skip for older versions
|
|
||||||
@needs_mongodb_v3
|
|
||||||
def test_near_and_min_distance(self):
|
|
||||||
"""Ensure the "min_distance" operator works alongside the "near"
|
|
||||||
operator.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data()
|
|
||||||
|
|
||||||
# find events at least 10 degrees away of san francisco
|
|
||||||
point = [-122.415579, 37.7566023]
|
|
||||||
events = self.Event.objects(location__near=point,
|
|
||||||
location__min_distance=10)
|
|
||||||
self.assertEqual(events.count(), 2)
|
|
||||||
|
|
||||||
def test_within_distance(self):
|
|
||||||
"""Make sure the "within_distance" operator works."""
|
|
||||||
event1, event2, event3 = self._create_event_data()
|
|
||||||
|
|
||||||
# find events within 5 degrees of pitchfork office, chicago
|
# find events within 5 degrees of pitchfork office, chicago
|
||||||
point_and_distance = [[-87.67892, 41.9120459], 5]
|
point_and_distance = [[-87.67892, 41.9120459], 5]
|
||||||
events = self.Event.objects(
|
events = Event.objects(location__within_distance=point_and_distance)
|
||||||
location__within_distance=point_and_distance)
|
|
||||||
self.assertEqual(events.count(), 2)
|
self.assertEqual(events.count(), 2)
|
||||||
events = list(events)
|
events = list(events)
|
||||||
self.assertTrue(event2 not in events)
|
self.assertTrue(event2 not in events)
|
||||||
self.assertTrue(event1 in events)
|
self.assertTrue(event1 in events)
|
||||||
self.assertTrue(event3 in events)
|
self.assertTrue(event3 in events)
|
||||||
|
|
||||||
|
# ensure ordering is respected by "near"
|
||||||
|
events = Event.objects(location__near=[-87.67892, 41.9120459])
|
||||||
|
events = events.order_by("-date")
|
||||||
|
self.assertEqual(events.count(), 3)
|
||||||
|
self.assertEqual(list(events), [event3, event1, event2])
|
||||||
|
|
||||||
|
# find events within 10 degrees of san francisco
|
||||||
|
point = [-122.415579, 37.7566023]
|
||||||
|
events = Event.objects(location__near=point, location__max_distance=10)
|
||||||
|
self.assertEqual(events.count(), 1)
|
||||||
|
self.assertEqual(events[0], event2)
|
||||||
|
|
||||||
|
# find events at least 10 degrees away of san francisco
|
||||||
|
point = [-122.415579, 37.7566023]
|
||||||
|
events = Event.objects(location__near=point, location__min_distance=10)
|
||||||
|
# The following real test passes on MongoDB 3 but minDistance seems
|
||||||
|
# buggy on older MongoDB versions
|
||||||
|
if get_connection().server_info()['versionArray'][0] > 2:
|
||||||
|
self.assertEqual(events.count(), 2)
|
||||||
|
else:
|
||||||
|
self.assertTrue(events.count() >= 2)
|
||||||
|
|
||||||
# find events within 10 degrees of san francisco
|
# find events within 10 degrees of san francisco
|
||||||
point_and_distance = [[-122.415579, 37.7566023], 10]
|
point_and_distance = [[-122.415579, 37.7566023], 10]
|
||||||
events = self.Event.objects(
|
events = Event.objects(location__within_distance=point_and_distance)
|
||||||
location__within_distance=point_and_distance)
|
|
||||||
self.assertEqual(events.count(), 1)
|
self.assertEqual(events.count(), 1)
|
||||||
self.assertEqual(events[0], event2)
|
self.assertEqual(events[0], event2)
|
||||||
|
|
||||||
# find events within 1 degree of greenpoint, broolyn, nyc, ny
|
# find events within 1 degree of greenpoint, broolyn, nyc, ny
|
||||||
point_and_distance = [[-73.9509714, 40.7237134], 1]
|
point_and_distance = [[-73.9509714, 40.7237134], 1]
|
||||||
events = self.Event.objects(
|
events = Event.objects(location__within_distance=point_and_distance)
|
||||||
location__within_distance=point_and_distance)
|
|
||||||
self.assertEqual(events.count(), 0)
|
self.assertEqual(events.count(), 0)
|
||||||
|
|
||||||
# ensure ordering is respected by "within_distance"
|
# ensure ordering is respected by "within_distance"
|
||||||
point_and_distance = [[-87.67892, 41.9120459], 10]
|
point_and_distance = [[-87.67892, 41.9120459], 10]
|
||||||
events = self.Event.objects(
|
events = Event.objects(location__within_distance=point_and_distance)
|
||||||
location__within_distance=point_and_distance)
|
|
||||||
events = events.order_by("-date")
|
events = events.order_by("-date")
|
||||||
self.assertEqual(events.count(), 2)
|
self.assertEqual(events.count(), 2)
|
||||||
self.assertEqual(events[0], event3)
|
self.assertEqual(events[0], event3)
|
||||||
|
|
||||||
def test_within_box(self):
|
|
||||||
"""Ensure the "within_box" operator works."""
|
|
||||||
event1, event2, event3 = self._create_event_data()
|
|
||||||
|
|
||||||
# check that within_box works
|
# check that within_box works
|
||||||
box = [(-125.0, 35.0), (-100.0, 40.0)]
|
box = [(-125.0, 35.0), (-100.0, 40.0)]
|
||||||
events = self.Event.objects(location__within_box=box)
|
events = Event.objects(location__within_box=box)
|
||||||
self.assertEqual(events.count(), 1)
|
self.assertEqual(events.count(), 1)
|
||||||
self.assertEqual(events[0].id, event2.id)
|
self.assertEqual(events[0].id, event2.id)
|
||||||
|
|
||||||
def test_within_polygon(self):
|
|
||||||
"""Ensure the "within_polygon" operator works."""
|
|
||||||
event1, event2, event3 = self._create_event_data()
|
|
||||||
|
|
||||||
polygon = [
|
polygon = [
|
||||||
(-87.694445, 41.912114),
|
(-87.694445, 41.912114),
|
||||||
(-87.69084, 41.919395),
|
(-87.69084, 41.919395),
|
||||||
@@ -141,7 +107,7 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
(-87.654276, 41.911731),
|
(-87.654276, 41.911731),
|
||||||
(-87.656164, 41.898061),
|
(-87.656164, 41.898061),
|
||||||
]
|
]
|
||||||
events = self.Event.objects(location__within_polygon=polygon)
|
events = Event.objects(location__within_polygon=polygon)
|
||||||
self.assertEqual(events.count(), 1)
|
self.assertEqual(events.count(), 1)
|
||||||
self.assertEqual(events[0].id, event1.id)
|
self.assertEqual(events[0].id, event1.id)
|
||||||
|
|
||||||
@@ -150,151 +116,13 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
(-1.225891, 52.792797),
|
(-1.225891, 52.792797),
|
||||||
(-4.40094, 53.389881)
|
(-4.40094, 53.389881)
|
||||||
]
|
]
|
||||||
events = self.Event.objects(location__within_polygon=polygon2)
|
events = Event.objects(location__within_polygon=polygon2)
|
||||||
self.assertEqual(events.count(), 0)
|
self.assertEqual(events.count(), 0)
|
||||||
|
|
||||||
def test_2dsphere_near(self):
|
def test_geo_spatial_embedded(self):
|
||||||
"""Make sure the "near" operator works with a PointField, which
|
|
||||||
corresponds to a 2dsphere index.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data(
|
|
||||||
point_field_class=PointField
|
|
||||||
)
|
|
||||||
|
|
||||||
# find all events "near" pitchfork office, chicago.
|
|
||||||
# note that "near" will show the san francisco event, too,
|
|
||||||
# although it sorts to last.
|
|
||||||
events = self.Event.objects(location__near=[-87.67892, 41.9120459])
|
|
||||||
self.assertEqual(events.count(), 3)
|
|
||||||
self.assertEqual(list(events), [event1, event3, event2])
|
|
||||||
|
|
||||||
# ensure ordering is respected by "near"
|
|
||||||
events = self.Event.objects(location__near=[-87.67892, 41.9120459])
|
|
||||||
events = events.order_by("-date")
|
|
||||||
self.assertEqual(events.count(), 3)
|
|
||||||
self.assertEqual(list(events), [event3, event1, event2])
|
|
||||||
|
|
||||||
def test_2dsphere_near_and_max_distance(self):
|
|
||||||
"""Ensure the "max_distance" operator works alongside the "near"
|
|
||||||
operator with a 2dsphere index.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data(
|
|
||||||
point_field_class=PointField
|
|
||||||
)
|
|
||||||
|
|
||||||
# find events within 10km of san francisco
|
|
||||||
point = [-122.415579, 37.7566023]
|
|
||||||
events = self.Event.objects(location__near=point,
|
|
||||||
location__max_distance=10000)
|
|
||||||
self.assertEqual(events.count(), 1)
|
|
||||||
self.assertEqual(events[0], event2)
|
|
||||||
|
|
||||||
# find events within 1km of greenpoint, broolyn, nyc, ny
|
|
||||||
events = self.Event.objects(location__near=[-73.9509714, 40.7237134],
|
|
||||||
location__max_distance=1000)
|
|
||||||
self.assertEqual(events.count(), 0)
|
|
||||||
|
|
||||||
# ensure ordering is respected by "near"
|
|
||||||
events = self.Event.objects(
|
|
||||||
location__near=[-87.67892, 41.9120459],
|
|
||||||
location__max_distance=10000
|
|
||||||
).order_by("-date")
|
|
||||||
self.assertEqual(events.count(), 2)
|
|
||||||
self.assertEqual(events[0], event3)
|
|
||||||
|
|
||||||
def test_2dsphere_geo_within_box(self):
|
|
||||||
"""Ensure the "geo_within_box" operator works with a 2dsphere
|
|
||||||
index.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data(
|
|
||||||
point_field_class=PointField
|
|
||||||
)
|
|
||||||
|
|
||||||
# check that within_box works
|
|
||||||
box = [(-125.0, 35.0), (-100.0, 40.0)]
|
|
||||||
events = self.Event.objects(location__geo_within_box=box)
|
|
||||||
self.assertEqual(events.count(), 1)
|
|
||||||
self.assertEqual(events[0].id, event2.id)
|
|
||||||
|
|
||||||
def test_2dsphere_geo_within_polygon(self):
|
|
||||||
"""Ensure the "geo_within_polygon" operator works with a
|
|
||||||
2dsphere index.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data(
|
|
||||||
point_field_class=PointField
|
|
||||||
)
|
|
||||||
|
|
||||||
polygon = [
|
|
||||||
(-87.694445, 41.912114),
|
|
||||||
(-87.69084, 41.919395),
|
|
||||||
(-87.681742, 41.927186),
|
|
||||||
(-87.654276, 41.911731),
|
|
||||||
(-87.656164, 41.898061),
|
|
||||||
]
|
|
||||||
events = self.Event.objects(location__geo_within_polygon=polygon)
|
|
||||||
self.assertEqual(events.count(), 1)
|
|
||||||
self.assertEqual(events[0].id, event1.id)
|
|
||||||
|
|
||||||
polygon2 = [
|
|
||||||
(-1.742249, 54.033586),
|
|
||||||
(-1.225891, 52.792797),
|
|
||||||
(-4.40094, 53.389881)
|
|
||||||
]
|
|
||||||
events = self.Event.objects(location__geo_within_polygon=polygon2)
|
|
||||||
self.assertEqual(events.count(), 0)
|
|
||||||
|
|
||||||
# $minDistance was added in MongoDB v2.6, but continued being buggy
|
|
||||||
# until v3.0; skip for older versions
|
|
||||||
@needs_mongodb_v3
|
|
||||||
def test_2dsphere_near_and_min_max_distance(self):
|
|
||||||
"""Ensure "min_distace" and "max_distance" operators work well
|
|
||||||
together with the "near" operator in a 2dsphere index.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data(
|
|
||||||
point_field_class=PointField
|
|
||||||
)
|
|
||||||
|
|
||||||
# ensure min_distance and max_distance combine well
|
|
||||||
events = self.Event.objects(
|
|
||||||
location__near=[-87.67892, 41.9120459],
|
|
||||||
location__min_distance=1000,
|
|
||||||
location__max_distance=10000
|
|
||||||
).order_by("-date")
|
|
||||||
self.assertEqual(events.count(), 1)
|
|
||||||
self.assertEqual(events[0], event3)
|
|
||||||
|
|
||||||
# ensure ordering is respected by "near" with "min_distance"
|
|
||||||
events = self.Event.objects(
|
|
||||||
location__near=[-87.67892, 41.9120459],
|
|
||||||
location__min_distance=10000
|
|
||||||
).order_by("-date")
|
|
||||||
self.assertEqual(events.count(), 1)
|
|
||||||
self.assertEqual(events[0], event2)
|
|
||||||
|
|
||||||
def test_2dsphere_geo_within_center(self):
|
|
||||||
"""Make sure the "geo_within_center" operator works with a
|
|
||||||
2dsphere index.
|
|
||||||
"""
|
|
||||||
event1, event2, event3 = self._create_event_data(
|
|
||||||
point_field_class=PointField
|
|
||||||
)
|
|
||||||
|
|
||||||
# find events within 5 degrees of pitchfork office, chicago
|
|
||||||
point_and_distance = [[-87.67892, 41.9120459], 2]
|
|
||||||
events = self.Event.objects(
|
|
||||||
location__geo_within_center=point_and_distance)
|
|
||||||
self.assertEqual(events.count(), 2)
|
|
||||||
events = list(events)
|
|
||||||
self.assertTrue(event2 not in events)
|
|
||||||
self.assertTrue(event1 in events)
|
|
||||||
self.assertTrue(event3 in events)
|
|
||||||
|
|
||||||
def _test_embedded(self, point_field_class):
|
|
||||||
"""Helper test method ensuring given point field class works
|
|
||||||
well in an embedded document.
|
|
||||||
"""
|
|
||||||
class Venue(EmbeddedDocument):
|
class Venue(EmbeddedDocument):
|
||||||
location = point_field_class()
|
location = GeoPointField()
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
class Event(Document):
|
class Event(Document):
|
||||||
@@ -320,18 +148,16 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
self.assertEqual(events.count(), 3)
|
self.assertEqual(events.count(), 3)
|
||||||
self.assertEqual(list(events), [event1, event3, event2])
|
self.assertEqual(list(events), [event1, event3, event2])
|
||||||
|
|
||||||
def test_geo_spatial_embedded(self):
|
|
||||||
"""Make sure GeoPointField works properly in an embedded document."""
|
|
||||||
self._test_embedded(point_field_class=GeoPointField)
|
|
||||||
|
|
||||||
def test_2dsphere_point_embedded(self):
|
|
||||||
"""Make sure PointField works properly in an embedded document."""
|
|
||||||
self._test_embedded(point_field_class=PointField)
|
|
||||||
|
|
||||||
# Needs MongoDB > 2.6.4 https://jira.mongodb.org/browse/SERVER-14039
|
|
||||||
@needs_mongodb_v3
|
|
||||||
def test_spherical_geospatial_operators(self):
|
def test_spherical_geospatial_operators(self):
|
||||||
"""Ensure that spherical geospatial queries are working."""
|
"""Ensure that spherical geospatial queries are working
|
||||||
|
"""
|
||||||
|
# Needs MongoDB > 2.6.4 https://jira.mongodb.org/browse/SERVER-14039
|
||||||
|
connection = get_connection()
|
||||||
|
info = connection.test.command('buildInfo')
|
||||||
|
mongodb_version = tuple([int(i) for i in info['version'].split('.')])
|
||||||
|
if mongodb_version < (2, 6, 4):
|
||||||
|
raise SkipTest("Need MongoDB version 2.6.4+")
|
||||||
|
|
||||||
class Point(Document):
|
class Point(Document):
|
||||||
location = GeoPointField()
|
location = GeoPointField()
|
||||||
|
|
||||||
@@ -351,10 +177,7 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
|
|
||||||
# Same behavior for _within_spherical_distance
|
# Same behavior for _within_spherical_distance
|
||||||
points = Point.objects(
|
points = Point.objects(
|
||||||
location__within_spherical_distance=[
|
location__within_spherical_distance=[[-122, 37.5], 60 / earth_radius]
|
||||||
[-122, 37.5],
|
|
||||||
60 / earth_radius
|
|
||||||
]
|
|
||||||
)
|
)
|
||||||
self.assertEqual(points.count(), 2)
|
self.assertEqual(points.count(), 2)
|
||||||
|
|
||||||
@@ -371,9 +194,14 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
# Test query works with min_distance, being farer from one point
|
# Test query works with min_distance, being farer from one point
|
||||||
points = Point.objects(location__near_sphere=[-122, 37.8],
|
points = Point.objects(location__near_sphere=[-122, 37.8],
|
||||||
location__min_distance=60 / earth_radius)
|
location__min_distance=60 / earth_radius)
|
||||||
self.assertEqual(points.count(), 1)
|
# The following real test passes on MongoDB 3 but minDistance seems
|
||||||
far_point = points.first()
|
# buggy on older MongoDB versions
|
||||||
self.assertNotEqual(close_point, far_point)
|
if get_connection().server_info()['versionArray'][0] > 2:
|
||||||
|
self.assertEqual(points.count(), 1)
|
||||||
|
far_point = points.first()
|
||||||
|
self.assertNotEqual(close_point, far_point)
|
||||||
|
else:
|
||||||
|
self.assertTrue(points.count() >= 1)
|
||||||
|
|
||||||
# Finds both points, but orders the north point first because it's
|
# Finds both points, but orders the north point first because it's
|
||||||
# closer to the reference point to the north.
|
# closer to the reference point to the north.
|
||||||
@@ -392,15 +220,141 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
# Finds only one point because only the first point is within 60km of
|
# Finds only one point because only the first point is within 60km of
|
||||||
# the reference point to the south.
|
# the reference point to the south.
|
||||||
points = Point.objects(
|
points = Point.objects(
|
||||||
location__within_spherical_distance=[
|
location__within_spherical_distance=[[-122, 36.5], 60/earth_radius])
|
||||||
[-122, 36.5],
|
|
||||||
60 / earth_radius
|
|
||||||
]
|
|
||||||
)
|
|
||||||
self.assertEqual(points.count(), 1)
|
self.assertEqual(points.count(), 1)
|
||||||
self.assertEqual(points[0].id, south_point.id)
|
self.assertEqual(points[0].id, south_point.id)
|
||||||
|
|
||||||
|
def test_2dsphere_point(self):
|
||||||
|
|
||||||
|
class Event(Document):
|
||||||
|
title = StringField()
|
||||||
|
date = DateTimeField()
|
||||||
|
location = PointField()
|
||||||
|
|
||||||
|
def __unicode__(self):
|
||||||
|
return self.title
|
||||||
|
|
||||||
|
Event.drop_collection()
|
||||||
|
|
||||||
|
event1 = Event(title="Coltrane Motion @ Double Door",
|
||||||
|
date=datetime.now() - timedelta(days=1),
|
||||||
|
location=[-87.677137, 41.909889])
|
||||||
|
event1.save()
|
||||||
|
event2 = Event(title="Coltrane Motion @ Bottom of the Hill",
|
||||||
|
date=datetime.now() - timedelta(days=10),
|
||||||
|
location=[-122.4194155, 37.7749295]).save()
|
||||||
|
event3 = Event(title="Coltrane Motion @ Empty Bottle",
|
||||||
|
date=datetime.now(),
|
||||||
|
location=[-87.686638, 41.900474]).save()
|
||||||
|
|
||||||
|
# find all events "near" pitchfork office, chicago.
|
||||||
|
# note that "near" will show the san francisco event, too,
|
||||||
|
# although it sorts to last.
|
||||||
|
events = Event.objects(location__near=[-87.67892, 41.9120459])
|
||||||
|
self.assertEqual(events.count(), 3)
|
||||||
|
self.assertEqual(list(events), [event1, event3, event2])
|
||||||
|
|
||||||
|
# find events within 5 degrees of pitchfork office, chicago
|
||||||
|
point_and_distance = [[-87.67892, 41.9120459], 2]
|
||||||
|
events = Event.objects(location__geo_within_center=point_and_distance)
|
||||||
|
self.assertEqual(events.count(), 2)
|
||||||
|
events = list(events)
|
||||||
|
self.assertTrue(event2 not in events)
|
||||||
|
self.assertTrue(event1 in events)
|
||||||
|
self.assertTrue(event3 in events)
|
||||||
|
|
||||||
|
# ensure ordering is respected by "near"
|
||||||
|
events = Event.objects(location__near=[-87.67892, 41.9120459])
|
||||||
|
events = events.order_by("-date")
|
||||||
|
self.assertEqual(events.count(), 3)
|
||||||
|
self.assertEqual(list(events), [event3, event1, event2])
|
||||||
|
|
||||||
|
# find events within 10km of san francisco
|
||||||
|
point = [-122.415579, 37.7566023]
|
||||||
|
events = Event.objects(location__near=point, location__max_distance=10000)
|
||||||
|
self.assertEqual(events.count(), 1)
|
||||||
|
self.assertEqual(events[0], event2)
|
||||||
|
|
||||||
|
# find events within 1km of greenpoint, broolyn, nyc, ny
|
||||||
|
events = Event.objects(location__near=[-73.9509714, 40.7237134], location__max_distance=1000)
|
||||||
|
self.assertEqual(events.count(), 0)
|
||||||
|
|
||||||
|
# ensure ordering is respected by "near"
|
||||||
|
events = Event.objects(location__near=[-87.67892, 41.9120459],
|
||||||
|
location__max_distance=10000).order_by("-date")
|
||||||
|
self.assertEqual(events.count(), 2)
|
||||||
|
self.assertEqual(events[0], event3)
|
||||||
|
|
||||||
|
# ensure min_distance and max_distance combine well
|
||||||
|
events = Event.objects(location__near=[-87.67892, 41.9120459],
|
||||||
|
location__min_distance=1000,
|
||||||
|
location__max_distance=10000).order_by("-date")
|
||||||
|
self.assertEqual(events.count(), 1)
|
||||||
|
self.assertEqual(events[0], event3)
|
||||||
|
|
||||||
|
# ensure ordering is respected by "near"
|
||||||
|
events = Event.objects(location__near=[-87.67892, 41.9120459],
|
||||||
|
# location__min_distance=10000
|
||||||
|
location__min_distance=10000).order_by("-date")
|
||||||
|
self.assertEqual(events.count(), 1)
|
||||||
|
self.assertEqual(events[0], event2)
|
||||||
|
|
||||||
|
# check that within_box works
|
||||||
|
box = [(-125.0, 35.0), (-100.0, 40.0)]
|
||||||
|
events = Event.objects(location__geo_within_box=box)
|
||||||
|
self.assertEqual(events.count(), 1)
|
||||||
|
self.assertEqual(events[0].id, event2.id)
|
||||||
|
|
||||||
|
polygon = [
|
||||||
|
(-87.694445, 41.912114),
|
||||||
|
(-87.69084, 41.919395),
|
||||||
|
(-87.681742, 41.927186),
|
||||||
|
(-87.654276, 41.911731),
|
||||||
|
(-87.656164, 41.898061),
|
||||||
|
]
|
||||||
|
events = Event.objects(location__geo_within_polygon=polygon)
|
||||||
|
self.assertEqual(events.count(), 1)
|
||||||
|
self.assertEqual(events[0].id, event1.id)
|
||||||
|
|
||||||
|
polygon2 = [
|
||||||
|
(-1.742249, 54.033586),
|
||||||
|
(-1.225891, 52.792797),
|
||||||
|
(-4.40094, 53.389881)
|
||||||
|
]
|
||||||
|
events = Event.objects(location__geo_within_polygon=polygon2)
|
||||||
|
self.assertEqual(events.count(), 0)
|
||||||
|
|
||||||
|
def test_2dsphere_point_embedded(self):
|
||||||
|
|
||||||
|
class Venue(EmbeddedDocument):
|
||||||
|
location = GeoPointField()
|
||||||
|
name = StringField()
|
||||||
|
|
||||||
|
class Event(Document):
|
||||||
|
title = StringField()
|
||||||
|
venue = EmbeddedDocumentField(Venue)
|
||||||
|
|
||||||
|
Event.drop_collection()
|
||||||
|
|
||||||
|
venue1 = Venue(name="The Rock", location=[-87.677137, 41.909889])
|
||||||
|
venue2 = Venue(name="The Bridge", location=[-122.4194155, 37.7749295])
|
||||||
|
|
||||||
|
event1 = Event(title="Coltrane Motion @ Double Door",
|
||||||
|
venue=venue1).save()
|
||||||
|
event2 = Event(title="Coltrane Motion @ Bottom of the Hill",
|
||||||
|
venue=venue2).save()
|
||||||
|
event3 = Event(title="Coltrane Motion @ Empty Bottle",
|
||||||
|
venue=venue1).save()
|
||||||
|
|
||||||
|
# find all events "near" pitchfork office, chicago.
|
||||||
|
# note that "near" will show the san francisco event, too,
|
||||||
|
# although it sorts to last.
|
||||||
|
events = Event.objects(venue__location__near=[-87.67892, 41.9120459])
|
||||||
|
self.assertEqual(events.count(), 3)
|
||||||
|
self.assertEqual(list(events), [event1, event3, event2])
|
||||||
|
|
||||||
def test_linestring(self):
|
def test_linestring(self):
|
||||||
|
|
||||||
class Road(Document):
|
class Road(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
line = LineStringField()
|
line = LineStringField()
|
||||||
@@ -456,6 +410,7 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
self.assertEqual(1, roads)
|
self.assertEqual(1, roads)
|
||||||
|
|
||||||
def test_polygon(self):
|
def test_polygon(self):
|
||||||
|
|
||||||
class Road(Document):
|
class Road(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
poly = PolygonField()
|
poly = PolygonField()
|
||||||
@@ -552,6 +507,5 @@ class GeoQueriesTest(MongoDBTestCase):
|
|||||||
loc = Location.objects.as_pymongo()[0]
|
loc = Location.objects.as_pymongo()[0]
|
||||||
self.assertEqual(loc["poly"], {"type": "Polygon", "coordinates": [[[40, 4], [40, 6], [41, 6], [40, 4]]]})
|
self.assertEqual(loc["poly"], {"type": "Polygon", "coordinates": [[[40, 4], [40, 6], [41, 6], [40, 4]]]})
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -19,9 +19,6 @@ from mongoengine.python_support import IS_PYMONGO_3
|
|||||||
from mongoengine.queryset import (DoesNotExist, MultipleObjectsReturned,
|
from mongoengine.queryset import (DoesNotExist, MultipleObjectsReturned,
|
||||||
QuerySet, QuerySetManager, queryset_manager)
|
QuerySet, QuerySetManager, queryset_manager)
|
||||||
|
|
||||||
from tests.utils import needs_mongodb_v26, skip_pymongo3
|
|
||||||
|
|
||||||
|
|
||||||
__all__ = ("QuerySetTest",)
|
__all__ = ("QuerySetTest",)
|
||||||
|
|
||||||
|
|
||||||
@@ -35,6 +32,37 @@ class db_ops_tracker(query_counter):
|
|||||||
return list(self.db.system.profile.find(ignore_query))
|
return list(self.db.system.profile.find(ignore_query))
|
||||||
|
|
||||||
|
|
||||||
|
def skip_older_mongodb(f):
|
||||||
|
def _inner(*args, **kwargs):
|
||||||
|
connection = get_connection()
|
||||||
|
info = connection.test.command('buildInfo')
|
||||||
|
mongodb_version = tuple([int(i) for i in info['version'].split('.')])
|
||||||
|
|
||||||
|
if mongodb_version < (2, 6):
|
||||||
|
raise SkipTest("Need MongoDB version 2.6+")
|
||||||
|
|
||||||
|
return f(*args, **kwargs)
|
||||||
|
|
||||||
|
_inner.__name__ = f.__name__
|
||||||
|
_inner.__doc__ = f.__doc__
|
||||||
|
|
||||||
|
return _inner
|
||||||
|
|
||||||
|
|
||||||
|
def skip_pymongo3(f):
|
||||||
|
def _inner(*args, **kwargs):
|
||||||
|
|
||||||
|
if IS_PYMONGO_3:
|
||||||
|
raise SkipTest("Useless with PyMongo 3+")
|
||||||
|
|
||||||
|
return f(*args, **kwargs)
|
||||||
|
|
||||||
|
_inner.__name__ = f.__name__
|
||||||
|
_inner.__doc__ = f.__doc__
|
||||||
|
|
||||||
|
return _inner
|
||||||
|
|
||||||
|
|
||||||
class QuerySetTest(unittest.TestCase):
|
class QuerySetTest(unittest.TestCase):
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
@@ -571,23 +599,16 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
self.assertEqual(post.comments[0].by, 'joe')
|
self.assertEqual(post.comments[0].by, 'joe')
|
||||||
self.assertEqual(post.comments[0].votes.score, 4)
|
self.assertEqual(post.comments[0].votes.score, 4)
|
||||||
|
|
||||||
@needs_mongodb_v26
|
|
||||||
def test_update_min_max(self):
|
def test_update_min_max(self):
|
||||||
class Scores(Document):
|
class Scores(Document):
|
||||||
high_score = IntField()
|
high_score = IntField()
|
||||||
low_score = IntField()
|
low_score = IntField()
|
||||||
|
scores = Scores(high_score=800, low_score=200)
|
||||||
scores = Scores.objects.create(high_score=800, low_score=200)
|
scores.save()
|
||||||
|
|
||||||
Scores.objects(id=scores.id).update(min__low_score=150)
|
Scores.objects(id=scores.id).update(min__low_score=150)
|
||||||
self.assertEqual(Scores.objects.get(id=scores.id).low_score, 150)
|
self.assertEqual(Scores.objects(id=scores.id).get().low_score, 150)
|
||||||
Scores.objects(id=scores.id).update(min__low_score=250)
|
Scores.objects(id=scores.id).update(min__low_score=250)
|
||||||
self.assertEqual(Scores.objects.get(id=scores.id).low_score, 150)
|
self.assertEqual(Scores.objects(id=scores.id).get().low_score, 150)
|
||||||
|
|
||||||
Scores.objects(id=scores.id).update(max__high_score=1000)
|
|
||||||
self.assertEqual(Scores.objects.get(id=scores.id).high_score, 1000)
|
|
||||||
Scores.objects(id=scores.id).update(max__high_score=500)
|
|
||||||
self.assertEqual(Scores.objects.get(id=scores.id).high_score, 1000)
|
|
||||||
|
|
||||||
def test_updates_can_have_match_operators(self):
|
def test_updates_can_have_match_operators(self):
|
||||||
|
|
||||||
@@ -991,7 +1012,7 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
self.assertEqual(person.name, "User A")
|
self.assertEqual(person.name, "User A")
|
||||||
self.assertEqual(person.age, 20)
|
self.assertEqual(person.age, 20)
|
||||||
|
|
||||||
@needs_mongodb_v26
|
@skip_older_mongodb
|
||||||
@skip_pymongo3
|
@skip_pymongo3
|
||||||
def test_cursor_args(self):
|
def test_cursor_args(self):
|
||||||
"""Ensures the cursor args can be set as expected
|
"""Ensures the cursor args can be set as expected
|
||||||
@@ -3108,7 +3129,7 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
|
|
||||||
self.assertEqual(Foo.objects.distinct("bar"), [bar])
|
self.assertEqual(Foo.objects.distinct("bar"), [bar])
|
||||||
|
|
||||||
@needs_mongodb_v26
|
@skip_older_mongodb
|
||||||
def test_text_indexes(self):
|
def test_text_indexes(self):
|
||||||
class News(Document):
|
class News(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -3195,7 +3216,7 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
'brasil').order_by('$text_score').first()
|
'brasil').order_by('$text_score').first()
|
||||||
self.assertEqual(item.get_text_score(), max_text_score)
|
self.assertEqual(item.get_text_score(), max_text_score)
|
||||||
|
|
||||||
@needs_mongodb_v26
|
@skip_older_mongodb
|
||||||
def test_distinct_handles_references_to_alias(self):
|
def test_distinct_handles_references_to_alias(self):
|
||||||
register_connection('testdb', 'mongoenginetest2')
|
register_connection('testdb', 'mongoenginetest2')
|
||||||
|
|
||||||
@@ -4870,7 +4891,6 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
self.assertTrue(Person.objects._has_data(),
|
self.assertTrue(Person.objects._has_data(),
|
||||||
'Cursor has data and returned False')
|
'Cursor has data and returned False')
|
||||||
|
|
||||||
@needs_mongodb_v26
|
|
||||||
def test_queryset_aggregation_framework(self):
|
def test_queryset_aggregation_framework(self):
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -4905,13 +4925,17 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
{'_id': p1.pk, 'name': "ISABELLA LUANNA"}
|
{'_id': p1.pk, 'name': "ISABELLA LUANNA"}
|
||||||
])
|
])
|
||||||
|
|
||||||
data = Person.objects(age__gte=17, age__lte=40).order_by('-age').aggregate({
|
data = Person.objects(
|
||||||
'$group': {
|
age__gte=17, age__lte=40).order_by('-age').aggregate(
|
||||||
'_id': None,
|
{'$group': {
|
||||||
'total': {'$sum': 1},
|
'_id': None,
|
||||||
'avg': {'$avg': '$age'}
|
'total': {'$sum': 1},
|
||||||
}
|
'avg': {'$avg': '$age'}
|
||||||
})
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
)
|
||||||
|
|
||||||
self.assertEqual(list(data), [
|
self.assertEqual(list(data), [
|
||||||
{'_id': None, 'avg': 29, 'total': 2}
|
{'_id': None, 'avg': 29, 'total': 2}
|
||||||
])
|
])
|
||||||
@@ -4952,16 +4976,28 @@ class QuerySetTest(unittest.TestCase):
|
|||||||
self.assertEquals(Animal.objects(folded_ears=True).count(), 1)
|
self.assertEquals(Animal.objects(folded_ears=True).count(), 1)
|
||||||
self.assertEquals(Animal.objects(whiskers_length=5.1).count(), 1)
|
self.assertEquals(Animal.objects(whiskers_length=5.1).count(), 1)
|
||||||
|
|
||||||
def test_loop_over_invalid_id_does_not_crash(self):
|
def test_loop_via_invalid_id_does_not_crash(self):
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
Person.objects.delete()
|
||||||
Person.drop_collection()
|
Person._get_collection().update({"name": "a"}, {"$set": {"_id": ""}}, upsert=True)
|
||||||
|
|
||||||
Person._get_collection().insert({'name': 'a', 'id': ''})
|
|
||||||
for p in Person.objects():
|
for p in Person.objects():
|
||||||
self.assertEqual(p.name, 'a')
|
self.assertEqual(p.name, 'a')
|
||||||
|
|
||||||
|
def test_last_field_name_like_operator(self):
|
||||||
|
class EmbeddedItem(EmbeddedDocument):
|
||||||
|
type = StringField()
|
||||||
|
|
||||||
|
class Doc(Document):
|
||||||
|
item = EmbeddedDocumentField(EmbeddedItem)
|
||||||
|
|
||||||
|
Doc.drop_collection()
|
||||||
|
|
||||||
|
doc = Doc(item=EmbeddedItem(type="axe"))
|
||||||
|
doc.save()
|
||||||
|
|
||||||
|
self.assertEqual(1, Doc.objects(item__type__="axe").count())
|
||||||
|
|
||||||
def test_len_during_iteration(self):
|
def test_len_during_iteration(self):
|
||||||
"""Tests that calling len on a queyset during iteration doesn't
|
"""Tests that calling len on a queyset during iteration doesn't
|
||||||
stop paging.
|
stop paging.
|
||||||
|
@@ -35,7 +35,8 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
mongoengine.connection._dbs = {}
|
mongoengine.connection._dbs = {}
|
||||||
|
|
||||||
def test_connect(self):
|
def test_connect(self):
|
||||||
"""Ensure that the connect() method works properly."""
|
"""Ensure that the connect() method works properly.
|
||||||
|
"""
|
||||||
connect('mongoenginetest')
|
connect('mongoenginetest')
|
||||||
|
|
||||||
conn = get_connection()
|
conn = get_connection()
|
||||||
@@ -145,7 +146,8 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
self.assertEqual(expected_connection, actual_connection)
|
self.assertEqual(expected_connection, actual_connection)
|
||||||
|
|
||||||
def test_connect_uri(self):
|
def test_connect_uri(self):
|
||||||
"""Ensure that the connect() method works properly with URIs."""
|
"""Ensure that the connect() method works properly with uri's
|
||||||
|
"""
|
||||||
c = connect(db='mongoenginetest', alias='admin')
|
c = connect(db='mongoenginetest', alias='admin')
|
||||||
c.admin.system.users.remove({})
|
c.admin.system.users.remove({})
|
||||||
c.mongoenginetest.system.users.remove({})
|
c.mongoenginetest.system.users.remove({})
|
||||||
@@ -198,6 +200,19 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
self.assertTrue(isinstance(db, pymongo.database.Database))
|
self.assertTrue(isinstance(db, pymongo.database.Database))
|
||||||
self.assertEqual(db.name, 'test')
|
self.assertEqual(db.name, 'test')
|
||||||
|
|
||||||
|
def test_connect_uri_with_replicaset(self):
|
||||||
|
"""Ensure connect() works when specifying a replicaSet."""
|
||||||
|
if IS_PYMONGO_3:
|
||||||
|
c = connect(host='mongodb://localhost/test?replicaSet=local-rs')
|
||||||
|
db = get_db()
|
||||||
|
self.assertTrue(isinstance(db, pymongo.database.Database))
|
||||||
|
self.assertEqual(db.name, 'test')
|
||||||
|
else:
|
||||||
|
# PyMongo < v3.x raises an exception:
|
||||||
|
# "localhost:27017 is not a member of replica set local-rs"
|
||||||
|
with self.assertRaises(MongoEngineConnectionError):
|
||||||
|
c = connect(host='mongodb://localhost/test?replicaSet=local-rs')
|
||||||
|
|
||||||
def test_uri_without_credentials_doesnt_override_conn_settings(self):
|
def test_uri_without_credentials_doesnt_override_conn_settings(self):
|
||||||
"""Ensure connect() uses the username & password params if the URI
|
"""Ensure connect() uses the username & password params if the URI
|
||||||
doesn't explicitly specify them.
|
doesn't explicitly specify them.
|
||||||
@@ -212,8 +227,9 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
self.assertRaises(OperationFailure, get_db)
|
self.assertRaises(OperationFailure, get_db)
|
||||||
|
|
||||||
def test_connect_uri_with_authsource(self):
|
def test_connect_uri_with_authsource(self):
|
||||||
"""Ensure that the connect() method works well with `authSource`
|
"""Ensure that the connect() method works well with
|
||||||
option in the URI.
|
the option `authSource` in URI.
|
||||||
|
This feature was introduced in MongoDB 2.4 and removed in 2.6
|
||||||
"""
|
"""
|
||||||
# Create users
|
# Create users
|
||||||
c = connect('mongoenginetest')
|
c = connect('mongoenginetest')
|
||||||
@@ -222,31 +238,30 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
|
|
||||||
# Authentication fails without "authSource"
|
# Authentication fails without "authSource"
|
||||||
if IS_PYMONGO_3:
|
if IS_PYMONGO_3:
|
||||||
test_conn = connect(
|
test_conn = connect('mongoenginetest', alias='test1',
|
||||||
'mongoenginetest', alias='test1',
|
host='mongodb://username2:password@localhost/mongoenginetest')
|
||||||
host='mongodb://username2:password@localhost/mongoenginetest'
|
|
||||||
)
|
|
||||||
self.assertRaises(OperationFailure, test_conn.server_info)
|
self.assertRaises(OperationFailure, test_conn.server_info)
|
||||||
else:
|
else:
|
||||||
self.assertRaises(
|
self.assertRaises(
|
||||||
MongoEngineConnectionError,
|
MongoEngineConnectionError, connect, 'mongoenginetest',
|
||||||
connect, 'mongoenginetest', alias='test1',
|
alias='test1',
|
||||||
host='mongodb://username2:password@localhost/mongoenginetest'
|
host='mongodb://username2:password@localhost/mongoenginetest'
|
||||||
)
|
)
|
||||||
self.assertRaises(MongoEngineConnectionError, get_db, 'test1')
|
self.assertRaises(MongoEngineConnectionError, get_db, 'test1')
|
||||||
|
|
||||||
# Authentication succeeds with "authSource"
|
# Authentication succeeds with "authSource"
|
||||||
authd_conn = connect(
|
connect(
|
||||||
'mongoenginetest', alias='test2',
|
'mongoenginetest', alias='test2',
|
||||||
host=('mongodb://username2:password@localhost/'
|
host=('mongodb://username2:password@localhost/'
|
||||||
'mongoenginetest?authSource=admin')
|
'mongoenginetest?authSource=admin')
|
||||||
)
|
)
|
||||||
|
# This will fail starting from MongoDB 2.6+
|
||||||
db = get_db('test2')
|
db = get_db('test2')
|
||||||
self.assertTrue(isinstance(db, pymongo.database.Database))
|
self.assertTrue(isinstance(db, pymongo.database.Database))
|
||||||
self.assertEqual(db.name, 'mongoenginetest')
|
self.assertEqual(db.name, 'mongoenginetest')
|
||||||
|
|
||||||
# Clear all users
|
# Clear all users
|
||||||
authd_conn.admin.system.users.remove({})
|
c.admin.system.users.remove({})
|
||||||
|
|
||||||
def test_register_connection(self):
|
def test_register_connection(self):
|
||||||
"""Ensure that connections with different aliases may be registered.
|
"""Ensure that connections with different aliases may be registered.
|
||||||
@@ -319,38 +334,6 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
self.assertEqual(dict(conn1.write_concern), {'w': 1, 'j': True})
|
self.assertEqual(dict(conn1.write_concern), {'w': 1, 'j': True})
|
||||||
self.assertEqual(dict(conn2.write_concern), {'w': 1, 'j': True})
|
self.assertEqual(dict(conn2.write_concern), {'w': 1, 'j': True})
|
||||||
|
|
||||||
def test_connect_with_replicaset_via_uri(self):
|
|
||||||
"""Ensure connect() works when specifying a replicaSet via the
|
|
||||||
MongoDB URI.
|
|
||||||
"""
|
|
||||||
if IS_PYMONGO_3:
|
|
||||||
c = connect(host='mongodb://localhost/test?replicaSet=local-rs')
|
|
||||||
db = get_db()
|
|
||||||
self.assertTrue(isinstance(db, pymongo.database.Database))
|
|
||||||
self.assertEqual(db.name, 'test')
|
|
||||||
else:
|
|
||||||
# PyMongo < v3.x raises an exception:
|
|
||||||
# "localhost:27017 is not a member of replica set local-rs"
|
|
||||||
with self.assertRaises(MongoEngineConnectionError):
|
|
||||||
c = connect(host='mongodb://localhost/test?replicaSet=local-rs')
|
|
||||||
|
|
||||||
def test_connect_with_replicaset_via_kwargs(self):
|
|
||||||
"""Ensure connect() works when specifying a replicaSet via the
|
|
||||||
connection kwargs
|
|
||||||
"""
|
|
||||||
if IS_PYMONGO_3:
|
|
||||||
c = connect(replicaset='local-rs')
|
|
||||||
self.assertEqual(c._MongoClient__options.replica_set_name,
|
|
||||||
'local-rs')
|
|
||||||
db = get_db()
|
|
||||||
self.assertTrue(isinstance(db, pymongo.database.Database))
|
|
||||||
self.assertEqual(db.name, 'test')
|
|
||||||
else:
|
|
||||||
# PyMongo < v3.x raises an exception:
|
|
||||||
# "localhost:27017 is not a member of replica set local-rs"
|
|
||||||
with self.assertRaises(MongoEngineConnectionError):
|
|
||||||
c = connect(replicaset='local-rs')
|
|
||||||
|
|
||||||
def test_datetime(self):
|
def test_datetime(self):
|
||||||
connect('mongoenginetest', tz_aware=True)
|
connect('mongoenginetest', tz_aware=True)
|
||||||
d = datetime.datetime(2010, 5, 5, tzinfo=utc)
|
d = datetime.datetime(2010, 5, 5, tzinfo=utc)
|
||||||
|
@@ -2,15 +2,10 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
from collections import OrderedDict
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.context_managers import query_counter
|
from mongoengine.context_managers import query_counter
|
||||||
from mongoengine.python_support import IS_PYMONGO_3
|
|
||||||
from mongoengine.base import TopLevelDocumentMetaclass
|
|
||||||
if IS_PYMONGO_3:
|
|
||||||
from bson import CodecOptions
|
|
||||||
|
|
||||||
|
|
||||||
class FieldTest(unittest.TestCase):
|
class FieldTest(unittest.TestCase):
|
||||||
@@ -1292,70 +1287,5 @@ class FieldTest(unittest.TestCase):
|
|||||||
|
|
||||||
self.assertEqual(q, 2)
|
self.assertEqual(q, 2)
|
||||||
|
|
||||||
def test_dynamic_field_dereference(self):
|
|
||||||
class Merchandise(Document):
|
|
||||||
name = StringField()
|
|
||||||
price = IntField()
|
|
||||||
|
|
||||||
class Store(Document):
|
|
||||||
merchandises = DynamicField()
|
|
||||||
|
|
||||||
Merchandise.drop_collection()
|
|
||||||
Store.drop_collection()
|
|
||||||
|
|
||||||
merchandises = {
|
|
||||||
'#1': Merchandise(name='foo', price=100).save(),
|
|
||||||
'#2': Merchandise(name='bar', price=120).save(),
|
|
||||||
'#3': Merchandise(name='baz', price=110).save(),
|
|
||||||
}
|
|
||||||
Store(merchandises=merchandises).save()
|
|
||||||
|
|
||||||
store = Store.objects().first()
|
|
||||||
for obj in store.merchandises.values():
|
|
||||||
self.assertFalse(isinstance(obj, Merchandise))
|
|
||||||
|
|
||||||
store.select_related()
|
|
||||||
for obj in store.merchandises.values():
|
|
||||||
self.assertTrue(isinstance(obj, Merchandise))
|
|
||||||
|
|
||||||
def test_dynamic_field_dereference_with_ordering_guarantee_on_pymongo3(self):
|
|
||||||
# This is because 'codec_options' is supported on pymongo3 or later
|
|
||||||
if IS_PYMONGO_3:
|
|
||||||
class OrderedDocument(Document):
|
|
||||||
my_metaclass = TopLevelDocumentMetaclass
|
|
||||||
__metaclass__ = TopLevelDocumentMetaclass
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def _get_collection(cls):
|
|
||||||
collection = super(OrderedDocument, cls)._get_collection()
|
|
||||||
opts = CodecOptions(document_class=OrderedDict)
|
|
||||||
|
|
||||||
return collection.with_options(codec_options=opts)
|
|
||||||
|
|
||||||
class Merchandise(Document):
|
|
||||||
name = StringField()
|
|
||||||
price = IntField()
|
|
||||||
|
|
||||||
class Store(OrderedDocument):
|
|
||||||
merchandises = DynamicField(container_class=OrderedDict)
|
|
||||||
|
|
||||||
Merchandise.drop_collection()
|
|
||||||
Store.drop_collection()
|
|
||||||
|
|
||||||
merchandises = OrderedDict()
|
|
||||||
merchandises['#1'] = Merchandise(name='foo', price=100).save()
|
|
||||||
merchandises['#2'] = Merchandise(name='bar', price=120).save()
|
|
||||||
merchandises['#3'] = Merchandise(name='baz', price=110).save()
|
|
||||||
|
|
||||||
Store(merchandises=merchandises).save()
|
|
||||||
|
|
||||||
store = Store.objects().first()
|
|
||||||
|
|
||||||
store.select_related()
|
|
||||||
|
|
||||||
# confirms that the load data order is same with the one at storing
|
|
||||||
self.assertTrue(type(store.merchandises), OrderedDict)
|
|
||||||
self.assertEqual(','.join(store.merchandises.keys()), '#1,#2,#3')
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -1,11 +1,7 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from nose.plugins.skip import SkipTest
|
|
||||||
|
|
||||||
from mongoengine import connect
|
from mongoengine import connect
|
||||||
from mongoengine.connection import get_db, get_connection
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.python_support import IS_PYMONGO_3
|
|
||||||
|
|
||||||
|
|
||||||
MONGO_TEST_DB = 'mongoenginetest'
|
MONGO_TEST_DB = 'mongoenginetest'
|
||||||
|
|
||||||
@@ -24,55 +20,3 @@ class MongoDBTestCase(unittest.TestCase):
|
|||||||
@classmethod
|
@classmethod
|
||||||
def tearDownClass(cls):
|
def tearDownClass(cls):
|
||||||
cls._connection.drop_database(MONGO_TEST_DB)
|
cls._connection.drop_database(MONGO_TEST_DB)
|
||||||
|
|
||||||
|
|
||||||
def get_mongodb_version():
|
|
||||||
"""Return the version tuple of the MongoDB server that the default
|
|
||||||
connection is connected to.
|
|
||||||
"""
|
|
||||||
return tuple(get_connection().server_info()['versionArray'])
|
|
||||||
|
|
||||||
def _decorated_with_ver_requirement(func, ver_tuple):
|
|
||||||
"""Return a given function decorated with the version requirement
|
|
||||||
for a particular MongoDB version tuple.
|
|
||||||
"""
|
|
||||||
def _inner(*args, **kwargs):
|
|
||||||
mongodb_ver = get_mongodb_version()
|
|
||||||
if mongodb_ver >= ver_tuple:
|
|
||||||
return func(*args, **kwargs)
|
|
||||||
|
|
||||||
raise SkipTest('Needs MongoDB v{}+'.format(
|
|
||||||
'.'.join([str(v) for v in ver_tuple])
|
|
||||||
))
|
|
||||||
|
|
||||||
_inner.__name__ = func.__name__
|
|
||||||
_inner.__doc__ = func.__doc__
|
|
||||||
|
|
||||||
return _inner
|
|
||||||
|
|
||||||
def needs_mongodb_v26(func):
|
|
||||||
"""Raise a SkipTest exception if we're working with MongoDB version
|
|
||||||
lower than v2.6.
|
|
||||||
"""
|
|
||||||
return _decorated_with_ver_requirement(func, (2, 6))
|
|
||||||
|
|
||||||
def needs_mongodb_v3(func):
|
|
||||||
"""Raise a SkipTest exception if we're working with MongoDB version
|
|
||||||
lower than v3.0.
|
|
||||||
"""
|
|
||||||
return _decorated_with_ver_requirement(func, (3, 0))
|
|
||||||
|
|
||||||
def skip_pymongo3(f):
|
|
||||||
"""Raise a SkipTest exception if we're running a test against
|
|
||||||
PyMongo v3.x.
|
|
||||||
"""
|
|
||||||
def _inner(*args, **kwargs):
|
|
||||||
if IS_PYMONGO_3:
|
|
||||||
raise SkipTest("Useless with PyMongo 3+")
|
|
||||||
return f(*args, **kwargs)
|
|
||||||
|
|
||||||
_inner.__name__ = f.__name__
|
|
||||||
_inner.__doc__ = f.__doc__
|
|
||||||
|
|
||||||
return _inner
|
|
||||||
|
|
||||||
|
13
tox.ini
13
tox.ini
@@ -1,5 +1,5 @@
|
|||||||
[tox]
|
[tox]
|
||||||
envlist = {py27,py35,pypy,pypy3}-{mg27,mg28,mg30}
|
envlist = {py26,py27,py33,py34,py35,pypy,pypy3}-{mg27,mg28},flake8
|
||||||
|
|
||||||
[testenv]
|
[testenv]
|
||||||
commands =
|
commands =
|
||||||
@@ -7,7 +7,16 @@ commands =
|
|||||||
deps =
|
deps =
|
||||||
nose
|
nose
|
||||||
mg27: PyMongo<2.8
|
mg27: PyMongo<2.8
|
||||||
mg28: PyMongo>=2.8,<2.9
|
mg28: PyMongo>=2.8,<3.0
|
||||||
mg30: PyMongo>=3.0
|
mg30: PyMongo>=3.0
|
||||||
|
mgdev: https://github.com/mongodb/mongo-python-driver/tarball/master
|
||||||
setenv =
|
setenv =
|
||||||
PYTHON_EGG_CACHE = {envdir}/python-eggs
|
PYTHON_EGG_CACHE = {envdir}/python-eggs
|
||||||
|
passenv = windir
|
||||||
|
|
||||||
|
[testenv:flake8]
|
||||||
|
deps =
|
||||||
|
flake8
|
||||||
|
flake8-import-order
|
||||||
|
commands =
|
||||||
|
flake8
|
||||||
|
Reference in New Issue
Block a user