Compare commits
1 Commits
v0.23.1
...
bagerard-p
Author | SHA1 | Date | |
---|---|---|---|
|
d73ca6f90d |
143
.github/workflows/github-actions.yml
vendored
143
.github/workflows/github-actions.yml
vendored
@@ -1,143 +0,0 @@
|
|||||||
name: MongoengineCI
|
|
||||||
on:
|
|
||||||
# All PR
|
|
||||||
pull_request:
|
|
||||||
# master branch merge
|
|
||||||
push:
|
|
||||||
branches:
|
|
||||||
- master
|
|
||||||
# release tags
|
|
||||||
create:
|
|
||||||
tags:
|
|
||||||
- 'v[0-9]+\.[0-9]+\.[0-9]+*'
|
|
||||||
env:
|
|
||||||
MONGODB_3_6: 3.6.14
|
|
||||||
MONGODB_4_0: 4.0.23
|
|
||||||
MONGODB_4_2: 4.2
|
|
||||||
MONGODB_4_4: 4.4
|
|
||||||
|
|
||||||
PYMONGO_3_4: 3.4
|
|
||||||
PYMONGO_3_6: 3.6
|
|
||||||
PYMONGO_3_9: 3.9
|
|
||||||
PYMONGO_3_11: 3.11
|
|
||||||
|
|
||||||
MAIN_PYTHON_VERSION: 3.7
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
linting:
|
|
||||||
# Run pre-commit (https://pre-commit.com/)
|
|
||||||
# which runs pre-configured linter & autoformatter
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
- name: Set up Python 3.7
|
|
||||||
uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: 3.7
|
|
||||||
- run: bash .github/workflows/install_ci_python_dep.sh
|
|
||||||
- run: pre-commit run -a
|
|
||||||
|
|
||||||
test:
|
|
||||||
# Test suite run against recent python versions
|
|
||||||
# and against a few combination of MongoDB and pymongo
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
strategy:
|
|
||||||
fail-fast: false
|
|
||||||
matrix:
|
|
||||||
python-version: [3.6, 3.7, 3.8, 3.9, pypy3]
|
|
||||||
MONGODB: [$MONGODB_4_0]
|
|
||||||
PYMONGO: [$PYMONGO_3_11]
|
|
||||||
include:
|
|
||||||
- python-version: 3.7
|
|
||||||
MONGODB: $MONGODB_3_6
|
|
||||||
PYMONGO: $PYMONGO_3_9
|
|
||||||
- python-version: 3.7
|
|
||||||
MONGODB: $MONGODB_4_2
|
|
||||||
PYMONGO: $PYMONGO_3_6
|
|
||||||
- python-version: 3.7
|
|
||||||
MONGODB: $MONGODB_4_4
|
|
||||||
PYMONGO: $PYMONGO_3_11
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
- name: Set up Python ${{ matrix.python-version }}
|
|
||||||
uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: ${{ matrix.python-version }}
|
|
||||||
- name: install mongo and ci dependencies
|
|
||||||
run: |
|
|
||||||
bash .github/workflows/install_mongo.sh ${{ matrix.MONGODB }}
|
|
||||||
bash .github/workflows/install_ci_python_dep.sh
|
|
||||||
bash .github/workflows/start_mongo.sh ${{ matrix.MONGODB }}
|
|
||||||
- name: tox dry-run (to pre-install venv)
|
|
||||||
run: tox -e $(echo py${{ matrix.python-version }}-mg${{ matrix.PYMONGO }} | tr -d . | sed -e 's/pypypy/pypy/') -- -a "-k=test_ci_placeholder"
|
|
||||||
- name: Run test suite
|
|
||||||
run: tox -e $(echo py${{ matrix.python-version }}-mg${{ matrix.PYMONGO }} | tr -d . | sed -e 's/pypypy/pypy/') -- -a "--cov=mongoengine"
|
|
||||||
- name: Send coverage to Coveralls
|
|
||||||
env:
|
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
COVERALLS_SERVICE_NAME: github
|
|
||||||
if: ${{ matrix.python-version == env.MAIN_PYTHON_VERSION }}
|
|
||||||
run: coveralls
|
|
||||||
|
|
||||||
build_doc_dryrun:
|
|
||||||
# ensures that readthedocs can be built continuously
|
|
||||||
# to avoid that it breaks when new releases are being created
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v2
|
|
||||||
- name: Set up Python ${{ matrix.python-version }}
|
|
||||||
uses: actions/setup-python@v2
|
|
||||||
with:
|
|
||||||
python-version: 3.7
|
|
||||||
- name: install python dep
|
|
||||||
run: |
|
|
||||||
pip install -e .
|
|
||||||
pip install -r docs/requirements.txt
|
|
||||||
- name: build doc
|
|
||||||
run: |
|
|
||||||
cd docs
|
|
||||||
make html-readthedocs
|
|
||||||
|
|
||||||
build-n-publish-dummy:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
needs: [linting, test, build_doc_dryrun]
|
|
||||||
if: github.event_name != 'pull_request'
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@master
|
|
||||||
- name: Set up Python 3.7
|
|
||||||
uses: actions/setup-python@v1
|
|
||||||
with:
|
|
||||||
python-version: 3.7
|
|
||||||
- name: build dummy wheel for test-pypi
|
|
||||||
run: |
|
|
||||||
pip install wheel
|
|
||||||
python setup.py egg_info -b ".dev`date '+%Y%m%d%H%M%S'`" build sdist bdist_wheel
|
|
||||||
# - name: publish test-pypi
|
|
||||||
# # Although working and recommended, test-pypi has a limit
|
|
||||||
# # in the size of projects so it's better to avoid publishing
|
|
||||||
# # until there is a way to garbage collect these dummy releases
|
|
||||||
# uses: pypa/gh-action-pypi-publish@master
|
|
||||||
# with:
|
|
||||||
# password: ${{ secrets.test_pypi_token }}
|
|
||||||
# repository_url: https://test.pypi.org/legacy/
|
|
||||||
|
|
||||||
build-n-publish:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
needs: [linting, test, build_doc_dryrun, build-n-publish-dummy]
|
|
||||||
if: github.event_name == 'create' && startsWith(github.ref, 'refs/tags/v')
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@master
|
|
||||||
- name: Set up Python 3.7
|
|
||||||
uses: actions/setup-python@v1
|
|
||||||
with:
|
|
||||||
python-version: 3.7
|
|
||||||
# todo separate build from publish
|
|
||||||
# https://stackoverflow.com/questions/59349905/which-properties-does-github-event-in-a-github-workflow-have
|
|
||||||
- name: build dummy wheel for test-pypi
|
|
||||||
run: |
|
|
||||||
pip install wheel
|
|
||||||
python setup.py sdist bdist_wheel
|
|
||||||
- name: publish pypi
|
|
||||||
uses: pypa/gh-action-pypi-publish@master
|
|
||||||
with:
|
|
||||||
password: ${{ secrets.pypi_token }}
|
|
5
.github/workflows/install_ci_python_dep.sh
vendored
5
.github/workflows/install_ci_python_dep.sh
vendored
@@ -1,5 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
pip install --upgrade pip
|
|
||||||
pip install coveralls
|
|
||||||
pip install pre-commit
|
|
||||||
pip install tox
|
|
18
.github/workflows/install_mongo.sh
vendored
18
.github/workflows/install_mongo.sh
vendored
@@ -1,18 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
MONGODB=$1
|
|
||||||
|
|
||||||
# Mongo > 4.0 follows different name convention for download links
|
|
||||||
mongo_build=mongodb-linux-x86_64-${MONGODB}
|
|
||||||
|
|
||||||
if [[ "$MONGODB" == *"4.2"* ]]; then
|
|
||||||
mongo_build=mongodb-linux-x86_64-ubuntu1804-v${MONGODB}-latest
|
|
||||||
elif [[ "$MONGODB" == *"4.4"* ]]; then
|
|
||||||
mongo_build=mongodb-linux-x86_64-ubuntu1804-v${MONGODB}-latest
|
|
||||||
fi
|
|
||||||
|
|
||||||
wget http://fastdl.mongodb.org/linux/$mongo_build.tgz
|
|
||||||
tar xzf $mongo_build.tgz
|
|
||||||
|
|
||||||
mongodb_dir=$(find ${PWD}/ -type d -name "mongodb-linux-x86_64*")
|
|
||||||
$mongodb_dir/bin/mongod --version
|
|
33
.github/workflows/main.yml
vendored
Normal file
33
.github/workflows/main.yml
vendored
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# This is a basic workflow to help you get started with Actions
|
||||||
|
|
||||||
|
name: CI
|
||||||
|
|
||||||
|
# Controls when the action will run. Triggers the workflow on push or pull request
|
||||||
|
# events but only for the master branch
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches: [ master ]
|
||||||
|
pull_request:
|
||||||
|
branches: [ master ]
|
||||||
|
|
||||||
|
# A workflow run is made up of one or more jobs that can run sequentially or in parallel
|
||||||
|
jobs:
|
||||||
|
# This workflow contains a single job called "build"
|
||||||
|
build:
|
||||||
|
# The type of runner that the job will run on
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
# Steps represent a sequence of tasks that will be executed as part of the job
|
||||||
|
steps:
|
||||||
|
# Checks-out your repository under $GITHUB_WORKSPACE, so your job can access it
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
|
||||||
|
# Runs a single command using the runners shell
|
||||||
|
- name: Run a one-line script
|
||||||
|
run: echo Hello, world!
|
||||||
|
|
||||||
|
# Runs a set of commands using the runners shell
|
||||||
|
- name: Run a multi-line script
|
||||||
|
run: |
|
||||||
|
echo Add other actions to build,
|
||||||
|
echo test, and deploy your project.
|
9
.github/workflows/start_mongo.sh
vendored
9
.github/workflows/start_mongo.sh
vendored
@@ -1,9 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
MONGODB=$1
|
|
||||||
|
|
||||||
mongodb_dir=$(find ${PWD}/ -type d -name "mongodb-linux-x86_64*")
|
|
||||||
|
|
||||||
mkdir $mongodb_dir/data
|
|
||||||
$mongodb_dir/bin/mongod --dbpath $mongodb_dir/data --logpath $mongodb_dir/mongodb.log --fork
|
|
||||||
mongo --eval 'db.version();' # Make sure mongo is awake
|
|
17
.landscape.yml
Normal file
17
.landscape.yml
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
pylint:
|
||||||
|
disable:
|
||||||
|
# We use this a lot (e.g. via document._meta)
|
||||||
|
- protected-access
|
||||||
|
|
||||||
|
options:
|
||||||
|
additional-builtins:
|
||||||
|
# add long as valid built-ins.
|
||||||
|
- long
|
||||||
|
|
||||||
|
pyflakes:
|
||||||
|
disable:
|
||||||
|
# undefined variables are already covered by pylint (and exclude long)
|
||||||
|
- F821
|
||||||
|
|
||||||
|
ignore-paths:
|
||||||
|
- benchmark.py
|
@@ -1,26 +1,12 @@
|
|||||||
fail_fast: false
|
fail_fast: false
|
||||||
repos:
|
repos:
|
||||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
- repo: https://github.com/ambv/black
|
||||||
rev: v3.4.0
|
rev: 19.10b0
|
||||||
hooks:
|
hooks:
|
||||||
- id: check-merge-conflict
|
- id: black
|
||||||
- id: debug-statements
|
- repo: https://gitlab.com/pycqa/flake8
|
||||||
- id: trailing-whitespace
|
rev: 3.8.0a2
|
||||||
- id: end-of-file-fixer
|
hooks:
|
||||||
- repo: https://github.com/ambv/black
|
- id: flake8
|
||||||
rev: 21.4b2
|
additional_dependencies:
|
||||||
hooks:
|
- flake8-import-order
|
||||||
- id: black
|
|
||||||
- repo: https://gitlab.com/pycqa/flake8
|
|
||||||
rev: 3.9.1
|
|
||||||
hooks:
|
|
||||||
- id: flake8
|
|
||||||
- repo: https://github.com/asottile/pyupgrade
|
|
||||||
rev: v2.14.0
|
|
||||||
hooks:
|
|
||||||
- id: pyupgrade
|
|
||||||
args: [--py36-plus]
|
|
||||||
- repo: https://github.com/pycqa/isort
|
|
||||||
rev: 5.8.0
|
|
||||||
hooks:
|
|
||||||
- id: isort
|
|
||||||
|
107
.travis.yml
Normal file
107
.travis.yml
Normal file
@@ -0,0 +1,107 @@
|
|||||||
|
# For full coverage, we'd have to test all supported Python, MongoDB, and
|
||||||
|
# PyMongo combinations. However, that would result in an overly long build
|
||||||
|
# with a very large number of jobs, hence we only test a subset of all the
|
||||||
|
# combinations.
|
||||||
|
# * Python3.7, MongoDB v3.4 & the latest PyMongo v3.x is currently the "main" setup,
|
||||||
|
# Other combinations are tested. See below for the details or check the travis jobs
|
||||||
|
|
||||||
|
# We should periodically check MongoDB Server versions supported by MongoDB
|
||||||
|
# Inc., add newly released versions to the test matrix, and remove versions
|
||||||
|
# which have reached their End of Life. See:
|
||||||
|
# 1. https://www.mongodb.com/support-policy.
|
||||||
|
# 2. https://docs.mongodb.com/ecosystem/drivers/driver-compatibility-reference/#python-driver-compatibility
|
||||||
|
#
|
||||||
|
# Reminder: Update README.rst if you change MongoDB versions we test.
|
||||||
|
|
||||||
|
language: python
|
||||||
|
dist: xenial
|
||||||
|
python:
|
||||||
|
- 3.6
|
||||||
|
- 3.7
|
||||||
|
- 3.8
|
||||||
|
- 3.9
|
||||||
|
- pypy3
|
||||||
|
|
||||||
|
env:
|
||||||
|
global:
|
||||||
|
- MONGODB_3_4=3.4.19
|
||||||
|
- MONGODB_3_6=3.6.13
|
||||||
|
- MONGODB_4_0=4.0.13
|
||||||
|
|
||||||
|
- PYMONGO_3_4=3.4
|
||||||
|
- PYMONGO_3_6=3.6
|
||||||
|
- PYMONGO_3_9=3.9
|
||||||
|
- PYMONGO_3_11=3.11
|
||||||
|
|
||||||
|
- MAIN_PYTHON_VERSION=3.7
|
||||||
|
matrix:
|
||||||
|
- MONGODB=${MONGODB_3_4} PYMONGO=${PYMONGO_3_11}
|
||||||
|
|
||||||
|
matrix:
|
||||||
|
# Finish the build as soon as one job fails
|
||||||
|
fast_finish: true
|
||||||
|
|
||||||
|
include:
|
||||||
|
- python: 3.7
|
||||||
|
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_6}
|
||||||
|
- python: 3.7
|
||||||
|
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_9}
|
||||||
|
- python: 3.7
|
||||||
|
env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_11}
|
||||||
|
- python: 3.8
|
||||||
|
env: MONGODB=${MONGODB_4_0} PYMONGO=${PYMONGO_3_11}
|
||||||
|
|
||||||
|
install:
|
||||||
|
# Install Mongo
|
||||||
|
- wget http://fastdl.mongodb.org/linux/mongodb-linux-x86_64-${MONGODB}.tgz
|
||||||
|
- tar xzf mongodb-linux-x86_64-${MONGODB}.tgz
|
||||||
|
- ${PWD}/mongodb-linux-x86_64-${MONGODB}/bin/mongod --version
|
||||||
|
# Install Python dependencies.
|
||||||
|
- pip install --upgrade pip
|
||||||
|
- pip install coveralls
|
||||||
|
- pip install pre-commit
|
||||||
|
- pip install tox
|
||||||
|
# tox dryrun to setup the tox venv (we run a mock test).
|
||||||
|
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "-k=test_ci_placeholder"
|
||||||
|
|
||||||
|
before_script:
|
||||||
|
- mkdir ${PWD}/mongodb-linux-x86_64-${MONGODB}/data
|
||||||
|
- ${PWD}/mongodb-linux-x86_64-${MONGODB}/bin/mongod --dbpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/data --logpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/mongodb.log --fork
|
||||||
|
# Run pre-commit hooks (black, flake8, etc) on entire codebase
|
||||||
|
- if [[ $TRAVIS_PYTHON_VERSION == $MAIN_PYTHON_VERSION ]]; then pre-commit run -a; else echo "pre-commit checks only runs on py37"; fi
|
||||||
|
- mongo --eval 'db.version();' # Make sure mongo is awake
|
||||||
|
|
||||||
|
script:
|
||||||
|
- tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "--cov=mongoengine"
|
||||||
|
|
||||||
|
after_success:
|
||||||
|
- if [[ $TRAVIS_PYTHON_VERSION == $MAIN_PYTHON_VERSION ]]; then coveralls --verbose; else echo "coveralls only sent for py37"; fi
|
||||||
|
|
||||||
|
notifications:
|
||||||
|
irc: irc.freenode.org#mongoengine
|
||||||
|
|
||||||
|
# Only run builds on the master branch and GitHub releases (tagged as vX.Y.Z)
|
||||||
|
branches:
|
||||||
|
only:
|
||||||
|
- master
|
||||||
|
- /^v.*$/
|
||||||
|
|
||||||
|
# Whenever a new release is created via GitHub, publish it on PyPI.
|
||||||
|
deploy:
|
||||||
|
provider: pypi
|
||||||
|
user: the_drow
|
||||||
|
password:
|
||||||
|
secure: QMyatmWBnC6ZN3XLW2+fTBDU4LQcp1m/LjR2/0uamyeUzWKdlOoh/Wx5elOgLwt/8N9ppdPeG83ose1jOz69l5G0MUMjv8n/RIcMFSpCT59tGYqn3kh55b0cIZXFT9ar+5cxlif6a5rS72IHm5li7QQyxexJIII6Uxp0kpvUmek=
|
||||||
|
|
||||||
|
# Create a source distribution and a pure python wheel for faster installs.
|
||||||
|
distributions: "sdist bdist_wheel"
|
||||||
|
|
||||||
|
# Only deploy on tagged commits (aka GitHub releases) and only for the parent
|
||||||
|
# repo's builds running Python v3.7 along with PyMongo v3.x and MongoDB v3.4.
|
||||||
|
# We run Travis against many different Python, PyMongo, and MongoDB versions
|
||||||
|
# and we don't want the deploy to occur multiple times).
|
||||||
|
on:
|
||||||
|
tags: true
|
||||||
|
repo: MongoEngine/mongoengine
|
||||||
|
condition: ($PYMONGO = ${PYMONGO_3_11}) && ($MONGODB = ${MONGODB_3_4})
|
||||||
|
python: 3.7
|
108
.travis_.yml
108
.travis_.yml
@@ -1,108 +0,0 @@
|
|||||||
## For full coverage, we'd have to test all supported Python, MongoDB, and
|
|
||||||
## PyMongo combinations. However, that would result in an overly long build
|
|
||||||
## with a very large number of jobs, hence we only test a subset of all the
|
|
||||||
## combinations.
|
|
||||||
## * Python3.7, MongoDB v3.4 & the latest PyMongo v3.x is currently the "main" setup,
|
|
||||||
## Other combinations are tested. See below for the details or check the travis jobs
|
|
||||||
#
|
|
||||||
## We should periodically check MongoDB Server versions supported by MongoDB
|
|
||||||
## Inc., add newly released versions to the test matrix, and remove versions
|
|
||||||
## which have reached their End of Life. See:
|
|
||||||
## 1. https://www.mongodb.com/support-policy.
|
|
||||||
## 2. https://docs.mongodb.com/ecosystem/drivers/driver-compatibility-reference/#python-driver-compatibility
|
|
||||||
##
|
|
||||||
## Reminder: Update README.rst if you change MongoDB versions we test.
|
|
||||||
#
|
|
||||||
#language: python
|
|
||||||
#dist: xenial
|
|
||||||
#python:
|
|
||||||
# - 3.6
|
|
||||||
# - 3.7
|
|
||||||
# - 3.8
|
|
||||||
# - 3.9
|
|
||||||
# - pypy3
|
|
||||||
#
|
|
||||||
#env:
|
|
||||||
# global:
|
|
||||||
# - MONGODB_3_4=3.4.19
|
|
||||||
# - MONGODB_3_6=3.6.13
|
|
||||||
# - MONGODB_4_0=4.0.13
|
|
||||||
#
|
|
||||||
# - PYMONGO_3_4=3.4
|
|
||||||
# - PYMONGO_3_6=3.6
|
|
||||||
# - PYMONGO_3_9=3.9
|
|
||||||
# - PYMONGO_3_11=3.11
|
|
||||||
#
|
|
||||||
# - MAIN_PYTHON_VERSION=3.7
|
|
||||||
# matrix:
|
|
||||||
# - MONGODB=${MONGODB_3_4} PYMONGO=${PYMONGO_3_11}
|
|
||||||
#
|
|
||||||
#matrix:
|
|
||||||
# # Finish the build as soon as one job fails
|
|
||||||
# fast_finish: true
|
|
||||||
#
|
|
||||||
# include:
|
|
||||||
# - python: 3.7
|
|
||||||
# env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_6}
|
|
||||||
# - python: 3.7
|
|
||||||
# env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_9}
|
|
||||||
# - python: 3.7
|
|
||||||
# env: MONGODB=${MONGODB_3_6} PYMONGO=${PYMONGO_3_11}
|
|
||||||
# - python: 3.8
|
|
||||||
# env: MONGODB=${MONGODB_4_0} PYMONGO=${PYMONGO_3_11}
|
|
||||||
#
|
|
||||||
#install:
|
|
||||||
# # Install Mongo
|
|
||||||
# - wget http://fastdl.mongodb.org/linux/mongodb-linux-x86_64-${MONGODB}.tgz
|
|
||||||
# - tar xzf mongodb-linux-x86_64-${MONGODB}.tgz
|
|
||||||
# - ${PWD}/mongodb-linux-x86_64-${MONGODB}/bin/mongod --version
|
|
||||||
# # Install Python dependencies.
|
|
||||||
# - pip install --upgrade pip
|
|
||||||
# - pip install coveralls
|
|
||||||
# - pip install pre-commit
|
|
||||||
# - pip install tox
|
|
||||||
# # tox dryrun to setup the tox venv (we run a mock test).
|
|
||||||
# - tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "-k=test_ci_placeholder"
|
|
||||||
#
|
|
||||||
#before_script:
|
|
||||||
# - mkdir ${PWD}/mongodb-linux-x86_64-${MONGODB}/data
|
|
||||||
# - ${PWD}/mongodb-linux-x86_64-${MONGODB}/bin/mongod --dbpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/data --logpath ${PWD}/mongodb-linux-x86_64-${MONGODB}/mongodb.log --fork
|
|
||||||
# # Run pre-commit hooks (black, flake8, etc) on entire codebase
|
|
||||||
# - if [[ $TRAVIS_PYTHON_VERSION == $MAIN_PYTHON_VERSION ]]; then pre-commit run -a; else echo "pre-commit checks only runs on py37"; fi
|
|
||||||
# - mongo --eval 'db.version();' # Make sure mongo is awake
|
|
||||||
#
|
|
||||||
#script:
|
|
||||||
# - tox -e $(echo py$TRAVIS_PYTHON_VERSION-mg$PYMONGO | tr -d . | sed -e 's/pypypy/pypy/') -- -a "--cov=mongoengine"
|
|
||||||
#
|
|
||||||
#after_success:
|
|
||||||
# - if [[ $TRAVIS_PYTHON_VERSION == $MAIN_PYTHON_VERSION ]]; then coveralls --verbose; else echo "coveralls only sent for py37"; fi
|
|
||||||
#
|
|
||||||
#notifications:
|
|
||||||
# irc: irc.freenode.org#mongoengine
|
|
||||||
#
|
|
||||||
## Only run builds on the master branch and GitHub releases (tagged as vX.Y.Z)
|
|
||||||
#branches:
|
|
||||||
# # Only run builds on the master branch and GitHub releases (tagged as vX.Y.Z)
|
|
||||||
# only:
|
|
||||||
# - master
|
|
||||||
# - /^v.*$/
|
|
||||||
#
|
|
||||||
## Whenever a new release is created via GitHub, publish it on PyPI.
|
|
||||||
#deploy:
|
|
||||||
# provider: pypi
|
|
||||||
# user: the_drow
|
|
||||||
# password:
|
|
||||||
# secure: QMyatmWBnC6ZN3XLW2+fTBDU4LQcp1m/LjR2/0uamyeUzWKdlOoh/Wx5elOgLwt/8N9ppdPeG83ose1jOz69l5G0MUMjv8n/RIcMFSpCT59tGYqn3kh55b0cIZXFT9ar+5cxlif6a5rS72IHm5li7QQyxexJIII6Uxp0kpvUmek=
|
|
||||||
#
|
|
||||||
# # Create a source distribution and a pure python wheel for faster installs.
|
|
||||||
# distributions: "sdist bdist_wheel"
|
|
||||||
#
|
|
||||||
# # Only deploy on tagged commits (aka GitHub releases) and only for the parent
|
|
||||||
# # repo's builds running Python v3.7 along with PyMongo v3.x and MongoDB v3.4.
|
|
||||||
# # We run Travis against many different Python, PyMongo, and MongoDB versions
|
|
||||||
# # and we don't want the deploy to occur multiple times).
|
|
||||||
# on:
|
|
||||||
# tags: true
|
|
||||||
# repo: MongoEngine/mongoengine
|
|
||||||
# condition: ($PYMONGO = ${PYMONGO_3_11}) && ($MONGODB = ${MONGODB_3_4})
|
|
||||||
# python: 3.7
|
|
1
AUTHORS
1
AUTHORS
@@ -259,4 +259,3 @@ that much better:
|
|||||||
* Agustin Barto (https://github.com/abarto)
|
* Agustin Barto (https://github.com/abarto)
|
||||||
* Stankiewicz Mateusz (https://github.com/mas15)
|
* Stankiewicz Mateusz (https://github.com/mas15)
|
||||||
* Felix Schultheiß (https://github.com/felix-smashdocs)
|
* Felix Schultheiß (https://github.com/felix-smashdocs)
|
||||||
* Jan Stein (https://github.com/janste63)
|
|
||||||
|
@@ -35,8 +35,8 @@ Travis runs the tests against the main Python 3.x versions.
|
|||||||
Style Guide
|
Style Guide
|
||||||
-----------
|
-----------
|
||||||
|
|
||||||
MongoEngine's codebase is auto-formatted with `black <https://github.com/python/black>`_, imports are ordered with `isort <https://pycqa.github.io/isort/>`_
|
MongoEngine's codebase is formatted with `black <https://github.com/python/black>`_, other tools like
|
||||||
and other tools like flake8 are also used. Those tools will run as part of the CI and will fail in case the code is not formatted properly.
|
flake8 are also used. Those tools will run as part of the CI and will fail in case the code is not formatted properly.
|
||||||
|
|
||||||
To install all development tools, simply run the following commands:
|
To install all development tools, simply run the following commands:
|
||||||
|
|
||||||
@@ -58,10 +58,6 @@ To enable ``pre-commit`` simply run:
|
|||||||
See the ``.pre-commit-config.yaml`` configuration file for more information
|
See the ``.pre-commit-config.yaml`` configuration file for more information
|
||||||
on how it works.
|
on how it works.
|
||||||
|
|
||||||
pre-commit will now run upon every commit and will reject anything that doesn't comply.
|
|
||||||
|
|
||||||
You can also run all the checks with ``pre-commit run -a``, this is what is used in the CI.
|
|
||||||
|
|
||||||
Testing
|
Testing
|
||||||
-------
|
-------
|
||||||
|
|
||||||
|
@@ -12,6 +12,10 @@ MongoEngine
|
|||||||
.. image:: https://coveralls.io/repos/github/MongoEngine/mongoengine/badge.svg?branch=master
|
.. image:: https://coveralls.io/repos/github/MongoEngine/mongoengine/badge.svg?branch=master
|
||||||
:target: https://coveralls.io/github/MongoEngine/mongoengine?branch=master
|
:target: https://coveralls.io/github/MongoEngine/mongoengine?branch=master
|
||||||
|
|
||||||
|
.. image:: https://landscape.io/github/MongoEngine/mongoengine/master/landscape.svg?style=flat
|
||||||
|
:target: https://landscape.io/github/MongoEngine/mongoengine/master
|
||||||
|
:alt: Code Health
|
||||||
|
|
||||||
.. image:: https://img.shields.io/badge/code%20style-black-000000.svg
|
.. image:: https://img.shields.io/badge/code%20style-black-000000.svg
|
||||||
:target: https://github.com/ambv/black
|
:target: https://github.com/ambv/black
|
||||||
|
|
||||||
|
@@ -45,7 +45,7 @@ def test_basic():
|
|||||||
|
|
||||||
print(
|
print(
|
||||||
"Doc setattr: %.3fus"
|
"Doc setattr: %.3fus"
|
||||||
% (timeit(lambda: setattr(b, "name", "New name"), 10000) * 10 ** 6) # noqa B010
|
% (timeit(lambda: setattr(b, "name", "New name"), 10000) * 10 ** 6)
|
||||||
)
|
)
|
||||||
|
|
||||||
print("Doc to mongo: %.3fus" % (timeit(b.to_mongo, 1000) * 10 ** 6))
|
print("Doc to mongo: %.3fus" % (timeit(b.to_mongo, 1000) * 10 ** 6))
|
||||||
|
@@ -31,7 +31,7 @@ myNoddys = noddy.find()
|
|||||||
print("-" * 100)
|
print("-" * 100)
|
||||||
print("PyMongo: Creating 10000 dictionaries.")
|
print("PyMongo: Creating 10000 dictionaries.")
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
stmt = """
|
stmt = """
|
||||||
from pymongo import MongoClient, WriteConcern
|
from pymongo import MongoClient, WriteConcern
|
||||||
@@ -54,7 +54,7 @@ myNoddys = noddy.find()
|
|||||||
print("-" * 100)
|
print("-" * 100)
|
||||||
print('PyMongo: Creating 10000 dictionaries (write_concern={"w": 0}).')
|
print('PyMongo: Creating 10000 dictionaries (write_concern={"w": 0}).')
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
setup = """
|
setup = """
|
||||||
from pymongo import MongoClient
|
from pymongo import MongoClient
|
||||||
@@ -84,7 +84,7 @@ myNoddys = Noddy.objects()
|
|||||||
print("-" * 100)
|
print("-" * 100)
|
||||||
print("MongoEngine: Creating 10000 dictionaries.")
|
print("MongoEngine: Creating 10000 dictionaries.")
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
stmt = """
|
stmt = """
|
||||||
for i in range(10000):
|
for i in range(10000):
|
||||||
@@ -102,7 +102,7 @@ myNoddys = Noddy.objects()
|
|||||||
print("-" * 100)
|
print("-" * 100)
|
||||||
print("MongoEngine: Creating 10000 dictionaries (using a single field assignment).")
|
print("MongoEngine: Creating 10000 dictionaries (using a single field assignment).")
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
stmt = """
|
stmt = """
|
||||||
for i in range(10000):
|
for i in range(10000):
|
||||||
@@ -118,7 +118,7 @@ myNoddys = Noddy.objects()
|
|||||||
print("-" * 100)
|
print("-" * 100)
|
||||||
print('MongoEngine: Creating 10000 dictionaries (write_concern={"w": 0}).')
|
print('MongoEngine: Creating 10000 dictionaries (write_concern={"w": 0}).')
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
stmt = """
|
stmt = """
|
||||||
for i in range(10000):
|
for i in range(10000):
|
||||||
@@ -136,7 +136,7 @@ myNoddys = Noddy.objects()
|
|||||||
'MongoEngine: Creating 10000 dictionaries (write_concern={"w": 0}, validate=False).'
|
'MongoEngine: Creating 10000 dictionaries (write_concern={"w": 0}, validate=False).'
|
||||||
)
|
)
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
stmt = """
|
stmt = """
|
||||||
for i in range(10000):
|
for i in range(10000):
|
||||||
@@ -154,7 +154,7 @@ myNoddys = Noddy.objects()
|
|||||||
'MongoEngine: Creating 10000 dictionaries (force_insert=True, write_concern={"w": 0}, validate=False).'
|
'MongoEngine: Creating 10000 dictionaries (force_insert=True, write_concern={"w": 0}, validate=False).'
|
||||||
)
|
)
|
||||||
t = timeit.Timer(stmt=stmt, setup=setup)
|
t = timeit.Timer(stmt=stmt, setup=setup)
|
||||||
print(f"{t.timeit(1)}s")
|
print("{}s".format(t.timeit(1)))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
@@ -33,14 +33,8 @@ clean:
|
|||||||
html:
|
html:
|
||||||
$(SPHINXBUILD) -b html $(ALLSPHINXOPTS) $(BUILDDIR)/html
|
$(SPHINXBUILD) -b html $(ALLSPHINXOPTS) $(BUILDDIR)/html
|
||||||
@echo
|
@echo
|
||||||
@echo "Build finished. Check $(BUILDDIR)/html/index.html"
|
|
||||||
|
|
||||||
html-readthedocs:
|
|
||||||
$(SPHINXBUILD) -T -E -b readthedocs $(ALLSPHINXOPTS) $(BUILDDIR)/html
|
|
||||||
@echo
|
|
||||||
@echo "Build finished. The HTML pages are in $(BUILDDIR)/html."
|
@echo "Build finished. The HTML pages are in $(BUILDDIR)/html."
|
||||||
|
|
||||||
|
|
||||||
dirhtml:
|
dirhtml:
|
||||||
$(SPHINXBUILD) -b dirhtml $(ALLSPHINXOPTS) $(BUILDDIR)/dirhtml
|
$(SPHINXBUILD) -b dirhtml $(ALLSPHINXOPTS) $(BUILDDIR)/dirhtml
|
||||||
@echo
|
@echo
|
||||||
|
@@ -1,5 +1,4 @@
|
|||||||
|
|
||||||
|
|
||||||
=========
|
=========
|
||||||
Changelog
|
Changelog
|
||||||
=========
|
=========
|
||||||
@@ -7,45 +6,17 @@ Changelog
|
|||||||
Development
|
Development
|
||||||
===========
|
===========
|
||||||
- (Fill this out as you fix issues and develop your features).
|
- (Fill this out as you fix issues and develop your features).
|
||||||
|
|
||||||
Changes in 0.23.1
|
|
||||||
===========
|
|
||||||
- Bug fix: ignore LazyReferenceFields when clearing _changed_fields #2484
|
|
||||||
- Improve connection doc #2481
|
|
||||||
|
|
||||||
Changes in 0.23.0
|
|
||||||
=================
|
|
||||||
- Bugfix: manually setting SequenceField in DynamicDocument doesn't increment the counter #2471
|
|
||||||
- Add MongoDB 4.2 and 4.4 to CI
|
|
||||||
- Add support for allowDiskUse on querysets #2468
|
|
||||||
|
|
||||||
Changes in 0.22.1
|
|
||||||
=================
|
|
||||||
- Declare that Py3.5 is not supported in package metadata #2449
|
|
||||||
- Moved CI from Travis to Github-Actions
|
|
||||||
|
|
||||||
Changes in 0.22.0
|
|
||||||
=================
|
|
||||||
- Fix LazyReferenceField dereferencing in embedded documents #2426
|
|
||||||
- Fix regarding the recent use of Cursor.__spec in .count() that was interfering with mongomock #2425
|
|
||||||
- Drop support for Python 3.5 by introducing f-strings in the codebase
|
|
||||||
|
|
||||||
Changes in 0.21.0
|
|
||||||
=================
|
|
||||||
- Bug fix in DynamicDocument which is not parsing known fields in constructor like Document do #2412
|
- Bug fix in DynamicDocument which is not parsing known fields in constructor like Document do #2412
|
||||||
- When using pymongo >= 3.7, make use of Collection.count_documents instead of Collection.count
|
- When using pymongo >= 3.7, make use of Collection.count_documents instead of Collection.count
|
||||||
and Cursor.count that got deprecated in pymongo >= 3.7.
|
and Cursor.count that got deprecated in pymongo >= 3.7.
|
||||||
This should have a negative impact on performance of count see Issue #2219
|
This should have a negative impact on performance of count see Issue #2219
|
||||||
- Fix a bug that made the queryset drop the read_preference after clone().
|
- Fix a bug that made the queryset drop the read_preference after clone().
|
||||||
- Remove Py3.5 from CI as it reached EOL and add Python 3.9
|
- Remove Py3.5 from CI as it reached EOL and add Python 3.9
|
||||||
- Fix some issues related with db_field/field conflict in constructor #2414
|
- Fix the behavior of Doc.objects.limit(0) which should return all documents (similar to mongodb) #2311
|
||||||
- BREAKING CHANGE: Fix the behavior of Doc.objects.limit(0) which should return all documents (similar to mongodb) #2311
|
|
||||||
- Bug fix in ListField when updating the first item, it was saving the whole list, instead of
|
- Bug fix in ListField when updating the first item, it was saving the whole list, instead of
|
||||||
just replacing the first item (as usually done when updating 1 item of the list) #2392
|
just replacing the first item (as it's usually done) #2392
|
||||||
- Add EnumField: ``mongoengine.fields.EnumField``
|
- Add EnumField: ``mongoengine.fields.EnumField``
|
||||||
- Refactoring - Remove useless code related to Document.__only_fields and Queryset.only_fields
|
- Refactoring - Remove useless code related to Document.__only_fields and Queryset.only_fields
|
||||||
- Fix query transformation regarding special operators #2365
|
|
||||||
- Bug Fix: Document.save() fails when shard_key is not _id #2154
|
|
||||||
|
|
||||||
Changes in 0.20.0
|
Changes in 0.20.0
|
||||||
=================
|
=================
|
||||||
|
@@ -26,7 +26,7 @@ sys.path.insert(0, os.path.abspath(".."))
|
|||||||
|
|
||||||
# Add any Sphinx extension module names here, as strings. They can be extensions
|
# Add any Sphinx extension module names here, as strings. They can be extensions
|
||||||
# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
|
# coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
|
||||||
extensions = ["sphinx.ext.autodoc", "sphinx.ext.todo", "readthedocs_ext.readthedocs"]
|
extensions = ["sphinx.ext.autodoc", "sphinx.ext.todo"]
|
||||||
|
|
||||||
# Add any paths that contain templates here, relative to this directory.
|
# Add any paths that contain templates here, relative to this directory.
|
||||||
templates_path = ["_templates"]
|
templates_path = ["_templates"]
|
||||||
@@ -41,8 +41,8 @@ source_suffix = ".rst"
|
|||||||
master_doc = "index"
|
master_doc = "index"
|
||||||
|
|
||||||
# General information about the project.
|
# General information about the project.
|
||||||
project = "MongoEngine"
|
project = u"MongoEngine"
|
||||||
copyright = "2009, MongoEngine Authors" # noqa: A001
|
copyright = u"2009, MongoEngine Authors"
|
||||||
|
|
||||||
# The version info for the project you're documenting, acts as replacement for
|
# The version info for the project you're documenting, acts as replacement for
|
||||||
# |version| and |release|, also used in various other places throughout the
|
# |version| and |release|, also used in various other places throughout the
|
||||||
|
@@ -5,7 +5,7 @@ Connecting to MongoDB
|
|||||||
=====================
|
=====================
|
||||||
|
|
||||||
Connections in MongoEngine are registered globally and are identified with aliases.
|
Connections in MongoEngine are registered globally and are identified with aliases.
|
||||||
If no ``alias`` is provided during the connection, it will use "default" as alias.
|
If no `alias` is provided during the connection, it will use "default" as alias.
|
||||||
|
|
||||||
To connect to a running instance of :program:`mongod`, use the :func:`~mongoengine.connect`
|
To connect to a running instance of :program:`mongod`, use the :func:`~mongoengine.connect`
|
||||||
function. The first argument is the name of the database to connect to::
|
function. The first argument is the name of the database to connect to::
|
||||||
@@ -14,66 +14,27 @@ function. The first argument is the name of the database to connect to::
|
|||||||
connect('project1')
|
connect('project1')
|
||||||
|
|
||||||
By default, MongoEngine assumes that the :program:`mongod` instance is running
|
By default, MongoEngine assumes that the :program:`mongod` instance is running
|
||||||
on **localhost** on port **27017**.
|
on **localhost** on port **27017**. If MongoDB is running elsewhere, you should
|
||||||
|
provide the :attr:`host` and :attr:`port` arguments to
|
||||||
|
:func:`~mongoengine.connect`::
|
||||||
|
|
||||||
If MongoDB is running elsewhere, you need to provide details on how to connect. There are two ways of
|
connect('project1', host='192.168.1.35', port=12345)
|
||||||
doing this. Using a connection string in URI format (**this is the preferred method**) or individual attributes
|
|
||||||
provided as keyword arguments.
|
|
||||||
|
|
||||||
Connect with URI string
|
|
||||||
=======================
|
|
||||||
|
|
||||||
When using a connection string in URI format you should specify the connection details
|
|
||||||
as the :attr:`host` to :func:`~mongoengine.connect`. In a web application context for instance, the URI
|
|
||||||
is typically read from the config file::
|
|
||||||
|
|
||||||
connect(host="mongodb://127.0.0.1:27017/my_db")
|
|
||||||
|
|
||||||
If the database requires authentication, you can specify it in the
|
|
||||||
URI. As each database can have its own users configured, you need to tell MongoDB
|
|
||||||
where to look for the user you are working with, that's what the ``?authSource=admin`` bit
|
|
||||||
of the MongoDB connection string is for::
|
|
||||||
|
|
||||||
# Connects to 'my_db' database by authenticating
|
|
||||||
# with given credentials against the 'admin' database (by default as authSource isn't provided)
|
|
||||||
connect(host="mongodb://my_user:my_password@127.0.0.1:27017/my_db")
|
|
||||||
|
|
||||||
# Equivalent to previous connection but explicitly states that
|
|
||||||
# it should use admin as the authentication source database
|
|
||||||
connect(host="mongodb://my_user:my_password@hostname:port/my_db?authSource=admin")
|
|
||||||
|
|
||||||
# Connects to 'my_db' database by authenticating
|
|
||||||
# with given credentials against that same database
|
|
||||||
connect(host="mongodb://my_user:my_password@127.0.0.1:27017/my_db?authSource=my_db")
|
|
||||||
|
|
||||||
The URI string can also be used to configure advanced parameters like ssl, replicaSet, etc. For more
|
|
||||||
information or example about URI string, you can refer to the `official doc <https://docs.mongodb.com/manual/reference/connection-string/>`_::
|
|
||||||
|
|
||||||
connect(host="mongodb://my_user:my_password@127.0.0.1:27017/my_db?authSource=admin&ssl=true&replicaSet=globaldb")
|
|
||||||
|
|
||||||
.. note:: URI containing SRV records (e.g "mongodb+srv://server.example.com/") can be used as well
|
|
||||||
|
|
||||||
Connect with keyword attributes
|
|
||||||
===============================
|
|
||||||
|
|
||||||
The second option for specifying the connection details is to provide the information as keyword
|
|
||||||
attributes to :func:`~mongoengine.connect`::
|
|
||||||
|
|
||||||
connect('my_db', host='127.0.0.1', port=27017)
|
|
||||||
|
|
||||||
If the database requires authentication, :attr:`username`, :attr:`password`
|
If the database requires authentication, :attr:`username`, :attr:`password`
|
||||||
and :attr:`authentication_source` arguments should be provided::
|
and :attr:`authentication_source` arguments should be provided::
|
||||||
|
|
||||||
connect('my_db', username='my_user', password='my_password', authentication_source='admin')
|
connect('project1', username='webapp', password='pwd123', authentication_source='admin')
|
||||||
|
|
||||||
The set of attributes that :func:`~mongoengine.connect` recognizes includes but is not limited to:
|
URI style connections are also supported -- just supply the URI as
|
||||||
:attr:`host`, :attr:`port`, :attr:`read_preference`, :attr:`username`, :attr:`password`, :attr:`authentication_source`, :attr:`authentication_mechanism`,
|
the :attr:`host` to
|
||||||
:attr:`replicaset`, :attr:`tls`, etc. Most of the parameters accepted by `pymongo.MongoClient <https://pymongo.readthedocs.io/en/stable/api/pymongo/mongo_client.html#pymongo.mongo_client.MongoClient>`_
|
:func:`~mongoengine.connect`::
|
||||||
can be used with :func:`~mongoengine.connect` and will simply be forwarded when instantiating the `pymongo.MongoClient`.
|
|
||||||
|
connect('project1', host='mongodb://localhost/database_name')
|
||||||
|
|
||||||
|
.. note:: URI containing SRV records (e.g mongodb+srv://server.example.com/) can be used as well as the :attr:`host`
|
||||||
|
|
||||||
.. note:: Database, username and password from URI string overrides
|
.. note:: Database, username and password from URI string overrides
|
||||||
corresponding parameters in :func:`~mongoengine.connect`, this should
|
corresponding parameters in :func:`~mongoengine.connect`: ::
|
||||||
obviously be avoided: ::
|
|
||||||
|
|
||||||
connect(
|
connect(
|
||||||
db='test',
|
db='test',
|
||||||
@@ -82,19 +43,28 @@ can be used with :func:`~mongoengine.connect` and will simply be forwarded when
|
|||||||
host='mongodb://admin:qwerty@localhost/production'
|
host='mongodb://admin:qwerty@localhost/production'
|
||||||
)
|
)
|
||||||
|
|
||||||
will establish connection to ``production`` database using ``admin`` username and ``qwerty`` password.
|
will establish connection to ``production`` database using
|
||||||
|
``admin`` username and ``qwerty`` password.
|
||||||
|
|
||||||
.. note:: Calling :func:`~mongoengine.connect` without argument will establish
|
.. note:: Calling :func:`~mongoengine.connect` without argument will establish
|
||||||
a connection to the "test" database by default
|
a connection to the "test" database by default
|
||||||
|
|
||||||
Read Preferences
|
Replica Sets
|
||||||
================
|
============
|
||||||
|
|
||||||
As stated above, Read preferences are supported through the connection but also via individual
|
MongoEngine supports connecting to replica sets::
|
||||||
|
|
||||||
|
from mongoengine import connect
|
||||||
|
|
||||||
|
# Regular connect
|
||||||
|
connect('dbname', replicaset='rs-name')
|
||||||
|
|
||||||
|
# MongoDB URI-style connect
|
||||||
|
connect(host='mongodb://localhost/dbname?replicaSet=rs-name')
|
||||||
|
|
||||||
|
Read preferences are supported through the connection or via individual
|
||||||
queries by passing the read_preference ::
|
queries by passing the read_preference ::
|
||||||
|
|
||||||
from pymongo import ReadPreference
|
|
||||||
|
|
||||||
Bar.objects().read_preference(ReadPreference.PRIMARY)
|
Bar.objects().read_preference(ReadPreference.PRIMARY)
|
||||||
Bar.objects(read_preference=ReadPreference.PRIMARY)
|
Bar.objects(read_preference=ReadPreference.PRIMARY)
|
||||||
|
|
||||||
|
@@ -290,12 +290,12 @@ as the constructor's argument::
|
|||||||
content = StringField()
|
content = StringField()
|
||||||
|
|
||||||
|
|
||||||
.. _many-to-many-with-listfields:
|
.. _one-to-many-with-listfields:
|
||||||
|
|
||||||
Many to Many with ListFields
|
One to Many with ListFields
|
||||||
'''''''''''''''''''''''''''
|
'''''''''''''''''''''''''''
|
||||||
|
|
||||||
If you are implementing a many to many relationship via a list of references,
|
If you are implementing a one to many relationship via a list of references,
|
||||||
then the references are stored as DBRefs and to query you need to pass an
|
then the references are stored as DBRefs and to query you need to pass an
|
||||||
instance of the object to the query::
|
instance of the object to the query::
|
||||||
|
|
||||||
@@ -432,10 +432,10 @@ Document collections
|
|||||||
====================
|
====================
|
||||||
Document classes that inherit **directly** from :class:`~mongoengine.Document`
|
Document classes that inherit **directly** from :class:`~mongoengine.Document`
|
||||||
will have their own **collection** in the database. The name of the collection
|
will have their own **collection** in the database. The name of the collection
|
||||||
is by default the name of the class converted to snake_case (e.g if your Document class
|
is by default the name of the class, converted to lowercase (so in the example
|
||||||
is named `CompanyUser`, the corresponding collection would be `company_user`). If you need
|
above, the collection would be called `page`). If you need to change the name
|
||||||
to change the name of the collection (e.g. to use MongoEngine with an existing database),
|
of the collection (e.g. to use MongoEngine with an existing database), then
|
||||||
then create a class dictionary attribute called :attr:`meta` on your document, and
|
create a class dictionary attribute called :attr:`meta` on your document, and
|
||||||
set :attr:`collection` to the name of the collection that you want your
|
set :attr:`collection` to the name of the collection that you want your
|
||||||
document class to use::
|
document class to use::
|
||||||
|
|
||||||
|
@@ -2,6 +2,8 @@
|
|||||||
GridFS
|
GridFS
|
||||||
======
|
======
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
|
||||||
Writing
|
Writing
|
||||||
-------
|
-------
|
||||||
|
|
||||||
|
@@ -14,6 +14,5 @@ User Guide
|
|||||||
gridfs
|
gridfs
|
||||||
signals
|
signals
|
||||||
text-indexes
|
text-indexes
|
||||||
migration
|
|
||||||
logging-monitoring
|
logging-monitoring
|
||||||
mongomock
|
mongomock
|
||||||
|
@@ -1,267 +0,0 @@
|
|||||||
===================
|
|
||||||
Documents migration
|
|
||||||
===================
|
|
||||||
|
|
||||||
The structure of your documents and their associated mongoengine schemas are likely
|
|
||||||
to change over the lifetime of an application. This section provides guidance and
|
|
||||||
recommendations on how to deal with migrations.
|
|
||||||
|
|
||||||
Due to the very flexible nature of mongodb, migrations of models aren't trivial and
|
|
||||||
for people that know about `alembic` for `sqlalchemy`, there is unfortunately no equivalent
|
|
||||||
library that will manage the migration in an automatic fashion for mongoengine.
|
|
||||||
|
|
||||||
Example 1: Addition of a field
|
|
||||||
==============================
|
|
||||||
|
|
||||||
Let's start by taking a simple example of a model change and review the different option you
|
|
||||||
have to deal with the migration.
|
|
||||||
|
|
||||||
Let's assume we start with the following schema and save an instance:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
class User(Document):
|
|
||||||
name = StringField()
|
|
||||||
|
|
||||||
User(name="John Doe").save()
|
|
||||||
|
|
||||||
# print the objects as they exist in mongodb
|
|
||||||
print(User.objects().as_pymongo()) # [{u'_id': ObjectId('5d06b9c3d7c1f18db3e7c874'), u'name': u'John Doe'}]
|
|
||||||
|
|
||||||
On the next version of your application, let's now assume that a new field `enabled` gets added to the
|
|
||||||
existing ``User`` model with a `default=True`. Thus you simply update the ``User`` class to the following:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
class User(Document):
|
|
||||||
name = StringField(required=True)
|
|
||||||
enabled = BooleanField(default=True)
|
|
||||||
|
|
||||||
Without applying any migration, we now reload an object from the database into the ``User`` class
|
|
||||||
and checks its `enabled` attribute:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
assert User.objects.count() == 1
|
|
||||||
user = User.objects().first()
|
|
||||||
assert user.enabled is True
|
|
||||||
assert User.objects(enabled=True).count() == 0 # uh?
|
|
||||||
assert User.objects(enabled=False).count() == 0 # uh?
|
|
||||||
|
|
||||||
# this is consistent with what we have in the database
|
|
||||||
# in fact, 'enabled' does not exist
|
|
||||||
print(User.objects().as_pymongo().first()) # {u'_id': ObjectId('5d06b9c3d7c1f18db3e7c874'), u'name': u'John'}
|
|
||||||
assert User.objects(enabled=None).count() == 1
|
|
||||||
|
|
||||||
As you can see, even if the document wasn't updated, mongoengine applies the default value seamlessly when it
|
|
||||||
loads the pymongo dict into a ``User`` instance. At first sight it looks like you don't need to migrate the
|
|
||||||
existing documents when adding new fields but this actually leads to inconsistencies when it comes to querying.
|
|
||||||
|
|
||||||
In fact, when querying, mongoengine isn't trying to account for the default value of the new field and so
|
|
||||||
if you don't actually migrate the existing documents, you are taking a risk that querying/updating
|
|
||||||
will be missing relevant record.
|
|
||||||
|
|
||||||
When adding fields/modifying default values, you can use any of the following to do the migration
|
|
||||||
as a standalone script:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
# Use mongoengine to set a default value for a given field
|
|
||||||
User.objects().update(enabled=True)
|
|
||||||
# or use pymongo
|
|
||||||
user_coll = User._get_collection()
|
|
||||||
user_coll.update_many({}, {'$set': {'enabled': True}})
|
|
||||||
|
|
||||||
|
|
||||||
Example 2: Inheritance change
|
|
||||||
=============================
|
|
||||||
|
|
||||||
Let's consider the following example:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
class Human(Document):
|
|
||||||
name = StringField()
|
|
||||||
meta = {"allow_inheritance": True}
|
|
||||||
|
|
||||||
class Jedi(Human):
|
|
||||||
dark_side = BooleanField()
|
|
||||||
light_saber_color = StringField()
|
|
||||||
|
|
||||||
Jedi(name="Darth Vader", dark_side=True, light_saber_color="red").save()
|
|
||||||
Jedi(name="Obi Wan Kenobi", dark_side=False, light_saber_color="blue").save()
|
|
||||||
|
|
||||||
assert Human.objects.count() == 2
|
|
||||||
assert Jedi.objects.count() == 2
|
|
||||||
|
|
||||||
# Let's check how these documents got stored in mongodb
|
|
||||||
print(Jedi.objects.as_pymongo())
|
|
||||||
# [
|
|
||||||
# {'_id': ObjectId('5fac4aaaf61d7fb06046e0f9'), '_cls': 'Human.Jedi', 'name': 'Darth Vader', 'dark_side': True, 'light_saber_color': 'red'},
|
|
||||||
# {'_id': ObjectId('5fac4ac4f61d7fb06046e0fa'), '_cls': 'Human.Jedi', 'name': 'Obi Wan Kenobi', 'dark_side': False, 'light_saber_color': 'blue'}
|
|
||||||
# ]
|
|
||||||
|
|
||||||
As you can observe, when you use inheritance, MongoEngine stores a field named '_cls' behind the scene to keep
|
|
||||||
track of the Document class.
|
|
||||||
|
|
||||||
Let's now take the scenario that you want to refactor the inheritance schema and:
|
|
||||||
- Have the Jedi's with dark_side=True/False become GoodJedi's/DarkSith
|
|
||||||
- get rid of the 'dark_side' field
|
|
||||||
|
|
||||||
move to the following schemas:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
# unchanged
|
|
||||||
class Human(Document):
|
|
||||||
name = StringField()
|
|
||||||
meta = {"allow_inheritance": True}
|
|
||||||
|
|
||||||
# attribute 'dark_side' removed
|
|
||||||
class GoodJedi(Human):
|
|
||||||
light_saber_color = StringField()
|
|
||||||
|
|
||||||
# new class
|
|
||||||
class BadSith(Human):
|
|
||||||
light_saber_color = StringField()
|
|
||||||
|
|
||||||
MongoEngine doesn't know about the change or how to map them with the existing data
|
|
||||||
so if you don't apply any migration, you will observe a strange behavior, as if the collection was suddenly
|
|
||||||
empty.
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
# As a reminder, the documents that we inserted
|
|
||||||
# have the _cls field = 'Human.Jedi'
|
|
||||||
|
|
||||||
# Following has no match
|
|
||||||
# because the query that is used behind the scene is
|
|
||||||
# filtering on {'_cls': 'Human.GoodJedi'}
|
|
||||||
assert GoodJedi.objects().count() == 0
|
|
||||||
|
|
||||||
# Following has also no match
|
|
||||||
# because it is filtering on {'_cls': {'$in': ('Human', 'Human.GoodJedi', 'Human.BadSith')}}
|
|
||||||
# which has no match
|
|
||||||
assert Human.objects.count() == 0
|
|
||||||
assert Human.objects.first() is None
|
|
||||||
|
|
||||||
# If we bypass MongoEngine and make use of underlying driver (PyMongo)
|
|
||||||
# we can see that the documents are there
|
|
||||||
humans_coll = Human._get_collection()
|
|
||||||
assert humans_coll.count_documents({}) == 2
|
|
||||||
# print first document
|
|
||||||
print(humans_coll.find_one())
|
|
||||||
# {'_id': ObjectId('5fac4aaaf61d7fb06046e0f9'), '_cls': 'Human.Jedi', 'name': 'Darth Vader', 'dark_side': True, 'light_saber_color': 'red'}
|
|
||||||
|
|
||||||
As you can see, first obvious problem is that we need to modify '_cls' values based on existing values of
|
|
||||||
'dark_side' documents.
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
humans_coll = Human._get_collection()
|
|
||||||
old_class = 'Human.Jedi'
|
|
||||||
good_jedi_class = 'Human.GoodJedi'
|
|
||||||
bad_sith_class = 'Human.BadSith'
|
|
||||||
humans_coll.update_many({'_cls': old_class, 'dark_side': False}, {'$set': {'_cls': good_jedi_class}})
|
|
||||||
humans_coll.update_many({'_cls': old_class, 'dark_side': True}, {'$set': {'_cls': bad_sith_class}})
|
|
||||||
|
|
||||||
Let's now check if querying improved in MongoEngine:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
assert GoodJedi.objects().count() == 1 # Hoorah!
|
|
||||||
assert BadSith.objects().count() == 1 # Hoorah!
|
|
||||||
assert Human.objects.count() == 2 # Hoorah!
|
|
||||||
|
|
||||||
# let's now check that documents load correctly
|
|
||||||
jedi = GoodJedi.objects().first()
|
|
||||||
# raises FieldDoesNotExist: The fields "{'dark_side'}" do not exist on the document "Human.GoodJedi"
|
|
||||||
|
|
||||||
In fact we only took care of renaming the _cls values but we havn't removed the 'dark_side' fields
|
|
||||||
which does not exist anymore on the GoodJedi's and BadSith's models.
|
|
||||||
Let's remove the field from the collections:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
humans_coll = Human._get_collection()
|
|
||||||
humans_coll.update_many({}, {'$unset': {'dark_side': 1}})
|
|
||||||
|
|
||||||
.. note:: We did this migration in 2 different steps for the sake of example but it could have been combined
|
|
||||||
with the migration of the _cls fields: ::
|
|
||||||
|
|
||||||
humans_coll.update_many(
|
|
||||||
{'_cls': old_class, 'dark_side': False},
|
|
||||||
{
|
|
||||||
'$set': {'_cls': good_jedi_class},
|
|
||||||
'$unset': {'dark_side': 1}
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
And verify that the documents now load correctly:
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
jedi = GoodJedi.objects().first()
|
|
||||||
assert jedi.name == "Obi Wan Kenobi"
|
|
||||||
|
|
||||||
sith = BadSith.objects().first()
|
|
||||||
assert sith.name == "Darth Vader"
|
|
||||||
|
|
||||||
|
|
||||||
An other way of dealing with this migration is to iterate over
|
|
||||||
the documents and update/replace them one by one. This is way slower but
|
|
||||||
it is often useful for complex migrations of Document models.
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
for doc in humans_coll.find():
|
|
||||||
if doc['_cls'] == 'Human.Jedi':
|
|
||||||
doc['_cls'] = 'Human.BadSith' if doc['dark_side'] else 'Human.GoodJedi'
|
|
||||||
doc.pop('dark_side')
|
|
||||||
humans_coll.replace_one({'_id': doc['_id']}, doc)
|
|
||||||
|
|
||||||
.. warning:: Be aware of this `flaw <https://groups.google.com/g/mongodb-user/c/AFC1ia7MHzk>`_ if you modify documents while iterating
|
|
||||||
|
|
||||||
Recommendations
|
|
||||||
===============
|
|
||||||
|
|
||||||
- Write migration scripts whenever you do changes to the model schemas
|
|
||||||
- Using :class:`~mongoengine.DynamicDocument` or ``meta = {"strict": False}`` may help to avoid some migrations or to have the 2 versions of your application to co-exist.
|
|
||||||
- Write post-processing checks to verify that migrations script worked. See below
|
|
||||||
|
|
||||||
Post-processing checks
|
|
||||||
======================
|
|
||||||
|
|
||||||
The following recipe can be used to sanity check a Document collection after you applied migration.
|
|
||||||
It does not make any assumption on what was migrated, it will fetch 1000 objects randomly and
|
|
||||||
run some quick checks on the documents to make sure the document looks OK. As it is, it will fail
|
|
||||||
on the first occurrence of an error but this is something that can be adapted based on your needs.
|
|
||||||
|
|
||||||
.. code-block:: python
|
|
||||||
|
|
||||||
def get_random_oids(collection, sample_size):
|
|
||||||
pipeline = [{"$project": {'_id': 1}}, {"$sample": {"size": sample_size}}]
|
|
||||||
return [s['_id'] for s in collection.aggregate(pipeline)]
|
|
||||||
|
|
||||||
def get_random_documents(DocCls, sample_size):
|
|
||||||
doc_collection = DocCls._get_collection()
|
|
||||||
random_oids = get_random_oids(doc_collection, sample_size)
|
|
||||||
return DocCls.objects(id__in=random_oids)
|
|
||||||
|
|
||||||
def check_documents(DocCls, sample_size):
|
|
||||||
for doc in get_random_documents(DocCls, sample_size):
|
|
||||||
# general validation (types and values)
|
|
||||||
doc.validate()
|
|
||||||
|
|
||||||
# load all subfields,
|
|
||||||
# this may trigger additional queries if you have ReferenceFields
|
|
||||||
# so it may be slow
|
|
||||||
for field in doc._fields:
|
|
||||||
try:
|
|
||||||
getattr(doc, field)
|
|
||||||
except Exception:
|
|
||||||
LOG.warning(f"Could not load field {field} in Document {doc.id}")
|
|
||||||
raise
|
|
||||||
|
|
||||||
check_documents(Human, sample_size=1000)
|
|
@@ -120,3 +120,4 @@ the validation and cleaning of a document when you call :meth:`~mongoengine.docu
|
|||||||
Person(age=1000).save(validate=False)
|
Person(age=1000).save(validate=False)
|
||||||
person = Person.objects.first()
|
person = Person.objects.first()
|
||||||
assert person.age == 1000
|
assert person.age == 1000
|
||||||
|
|
||||||
|
@@ -1,3 +1,3 @@
|
|||||||
Sphinx==3.3.0
|
pymongo>=3.11
|
||||||
|
Sphinx==3.2.1
|
||||||
sphinx-rtd-theme==0.5.0
|
sphinx-rtd-theme==0.5.0
|
||||||
readthedocs-sphinx-ext==2.1.1
|
|
||||||
|
@@ -1,23 +1,22 @@
|
|||||||
# Import submodules so that we can expose their __all__
|
# Import submodules so that we can expose their __all__
|
||||||
from mongoengine import (
|
from mongoengine import connection
|
||||||
connection,
|
from mongoengine import document
|
||||||
document,
|
from mongoengine import errors
|
||||||
errors,
|
from mongoengine import fields
|
||||||
fields,
|
from mongoengine import queryset
|
||||||
queryset,
|
from mongoengine import signals
|
||||||
signals,
|
|
||||||
)
|
|
||||||
|
|
||||||
# Import everything from each submodule so that it can be accessed via
|
# Import everything from each submodule so that it can be accessed via
|
||||||
# mongoengine, e.g. instead of `from mongoengine.connection import connect`,
|
# mongoengine, e.g. instead of `from mongoengine.connection import connect`,
|
||||||
# users can simply use `from mongoengine import connect`, or even
|
# users can simply use `from mongoengine import connect`, or even
|
||||||
# `from mongoengine import *` and then `connect('testdb')`.
|
# `from mongoengine import *` and then `connect('testdb')`.
|
||||||
from mongoengine.connection import * # noqa: F401
|
from mongoengine.connection import *
|
||||||
from mongoengine.document import * # noqa: F401
|
from mongoengine.document import *
|
||||||
from mongoengine.errors import * # noqa: F401
|
from mongoengine.errors import *
|
||||||
from mongoengine.fields import * # noqa: F401
|
from mongoengine.fields import *
|
||||||
from mongoengine.queryset import * # noqa: F401
|
from mongoengine.queryset import *
|
||||||
from mongoengine.signals import * # noqa: F401
|
from mongoengine.signals import *
|
||||||
|
|
||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
list(document.__all__)
|
list(document.__all__)
|
||||||
@@ -29,7 +28,7 @@ __all__ = (
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
VERSION = (0, 23, 1)
|
VERSION = (0, 20, 0)
|
||||||
|
|
||||||
|
|
||||||
def get_version():
|
def get_version():
|
||||||
|
@@ -67,11 +67,11 @@ class BaseDict(dict):
|
|||||||
if isinstance(value, EmbeddedDocument) and value._instance is None:
|
if isinstance(value, EmbeddedDocument) and value._instance is None:
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
||||||
value = BaseDict(value, None, f"{self._name}.{key}")
|
value = BaseDict(value, None, "{}.{}".format(self._name, key))
|
||||||
super().__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, list) and not isinstance(value, BaseList):
|
elif isinstance(value, list) and not isinstance(value, BaseList):
|
||||||
value = BaseList(value, None, f"{self._name}.{key}")
|
value = BaseList(value, None, "{}.{}".format(self._name, key))
|
||||||
super().__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
return value
|
return value
|
||||||
@@ -97,7 +97,7 @@ class BaseDict(dict):
|
|||||||
def _mark_as_changed(self, key=None):
|
def _mark_as_changed(self, key=None):
|
||||||
if hasattr(self._instance, "_mark_as_changed"):
|
if hasattr(self._instance, "_mark_as_changed"):
|
||||||
if key:
|
if key:
|
||||||
self._instance._mark_as_changed(f"{self._name}.{key}")
|
self._instance._mark_as_changed("{}.{}".format(self._name, key))
|
||||||
else:
|
else:
|
||||||
self._instance._mark_as_changed(self._name)
|
self._instance._mark_as_changed(self._name)
|
||||||
|
|
||||||
@@ -133,12 +133,12 @@ class BaseList(list):
|
|||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
elif isinstance(value, dict) and not isinstance(value, BaseDict):
|
||||||
# Replace dict by BaseDict
|
# Replace dict by BaseDict
|
||||||
value = BaseDict(value, None, f"{self._name}.{key}")
|
value = BaseDict(value, None, "{}.{}".format(self._name, key))
|
||||||
super().__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
elif isinstance(value, list) and not isinstance(value, BaseList):
|
elif isinstance(value, list) and not isinstance(value, BaseList):
|
||||||
# Replace list by BaseList
|
# Replace list by BaseList
|
||||||
value = BaseList(value, None, f"{self._name}.{key}")
|
value = BaseList(value, None, "{}.{}".format(self._name, key))
|
||||||
super().__setitem__(key, value)
|
super().__setitem__(key, value)
|
||||||
value._instance = self._instance
|
value._instance = self._instance
|
||||||
return value
|
return value
|
||||||
@@ -180,7 +180,9 @@ class BaseList(list):
|
|||||||
def _mark_as_changed(self, key=None):
|
def _mark_as_changed(self, key=None):
|
||||||
if hasattr(self._instance, "_mark_as_changed"):
|
if hasattr(self._instance, "_mark_as_changed"):
|
||||||
if key is not None:
|
if key is not None:
|
||||||
self._instance._mark_as_changed(f"{self._name}.{key % len(self)}")
|
self._instance._mark_as_changed(
|
||||||
|
"{}.{}".format(self._name, key % len(self))
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
self._instance._mark_as_changed(self._name)
|
self._instance._mark_as_changed(self._name)
|
||||||
|
|
||||||
@@ -427,7 +429,7 @@ class StrictDict:
|
|||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return "{%s}" % ", ".join(
|
return "{%s}" % ", ".join(
|
||||||
f'"{k!s}": {v!r}' for k, v in self.items()
|
'"{!s}": {!r}'.format(k, v) for k, v in self.items()
|
||||||
)
|
)
|
||||||
|
|
||||||
cls._classes[allowed_keys] = SpecificStrictDict
|
cls._classes[allowed_keys] = SpecificStrictDict
|
||||||
@@ -470,4 +472,4 @@ class LazyReference(DBRef):
|
|||||||
raise AttributeError()
|
raise AttributeError()
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return f"<LazyReference({self.document_type}, {self.pk!r})>"
|
return "<LazyReference({}, {!r})>".format(self.document_type, self.pk)
|
||||||
|
@@ -1,9 +1,10 @@
|
|||||||
import copy
|
import copy
|
||||||
|
|
||||||
import numbers
|
import numbers
|
||||||
from functools import partial
|
from functools import partial
|
||||||
|
|
||||||
|
from bson import DBRef, ObjectId, SON, json_util
|
||||||
import pymongo
|
import pymongo
|
||||||
from bson import SON, DBRef, ObjectId, json_util
|
|
||||||
|
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
from mongoengine.base.common import get_document
|
from mongoengine.base.common import get_document
|
||||||
@@ -88,7 +89,9 @@ class BaseDocument:
|
|||||||
list(self._fields.keys()) + ["id", "pk", "_cls", "_text_score"]
|
list(self._fields.keys()) + ["id", "pk", "_cls", "_text_score"]
|
||||||
)
|
)
|
||||||
if _undefined_fields:
|
if _undefined_fields:
|
||||||
msg = f'The fields "{_undefined_fields}" do not exist on the document "{self._class_name}"'
|
msg = ('The fields "{}" do not exist on the document "{}"').format(
|
||||||
|
_undefined_fields, self._class_name
|
||||||
|
)
|
||||||
raise FieldDoesNotExist(msg)
|
raise FieldDoesNotExist(msg)
|
||||||
|
|
||||||
if self.STRICT and not self._dynamic:
|
if self.STRICT and not self._dynamic:
|
||||||
@@ -98,13 +101,12 @@ class BaseDocument:
|
|||||||
|
|
||||||
self._dynamic_fields = SON()
|
self._dynamic_fields = SON()
|
||||||
|
|
||||||
# Assign default values for fields
|
# Assign default values to the instance.
|
||||||
# not set in the constructor
|
for key, field in self._fields.items():
|
||||||
for field_name in self._fields:
|
if self._db_field_map.get(key, key) in values:
|
||||||
if field_name in values:
|
|
||||||
continue
|
continue
|
||||||
value = getattr(self, field_name, None)
|
value = getattr(self, key, None)
|
||||||
setattr(self, field_name, value)
|
setattr(self, key, value)
|
||||||
|
|
||||||
if "_cls" not in values:
|
if "_cls" not in values:
|
||||||
self._cls = self._class_name
|
self._cls = self._class_name
|
||||||
@@ -113,6 +115,7 @@ class BaseDocument:
|
|||||||
dynamic_data = {}
|
dynamic_data = {}
|
||||||
FileField = _import_class("FileField")
|
FileField = _import_class("FileField")
|
||||||
for key, value in values.items():
|
for key, value in values.items():
|
||||||
|
key = self._reverse_db_field_map.get(key, key)
|
||||||
field = self._fields.get(key)
|
field = self._fields.get(key)
|
||||||
if field or key in ("id", "pk", "_cls"):
|
if field or key in ("id", "pk", "_cls"):
|
||||||
if __auto_convert and value is not None:
|
if __auto_convert and value is not None:
|
||||||
@@ -155,7 +158,7 @@ class BaseDocument:
|
|||||||
# Handle dynamic data only if an initialised dynamic document
|
# Handle dynamic data only if an initialised dynamic document
|
||||||
if self._dynamic and not self._dynamic_lock:
|
if self._dynamic and not self._dynamic_lock:
|
||||||
|
|
||||||
if name not in self._fields_ordered and not name.startswith("_"):
|
if not hasattr(self, name) and not name.startswith("_"):
|
||||||
DynamicField = _import_class("DynamicField")
|
DynamicField = _import_class("DynamicField")
|
||||||
field = DynamicField(db_field=name, null=True)
|
field = DynamicField(db_field=name, null=True)
|
||||||
field.name = name
|
field.name = name
|
||||||
@@ -228,10 +231,10 @@ class BaseDocument:
|
|||||||
setattr(self, k, data[k])
|
setattr(self, k, data[k])
|
||||||
if "_fields_ordered" in data:
|
if "_fields_ordered" in data:
|
||||||
if self._dynamic:
|
if self._dynamic:
|
||||||
self._fields_ordered = data["_fields_ordered"]
|
setattr(self, "_fields_ordered", data["_fields_ordered"])
|
||||||
else:
|
else:
|
||||||
_super_fields_ordered = type(self)._fields_ordered
|
_super_fields_ordered = type(self)._fields_ordered
|
||||||
self._fields_ordered = _super_fields_ordered
|
setattr(self, "_fields_ordered", _super_fields_ordered)
|
||||||
|
|
||||||
dynamic_fields = data.get("_dynamic_fields") or SON()
|
dynamic_fields = data.get("_dynamic_fields") or SON()
|
||||||
for k in dynamic_fields.keys():
|
for k in dynamic_fields.keys():
|
||||||
@@ -241,7 +244,8 @@ class BaseDocument:
|
|||||||
return iter(self._fields_ordered)
|
return iter(self._fields_ordered)
|
||||||
|
|
||||||
def __getitem__(self, name):
|
def __getitem__(self, name):
|
||||||
"""Dictionary-style field access, return a field's value if present."""
|
"""Dictionary-style field access, return a field's value if present.
|
||||||
|
"""
|
||||||
try:
|
try:
|
||||||
if name in self._fields_ordered:
|
if name in self._fields_ordered:
|
||||||
return getattr(self, name)
|
return getattr(self, name)
|
||||||
@@ -250,7 +254,8 @@ class BaseDocument:
|
|||||||
raise KeyError(name)
|
raise KeyError(name)
|
||||||
|
|
||||||
def __setitem__(self, name, value):
|
def __setitem__(self, name, value):
|
||||||
"""Dictionary-style field access, set a field's value."""
|
"""Dictionary-style field access, set a field's value.
|
||||||
|
"""
|
||||||
# Ensure that the field exists before settings its value
|
# Ensure that the field exists before settings its value
|
||||||
if not self._dynamic and name not in self._fields:
|
if not self._dynamic and name not in self._fields:
|
||||||
raise KeyError(name)
|
raise KeyError(name)
|
||||||
@@ -272,7 +277,7 @@ class BaseDocument:
|
|||||||
except (UnicodeEncodeError, UnicodeDecodeError):
|
except (UnicodeEncodeError, UnicodeDecodeError):
|
||||||
u = "[Bad Unicode data]"
|
u = "[Bad Unicode data]"
|
||||||
repr_type = str if u is None else type(u)
|
repr_type = str if u is None else type(u)
|
||||||
return repr_type(f"<{self.__class__.__name__}: {u}>")
|
return repr_type("<{}: {}>".format(self.__class__.__name__, u))
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
# TODO this could be simpler?
|
# TODO this could be simpler?
|
||||||
@@ -428,7 +433,7 @@ class BaseDocument:
|
|||||||
pk = self.pk
|
pk = self.pk
|
||||||
elif self._instance and hasattr(self._instance, "pk"):
|
elif self._instance and hasattr(self._instance, "pk"):
|
||||||
pk = self._instance.pk
|
pk = self._instance.pk
|
||||||
message = f"ValidationError ({self._class_name}:{pk}) "
|
message = "ValidationError ({}:{}) ".format(self._class_name, pk)
|
||||||
raise ValidationError(message, errors=errors)
|
raise ValidationError(message, errors=errors)
|
||||||
|
|
||||||
def to_json(self, *args, **kwargs):
|
def to_json(self, *args, **kwargs):
|
||||||
@@ -501,7 +506,7 @@ class BaseDocument:
|
|||||||
if "." in key:
|
if "." in key:
|
||||||
key, rest = key.split(".", 1)
|
key, rest = key.split(".", 1)
|
||||||
key = self._db_field_map.get(key, key)
|
key = self._db_field_map.get(key, key)
|
||||||
key = f"{key}.{rest}"
|
key = "{}.{}".format(key, rest)
|
||||||
else:
|
else:
|
||||||
key = self._db_field_map.get(key, key)
|
key = self._db_field_map.get(key, key)
|
||||||
|
|
||||||
@@ -573,7 +578,7 @@ class BaseDocument:
|
|||||||
else:
|
else:
|
||||||
iterator = data.items()
|
iterator = data.items()
|
||||||
|
|
||||||
for _index_or_key, value in iterator:
|
for index_or_key, value in iterator:
|
||||||
if hasattr(value, "_get_changed_fields") and not isinstance(
|
if hasattr(value, "_get_changed_fields") and not isinstance(
|
||||||
value, Document
|
value, Document
|
||||||
): # don't follow references
|
): # don't follow references
|
||||||
@@ -597,7 +602,7 @@ class BaseDocument:
|
|||||||
iterator = data.items()
|
iterator = data.items()
|
||||||
|
|
||||||
for index_or_key, value in iterator:
|
for index_or_key, value in iterator:
|
||||||
item_key = f"{base_key}{index_or_key}."
|
item_key = "{}{}.".format(base_key, index_or_key)
|
||||||
# don't check anything lower if this key is already marked
|
# don't check anything lower if this key is already marked
|
||||||
# as changed.
|
# as changed.
|
||||||
if item_key[:-1] in changed_fields:
|
if item_key[:-1] in changed_fields:
|
||||||
@@ -605,18 +610,17 @@ class BaseDocument:
|
|||||||
|
|
||||||
if hasattr(value, "_get_changed_fields"):
|
if hasattr(value, "_get_changed_fields"):
|
||||||
changed = value._get_changed_fields()
|
changed = value._get_changed_fields()
|
||||||
changed_fields += [f"{item_key}{k}" for k in changed if k]
|
changed_fields += ["{}{}".format(item_key, k) for k in changed if k]
|
||||||
elif isinstance(value, (list, tuple, dict)):
|
elif isinstance(value, (list, tuple, dict)):
|
||||||
BaseDocument._nestable_types_changed_fields(
|
BaseDocument._nestable_types_changed_fields(
|
||||||
changed_fields, item_key, value
|
changed_fields, item_key, value
|
||||||
)
|
)
|
||||||
|
|
||||||
def _get_changed_fields(self):
|
def _get_changed_fields(self):
|
||||||
"""Return a list of all fields that have explicitly been changed."""
|
"""Return a list of all fields that have explicitly been changed.
|
||||||
|
"""
|
||||||
EmbeddedDocument = _import_class("EmbeddedDocument")
|
EmbeddedDocument = _import_class("EmbeddedDocument")
|
||||||
LazyReferenceField = _import_class("LazyReferenceField")
|
|
||||||
ReferenceField = _import_class("ReferenceField")
|
ReferenceField = _import_class("ReferenceField")
|
||||||
GenericLazyReferenceField = _import_class("GenericLazyReferenceField")
|
|
||||||
GenericReferenceField = _import_class("GenericReferenceField")
|
GenericReferenceField = _import_class("GenericReferenceField")
|
||||||
SortedListField = _import_class("SortedListField")
|
SortedListField = _import_class("SortedListField")
|
||||||
|
|
||||||
@@ -639,16 +643,10 @@ class BaseDocument:
|
|||||||
if isinstance(data, EmbeddedDocument):
|
if isinstance(data, EmbeddedDocument):
|
||||||
# Find all embedded fields that have been changed
|
# Find all embedded fields that have been changed
|
||||||
changed = data._get_changed_fields()
|
changed = data._get_changed_fields()
|
||||||
changed_fields += [f"{key}{k}" for k in changed if k]
|
changed_fields += ["{}{}".format(key, k) for k in changed if k]
|
||||||
elif isinstance(data, (list, tuple, dict)):
|
elif isinstance(data, (list, tuple, dict)):
|
||||||
if hasattr(field, "field") and isinstance(
|
if hasattr(field, "field") and isinstance(
|
||||||
field.field,
|
field.field, (ReferenceField, GenericReferenceField)
|
||||||
(
|
|
||||||
LazyReferenceField,
|
|
||||||
ReferenceField,
|
|
||||||
GenericLazyReferenceField,
|
|
||||||
GenericReferenceField,
|
|
||||||
),
|
|
||||||
):
|
):
|
||||||
continue
|
continue
|
||||||
elif isinstance(field, SortedListField) and field._ordering:
|
elif isinstance(field, SortedListField) and field._ordering:
|
||||||
@@ -752,7 +750,7 @@ class BaseDocument:
|
|||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _from_son(cls, son, _auto_dereference=True, created=False):
|
def _from_son(cls, son, _auto_dereference=True, created=False):
|
||||||
"""Create an instance of a Document (subclass) from a PyMongo SON (dict)"""
|
"""Create an instance of a Document (subclass) from a PyMongo SON."""
|
||||||
if son and not isinstance(son, dict):
|
if son and not isinstance(son, dict):
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
"The source SON object needs to be of type 'dict' but a '%s' was found"
|
"The source SON object needs to be of type 'dict' but a '%s' was found"
|
||||||
@@ -765,8 +763,6 @@ class BaseDocument:
|
|||||||
|
|
||||||
# Convert SON to a data dict, making sure each key is a string and
|
# Convert SON to a data dict, making sure each key is a string and
|
||||||
# corresponds to the right db field.
|
# corresponds to the right db field.
|
||||||
# This is needed as _from_son is currently called both from BaseDocument.__init__
|
|
||||||
# and from EmbeddedDocumentField.to_python
|
|
||||||
data = {}
|
data = {}
|
||||||
for key, value in son.items():
|
for key, value in son.items():
|
||||||
key = str(key)
|
key = str(key)
|
||||||
@@ -797,10 +793,11 @@ class BaseDocument:
|
|||||||
errors_dict[field_name] = e
|
errors_dict[field_name] = e
|
||||||
|
|
||||||
if errors_dict:
|
if errors_dict:
|
||||||
errors = "\n".join([f"Field '{k}' - {v}" for k, v in errors_dict.items()])
|
errors = "\n".join(
|
||||||
|
["Field '{}' - {}".format(k, v) for k, v in errors_dict.items()]
|
||||||
|
)
|
||||||
msg = "Invalid data to create a `{}` instance.\n{}".format(
|
msg = "Invalid data to create a `{}` instance.\n{}".format(
|
||||||
cls._class_name,
|
cls._class_name, errors,
|
||||||
errors,
|
|
||||||
)
|
)
|
||||||
raise InvalidDocumentError(msg)
|
raise InvalidDocumentError(msg)
|
||||||
|
|
||||||
@@ -968,7 +965,10 @@ class BaseDocument:
|
|||||||
unique_fields += unique_with
|
unique_fields += unique_with
|
||||||
|
|
||||||
# Add the new index to the list
|
# Add the new index to the list
|
||||||
fields = [(f"{namespace}{f}", pymongo.ASCENDING) for f in unique_fields]
|
fields = [
|
||||||
|
("{}{}".format(namespace, f), pymongo.ASCENDING)
|
||||||
|
for f in unique_fields
|
||||||
|
]
|
||||||
index = {"fields": fields, "unique": True, "sparse": sparse}
|
index = {"fields": fields, "unique": True, "sparse": sparse}
|
||||||
unique_indexes.append(index)
|
unique_indexes.append(index)
|
||||||
|
|
||||||
@@ -1004,7 +1004,9 @@ class BaseDocument:
|
|||||||
"PolygonField",
|
"PolygonField",
|
||||||
)
|
)
|
||||||
|
|
||||||
geo_field_types = tuple(_import_class(field) for field in geo_field_type_names)
|
geo_field_types = tuple(
|
||||||
|
[_import_class(field) for field in geo_field_type_names]
|
||||||
|
)
|
||||||
|
|
||||||
for field in cls._fields.values():
|
for field in cls._fields.values():
|
||||||
if not isinstance(field, geo_field_types):
|
if not isinstance(field, geo_field_types):
|
||||||
@@ -1022,7 +1024,7 @@ class BaseDocument:
|
|||||||
elif field._geo_index:
|
elif field._geo_index:
|
||||||
field_name = field.db_field
|
field_name = field.db_field
|
||||||
if parent_field:
|
if parent_field:
|
||||||
field_name = f"{parent_field}.{field_name}"
|
field_name = "{}.{}".format(parent_field, field_name)
|
||||||
geo_indices.append({"fields": [(field_name, field._geo_index)]})
|
geo_indices.append({"fields": [(field_name, field._geo_index)]})
|
||||||
|
|
||||||
return geo_indices
|
return geo_indices
|
||||||
@@ -1160,7 +1162,8 @@ class BaseDocument:
|
|||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def _translate_field_name(cls, field, sep="."):
|
def _translate_field_name(cls, field, sep="."):
|
||||||
"""Translate a field attribute name to a database field name."""
|
"""Translate a field attribute name to a database field name.
|
||||||
|
"""
|
||||||
parts = field.split(sep)
|
parts = field.split(sep)
|
||||||
parts = [f.db_field for f in cls._lookup_field(parts)]
|
parts = [f.db_field for f in cls._lookup_field(parts)]
|
||||||
return ".".join(parts)
|
return ".".join(parts)
|
||||||
|
@@ -1,15 +1,12 @@
|
|||||||
import operator
|
import operator
|
||||||
|
import warnings
|
||||||
import weakref
|
import weakref
|
||||||
|
|
||||||
|
from bson import DBRef, ObjectId, SON
|
||||||
import pymongo
|
import pymongo
|
||||||
from bson import SON, DBRef, ObjectId
|
|
||||||
|
|
||||||
from mongoengine.base.common import UPDATE_OPERATORS
|
from mongoengine.base.common import UPDATE_OPERATORS
|
||||||
from mongoengine.base.datastructures import (
|
from mongoengine.base.datastructures import BaseDict, BaseList, EmbeddedDocumentList
|
||||||
BaseDict,
|
|
||||||
BaseList,
|
|
||||||
EmbeddedDocumentList,
|
|
||||||
)
|
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.errors import DeprecatedError, ValidationError
|
from mongoengine.errors import DeprecatedError, ValidationError
|
||||||
|
|
||||||
@@ -19,9 +16,11 @@ __all__ = ("BaseField", "ComplexBaseField", "ObjectIdField", "GeoJsonBaseField")
|
|||||||
class BaseField:
|
class BaseField:
|
||||||
"""A base class for fields in a MongoDB document. Instances of this class
|
"""A base class for fields in a MongoDB document. Instances of this class
|
||||||
may be added to subclasses of `Document` to define a document's schema.
|
may be added to subclasses of `Document` to define a document's schema.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.5 - added verbose and help text
|
||||||
"""
|
"""
|
||||||
|
|
||||||
name = None # set in TopLevelDocumentMetaclass
|
name = None
|
||||||
_geo_index = False
|
_geo_index = False
|
||||||
_auto_gen = False # Call `generate` to generate a value
|
_auto_gen = False # Call `generate` to generate a value
|
||||||
_auto_dereference = True
|
_auto_dereference = True
|
||||||
@@ -44,7 +43,7 @@ class BaseField:
|
|||||||
choices=None,
|
choices=None,
|
||||||
null=False,
|
null=False,
|
||||||
sparse=False,
|
sparse=False,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
:param db_field: The database field to store this field in
|
:param db_field: The database field to store this field in
|
||||||
@@ -121,7 +120,8 @@ class BaseField:
|
|||||||
BaseField.creation_counter += 1
|
BaseField.creation_counter += 1
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
"""Descriptor for retrieving a value from a field in a document."""
|
"""Descriptor for retrieving a value from a field in a document.
|
||||||
|
"""
|
||||||
if instance is None:
|
if instance is None:
|
||||||
# Document class being used rather than a document object
|
# Document class being used rather than a document object
|
||||||
return self
|
return self
|
||||||
@@ -265,22 +265,11 @@ class ComplexBaseField(BaseField):
|
|||||||
Allows for nesting of embedded documents inside complex types.
|
Allows for nesting of embedded documents inside complex types.
|
||||||
Handles the lazy dereferencing of a queryset by lazily dereferencing all
|
Handles the lazy dereferencing of a queryset by lazily dereferencing all
|
||||||
items in a list / dict rather than one at a time.
|
items in a list / dict rather than one at a time.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, field=None, **kwargs):
|
field = None
|
||||||
self.field = field
|
|
||||||
super().__init__(**kwargs)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _lazy_load_refs(instance, name, ref_values, *, max_depth):
|
|
||||||
_dereference = _import_class("DeReference")()
|
|
||||||
documents = _dereference(
|
|
||||||
ref_values,
|
|
||||||
max_depth=max_depth,
|
|
||||||
instance=instance,
|
|
||||||
name=name,
|
|
||||||
)
|
|
||||||
return documents
|
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
"""Descriptor to automatically dereference references."""
|
"""Descriptor to automatically dereference references."""
|
||||||
@@ -299,15 +288,19 @@ class ComplexBaseField(BaseField):
|
|||||||
or isinstance(self.field, (GenericReferenceField, ReferenceField))
|
or isinstance(self.field, (GenericReferenceField, ReferenceField))
|
||||||
)
|
)
|
||||||
|
|
||||||
|
_dereference = _import_class("DeReference")()
|
||||||
|
|
||||||
if (
|
if (
|
||||||
instance._initialised
|
instance._initialised
|
||||||
and dereference
|
and dereference
|
||||||
and instance._data.get(self.name)
|
and instance._data.get(self.name)
|
||||||
and not getattr(instance._data[self.name], "_dereferenced", False)
|
and not getattr(instance._data[self.name], "_dereferenced", False)
|
||||||
):
|
):
|
||||||
ref_values = instance._data.get(self.name)
|
instance._data[self.name] = _dereference(
|
||||||
instance._data[self.name] = self._lazy_load_refs(
|
instance._data.get(self.name),
|
||||||
ref_values=ref_values, instance=instance, name=self.name, max_depth=1
|
max_depth=1,
|
||||||
|
instance=instance,
|
||||||
|
name=self.name,
|
||||||
)
|
)
|
||||||
if hasattr(instance._data[self.name], "_dereferenced"):
|
if hasattr(instance._data[self.name], "_dereferenced"):
|
||||||
instance._data[self.name]._dereferenced = True
|
instance._data[self.name]._dereferenced = True
|
||||||
@@ -333,9 +326,7 @@ class ComplexBaseField(BaseField):
|
|||||||
and isinstance(value, (BaseList, BaseDict))
|
and isinstance(value, (BaseList, BaseDict))
|
||||||
and not value._dereferenced
|
and not value._dereferenced
|
||||||
):
|
):
|
||||||
value = self._lazy_load_refs(
|
value = _dereference(value, max_depth=1, instance=instance, name=self.name)
|
||||||
ref_values=value, instance=instance, name=self.name, max_depth=1
|
|
||||||
)
|
|
||||||
value._dereferenced = True
|
value._dereferenced = True
|
||||||
instance._data[self.name] = value
|
instance._data[self.name] = value
|
||||||
|
|
||||||
@@ -478,7 +469,9 @@ class ComplexBaseField(BaseField):
|
|||||||
|
|
||||||
if errors:
|
if errors:
|
||||||
field_class = self.field.__class__.__name__
|
field_class = self.field.__class__.__name__
|
||||||
self.error(f"Invalid {field_class} item ({value})", errors=errors)
|
self.error(
|
||||||
|
"Invalid {} item ({})".format(field_class, value), errors=errors
|
||||||
|
)
|
||||||
# Don't allow empty values if required
|
# Don't allow empty values if required
|
||||||
if self.required and not value:
|
if self.required and not value:
|
||||||
self.error("Field is required and cannot be empty")
|
self.error("Field is required and cannot be empty")
|
||||||
@@ -527,7 +520,10 @@ class ObjectIdField(BaseField):
|
|||||||
|
|
||||||
|
|
||||||
class GeoJsonBaseField(BaseField):
|
class GeoJsonBaseField(BaseField):
|
||||||
"""A geo json field storing a geojson style object."""
|
"""A geo json field storing a geojson style object.
|
||||||
|
|
||||||
|
.. versionadded:: 0.8
|
||||||
|
"""
|
||||||
|
|
||||||
_geo_index = pymongo.GEOSPHERE
|
_geo_index = pymongo.GEOSPHERE
|
||||||
_type = "GeoBase"
|
_type = "GeoBase"
|
||||||
@@ -547,7 +543,7 @@ class GeoJsonBaseField(BaseField):
|
|||||||
if isinstance(value, dict):
|
if isinstance(value, dict):
|
||||||
if set(value.keys()) == {"type", "coordinates"}:
|
if set(value.keys()) == {"type", "coordinates"}:
|
||||||
if value["type"] != self._type:
|
if value["type"] != self._type:
|
||||||
self.error(f'{self._name} type must be "{self._type}"')
|
self.error('{} type must be "{}"'.format(self._name, self._type))
|
||||||
return self.validate(value["coordinates"])
|
return self.validate(value["coordinates"])
|
||||||
else:
|
else:
|
||||||
self.error(
|
self.error(
|
||||||
|
@@ -2,11 +2,7 @@ import itertools
|
|||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
from mongoengine.base.common import _document_registry
|
from mongoengine.base.common import _document_registry
|
||||||
from mongoengine.base.fields import (
|
from mongoengine.base.fields import BaseField, ComplexBaseField, ObjectIdField
|
||||||
BaseField,
|
|
||||||
ComplexBaseField,
|
|
||||||
ObjectIdField,
|
|
||||||
)
|
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.errors import InvalidDocumentError
|
from mongoengine.errors import InvalidDocumentError
|
||||||
from mongoengine.queryset import (
|
from mongoengine.queryset import (
|
||||||
@@ -16,6 +12,7 @@ from mongoengine.queryset import (
|
|||||||
QuerySetManager,
|
QuerySetManager,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
__all__ = ("DocumentMetaclass", "TopLevelDocumentMetaclass")
|
__all__ = ("DocumentMetaclass", "TopLevelDocumentMetaclass")
|
||||||
|
|
||||||
|
|
||||||
@@ -340,7 +337,7 @@ class TopLevelDocumentMetaclass(DocumentMetaclass):
|
|||||||
# allow_inheritance to False. If the base Document allows inheritance,
|
# allow_inheritance to False. If the base Document allows inheritance,
|
||||||
# none of its subclasses can override allow_inheritance to False.
|
# none of its subclasses can override allow_inheritance to False.
|
||||||
simple_class = all(
|
simple_class = all(
|
||||||
b._meta.get("abstract") for b in flattened_bases if hasattr(b, "_meta")
|
[b._meta.get("abstract") for b in flattened_bases if hasattr(b, "_meta")]
|
||||||
)
|
)
|
||||||
if (
|
if (
|
||||||
not simple_class
|
not simple_class
|
||||||
@@ -442,8 +439,8 @@ class TopLevelDocumentMetaclass(DocumentMetaclass):
|
|||||||
|
|
||||||
id_basename, id_db_basename, i = ("auto_id", "_auto_id", 0)
|
id_basename, id_db_basename, i = ("auto_id", "_auto_id", 0)
|
||||||
for i in itertools.count():
|
for i in itertools.count():
|
||||||
id_name = f"{id_basename}_{i}"
|
id_name = "{}_{}".format(id_basename, i)
|
||||||
id_db_name = f"{id_db_basename}_{i}"
|
id_db_name = "{}_{}".format(id_db_basename, i)
|
||||||
if id_name not in existing_fields and id_db_name not in existing_db_fields:
|
if id_name not in existing_fields and id_db_name not in existing_db_fields:
|
||||||
return id_name, id_db_name
|
return id_name, id_db_name
|
||||||
|
|
||||||
|
@@ -54,7 +54,7 @@ def _get_connection_settings(
|
|||||||
password=None,
|
password=None,
|
||||||
authentication_source=None,
|
authentication_source=None,
|
||||||
authentication_mechanism=None,
|
authentication_mechanism=None,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""Get the connection settings as a dict
|
"""Get the connection settings as a dict
|
||||||
|
|
||||||
@@ -74,6 +74,8 @@ def _get_connection_settings(
|
|||||||
: param kwargs: ad-hoc parameters to be passed into the pymongo driver,
|
: param kwargs: ad-hoc parameters to be passed into the pymongo driver,
|
||||||
for example maxpoolsize, tz_aware, etc. See the documentation
|
for example maxpoolsize, tz_aware, etc. See the documentation
|
||||||
for pymongo's `MongoClient` for a full list.
|
for pymongo's `MongoClient` for a full list.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.10.6 - added mongomock support
|
||||||
"""
|
"""
|
||||||
conn_settings = {
|
conn_settings = {
|
||||||
"name": name or db or DEFAULT_DATABASE_NAME,
|
"name": name or db or DEFAULT_DATABASE_NAME,
|
||||||
@@ -177,7 +179,7 @@ def register_connection(
|
|||||||
password=None,
|
password=None,
|
||||||
authentication_source=None,
|
authentication_source=None,
|
||||||
authentication_mechanism=None,
|
authentication_mechanism=None,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""Register the connection settings.
|
"""Register the connection settings.
|
||||||
|
|
||||||
@@ -199,6 +201,8 @@ def register_connection(
|
|||||||
: param kwargs: ad-hoc parameters to be passed into the pymongo driver,
|
: param kwargs: ad-hoc parameters to be passed into the pymongo driver,
|
||||||
for example maxpoolsize, tz_aware, etc. See the documentation
|
for example maxpoolsize, tz_aware, etc. See the documentation
|
||||||
for pymongo's `MongoClient` for a full list.
|
for pymongo's `MongoClient` for a full list.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.10.6 - added mongomock support
|
||||||
"""
|
"""
|
||||||
conn_settings = _get_connection_settings(
|
conn_settings = _get_connection_settings(
|
||||||
db=db,
|
db=db,
|
||||||
@@ -210,15 +214,15 @@ def register_connection(
|
|||||||
password=password,
|
password=password,
|
||||||
authentication_source=authentication_source,
|
authentication_source=authentication_source,
|
||||||
authentication_mechanism=authentication_mechanism,
|
authentication_mechanism=authentication_mechanism,
|
||||||
**kwargs,
|
**kwargs
|
||||||
)
|
)
|
||||||
_connection_settings[alias] = conn_settings
|
_connection_settings[alias] = conn_settings
|
||||||
|
|
||||||
|
|
||||||
def disconnect(alias=DEFAULT_CONNECTION_NAME):
|
def disconnect(alias=DEFAULT_CONNECTION_NAME):
|
||||||
"""Close the connection with a given alias."""
|
"""Close the connection with a given alias."""
|
||||||
from mongoengine import Document
|
|
||||||
from mongoengine.base.common import _get_documents_by_db
|
from mongoengine.base.common import _get_documents_by_db
|
||||||
|
from mongoengine import Document
|
||||||
|
|
||||||
if alias in _connections:
|
if alias in _connections:
|
||||||
get_connection(alias=alias).close()
|
get_connection(alias=alias).close()
|
||||||
@@ -313,7 +317,7 @@ def _create_connection(alias, connection_class, **connection_settings):
|
|||||||
try:
|
try:
|
||||||
return connection_class(**connection_settings)
|
return connection_class(**connection_settings)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise ConnectionFailure(f"Cannot connect to database {alias} :\n{e}")
|
raise ConnectionFailure("Cannot connect to database {} :\n{}".format(alias, e))
|
||||||
|
|
||||||
|
|
||||||
def _find_existing_connection(connection_settings):
|
def _find_existing_connection(connection_settings):
|
||||||
@@ -382,6 +386,8 @@ def connect(db=None, alias=DEFAULT_CONNECTION_NAME, **kwargs):
|
|||||||
|
|
||||||
See the docstring for `register_connection` for more details about all
|
See the docstring for `register_connection` for more details about all
|
||||||
supported kwargs.
|
supported kwargs.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.6 - added multiple database support.
|
||||||
"""
|
"""
|
||||||
if alias in _connections:
|
if alias in _connections:
|
||||||
prev_conn_setting = _connection_settings[alias]
|
prev_conn_setting = _connection_settings[alias]
|
||||||
|
@@ -184,7 +184,8 @@ class query_counter:
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, alias=DEFAULT_CONNECTION_NAME):
|
def __init__(self, alias=DEFAULT_CONNECTION_NAME):
|
||||||
"""Construct the query_counter"""
|
"""Construct the query_counter
|
||||||
|
"""
|
||||||
self.db = get_db(alias=alias)
|
self.db = get_db(alias=alias)
|
||||||
self.initial_profiling_level = None
|
self.initial_profiling_level = None
|
||||||
self._ctx_query_counter = 0 # number of queries issued by the context
|
self._ctx_query_counter = 0 # number of queries issued by the context
|
||||||
|
@@ -1,4 +1,4 @@
|
|||||||
from bson import SON, DBRef
|
from bson import DBRef, SON
|
||||||
|
|
||||||
from mongoengine.base import (
|
from mongoengine.base import (
|
||||||
BaseDict,
|
BaseDict,
|
||||||
@@ -10,12 +10,7 @@ from mongoengine.base import (
|
|||||||
from mongoengine.base.datastructures import LazyReference
|
from mongoengine.base.datastructures import LazyReference
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.document import Document, EmbeddedDocument
|
from mongoengine.document import Document, EmbeddedDocument
|
||||||
from mongoengine.fields import (
|
from mongoengine.fields import DictField, ListField, MapField, ReferenceField
|
||||||
DictField,
|
|
||||||
ListField,
|
|
||||||
MapField,
|
|
||||||
ReferenceField,
|
|
||||||
)
|
|
||||||
from mongoengine.queryset import QuerySet
|
from mongoengine.queryset import QuerySet
|
||||||
|
|
||||||
|
|
||||||
@@ -56,10 +51,10 @@ class DeReference:
|
|||||||
doc_type = doc_type.document_type
|
doc_type = doc_type.document_type
|
||||||
is_list = not hasattr(items, "items")
|
is_list = not hasattr(items, "items")
|
||||||
|
|
||||||
if is_list and all(i.__class__ == doc_type for i in items):
|
if is_list and all([i.__class__ == doc_type for i in items]):
|
||||||
return items
|
return items
|
||||||
elif not is_list and all(
|
elif not is_list and all(
|
||||||
i.__class__ == doc_type for i in items.values()
|
[i.__class__ == doc_type for i in items.values()]
|
||||||
):
|
):
|
||||||
return items
|
return items
|
||||||
elif not field.dbref:
|
elif not field.dbref:
|
||||||
@@ -162,7 +157,8 @@ class DeReference:
|
|||||||
return reference_map
|
return reference_map
|
||||||
|
|
||||||
def _fetch_objects(self, doc_type=None):
|
def _fetch_objects(self, doc_type=None):
|
||||||
"""Fetch all references and convert to their document objects"""
|
"""Fetch all references and convert to their document objects
|
||||||
|
"""
|
||||||
object_map = {}
|
object_map = {}
|
||||||
for collection, dbrefs in self.reference_map.items():
|
for collection, dbrefs in self.reference_map.items():
|
||||||
|
|
||||||
@@ -276,12 +272,12 @@ class DeReference:
|
|||||||
(v["_ref"].collection, v["_ref"].id), v
|
(v["_ref"].collection, v["_ref"].id), v
|
||||||
)
|
)
|
||||||
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
||||||
item_name = f"{name}.{k}.{field_name}"
|
item_name = "{}.{}.{}".format(name, k, field_name)
|
||||||
data[k]._data[field_name] = self._attach_objects(
|
data[k]._data[field_name] = self._attach_objects(
|
||||||
v, depth, instance=instance, name=item_name
|
v, depth, instance=instance, name=item_name
|
||||||
)
|
)
|
||||||
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
elif isinstance(v, (dict, list, tuple)) and depth <= self.max_depth:
|
||||||
item_name = f"{name}.{k}" if name else name
|
item_name = "{}.{}".format(name, k) if name else name
|
||||||
data[k] = self._attach_objects(
|
data[k] = self._attach_objects(
|
||||||
v, depth - 1, instance=instance, name=item_name
|
v, depth - 1, instance=instance, name=item_name
|
||||||
)
|
)
|
||||||
|
@@ -1,7 +1,8 @@
|
|||||||
import re
|
import re
|
||||||
|
import warnings
|
||||||
|
|
||||||
import pymongo
|
|
||||||
from bson.dbref import DBRef
|
from bson.dbref import DBRef
|
||||||
|
import pymongo
|
||||||
from pymongo.read_preferences import ReadPreference
|
from pymongo.read_preferences import ReadPreference
|
||||||
|
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
@@ -16,23 +17,14 @@ from mongoengine.base import (
|
|||||||
)
|
)
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
||||||
from mongoengine.context_managers import (
|
from mongoengine.context_managers import set_write_concern, switch_collection, switch_db
|
||||||
set_write_concern,
|
|
||||||
switch_collection,
|
|
||||||
switch_db,
|
|
||||||
)
|
|
||||||
from mongoengine.errors import (
|
from mongoengine.errors import (
|
||||||
InvalidDocumentError,
|
InvalidDocumentError,
|
||||||
InvalidQueryError,
|
InvalidQueryError,
|
||||||
SaveConditionError,
|
SaveConditionError,
|
||||||
)
|
)
|
||||||
from mongoengine.pymongo_support import list_collection_names
|
from mongoengine.pymongo_support import list_collection_names
|
||||||
from mongoengine.queryset import (
|
from mongoengine.queryset import NotUniqueError, OperationError, QuerySet, transform
|
||||||
NotUniqueError,
|
|
||||||
OperationError,
|
|
||||||
QuerySet,
|
|
||||||
transform,
|
|
||||||
)
|
|
||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
"Document",
|
"Document",
|
||||||
@@ -118,7 +110,7 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
|
|
||||||
By default, the MongoDB collection used to store documents created using a
|
By default, the MongoDB collection used to store documents created using a
|
||||||
:class:`~mongoengine.Document` subclass will be the name of the subclass
|
:class:`~mongoengine.Document` subclass will be the name of the subclass
|
||||||
converted to snake_case. A different collection may be specified by
|
converted to lowercase. A different collection may be specified by
|
||||||
providing :attr:`collection` to the :attr:`meta` dictionary in the class
|
providing :attr:`collection` to the :attr:`meta` dictionary in the class
|
||||||
definition.
|
definition.
|
||||||
|
|
||||||
@@ -332,7 +324,7 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
_refs=None,
|
_refs=None,
|
||||||
save_condition=None,
|
save_condition=None,
|
||||||
signal_kwargs=None,
|
signal_kwargs=None,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""Save the :class:`~mongoengine.Document` to the database. If the
|
"""Save the :class:`~mongoengine.Document` to the database. If the
|
||||||
document already exists, it will be updated, otherwise it will be
|
document already exists, it will be updated, otherwise it will be
|
||||||
@@ -375,6 +367,15 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
meta['cascade'] = True. Also you can pass different kwargs to
|
meta['cascade'] = True. Also you can pass different kwargs to
|
||||||
the cascade save using cascade_kwargs which overwrites the
|
the cascade save using cascade_kwargs which overwrites the
|
||||||
existing kwargs with custom values.
|
existing kwargs with custom values.
|
||||||
|
.. versionchanged:: 0.8.5
|
||||||
|
Optional save_condition that only overwrites existing documents
|
||||||
|
if the condition is satisfied in the current db record.
|
||||||
|
.. versionchanged:: 0.10
|
||||||
|
:class:`OperationError` exception raised if save_condition fails.
|
||||||
|
.. versionchanged:: 0.10.1
|
||||||
|
:class: save_condition failure now raises a `SaveConditionError`
|
||||||
|
.. versionchanged:: 0.10.7
|
||||||
|
Add signal_kwargs argument
|
||||||
"""
|
"""
|
||||||
signal_kwargs = signal_kwargs or {}
|
signal_kwargs = signal_kwargs or {}
|
||||||
|
|
||||||
@@ -563,7 +564,7 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
if not getattr(ref, "_changed_fields", True):
|
if not getattr(ref, "_changed_fields", True):
|
||||||
continue
|
continue
|
||||||
|
|
||||||
ref_id = f"{ref.__class__.__name__},{str(ref._data)}"
|
ref_id = "{},{}".format(ref.__class__.__name__, str(ref._data))
|
||||||
if ref and ref_id not in _refs:
|
if ref and ref_id not in _refs:
|
||||||
_refs.append(ref_id)
|
_refs.append(ref_id)
|
||||||
kwargs["_refs"] = _refs
|
kwargs["_refs"] = _refs
|
||||||
@@ -629,6 +630,9 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
For example, ``save(..., w: 2, fsync: True)`` will
|
For example, ``save(..., w: 2, fsync: True)`` will
|
||||||
wait until at least two servers have recorded the write and
|
wait until at least two servers have recorded the write and
|
||||||
will force an fsync on the primary server.
|
will force an fsync on the primary server.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.10.7
|
||||||
|
Add signal_kwargs argument
|
||||||
"""
|
"""
|
||||||
signal_kwargs = signal_kwargs or {}
|
signal_kwargs = signal_kwargs or {}
|
||||||
signals.pre_delete.send(self.__class__, document=self, **signal_kwargs)
|
signals.pre_delete.send(self.__class__, document=self, **signal_kwargs)
|
||||||
@@ -710,6 +714,8 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
def select_related(self, max_depth=1):
|
def select_related(self, max_depth=1):
|
||||||
"""Handles dereferencing of :class:`~bson.dbref.DBRef` objects to
|
"""Handles dereferencing of :class:`~bson.dbref.DBRef` objects to
|
||||||
a maximum depth in order to cut down the number queries to mongodb.
|
a maximum depth in order to cut down the number queries to mongodb.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
DeReference = _import_class("DeReference")
|
DeReference = _import_class("DeReference")
|
||||||
DeReference()([self], max_depth + 1)
|
DeReference()([self], max_depth + 1)
|
||||||
@@ -720,6 +726,10 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
|
|
||||||
:param fields: (optional) args list of fields to reload
|
:param fields: (optional) args list of fields to reload
|
||||||
:param max_depth: (optional) depth of dereferencing to follow
|
:param max_depth: (optional) depth of dereferencing to follow
|
||||||
|
|
||||||
|
.. versionadded:: 0.1.2
|
||||||
|
.. versionchanged:: 0.6 Now chainable
|
||||||
|
.. versionchanged:: 0.9 Can provide specific fields to reload
|
||||||
"""
|
"""
|
||||||
max_depth = 1
|
max_depth = 1
|
||||||
if fields and isinstance(fields[0], int):
|
if fields and isinstance(fields[0], int):
|
||||||
@@ -821,6 +831,9 @@ class Document(BaseDocument, metaclass=TopLevelDocumentMetaclass):
|
|||||||
|
|
||||||
Raises :class:`OperationError` if the document has no collection set
|
Raises :class:`OperationError` if the document has no collection set
|
||||||
(i.g. if it is `abstract`)
|
(i.g. if it is `abstract`)
|
||||||
|
|
||||||
|
.. versionchanged:: 0.10.7
|
||||||
|
:class:`OperationError` exception raised if no collection available
|
||||||
"""
|
"""
|
||||||
coll_name = cls._get_collection_name()
|
coll_name = cls._get_collection_name()
|
||||||
if not coll_name:
|
if not coll_name:
|
||||||
@@ -1075,6 +1088,8 @@ class MapReduceDocument:
|
|||||||
an ``ObjectId`` found in the given ``collection``,
|
an ``ObjectId`` found in the given ``collection``,
|
||||||
the object can be accessed via the ``object`` property.
|
the object can be accessed via the ``object`` property.
|
||||||
:param value: The result(s) for this key.
|
:param value: The result(s) for this key.
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, document, collection, key, value):
|
def __init__(self, document, collection, key, value):
|
||||||
|
@@ -1,5 +1,6 @@
|
|||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
|
|
||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
"NotRegistered",
|
"NotRegistered",
|
||||||
"InvalidDocumentError",
|
"InvalidDocumentError",
|
||||||
@@ -93,7 +94,7 @@ class ValidationError(AssertionError):
|
|||||||
return str(self.message)
|
return str(self.message)
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return f"{self.__class__.__name__}({self.message},)"
|
return "{}({},)".format(self.__class__.__name__, self.message)
|
||||||
|
|
||||||
def __getattribute__(self, name):
|
def __getattribute__(self, name):
|
||||||
message = super().__getattribute__(name)
|
message = super().__getattribute__(name)
|
||||||
@@ -101,7 +102,7 @@ class ValidationError(AssertionError):
|
|||||||
if self.field_name:
|
if self.field_name:
|
||||||
message = "%s" % message
|
message = "%s" % message
|
||||||
if self.errors:
|
if self.errors:
|
||||||
message = f"{message}({self._format_errors()})"
|
message = "{}({})".format(message, self._format_errors())
|
||||||
return message
|
return message
|
||||||
|
|
||||||
def _get_message(self):
|
def _get_message(self):
|
||||||
@@ -146,13 +147,13 @@ class ValidationError(AssertionError):
|
|||||||
elif isinstance(value, dict):
|
elif isinstance(value, dict):
|
||||||
value = " ".join([generate_key(v, k) for k, v in value.items()])
|
value = " ".join([generate_key(v, k) for k, v in value.items()])
|
||||||
|
|
||||||
results = f"{prefix}.{value}" if prefix else value
|
results = "{}.{}".format(prefix, value) if prefix else value
|
||||||
return results
|
return results
|
||||||
|
|
||||||
error_dict = defaultdict(list)
|
error_dict = defaultdict(list)
|
||||||
for k, v in self.to_dict().items():
|
for k, v in self.to_dict().items():
|
||||||
error_dict[generate_key(v)].append(k)
|
error_dict[generate_key(v)].append(k)
|
||||||
return " ".join([f"{k}: {v}" for k, v in error_dict.items()])
|
return " ".join(["{}: {}".format(k, v) for k, v in error_dict.items()])
|
||||||
|
|
||||||
|
|
||||||
class DeprecatedError(Exception):
|
class DeprecatedError(Exception):
|
||||||
|
@@ -1,6 +1,5 @@
|
|||||||
import datetime
|
import datetime
|
||||||
import decimal
|
import decimal
|
||||||
import inspect
|
|
||||||
import itertools
|
import itertools
|
||||||
import re
|
import re
|
||||||
import socket
|
import socket
|
||||||
@@ -9,10 +8,10 @@ import uuid
|
|||||||
from io import BytesIO
|
from io import BytesIO
|
||||||
from operator import itemgetter
|
from operator import itemgetter
|
||||||
|
|
||||||
|
from bson import Binary, DBRef, ObjectId, SON
|
||||||
|
from bson.int64 import Int64
|
||||||
import gridfs
|
import gridfs
|
||||||
import pymongo
|
import pymongo
|
||||||
from bson import SON, Binary, DBRef, ObjectId
|
|
||||||
from bson.int64 import Int64
|
|
||||||
from pymongo import ReturnDocument
|
from pymongo import ReturnDocument
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@@ -22,6 +21,7 @@ except ImportError:
|
|||||||
else:
|
else:
|
||||||
import dateutil.parser
|
import dateutil.parser
|
||||||
|
|
||||||
|
|
||||||
from mongoengine.base import (
|
from mongoengine.base import (
|
||||||
BaseDocument,
|
BaseDocument,
|
||||||
BaseField,
|
BaseField,
|
||||||
@@ -35,11 +35,8 @@ from mongoengine.base.utils import LazyRegexCompiler
|
|||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
from mongoengine.connection import DEFAULT_CONNECTION_NAME, get_db
|
||||||
from mongoengine.document import Document, EmbeddedDocument
|
from mongoengine.document import Document, EmbeddedDocument
|
||||||
from mongoengine.errors import (
|
from mongoengine.errors import DoesNotExist, InvalidQueryError, ValidationError
|
||||||
DoesNotExist,
|
from mongoengine.mongodb_support import MONGODB_36, get_mongodb_version
|
||||||
InvalidQueryError,
|
|
||||||
ValidationError,
|
|
||||||
)
|
|
||||||
from mongoengine.queryset import DO_NOTHING
|
from mongoengine.queryset import DO_NOTHING
|
||||||
from mongoengine.queryset.base import BaseQuerySet
|
from mongoengine.queryset.base import BaseQuerySet
|
||||||
from mongoengine.queryset.transform import STRING_OPERATORS
|
from mongoengine.queryset.transform import STRING_OPERATORS
|
||||||
@@ -104,12 +101,6 @@ class StringField(BaseField):
|
|||||||
"""A unicode string field."""
|
"""A unicode string field."""
|
||||||
|
|
||||||
def __init__(self, regex=None, max_length=None, min_length=None, **kwargs):
|
def __init__(self, regex=None, max_length=None, min_length=None, **kwargs):
|
||||||
"""
|
|
||||||
:param regex: (optional) A string pattern that will be applied during validation
|
|
||||||
:param max_length: (optional) A max length that will be applied during validation
|
|
||||||
:param min_length: (optional) A min length that will be applied during validation
|
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
"""
|
|
||||||
self.regex = re.compile(regex) if regex else None
|
self.regex = re.compile(regex) if regex else None
|
||||||
self.max_length = max_length
|
self.max_length = max_length
|
||||||
self.min_length = min_length
|
self.min_length = min_length
|
||||||
@@ -165,7 +156,10 @@ class StringField(BaseField):
|
|||||||
|
|
||||||
|
|
||||||
class URLField(StringField):
|
class URLField(StringField):
|
||||||
"""A field that validates input as an URL."""
|
"""A field that validates input as an URL.
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
|
"""
|
||||||
|
|
||||||
_URL_REGEX = LazyRegexCompiler(
|
_URL_REGEX = LazyRegexCompiler(
|
||||||
r"^(?:[a-z0-9\.\-]*)://" # scheme is validated separately
|
r"^(?:[a-z0-9\.\-]*)://" # scheme is validated separately
|
||||||
@@ -180,11 +174,6 @@ class URLField(StringField):
|
|||||||
_URL_SCHEMES = ["http", "https", "ftp", "ftps"]
|
_URL_SCHEMES = ["http", "https", "ftp", "ftps"]
|
||||||
|
|
||||||
def __init__(self, url_regex=None, schemes=None, **kwargs):
|
def __init__(self, url_regex=None, schemes=None, **kwargs):
|
||||||
"""
|
|
||||||
:param url_regex: (optional) Overwrite the default regex used for validation
|
|
||||||
:param schemes: (optional) Overwrite the default URL schemes that are allowed
|
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.StringField`
|
|
||||||
"""
|
|
||||||
self.url_regex = url_regex or self._URL_REGEX
|
self.url_regex = url_regex or self._URL_REGEX
|
||||||
self.schemes = schemes or self._URL_SCHEMES
|
self.schemes = schemes or self._URL_SCHEMES
|
||||||
super().__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
@@ -193,15 +182,18 @@ class URLField(StringField):
|
|||||||
# Check first if the scheme is valid
|
# Check first if the scheme is valid
|
||||||
scheme = value.split("://")[0].lower()
|
scheme = value.split("://")[0].lower()
|
||||||
if scheme not in self.schemes:
|
if scheme not in self.schemes:
|
||||||
self.error(f"Invalid scheme {scheme} in URL: {value}")
|
self.error("Invalid scheme {} in URL: {}".format(scheme, value))
|
||||||
|
|
||||||
# Then check full URL
|
# Then check full URL
|
||||||
if not self.url_regex.match(value):
|
if not self.url_regex.match(value):
|
||||||
self.error(f"Invalid URL: {value}")
|
self.error("Invalid URL: {}".format(value))
|
||||||
|
|
||||||
|
|
||||||
class EmailField(StringField):
|
class EmailField(StringField):
|
||||||
"""A field that validates input as an email address."""
|
"""A field that validates input as an email address.
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
"""
|
||||||
|
|
||||||
USER_REGEX = LazyRegexCompiler(
|
USER_REGEX = LazyRegexCompiler(
|
||||||
# `dot-atom` defined in RFC 5322 Section 3.2.3.
|
# `dot-atom` defined in RFC 5322 Section 3.2.3.
|
||||||
@@ -235,13 +227,18 @@ class EmailField(StringField):
|
|||||||
allow_utf8_user=False,
|
allow_utf8_user=False,
|
||||||
allow_ip_domain=False,
|
allow_ip_domain=False,
|
||||||
*args,
|
*args,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""
|
"""Initialize the EmailField.
|
||||||
:param domain_whitelist: (optional) list of valid domain names applied during validation
|
|
||||||
:param allow_utf8_user: Allow user part of the email to contain utf8 char
|
Args:
|
||||||
:param allow_ip_domain: Allow domain part of the email to be an IPv4 or IPv6 address
|
domain_whitelist (list) - list of otherwise invalid domain
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.StringField`
|
names which you'd like to support.
|
||||||
|
allow_utf8_user (bool) - if True, the user part of the email
|
||||||
|
address can contain UTF8 characters.
|
||||||
|
False by default.
|
||||||
|
allow_ip_domain (bool) - if True, the domain part of the email
|
||||||
|
can be a valid IPv4 or IPv6 address.
|
||||||
"""
|
"""
|
||||||
self.domain_whitelist = domain_whitelist or []
|
self.domain_whitelist = domain_whitelist or []
|
||||||
self.allow_utf8_user = allow_utf8_user
|
self.allow_utf8_user = allow_utf8_user
|
||||||
@@ -313,11 +310,6 @@ class IntField(BaseField):
|
|||||||
"""32-bit integer field."""
|
"""32-bit integer field."""
|
||||||
|
|
||||||
def __init__(self, min_value=None, max_value=None, **kwargs):
|
def __init__(self, min_value=None, max_value=None, **kwargs):
|
||||||
"""
|
|
||||||
:param min_value: (optional) A min value that will be applied during validation
|
|
||||||
:param max_value: (optional) A max value that will be applied during validation
|
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
"""
|
|
||||||
self.min_value, self.max_value = min_value, max_value
|
self.min_value, self.max_value = min_value, max_value
|
||||||
super().__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
@@ -351,11 +343,6 @@ class LongField(BaseField):
|
|||||||
"""64-bit integer field. (Equivalent to IntField since the support to Python2 was dropped)"""
|
"""64-bit integer field. (Equivalent to IntField since the support to Python2 was dropped)"""
|
||||||
|
|
||||||
def __init__(self, min_value=None, max_value=None, **kwargs):
|
def __init__(self, min_value=None, max_value=None, **kwargs):
|
||||||
"""
|
|
||||||
:param min_value: (optional) A min value that will be applied during validation
|
|
||||||
:param max_value: (optional) A max value that will be applied during validation
|
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
"""
|
|
||||||
self.min_value, self.max_value = min_value, max_value
|
self.min_value, self.max_value = min_value, max_value
|
||||||
super().__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
@@ -392,11 +379,6 @@ class FloatField(BaseField):
|
|||||||
"""Floating point number field."""
|
"""Floating point number field."""
|
||||||
|
|
||||||
def __init__(self, min_value=None, max_value=None, **kwargs):
|
def __init__(self, min_value=None, max_value=None, **kwargs):
|
||||||
"""
|
|
||||||
:param min_value: (optional) A min value that will be applied during validation
|
|
||||||
:param max_value: (optional) A max value that will be applied during validation
|
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
"""
|
|
||||||
self.min_value, self.max_value = min_value, max_value
|
self.min_value, self.max_value = min_value, max_value
|
||||||
super().__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
@@ -433,6 +415,9 @@ class FloatField(BaseField):
|
|||||||
class DecimalField(BaseField):
|
class DecimalField(BaseField):
|
||||||
"""Fixed-point decimal number field. Stores the value as a float by default unless `force_string` is used.
|
"""Fixed-point decimal number field. Stores the value as a float by default unless `force_string` is used.
|
||||||
If using floats, beware of Decimal to float conversion (potential precision loss)
|
If using floats, beware of Decimal to float conversion (potential precision loss)
|
||||||
|
|
||||||
|
.. versionchanged:: 0.8
|
||||||
|
.. versionadded:: 0.3
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
@@ -442,11 +427,11 @@ class DecimalField(BaseField):
|
|||||||
force_string=False,
|
force_string=False,
|
||||||
precision=2,
|
precision=2,
|
||||||
rounding=decimal.ROUND_HALF_UP,
|
rounding=decimal.ROUND_HALF_UP,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
:param min_value: (optional) A min value that will be applied during validation
|
:param min_value: Validation rule for the minimum acceptable value.
|
||||||
:param max_value: (optional) A max value that will be applied during validation
|
:param max_value: Validation rule for the maximum acceptable value.
|
||||||
:param force_string: Store the value as a string (instead of a float).
|
:param force_string: Store the value as a string (instead of a float).
|
||||||
Be aware that this affects query sorting and operation like lte, gte (as string comparison is applied)
|
Be aware that this affects query sorting and operation like lte, gte (as string comparison is applied)
|
||||||
and some query operator won't work (e.g. inc, dec)
|
and some query operator won't work (e.g. inc, dec)
|
||||||
@@ -463,7 +448,7 @@ class DecimalField(BaseField):
|
|||||||
- decimal.ROUND_05UP (away from zero if last digit after rounding towards zero would have been 0 or 5; otherwise towards zero)
|
- decimal.ROUND_05UP (away from zero if last digit after rounding towards zero would have been 0 or 5; otherwise towards zero)
|
||||||
|
|
||||||
Defaults to: ``decimal.ROUND_HALF_UP``
|
Defaults to: ``decimal.ROUND_HALF_UP``
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
"""
|
"""
|
||||||
self.min_value = min_value
|
self.min_value = min_value
|
||||||
self.max_value = max_value
|
self.max_value = max_value
|
||||||
@@ -513,12 +498,15 @@ class DecimalField(BaseField):
|
|||||||
|
|
||||||
|
|
||||||
class BooleanField(BaseField):
|
class BooleanField(BaseField):
|
||||||
"""Boolean field type."""
|
"""Boolean field type.
|
||||||
|
|
||||||
|
.. versionadded:: 0.1.2
|
||||||
|
"""
|
||||||
|
|
||||||
def to_python(self, value):
|
def to_python(self, value):
|
||||||
try:
|
try:
|
||||||
value = bool(value)
|
value = bool(value)
|
||||||
except (ValueError, TypeError):
|
except ValueError:
|
||||||
pass
|
pass
|
||||||
return value
|
return value
|
||||||
|
|
||||||
@@ -558,13 +546,12 @@ class DateTimeField(BaseField):
|
|||||||
if callable(value):
|
if callable(value):
|
||||||
return value()
|
return value()
|
||||||
|
|
||||||
if isinstance(value, str):
|
if not isinstance(value, str):
|
||||||
return self._parse_datetime(value)
|
|
||||||
else:
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
@staticmethod
|
return self._parse_datetime(value)
|
||||||
def _parse_datetime(value):
|
|
||||||
|
def _parse_datetime(self, value):
|
||||||
# Attempt to parse a datetime from a string
|
# Attempt to parse a datetime from a string
|
||||||
value = value.strip()
|
value = value.strip()
|
||||||
if not value:
|
if not value:
|
||||||
@@ -640,12 +627,13 @@ class ComplexDateTimeField(StringField):
|
|||||||
keyword when initializing the field.
|
keyword when initializing the field.
|
||||||
|
|
||||||
Note: To default the field to the current datetime, use: DateTimeField(default=datetime.utcnow)
|
Note: To default the field to the current datetime, use: DateTimeField(default=datetime.utcnow)
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, separator=",", **kwargs):
|
def __init__(self, separator=",", **kwargs):
|
||||||
"""
|
"""
|
||||||
:param separator: Allows to customize the separator used for storage (default ``,``)
|
:param separator: Allows to customize the separator used for storage (default ``,``)
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.StringField`
|
|
||||||
"""
|
"""
|
||||||
self.separator = separator
|
self.separator = separator
|
||||||
self.format = separator.join(["%Y", "%m", "%d", "%H", "%M", "%S", "%f"])
|
self.format = separator.join(["%Y", "%m", "%d", "%H", "%M", "%S", "%f"])
|
||||||
@@ -918,16 +906,17 @@ class ListField(ComplexBaseField):
|
|||||||
"""A list field that wraps a standard field, allowing multiple instances
|
"""A list field that wraps a standard field, allowing multiple instances
|
||||||
of the field to be used as a list in the database.
|
of the field to be used as a list in the database.
|
||||||
|
|
||||||
If using with ReferenceFields see: :ref:`many-to-many-with-listfields`
|
If using with ReferenceFields see: :ref:`one-to-many-with-listfields`
|
||||||
|
|
||||||
.. note::
|
.. note::
|
||||||
Required means it cannot be empty - as the default for ListFields is []
|
Required means it cannot be empty - as the default for ListFields is []
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, field=None, max_length=None, **kwargs):
|
def __init__(self, field=None, max_length=None, **kwargs):
|
||||||
|
self.field = field
|
||||||
self.max_length = max_length
|
self.max_length = max_length
|
||||||
kwargs.setdefault("default", lambda: [])
|
kwargs.setdefault("default", lambda: [])
|
||||||
super().__init__(field=field, **kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
if instance is None:
|
if instance is None:
|
||||||
@@ -986,13 +975,16 @@ class EmbeddedDocumentListField(ListField):
|
|||||||
.. note::
|
.. note::
|
||||||
The only valid list values are subclasses of
|
The only valid list values are subclasses of
|
||||||
:class:`~mongoengine.EmbeddedDocument`.
|
:class:`~mongoengine.EmbeddedDocument`.
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, document_type, **kwargs):
|
def __init__(self, document_type, **kwargs):
|
||||||
"""
|
"""
|
||||||
:param document_type: The type of
|
:param document_type: The type of
|
||||||
:class:`~mongoengine.EmbeddedDocument` the list will hold.
|
:class:`~mongoengine.EmbeddedDocument` the list will hold.
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.ListField`
|
:param kwargs: Keyword arguments passed directly into the parent
|
||||||
|
:class:`~mongoengine.ListField`.
|
||||||
"""
|
"""
|
||||||
super().__init__(field=EmbeddedDocumentField(document_type), **kwargs)
|
super().__init__(field=EmbeddedDocumentField(document_type), **kwargs)
|
||||||
|
|
||||||
@@ -1007,11 +999,19 @@ class SortedListField(ListField):
|
|||||||
save the whole list then other processes trying to save the whole list
|
save the whole list then other processes trying to save the whole list
|
||||||
as well could overwrite changes. The safest way to append to a list is
|
as well could overwrite changes. The safest way to append to a list is
|
||||||
to perform a push operation.
|
to perform a push operation.
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
.. versionchanged:: 0.6 - added reverse keyword
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
_ordering = None
|
||||||
|
_order_reverse = False
|
||||||
|
|
||||||
def __init__(self, field, **kwargs):
|
def __init__(self, field, **kwargs):
|
||||||
self._ordering = kwargs.pop("ordering", None)
|
if "ordering" in kwargs.keys():
|
||||||
self._order_reverse = kwargs.pop("reverse", False)
|
self._ordering = kwargs.pop("ordering")
|
||||||
|
if "reverse" in kwargs.keys():
|
||||||
|
self._order_reverse = kwargs.pop("reverse")
|
||||||
super().__init__(field, **kwargs)
|
super().__init__(field, **kwargs)
|
||||||
|
|
||||||
def to_mongo(self, value, use_db_field=True, fields=None):
|
def to_mongo(self, value, use_db_field=True, fields=None):
|
||||||
@@ -1032,6 +1032,17 @@ def key_not_string(d):
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
def key_has_dot_or_dollar(d):
|
||||||
|
"""Helper function to recursively determine if any key in a
|
||||||
|
dictionary contains a dot or a dollar sign.
|
||||||
|
"""
|
||||||
|
for k, v in d.items():
|
||||||
|
if ("." in k or k.startswith("$")) or (
|
||||||
|
isinstance(v, dict) and key_has_dot_or_dollar(v)
|
||||||
|
):
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
def key_starts_with_dollar(d):
|
def key_starts_with_dollar(d):
|
||||||
"""Helper function to recursively determine if any key in a
|
"""Helper function to recursively determine if any key in a
|
||||||
dictionary starts with a dollar
|
dictionary starts with a dollar
|
||||||
@@ -1047,13 +1058,17 @@ class DictField(ComplexBaseField):
|
|||||||
|
|
||||||
.. note::
|
.. note::
|
||||||
Required means it cannot be empty - as the default for DictFields is {}
|
Required means it cannot be empty - as the default for DictFields is {}
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
|
.. versionchanged:: 0.5 - Can now handle complex / varying types of data
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, field=None, *args, **kwargs):
|
def __init__(self, field=None, *args, **kwargs):
|
||||||
|
self.field = field
|
||||||
self._auto_dereference = False
|
self._auto_dereference = False
|
||||||
|
|
||||||
kwargs.setdefault("default", lambda: {})
|
kwargs.setdefault("default", lambda: {})
|
||||||
super().__init__(*args, field=field, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
"""Make sure that a list of valid fields is being used."""
|
"""Make sure that a list of valid fields is being used."""
|
||||||
@@ -1109,6 +1124,8 @@ class MapField(DictField):
|
|||||||
"""A field that maps a name to a specified field type. Similar to
|
"""A field that maps a name to a specified field type. Similar to
|
||||||
a DictField, except the 'value' of each item must match the specified
|
a DictField, except the 'value' of each item must match the specified
|
||||||
field type.
|
field type.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, field=None, *args, **kwargs):
|
def __init__(self, field=None, *args, **kwargs):
|
||||||
@@ -1156,6 +1173,8 @@ class ReferenceField(BaseField):
|
|||||||
org = ReferenceField('Org', reverse_delete_rule=CASCADE)
|
org = ReferenceField('Org', reverse_delete_rule=CASCADE)
|
||||||
|
|
||||||
User.register_delete_rule(Org, 'owner', DENY)
|
User.register_delete_rule(Org, 'owner', DENY)
|
||||||
|
|
||||||
|
.. versionchanged:: 0.5 added `reverse_delete_rule`
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
@@ -1163,12 +1182,10 @@ class ReferenceField(BaseField):
|
|||||||
):
|
):
|
||||||
"""Initialises the Reference Field.
|
"""Initialises the Reference Field.
|
||||||
|
|
||||||
:param document_type: The type of Document that will be referenced
|
|
||||||
:param dbref: Store the reference as :class:`~pymongo.dbref.DBRef`
|
:param dbref: Store the reference as :class:`~pymongo.dbref.DBRef`
|
||||||
or as the :class:`~pymongo.objectid.ObjectId`.
|
or as the :class:`~pymongo.objectid.ObjectId`.id .
|
||||||
:param reverse_delete_rule: Determines what to do when the referring
|
:param reverse_delete_rule: Determines what to do when the referring
|
||||||
object is deleted
|
object is deleted
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
|
|
||||||
.. note ::
|
.. note ::
|
||||||
A reference to an abstract document type is always stored as a
|
A reference to an abstract document type is always stored as a
|
||||||
@@ -1197,14 +1214,6 @@ class ReferenceField(BaseField):
|
|||||||
self.document_type_obj = get_document(self.document_type_obj)
|
self.document_type_obj = get_document(self.document_type_obj)
|
||||||
return self.document_type_obj
|
return self.document_type_obj
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _lazy_load_ref(ref_cls, dbref):
|
|
||||||
dereferenced_son = ref_cls._get_db().dereference(dbref)
|
|
||||||
if dereferenced_son is None:
|
|
||||||
raise DoesNotExist(f"Trying to dereference unknown document {dbref}")
|
|
||||||
|
|
||||||
return ref_cls._from_son(dereferenced_son)
|
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
"""Descriptor to allow lazy dereferencing."""
|
"""Descriptor to allow lazy dereferencing."""
|
||||||
if instance is None:
|
if instance is None:
|
||||||
@@ -1212,17 +1221,20 @@ class ReferenceField(BaseField):
|
|||||||
return self
|
return self
|
||||||
|
|
||||||
# Get value from document instance if available
|
# Get value from document instance if available
|
||||||
ref_value = instance._data.get(self.name)
|
value = instance._data.get(self.name)
|
||||||
auto_dereference = instance._fields[self.name]._auto_dereference
|
auto_dereference = instance._fields[self.name]._auto_dereference
|
||||||
# Dereference DBRefs
|
# Dereference DBRefs
|
||||||
if auto_dereference and isinstance(ref_value, DBRef):
|
if auto_dereference and isinstance(value, DBRef):
|
||||||
if hasattr(ref_value, "cls"):
|
if hasattr(value, "cls"):
|
||||||
# Dereference using the class type specified in the reference
|
# Dereference using the class type specified in the reference
|
||||||
cls = get_document(ref_value.cls)
|
cls = get_document(value.cls)
|
||||||
else:
|
else:
|
||||||
cls = self.document_type
|
cls = self.document_type
|
||||||
|
dereferenced = cls._get_db().dereference(value)
|
||||||
instance._data[self.name] = self._lazy_load_ref(cls, ref_value)
|
if dereferenced is None:
|
||||||
|
raise DoesNotExist("Trying to dereference unknown document %s" % value)
|
||||||
|
else:
|
||||||
|
instance._data[self.name] = cls._from_son(dereferenced)
|
||||||
|
|
||||||
return super().__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
@@ -1295,22 +1307,24 @@ class ReferenceField(BaseField):
|
|||||||
|
|
||||||
|
|
||||||
class CachedReferenceField(BaseField):
|
class CachedReferenceField(BaseField):
|
||||||
"""A referencefield with cache fields to purpose pseudo-joins"""
|
"""
|
||||||
|
A referencefield with cache fields to purpose pseudo-joins
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
|
"""
|
||||||
|
|
||||||
def __init__(self, document_type, fields=None, auto_sync=True, **kwargs):
|
def __init__(self, document_type, fields=None, auto_sync=True, **kwargs):
|
||||||
"""Initialises the Cached Reference Field.
|
"""Initialises the Cached Reference Field.
|
||||||
|
|
||||||
:param document_type: The type of Document that will be referenced
|
|
||||||
:param fields: A list of fields to be cached in document
|
:param fields: A list of fields to be cached in document
|
||||||
:param auto_sync: if True documents are auto updated
|
:param auto_sync: if True documents are auto updated.
|
||||||
:param kwargs: Keyword arguments passed into the parent :class:`~mongoengine.BaseField`
|
|
||||||
"""
|
"""
|
||||||
if fields is None:
|
if fields is None:
|
||||||
fields = []
|
fields = []
|
||||||
|
|
||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
if not isinstance(document_type, str) and not (
|
if not isinstance(document_type, str) and not issubclass(
|
||||||
inspect.isclass(document_type) and issubclass(document_type, Document)
|
document_type, Document
|
||||||
):
|
):
|
||||||
self.error(
|
self.error(
|
||||||
"Argument to CachedReferenceField constructor must be a"
|
"Argument to CachedReferenceField constructor must be a"
|
||||||
@@ -1332,7 +1346,7 @@ class CachedReferenceField(BaseField):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
update_kwargs = {
|
update_kwargs = {
|
||||||
f"set__{self.name}__{key}": val
|
"set__{}__{}".format(self.name, key): val
|
||||||
for key, val in document._delta()[0].items()
|
for key, val in document._delta()[0].items()
|
||||||
if key in self.fields
|
if key in self.fields
|
||||||
}
|
}
|
||||||
@@ -1361,14 +1375,6 @@ class CachedReferenceField(BaseField):
|
|||||||
self.document_type_obj = get_document(self.document_type_obj)
|
self.document_type_obj = get_document(self.document_type_obj)
|
||||||
return self.document_type_obj
|
return self.document_type_obj
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _lazy_load_ref(ref_cls, dbref):
|
|
||||||
dereferenced_son = ref_cls._get_db().dereference(dbref)
|
|
||||||
if dereferenced_son is None:
|
|
||||||
raise DoesNotExist(f"Trying to dereference unknown document {dbref}")
|
|
||||||
|
|
||||||
return ref_cls._from_son(dereferenced_son)
|
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
if instance is None:
|
if instance is None:
|
||||||
# Document class being used rather than a document object
|
# Document class being used rather than a document object
|
||||||
@@ -1380,7 +1386,11 @@ class CachedReferenceField(BaseField):
|
|||||||
|
|
||||||
# Dereference DBRefs
|
# Dereference DBRefs
|
||||||
if auto_dereference and isinstance(value, DBRef):
|
if auto_dereference and isinstance(value, DBRef):
|
||||||
instance._data[self.name] = self._lazy_load_ref(self.document_type, value)
|
dereferenced = self.document_type._get_db().dereference(value)
|
||||||
|
if dereferenced is None:
|
||||||
|
raise DoesNotExist("Trying to dereference unknown document %s" % value)
|
||||||
|
else:
|
||||||
|
instance._data[self.name] = self.document_type._from_son(dereferenced)
|
||||||
|
|
||||||
return super().__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
@@ -1475,6 +1485,8 @@ class GenericReferenceField(BaseField):
|
|||||||
it.
|
it.
|
||||||
|
|
||||||
* You can use the choices param to limit the acceptable Document types
|
* You can use the choices param to limit the acceptable Document types
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
@@ -1505,14 +1517,6 @@ class GenericReferenceField(BaseField):
|
|||||||
value = value._class_name
|
value = value._class_name
|
||||||
super()._validate_choices(value)
|
super()._validate_choices(value)
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _lazy_load_ref(ref_cls, dbref):
|
|
||||||
dereferenced_son = ref_cls._get_db().dereference(dbref)
|
|
||||||
if dereferenced_son is None:
|
|
||||||
raise DoesNotExist(f"Trying to dereference unknown document {dbref}")
|
|
||||||
|
|
||||||
return ref_cls._from_son(dereferenced_son)
|
|
||||||
|
|
||||||
def __get__(self, instance, owner):
|
def __get__(self, instance, owner):
|
||||||
if instance is None:
|
if instance is None:
|
||||||
return self
|
return self
|
||||||
@@ -1520,9 +1524,12 @@ class GenericReferenceField(BaseField):
|
|||||||
value = instance._data.get(self.name)
|
value = instance._data.get(self.name)
|
||||||
|
|
||||||
auto_dereference = instance._fields[self.name]._auto_dereference
|
auto_dereference = instance._fields[self.name]._auto_dereference
|
||||||
if auto_dereference and isinstance(value, dict):
|
if auto_dereference and isinstance(value, (dict, SON)):
|
||||||
doc_cls = get_document(value["_cls"])
|
dereferenced = self.dereference(value)
|
||||||
instance._data[self.name] = self._lazy_load_ref(doc_cls, value["_ref"])
|
if dereferenced is None:
|
||||||
|
raise DoesNotExist("Trying to dereference unknown document %s" % value)
|
||||||
|
else:
|
||||||
|
instance._data[self.name] = dereferenced
|
||||||
|
|
||||||
return super().__get__(instance, owner)
|
return super().__get__(instance, owner)
|
||||||
|
|
||||||
@@ -1541,6 +1548,14 @@ class GenericReferenceField(BaseField):
|
|||||||
" saved to the database"
|
" saved to the database"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
def dereference(self, value):
|
||||||
|
doc_cls = get_document(value["_cls"])
|
||||||
|
reference = value["_ref"]
|
||||||
|
doc = doc_cls._get_db().dereference(reference)
|
||||||
|
if doc is not None:
|
||||||
|
doc = doc_cls._from_son(doc)
|
||||||
|
return doc
|
||||||
|
|
||||||
def to_mongo(self, document):
|
def to_mongo(self, document):
|
||||||
if document is None:
|
if document is None:
|
||||||
return None
|
return None
|
||||||
@@ -1641,7 +1656,7 @@ class EnumField(BaseField):
|
|||||||
"'choices' can't be set on EnumField, "
|
"'choices' can't be set on EnumField, "
|
||||||
"it is implicitly set as the enum class"
|
"it is implicitly set as the enum class"
|
||||||
)
|
)
|
||||||
kwargs["choices"] = list(self._enum_cls) # Implicit validator
|
kwargs["choices"] = list(self._enum_cls)
|
||||||
super().__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
def __set__(self, instance, value):
|
def __set__(self, instance, value):
|
||||||
@@ -1658,6 +1673,13 @@ class EnumField(BaseField):
|
|||||||
return value.value
|
return value.value
|
||||||
return value
|
return value
|
||||||
|
|
||||||
|
def validate(self, value):
|
||||||
|
if value and not isinstance(value, self._enum_cls):
|
||||||
|
try:
|
||||||
|
self._enum_cls(value)
|
||||||
|
except Exception as e:
|
||||||
|
self.error(str(e))
|
||||||
|
|
||||||
def prepare_query_value(self, op, value):
|
def prepare_query_value(self, op, value):
|
||||||
if value is None:
|
if value is None:
|
||||||
return value
|
return value
|
||||||
@@ -1669,7 +1691,12 @@ class GridFSError(Exception):
|
|||||||
|
|
||||||
|
|
||||||
class GridFSProxy:
|
class GridFSProxy:
|
||||||
"""Proxy object to handle writing and reading of files to and from GridFS"""
|
"""Proxy object to handle writing and reading of files to and from GridFS
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
.. versionchanged:: 0.5 - added optional size param to read
|
||||||
|
.. versionchanged:: 0.6 - added collection name param
|
||||||
|
"""
|
||||||
|
|
||||||
_fs = None
|
_fs = None
|
||||||
|
|
||||||
@@ -1727,12 +1754,12 @@ class GridFSProxy:
|
|||||||
return self.__copy__()
|
return self.__copy__()
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
return f"<{self.__class__.__name__}: {self.grid_id}>"
|
return "<{}: {}>".format(self.__class__.__name__, self.grid_id)
|
||||||
|
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
gridout = self.get()
|
gridout = self.get()
|
||||||
filename = gridout.filename if gridout else "<no file>"
|
filename = getattr(gridout, "filename") if gridout else "<no file>"
|
||||||
return f"<{self.__class__.__name__}: {filename} ({self.grid_id})>"
|
return "<{}: {} ({})>".format(self.__class__.__name__, filename, self.grid_id)
|
||||||
|
|
||||||
def __eq__(self, other):
|
def __eq__(self, other):
|
||||||
if isinstance(other, GridFSProxy):
|
if isinstance(other, GridFSProxy):
|
||||||
@@ -1831,7 +1858,12 @@ class GridFSProxy:
|
|||||||
|
|
||||||
|
|
||||||
class FileField(BaseField):
|
class FileField(BaseField):
|
||||||
"""A GridFS storage field."""
|
"""A GridFS storage field.
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
.. versionchanged:: 0.5 added optional size param for read
|
||||||
|
.. versionchanged:: 0.6 added db_alias for multidb support
|
||||||
|
"""
|
||||||
|
|
||||||
proxy_class = GridFSProxy
|
proxy_class = GridFSProxy
|
||||||
|
|
||||||
@@ -1913,7 +1945,11 @@ class FileField(BaseField):
|
|||||||
|
|
||||||
|
|
||||||
class ImageGridFsProxy(GridFSProxy):
|
class ImageGridFsProxy(GridFSProxy):
|
||||||
"""Proxy for ImageField"""
|
"""
|
||||||
|
Proxy for ImageField
|
||||||
|
|
||||||
|
versionadded: 0.6
|
||||||
|
"""
|
||||||
|
|
||||||
def put(self, file_obj, **kwargs):
|
def put(self, file_obj, **kwargs):
|
||||||
"""
|
"""
|
||||||
@@ -2047,6 +2083,8 @@ class ImageField(FileField):
|
|||||||
:param size: max size to store images, provided as (width, height, force)
|
:param size: max size to store images, provided as (width, height, force)
|
||||||
if larger, it will be automatically resized (ex: size=(800, 600, True))
|
if larger, it will be automatically resized (ex: size=(800, 600, True))
|
||||||
:param thumbnail_size: size to generate a thumbnail, provided as (width, height, force)
|
:param thumbnail_size: size to generate a thumbnail, provided as (width, height, force)
|
||||||
|
|
||||||
|
.. versionadded:: 0.6
|
||||||
"""
|
"""
|
||||||
|
|
||||||
proxy_class = ImageGridFsProxy
|
proxy_class = ImageGridFsProxy
|
||||||
@@ -2094,6 +2132,9 @@ class SequenceField(BaseField):
|
|||||||
In case the counter is defined in the abstract document, it will be
|
In case the counter is defined in the abstract document, it will be
|
||||||
common to all inherited documents and the default sequence name will
|
common to all inherited documents and the default sequence name will
|
||||||
be the class name of the abstract document.
|
be the class name of the abstract document.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
|
.. versionchanged:: 0.8 added `value_decorator`
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_auto_gen = True
|
_auto_gen = True
|
||||||
@@ -2107,7 +2148,7 @@ class SequenceField(BaseField):
|
|||||||
sequence_name=None,
|
sequence_name=None,
|
||||||
value_decorator=None,
|
value_decorator=None,
|
||||||
*args,
|
*args,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
self.collection_name = collection_name or self.COLLECTION_NAME
|
self.collection_name = collection_name or self.COLLECTION_NAME
|
||||||
self.db_alias = db_alias or DEFAULT_CONNECTION_NAME
|
self.db_alias = db_alias or DEFAULT_CONNECTION_NAME
|
||||||
@@ -2122,7 +2163,7 @@ class SequenceField(BaseField):
|
|||||||
Generate and Increment the counter
|
Generate and Increment the counter
|
||||||
"""
|
"""
|
||||||
sequence_name = self.get_sequence_name()
|
sequence_name = self.get_sequence_name()
|
||||||
sequence_id = f"{sequence_name}.{self.name}"
|
sequence_id = "{}.{}".format(sequence_name, self.name)
|
||||||
collection = get_db(alias=self.db_alias)[self.collection_name]
|
collection = get_db(alias=self.db_alias)[self.collection_name]
|
||||||
|
|
||||||
counter = collection.find_one_and_update(
|
counter = collection.find_one_and_update(
|
||||||
@@ -2136,7 +2177,7 @@ class SequenceField(BaseField):
|
|||||||
def set_next_value(self, value):
|
def set_next_value(self, value):
|
||||||
"""Helper method to set the next sequence value"""
|
"""Helper method to set the next sequence value"""
|
||||||
sequence_name = self.get_sequence_name()
|
sequence_name = self.get_sequence_name()
|
||||||
sequence_id = f"{sequence_name}.{self.name}"
|
sequence_id = "{}.{}".format(sequence_name, self.name)
|
||||||
collection = get_db(alias=self.db_alias)[self.collection_name]
|
collection = get_db(alias=self.db_alias)[self.collection_name]
|
||||||
counter = collection.find_one_and_update(
|
counter = collection.find_one_and_update(
|
||||||
filter={"_id": sequence_id},
|
filter={"_id": sequence_id},
|
||||||
@@ -2153,7 +2194,7 @@ class SequenceField(BaseField):
|
|||||||
as it is only fixed on set.
|
as it is only fixed on set.
|
||||||
"""
|
"""
|
||||||
sequence_name = self.get_sequence_name()
|
sequence_name = self.get_sequence_name()
|
||||||
sequence_id = f"{sequence_name}.{self.name}"
|
sequence_id = "{}.{}".format(sequence_name, self.name)
|
||||||
collection = get_db(alias=self.db_alias)[self.collection_name]
|
collection = get_db(alias=self.db_alias)[self.collection_name]
|
||||||
data = collection.find_one({"_id": sequence_id})
|
data = collection.find_one({"_id": sequence_id})
|
||||||
|
|
||||||
@@ -2206,7 +2247,10 @@ class SequenceField(BaseField):
|
|||||||
|
|
||||||
|
|
||||||
class UUIDField(BaseField):
|
class UUIDField(BaseField):
|
||||||
"""A UUID field."""
|
"""A UUID field.
|
||||||
|
|
||||||
|
.. versionadded:: 0.6
|
||||||
|
"""
|
||||||
|
|
||||||
_binary = None
|
_binary = None
|
||||||
|
|
||||||
@@ -2215,6 +2259,9 @@ class UUIDField(BaseField):
|
|||||||
Store UUID data in the database
|
Store UUID data in the database
|
||||||
|
|
||||||
:param binary: if False store as a string.
|
:param binary: if False store as a string.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.8.0
|
||||||
|
.. versionchanged:: 0.6.19
|
||||||
"""
|
"""
|
||||||
self._binary = binary
|
self._binary = binary
|
||||||
super().__init__(**kwargs)
|
super().__init__(**kwargs)
|
||||||
@@ -2259,6 +2306,8 @@ class GeoPointField(BaseField):
|
|||||||
representing a geo point. It admits 2d indexes but not "2dsphere" indexes
|
representing a geo point. It admits 2d indexes but not "2dsphere" indexes
|
||||||
in MongoDB > 2.4 which are more natural for modeling geospatial points.
|
in MongoDB > 2.4 which are more natural for modeling geospatial points.
|
||||||
See :ref:`geospatial-indexes`
|
See :ref:`geospatial-indexes`
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_geo_index = pymongo.GEO2D
|
_geo_index = pymongo.GEO2D
|
||||||
@@ -2290,6 +2339,8 @@ class PointField(GeoJsonBaseField):
|
|||||||
to set the value.
|
to set the value.
|
||||||
|
|
||||||
Requires mongodb >= 2.4
|
Requires mongodb >= 2.4
|
||||||
|
|
||||||
|
.. versionadded:: 0.8
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_type = "Point"
|
_type = "Point"
|
||||||
@@ -2308,6 +2359,8 @@ class LineStringField(GeoJsonBaseField):
|
|||||||
You can either pass a dict with the full information or a list of points.
|
You can either pass a dict with the full information or a list of points.
|
||||||
|
|
||||||
Requires mongodb >= 2.4
|
Requires mongodb >= 2.4
|
||||||
|
|
||||||
|
.. versionadded:: 0.8
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_type = "LineString"
|
_type = "LineString"
|
||||||
@@ -2329,6 +2382,8 @@ class PolygonField(GeoJsonBaseField):
|
|||||||
holes.
|
holes.
|
||||||
|
|
||||||
Requires mongodb >= 2.4
|
Requires mongodb >= 2.4
|
||||||
|
|
||||||
|
.. versionadded:: 0.8
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_type = "Polygon"
|
_type = "Polygon"
|
||||||
@@ -2348,6 +2403,8 @@ class MultiPointField(GeoJsonBaseField):
|
|||||||
to set the value.
|
to set the value.
|
||||||
|
|
||||||
Requires mongodb >= 2.6
|
Requires mongodb >= 2.6
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_type = "MultiPoint"
|
_type = "MultiPoint"
|
||||||
@@ -2367,6 +2424,8 @@ class MultiLineStringField(GeoJsonBaseField):
|
|||||||
You can either pass a dict with the full information or a list of points.
|
You can either pass a dict with the full information or a list of points.
|
||||||
|
|
||||||
Requires mongodb >= 2.6
|
Requires mongodb >= 2.6
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_type = "MultiLineString"
|
_type = "MultiLineString"
|
||||||
@@ -2393,6 +2452,8 @@ class MultiPolygonField(GeoJsonBaseField):
|
|||||||
of Polygons.
|
of Polygons.
|
||||||
|
|
||||||
Requires mongodb >= 2.6
|
Requires mongodb >= 2.6
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
|
|
||||||
_type = "MultiPolygon"
|
_type = "MultiPolygon"
|
||||||
@@ -2405,6 +2466,8 @@ class LazyReferenceField(BaseField):
|
|||||||
Instead, access will return a :class:`~mongoengine.base.LazyReference` class
|
Instead, access will return a :class:`~mongoengine.base.LazyReference` class
|
||||||
instance, allowing access to `pk` or manual dereference by using
|
instance, allowing access to `pk` or manual dereference by using
|
||||||
``fetch()`` method.
|
``fetch()`` method.
|
||||||
|
|
||||||
|
.. versionadded:: 0.15
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
@@ -2413,7 +2476,7 @@ class LazyReferenceField(BaseField):
|
|||||||
passthrough=False,
|
passthrough=False,
|
||||||
dbref=False,
|
dbref=False,
|
||||||
reverse_delete_rule=DO_NOTHING,
|
reverse_delete_rule=DO_NOTHING,
|
||||||
**kwargs,
|
**kwargs
|
||||||
):
|
):
|
||||||
"""Initialises the Reference Field.
|
"""Initialises the Reference Field.
|
||||||
|
|
||||||
@@ -2423,7 +2486,7 @@ class LazyReferenceField(BaseField):
|
|||||||
object is deleted
|
object is deleted
|
||||||
:param passthrough: When trying to access unknown fields, the
|
:param passthrough: When trying to access unknown fields, the
|
||||||
:class:`~mongoengine.base.datastructure.LazyReference` instance will
|
:class:`~mongoengine.base.datastructure.LazyReference` instance will
|
||||||
automatically call `fetch()` and try to retrieve the field on the fetched
|
automatically call `fetch()` and try to retrive the field on the fetched
|
||||||
document. Note this only work getting field (not setting or deleting).
|
document. Note this only work getting field (not setting or deleting).
|
||||||
"""
|
"""
|
||||||
# XXX ValidationError raised outside of the "validate" method.
|
# XXX ValidationError raised outside of the "validate" method.
|
||||||
@@ -2507,7 +2570,6 @@ class LazyReferenceField(BaseField):
|
|||||||
if not isinstance(value, (DBRef, Document, EmbeddedDocument)):
|
if not isinstance(value, (DBRef, Document, EmbeddedDocument)):
|
||||||
collection = self.document_type._get_collection_name()
|
collection = self.document_type._get_collection_name()
|
||||||
value = DBRef(collection, self.document_type.id.to_python(value))
|
value = DBRef(collection, self.document_type.id.to_python(value))
|
||||||
value = self.build_lazyref(value)
|
|
||||||
return value
|
return value
|
||||||
|
|
||||||
def validate(self, value):
|
def validate(self, value):
|
||||||
@@ -2568,6 +2630,8 @@ class GenericLazyReferenceField(GenericReferenceField):
|
|||||||
it.
|
it.
|
||||||
|
|
||||||
* You can use the choices param to limit the acceptable Document types
|
* You can use the choices param to limit the acceptable Document types
|
||||||
|
|
||||||
|
.. versionadded:: 0.15
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
def __init__(self, *args, **kwargs):
|
||||||
|
@@ -3,12 +3,11 @@ Helper functions, constants, and types to aid with MongoDB version support
|
|||||||
"""
|
"""
|
||||||
from mongoengine.connection import get_connection
|
from mongoengine.connection import get_connection
|
||||||
|
|
||||||
|
|
||||||
# Constant that can be used to compare the version retrieved with
|
# Constant that can be used to compare the version retrieved with
|
||||||
# get_mongodb_version()
|
# get_mongodb_version()
|
||||||
MONGODB_34 = (3, 4)
|
MONGODB_34 = (3, 4)
|
||||||
MONGODB_36 = (3, 6)
|
MONGODB_36 = (3, 6)
|
||||||
MONGODB_42 = (4, 2)
|
|
||||||
MONGODB_44 = (4, 4)
|
|
||||||
|
|
||||||
|
|
||||||
def get_mongodb_version():
|
def get_mongodb_version():
|
||||||
|
@@ -14,7 +14,8 @@ IS_PYMONGO_GTE_37 = PYMONGO_VERSION >= _PYMONGO_37
|
|||||||
def count_documents(
|
def count_documents(
|
||||||
collection, filter, skip=None, limit=None, hint=None, collation=None
|
collection, filter, skip=None, limit=None, hint=None, collation=None
|
||||||
):
|
):
|
||||||
"""Pymongo>3.7 deprecates count in favour of count_documents"""
|
"""Pymongo>3.7 deprecates count in favour of count_documents
|
||||||
|
"""
|
||||||
if limit == 0:
|
if limit == 0:
|
||||||
return 0 # Pymongo raises an OperationFailure if called with limit=0
|
return 0 # Pymongo raises an OperationFailure if called with limit=0
|
||||||
|
|
||||||
|
@@ -2,12 +2,13 @@ import copy
|
|||||||
import itertools
|
import itertools
|
||||||
import re
|
import re
|
||||||
import warnings
|
import warnings
|
||||||
|
|
||||||
from collections.abc import Mapping
|
from collections.abc import Mapping
|
||||||
|
|
||||||
import pymongo
|
|
||||||
import pymongo.errors
|
|
||||||
from bson import SON, json_util
|
from bson import SON, json_util
|
||||||
from bson.code import Code
|
from bson.code import Code
|
||||||
|
import pymongo
|
||||||
|
import pymongo.errors
|
||||||
from pymongo.collection import ReturnDocument
|
from pymongo.collection import ReturnDocument
|
||||||
from pymongo.common import validate_read_preference
|
from pymongo.common import validate_read_preference
|
||||||
from pymongo.read_concern import ReadConcern
|
from pymongo.read_concern import ReadConcern
|
||||||
@@ -33,6 +34,7 @@ from mongoengine.queryset import transform
|
|||||||
from mongoengine.queryset.field_list import QueryFieldList
|
from mongoengine.queryset.field_list import QueryFieldList
|
||||||
from mongoengine.queryset.visitor import Q, QNode
|
from mongoengine.queryset.visitor import Q, QNode
|
||||||
|
|
||||||
|
|
||||||
__all__ = ("BaseQuerySet", "DO_NOTHING", "NULLIFY", "CASCADE", "DENY", "PULL")
|
__all__ = ("BaseQuerySet", "DO_NOTHING", "NULLIFY", "CASCADE", "DENY", "PULL")
|
||||||
|
|
||||||
# Delete rules
|
# Delete rules
|
||||||
@@ -62,7 +64,6 @@ class BaseQuerySet:
|
|||||||
self._ordering = None
|
self._ordering = None
|
||||||
self._snapshot = False
|
self._snapshot = False
|
||||||
self._timeout = True
|
self._timeout = True
|
||||||
self._allow_disk_use = False
|
|
||||||
self._read_preference = None
|
self._read_preference = None
|
||||||
self._read_concern = None
|
self._read_concern = None
|
||||||
self._iter = False
|
self._iter = False
|
||||||
@@ -188,8 +189,7 @@ class BaseQuerySet:
|
|||||||
if queryset._scalar:
|
if queryset._scalar:
|
||||||
return queryset._get_scalar(
|
return queryset._get_scalar(
|
||||||
queryset._document._from_son(
|
queryset._document._from_son(
|
||||||
queryset._cursor[key],
|
queryset._cursor[key], _auto_dereference=self._auto_dereference,
|
||||||
_auto_dereference=self._auto_dereference,
|
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -197,8 +197,7 @@ class BaseQuerySet:
|
|||||||
return queryset._cursor[key]
|
return queryset._cursor[key]
|
||||||
|
|
||||||
return queryset._document._from_son(
|
return queryset._document._from_son(
|
||||||
queryset._cursor[key],
|
queryset._cursor[key], _auto_dereference=self._auto_dereference,
|
||||||
_auto_dereference=self._auto_dereference,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
raise TypeError("Provide a slice or an integer index")
|
raise TypeError("Provide a slice or an integer index")
|
||||||
@@ -257,6 +256,8 @@ class BaseQuerySet:
|
|||||||
`DocumentName.MultipleObjectsReturned` exception if multiple results
|
`DocumentName.MultipleObjectsReturned` exception if multiple results
|
||||||
and :class:`~mongoengine.queryset.DoesNotExist` or
|
and :class:`~mongoengine.queryset.DoesNotExist` or
|
||||||
`DocumentName.DoesNotExist` if no results are found.
|
`DocumentName.DoesNotExist` if no results are found.
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset = queryset.order_by().limit(2)
|
queryset = queryset.order_by().limit(2)
|
||||||
@@ -280,7 +281,10 @@ class BaseQuerySet:
|
|||||||
)
|
)
|
||||||
|
|
||||||
def create(self, **kwargs):
|
def create(self, **kwargs):
|
||||||
"""Create new object. Returns the saved object instance."""
|
"""Create new object. Returns the saved object instance.
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
"""
|
||||||
return self._document(**kwargs).save(force_insert=True)
|
return self._document(**kwargs).save(force_insert=True)
|
||||||
|
|
||||||
def first(self):
|
def first(self):
|
||||||
@@ -312,6 +316,10 @@ class BaseQuerySet:
|
|||||||
|
|
||||||
By default returns document instances, set ``load_bulk`` to False to
|
By default returns document instances, set ``load_bulk`` to False to
|
||||||
return just ``ObjectIds``
|
return just ``ObjectIds``
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
|
.. versionchanged:: 0.10.7
|
||||||
|
Add signal_kwargs argument
|
||||||
"""
|
"""
|
||||||
Document = _import_class("Document")
|
Document = _import_class("Document")
|
||||||
|
|
||||||
@@ -419,8 +427,8 @@ class BaseQuerySet:
|
|||||||
|
|
||||||
count = count_documents(
|
count = count_documents(
|
||||||
collection=self._cursor.collection,
|
collection=self._cursor.collection,
|
||||||
filter=self._query,
|
filter=self._cursor._Cursor__spec,
|
||||||
**kwargs,
|
**kwargs
|
||||||
)
|
)
|
||||||
|
|
||||||
self._cursor_obj = None
|
self._cursor_obj = None
|
||||||
@@ -524,7 +532,7 @@ class BaseQuerySet:
|
|||||||
write_concern=None,
|
write_concern=None,
|
||||||
read_concern=None,
|
read_concern=None,
|
||||||
full_result=False,
|
full_result=False,
|
||||||
**update,
|
**update
|
||||||
):
|
):
|
||||||
"""Perform an atomic update on the fields matched by the query.
|
"""Perform an atomic update on the fields matched by the query.
|
||||||
|
|
||||||
@@ -542,6 +550,8 @@ class BaseQuerySet:
|
|||||||
:param update: Django-style update keyword arguments
|
:param update: Django-style update keyword arguments
|
||||||
|
|
||||||
:returns the number of updated documents (unless ``full_result`` is True)
|
:returns the number of updated documents (unless ``full_result`` is True)
|
||||||
|
|
||||||
|
.. versionadded:: 0.2
|
||||||
"""
|
"""
|
||||||
if not update and not upsert:
|
if not update and not upsert:
|
||||||
raise OperationError("No update parameters, would remove data")
|
raise OperationError("No update parameters, would remove data")
|
||||||
@@ -593,6 +603,8 @@ class BaseQuerySet:
|
|||||||
:param update: Django-style update keyword arguments
|
:param update: Django-style update keyword arguments
|
||||||
|
|
||||||
:returns the new or overwritten document
|
:returns the new or overwritten document
|
||||||
|
|
||||||
|
.. versionadded:: 0.10.2
|
||||||
"""
|
"""
|
||||||
|
|
||||||
atomic_update = self.update(
|
atomic_update = self.update(
|
||||||
@@ -601,7 +613,7 @@ class BaseQuerySet:
|
|||||||
write_concern=write_concern,
|
write_concern=write_concern,
|
||||||
read_concern=read_concern,
|
read_concern=read_concern,
|
||||||
full_result=True,
|
full_result=True,
|
||||||
**update,
|
**update
|
||||||
)
|
)
|
||||||
|
|
||||||
if atomic_update.raw_result["updatedExisting"]:
|
if atomic_update.raw_result["updatedExisting"]:
|
||||||
@@ -626,13 +638,14 @@ class BaseQuerySet:
|
|||||||
:param update: Django-style update keyword arguments
|
:param update: Django-style update keyword arguments
|
||||||
full_result
|
full_result
|
||||||
:returns the number of updated documents (unless ``full_result`` is True)
|
:returns the number of updated documents (unless ``full_result`` is True)
|
||||||
|
.. versionadded:: 0.2
|
||||||
"""
|
"""
|
||||||
return self.update(
|
return self.update(
|
||||||
upsert=upsert,
|
upsert=upsert,
|
||||||
multi=False,
|
multi=False,
|
||||||
write_concern=write_concern,
|
write_concern=write_concern,
|
||||||
full_result=full_result,
|
full_result=full_result,
|
||||||
**update,
|
**update
|
||||||
)
|
)
|
||||||
|
|
||||||
def modify(
|
def modify(
|
||||||
@@ -657,6 +670,8 @@ class BaseQuerySet:
|
|||||||
:param new: return updated rather than original document
|
:param new: return updated rather than original document
|
||||||
(default ``False``)
|
(default ``False``)
|
||||||
:param update: Django-style update keyword arguments
|
:param update: Django-style update keyword arguments
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if remove and new:
|
if remove and new:
|
||||||
@@ -690,7 +705,7 @@ class BaseQuerySet:
|
|||||||
upsert=upsert,
|
upsert=upsert,
|
||||||
sort=sort,
|
sort=sort,
|
||||||
return_document=return_doc,
|
return_document=return_doc,
|
||||||
**self._cursor_args,
|
**self._cursor_args
|
||||||
)
|
)
|
||||||
except pymongo.errors.DuplicateKeyError as err:
|
except pymongo.errors.DuplicateKeyError as err:
|
||||||
raise NotUniqueError("Update failed (%s)" % err)
|
raise NotUniqueError("Update failed (%s)" % err)
|
||||||
@@ -712,6 +727,8 @@ class BaseQuerySet:
|
|||||||
`None` if no document exists with that id.
|
`None` if no document exists with that id.
|
||||||
|
|
||||||
:param object_id: the value for the id of the document to look up
|
:param object_id: the value for the id of the document to look up
|
||||||
|
|
||||||
|
.. versionchanged:: 0.6 Raises InvalidQueryError if filter has been set
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
if not queryset._query_obj.empty:
|
if not queryset._query_obj.empty:
|
||||||
@@ -720,11 +737,13 @@ class BaseQuerySet:
|
|||||||
return queryset.filter(pk=object_id).first()
|
return queryset.filter(pk=object_id).first()
|
||||||
|
|
||||||
def in_bulk(self, object_ids):
|
def in_bulk(self, object_ids):
|
||||||
""" "Retrieve a set of documents by their ids.
|
""""Retrieve a set of documents by their ids.
|
||||||
|
|
||||||
:param object_ids: a list or tuple of ObjectId's
|
:param object_ids: a list or tuple of ObjectId's
|
||||||
:rtype: dict of ObjectId's as keys and collection-specific
|
:rtype: dict of ObjectId's as keys and collection-specific
|
||||||
Document subclasses as values.
|
Document subclasses as values.
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
"""
|
"""
|
||||||
doc_map = {}
|
doc_map = {}
|
||||||
|
|
||||||
@@ -738,8 +757,7 @@ class BaseQuerySet:
|
|||||||
else:
|
else:
|
||||||
for doc in docs:
|
for doc in docs:
|
||||||
doc_map[doc["_id"]] = self._document._from_son(
|
doc_map[doc["_id"]] = self._document._from_son(
|
||||||
doc,
|
doc, _auto_dereference=self._auto_dereference,
|
||||||
_auto_dereference=self._auto_dereference,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
return doc_map
|
return doc_map
|
||||||
@@ -767,6 +785,8 @@ class BaseQuerySet:
|
|||||||
evaluated against if you are using more than one database.
|
evaluated against if you are using more than one database.
|
||||||
|
|
||||||
:param alias: The database alias
|
:param alias: The database alias
|
||||||
|
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
|
|
||||||
with switch_db(self._document, alias) as cls:
|
with switch_db(self._document, alias) as cls:
|
||||||
@@ -798,7 +818,6 @@ class BaseQuerySet:
|
|||||||
"_ordering",
|
"_ordering",
|
||||||
"_snapshot",
|
"_snapshot",
|
||||||
"_timeout",
|
"_timeout",
|
||||||
"_allow_disk_use",
|
|
||||||
"_read_preference",
|
"_read_preference",
|
||||||
"_read_concern",
|
"_read_concern",
|
||||||
"_iter",
|
"_iter",
|
||||||
@@ -829,6 +848,8 @@ class BaseQuerySet:
|
|||||||
"""Handles dereferencing of :class:`~bson.dbref.DBRef` objects or
|
"""Handles dereferencing of :class:`~bson.dbref.DBRef` objects or
|
||||||
:class:`~bson.object_id.ObjectId` a maximum depth in order to cut down
|
:class:`~bson.object_id.ObjectId` a maximum depth in order to cut down
|
||||||
the number queries to mongodb.
|
the number queries to mongodb.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
# Make select related work the same for querysets
|
# Make select related work the same for querysets
|
||||||
max_depth += 1
|
max_depth += 1
|
||||||
@@ -877,6 +898,8 @@ class BaseQuerySet:
|
|||||||
|
|
||||||
Hinting will not do anything if the corresponding index does not exist.
|
Hinting will not do anything if the corresponding index does not exist.
|
||||||
The last hint applied to this cursor takes precedence over all others.
|
The last hint applied to this cursor takes precedence over all others.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset._hint = index
|
queryset._hint = index
|
||||||
@@ -938,6 +961,10 @@ class BaseQuerySet:
|
|||||||
|
|
||||||
.. note:: This is a command and won't take ordering or limit into
|
.. note:: This is a command and won't take ordering or limit into
|
||||||
account.
|
account.
|
||||||
|
|
||||||
|
.. versionadded:: 0.4
|
||||||
|
.. versionchanged:: 0.5 - Fixed handling references
|
||||||
|
.. versionchanged:: 0.6 - Improved db_field refrence handling
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
|
|
||||||
@@ -1001,6 +1028,9 @@ class BaseQuerySet:
|
|||||||
field filters.
|
field filters.
|
||||||
|
|
||||||
:param fields: fields to include
|
:param fields: fields to include
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
|
.. versionchanged:: 0.5 - Added subfield support
|
||||||
"""
|
"""
|
||||||
fields = {f: QueryFieldList.ONLY for f in fields}
|
fields = {f: QueryFieldList.ONLY for f in fields}
|
||||||
return self.fields(True, **fields)
|
return self.fields(True, **fields)
|
||||||
@@ -1019,6 +1049,8 @@ class BaseQuerySet:
|
|||||||
field filters.
|
field filters.
|
||||||
|
|
||||||
:param fields: fields to exclude
|
:param fields: fields to exclude
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
fields = {f: QueryFieldList.EXCLUDE for f in fields}
|
fields = {f: QueryFieldList.EXCLUDE for f in fields}
|
||||||
return self.fields(**fields)
|
return self.fields(**fields)
|
||||||
@@ -1045,6 +1077,8 @@ class BaseQuerySet:
|
|||||||
|
|
||||||
:param kwargs: A set of keyword arguments identifying what to
|
:param kwargs: A set of keyword arguments identifying what to
|
||||||
include, exclude, or slice.
|
include, exclude, or slice.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# Check for an operator and transform to mongo-style if there is
|
# Check for an operator and transform to mongo-style if there is
|
||||||
@@ -1086,6 +1120,8 @@ class BaseQuerySet:
|
|||||||
.exclude(). ::
|
.exclude(). ::
|
||||||
|
|
||||||
post = BlogPost.objects.exclude('comments').all_fields()
|
post = BlogPost.objects.exclude('comments').all_fields()
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset._loaded_fields = QueryFieldList(
|
queryset._loaded_fields = QueryFieldList(
|
||||||
@@ -1158,6 +1194,9 @@ class BaseQuerySet:
|
|||||||
"""Enable or disable snapshot mode when querying.
|
"""Enable or disable snapshot mode when querying.
|
||||||
|
|
||||||
:param enabled: whether or not snapshot mode is enabled
|
:param enabled: whether or not snapshot mode is enabled
|
||||||
|
|
||||||
|
..versionchanged:: 0.5 - made chainable
|
||||||
|
.. deprecated:: Ignored with PyMongo 3+
|
||||||
"""
|
"""
|
||||||
msg = "snapshot is deprecated as it has no impact when using PyMongo 3+."
|
msg = "snapshot is deprecated as it has no impact when using PyMongo 3+."
|
||||||
warnings.warn(msg, DeprecationWarning)
|
warnings.warn(msg, DeprecationWarning)
|
||||||
@@ -1165,20 +1204,12 @@ class BaseQuerySet:
|
|||||||
queryset._snapshot = enabled
|
queryset._snapshot = enabled
|
||||||
return queryset
|
return queryset
|
||||||
|
|
||||||
def allow_disk_use(self, enabled):
|
|
||||||
"""Enable or disable the use of temporary files on disk while processing a blocking sort operation.
|
|
||||||
(To store data exceeding the 100 megabyte system memory limit)
|
|
||||||
|
|
||||||
:param enabled: whether or not temporary files on disk are used
|
|
||||||
"""
|
|
||||||
queryset = self.clone()
|
|
||||||
queryset._allow_disk_use = enabled
|
|
||||||
return queryset
|
|
||||||
|
|
||||||
def timeout(self, enabled):
|
def timeout(self, enabled):
|
||||||
"""Enable or disable the default mongod timeout when querying. (no_cursor_timeout option)
|
"""Enable or disable the default mongod timeout when querying. (no_cursor_timeout option)
|
||||||
|
|
||||||
:param enabled: whether or not the timeout is used
|
:param enabled: whether or not the timeout is used
|
||||||
|
|
||||||
|
..versionchanged:: 0.5 - made chainable
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset._timeout = enabled
|
queryset._timeout = enabled
|
||||||
@@ -1203,7 +1234,7 @@ class BaseQuerySet:
|
|||||||
preference.
|
preference.
|
||||||
"""
|
"""
|
||||||
if read_concern is not None and not isinstance(read_concern, Mapping):
|
if read_concern is not None and not isinstance(read_concern, Mapping):
|
||||||
raise TypeError(f"{read_concern!r} is not a valid read concern.")
|
raise TypeError("%r is not a valid read concern." % (read_concern,))
|
||||||
|
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
queryset._read_concern = (
|
queryset._read_concern = (
|
||||||
@@ -1277,6 +1308,7 @@ class BaseQuerySet:
|
|||||||
parameter will be removed shortly
|
parameter will be removed shortly
|
||||||
:param kwargs: (optional) kwargs dictionary to be passed to pymongo's aggregate call
|
:param kwargs: (optional) kwargs dictionary to be passed to pymongo's aggregate call
|
||||||
See https://api.mongodb.com/python/current/api/pymongo/collection.html#pymongo.collection.Collection.aggregate
|
See https://api.mongodb.com/python/current/api/pymongo/collection.html#pymongo.collection.Collection.aggregate
|
||||||
|
.. versionadded:: 0.9
|
||||||
"""
|
"""
|
||||||
using_deprecated_interface = isinstance(pipeline, dict) or bool(suppl_pipeline)
|
using_deprecated_interface = isinstance(pipeline, dict) or bool(suppl_pipeline)
|
||||||
user_pipeline = [pipeline] if isinstance(pipeline, dict) else list(pipeline)
|
user_pipeline = [pipeline] if isinstance(pipeline, dict) else list(pipeline)
|
||||||
@@ -1348,23 +1380,32 @@ class BaseQuerySet:
|
|||||||
Map/Reduce changed in server version **>= 1.7.4**. The PyMongo
|
Map/Reduce changed in server version **>= 1.7.4**. The PyMongo
|
||||||
:meth:`~pymongo.collection.Collection.map_reduce` helper requires
|
:meth:`~pymongo.collection.Collection.map_reduce` helper requires
|
||||||
PyMongo version **>= 1.11**.
|
PyMongo version **>= 1.11**.
|
||||||
|
|
||||||
|
.. versionchanged:: 0.5
|
||||||
|
- removed ``keep_temp`` keyword argument, which was only relevant
|
||||||
|
for MongoDB server versions older than 1.7.4
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
|
|
||||||
MapReduceDocument = _import_class("MapReduceDocument")
|
MapReduceDocument = _import_class("MapReduceDocument")
|
||||||
|
|
||||||
|
if not hasattr(self._collection, "map_reduce"):
|
||||||
|
raise NotImplementedError("Requires MongoDB >= 1.7.1")
|
||||||
|
|
||||||
map_f_scope = {}
|
map_f_scope = {}
|
||||||
if isinstance(map_f, Code):
|
if isinstance(map_f, Code):
|
||||||
map_f_scope = map_f.scope
|
map_f_scope = map_f.scope
|
||||||
map_f = str(map_f)
|
map_f = str(map_f)
|
||||||
map_f = Code(queryset._sub_js_fields(map_f), map_f_scope or None)
|
map_f = Code(queryset._sub_js_fields(map_f), map_f_scope)
|
||||||
|
|
||||||
reduce_f_scope = {}
|
reduce_f_scope = {}
|
||||||
if isinstance(reduce_f, Code):
|
if isinstance(reduce_f, Code):
|
||||||
reduce_f_scope = reduce_f.scope
|
reduce_f_scope = reduce_f.scope
|
||||||
reduce_f = str(reduce_f)
|
reduce_f = str(reduce_f)
|
||||||
reduce_f_code = queryset._sub_js_fields(reduce_f)
|
reduce_f_code = queryset._sub_js_fields(reduce_f)
|
||||||
reduce_f = Code(reduce_f_code, reduce_f_scope or None)
|
reduce_f = Code(reduce_f_code, reduce_f_scope)
|
||||||
|
|
||||||
mr_args = {"query": queryset._query}
|
mr_args = {"query": queryset._query}
|
||||||
|
|
||||||
@@ -1374,7 +1415,7 @@ class BaseQuerySet:
|
|||||||
finalize_f_scope = finalize_f.scope
|
finalize_f_scope = finalize_f.scope
|
||||||
finalize_f = str(finalize_f)
|
finalize_f = str(finalize_f)
|
||||||
finalize_f_code = queryset._sub_js_fields(finalize_f)
|
finalize_f_code = queryset._sub_js_fields(finalize_f)
|
||||||
finalize_f = Code(finalize_f_code, finalize_f_scope or None)
|
finalize_f = Code(finalize_f_code, finalize_f_scope)
|
||||||
mr_args["finalize"] = finalize_f
|
mr_args["finalize"] = finalize_f
|
||||||
|
|
||||||
if scope:
|
if scope:
|
||||||
@@ -1481,6 +1522,8 @@ class BaseQuerySet:
|
|||||||
.. note:: When using this mode of query, the database will call your
|
.. note:: When using this mode of query, the database will call your
|
||||||
function, or evaluate your predicate clause, for each object
|
function, or evaluate your predicate clause, for each object
|
||||||
in the collection.
|
in the collection.
|
||||||
|
|
||||||
|
.. versionadded:: 0.5
|
||||||
"""
|
"""
|
||||||
queryset = self.clone()
|
queryset = self.clone()
|
||||||
where_clause = queryset._sub_js_fields(where_clause)
|
where_clause = queryset._sub_js_fields(where_clause)
|
||||||
@@ -1557,6 +1600,9 @@ class BaseQuerySet:
|
|||||||
:param field: the field to use
|
:param field: the field to use
|
||||||
:param normalize: normalize the results so they add to 1.0
|
:param normalize: normalize the results so they add to 1.0
|
||||||
:param map_reduce: Use map_reduce over exec_js
|
:param map_reduce: Use map_reduce over exec_js
|
||||||
|
|
||||||
|
.. versionchanged:: 0.5 defaults to map_reduce and can handle embedded
|
||||||
|
document lookups
|
||||||
"""
|
"""
|
||||||
if map_reduce:
|
if map_reduce:
|
||||||
return self._item_frequencies_map_reduce(field, normalize=normalize)
|
return self._item_frequencies_map_reduce(field, normalize=normalize)
|
||||||
@@ -1565,7 +1611,8 @@ class BaseQuerySet:
|
|||||||
# Iterator helpers
|
# Iterator helpers
|
||||||
|
|
||||||
def __next__(self):
|
def __next__(self):
|
||||||
"""Wrap the result in a :class:`~mongoengine.Document` object."""
|
"""Wrap the result in a :class:`~mongoengine.Document` object.
|
||||||
|
"""
|
||||||
if self._none or self._empty:
|
if self._none or self._empty:
|
||||||
raise StopIteration
|
raise StopIteration
|
||||||
|
|
||||||
@@ -1575,8 +1622,7 @@ class BaseQuerySet:
|
|||||||
return raw_doc
|
return raw_doc
|
||||||
|
|
||||||
doc = self._document._from_son(
|
doc = self._document._from_son(
|
||||||
raw_doc,
|
raw_doc, _auto_dereference=self._auto_dereference,
|
||||||
_auto_dereference=self._auto_dereference,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
if self._scalar:
|
if self._scalar:
|
||||||
@@ -1585,7 +1631,10 @@ class BaseQuerySet:
|
|||||||
return doc
|
return doc
|
||||||
|
|
||||||
def rewind(self):
|
def rewind(self):
|
||||||
"""Rewind the cursor to its unevaluated state."""
|
"""Rewind the cursor to its unevaluated state.
|
||||||
|
|
||||||
|
.. versionadded:: 0.3
|
||||||
|
"""
|
||||||
self._iter = False
|
self._iter = False
|
||||||
self._cursor.rewind()
|
self._cursor.rewind()
|
||||||
|
|
||||||
@@ -1611,9 +1660,6 @@ class BaseQuerySet:
|
|||||||
if not self._timeout:
|
if not self._timeout:
|
||||||
cursor_args["no_cursor_timeout"] = True
|
cursor_args["no_cursor_timeout"] = True
|
||||||
|
|
||||||
if self._allow_disk_use:
|
|
||||||
cursor_args["allow_disk_use"] = True
|
|
||||||
|
|
||||||
if self._loaded_fields:
|
if self._loaded_fields:
|
||||||
cursor_args[fields_name] = self._loaded_fields.as_dict()
|
cursor_args[fields_name] = self._loaded_fields.as_dict()
|
||||||
|
|
||||||
|
@@ -1,5 +1,4 @@
|
|||||||
from functools import partial
|
from functools import partial
|
||||||
|
|
||||||
from mongoengine.queryset.queryset import QuerySet
|
from mongoengine.queryset.queryset import QuerySet
|
||||||
|
|
||||||
__all__ = ("queryset_manager", "QuerySetManager")
|
__all__ = ("queryset_manager", "QuerySetManager")
|
||||||
|
@@ -1,11 +1,11 @@
|
|||||||
from mongoengine.errors import OperationError
|
from mongoengine.errors import OperationError
|
||||||
from mongoengine.queryset.base import (
|
from mongoengine.queryset.base import (
|
||||||
|
BaseQuerySet,
|
||||||
CASCADE,
|
CASCADE,
|
||||||
DENY,
|
DENY,
|
||||||
DO_NOTHING,
|
DO_NOTHING,
|
||||||
NULLIFY,
|
NULLIFY,
|
||||||
PULL,
|
PULL,
|
||||||
BaseQuerySet,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
__all__ = (
|
__all__ = (
|
||||||
@@ -150,7 +150,10 @@ class QuerySet(BaseQuerySet):
|
|||||||
return self._len
|
return self._len
|
||||||
|
|
||||||
def no_cache(self):
|
def no_cache(self):
|
||||||
"""Convert to a non-caching queryset"""
|
"""Convert to a non-caching queryset
|
||||||
|
|
||||||
|
.. versionadded:: 0.8.3 Convert to non caching queryset
|
||||||
|
"""
|
||||||
if self._result_cache is not None:
|
if self._result_cache is not None:
|
||||||
raise OperationError("QuerySet already cached")
|
raise OperationError("QuerySet already cached")
|
||||||
|
|
||||||
@@ -161,11 +164,17 @@ class QuerySetNoCache(BaseQuerySet):
|
|||||||
"""A non caching QuerySet"""
|
"""A non caching QuerySet"""
|
||||||
|
|
||||||
def cache(self):
|
def cache(self):
|
||||||
"""Convert to a caching queryset"""
|
"""Convert to a caching queryset
|
||||||
|
|
||||||
|
.. versionadded:: 0.8.3 Convert to caching queryset
|
||||||
|
"""
|
||||||
return self._clone_into(QuerySet(self._document, self._collection))
|
return self._clone_into(QuerySet(self._document, self._collection))
|
||||||
|
|
||||||
def __repr__(self):
|
def __repr__(self):
|
||||||
"""Provides the string representation of the QuerySet"""
|
"""Provides the string representation of the QuerySet
|
||||||
|
|
||||||
|
.. versionchanged:: 0.6.13 Now doesnt modify the cursor
|
||||||
|
"""
|
||||||
if self._iter:
|
if self._iter:
|
||||||
return ".. queryset mid-iteration .."
|
return ".. queryset mid-iteration .."
|
||||||
|
|
||||||
|
@@ -1,8 +1,8 @@
|
|||||||
from collections import defaultdict
|
from collections import defaultdict
|
||||||
|
|
||||||
import pymongo
|
from bson import ObjectId, SON
|
||||||
from bson import SON, ObjectId
|
|
||||||
from bson.dbref import DBRef
|
from bson.dbref import DBRef
|
||||||
|
import pymongo
|
||||||
|
|
||||||
from mongoengine.base import UPDATE_OPERATORS
|
from mongoengine.base import UPDATE_OPERATORS
|
||||||
from mongoengine.common import _import_class
|
from mongoengine.common import _import_class
|
||||||
|
@@ -13,14 +13,17 @@ def warn_empty_is_deprecated():
|
|||||||
|
|
||||||
|
|
||||||
class QNodeVisitor:
|
class QNodeVisitor:
|
||||||
"""Base visitor class for visiting Q-object nodes in a query tree."""
|
"""Base visitor class for visiting Q-object nodes in a query tree.
|
||||||
|
"""
|
||||||
|
|
||||||
def visit_combination(self, combination):
|
def visit_combination(self, combination):
|
||||||
"""Called by QCombination objects."""
|
"""Called by QCombination objects.
|
||||||
|
"""
|
||||||
return combination
|
return combination
|
||||||
|
|
||||||
def visit_query(self, query):
|
def visit_query(self, query):
|
||||||
"""Called by (New)Q objects."""
|
"""Called by (New)Q objects.
|
||||||
|
"""
|
||||||
return query
|
return query
|
||||||
|
|
||||||
|
|
||||||
@@ -46,7 +49,8 @@ class SimplificationVisitor(QNodeVisitor):
|
|||||||
return combination
|
return combination
|
||||||
|
|
||||||
def _query_conjunction(self, queries):
|
def _query_conjunction(self, queries):
|
||||||
"""Merges query dicts - effectively &ing them together."""
|
"""Merges query dicts - effectively &ing them together.
|
||||||
|
"""
|
||||||
query_ops = set()
|
query_ops = set()
|
||||||
combined_query = {}
|
combined_query = {}
|
||||||
for query in queries:
|
for query in queries:
|
||||||
|
@@ -1,5 +1,6 @@
|
|||||||
black
|
black
|
||||||
flake8
|
flake8
|
||||||
|
flake8-import-order
|
||||||
pre-commit
|
pre-commit
|
||||||
pytest
|
pytest
|
||||||
ipdb
|
ipdb
|
||||||
|
12
setup.cfg
12
setup.cfg
@@ -1,18 +1,10 @@
|
|||||||
[flake8]
|
[flake8]
|
||||||
ignore=E501,F403,F405,I201,I202,W504,W605,W503,B007
|
ignore=E501,F401,F403,F405,I201,I202,W504, W605, W503
|
||||||
exclude=build,dist,docs,venv,venv3,.tox,.eggs,tests
|
exclude=build,dist,docs,venv,venv3,.tox,.eggs,tests
|
||||||
max-complexity=47
|
max-complexity=47
|
||||||
|
application-import-names=mongoengine,tests
|
||||||
|
|
||||||
[tool:pytest]
|
[tool:pytest]
|
||||||
# Limits the discovery to tests directory
|
# Limits the discovery to tests directory
|
||||||
# avoids that it runs for instance the benchmark
|
# avoids that it runs for instance the benchmark
|
||||||
testpaths = tests
|
testpaths = tests
|
||||||
|
|
||||||
[isort]
|
|
||||||
known_first_party = mongoengine,tests
|
|
||||||
default_section = THIRDPARTY
|
|
||||||
multi_line_output = 3
|
|
||||||
include_trailing_comma = True
|
|
||||||
combine_as_imports = True
|
|
||||||
line_length = 70
|
|
||||||
ensure_newline_before_comments = 1
|
|
||||||
|
11
setup.py
11
setup.py
@@ -7,7 +7,7 @@ from setuptools.command.test import test as TestCommand
|
|||||||
|
|
||||||
# Hack to silence atexit traceback in newer python versions
|
# Hack to silence atexit traceback in newer python versions
|
||||||
try:
|
try:
|
||||||
import multiprocessing # noqa: F401
|
import multiprocessing
|
||||||
except ImportError:
|
except ImportError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
@@ -53,8 +53,8 @@ class PyTest(TestCommand):
|
|||||||
|
|
||||||
def run_tests(self):
|
def run_tests(self):
|
||||||
# import here, cause outside the eggs aren't loaded
|
# import here, cause outside the eggs aren't loaded
|
||||||
import pytest
|
|
||||||
from pkg_resources import _namespace_packages
|
from pkg_resources import _namespace_packages
|
||||||
|
import pytest
|
||||||
|
|
||||||
# Purge modules under test from sys.modules. The test loader will
|
# Purge modules under test from sys.modules. The test loader will
|
||||||
# re-import them from the build location. Required when 2to3 is used
|
# re-import them from the build location. Required when 2to3 is used
|
||||||
@@ -98,6 +98,7 @@ CLASSIFIERS = [
|
|||||||
"Operating System :: OS Independent",
|
"Operating System :: OS Independent",
|
||||||
"Programming Language :: Python",
|
"Programming Language :: Python",
|
||||||
"Programming Language :: Python :: 3",
|
"Programming Language :: Python :: 3",
|
||||||
|
"Programming Language :: Python :: 3.5",
|
||||||
"Programming Language :: Python :: 3.6",
|
"Programming Language :: Python :: 3.6",
|
||||||
"Programming Language :: Python :: 3.7",
|
"Programming Language :: Python :: 3.7",
|
||||||
"Programming Language :: Python :: 3.8",
|
"Programming Language :: Python :: 3.8",
|
||||||
@@ -112,9 +113,9 @@ extra_opts = {
|
|||||||
"tests_require": [
|
"tests_require": [
|
||||||
"pytest<5.0",
|
"pytest<5.0",
|
||||||
"pytest-cov",
|
"pytest-cov",
|
||||||
"coverage",
|
"coverage<5.0", # recent coverage switched to sqlite format for the .coverage file which isn't handled properly by coveralls
|
||||||
"blinker",
|
"blinker",
|
||||||
"Pillow>=7.0.0",
|
"Pillow>=2.0.0, <7.0.0", # 7.0.0 dropped Python2 support
|
||||||
],
|
],
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -139,7 +140,7 @@ setup(
|
|||||||
long_description=LONG_DESCRIPTION,
|
long_description=LONG_DESCRIPTION,
|
||||||
platforms=["any"],
|
platforms=["any"],
|
||||||
classifiers=CLASSIFIERS,
|
classifiers=CLASSIFIERS,
|
||||||
python_requires=">=3.6",
|
python_requires=">=3.5",
|
||||||
install_requires=["pymongo>=3.4, <4.0"],
|
install_requires=["pymongo>=3.4, <4.0"],
|
||||||
cmdclass={"test": PyTest},
|
cmdclass={"test": PyTest},
|
||||||
**extra_opts
|
**extra_opts
|
||||||
|
@@ -26,14 +26,16 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
self.db.drop_collection(collection)
|
self.db.drop_collection(collection)
|
||||||
|
|
||||||
def test_definition(self):
|
def test_definition(self):
|
||||||
"""Ensure that document may be defined using fields."""
|
"""Ensure that document may be defined using fields.
|
||||||
|
"""
|
||||||
assert ["_cls", "age", "id", "name"] == sorted(self.Person._fields.keys())
|
assert ["_cls", "age", "id", "name"] == sorted(self.Person._fields.keys())
|
||||||
assert ["IntField", "ObjectIdField", "StringField", "StringField"] == sorted(
|
assert ["IntField", "ObjectIdField", "StringField", "StringField"] == sorted(
|
||||||
x.__class__.__name__ for x in self.Person._fields.values()
|
[x.__class__.__name__ for x in self.Person._fields.values()]
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_get_db(self):
|
def test_get_db(self):
|
||||||
"""Ensure that get_db returns the expected db."""
|
"""Ensure that get_db returns the expected db.
|
||||||
|
"""
|
||||||
db = self.Person._get_db()
|
db = self.Person._get_db()
|
||||||
assert self.db == db
|
assert self.db == db
|
||||||
|
|
||||||
@@ -45,13 +47,15 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert collection_name == self.Person._get_collection_name()
|
assert collection_name == self.Person._get_collection_name()
|
||||||
|
|
||||||
def test_get_collection(self):
|
def test_get_collection(self):
|
||||||
"""Ensure that get_collection returns the expected collection."""
|
"""Ensure that get_collection returns the expected collection.
|
||||||
|
"""
|
||||||
collection_name = "person"
|
collection_name = "person"
|
||||||
collection = self.Person._get_collection()
|
collection = self.Person._get_collection()
|
||||||
assert self.db[collection_name] == collection
|
assert self.db[collection_name] == collection
|
||||||
|
|
||||||
def test_drop_collection(self):
|
def test_drop_collection(self):
|
||||||
"""Ensure that the collection may be dropped from the database."""
|
"""Ensure that the collection may be dropped from the database.
|
||||||
|
"""
|
||||||
collection_name = "person"
|
collection_name = "person"
|
||||||
self.Person(name="Test").save()
|
self.Person(name="Test").save()
|
||||||
assert collection_name in list_collection_names(self.db)
|
assert collection_name in list_collection_names(self.db)
|
||||||
@@ -73,7 +77,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert self.Person._meta["delete_rules"] == {(Job, "employee"): NULLIFY}
|
assert self.Person._meta["delete_rules"] == {(Job, "employee"): NULLIFY}
|
||||||
|
|
||||||
def test_compare_indexes(self):
|
def test_compare_indexes(self):
|
||||||
"""Ensure that the indexes are properly created and that
|
""" Ensure that the indexes are properly created and that
|
||||||
compare_indexes identifies the missing/extra indexes
|
compare_indexes identifies the missing/extra indexes
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@@ -106,7 +110,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
}
|
}
|
||||||
|
|
||||||
def test_compare_indexes_inheritance(self):
|
def test_compare_indexes_inheritance(self):
|
||||||
"""Ensure that the indexes are properly created and that
|
""" Ensure that the indexes are properly created and that
|
||||||
compare_indexes identifies the missing/extra indexes for subclassed
|
compare_indexes identifies the missing/extra indexes for subclassed
|
||||||
documents (_cls included)
|
documents (_cls included)
|
||||||
"""
|
"""
|
||||||
@@ -146,7 +150,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
}
|
}
|
||||||
|
|
||||||
def test_compare_indexes_multiple_subclasses(self):
|
def test_compare_indexes_multiple_subclasses(self):
|
||||||
"""Ensure that compare_indexes behaves correctly if called from a
|
""" Ensure that compare_indexes behaves correctly if called from a
|
||||||
class, which base class has multiple subclasses
|
class, which base class has multiple subclasses
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@@ -177,7 +181,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert BlogPostWithCustomField.compare_indexes() == {"missing": [], "extra": []}
|
assert BlogPostWithCustomField.compare_indexes() == {"missing": [], "extra": []}
|
||||||
|
|
||||||
def test_compare_indexes_for_text_indexes(self):
|
def test_compare_indexes_for_text_indexes(self):
|
||||||
"""Ensure that compare_indexes behaves correctly for text indexes"""
|
""" Ensure that compare_indexes behaves correctly for text indexes """
|
||||||
|
|
||||||
class Doc(Document):
|
class Doc(Document):
|
||||||
a = StringField()
|
a = StringField()
|
||||||
@@ -199,7 +203,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert actual == expected
|
assert actual == expected
|
||||||
|
|
||||||
def test_list_indexes_inheritance(self):
|
def test_list_indexes_inheritance(self):
|
||||||
"""ensure that all of the indexes are listed regardless of the super-
|
""" ensure that all of the indexes are listed regardless of the super-
|
||||||
or sub-class that we call it from
|
or sub-class that we call it from
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@@ -231,7 +235,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert BlogPost.list_indexes() == [
|
assert BlogPost.list_indexes() == [
|
||||||
[("_cls", 1), ("author", 1), ("tags", 1)],
|
[("_cls", 1), ("author", 1), ("tags", 1)],
|
||||||
[("_cls", 1), ("author", 1), ("tags", 1), ("extra_text", 1)],
|
[("_cls", 1), ("author", 1), ("tags", 1), ("extra_text", 1)],
|
||||||
[("_id", 1)],
|
[(u"_id", 1)],
|
||||||
[("_cls", 1)],
|
[("_cls", 1)],
|
||||||
]
|
]
|
||||||
|
|
||||||
@@ -256,7 +260,8 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert Vaccine._meta["delete_rules"][(Cat, "vaccine_made")] == PULL
|
assert Vaccine._meta["delete_rules"][(Cat, "vaccine_made")] == PULL
|
||||||
|
|
||||||
def test_collection_naming(self):
|
def test_collection_naming(self):
|
||||||
"""Ensure that a collection with a specified name may be used."""
|
"""Ensure that a collection with a specified name may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class DefaultNamingTest(Document):
|
class DefaultNamingTest(Document):
|
||||||
pass
|
pass
|
||||||
@@ -288,7 +293,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert "wibble" == InheritedAbstractNamingTest._get_collection_name()
|
assert "wibble" == InheritedAbstractNamingTest._get_collection_name()
|
||||||
|
|
||||||
# Mixin tests
|
# Mixin tests
|
||||||
class BaseMixin:
|
class BaseMixin(object):
|
||||||
meta = {"collection": lambda c: c.__name__.lower()}
|
meta = {"collection": lambda c: c.__name__.lower()}
|
||||||
|
|
||||||
class OldMixinNamingConvention(Document, BaseMixin):
|
class OldMixinNamingConvention(Document, BaseMixin):
|
||||||
@@ -299,7 +304,7 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
== OldMixinNamingConvention._get_collection_name()
|
== OldMixinNamingConvention._get_collection_name()
|
||||||
)
|
)
|
||||||
|
|
||||||
class BaseMixin:
|
class BaseMixin(object):
|
||||||
meta = {"collection": lambda c: c.__name__.lower()}
|
meta = {"collection": lambda c: c.__name__.lower()}
|
||||||
|
|
||||||
class BaseDocument(Document, BaseMixin):
|
class BaseDocument(Document, BaseMixin):
|
||||||
@@ -311,7 +316,8 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert "basedocument" == MyDocument._get_collection_name()
|
assert "basedocument" == MyDocument._get_collection_name()
|
||||||
|
|
||||||
def test_custom_collection_name_operations(self):
|
def test_custom_collection_name_operations(self):
|
||||||
"""Ensure that a collection with a specified name is used as expected."""
|
"""Ensure that a collection with a specified name is used as expected.
|
||||||
|
"""
|
||||||
collection_name = "personCollTest"
|
collection_name = "personCollTest"
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
@@ -331,7 +337,8 @@ class TestClassMethods(unittest.TestCase):
|
|||||||
assert collection_name not in list_collection_names(self.db)
|
assert collection_name not in list_collection_names(self.db)
|
||||||
|
|
||||||
def test_collection_name_and_primary(self):
|
def test_collection_name_and_primary(self):
|
||||||
"""Ensure that a collection with a specified name may be used."""
|
"""Ensure that a collection with a specified name may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField(primary_key=True)
|
name = StringField(primary_key=True)
|
||||||
|
@@ -1,7 +1,6 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from bson import SON
|
from bson import SON
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.pymongo_support import list_collection_names
|
from mongoengine.pymongo_support import list_collection_names
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
@@ -9,7 +8,7 @@ from tests.utils import MongoDBTestCase
|
|||||||
|
|
||||||
class TestDelta(MongoDBTestCase):
|
class TestDelta(MongoDBTestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
super().setUp()
|
super(TestDelta, self).setUp()
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -644,10 +643,7 @@ class TestDelta(MongoDBTestCase):
|
|||||||
doc.save()
|
doc.save()
|
||||||
doc = doc.reload(10)
|
doc = doc.reload(10)
|
||||||
|
|
||||||
assert doc._delta() == (
|
assert doc._delta() == ({}, {},)
|
||||||
{},
|
|
||||||
{},
|
|
||||||
)
|
|
||||||
del doc.embedded_field.list_field[2].list_field
|
del doc.embedded_field.list_field[2].list_field
|
||||||
assert doc._delta() == (
|
assert doc._delta() == (
|
||||||
{},
|
{},
|
||||||
|
@@ -10,7 +10,7 @@ __all__ = ("TestDynamicDocument",)
|
|||||||
|
|
||||||
class TestDynamicDocument(MongoDBTestCase):
|
class TestDynamicDocument(MongoDBTestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
super().setUp()
|
super(TestDynamicDocument, self).setUp()
|
||||||
|
|
||||||
class Person(DynamicDocument):
|
class Person(DynamicDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -118,17 +118,17 @@ class TestDynamicDocument(MongoDBTestCase):
|
|||||||
p.save()
|
p.save()
|
||||||
|
|
||||||
raw_p = Person.objects.as_pymongo().get(id=p.id)
|
raw_p = Person.objects.as_pymongo().get(id=p.id)
|
||||||
assert raw_p == {"_cls": "Person", "_id": p.id, "name": "Dean"}
|
assert raw_p == {"_cls": u"Person", "_id": p.id, "name": u"Dean"}
|
||||||
|
|
||||||
p.name = "OldDean"
|
p.name = "OldDean"
|
||||||
p.newattr = "garbage"
|
p.newattr = "garbage"
|
||||||
p.save()
|
p.save()
|
||||||
raw_p = Person.objects.as_pymongo().get(id=p.id)
|
raw_p = Person.objects.as_pymongo().get(id=p.id)
|
||||||
assert raw_p == {
|
assert raw_p == {
|
||||||
"_cls": "Person",
|
"_cls": u"Person",
|
||||||
"_id": p.id,
|
"_id": p.id,
|
||||||
"name": "OldDean",
|
"name": "OldDean",
|
||||||
"newattr": "garbage",
|
"newattr": u"garbage",
|
||||||
}
|
}
|
||||||
|
|
||||||
def test_fields_containing_underscore(self):
|
def test_fields_containing_underscore(self):
|
||||||
@@ -144,14 +144,14 @@ class TestDynamicDocument(MongoDBTestCase):
|
|||||||
p.save()
|
p.save()
|
||||||
|
|
||||||
raw_p = WeirdPerson.objects.as_pymongo().get(id=p.id)
|
raw_p = WeirdPerson.objects.as_pymongo().get(id=p.id)
|
||||||
assert raw_p == {"_id": p.id, "_name": "Dean", "name": "Dean"}
|
assert raw_p == {"_id": p.id, "_name": u"Dean", "name": u"Dean"}
|
||||||
|
|
||||||
p.name = "OldDean"
|
p.name = "OldDean"
|
||||||
p._name = "NewDean"
|
p._name = "NewDean"
|
||||||
p._newattr1 = "garbage" # Unknown fields won't be added
|
p._newattr1 = "garbage" # Unknown fields won't be added
|
||||||
p.save()
|
p.save()
|
||||||
raw_p = WeirdPerson.objects.as_pymongo().get(id=p.id)
|
raw_p = WeirdPerson.objects.as_pymongo().get(id=p.id)
|
||||||
assert raw_p == {"_id": p.id, "_name": "NewDean", "name": "OldDean"}
|
assert raw_p == {"_id": p.id, "_name": u"NewDean", "name": u"OldDean"}
|
||||||
|
|
||||||
def test_dynamic_document_queries(self):
|
def test_dynamic_document_queries(self):
|
||||||
"""Ensure we can query dynamic fields"""
|
"""Ensure we can query dynamic fields"""
|
||||||
|
@@ -1,16 +1,12 @@
|
|||||||
import unittest
|
import unittest
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
import pytest
|
|
||||||
from pymongo.collation import Collation
|
from pymongo.collation import Collation
|
||||||
from pymongo.errors import OperationFailure
|
from pymongo.errors import OperationFailure
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.mongodb_support import (
|
|
||||||
MONGODB_42,
|
|
||||||
get_mongodb_version,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class TestIndexes(unittest.TestCase):
|
class TestIndexes(unittest.TestCase):
|
||||||
@@ -175,7 +171,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert MyDoc._meta["index_specs"] == [{"fields": [("keywords", 1)]}]
|
assert MyDoc._meta["index_specs"] == [{"fields": [("keywords", 1)]}]
|
||||||
|
|
||||||
def test_embedded_document_index_meta(self):
|
def test_embedded_document_index_meta(self):
|
||||||
"""Ensure that embedded document indexes are created explicitly"""
|
"""Ensure that embedded document indexes are created explicitly
|
||||||
|
"""
|
||||||
|
|
||||||
class Rank(EmbeddedDocument):
|
class Rank(EmbeddedDocument):
|
||||||
title = StringField(required=True)
|
title = StringField(required=True)
|
||||||
@@ -197,7 +194,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert [("rank.title", 1)] in info
|
assert [("rank.title", 1)] in info
|
||||||
|
|
||||||
def test_explicit_geo2d_index(self):
|
def test_explicit_geo2d_index(self):
|
||||||
"""Ensure that geo2d indexes work when created via meta[indexes]"""
|
"""Ensure that geo2d indexes work when created via meta[indexes]
|
||||||
|
"""
|
||||||
|
|
||||||
class Place(Document):
|
class Place(Document):
|
||||||
location = DictField()
|
location = DictField()
|
||||||
@@ -211,7 +209,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert [("location.point", "2d")] in info
|
assert [("location.point", "2d")] in info
|
||||||
|
|
||||||
def test_explicit_geo2d_index_embedded(self):
|
def test_explicit_geo2d_index_embedded(self):
|
||||||
"""Ensure that geo2d indexes work when created via meta[indexes]"""
|
"""Ensure that geo2d indexes work when created via meta[indexes]
|
||||||
|
"""
|
||||||
|
|
||||||
class EmbeddedLocation(EmbeddedDocument):
|
class EmbeddedLocation(EmbeddedDocument):
|
||||||
location = DictField()
|
location = DictField()
|
||||||
@@ -230,7 +229,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert [("current.location.point", "2d")] in info
|
assert [("current.location.point", "2d")] in info
|
||||||
|
|
||||||
def test_explicit_geosphere_index(self):
|
def test_explicit_geosphere_index(self):
|
||||||
"""Ensure that geosphere indexes work when created via meta[indexes]"""
|
"""Ensure that geosphere indexes work when created via meta[indexes]
|
||||||
|
"""
|
||||||
|
|
||||||
class Place(Document):
|
class Place(Document):
|
||||||
location = DictField()
|
location = DictField()
|
||||||
@@ -246,7 +246,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert [("location.point", "2dsphere")] in info
|
assert [("location.point", "2dsphere")] in info
|
||||||
|
|
||||||
def test_explicit_geohaystack_index(self):
|
def test_explicit_geohaystack_index(self):
|
||||||
"""Ensure that geohaystack indexes work when created via meta[indexes]"""
|
"""Ensure that geohaystack indexes work when created via meta[indexes]
|
||||||
|
"""
|
||||||
pytest.skip(
|
pytest.skip(
|
||||||
"GeoHaystack index creation is not supported for now"
|
"GeoHaystack index creation is not supported for now"
|
||||||
"from meta, as it requires a bucketSize parameter."
|
"from meta, as it requires a bucketSize parameter."
|
||||||
@@ -267,7 +268,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert [("location.point", "geoHaystack")] in info
|
assert [("location.point", "geoHaystack")] in info
|
||||||
|
|
||||||
def test_create_geohaystack_index(self):
|
def test_create_geohaystack_index(self):
|
||||||
"""Ensure that geohaystack indexes can be created"""
|
"""Ensure that geohaystack indexes can be created
|
||||||
|
"""
|
||||||
|
|
||||||
class Place(Document):
|
class Place(Document):
|
||||||
location = DictField()
|
location = DictField()
|
||||||
@@ -362,7 +364,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert sorted(info.keys()) == ["_cls_1_user_guid_1", "_id_"]
|
assert sorted(info.keys()) == ["_cls_1_user_guid_1", "_id_"]
|
||||||
|
|
||||||
def test_embedded_document_index(self):
|
def test_embedded_document_index(self):
|
||||||
"""Tests settings an index on an embedded document"""
|
"""Tests settings an index on an embedded document
|
||||||
|
"""
|
||||||
|
|
||||||
class Date(EmbeddedDocument):
|
class Date(EmbeddedDocument):
|
||||||
year = IntField(db_field="yr")
|
year = IntField(db_field="yr")
|
||||||
@@ -379,7 +382,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert sorted(info.keys()) == ["_id_", "date.yr_-1"]
|
assert sorted(info.keys()) == ["_id_", "date.yr_-1"]
|
||||||
|
|
||||||
def test_list_embedded_document_index(self):
|
def test_list_embedded_document_index(self):
|
||||||
"""Ensure list embedded documents can be indexed"""
|
"""Ensure list embedded documents can be indexed
|
||||||
|
"""
|
||||||
|
|
||||||
class Tag(EmbeddedDocument):
|
class Tag(EmbeddedDocument):
|
||||||
name = StringField(db_field="tag")
|
name = StringField(db_field="tag")
|
||||||
@@ -415,7 +419,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert sorted(info.keys()) == ["_cls_1", "_id_"]
|
assert sorted(info.keys()) == ["_cls_1", "_id_"]
|
||||||
|
|
||||||
def test_covered_index(self):
|
def test_covered_index(self):
|
||||||
"""Ensure that covered indexes can be used"""
|
"""Ensure that covered indexes can be used
|
||||||
|
"""
|
||||||
|
|
||||||
class Test(Document):
|
class Test(Document):
|
||||||
a = IntField()
|
a = IntField()
|
||||||
@@ -456,11 +461,9 @@ class TestIndexes(unittest.TestCase):
|
|||||||
.get("stage")
|
.get("stage")
|
||||||
== "IXSCAN"
|
== "IXSCAN"
|
||||||
)
|
)
|
||||||
mongo_db = get_mongodb_version()
|
|
||||||
PROJECTION_STR = "PROJECTION" if mongo_db < MONGODB_42 else "PROJECTION_COVERED"
|
|
||||||
assert (
|
assert (
|
||||||
query_plan.get("queryPlanner").get("winningPlan").get("stage")
|
query_plan.get("queryPlanner").get("winningPlan").get("stage")
|
||||||
== PROJECTION_STR
|
== "PROJECTION"
|
||||||
)
|
)
|
||||||
|
|
||||||
query_plan = Test.objects(a=1).explain()
|
query_plan = Test.objects(a=1).explain()
|
||||||
@@ -555,7 +558,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
assert [x.name for x in query_result] == sorted(names)
|
assert [x.name for x in query_result] == sorted(names)
|
||||||
|
|
||||||
def test_unique(self):
|
def test_unique(self):
|
||||||
"""Ensure that uniqueness constraints are applied to fields."""
|
"""Ensure that uniqueness constraints are applied to fields.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -603,7 +607,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_unique_with(self):
|
def test_unique_with(self):
|
||||||
"""Ensure that unique_with constraints are applied to fields."""
|
"""Ensure that unique_with constraints are applied to fields.
|
||||||
|
"""
|
||||||
|
|
||||||
class Date(EmbeddedDocument):
|
class Date(EmbeddedDocument):
|
||||||
year = IntField(db_field="yr")
|
year = IntField(db_field="yr")
|
||||||
@@ -628,7 +633,8 @@ class TestIndexes(unittest.TestCase):
|
|||||||
post3.save()
|
post3.save()
|
||||||
|
|
||||||
def test_unique_embedded_document(self):
|
def test_unique_embedded_document(self):
|
||||||
"""Ensure that uniqueness constraints are applied to fields on embedded documents."""
|
"""Ensure that uniqueness constraints are applied to fields on embedded documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class SubDocument(EmbeddedDocument):
|
class SubDocument(EmbeddedDocument):
|
||||||
year = IntField(db_field="yr")
|
year = IntField(db_field="yr")
|
||||||
|
@@ -45,7 +45,8 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
test_doc.delete()
|
test_doc.delete()
|
||||||
|
|
||||||
def test_superclasses(self):
|
def test_superclasses(self):
|
||||||
"""Ensure that the correct list of superclasses is assembled."""
|
"""Ensure that the correct list of superclasses is assembled.
|
||||||
|
"""
|
||||||
|
|
||||||
class Animal(Document):
|
class Animal(Document):
|
||||||
meta = {"allow_inheritance": True}
|
meta = {"allow_inheritance": True}
|
||||||
@@ -215,7 +216,8 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
assert Pike._subclasses == ("Animal.Fish.Pike",)
|
assert Pike._subclasses == ("Animal.Fish.Pike",)
|
||||||
|
|
||||||
def test_inheritance_meta_data(self):
|
def test_inheritance_meta_data(self):
|
||||||
"""Ensure that document may inherit fields from a superclass document."""
|
"""Ensure that document may inherit fields from a superclass document.
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -232,7 +234,8 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
assert Employee._get_collection_name() == Person._get_collection_name()
|
assert Employee._get_collection_name() == Person._get_collection_name()
|
||||||
|
|
||||||
def test_inheritance_to_mongo_keys(self):
|
def test_inheritance_to_mongo_keys(self):
|
||||||
"""Ensure that document may inherit fields from a superclass document."""
|
"""Ensure that document may inherit fields from a superclass document.
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -256,7 +259,7 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
assert Employee._get_collection_name() == Person._get_collection_name()
|
assert Employee._get_collection_name() == Person._get_collection_name()
|
||||||
|
|
||||||
def test_indexes_and_multiple_inheritance(self):
|
def test_indexes_and_multiple_inheritance(self):
|
||||||
"""Ensure that all of the indexes are created for a document with
|
""" Ensure that all of the indexes are created for a document with
|
||||||
multiple inheritance.
|
multiple inheritance.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
@@ -280,11 +283,14 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
C.ensure_indexes()
|
C.ensure_indexes()
|
||||||
|
|
||||||
assert sorted(
|
assert sorted(
|
||||||
idx["key"] for idx in C._get_collection().index_information().values()
|
[idx["key"] for idx in C._get_collection().index_information().values()]
|
||||||
) == sorted([[("_cls", 1), ("b", 1)], [("_id", 1)], [("_cls", 1), ("a", 1)]])
|
) == sorted(
|
||||||
|
[[(u"_cls", 1), (u"b", 1)], [(u"_id", 1)], [(u"_cls", 1), (u"a", 1)]]
|
||||||
|
)
|
||||||
|
|
||||||
def test_polymorphic_queries(self):
|
def test_polymorphic_queries(self):
|
||||||
"""Ensure that the correct subclasses are returned from a query"""
|
"""Ensure that the correct subclasses are returned from a query
|
||||||
|
"""
|
||||||
|
|
||||||
class Animal(Document):
|
class Animal(Document):
|
||||||
meta = {"allow_inheritance": True}
|
meta = {"allow_inheritance": True}
|
||||||
@@ -341,7 +347,8 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
assert "_cls" not in obj
|
assert "_cls" not in obj
|
||||||
|
|
||||||
def test_cant_turn_off_inheritance_on_subclass(self):
|
def test_cant_turn_off_inheritance_on_subclass(self):
|
||||||
"""Ensure if inheritance is on in a subclass you cant turn it off."""
|
"""Ensure if inheritance is on in a subclass you cant turn it off.
|
||||||
|
"""
|
||||||
|
|
||||||
class Animal(Document):
|
class Animal(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -467,7 +474,7 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
assert city.pk is None
|
assert city.pk is None
|
||||||
# TODO: expected error? Shouldn't we create a new error type?
|
# TODO: expected error? Shouldn't we create a new error type?
|
||||||
with pytest.raises(KeyError):
|
with pytest.raises(KeyError):
|
||||||
city.pk = 1
|
setattr(city, "pk", 1)
|
||||||
|
|
||||||
def test_allow_inheritance_embedded_document(self):
|
def test_allow_inheritance_embedded_document(self):
|
||||||
"""Ensure embedded documents respect inheritance."""
|
"""Ensure embedded documents respect inheritance."""
|
||||||
@@ -491,7 +498,8 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
assert "_cls" in doc.to_mongo()
|
assert "_cls" in doc.to_mongo()
|
||||||
|
|
||||||
def test_document_inheritance(self):
|
def test_document_inheritance(self):
|
||||||
"""Ensure mutliple inheritance of abstract documents"""
|
"""Ensure mutliple inheritance of abstract documents
|
||||||
|
"""
|
||||||
|
|
||||||
class DateCreatedDocument(Document):
|
class DateCreatedDocument(Document):
|
||||||
meta = {"allow_inheritance": True, "abstract": True}
|
meta = {"allow_inheritance": True, "abstract": True}
|
||||||
@@ -499,8 +507,13 @@ class TestInheritance(MongoDBTestCase):
|
|||||||
class DateUpdatedDocument(Document):
|
class DateUpdatedDocument(Document):
|
||||||
meta = {"allow_inheritance": True, "abstract": True}
|
meta = {"allow_inheritance": True, "abstract": True}
|
||||||
|
|
||||||
class MyDocument(DateCreatedDocument, DateUpdatedDocument):
|
try:
|
||||||
pass
|
|
||||||
|
class MyDocument(DateCreatedDocument, DateUpdatedDocument):
|
||||||
|
pass
|
||||||
|
|
||||||
|
except Exception:
|
||||||
|
assert False, "Couldn't create MyDocument class"
|
||||||
|
|
||||||
def test_abstract_documents(self):
|
def test_abstract_documents(self):
|
||||||
"""Ensure that a document superclass can be marked as abstract
|
"""Ensure that a document superclass can be marked as abstract
|
||||||
|
@@ -6,9 +6,9 @@ import weakref
|
|||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
|
||||||
import bson
|
import bson
|
||||||
import pytest
|
|
||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
from pymongo.errors import DuplicateKeyError
|
from pymongo.errors import DuplicateKeyError
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine import signals
|
from mongoengine import signals
|
||||||
@@ -23,11 +23,7 @@ from mongoengine.errors import (
|
|||||||
NotUniqueError,
|
NotUniqueError,
|
||||||
SaveConditionError,
|
SaveConditionError,
|
||||||
)
|
)
|
||||||
from mongoengine.mongodb_support import (
|
from mongoengine.mongodb_support import MONGODB_34, MONGODB_36, get_mongodb_version
|
||||||
MONGODB_34,
|
|
||||||
MONGODB_36,
|
|
||||||
get_mongodb_version,
|
|
||||||
)
|
|
||||||
from mongoengine.pymongo_support import list_collection_names
|
from mongoengine.pymongo_support import list_collection_names
|
||||||
from mongoengine.queryset import NULLIFY, Q
|
from mongoengine.queryset import NULLIFY, Q
|
||||||
from tests import fixtures
|
from tests import fixtures
|
||||||
@@ -164,7 +160,8 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
Log.objects
|
Log.objects
|
||||||
|
|
||||||
def test_repr(self):
|
def test_repr(self):
|
||||||
"""Ensure that unicode representation works"""
|
"""Ensure that unicode representation works
|
||||||
|
"""
|
||||||
|
|
||||||
class Article(Document):
|
class Article(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -172,7 +169,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
def __unicode__(self):
|
def __unicode__(self):
|
||||||
return self.title
|
return self.title
|
||||||
|
|
||||||
doc = Article(title="привет мир")
|
doc = Article(title=u"привет мир")
|
||||||
|
|
||||||
assert "<Article: привет мир>" == repr(doc)
|
assert "<Article: привет мир>" == repr(doc)
|
||||||
|
|
||||||
@@ -185,7 +182,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
def __str__(self):
|
def __str__(self):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
doc = Article(title="привет мир")
|
doc = Article(title=u"привет мир")
|
||||||
assert "<Article: None>" == repr(doc)
|
assert "<Article: None>" == repr(doc)
|
||||||
|
|
||||||
def test_queryset_resurrects_dropped_collection(self):
|
def test_queryset_resurrects_dropped_collection(self):
|
||||||
@@ -525,9 +522,9 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
query_op = q.db.system.profile.find({"ns": "mongoenginetest.animal"})[0]
|
query_op = q.db.system.profile.find({"ns": "mongoenginetest.animal"})[0]
|
||||||
assert query_op["op"] == "update"
|
assert query_op["op"] == "update"
|
||||||
if mongo_db <= MONGODB_34:
|
if mongo_db <= MONGODB_34:
|
||||||
assert set(query_op["query"].keys()) == {"_id", "is_mammal"}
|
assert set(query_op["query"].keys()) == set(["_id", "is_mammal"])
|
||||||
else:
|
else:
|
||||||
assert set(query_op["command"]["q"].keys()) == {"_id", "is_mammal"}
|
assert set(query_op["command"]["q"].keys()) == set(["_id", "is_mammal"])
|
||||||
|
|
||||||
Animal.drop_collection()
|
Animal.drop_collection()
|
||||||
|
|
||||||
@@ -550,7 +547,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
query_op = q.db.system.profile.find({"ns": "mongoenginetest.animal"})[0]
|
query_op = q.db.system.profile.find({"ns": "mongoenginetest.animal"})[0]
|
||||||
assert query_op["op"] == "command"
|
assert query_op["op"] == "command"
|
||||||
assert query_op["command"]["findAndModify"] == "animal"
|
assert query_op["command"]["findAndModify"] == "animal"
|
||||||
assert set(query_op["command"]["query"].keys()) == {"_id", "is_mammal"}
|
assert set(query_op["command"]["query"].keys()) == set(["_id", "is_mammal"])
|
||||||
|
|
||||||
Animal.drop_collection()
|
Animal.drop_collection()
|
||||||
|
|
||||||
@@ -1432,11 +1429,11 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
coll = self.Person._get_collection()
|
coll = self.Person._get_collection()
|
||||||
doc = self.Person(name="John").save()
|
doc = self.Person(name="John").save()
|
||||||
raw_doc = coll.find_one({"_id": doc.pk})
|
raw_doc = coll.find_one({"_id": doc.pk})
|
||||||
assert set(raw_doc.keys()) == {"_id", "_cls", "name"}
|
assert set(raw_doc.keys()) == set(["_id", "_cls", "name"])
|
||||||
|
|
||||||
doc.update(rename__name="first_name")
|
doc.update(rename__name="first_name")
|
||||||
raw_doc = coll.find_one({"_id": doc.pk})
|
raw_doc = coll.find_one({"_id": doc.pk})
|
||||||
assert set(raw_doc.keys()) == {"_id", "_cls", "first_name"}
|
assert set(raw_doc.keys()) == set(["_id", "_cls", "first_name"])
|
||||||
assert raw_doc["first_name"] == "John"
|
assert raw_doc["first_name"] == "John"
|
||||||
|
|
||||||
def test_inserts_if_you_set_the_pk(self):
|
def test_inserts_if_you_set_the_pk(self):
|
||||||
@@ -1556,7 +1553,8 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
assert site.page.log_message == "Error: Dummy message"
|
assert site.page.log_message == "Error: Dummy message"
|
||||||
|
|
||||||
def test_update_list_field(self):
|
def test_update_list_field(self):
|
||||||
"""Test update on `ListField` with $pull + $in."""
|
"""Test update on `ListField` with $pull + $in.
|
||||||
|
"""
|
||||||
|
|
||||||
class Doc(Document):
|
class Doc(Document):
|
||||||
foo = ListField(StringField())
|
foo = ListField(StringField())
|
||||||
@@ -2045,7 +2043,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
assert promoted_employee.details is None
|
assert promoted_employee.details is None
|
||||||
|
|
||||||
def test_object_mixins(self):
|
def test_object_mixins(self):
|
||||||
class NameMixin:
|
class NameMixin(object):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
class Foo(EmbeddedDocument, NameMixin):
|
class Foo(EmbeddedDocument, NameMixin):
|
||||||
@@ -2059,7 +2057,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
assert ["id", "name", "widgets"] == sorted(Bar._fields.keys())
|
assert ["id", "name", "widgets"] == sorted(Bar._fields.keys())
|
||||||
|
|
||||||
def test_mixin_inheritance(self):
|
def test_mixin_inheritance(self):
|
||||||
class BaseMixIn:
|
class BaseMixIn(object):
|
||||||
count = IntField()
|
count = IntField()
|
||||||
data = StringField()
|
data = StringField()
|
||||||
|
|
||||||
@@ -2818,13 +2816,15 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
register_connection("testdb-2", "mongoenginetest2")
|
register_connection("testdb-2", "mongoenginetest2")
|
||||||
|
|
||||||
class A(Document):
|
class A(Document):
|
||||||
"""Uses default db_alias"""
|
"""Uses default db_alias
|
||||||
|
"""
|
||||||
|
|
||||||
name = StringField()
|
name = StringField()
|
||||||
meta = {"allow_inheritance": True}
|
meta = {"allow_inheritance": True}
|
||||||
|
|
||||||
class B(A):
|
class B(A):
|
||||||
"""Uses testdb-2 db_alias"""
|
"""Uses testdb-2 db_alias
|
||||||
|
"""
|
||||||
|
|
||||||
meta = {"db_alias": "testdb-2"}
|
meta = {"db_alias": "testdb-2"}
|
||||||
|
|
||||||
@@ -2904,32 +2904,50 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
# Checks
|
# Checks
|
||||||
assert ",".join([str(b) for b in Book.objects.all()]) == "1,2,3,4,5,6,7,8,9"
|
assert ",".join([str(b) for b in Book.objects.all()]) == "1,2,3,4,5,6,7,8,9"
|
||||||
# bob related books
|
# bob related books
|
||||||
bob_books_qs = Book.objects.filter(
|
assert (
|
||||||
Q(extra__a=bob) | Q(author=bob) | Q(extra__b=bob)
|
",".join(
|
||||||
|
[
|
||||||
|
str(b)
|
||||||
|
for b in Book.objects.filter(
|
||||||
|
Q(extra__a=bob) | Q(author=bob) | Q(extra__b=bob)
|
||||||
|
)
|
||||||
|
]
|
||||||
|
)
|
||||||
|
== "1,2,3,4"
|
||||||
)
|
)
|
||||||
assert [str(b) for b in bob_books_qs] == ["1", "2", "3", "4"]
|
|
||||||
assert bob_books_qs.count() == 4
|
|
||||||
|
|
||||||
# Susan & Karl related books
|
# Susan & Karl related books
|
||||||
susan_karl_books_qs = Book.objects.filter(
|
assert (
|
||||||
Q(extra__a__all=[karl, susan])
|
",".join(
|
||||||
| Q(author__all=[karl, susan])
|
[
|
||||||
| Q(extra__b__all=[karl.to_dbref(), susan.to_dbref()])
|
str(b)
|
||||||
|
for b in Book.objects.filter(
|
||||||
|
Q(extra__a__all=[karl, susan])
|
||||||
|
| Q(author__all=[karl, susan])
|
||||||
|
| Q(extra__b__all=[karl.to_dbref(), susan.to_dbref()])
|
||||||
|
)
|
||||||
|
]
|
||||||
|
)
|
||||||
|
== "1"
|
||||||
)
|
)
|
||||||
assert [str(b) for b in susan_karl_books_qs] == ["1"]
|
|
||||||
assert susan_karl_books_qs.count() == 1
|
|
||||||
|
|
||||||
# $Where
|
# $Where
|
||||||
custom_qs = Book.objects.filter(
|
assert (
|
||||||
__raw__={
|
u",".join(
|
||||||
"$where": """
|
[
|
||||||
|
str(b)
|
||||||
|
for b in Book.objects.filter(
|
||||||
|
__raw__={
|
||||||
|
"$where": """
|
||||||
function(){
|
function(){
|
||||||
return this.name == '1' ||
|
return this.name == '1' ||
|
||||||
this.name == '2';}"""
|
this.name == '2';}"""
|
||||||
}
|
}
|
||||||
|
)
|
||||||
|
]
|
||||||
|
)
|
||||||
|
== "1,2"
|
||||||
)
|
)
|
||||||
assert [str(b) for b in custom_qs] == ["1", "2"]
|
|
||||||
assert custom_qs.count() == 2
|
|
||||||
|
|
||||||
def test_switch_db_instance(self):
|
def test_switch_db_instance(self):
|
||||||
register_connection("testdb-1", "mongoenginetest2")
|
register_connection("testdb-1", "mongoenginetest2")
|
||||||
@@ -3289,7 +3307,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
for node_name, node in self.nodes.items():
|
for node_name, node in self.nodes.items():
|
||||||
node.expand()
|
node.expand()
|
||||||
node.save(*args, **kwargs)
|
node.save(*args, **kwargs)
|
||||||
super().save(*args, **kwargs)
|
super(NodesSystem, self).save(*args, **kwargs)
|
||||||
|
|
||||||
NodesSystem.drop_collection()
|
NodesSystem.drop_collection()
|
||||||
Node.drop_collection()
|
Node.drop_collection()
|
||||||
@@ -3594,7 +3612,8 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
assert u_from_db.height is None
|
assert u_from_db.height is None
|
||||||
|
|
||||||
def test_not_saved_eq(self):
|
def test_not_saved_eq(self):
|
||||||
"""Ensure we can compare documents not saved."""
|
"""Ensure we can compare documents not saved.
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
pass
|
pass
|
||||||
@@ -3738,7 +3757,7 @@ class TestDocumentInstance(MongoDBTestCase):
|
|||||||
_ = list(Jedi.objects) # Ensure a proper document loads without errors
|
_ = list(Jedi.objects) # Ensure a proper document loads without errors
|
||||||
|
|
||||||
# Forces a document with a wrong shape (may occur in case of migration)
|
# Forces a document with a wrong shape (may occur in case of migration)
|
||||||
value = "I_should_be_a_dict"
|
value = u"I_should_be_a_dict"
|
||||||
coll.insert_one({"light_saber": value})
|
coll.insert_one({"light_saber": value})
|
||||||
|
|
||||||
with pytest.raises(InvalidDocumentError) as exc_info:
|
with pytest.raises(InvalidDocumentError) as exc_info:
|
||||||
@@ -3803,95 +3822,5 @@ class ObjectKeyTestCase(MongoDBTestCase):
|
|||||||
assert book._object_key == {"pk": book.pk, "author__name": "Author"}
|
assert book._object_key == {"pk": book.pk, "author__name": "Author"}
|
||||||
|
|
||||||
|
|
||||||
class DBFieldMappingTest(MongoDBTestCase):
|
|
||||||
def setUp(self):
|
|
||||||
class Fields:
|
|
||||||
w1 = BooleanField(db_field="w2")
|
|
||||||
|
|
||||||
x1 = BooleanField(db_field="x2")
|
|
||||||
x2 = BooleanField(db_field="x3")
|
|
||||||
|
|
||||||
y1 = BooleanField(db_field="y0")
|
|
||||||
y2 = BooleanField(db_field="y1")
|
|
||||||
|
|
||||||
z1 = BooleanField(db_field="z2")
|
|
||||||
z2 = BooleanField(db_field="z1")
|
|
||||||
|
|
||||||
class Doc(Fields, Document):
|
|
||||||
pass
|
|
||||||
|
|
||||||
class DynDoc(Fields, DynamicDocument):
|
|
||||||
pass
|
|
||||||
|
|
||||||
self.Doc = Doc
|
|
||||||
self.DynDoc = DynDoc
|
|
||||||
|
|
||||||
def tearDown(self):
|
|
||||||
for collection in list_collection_names(self.db):
|
|
||||||
self.db.drop_collection(collection)
|
|
||||||
|
|
||||||
def test_setting_fields_in_constructor_of_strict_doc_uses_model_names(self):
|
|
||||||
doc = self.Doc(z1=True, z2=False)
|
|
||||||
assert doc.z1 is True
|
|
||||||
assert doc.z2 is False
|
|
||||||
|
|
||||||
def test_setting_fields_in_constructor_of_dyn_doc_uses_model_names(self):
|
|
||||||
doc = self.DynDoc(z1=True, z2=False)
|
|
||||||
assert doc.z1 is True
|
|
||||||
assert doc.z2 is False
|
|
||||||
|
|
||||||
def test_setting_unknown_field_in_constructor_of_dyn_doc_does_not_overwrite_model_fields(
|
|
||||||
self,
|
|
||||||
):
|
|
||||||
doc = self.DynDoc(w2=True)
|
|
||||||
assert doc.w1 is None
|
|
||||||
assert doc.w2 is True
|
|
||||||
|
|
||||||
def test_unknown_fields_of_strict_doc_do_not_overwrite_dbfields_1(self):
|
|
||||||
doc = self.Doc()
|
|
||||||
doc.w2 = True
|
|
||||||
doc.x3 = True
|
|
||||||
doc.y0 = True
|
|
||||||
doc.save()
|
|
||||||
reloaded = self.Doc.objects.get(id=doc.id)
|
|
||||||
assert reloaded.w1 is None
|
|
||||||
assert reloaded.x1 is None
|
|
||||||
assert reloaded.x2 is None
|
|
||||||
assert reloaded.y1 is None
|
|
||||||
assert reloaded.y2 is None
|
|
||||||
|
|
||||||
def test_dbfields_are_loaded_to_the_right_modelfield_for_strict_doc_2(self):
|
|
||||||
doc = self.Doc()
|
|
||||||
doc.x2 = True
|
|
||||||
doc.y2 = True
|
|
||||||
doc.z2 = True
|
|
||||||
doc.save()
|
|
||||||
reloaded = self.Doc.objects.get(id=doc.id)
|
|
||||||
assert (
|
|
||||||
reloaded.x1,
|
|
||||||
reloaded.x2,
|
|
||||||
reloaded.y1,
|
|
||||||
reloaded.y2,
|
|
||||||
reloaded.z1,
|
|
||||||
reloaded.z2,
|
|
||||||
) == (doc.x1, doc.x2, doc.y1, doc.y2, doc.z1, doc.z2)
|
|
||||||
|
|
||||||
def test_dbfields_are_loaded_to_the_right_modelfield_for_dyn_doc_2(self):
|
|
||||||
doc = self.DynDoc()
|
|
||||||
doc.x2 = True
|
|
||||||
doc.y2 = True
|
|
||||||
doc.z2 = True
|
|
||||||
doc.save()
|
|
||||||
reloaded = self.DynDoc.objects.get(id=doc.id)
|
|
||||||
assert (
|
|
||||||
reloaded.x1,
|
|
||||||
reloaded.x2,
|
|
||||||
reloaded.y1,
|
|
||||||
reloaded.y2,
|
|
||||||
reloaded.z1,
|
|
||||||
reloaded.z2,
|
|
||||||
) == (doc.x1, doc.x2, doc.y1, doc.y2, doc.z1, doc.z2)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -1,7 +1,7 @@
|
|||||||
import unittest
|
import unittest
|
||||||
import uuid
|
import uuid
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
from bson import ObjectId
|
from bson import ObjectId
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
@@ -9,7 +9,8 @@ from tests.utils import MongoDBTestCase
|
|||||||
|
|
||||||
class TestValidatorError(MongoDBTestCase):
|
class TestValidatorError(MongoDBTestCase):
|
||||||
def test_to_dict(self):
|
def test_to_dict(self):
|
||||||
"""Ensure a ValidationError handles error to_dict correctly."""
|
"""Ensure a ValidationError handles error to_dict correctly.
|
||||||
|
"""
|
||||||
error = ValidationError("root")
|
error = ValidationError("root")
|
||||||
assert error.to_dict() == {}
|
assert error.to_dict() == {}
|
||||||
|
|
||||||
@@ -89,7 +90,8 @@ class TestValidatorError(MongoDBTestCase):
|
|||||||
p.validate()
|
p.validate()
|
||||||
|
|
||||||
def test_embedded_document_validation(self):
|
def test_embedded_document_validation(self):
|
||||||
"""Ensure that embedded documents may be validated."""
|
"""Ensure that embedded documents may be validated.
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
date = DateTimeField()
|
date = DateTimeField()
|
||||||
@@ -210,7 +212,10 @@ class TestValidatorError(MongoDBTestCase):
|
|||||||
child.reference = parent
|
child.reference = parent
|
||||||
|
|
||||||
# Saving the child should not raise a ValidationError
|
# Saving the child should not raise a ValidationError
|
||||||
child.save()
|
try:
|
||||||
|
child.save()
|
||||||
|
except ValidationError as e:
|
||||||
|
self.fail("ValidationError raised: %s" % e.message)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
@@ -1,7 +1,7 @@
|
|||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
import pytest
|
|
||||||
from bson import Binary
|
from bson import Binary
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
@@ -13,13 +13,14 @@ BIN_VALUE = "\xa9\xf3\x8d(\xd7\x03\x84\xb4k[\x0f\xe3\xa2\x19\x85p[J\xa3\xd2>\xde
|
|||||||
|
|
||||||
class TestBinaryField(MongoDBTestCase):
|
class TestBinaryField(MongoDBTestCase):
|
||||||
def test_binary_fields(self):
|
def test_binary_fields(self):
|
||||||
"""Ensure that binary fields can be stored and retrieved."""
|
"""Ensure that binary fields can be stored and retrieved.
|
||||||
|
"""
|
||||||
|
|
||||||
class Attachment(Document):
|
class Attachment(Document):
|
||||||
content_type = StringField()
|
content_type = StringField()
|
||||||
blob = BinaryField()
|
blob = BinaryField()
|
||||||
|
|
||||||
BLOB = b"\xe6\x00\xc4\xff\x07"
|
BLOB = "\xe6\x00\xc4\xff\x07".encode("latin-1")
|
||||||
MIME_TYPE = "application/octet-stream"
|
MIME_TYPE = "application/octet-stream"
|
||||||
|
|
||||||
Attachment.drop_collection()
|
Attachment.drop_collection()
|
||||||
@@ -32,7 +33,8 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
assert BLOB == bytes(attachment_1.blob)
|
assert BLOB == bytes(attachment_1.blob)
|
||||||
|
|
||||||
def test_validation_succeeds(self):
|
def test_validation_succeeds(self):
|
||||||
"""Ensure that valid values can be assigned to binary fields."""
|
"""Ensure that valid values can be assigned to binary fields.
|
||||||
|
"""
|
||||||
|
|
||||||
class AttachmentRequired(Document):
|
class AttachmentRequired(Document):
|
||||||
blob = BinaryField(required=True)
|
blob = BinaryField(required=True)
|
||||||
@@ -43,11 +45,11 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
attachment_required = AttachmentRequired()
|
attachment_required = AttachmentRequired()
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
attachment_required.validate()
|
attachment_required.validate()
|
||||||
attachment_required.blob = Binary(b"\xe6\x00\xc4\xff\x07")
|
attachment_required.blob = Binary("\xe6\x00\xc4\xff\x07".encode("latin-1"))
|
||||||
attachment_required.validate()
|
attachment_required.validate()
|
||||||
|
|
||||||
_5_BYTES = b"\xe6\x00\xc4\xff\x07"
|
_5_BYTES = "\xe6\x00\xc4\xff\x07".encode("latin-1")
|
||||||
_4_BYTES = b"\xe6\x00\xc4\xff"
|
_4_BYTES = "\xe6\x00\xc4\xff".encode("latin-1")
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
AttachmentSizeLimit(blob=_5_BYTES).validate()
|
AttachmentSizeLimit(blob=_5_BYTES).validate()
|
||||||
AttachmentSizeLimit(blob=_4_BYTES).validate()
|
AttachmentSizeLimit(blob=_4_BYTES).validate()
|
||||||
@@ -58,7 +60,7 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
class Attachment(Document):
|
class Attachment(Document):
|
||||||
blob = BinaryField()
|
blob = BinaryField()
|
||||||
|
|
||||||
for invalid_data in (2, "Im_a_unicode", ["some_str"]):
|
for invalid_data in (2, u"Im_a_unicode", ["some_str"]):
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
Attachment(blob=invalid_data).validate()
|
Attachment(blob=invalid_data).validate()
|
||||||
|
|
||||||
@@ -129,7 +131,7 @@ class TestBinaryField(MongoDBTestCase):
|
|||||||
|
|
||||||
MyDocument.drop_collection()
|
MyDocument.drop_collection()
|
||||||
|
|
||||||
bin_data = b"\xe6\x00\xc4\xff\x07"
|
bin_data = "\xe6\x00\xc4\xff\x07".encode("latin-1")
|
||||||
doc = MyDocument(bin_field=bin_data).save()
|
doc = MyDocument(bin_field=bin_data).save()
|
||||||
|
|
||||||
n_updated = MyDocument.objects(bin_field=bin_data).update_one(
|
n_updated = MyDocument.objects(bin_field=bin_data).update_one(
|
||||||
|
@@ -13,17 +13,6 @@ class TestBooleanField(MongoDBTestCase):
|
|||||||
person.save()
|
person.save()
|
||||||
assert get_as_pymongo(person) == {"_id": person.id, "admin": True}
|
assert get_as_pymongo(person) == {"_id": person.id, "admin": True}
|
||||||
|
|
||||||
def test_construction_does_not_fail_uncastable_value(self):
|
|
||||||
class BoolFail:
|
|
||||||
def __bool__(self):
|
|
||||||
return "bogus"
|
|
||||||
|
|
||||||
class Person(Document):
|
|
||||||
admin = BooleanField()
|
|
||||||
|
|
||||||
person = Person(admin=BoolFail())
|
|
||||||
person.admin == "bogus"
|
|
||||||
|
|
||||||
def test_validation(self):
|
def test_validation(self):
|
||||||
"""Ensure that invalid values cannot be assigned to boolean
|
"""Ensure that invalid values cannot be assigned to boolean
|
||||||
fields.
|
fields.
|
||||||
|
@@ -2,28 +2,11 @@ from decimal import Decimal
|
|||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import (
|
from mongoengine import *
|
||||||
CachedReferenceField,
|
|
||||||
DecimalField,
|
|
||||||
Document,
|
|
||||||
EmbeddedDocument,
|
|
||||||
EmbeddedDocumentField,
|
|
||||||
InvalidDocumentError,
|
|
||||||
ListField,
|
|
||||||
ReferenceField,
|
|
||||||
StringField,
|
|
||||||
ValidationError,
|
|
||||||
)
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
class TestCachedReferenceField(MongoDBTestCase):
|
class TestCachedReferenceField(MongoDBTestCase):
|
||||||
def test_constructor_fail_bad_document_type(self):
|
|
||||||
with pytest.raises(
|
|
||||||
ValidationError, match="must be a document class or a string"
|
|
||||||
):
|
|
||||||
CachedReferenceField(document_type=0)
|
|
||||||
|
|
||||||
def test_get_and_save(self):
|
def test_get_and_save(self):
|
||||||
"""
|
"""
|
||||||
Tests #1047: CachedReferenceField creates DBRefs on to_python,
|
Tests #1047: CachedReferenceField creates DBRefs on to_python,
|
||||||
@@ -207,9 +190,9 @@ class TestCachedReferenceField(MongoDBTestCase):
|
|||||||
|
|
||||||
assert dict(a2.to_mongo()) == {
|
assert dict(a2.to_mongo()) == {
|
||||||
"_id": a2.pk,
|
"_id": a2.pk,
|
||||||
"name": "Wilson Junior",
|
"name": u"Wilson Junior",
|
||||||
"tp": "pf",
|
"tp": u"pf",
|
||||||
"father": {"_id": a1.pk, "tp": "pj"},
|
"father": {"_id": a1.pk, "tp": u"pj"},
|
||||||
}
|
}
|
||||||
|
|
||||||
assert Person.objects(father=a1)._query == {"father._id": a1.pk}
|
assert Person.objects(father=a1)._query == {"father._id": a1.pk}
|
||||||
@@ -221,9 +204,9 @@ class TestCachedReferenceField(MongoDBTestCase):
|
|||||||
a2.reload()
|
a2.reload()
|
||||||
assert dict(a2.to_mongo()) == {
|
assert dict(a2.to_mongo()) == {
|
||||||
"_id": a2.pk,
|
"_id": a2.pk,
|
||||||
"name": "Wilson Junior",
|
"name": u"Wilson Junior",
|
||||||
"tp": "pf",
|
"tp": u"pf",
|
||||||
"father": {"_id": a1.pk, "tp": "pf"},
|
"father": {"_id": a1.pk, "tp": u"pf"},
|
||||||
}
|
}
|
||||||
|
|
||||||
def test_cached_reference_fields_on_embedded_documents(self):
|
def test_cached_reference_fields_on_embedded_documents(self):
|
||||||
|
@@ -6,6 +6,7 @@ import re
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -59,7 +60,7 @@ class ComplexDateTimeFieldTest(MongoDBTestCase):
|
|||||||
assert log == log1
|
assert log == log1
|
||||||
|
|
||||||
# Test string padding
|
# Test string padding
|
||||||
microsecond = map(int, (math.pow(10, x) for x in range(6)))
|
microsecond = map(int, [math.pow(10, x) for x in range(6)])
|
||||||
mm = dd = hh = ii = ss = [1, 10]
|
mm = dd = hh = ii = ss = [1, 10]
|
||||||
|
|
||||||
for values in itertools.product([2014], mm, dd, hh, ii, ss, microsecond):
|
for values in itertools.product([2014], mm, dd, hh, ii, ss, microsecond):
|
||||||
|
@@ -9,6 +9,7 @@ except ImportError:
|
|||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine import connection
|
from mongoengine import connection
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
|
@@ -2,11 +2,60 @@ from decimal import Decimal
|
|||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import DecimalField, Document, ValidationError
|
from mongoengine import *
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
class TestDecimalField(MongoDBTestCase):
|
class TestDecimalField(MongoDBTestCase):
|
||||||
|
def test_validation(self):
|
||||||
|
"""Ensure that invalid values cannot be assigned to decimal fields.
|
||||||
|
"""
|
||||||
|
|
||||||
|
class Person(Document):
|
||||||
|
height = DecimalField(min_value=Decimal("0.1"), max_value=Decimal("3.5"))
|
||||||
|
|
||||||
|
Person.drop_collection()
|
||||||
|
|
||||||
|
Person(height=Decimal("1.89")).save()
|
||||||
|
person = Person.objects.first()
|
||||||
|
assert person.height == Decimal("1.89")
|
||||||
|
|
||||||
|
person.height = "2.0"
|
||||||
|
person.save()
|
||||||
|
person.height = 0.01
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
person.height = Decimal("0.01")
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
person.height = Decimal("4.0")
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
person.height = "something invalid"
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
|
||||||
|
person_2 = Person(height="something invalid")
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person_2.validate()
|
||||||
|
|
||||||
|
def test_comparison(self):
|
||||||
|
class Person(Document):
|
||||||
|
money = DecimalField()
|
||||||
|
|
||||||
|
Person.drop_collection()
|
||||||
|
|
||||||
|
Person(money=6).save()
|
||||||
|
Person(money=7).save()
|
||||||
|
Person(money=8).save()
|
||||||
|
Person(money=10).save()
|
||||||
|
|
||||||
|
assert 2 == Person.objects(money__gt=Decimal("7")).count()
|
||||||
|
assert 2 == Person.objects(money__gt=7).count()
|
||||||
|
assert 2 == Person.objects(money__gt="7").count()
|
||||||
|
|
||||||
|
assert 3 == Person.objects(money__gte="7").count()
|
||||||
|
|
||||||
def test_storage(self):
|
def test_storage(self):
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
float_value = DecimalField(precision=4)
|
float_value = DecimalField(precision=4)
|
||||||
@@ -58,63 +107,3 @@ class TestDecimalField(MongoDBTestCase):
|
|||||||
for field_name in ["float_value", "string_value"]:
|
for field_name in ["float_value", "string_value"]:
|
||||||
actual = list(Person.objects().scalar(field_name))
|
actual = list(Person.objects().scalar(field_name))
|
||||||
assert expected == actual
|
assert expected == actual
|
||||||
|
|
||||||
def test_save_none(self):
|
|
||||||
class Person(Document):
|
|
||||||
value = DecimalField()
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
|
|
||||||
person = Person(value=None)
|
|
||||||
assert person.value is None
|
|
||||||
person.save()
|
|
||||||
fetched_person = Person.objects.first()
|
|
||||||
fetched_person.value is None
|
|
||||||
|
|
||||||
def test_validation(self):
|
|
||||||
"""Ensure that invalid values cannot be assigned to decimal fields."""
|
|
||||||
|
|
||||||
class Person(Document):
|
|
||||||
height = DecimalField(min_value=Decimal("0.1"), max_value=Decimal("3.5"))
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
|
|
||||||
Person(height=Decimal("1.89")).save()
|
|
||||||
person = Person.objects.first()
|
|
||||||
assert person.height == Decimal("1.89")
|
|
||||||
|
|
||||||
person.height = "2.0"
|
|
||||||
person.save()
|
|
||||||
person.height = 0.01
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person.validate()
|
|
||||||
person.height = Decimal("0.01")
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person.validate()
|
|
||||||
person.height = Decimal("4.0")
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person.validate()
|
|
||||||
person.height = "something invalid"
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person.validate()
|
|
||||||
|
|
||||||
person_2 = Person(height="something invalid")
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person_2.validate()
|
|
||||||
|
|
||||||
def test_comparison(self):
|
|
||||||
class Person(Document):
|
|
||||||
money = DecimalField()
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
|
|
||||||
Person(money=6).save()
|
|
||||||
Person(money=7).save()
|
|
||||||
Person(money=8).save()
|
|
||||||
Person(money=10).save()
|
|
||||||
|
|
||||||
assert 2 == Person.objects(money__gt=Decimal("7")).count()
|
|
||||||
assert 2 == Person.objects(money__gt=7).count()
|
|
||||||
assert 2 == Person.objects(money__gt="7").count()
|
|
||||||
|
|
||||||
assert 3 == Person.objects(money__gte="7").count()
|
|
||||||
|
@@ -1,12 +1,10 @@
|
|||||||
import pytest
|
|
||||||
from bson import InvalidDocument
|
from bson import InvalidDocument
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.base import BaseDict
|
from mongoengine.base import BaseDict
|
||||||
from mongoengine.mongodb_support import (
|
from mongoengine.mongodb_support import MONGODB_36, get_mongodb_version
|
||||||
MONGODB_36,
|
|
||||||
get_mongodb_version,
|
|
||||||
)
|
|
||||||
from tests.utils import MongoDBTestCase, get_as_pymongo
|
from tests.utils import MongoDBTestCase, get_as_pymongo
|
||||||
|
|
||||||
|
|
||||||
|
@@ -1,6 +1,8 @@
|
|||||||
|
import sys
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import Document, EmailField, ValidationError
|
from mongoengine import *
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -28,11 +30,11 @@ class TestEmailField(MongoDBTestCase):
|
|||||||
user.validate()
|
user.validate()
|
||||||
|
|
||||||
# unicode domain
|
# unicode domain
|
||||||
user = User(email="user@пример.рф")
|
user = User(email=u"user@пример.рф")
|
||||||
user.validate()
|
user.validate()
|
||||||
|
|
||||||
# invalid unicode domain
|
# invalid unicode domain
|
||||||
user = User(email="user@пример")
|
user = User(email=u"user@пример")
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
user.validate()
|
user.validate()
|
||||||
|
|
||||||
@@ -46,7 +48,7 @@ class TestEmailField(MongoDBTestCase):
|
|||||||
email = EmailField()
|
email = EmailField()
|
||||||
|
|
||||||
# unicode user shouldn't validate by default...
|
# unicode user shouldn't validate by default...
|
||||||
user = User(email="Dörte@Sörensen.example.com")
|
user = User(email=u"Dörte@Sörensen.example.com")
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
user.validate()
|
user.validate()
|
||||||
|
|
||||||
@@ -54,7 +56,7 @@ class TestEmailField(MongoDBTestCase):
|
|||||||
class User(Document):
|
class User(Document):
|
||||||
email = EmailField(allow_utf8_user=True)
|
email = EmailField(allow_utf8_user=True)
|
||||||
|
|
||||||
user = User(email="Dörte@Sörensen.example.com")
|
user = User(email=u"Dörte@Sörensen.example.com")
|
||||||
user.validate()
|
user.validate()
|
||||||
|
|
||||||
def test_email_field_domain_whitelist(self):
|
def test_email_field_domain_whitelist(self):
|
||||||
|
@@ -12,6 +12,7 @@ from mongoengine import (
|
|||||||
StringField,
|
StringField,
|
||||||
ValidationError,
|
ValidationError,
|
||||||
)
|
)
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -73,7 +74,7 @@ class TestEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
Person.objects(settings__notexist="bar").first()
|
Person.objects(settings__notexist="bar").first()
|
||||||
assert str(exc_info.value) == 'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
with pytest.raises(LookUpError):
|
with pytest.raises(LookUpError):
|
||||||
Person.objects.only("settings.notexist")
|
Person.objects.only("settings.notexist")
|
||||||
@@ -109,7 +110,7 @@ class TestEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
assert Person.objects(settings__notexist="bar").first().id == p.id
|
assert Person.objects(settings__notexist="bar").first().id == p.id
|
||||||
assert str(exc_info.value) == 'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
# Test existing attribute
|
# Test existing attribute
|
||||||
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
||||||
@@ -317,7 +318,7 @@ class TestGenericEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
Person.objects(settings__notexist="bar").first()
|
Person.objects(settings__notexist="bar").first()
|
||||||
assert str(exc_info.value) == 'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
with pytest.raises(LookUpError):
|
with pytest.raises(LookUpError):
|
||||||
Person.objects.only("settings.notexist")
|
Person.objects.only("settings.notexist")
|
||||||
@@ -345,7 +346,7 @@ class TestGenericEmbeddedDocumentField(MongoDBTestCase):
|
|||||||
# Test non exiting attribute
|
# Test non exiting attribute
|
||||||
with pytest.raises(InvalidQueryError) as exc_info:
|
with pytest.raises(InvalidQueryError) as exc_info:
|
||||||
assert Person.objects(settings__notexist="bar").first().id == p.id
|
assert Person.objects(settings__notexist="bar").first().id == p.id
|
||||||
assert str(exc_info.value) == 'Cannot resolve field "notexist"'
|
assert str(exc_info.value) == u'Cannot resolve field "notexist"'
|
||||||
|
|
||||||
# Test existing attribute
|
# Test existing attribute
|
||||||
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
assert Person.objects(settings__base_foo="basefoo").first().id == p.id
|
||||||
|
@@ -1,9 +1,9 @@
|
|||||||
from enum import Enum
|
from enum import Enum
|
||||||
|
|
||||||
import pytest
|
|
||||||
from bson import InvalidDocument
|
from bson import InvalidDocument
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import Document, EnumField, ValidationError
|
from mongoengine import *
|
||||||
from tests.utils import MongoDBTestCase, get_as_pymongo
|
from tests.utils import MongoDBTestCase, get_as_pymongo
|
||||||
|
|
||||||
|
|
||||||
@@ -45,11 +45,6 @@ class TestStringEnumField(MongoDBTestCase):
|
|||||||
m.save()
|
m.save()
|
||||||
assert m.status == Status.DONE
|
assert m.status == Status.DONE
|
||||||
|
|
||||||
m.status = "wrong"
|
|
||||||
assert m.status == "wrong"
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
m.validate()
|
|
||||||
|
|
||||||
def test_set_default(self):
|
def test_set_default(self):
|
||||||
class ModelWithDefault(Document):
|
class ModelWithDefault(Document):
|
||||||
status = EnumField(Status, default=Status.DONE)
|
status = EnumField(Status, default=Status.DONE)
|
||||||
|
@@ -1,8 +1,8 @@
|
|||||||
import datetime
|
import datetime
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
|
from bson import DBRef, ObjectId, SON
|
||||||
import pytest
|
import pytest
|
||||||
from bson import SON, DBRef, ObjectId
|
|
||||||
|
|
||||||
from mongoengine import (
|
from mongoengine import (
|
||||||
BooleanField,
|
BooleanField,
|
||||||
@@ -34,12 +34,9 @@ from mongoengine import (
|
|||||||
StringField,
|
StringField,
|
||||||
ValidationError,
|
ValidationError,
|
||||||
)
|
)
|
||||||
from mongoengine.base import (
|
from mongoengine.base import BaseField, EmbeddedDocumentList, _document_registry
|
||||||
BaseField,
|
|
||||||
EmbeddedDocumentList,
|
|
||||||
_document_registry,
|
|
||||||
)
|
|
||||||
from mongoengine.errors import DeprecatedError
|
from mongoengine.errors import DeprecatedError
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -295,7 +292,7 @@ class TestField(MongoDBTestCase):
|
|||||||
HandleNoneFields.drop_collection()
|
HandleNoneFields.drop_collection()
|
||||||
|
|
||||||
doc = HandleNoneFields()
|
doc = HandleNoneFields()
|
||||||
doc.str_fld = "spam ham egg"
|
doc.str_fld = u"spam ham egg"
|
||||||
doc.int_fld = 42
|
doc.int_fld = 42
|
||||||
doc.flt_fld = 4.2
|
doc.flt_fld = 4.2
|
||||||
doc.com_dt_fld = datetime.datetime.utcnow()
|
doc.com_dt_fld = datetime.datetime.utcnow()
|
||||||
@@ -309,7 +306,7 @@ class TestField(MongoDBTestCase):
|
|||||||
)
|
)
|
||||||
assert res == 1
|
assert res == 1
|
||||||
|
|
||||||
# Retrieve data from db and verify it.
|
# Retrive data from db and verify it.
|
||||||
ret = HandleNoneFields.objects.all()[0]
|
ret = HandleNoneFields.objects.all()[0]
|
||||||
assert ret.str_fld is None
|
assert ret.str_fld is None
|
||||||
assert ret.int_fld is None
|
assert ret.int_fld is None
|
||||||
@@ -331,7 +328,7 @@ class TestField(MongoDBTestCase):
|
|||||||
HandleNoneFields.drop_collection()
|
HandleNoneFields.drop_collection()
|
||||||
|
|
||||||
doc = HandleNoneFields()
|
doc = HandleNoneFields()
|
||||||
doc.str_fld = "spam ham egg"
|
doc.str_fld = u"spam ham egg"
|
||||||
doc.int_fld = 42
|
doc.int_fld = 42
|
||||||
doc.flt_fld = 4.2
|
doc.flt_fld = 4.2
|
||||||
doc.comp_dt_fld = datetime.datetime.utcnow()
|
doc.comp_dt_fld = datetime.datetime.utcnow()
|
||||||
@@ -343,7 +340,7 @@ class TestField(MongoDBTestCase):
|
|||||||
{"$unset": {"str_fld": 1, "int_fld": 1, "flt_fld": 1, "comp_dt_fld": 1}},
|
{"$unset": {"str_fld": 1, "int_fld": 1, "flt_fld": 1, "comp_dt_fld": 1}},
|
||||||
)
|
)
|
||||||
|
|
||||||
# Retrieve data from db and verify it.
|
# Retrive data from db and verify it.
|
||||||
ret = HandleNoneFields.objects.first()
|
ret = HandleNoneFields.objects.first()
|
||||||
assert ret.str_fld is None
|
assert ret.str_fld is None
|
||||||
assert ret.int_fld is None
|
assert ret.int_fld is None
|
||||||
@@ -377,6 +374,34 @@ class TestField(MongoDBTestCase):
|
|||||||
person.id = str(ObjectId())
|
person.id = str(ObjectId())
|
||||||
person.validate()
|
person.validate()
|
||||||
|
|
||||||
|
def test_string_validation(self):
|
||||||
|
"""Ensure that invalid values cannot be assigned to string fields."""
|
||||||
|
|
||||||
|
class Person(Document):
|
||||||
|
name = StringField(max_length=20)
|
||||||
|
userid = StringField(r"[0-9a-z_]+$")
|
||||||
|
|
||||||
|
person = Person(name=34)
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
|
||||||
|
# Test regex validation on userid
|
||||||
|
person = Person(userid="test.User")
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
|
||||||
|
person.userid = "test_user"
|
||||||
|
assert person.userid == "test_user"
|
||||||
|
person.validate()
|
||||||
|
|
||||||
|
# Test max length validation on name
|
||||||
|
person = Person(name="Name that is more than twenty characters")
|
||||||
|
with pytest.raises(ValidationError):
|
||||||
|
person.validate()
|
||||||
|
|
||||||
|
person.name = "Shorter name"
|
||||||
|
person.validate()
|
||||||
|
|
||||||
def test_db_field_validation(self):
|
def test_db_field_validation(self):
|
||||||
"""Ensure that db_field doesn't accept invalid values."""
|
"""Ensure that db_field doesn't accept invalid values."""
|
||||||
|
|
||||||
@@ -401,9 +426,9 @@ class TestField(MongoDBTestCase):
|
|||||||
def test_list_validation(self):
|
def test_list_validation(self):
|
||||||
"""Ensure that a list field only accepts lists with valid elements."""
|
"""Ensure that a list field only accepts lists with valid elements."""
|
||||||
access_level_choices = (
|
access_level_choices = (
|
||||||
("a", "Administration"),
|
("a", u"Administration"),
|
||||||
("b", "Manager"),
|
("b", u"Manager"),
|
||||||
("c", "Staff"),
|
("c", u"Staff"),
|
||||||
)
|
)
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
@@ -451,7 +476,7 @@ class TestField(MongoDBTestCase):
|
|||||||
post.access_list = ["a", "b"]
|
post.access_list = ["a", "b"]
|
||||||
post.validate()
|
post.validate()
|
||||||
|
|
||||||
assert post.get_access_list_display() == "Administration, Manager"
|
assert post.get_access_list_display() == u"Administration, Manager"
|
||||||
|
|
||||||
post.comments = ["a"]
|
post.comments = ["a"]
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
@@ -519,7 +544,8 @@ class TestField(MongoDBTestCase):
|
|||||||
post.validate()
|
post.validate()
|
||||||
|
|
||||||
def test_sorted_list_sorting(self):
|
def test_sorted_list_sorting(self):
|
||||||
"""Ensure that a sorted list field properly sorts values."""
|
"""Ensure that a sorted list field properly sorts values.
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
order = IntField()
|
order = IntField()
|
||||||
@@ -635,7 +661,8 @@ class TestField(MongoDBTestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_list_field_manipulative_operators(self):
|
def test_list_field_manipulative_operators(self):
|
||||||
"""Ensure that ListField works with standard list operators that manipulate the list."""
|
"""Ensure that ListField works with standard list operators that manipulate the list.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
ref = StringField()
|
ref = StringField()
|
||||||
@@ -1332,9 +1359,9 @@ class TestField(MongoDBTestCase):
|
|||||||
foo.delete()
|
foo.delete()
|
||||||
bar = Bar.objects.get()
|
bar = Bar.objects.get()
|
||||||
with pytest.raises(DoesNotExist):
|
with pytest.raises(DoesNotExist):
|
||||||
bar.ref
|
getattr(bar, "ref")
|
||||||
with pytest.raises(DoesNotExist):
|
with pytest.raises(DoesNotExist):
|
||||||
bar.generic_ref
|
getattr(bar, "generic_ref")
|
||||||
|
|
||||||
# When auto_dereference is disabled, there is no trouble returning DBRef
|
# When auto_dereference is disabled, there is no trouble returning DBRef
|
||||||
bar = Bar.objects.get()
|
bar = Bar.objects.get()
|
||||||
@@ -1345,7 +1372,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert bar.generic_ref == {"_ref": expected, "_cls": "Foo"}
|
assert bar.generic_ref == {"_ref": expected, "_cls": "Foo"}
|
||||||
|
|
||||||
def test_list_item_dereference(self):
|
def test_list_item_dereference(self):
|
||||||
"""Ensure that DBRef items in ListFields are dereferenced."""
|
"""Ensure that DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1370,7 +1398,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert group_obj.members[1].name == user2.name
|
assert group_obj.members[1].name == user2.name
|
||||||
|
|
||||||
def test_recursive_reference(self):
|
def test_recursive_reference(self):
|
||||||
"""Ensure that ReferenceFields can reference their own documents."""
|
"""Ensure that ReferenceFields can reference their own documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class Employee(Document):
|
class Employee(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1397,7 +1426,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert peter.friends == friends
|
assert peter.friends == friends
|
||||||
|
|
||||||
def test_recursive_embedding(self):
|
def test_recursive_embedding(self):
|
||||||
"""Ensure that EmbeddedDocumentFields can contain their own documents."""
|
"""Ensure that EmbeddedDocumentFields can contain their own documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class TreeNode(EmbeddedDocument):
|
class TreeNode(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1473,7 +1503,8 @@ class TestField(MongoDBTestCase):
|
|||||||
AbstractDoc.drop_collection()
|
AbstractDoc.drop_collection()
|
||||||
|
|
||||||
def test_reference_class_with_abstract_parent(self):
|
def test_reference_class_with_abstract_parent(self):
|
||||||
"""Ensure that a class with an abstract parent can be referenced."""
|
"""Ensure that a class with an abstract parent can be referenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class Sibling(Document):
|
class Sibling(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1543,7 +1574,8 @@ class TestField(MongoDBTestCase):
|
|||||||
brother.save()
|
brother.save()
|
||||||
|
|
||||||
def test_generic_reference(self):
|
def test_generic_reference(self):
|
||||||
"""Ensure that a GenericReferenceField properly dereferences items."""
|
"""Ensure that a GenericReferenceField properly dereferences items.
|
||||||
|
"""
|
||||||
|
|
||||||
class Link(Document):
|
class Link(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -1582,7 +1614,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert isinstance(bm.bookmark_object, Link)
|
assert isinstance(bm.bookmark_object, Link)
|
||||||
|
|
||||||
def test_generic_reference_list(self):
|
def test_generic_reference_list(self):
|
||||||
"""Ensure that a ListField properly dereferences generic references."""
|
"""Ensure that a ListField properly dereferences generic references.
|
||||||
|
"""
|
||||||
|
|
||||||
class Link(Document):
|
class Link(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -1685,7 +1718,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert bm.bookmark_object == post_1
|
assert bm.bookmark_object == post_1
|
||||||
|
|
||||||
def test_generic_reference_string_choices(self):
|
def test_generic_reference_string_choices(self):
|
||||||
"""Ensure that a GenericReferenceField can handle choices as strings"""
|
"""Ensure that a GenericReferenceField can handle choices as strings
|
||||||
|
"""
|
||||||
|
|
||||||
class Link(Document):
|
class Link(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -1777,7 +1811,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert user.bookmarks == [post_1]
|
assert user.bookmarks == [post_1]
|
||||||
|
|
||||||
def test_generic_reference_list_item_modification(self):
|
def test_generic_reference_list_item_modification(self):
|
||||||
"""Ensure that modifications of related documents (through generic reference) don't influence on querying"""
|
"""Ensure that modifications of related documents (through generic reference) don't influence on querying
|
||||||
|
"""
|
||||||
|
|
||||||
class Post(Document):
|
class Post(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -1865,7 +1900,8 @@ class TestField(MongoDBTestCase):
|
|||||||
assert doc == doc2
|
assert doc == doc2
|
||||||
|
|
||||||
def test_choices_allow_using_sets_as_choices(self):
|
def test_choices_allow_using_sets_as_choices(self):
|
||||||
"""Ensure that sets can be used when setting choices"""
|
"""Ensure that sets can be used when setting choices
|
||||||
|
"""
|
||||||
|
|
||||||
class Shirt(Document):
|
class Shirt(Document):
|
||||||
size = StringField(choices={"M", "L"})
|
size = StringField(choices={"M", "L"})
|
||||||
@@ -1884,7 +1920,8 @@ class TestField(MongoDBTestCase):
|
|||||||
shirt.validate()
|
shirt.validate()
|
||||||
|
|
||||||
def test_choices_validation_accept_possible_value(self):
|
def test_choices_validation_accept_possible_value(self):
|
||||||
"""Ensure that value is in a container of allowed values."""
|
"""Ensure that value is in a container of allowed values.
|
||||||
|
"""
|
||||||
|
|
||||||
class Shirt(Document):
|
class Shirt(Document):
|
||||||
size = StringField(choices=("S", "M"))
|
size = StringField(choices=("S", "M"))
|
||||||
@@ -1893,7 +1930,8 @@ class TestField(MongoDBTestCase):
|
|||||||
shirt.validate()
|
shirt.validate()
|
||||||
|
|
||||||
def test_choices_validation_reject_unknown_value(self):
|
def test_choices_validation_reject_unknown_value(self):
|
||||||
"""Ensure that unallowed value are rejected upon validation"""
|
"""Ensure that unallowed value are rejected upon validation
|
||||||
|
"""
|
||||||
|
|
||||||
class Shirt(Document):
|
class Shirt(Document):
|
||||||
size = StringField(choices=("S", "M"))
|
size = StringField(choices=("S", "M"))
|
||||||
@@ -1951,7 +1989,8 @@ class TestField(MongoDBTestCase):
|
|||||||
shirt1.validate()
|
shirt1.validate()
|
||||||
|
|
||||||
def test_simple_choices_validation(self):
|
def test_simple_choices_validation(self):
|
||||||
"""Ensure that value is in a container of allowed values."""
|
"""Ensure that value is in a container of allowed values.
|
||||||
|
"""
|
||||||
|
|
||||||
class Shirt(Document):
|
class Shirt(Document):
|
||||||
size = StringField(max_length=3, choices=("S", "M", "L", "XL", "XXL"))
|
size = StringField(max_length=3, choices=("S", "M", "L", "XL", "XXL"))
|
||||||
@@ -2000,11 +2039,12 @@ class TestField(MongoDBTestCase):
|
|||||||
shirt.validate()
|
shirt.validate()
|
||||||
|
|
||||||
def test_simple_choices_validation_invalid_value(self):
|
def test_simple_choices_validation_invalid_value(self):
|
||||||
"""Ensure that error messages are correct."""
|
"""Ensure that error messages are correct.
|
||||||
|
"""
|
||||||
SIZES = ("S", "M", "L", "XL", "XXL")
|
SIZES = ("S", "M", "L", "XL", "XXL")
|
||||||
COLORS = (("R", "Red"), ("B", "Blue"))
|
COLORS = (("R", "Red"), ("B", "Blue"))
|
||||||
SIZE_MESSAGE = "Value must be one of ('S', 'M', 'L', 'XL', 'XXL')"
|
SIZE_MESSAGE = u"Value must be one of ('S', 'M', 'L', 'XL', 'XXL')"
|
||||||
COLOR_MESSAGE = "Value must be one of ['R', 'B']"
|
COLOR_MESSAGE = u"Value must be one of ['R', 'B']"
|
||||||
|
|
||||||
class Shirt(Document):
|
class Shirt(Document):
|
||||||
size = StringField(max_length=3, choices=SIZES)
|
size = StringField(max_length=3, choices=SIZES)
|
||||||
@@ -2067,7 +2107,7 @@ class TestField(MongoDBTestCase):
|
|||||||
assert "comments" in error_dict
|
assert "comments" in error_dict
|
||||||
assert 1 in error_dict["comments"]
|
assert 1 in error_dict["comments"]
|
||||||
assert "content" in error_dict["comments"][1]
|
assert "content" in error_dict["comments"][1]
|
||||||
assert error_dict["comments"][1]["content"] == "Field is required"
|
assert error_dict["comments"][1]["content"] == u"Field is required"
|
||||||
|
|
||||||
post.comments[1].content = "here we go"
|
post.comments[1].content = "here we go"
|
||||||
post.validate()
|
post.validate()
|
||||||
@@ -2079,7 +2119,7 @@ class TestField(MongoDBTestCase):
|
|||||||
|
|
||||||
class EnumField(BaseField):
|
class EnumField(BaseField):
|
||||||
def __init__(self, **kwargs):
|
def __init__(self, **kwargs):
|
||||||
super().__init__(**kwargs)
|
super(EnumField, self).__init__(**kwargs)
|
||||||
|
|
||||||
def to_mongo(self, value):
|
def to_mongo(self, value):
|
||||||
return value
|
return value
|
||||||
@@ -2232,13 +2272,6 @@ class TestField(MongoDBTestCase):
|
|||||||
with pytest.raises(FieldDoesNotExist):
|
with pytest.raises(FieldDoesNotExist):
|
||||||
Doc(bar="test")
|
Doc(bar="test")
|
||||||
|
|
||||||
def test_undefined_field_works_no_confusion_with_db_field(self):
|
|
||||||
class Doc(Document):
|
|
||||||
foo = StringField(db_field="bar")
|
|
||||||
|
|
||||||
with pytest.raises(FieldDoesNotExist):
|
|
||||||
Doc(bar="test")
|
|
||||||
|
|
||||||
|
|
||||||
class TestEmbeddedDocumentListField(MongoDBTestCase):
|
class TestEmbeddedDocumentListField(MongoDBTestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
@@ -2581,11 +2614,11 @@ class TestEmbeddedDocumentListField(MongoDBTestCase):
|
|||||||
"""
|
"""
|
||||||
post = self.BlogPost(
|
post = self.BlogPost(
|
||||||
comments=[
|
comments=[
|
||||||
self.Comments(author="user1", message="сообщение"),
|
self.Comments(author="user1", message=u"сообщение"),
|
||||||
self.Comments(author="user2", message="хабарлама"),
|
self.Comments(author="user2", message=u"хабарлама"),
|
||||||
]
|
]
|
||||||
).save()
|
).save()
|
||||||
assert post.comments.get(message="сообщение").author == "user1"
|
assert post.comments.get(message=u"сообщение").author == "user1"
|
||||||
|
|
||||||
def test_save(self):
|
def test_save(self):
|
||||||
"""
|
"""
|
||||||
|
@@ -11,7 +11,7 @@ from mongoengine import *
|
|||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from PIL import Image # noqa: F401
|
from PIL import Image
|
||||||
|
|
||||||
HAS_PIL = True
|
HAS_PIL = True
|
||||||
except ImportError:
|
except ImportError:
|
||||||
@@ -48,14 +48,15 @@ class TestFileField(MongoDBTestCase):
|
|||||||
DemoFile.objects.create()
|
DemoFile.objects.create()
|
||||||
|
|
||||||
def test_file_fields(self):
|
def test_file_fields(self):
|
||||||
"""Ensure that file fields can be written to and their data retrieved"""
|
"""Ensure that file fields can be written to and their data retrieved
|
||||||
|
"""
|
||||||
|
|
||||||
class PutFile(Document):
|
class PutFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
|
|
||||||
PutFile.drop_collection()
|
PutFile.drop_collection()
|
||||||
|
|
||||||
text = b"Hello, World!"
|
text = "Hello, World!".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
putfile = PutFile()
|
putfile = PutFile()
|
||||||
@@ -90,15 +91,16 @@ class TestFileField(MongoDBTestCase):
|
|||||||
result.the_file.delete()
|
result.the_file.delete()
|
||||||
|
|
||||||
def test_file_fields_stream(self):
|
def test_file_fields_stream(self):
|
||||||
"""Ensure that file fields can be written to and their data retrieved"""
|
"""Ensure that file fields can be written to and their data retrieved
|
||||||
|
"""
|
||||||
|
|
||||||
class StreamFile(Document):
|
class StreamFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
|
|
||||||
StreamFile.drop_collection()
|
StreamFile.drop_collection()
|
||||||
|
|
||||||
text = b"Hello, World!"
|
text = "Hello, World!".encode("latin-1")
|
||||||
more_text = b"Foo Bar"
|
more_text = "Foo Bar".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
streamfile = StreamFile()
|
streamfile = StreamFile()
|
||||||
@@ -133,8 +135,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
StreamFile.drop_collection()
|
StreamFile.drop_collection()
|
||||||
|
|
||||||
text = b"Hello, World!"
|
text = "Hello, World!".encode("latin-1")
|
||||||
more_text = b"Foo Bar"
|
more_text = "Foo Bar".encode("latin-1")
|
||||||
|
|
||||||
streamfile = StreamFile()
|
streamfile = StreamFile()
|
||||||
streamfile.save()
|
streamfile.save()
|
||||||
@@ -163,8 +165,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
class SetFile(Document):
|
class SetFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
|
|
||||||
text = b"Hello, World!"
|
text = "Hello, World!".encode("latin-1")
|
||||||
more_text = b"Foo Bar"
|
more_text = "Foo Bar".encode("latin-1")
|
||||||
|
|
||||||
SetFile.drop_collection()
|
SetFile.drop_collection()
|
||||||
|
|
||||||
@@ -192,7 +194,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
GridDocument.drop_collection()
|
GridDocument.drop_collection()
|
||||||
|
|
||||||
with tempfile.TemporaryFile() as f:
|
with tempfile.TemporaryFile() as f:
|
||||||
f.write(b"Hello World!")
|
f.write("Hello World!".encode("latin-1"))
|
||||||
f.flush()
|
f.flush()
|
||||||
|
|
||||||
# Test without default
|
# Test without default
|
||||||
@@ -209,7 +211,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
assert doc_b.the_file.grid_id == doc_c.the_file.grid_id
|
assert doc_b.the_file.grid_id == doc_c.the_file.grid_id
|
||||||
|
|
||||||
# Test with default
|
# Test with default
|
||||||
doc_d = GridDocument(the_file=b"")
|
doc_d = GridDocument(the_file="".encode("latin-1"))
|
||||||
doc_d.save()
|
doc_d.save()
|
||||||
|
|
||||||
doc_e = GridDocument.objects.with_id(doc_d.id)
|
doc_e = GridDocument.objects.with_id(doc_d.id)
|
||||||
@@ -226,7 +228,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
assert ["doc_b", "doc_e"] == grid_fs.list()
|
assert ["doc_b", "doc_e"] == grid_fs.list()
|
||||||
|
|
||||||
def test_file_uniqueness(self):
|
def test_file_uniqueness(self):
|
||||||
"""Ensure that each instance of a FileField is unique"""
|
"""Ensure that each instance of a FileField is unique
|
||||||
|
"""
|
||||||
|
|
||||||
class TestFile(Document):
|
class TestFile(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -235,7 +238,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
# First instance
|
# First instance
|
||||||
test_file = TestFile()
|
test_file = TestFile()
|
||||||
test_file.name = "Hello, World!"
|
test_file.name = "Hello, World!"
|
||||||
test_file.the_file.put(b"Hello, World!")
|
test_file.the_file.put("Hello, World!".encode("latin-1"))
|
||||||
test_file.save()
|
test_file.save()
|
||||||
|
|
||||||
# Second instance
|
# Second instance
|
||||||
@@ -282,7 +285,8 @@ class TestFileField(MongoDBTestCase):
|
|||||||
assert test_file.the_file.get().length == 4971
|
assert test_file.the_file.get().length == 4971
|
||||||
|
|
||||||
def test_file_boolean(self):
|
def test_file_boolean(self):
|
||||||
"""Ensure that a boolean test of a FileField indicates its presence"""
|
"""Ensure that a boolean test of a FileField indicates its presence
|
||||||
|
"""
|
||||||
|
|
||||||
class TestFile(Document):
|
class TestFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
@@ -291,7 +295,9 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
test_file = TestFile()
|
test_file = TestFile()
|
||||||
assert not bool(test_file.the_file)
|
assert not bool(test_file.the_file)
|
||||||
test_file.the_file.put(b"Hello, World!", content_type="text/plain")
|
test_file.the_file.put(
|
||||||
|
"Hello, World!".encode("latin-1"), content_type="text/plain"
|
||||||
|
)
|
||||||
test_file.save()
|
test_file.save()
|
||||||
assert bool(test_file.the_file)
|
assert bool(test_file.the_file)
|
||||||
|
|
||||||
@@ -308,12 +314,12 @@ class TestFileField(MongoDBTestCase):
|
|||||||
assert test_file.the_file not in [{"test": 1}]
|
assert test_file.the_file not in [{"test": 1}]
|
||||||
|
|
||||||
def test_file_disk_space(self):
|
def test_file_disk_space(self):
|
||||||
"""Test disk space usage when we delete/replace a file"""
|
""" Test disk space usage when we delete/replace a file """
|
||||||
|
|
||||||
class TestFile(Document):
|
class TestFile(Document):
|
||||||
the_file = FileField()
|
the_file = FileField()
|
||||||
|
|
||||||
text = b"Hello, World!"
|
text = "Hello, World!".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
testfile = TestFile()
|
testfile = TestFile()
|
||||||
@@ -357,7 +363,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
testfile.the_file.put(text, content_type=content_type, filename="hello")
|
testfile.the_file.put(text, content_type=content_type, filename="hello")
|
||||||
testfile.save()
|
testfile.save()
|
||||||
|
|
||||||
text = b"Bonjour, World!"
|
text = "Bonjour, World!".encode("latin-1")
|
||||||
testfile.the_file.replace(text, content_type=content_type, filename="hello")
|
testfile.the_file.replace(text, content_type=content_type, filename="hello")
|
||||||
testfile.save()
|
testfile.save()
|
||||||
|
|
||||||
@@ -381,7 +387,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
TestImage.drop_collection()
|
TestImage.drop_collection()
|
||||||
|
|
||||||
with tempfile.TemporaryFile() as f:
|
with tempfile.TemporaryFile() as f:
|
||||||
f.write(b"Hello World!")
|
f.write("Hello World!".encode("latin-1"))
|
||||||
f.flush()
|
f.flush()
|
||||||
|
|
||||||
t = TestImage()
|
t = TestImage()
|
||||||
@@ -423,7 +429,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
@require_pil
|
@require_pil
|
||||||
def test_image_field_resize(self):
|
def test_image_field_resize(self):
|
||||||
class TestImage(Document):
|
class TestImage(Document):
|
||||||
image = ImageField(size=(185, 37, True))
|
image = ImageField(size=(185, 37))
|
||||||
|
|
||||||
TestImage.drop_collection()
|
TestImage.drop_collection()
|
||||||
|
|
||||||
@@ -465,7 +471,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
@require_pil
|
@require_pil
|
||||||
def test_image_field_thumbnail(self):
|
def test_image_field_thumbnail(self):
|
||||||
class TestImage(Document):
|
class TestImage(Document):
|
||||||
image = ImageField(thumbnail_size=(92, 18, True))
|
image = ImageField(thumbnail_size=(92, 18))
|
||||||
|
|
||||||
TestImage.drop_collection()
|
TestImage.drop_collection()
|
||||||
|
|
||||||
@@ -497,21 +503,21 @@ class TestFileField(MongoDBTestCase):
|
|||||||
# First instance
|
# First instance
|
||||||
test_file = TestFile()
|
test_file = TestFile()
|
||||||
test_file.name = "Hello, World!"
|
test_file.name = "Hello, World!"
|
||||||
test_file.the_file.put(b"Hello, World!", name="hello.txt")
|
test_file.the_file.put("Hello, World!".encode("latin-1"), name="hello.txt")
|
||||||
test_file.save()
|
test_file.save()
|
||||||
|
|
||||||
data = get_db("test_files").macumba.files.find_one()
|
data = get_db("test_files").macumba.files.find_one()
|
||||||
assert data.get("name") == "hello.txt"
|
assert data.get("name") == "hello.txt"
|
||||||
|
|
||||||
test_file = TestFile.objects.first()
|
test_file = TestFile.objects.first()
|
||||||
assert test_file.the_file.read() == b"Hello, World!"
|
assert test_file.the_file.read() == "Hello, World!".encode("latin-1")
|
||||||
|
|
||||||
test_file = TestFile.objects.first()
|
test_file = TestFile.objects.first()
|
||||||
test_file.the_file = b"Hello, World!"
|
test_file.the_file = "Hello, World!".encode("latin-1")
|
||||||
test_file.save()
|
test_file.save()
|
||||||
|
|
||||||
test_file = TestFile.objects.first()
|
test_file = TestFile.objects.first()
|
||||||
assert test_file.the_file.read() == b"Hello, World!"
|
assert test_file.the_file.read() == "Hello, World!".encode("latin-1")
|
||||||
|
|
||||||
def test_copyable(self):
|
def test_copyable(self):
|
||||||
class PutFile(Document):
|
class PutFile(Document):
|
||||||
@@ -519,7 +525,7 @@ class TestFileField(MongoDBTestCase):
|
|||||||
|
|
||||||
PutFile.drop_collection()
|
PutFile.drop_collection()
|
||||||
|
|
||||||
text = b"Hello, World!"
|
text = "Hello, World!".encode("latin-1")
|
||||||
content_type = "text/plain"
|
content_type = "text/plain"
|
||||||
|
|
||||||
putfile = PutFile()
|
putfile = PutFile()
|
||||||
|
@@ -1,6 +1,7 @@
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -18,7 +19,8 @@ class TestFloatField(MongoDBTestCase):
|
|||||||
assert 1 == TestDocument.objects(float_fld__ne=1).count()
|
assert 1 == TestDocument.objects(float_fld__ne=1).count()
|
||||||
|
|
||||||
def test_validation(self):
|
def test_validation(self):
|
||||||
"""Ensure that invalid values cannot be assigned to float fields."""
|
"""Ensure that invalid values cannot be assigned to float fields.
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
height = FloatField(min_value=0.1, max_value=3.5)
|
height = FloatField(min_value=0.1, max_value=3.5)
|
||||||
|
@@ -8,7 +8,7 @@ class TestGeoField(MongoDBTestCase):
|
|||||||
def _test_for_expected_error(self, Cls, loc, expected):
|
def _test_for_expected_error(self, Cls, loc, expected):
|
||||||
try:
|
try:
|
||||||
Cls(loc=loc).validate()
|
Cls(loc=loc).validate()
|
||||||
self.fail(f"Should not validate the location {loc}")
|
self.fail("Should not validate the location {0}".format(loc))
|
||||||
except ValidationError as e:
|
except ValidationError as e:
|
||||||
assert expected == e.to_dict()["loc"]
|
assert expected == e.to_dict()["loc"]
|
||||||
|
|
||||||
@@ -290,7 +290,8 @@ class TestGeoField(MongoDBTestCase):
|
|||||||
Location(loc=[[[[1, 2], [3, 4], [5, 6], [1, 2]]]]).validate()
|
Location(loc=[[[[1, 2], [3, 4], [5, 6], [1, 2]]]]).validate()
|
||||||
|
|
||||||
def test_indexes_geopoint(self):
|
def test_indexes_geopoint(self):
|
||||||
"""Ensure that indexes are created automatically for GeoPointFields."""
|
"""Ensure that indexes are created automatically for GeoPointFields.
|
||||||
|
"""
|
||||||
|
|
||||||
class Event(Document):
|
class Event(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -316,7 +317,8 @@ class TestGeoField(MongoDBTestCase):
|
|||||||
assert geo_indicies == [{"fields": [("venue.location", "2d")]}]
|
assert geo_indicies == [{"fields": [("venue.location", "2d")]}]
|
||||||
|
|
||||||
def test_indexes_2dsphere(self):
|
def test_indexes_2dsphere(self):
|
||||||
"""Ensure that indexes are created automatically for GeoPointFields."""
|
"""Ensure that indexes are created automatically for GeoPointFields.
|
||||||
|
"""
|
||||||
|
|
||||||
class Event(Document):
|
class Event(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -330,7 +332,8 @@ class TestGeoField(MongoDBTestCase):
|
|||||||
assert {"fields": [("point", "2dsphere")]} in geo_indicies
|
assert {"fields": [("point", "2dsphere")]} in geo_indicies
|
||||||
|
|
||||||
def test_indexes_2dsphere_embedded(self):
|
def test_indexes_2dsphere_embedded(self):
|
||||||
"""Ensure that indexes are created automatically for GeoPointFields."""
|
"""Ensure that indexes are created automatically for GeoPointFields.
|
||||||
|
"""
|
||||||
|
|
||||||
class Venue(EmbeddedDocument):
|
class Venue(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
@@ -1,12 +1,14 @@
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
class TestIntField(MongoDBTestCase):
|
class TestIntField(MongoDBTestCase):
|
||||||
def test_int_validation(self):
|
def test_int_validation(self):
|
||||||
"""Ensure that invalid values cannot be assigned to int fields."""
|
"""Ensure that invalid values cannot be assigned to int fields.
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
age = IntField(min_value=0, max_value=110)
|
age = IntField(min_value=0, max_value=110)
|
||||||
|
@@ -1,9 +1,9 @@
|
|||||||
import pytest
|
|
||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.base import LazyReference
|
from mongoengine.base import LazyReference
|
||||||
from mongoengine.context_managers import query_counter
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -330,70 +330,6 @@ class TestLazyReferenceField(MongoDBTestCase):
|
|||||||
occ.in_embedded.in_list = [animal1.id, animal2.id]
|
occ.in_embedded.in_list = [animal1.id, animal2.id]
|
||||||
check_fields_type(occ)
|
check_fields_type(occ)
|
||||||
|
|
||||||
def test_lazy_reference_embedded_dereferencing(self):
|
|
||||||
# Test case for #2375
|
|
||||||
|
|
||||||
# -- Test documents
|
|
||||||
|
|
||||||
class Author(Document):
|
|
||||||
name = StringField()
|
|
||||||
|
|
||||||
class AuthorReference(EmbeddedDocument):
|
|
||||||
author = LazyReferenceField(Author)
|
|
||||||
|
|
||||||
class Book(Document):
|
|
||||||
authors = EmbeddedDocumentListField(AuthorReference)
|
|
||||||
|
|
||||||
# -- Cleanup
|
|
||||||
|
|
||||||
Author.drop_collection()
|
|
||||||
Book.drop_collection()
|
|
||||||
|
|
||||||
# -- Create test data
|
|
||||||
|
|
||||||
author_1 = Author(name="A1").save()
|
|
||||||
author_2 = Author(name="A2").save()
|
|
||||||
author_3 = Author(name="A3").save()
|
|
||||||
book = Book(
|
|
||||||
authors=[
|
|
||||||
AuthorReference(author=author_1),
|
|
||||||
AuthorReference(author=author_2),
|
|
||||||
AuthorReference(author=author_3),
|
|
||||||
]
|
|
||||||
).save()
|
|
||||||
|
|
||||||
with query_counter() as qc:
|
|
||||||
book = Book.objects.first()
|
|
||||||
# Accessing the list must not trigger dereferencing.
|
|
||||||
book.authors
|
|
||||||
assert qc == 1
|
|
||||||
|
|
||||||
for ref in book.authors:
|
|
||||||
with pytest.raises(AttributeError):
|
|
||||||
ref["author"].name
|
|
||||||
assert isinstance(ref.author, LazyReference)
|
|
||||||
assert isinstance(ref.author.id, ObjectId)
|
|
||||||
|
|
||||||
def test_lazy_reference_in_list_with_changed_element(self):
|
|
||||||
class Animal(Document):
|
|
||||||
name = StringField()
|
|
||||||
tag = StringField()
|
|
||||||
|
|
||||||
class Ocurrence(Document):
|
|
||||||
in_list = ListField(LazyReferenceField(Animal))
|
|
||||||
|
|
||||||
Animal.drop_collection()
|
|
||||||
Ocurrence.drop_collection()
|
|
||||||
|
|
||||||
animal1 = Animal(name="doggo").save()
|
|
||||||
|
|
||||||
animal1.tag = "blue"
|
|
||||||
|
|
||||||
occ = Ocurrence(in_list=[animal1]).save()
|
|
||||||
animal1.save()
|
|
||||||
assert isinstance(occ.in_list[0], LazyReference)
|
|
||||||
assert occ.in_list[0].pk == animal1.pk
|
|
||||||
|
|
||||||
|
|
||||||
class TestGenericLazyReferenceField(MongoDBTestCase):
|
class TestGenericLazyReferenceField(MongoDBTestCase):
|
||||||
def test_generic_lazy_reference_simple(self):
|
def test_generic_lazy_reference_simple(self):
|
||||||
|
@@ -1,28 +1,13 @@
|
|||||||
import pytest
|
|
||||||
from bson.int64 import Int64
|
from bson.int64 import Int64
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from tests.utils import MongoDBTestCase, get_as_pymongo
|
|
||||||
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
class TestLongField(MongoDBTestCase):
|
class TestLongField(MongoDBTestCase):
|
||||||
def test_storage(self):
|
|
||||||
class Person(Document):
|
|
||||||
value = LongField()
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
person = Person(value=5000)
|
|
||||||
person.save()
|
|
||||||
assert get_as_pymongo(person) == {"_id": person.id, "value": 5000}
|
|
||||||
|
|
||||||
def test_construction_does_not_fail_with_invalid_value(self):
|
|
||||||
class Person(Document):
|
|
||||||
value = LongField()
|
|
||||||
|
|
||||||
person = Person(value="not_an_int")
|
|
||||||
assert person.value == "not_an_int"
|
|
||||||
|
|
||||||
def test_long_field_is_considered_as_int64(self):
|
def test_long_field_is_considered_as_int64(self):
|
||||||
"""
|
"""
|
||||||
Tests that long fields are stored as long in mongo, even if long
|
Tests that long fields are stored as long in mongo, even if long
|
||||||
@@ -40,21 +25,25 @@ class TestLongField(MongoDBTestCase):
|
|||||||
assert isinstance(doc.some_long, int)
|
assert isinstance(doc.some_long, int)
|
||||||
|
|
||||||
def test_long_validation(self):
|
def test_long_validation(self):
|
||||||
"""Ensure that invalid values cannot be assigned to long fields."""
|
"""Ensure that invalid values cannot be assigned to long fields.
|
||||||
|
"""
|
||||||
|
|
||||||
class TestDocument(Document):
|
class TestDocument(Document):
|
||||||
value = LongField(min_value=0, max_value=110)
|
value = LongField(min_value=0, max_value=110)
|
||||||
|
|
||||||
TestDocument(value=50).validate()
|
doc = TestDocument()
|
||||||
|
doc.value = 50
|
||||||
|
doc.validate()
|
||||||
|
|
||||||
|
doc.value = -1
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
TestDocument(value=-1).validate()
|
doc.validate()
|
||||||
|
doc.value = 120
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
TestDocument(value=120).validate()
|
doc.validate()
|
||||||
|
doc.value = "ten"
|
||||||
with pytest.raises(ValidationError):
|
with pytest.raises(ValidationError):
|
||||||
TestDocument(value="ten").validate()
|
doc.validate()
|
||||||
|
|
||||||
def test_long_ne_operator(self):
|
def test_long_ne_operator(self):
|
||||||
class TestDocument(Document):
|
class TestDocument(Document):
|
||||||
@@ -65,5 +54,4 @@ class TestLongField(MongoDBTestCase):
|
|||||||
TestDocument(long_fld=None).save()
|
TestDocument(long_fld=None).save()
|
||||||
TestDocument(long_fld=1).save()
|
TestDocument(long_fld=1).save()
|
||||||
|
|
||||||
assert TestDocument.objects(long_fld__ne=None).count() == 1
|
assert 1 == TestDocument.objects(long_fld__ne=None).count()
|
||||||
assert TestDocument.objects(long_fld__ne=1).count() == 1
|
|
||||||
|
@@ -135,11 +135,11 @@ class TestMapField(MongoDBTestCase):
|
|||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
tree = BlogPost(info_dict={"éééé": {"description": "VALUE: éééé"}})
|
tree = BlogPost(info_dict={u"éééé": {"description": u"VALUE: éééé"}})
|
||||||
|
|
||||||
tree.save()
|
tree.save()
|
||||||
|
|
||||||
assert (
|
assert (
|
||||||
BlogPost.objects.get(id=tree.id).info_dict["éééé"].description
|
BlogPost.objects.get(id=tree.id).info_dict[u"éééé"].description
|
||||||
== "VALUE: éééé"
|
== u"VALUE: éééé"
|
||||||
)
|
)
|
||||||
|
@@ -1,5 +1,5 @@
|
|||||||
|
from bson import DBRef, SON
|
||||||
import pytest
|
import pytest
|
||||||
from bson import SON, DBRef
|
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
@@ -87,7 +87,7 @@ class TestReferenceField(MongoDBTestCase):
|
|||||||
parent = ReferenceField("self", dbref=False)
|
parent = ReferenceField("self", dbref=False)
|
||||||
|
|
||||||
p = Person(name="Steve", parent=DBRef("person", "abcdefghijklmnop"))
|
p = Person(name="Steve", parent=DBRef("person", "abcdefghijklmnop"))
|
||||||
assert p.to_mongo() == SON([("name", "Steve"), ("parent", "abcdefghijklmnop")])
|
assert p.to_mongo() == SON([("name", u"Steve"), ("parent", "abcdefghijklmnop")])
|
||||||
|
|
||||||
def test_objectid_reference_fields(self):
|
def test_objectid_reference_fields(self):
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
@@ -107,7 +107,8 @@ class TestReferenceField(MongoDBTestCase):
|
|||||||
assert p.parent == p1
|
assert p.parent == p1
|
||||||
|
|
||||||
def test_undefined_reference(self):
|
def test_undefined_reference(self):
|
||||||
"""Ensure that ReferenceFields may reference undefined Documents."""
|
"""Ensure that ReferenceFields may reference undefined Documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class Product(Document):
|
class Product(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
@@ -1,4 +1,5 @@
|
|||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -165,8 +166,8 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
ids = [i.id for i in Person.objects]
|
ids = [i.id for i in Person.objects]
|
||||||
assert ids == list(range(1, 11))
|
assert ids == list(range(1, 11))
|
||||||
|
|
||||||
_id = [i.id for i in Animal.objects]
|
id = [i.id for i in Animal.objects]
|
||||||
assert _id == list(range(1, 11))
|
assert id == list(range(1, 11))
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
||||||
assert c["next"] == 10
|
assert c["next"] == 10
|
||||||
@@ -273,25 +274,3 @@ class TestSequenceField(MongoDBTestCase):
|
|||||||
assert foo.counter == bar.counter
|
assert foo.counter == bar.counter
|
||||||
assert foo._fields["counter"].owner_document == Foo
|
assert foo._fields["counter"].owner_document == Foo
|
||||||
assert bar._fields["counter"].owner_document == Bar
|
assert bar._fields["counter"].owner_document == Bar
|
||||||
|
|
||||||
def test_sequence_setattr_not_incrementing_counter(self):
|
|
||||||
class Person(DynamicDocument):
|
|
||||||
id = SequenceField(primary_key=True)
|
|
||||||
name = StringField()
|
|
||||||
|
|
||||||
self.db["mongoengine.counters"].drop()
|
|
||||||
Person.drop_collection()
|
|
||||||
|
|
||||||
for x in range(10):
|
|
||||||
Person(name="Person %s" % x).save()
|
|
||||||
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
|
||||||
assert c["next"] == 10
|
|
||||||
|
|
||||||
# Setting SequenceField field value should not increment counter:
|
|
||||||
new_person = Person()
|
|
||||||
new_person.id = 1100
|
|
||||||
|
|
||||||
# Counter should still be at 10
|
|
||||||
c = self.db["mongoengine.counters"].find_one({"_id": "person.id"})
|
|
||||||
assert c["next"] == 10
|
|
||||||
|
@@ -1,43 +0,0 @@
|
|||||||
import pytest
|
|
||||||
|
|
||||||
from mongoengine import *
|
|
||||||
from tests.utils import MongoDBTestCase, get_as_pymongo
|
|
||||||
|
|
||||||
|
|
||||||
class TestStringField(MongoDBTestCase):
|
|
||||||
def test_storage(self):
|
|
||||||
class Person(Document):
|
|
||||||
name = StringField()
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
person = Person(name="test123")
|
|
||||||
person.save()
|
|
||||||
assert get_as_pymongo(person) == {"_id": person.id, "name": "test123"}
|
|
||||||
|
|
||||||
def test_validation(self):
|
|
||||||
class Person(Document):
|
|
||||||
name = StringField(max_length=20, min_length=2)
|
|
||||||
userid = StringField(r"[0-9a-z_]+$")
|
|
||||||
|
|
||||||
with pytest.raises(ValidationError, match="only accepts string values"):
|
|
||||||
Person(name=34).validate()
|
|
||||||
|
|
||||||
with pytest.raises(ValidationError, match="value is too short"):
|
|
||||||
Person(name="s").validate()
|
|
||||||
|
|
||||||
# Test regex validation on userid
|
|
||||||
person = Person(userid="test.User")
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person.validate()
|
|
||||||
|
|
||||||
person.userid = "test_user"
|
|
||||||
assert person.userid == "test_user"
|
|
||||||
person.validate()
|
|
||||||
|
|
||||||
# Test max length validation on name
|
|
||||||
person = Person(name="Name that is more than twenty characters")
|
|
||||||
with pytest.raises(ValidationError):
|
|
||||||
person.validate()
|
|
||||||
|
|
||||||
person = Person(name="a friendl name", userid="7a757668sqjdkqlsdkq")
|
|
||||||
person.validate()
|
|
@@ -26,7 +26,7 @@ class TestURLField(MongoDBTestCase):
|
|||||||
url = URLField()
|
url = URLField()
|
||||||
|
|
||||||
link = Link()
|
link = Link()
|
||||||
link.url = "http://привет.com"
|
link.url = u"http://привет.com"
|
||||||
|
|
||||||
# TODO fix URL validation - this *IS* a valid URL
|
# TODO fix URL validation - this *IS* a valid URL
|
||||||
# For now we just want to make sure that the error message is correct
|
# For now we just want to make sure that the error message is correct
|
||||||
@@ -34,11 +34,12 @@ class TestURLField(MongoDBTestCase):
|
|||||||
link.validate()
|
link.validate()
|
||||||
assert (
|
assert (
|
||||||
str(exc_info.value)
|
str(exc_info.value)
|
||||||
== "ValidationError (Link:None) (Invalid URL: http://\u043f\u0440\u0438\u0432\u0435\u0442.com: ['url'])"
|
== u"ValidationError (Link:None) (Invalid URL: http://\u043f\u0440\u0438\u0432\u0435\u0442.com: ['url'])"
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_url_scheme_validation(self):
|
def test_url_scheme_validation(self):
|
||||||
"""Ensure that URLFields validate urls with specific schemes properly."""
|
"""Ensure that URLFields validate urls with specific schemes properly.
|
||||||
|
"""
|
||||||
|
|
||||||
class Link(Document):
|
class Link(Document):
|
||||||
url = URLField()
|
url = URLField()
|
||||||
|
@@ -17,7 +17,8 @@ class TestUUIDField(MongoDBTestCase):
|
|||||||
assert get_as_pymongo(person) == {"_id": person.id, "api_key": str(uid)}
|
assert get_as_pymongo(person) == {"_id": person.id, "api_key": str(uid)}
|
||||||
|
|
||||||
def test_field_string(self):
|
def test_field_string(self):
|
||||||
"""Test UUID fields storing as String"""
|
"""Test UUID fields storing as String
|
||||||
|
"""
|
||||||
Person.drop_collection()
|
Person.drop_collection()
|
||||||
|
|
||||||
uu = uuid.uuid4()
|
uu = uuid.uuid4()
|
||||||
|
@@ -53,7 +53,7 @@ signals.post_save.connect(PickleSignalsTest.post_save, sender=PickleSignalsTest)
|
|||||||
signals.post_delete.connect(PickleSignalsTest.post_delete, sender=PickleSignalsTest)
|
signals.post_delete.connect(PickleSignalsTest.post_delete, sender=PickleSignalsTest)
|
||||||
|
|
||||||
|
|
||||||
class Mixin:
|
class Mixin(object):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
|
|
||||||
|
|
||||||
|
@@ -148,7 +148,8 @@ class TestOnlyExcludeAll(unittest.TestCase):
|
|||||||
assert qs._loaded_fields.as_dict() == {"c": {"$slice": 2}, "a": 1}
|
assert qs._loaded_fields.as_dict() == {"c": {"$slice": 2}, "a": 1}
|
||||||
|
|
||||||
def test_only(self):
|
def test_only(self):
|
||||||
"""Ensure that QuerySet.only only returns the requested fields."""
|
"""Ensure that QuerySet.only only returns the requested fields.
|
||||||
|
"""
|
||||||
person = self.Person(name="test", age=25)
|
person = self.Person(name="test", age=25)
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
@@ -364,7 +365,8 @@ class TestOnlyExcludeAll(unittest.TestCase):
|
|||||||
Email.drop_collection()
|
Email.drop_collection()
|
||||||
|
|
||||||
def test_slicing_fields(self):
|
def test_slicing_fields(self):
|
||||||
"""Ensure that query slicing an array works."""
|
"""Ensure that query slicing an array works.
|
||||||
|
"""
|
||||||
|
|
||||||
class Numbers(Document):
|
class Numbers(Document):
|
||||||
n = ListField(IntField())
|
n = ListField(IntField())
|
||||||
@@ -399,7 +401,8 @@ class TestOnlyExcludeAll(unittest.TestCase):
|
|||||||
assert numbers.n == [-5, -4, -3, -2, -1]
|
assert numbers.n == [-5, -4, -3, -2, -1]
|
||||||
|
|
||||||
def test_slicing_nested_fields(self):
|
def test_slicing_nested_fields(self):
|
||||||
"""Ensure that query slicing an embedded array works."""
|
"""Ensure that query slicing an embedded array works.
|
||||||
|
"""
|
||||||
|
|
||||||
class EmbeddedNumber(EmbeddedDocument):
|
class EmbeddedNumber(EmbeddedDocument):
|
||||||
n = ListField(IntField())
|
n = ListField(IntField())
|
||||||
|
@@ -2,6 +2,7 @@ import datetime
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
|
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -495,8 +496,8 @@ class TestGeoQueries(MongoDBTestCase):
|
|||||||
p.save()
|
p.save()
|
||||||
qs = Place.objects().only("location")
|
qs = Place.objects().only("location")
|
||||||
assert qs.as_pymongo()[0]["location"] == {
|
assert qs.as_pymongo()[0]["location"] == {
|
||||||
"type": "Point",
|
u"type": u"Point",
|
||||||
"coordinates": [24.946861267089844, 60.16311983618494],
|
u"coordinates": [24.946861267089844, 60.16311983618494],
|
||||||
}
|
}
|
||||||
|
|
||||||
def test_2dsphere_point_sets_correctly(self):
|
def test_2dsphere_point_sets_correctly(self):
|
||||||
|
@@ -1,12 +1,6 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from mongoengine import (
|
from mongoengine import Document, IntField, ListField, StringField, connect
|
||||||
Document,
|
|
||||||
IntField,
|
|
||||||
ListField,
|
|
||||||
StringField,
|
|
||||||
connect,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class Doc(Document):
|
class Doc(Document):
|
||||||
|
@@ -1,6 +1,8 @@
|
|||||||
import pickle
|
import pickle
|
||||||
|
import unittest
|
||||||
|
|
||||||
from mongoengine import Document, IntField, StringField
|
from mongoengine import Document, IntField, StringField
|
||||||
|
from mongoengine.connection import connect
|
||||||
from tests.utils import MongoDBTestCase
|
from tests.utils import MongoDBTestCase
|
||||||
|
|
||||||
|
|
||||||
@@ -16,15 +18,18 @@ class TestQuerysetPickable(MongoDBTestCase):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
super().setUp()
|
super(TestQuerysetPickable, self).setUp()
|
||||||
self.john = Person.objects.create(name="John", age=21)
|
self.john = Person.objects.create(name="John", age=21)
|
||||||
|
|
||||||
def test_picke_simple_qs(self):
|
def test_picke_simple_qs(self):
|
||||||
|
|
||||||
qs = Person.objects.all()
|
qs = Person.objects.all()
|
||||||
|
|
||||||
pickle.dumps(qs)
|
pickle.dumps(qs)
|
||||||
|
|
||||||
def _get_loaded(self, qs):
|
def _get_loaded(self, qs):
|
||||||
s = pickle.dumps(qs)
|
s = pickle.dumps(qs)
|
||||||
|
|
||||||
return pickle.loads(s)
|
return pickle.loads(s)
|
||||||
|
|
||||||
def test_unpickle(self):
|
def test_unpickle(self):
|
||||||
|
@@ -3,20 +3,18 @@ import unittest
|
|||||||
import uuid
|
import uuid
|
||||||
from decimal import Decimal
|
from decimal import Decimal
|
||||||
|
|
||||||
import pymongo
|
|
||||||
import pytest
|
|
||||||
from bson import DBRef, ObjectId
|
from bson import DBRef, ObjectId
|
||||||
|
import pymongo
|
||||||
|
from pymongo.read_concern import ReadConcern
|
||||||
from pymongo.read_preferences import ReadPreference
|
from pymongo.read_preferences import ReadPreference
|
||||||
from pymongo.results import UpdateResult
|
from pymongo.results import UpdateResult
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.connection import get_db
|
from mongoengine.connection import get_db
|
||||||
from mongoengine.context_managers import query_counter, switch_db
|
from mongoengine.context_managers import query_counter, switch_db
|
||||||
from mongoengine.errors import InvalidQueryError
|
from mongoengine.errors import InvalidQueryError
|
||||||
from mongoengine.mongodb_support import (
|
from mongoengine.mongodb_support import MONGODB_36, get_mongodb_version
|
||||||
MONGODB_36,
|
|
||||||
get_mongodb_version,
|
|
||||||
)
|
|
||||||
from mongoengine.queryset import (
|
from mongoengine.queryset import (
|
||||||
DoesNotExist,
|
DoesNotExist,
|
||||||
MultipleObjectsReturned,
|
MultipleObjectsReturned,
|
||||||
@@ -24,10 +22,6 @@ from mongoengine.queryset import (
|
|||||||
QuerySetManager,
|
QuerySetManager,
|
||||||
queryset_manager,
|
queryset_manager,
|
||||||
)
|
)
|
||||||
from tests.utils import (
|
|
||||||
requires_mongodb_gte_44,
|
|
||||||
requires_mongodb_lt_42,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class db_ops_tracker(query_counter):
|
class db_ops_tracker(query_counter):
|
||||||
@@ -66,7 +60,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
self.mongodb_version = get_mongodb_version()
|
self.mongodb_version = get_mongodb_version()
|
||||||
|
|
||||||
def test_initialisation(self):
|
def test_initialisation(self):
|
||||||
"""Ensure that a QuerySet is correctly initialised by QuerySetManager."""
|
"""Ensure that a QuerySet is correctly initialised by QuerySetManager.
|
||||||
|
"""
|
||||||
assert isinstance(self.Person.objects, QuerySet)
|
assert isinstance(self.Person.objects, QuerySet)
|
||||||
assert (
|
assert (
|
||||||
self.Person.objects._collection.name == self.Person._get_collection_name()
|
self.Person.objects._collection.name == self.Person._get_collection_name()
|
||||||
@@ -119,7 +114,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
def test_slicing_sets_empty_limit_skip(self):
|
def test_slicing_sets_empty_limit_skip(self):
|
||||||
self.Person.objects.insert(
|
self.Person.objects.insert(
|
||||||
[self.Person(name=f"User {i}", age=i) for i in range(5)],
|
[self.Person(name="User {}".format(i), age=i) for i in range(5)],
|
||||||
load_bulk=False,
|
load_bulk=False,
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -276,7 +271,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_find_one(self):
|
def test_find_one(self):
|
||||||
"""Ensure that a query using find_one returns a valid result."""
|
"""Ensure that a query using find_one returns a valid result.
|
||||||
|
"""
|
||||||
person1 = self.Person(name="User A", age=20)
|
person1 = self.Person(name="User A", age=20)
|
||||||
person1.save()
|
person1.save()
|
||||||
person2 = self.Person(name="User B", age=30)
|
person2 = self.Person(name="User B", age=30)
|
||||||
@@ -321,7 +317,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
self.Person.objects.get()
|
self.Person.objects.get()
|
||||||
|
|
||||||
def test_get_multiple_match_raises_multipleobjectsreturned(self):
|
def test_get_multiple_match_raises_multipleobjectsreturned(self):
|
||||||
"""Ensure that a query using ``get`` returns at most one result."""
|
"""Ensure that a query using ``get`` returns at most one result.
|
||||||
|
"""
|
||||||
assert self.Person.objects().count() == 0
|
assert self.Person.objects().count() == 0
|
||||||
|
|
||||||
person1 = self.Person(name="User A", age=20)
|
person1 = self.Person(name="User A", age=20)
|
||||||
@@ -354,7 +351,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert person == person3
|
assert person == person3
|
||||||
|
|
||||||
def test_find_array_position(self):
|
def test_find_array_position(self):
|
||||||
"""Ensure that query by array position works."""
|
"""Ensure that query by array position works.
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -455,7 +453,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
# test iterating over the result set
|
# test iterating over the result set
|
||||||
cnt = 0
|
cnt = 0
|
||||||
for _ in A.objects.batch_size(10):
|
for a in A.objects.batch_size(10):
|
||||||
cnt += 1
|
cnt += 1
|
||||||
assert cnt == 100
|
assert cnt == 100
|
||||||
|
|
||||||
@@ -463,7 +461,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
qs = A.objects.all()
|
qs = A.objects.all()
|
||||||
qs = qs.limit(10).batch_size(20).skip(91)
|
qs = qs.limit(10).batch_size(20).skip(91)
|
||||||
cnt = 0
|
cnt = 0
|
||||||
for _ in qs:
|
for a in qs:
|
||||||
cnt += 1
|
cnt += 1
|
||||||
assert cnt == 9
|
assert cnt == 9
|
||||||
|
|
||||||
@@ -1143,7 +1141,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
def test_repeated_iteration(self):
|
def test_repeated_iteration(self):
|
||||||
"""Ensure that QuerySet rewinds itself one iteration finishes."""
|
"""Ensure that QuerySet rewinds itself one iteration finishes.
|
||||||
|
"""
|
||||||
self.Person(name="Person 1").save()
|
self.Person(name="Person 1").save()
|
||||||
self.Person(name="Person 2").save()
|
self.Person(name="Person 2").save()
|
||||||
|
|
||||||
@@ -1152,7 +1151,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
people2 = [person for person in queryset]
|
people2 = [person for person in queryset]
|
||||||
|
|
||||||
# Check that it still works even if iteration is interrupted.
|
# Check that it still works even if iteration is interrupted.
|
||||||
for _person in queryset:
|
for person in queryset:
|
||||||
break
|
break
|
||||||
people3 = [person for person in queryset]
|
people3 = [person for person in queryset]
|
||||||
|
|
||||||
@@ -1188,11 +1187,12 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert "[<Doc: 1>, <Doc: 2>, <Doc: 3>]" == "%s" % docs
|
assert "[<Doc: 1>, <Doc: 2>, <Doc: 3>]" == "%s" % docs
|
||||||
|
|
||||||
assert docs.count(with_limit_and_skip=True) == 3
|
assert docs.count(with_limit_and_skip=True) == 3
|
||||||
for _ in docs:
|
for doc in docs:
|
||||||
assert ".. queryset mid-iteration .." == repr(docs)
|
assert ".. queryset mid-iteration .." == repr(docs)
|
||||||
|
|
||||||
def test_regex_query_shortcuts(self):
|
def test_regex_query_shortcuts(self):
|
||||||
"""Ensure that contains, startswith, endswith, etc work."""
|
"""Ensure that contains, startswith, endswith, etc work.
|
||||||
|
"""
|
||||||
person = self.Person(name="Guido van Rossum")
|
person = self.Person(name="Guido van Rossum")
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
@@ -1248,7 +1248,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert obj == person
|
assert obj == person
|
||||||
|
|
||||||
def test_not(self):
|
def test_not(self):
|
||||||
"""Ensure that the __not operator works as expected."""
|
"""Ensure that the __not operator works as expected.
|
||||||
|
"""
|
||||||
alice = self.Person(name="Alice", age=25)
|
alice = self.Person(name="Alice", age=25)
|
||||||
alice.save()
|
alice.save()
|
||||||
|
|
||||||
@@ -1259,7 +1260,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert obj is None
|
assert obj is None
|
||||||
|
|
||||||
def test_filter_chaining(self):
|
def test_filter_chaining(self):
|
||||||
"""Ensure filters can be chained together."""
|
"""Ensure filters can be chained together.
|
||||||
|
"""
|
||||||
|
|
||||||
class Blog(Document):
|
class Blog(Document):
|
||||||
id = StringField(primary_key=True)
|
id = StringField(primary_key=True)
|
||||||
@@ -1335,7 +1337,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert qs[i] == expected[i]
|
assert qs[i] == expected[i]
|
||||||
|
|
||||||
def test_ordering(self):
|
def test_ordering(self):
|
||||||
"""Ensure default ordering is applied and can be overridden."""
|
"""Ensure default ordering is applied and can be overridden.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -1405,7 +1408,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert ORDER_BY_KEY not in q.get_ops()[0][CMD_QUERY_KEY]
|
assert ORDER_BY_KEY not in q.get_ops()[0][CMD_QUERY_KEY]
|
||||||
|
|
||||||
def test_no_ordering_for_get(self):
|
def test_no_ordering_for_get(self):
|
||||||
"""Ensure that Doc.objects.get doesn't use any ordering."""
|
""" Ensure that Doc.objects.get doesn't use any ordering.
|
||||||
|
"""
|
||||||
ORDER_BY_KEY, CMD_QUERY_KEY = get_key_compat(self.mongodb_version)
|
ORDER_BY_KEY, CMD_QUERY_KEY = get_key_compat(self.mongodb_version)
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
@@ -1481,7 +1485,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert result.author is None
|
assert result.author is None
|
||||||
|
|
||||||
def test_find_dict_item(self):
|
def test_find_dict_item(self):
|
||||||
"""Ensure that DictField items may be found."""
|
"""Ensure that DictField items may be found.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
info = DictField()
|
info = DictField()
|
||||||
@@ -1496,9 +1501,9 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
@requires_mongodb_lt_42
|
|
||||||
def test_exec_js_query(self):
|
def test_exec_js_query(self):
|
||||||
"""Ensure that queries are properly formed for use in exec_js."""
|
"""Ensure that queries are properly formed for use in exec_js.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
hits = IntField()
|
hits = IntField()
|
||||||
@@ -1534,9 +1539,9 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
@requires_mongodb_lt_42
|
|
||||||
def test_exec_js_field_sub(self):
|
def test_exec_js_field_sub(self):
|
||||||
"""Ensure that field substitutions occur properly in exec_js functions."""
|
"""Ensure that field substitutions occur properly in exec_js functions.
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
content = StringField(db_field="body")
|
content = StringField(db_field="body")
|
||||||
@@ -1578,9 +1583,9 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
results = BlogPost.objects.exec_js(code)
|
results = BlogPost.objects.exec_js(code)
|
||||||
expected_results = [
|
expected_results = [
|
||||||
{"comment": "cool", "document": "post1"},
|
{u"comment": u"cool", u"document": u"post1"},
|
||||||
{"comment": "yay", "document": "post1"},
|
{u"comment": u"yay", u"document": u"post1"},
|
||||||
{"comment": "nice stuff", "document": "post2"},
|
{u"comment": u"nice stuff", u"document": u"post2"},
|
||||||
]
|
]
|
||||||
assert results == expected_results
|
assert results == expected_results
|
||||||
|
|
||||||
@@ -1592,7 +1597,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_delete(self):
|
def test_delete(self):
|
||||||
"""Ensure that documents are properly deleted from the database."""
|
"""Ensure that documents are properly deleted from the database.
|
||||||
|
"""
|
||||||
self.Person(name="User A", age=20).save()
|
self.Person(name="User A", age=20).save()
|
||||||
self.Person(name="User B", age=30).save()
|
self.Person(name="User B", age=30).save()
|
||||||
self.Person(name="User C", age=40).save()
|
self.Person(name="User C", age=40).save()
|
||||||
@@ -1606,7 +1612,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert self.Person.objects.count() == 0
|
assert self.Person.objects.count() == 0
|
||||||
|
|
||||||
def test_reverse_delete_rule_cascade(self):
|
def test_reverse_delete_rule_cascade(self):
|
||||||
"""Ensure cascading deletion of referring documents from the database."""
|
"""Ensure cascading deletion of referring documents from the database.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -1738,7 +1745,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert 0 == Category.objects.count()
|
assert 0 == Category.objects.count()
|
||||||
|
|
||||||
def test_reverse_delete_rule_nullify(self):
|
def test_reverse_delete_rule_nullify(self):
|
||||||
"""Ensure nullification of references to deleted documents."""
|
"""Ensure nullification of references to deleted documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class Category(Document):
|
class Category(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1834,7 +1842,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
self.Person.objects.delete()
|
self.Person.objects.delete()
|
||||||
|
|
||||||
def test_reverse_delete_rule_pull(self):
|
def test_reverse_delete_rule_pull(self):
|
||||||
"""Ensure pulling of references to deleted documents."""
|
"""Ensure pulling of references to deleted documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -1909,7 +1918,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert 8 == Log.objects.count()
|
assert 8 == Log.objects.count()
|
||||||
|
|
||||||
def test_delete_with_limit_handles_delete_rules(self):
|
def test_delete_with_limit_handles_delete_rules(self):
|
||||||
"""Ensure cascading deletion of referring documents from the database."""
|
"""Ensure cascading deletion of referring documents from the database.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -1941,7 +1951,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert del_result is None
|
assert del_result is None
|
||||||
|
|
||||||
def test_reference_field_find(self):
|
def test_reference_field_find(self):
|
||||||
"""Ensure cascading deletion of referring documents from the database."""
|
"""Ensure cascading deletion of referring documents from the database.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -1962,7 +1973,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert 1 == BlogPost.objects(author__in=["%s" % me.pk]).count()
|
assert 1 == BlogPost.objects(author__in=["%s" % me.pk]).count()
|
||||||
|
|
||||||
def test_reference_field_find_dbref(self):
|
def test_reference_field_find_dbref(self):
|
||||||
"""Ensure cascading deletion of referring documents from the database."""
|
"""Ensure cascading deletion of referring documents from the database.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
content = StringField()
|
content = StringField()
|
||||||
@@ -2046,7 +2058,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
BlogPost.objects.update_one(inc__review=0.1) # test with floats
|
BlogPost.objects.update_one(inc__review=0.1) # test with floats
|
||||||
|
|
||||||
def test_update_listfield_operator(self):
|
def test_update_listfield_operator(self):
|
||||||
"""Ensure that atomic updates work properly."""
|
"""Ensure that atomic updates work properly.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
tags = ListField(StringField())
|
tags = ListField(StringField())
|
||||||
@@ -2094,7 +2107,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert "title" not in pymongo_doc
|
assert "title" not in pymongo_doc
|
||||||
|
|
||||||
def test_update_push_with_position(self):
|
def test_update_push_with_position(self):
|
||||||
"""Ensure that the 'push' update with position works properly."""
|
"""Ensure that the 'push' update with position works properly.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
slug = StringField()
|
slug = StringField()
|
||||||
@@ -2119,7 +2133,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert post.tags == ["scala", "mongodb", "python", "java"]
|
assert post.tags == ["scala", "mongodb", "python", "java"]
|
||||||
|
|
||||||
def test_update_push_list_of_list(self):
|
def test_update_push_list_of_list(self):
|
||||||
"""Ensure that the 'push' update operation works in the list of list"""
|
"""Ensure that the 'push' update operation works in the list of list
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
slug = StringField()
|
slug = StringField()
|
||||||
@@ -2134,7 +2149,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert post.tags == [["value1", 123]]
|
assert post.tags == [["value1", 123]]
|
||||||
|
|
||||||
def test_update_push_and_pull_add_to_set(self):
|
def test_update_push_and_pull_add_to_set(self):
|
||||||
"""Ensure that the 'pull' update operation works correctly."""
|
"""Ensure that the 'pull' update operation works correctly.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
slug = StringField()
|
slug = StringField()
|
||||||
@@ -2243,7 +2259,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_pull_from_nested_embedded_using_in_nin(self):
|
def test_pull_from_nested_embedded_using_in_nin(self):
|
||||||
"""Ensure that the 'pull' update operation works on embedded documents using 'in' and 'nin' operators."""
|
"""Ensure that the 'pull' update operation works on embedded documents using 'in' and 'nin' operators.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(EmbeddedDocument):
|
class User(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -2474,7 +2491,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert doc.pk == doc.embedded[0]._instance.pk
|
assert doc.pk == doc.embedded[0]._instance.pk
|
||||||
|
|
||||||
def test_order_by(self):
|
def test_order_by(self):
|
||||||
"""Ensure that QuerySets may be ordered."""
|
"""Ensure that QuerySets may be ordered.
|
||||||
|
"""
|
||||||
self.Person(name="User B", age=40).save()
|
self.Person(name="User B", age=40).save()
|
||||||
self.Person(name="User A", age=20).save()
|
self.Person(name="User A", age=20).save()
|
||||||
self.Person(name="User C", age=30).save()
|
self.Person(name="User C", age=30).save()
|
||||||
@@ -2542,7 +2560,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
self.assertSequence(qs, expected)
|
self.assertSequence(qs, expected)
|
||||||
|
|
||||||
def test_order_by_chaining(self):
|
def test_order_by_chaining(self):
|
||||||
"""Ensure that an order_by query chains properly and allows .only()"""
|
"""Ensure that an order_by query chains properly and allows .only()
|
||||||
|
"""
|
||||||
self.Person(name="User B", age=40).save()
|
self.Person(name="User B", age=40).save()
|
||||||
self.Person(name="User A", age=20).save()
|
self.Person(name="User A", age=20).save()
|
||||||
self.Person(name="User C", age=30).save()
|
self.Person(name="User C", age=30).save()
|
||||||
@@ -2616,7 +2635,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert op[CMD_QUERY_KEY][COMMENT_KEY] == "looking for an adult"
|
assert op[CMD_QUERY_KEY][COMMENT_KEY] == "looking for an adult"
|
||||||
|
|
||||||
def test_map_reduce(self):
|
def test_map_reduce(self):
|
||||||
"""Ensure map/reduce is both mapping and reducing."""
|
"""Ensure map/reduce is both mapping and reducing.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -2668,8 +2688,6 @@ class TestQueryset(unittest.TestCase):
|
|||||||
title = StringField(primary_key=True)
|
title = StringField(primary_key=True)
|
||||||
tags = ListField(StringField())
|
tags = ListField(StringField())
|
||||||
|
|
||||||
BlogPost.drop_collection()
|
|
||||||
|
|
||||||
post1 = BlogPost(title="Post #1", tags=["mongodb", "mongoengine"])
|
post1 = BlogPost(title="Post #1", tags=["mongodb", "mongoengine"])
|
||||||
post2 = BlogPost(title="Post #2", tags=["django", "mongodb"])
|
post2 = BlogPost(title="Post #2", tags=["django", "mongodb"])
|
||||||
post3 = BlogPost(title="Post #3", tags=["hitchcock films"])
|
post3 = BlogPost(title="Post #3", tags=["hitchcock films"])
|
||||||
@@ -2698,15 +2716,12 @@ class TestQueryset(unittest.TestCase):
|
|||||||
}
|
}
|
||||||
"""
|
"""
|
||||||
|
|
||||||
results = BlogPost.objects.order_by("_id").map_reduce(
|
results = BlogPost.objects.map_reduce(map_f, reduce_f, "myresults")
|
||||||
map_f, reduce_f, "myresults2"
|
|
||||||
)
|
|
||||||
results = list(results)
|
results = list(results)
|
||||||
|
|
||||||
assert len(results) == 3
|
assert results[0].object == post1
|
||||||
assert results[0].object.id == post1.id
|
assert results[1].object == post2
|
||||||
assert results[1].object.id == post2.id
|
assert results[2].object == post3
|
||||||
assert results[2].object.id == post3.id
|
|
||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
@@ -2714,6 +2729,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
"""
|
"""
|
||||||
Test map/reduce custom output
|
Test map/reduce custom output
|
||||||
"""
|
"""
|
||||||
|
register_connection("test2", "mongoenginetest2")
|
||||||
|
|
||||||
class Family(Document):
|
class Family(Document):
|
||||||
id = IntField(primary_key=True)
|
id = IntField(primary_key=True)
|
||||||
@@ -2733,10 +2749,10 @@ class TestQueryset(unittest.TestCase):
|
|||||||
f1.save()
|
f1.save()
|
||||||
|
|
||||||
# persons of first family
|
# persons of first family
|
||||||
Person(id=1, family=f1, name="Wilson Jr", age=21).save()
|
Person(id=1, family=f1, name=u"Wilson Jr", age=21).save()
|
||||||
Person(id=2, family=f1, name="Wilson Father", age=45).save()
|
Person(id=2, family=f1, name=u"Wilson Father", age=45).save()
|
||||||
Person(id=3, family=f1, name="Eliana Costa", age=40).save()
|
Person(id=3, family=f1, name=u"Eliana Costa", age=40).save()
|
||||||
Person(id=4, family=f1, name="Tayza Mariana", age=17).save()
|
Person(id=4, family=f1, name=u"Tayza Mariana", age=17).save()
|
||||||
|
|
||||||
# creating second family
|
# creating second family
|
||||||
f2 = Family(id=2, log="Av prof frasc brunno")
|
f2 = Family(id=2, log="Av prof frasc brunno")
|
||||||
@@ -2786,7 +2802,6 @@ class TestQueryset(unittest.TestCase):
|
|||||||
family.persons.push(person);
|
family.persons.push(person);
|
||||||
family.totalAge += person.age;
|
family.totalAge += person.age;
|
||||||
});
|
});
|
||||||
family.persons.sort((a, b) => (a.age > b.age))
|
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -2815,10 +2830,10 @@ class TestQueryset(unittest.TestCase):
|
|||||||
"_id": 1,
|
"_id": 1,
|
||||||
"value": {
|
"value": {
|
||||||
"persons": [
|
"persons": [
|
||||||
{"age": 17, "name": "Tayza Mariana"},
|
{"age": 21, "name": u"Wilson Jr"},
|
||||||
{"age": 21, "name": "Wilson Jr"},
|
{"age": 45, "name": u"Wilson Father"},
|
||||||
{"age": 40, "name": "Eliana Costa"},
|
{"age": 40, "name": u"Eliana Costa"},
|
||||||
{"age": 45, "name": "Wilson Father"},
|
{"age": 17, "name": u"Tayza Mariana"},
|
||||||
],
|
],
|
||||||
"totalAge": 123,
|
"totalAge": 123,
|
||||||
},
|
},
|
||||||
@@ -2828,9 +2843,9 @@ class TestQueryset(unittest.TestCase):
|
|||||||
"_id": 2,
|
"_id": 2,
|
||||||
"value": {
|
"value": {
|
||||||
"persons": [
|
"persons": [
|
||||||
{"age": 10, "name": "Igor Gabriel"},
|
{"age": 16, "name": u"Isabella Luanna"},
|
||||||
{"age": 16, "name": "Isabella Luanna"},
|
{"age": 36, "name": u"Sandra Mara"},
|
||||||
{"age": 36, "name": "Sandra Mara"},
|
{"age": 10, "name": u"Igor Gabriel"},
|
||||||
],
|
],
|
||||||
"totalAge": 62,
|
"totalAge": 62,
|
||||||
},
|
},
|
||||||
@@ -2840,8 +2855,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
"_id": 3,
|
"_id": 3,
|
||||||
"value": {
|
"value": {
|
||||||
"persons": [
|
"persons": [
|
||||||
{"age": 25, "name": "Paula Leonel"},
|
{"age": 30, "name": u"Arthur WA"},
|
||||||
{"age": 30, "name": "Arthur WA"},
|
{"age": 25, "name": u"Paula Leonel"},
|
||||||
],
|
],
|
||||||
"totalAge": 55,
|
"totalAge": 55,
|
||||||
},
|
},
|
||||||
@@ -2973,7 +2988,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Link.drop_collection()
|
Link.drop_collection()
|
||||||
|
|
||||||
def test_item_frequencies(self):
|
def test_item_frequencies(self):
|
||||||
"""Ensure that item frequencies are properly generated from lists."""
|
"""Ensure that item frequencies are properly generated from lists.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
hits = IntField()
|
hits = IntField()
|
||||||
@@ -2987,7 +3003,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
def test_assertions(f):
|
def test_assertions(f):
|
||||||
f = {key: int(val) for key, val in f.items()}
|
f = {key: int(val) for key, val in f.items()}
|
||||||
assert {"music", "film", "actors", "watch"} == set(f.keys())
|
assert set(["music", "film", "actors", "watch"]) == set(f.keys())
|
||||||
assert f["music"] == 3
|
assert f["music"] == 3
|
||||||
assert f["actors"] == 2
|
assert f["actors"] == 2
|
||||||
assert f["watch"] == 2
|
assert f["watch"] == 2
|
||||||
@@ -3001,7 +3017,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
# Ensure query is taken into account
|
# Ensure query is taken into account
|
||||||
def test_assertions(f):
|
def test_assertions(f):
|
||||||
f = {key: int(val) for key, val in f.items()}
|
f = {key: int(val) for key, val in f.items()}
|
||||||
assert {"music", "actors", "watch"} == set(f.keys())
|
assert set(["music", "actors", "watch"]) == set(f.keys())
|
||||||
assert f["music"] == 2
|
assert f["music"] == 2
|
||||||
assert f["actors"] == 1
|
assert f["actors"] == 1
|
||||||
assert f["watch"] == 1
|
assert f["watch"] == 1
|
||||||
@@ -3029,7 +3045,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
# Check item_frequencies works for non-list fields
|
# Check item_frequencies works for non-list fields
|
||||||
def test_assertions(f):
|
def test_assertions(f):
|
||||||
assert {1, 2} == set(f.keys())
|
assert set([1, 2]) == set(f.keys())
|
||||||
assert f[1] == 1
|
assert f[1] == 1
|
||||||
assert f[2] == 2
|
assert f[2] == 2
|
||||||
|
|
||||||
@@ -3041,7 +3057,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_item_frequencies_on_embedded(self):
|
def test_item_frequencies_on_embedded(self):
|
||||||
"""Ensure that item frequencies are properly generated from lists."""
|
"""Ensure that item frequencies are properly generated from lists.
|
||||||
|
"""
|
||||||
|
|
||||||
class Phone(EmbeddedDocument):
|
class Phone(EmbeddedDocument):
|
||||||
number = StringField()
|
number = StringField()
|
||||||
@@ -3066,7 +3083,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
def test_assertions(f):
|
def test_assertions(f):
|
||||||
f = {key: int(val) for key, val in f.items()}
|
f = {key: int(val) for key, val in f.items()}
|
||||||
assert {"62-3331-1656", "62-3332-1656"} == set(f.keys())
|
assert set(["62-3331-1656", "62-3332-1656"]) == set(f.keys())
|
||||||
assert f["62-3331-1656"] == 2
|
assert f["62-3331-1656"] == 2
|
||||||
assert f["62-3332-1656"] == 1
|
assert f["62-3332-1656"] == 1
|
||||||
|
|
||||||
@@ -3078,7 +3095,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
# Ensure query is taken into account
|
# Ensure query is taken into account
|
||||||
def test_assertions(f):
|
def test_assertions(f):
|
||||||
f = {key: int(val) for key, val in f.items()}
|
f = {key: int(val) for key, val in f.items()}
|
||||||
assert {"62-3331-1656"} == set(f.keys())
|
assert set(["62-3331-1656"]) == set(f.keys())
|
||||||
assert f["62-3331-1656"] == 2
|
assert f["62-3331-1656"] == 2
|
||||||
|
|
||||||
exec_js = Person.objects(phone__number="62-3331-1656").item_frequencies(
|
exec_js = Person.objects(phone__number="62-3331-1656").item_frequencies(
|
||||||
@@ -3122,7 +3139,6 @@ class TestQueryset(unittest.TestCase):
|
|||||||
freq = Person.objects.item_frequencies("city", normalize=True, map_reduce=True)
|
freq = Person.objects.item_frequencies("city", normalize=True, map_reduce=True)
|
||||||
assert freq == {"CRB": 0.5, None: 0.5}
|
assert freq == {"CRB": 0.5, None: 0.5}
|
||||||
|
|
||||||
@requires_mongodb_lt_42
|
|
||||||
def test_item_frequencies_with_null_embedded(self):
|
def test_item_frequencies_with_null_embedded(self):
|
||||||
class Data(EmbeddedDocument):
|
class Data(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -3146,12 +3162,11 @@ class TestQueryset(unittest.TestCase):
|
|||||||
p.save()
|
p.save()
|
||||||
|
|
||||||
ot = Person.objects.item_frequencies("extra.tag", map_reduce=False)
|
ot = Person.objects.item_frequencies("extra.tag", map_reduce=False)
|
||||||
assert ot == {None: 1.0, "friend": 1.0}
|
assert ot == {None: 1.0, u"friend": 1.0}
|
||||||
|
|
||||||
ot = Person.objects.item_frequencies("extra.tag", map_reduce=True)
|
ot = Person.objects.item_frequencies("extra.tag", map_reduce=True)
|
||||||
assert ot == {None: 1.0, "friend": 1.0}
|
assert ot == {None: 1.0, u"friend": 1.0}
|
||||||
|
|
||||||
@requires_mongodb_lt_42
|
|
||||||
def test_item_frequencies_with_0_values(self):
|
def test_item_frequencies_with_0_values(self):
|
||||||
class Test(Document):
|
class Test(Document):
|
||||||
val = IntField()
|
val = IntField()
|
||||||
@@ -3166,7 +3181,6 @@ class TestQueryset(unittest.TestCase):
|
|||||||
ot = Test.objects.item_frequencies("val", map_reduce=False)
|
ot = Test.objects.item_frequencies("val", map_reduce=False)
|
||||||
assert ot == {0: 1}
|
assert ot == {0: 1}
|
||||||
|
|
||||||
@requires_mongodb_lt_42
|
|
||||||
def test_item_frequencies_with_False_values(self):
|
def test_item_frequencies_with_False_values(self):
|
||||||
class Test(Document):
|
class Test(Document):
|
||||||
val = BooleanField()
|
val = BooleanField()
|
||||||
@@ -3181,17 +3195,16 @@ class TestQueryset(unittest.TestCase):
|
|||||||
ot = Test.objects.item_frequencies("val", map_reduce=False)
|
ot = Test.objects.item_frequencies("val", map_reduce=False)
|
||||||
assert ot == {False: 1}
|
assert ot == {False: 1}
|
||||||
|
|
||||||
@requires_mongodb_lt_42
|
|
||||||
def test_item_frequencies_normalize(self):
|
def test_item_frequencies_normalize(self):
|
||||||
class Test(Document):
|
class Test(Document):
|
||||||
val = IntField()
|
val = IntField()
|
||||||
|
|
||||||
Test.drop_collection()
|
Test.drop_collection()
|
||||||
|
|
||||||
for _ in range(50):
|
for i in range(50):
|
||||||
Test(val=1).save()
|
Test(val=1).save()
|
||||||
|
|
||||||
for _ in range(20):
|
for i in range(20):
|
||||||
Test(val=2).save()
|
Test(val=2).save()
|
||||||
|
|
||||||
freqs = Test.objects.item_frequencies("val", map_reduce=False, normalize=True)
|
freqs = Test.objects.item_frequencies("val", map_reduce=False, normalize=True)
|
||||||
@@ -3201,7 +3214,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert freqs == {1: 50.0 / 70, 2: 20.0 / 70}
|
assert freqs == {1: 50.0 / 70, 2: 20.0 / 70}
|
||||||
|
|
||||||
def test_average(self):
|
def test_average(self):
|
||||||
"""Ensure that field can be averaged correctly."""
|
"""Ensure that field can be averaged correctly.
|
||||||
|
"""
|
||||||
self.Person(name="person", age=0).save()
|
self.Person(name="person", age=0).save()
|
||||||
assert int(self.Person.objects.average("age")) == 0
|
assert int(self.Person.objects.average("age")) == 0
|
||||||
|
|
||||||
@@ -3224,7 +3238,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
for i, weight in enumerate(ages):
|
for i, weight in enumerate(ages):
|
||||||
self.Person(
|
self.Person(
|
||||||
name=f"test meta{i}", person_meta=self.PersonMeta(weight=weight)
|
name="test meta%i", person_meta=self.PersonMeta(weight=weight)
|
||||||
).save()
|
).save()
|
||||||
|
|
||||||
assert (
|
assert (
|
||||||
@@ -3241,7 +3255,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert self.Person.objects.filter(age__gte=50).average("age") == avg
|
assert self.Person.objects.filter(age__gte=50).average("age") == avg
|
||||||
|
|
||||||
def test_sum(self):
|
def test_sum(self):
|
||||||
"""Ensure that field can be summed over correctly."""
|
"""Ensure that field can be summed over correctly.
|
||||||
|
"""
|
||||||
ages = [23, 54, 12, 94, 27]
|
ages = [23, 54, 12, 94, 27]
|
||||||
for i, age in enumerate(ages):
|
for i, age in enumerate(ages):
|
||||||
self.Person(name="test%s" % i, age=age).save()
|
self.Person(name="test%s" % i, age=age).save()
|
||||||
@@ -3263,7 +3278,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
# test summing over a filtered queryset
|
# test summing over a filtered queryset
|
||||||
assert self.Person.objects.filter(age__gte=50).sum("age") == sum(
|
assert self.Person.objects.filter(age__gte=50).sum("age") == sum(
|
||||||
a for a in ages if a >= 50
|
[a for a in ages if a >= 50]
|
||||||
)
|
)
|
||||||
|
|
||||||
def test_sum_over_db_field(self):
|
def test_sum_over_db_field(self):
|
||||||
@@ -3391,21 +3406,19 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert Doc.objects.sum("values") == 1360
|
assert Doc.objects.sum("values") == 1360
|
||||||
|
|
||||||
def test_distinct(self):
|
def test_distinct(self):
|
||||||
"""Ensure that the QuerySet.distinct method works."""
|
"""Ensure that the QuerySet.distinct method works.
|
||||||
|
"""
|
||||||
self.Person(name="Mr Orange", age=20).save()
|
self.Person(name="Mr Orange", age=20).save()
|
||||||
self.Person(name="Mr White", age=20).save()
|
self.Person(name="Mr White", age=20).save()
|
||||||
self.Person(name="Mr Orange", age=30).save()
|
self.Person(name="Mr Orange", age=30).save()
|
||||||
self.Person(name="Mr Pink", age=30).save()
|
self.Person(name="Mr Pink", age=30).save()
|
||||||
assert set(self.Person.objects.distinct("name")) == {
|
assert set(self.Person.objects.distinct("name")) == set(
|
||||||
"Mr Orange",
|
["Mr Orange", "Mr White", "Mr Pink"]
|
||||||
"Mr White",
|
)
|
||||||
"Mr Pink",
|
assert set(self.Person.objects.distinct("age")) == set([20, 30])
|
||||||
}
|
assert set(self.Person.objects(age=30).distinct("name")) == set(
|
||||||
assert set(self.Person.objects.distinct("age")) == {20, 30}
|
["Mr Orange", "Mr Pink"]
|
||||||
assert set(self.Person.objects(age=30).distinct("name")) == {
|
)
|
||||||
"Mr Orange",
|
|
||||||
"Mr Pink",
|
|
||||||
}
|
|
||||||
|
|
||||||
def test_distinct_handles_references(self):
|
def test_distinct_handles_references(self):
|
||||||
class Foo(Document):
|
class Foo(Document):
|
||||||
@@ -3466,8 +3479,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert count == 1
|
assert count == 1
|
||||||
|
|
||||||
News(
|
News(
|
||||||
title="As eleições no Brasil já estão em planejamento",
|
title=u"As eleições no Brasil já estão em planejamento",
|
||||||
content="A candidata dilma roussef já começa o teu planejamento",
|
content=u"A candidata dilma roussef já começa o teu planejamento",
|
||||||
is_active=False,
|
is_active=False,
|
||||||
).save()
|
).save()
|
||||||
|
|
||||||
@@ -3534,7 +3547,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert Foo.objects.distinct("bar") == [bar]
|
assert Foo.objects.distinct("bar") == [bar]
|
||||||
|
|
||||||
def test_distinct_handles_db_field(self):
|
def test_distinct_handles_db_field(self):
|
||||||
"""Ensure that distinct resolves field name to db_field as expected."""
|
"""Ensure that distinct resolves field name to db_field as expected.
|
||||||
|
"""
|
||||||
|
|
||||||
class Product(Document):
|
class Product(Document):
|
||||||
product_id = IntField(db_field="pid")
|
product_id = IntField(db_field="pid")
|
||||||
@@ -3545,8 +3559,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Product(product_id=2).save()
|
Product(product_id=2).save()
|
||||||
Product(product_id=1).save()
|
Product(product_id=1).save()
|
||||||
|
|
||||||
assert set(Product.objects.distinct("product_id")) == {1, 2}
|
assert set(Product.objects.distinct("product_id")) == set([1, 2])
|
||||||
assert set(Product.objects.distinct("pid")) == {1, 2}
|
assert set(Product.objects.distinct("pid")) == set([1, 2])
|
||||||
|
|
||||||
Product.drop_collection()
|
Product.drop_collection()
|
||||||
|
|
||||||
@@ -3568,8 +3582,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Book.objects.create(title="The Stories", authors=[mark_twain, john_tolkien])
|
Book.objects.create(title="The Stories", authors=[mark_twain, john_tolkien])
|
||||||
|
|
||||||
authors = Book.objects.distinct("authors")
|
authors = Book.objects.distinct("authors")
|
||||||
authors_names = {author.name for author in authors}
|
assert authors == [mark_twain, john_tolkien]
|
||||||
assert authors_names == {mark_twain.name, john_tolkien.name}
|
|
||||||
|
|
||||||
def test_distinct_ListField_EmbeddedDocumentField_EmbeddedDocumentField(self):
|
def test_distinct_ListField_EmbeddedDocumentField_EmbeddedDocumentField(self):
|
||||||
class Continent(EmbeddedDocument):
|
class Continent(EmbeddedDocument):
|
||||||
@@ -3606,8 +3619,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert country_list == [scotland, tibet]
|
assert country_list == [scotland, tibet]
|
||||||
|
|
||||||
continent_list = Book.objects.distinct("authors.country.continent")
|
continent_list = Book.objects.distinct("authors.country.continent")
|
||||||
continent_list_names = {c.continent_name for c in continent_list}
|
assert continent_list == [europe, asia]
|
||||||
assert continent_list_names == {europe.continent_name, asia.continent_name}
|
|
||||||
|
|
||||||
def test_distinct_ListField_ReferenceField(self):
|
def test_distinct_ListField_ReferenceField(self):
|
||||||
class Bar(Document):
|
class Bar(Document):
|
||||||
@@ -3632,7 +3644,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert Foo.objects.distinct("bar_lst") == [bar_1, bar_2]
|
assert Foo.objects.distinct("bar_lst") == [bar_1, bar_2]
|
||||||
|
|
||||||
def test_custom_manager(self):
|
def test_custom_manager(self):
|
||||||
"""Ensure that custom QuerySetManager instances work as expected."""
|
"""Ensure that custom QuerySetManager instances work as expected.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
tags = ListField(StringField())
|
tags = ListField(StringField())
|
||||||
@@ -3733,7 +3746,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert 1 == Bar.objects.count()
|
assert 1 == Bar.objects.count()
|
||||||
|
|
||||||
def test_query_value_conversion(self):
|
def test_query_value_conversion(self):
|
||||||
"""Ensure that query values are properly converted when necessary."""
|
"""Ensure that query values are properly converted when necessary.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
author = ReferenceField(self.Person)
|
author = ReferenceField(self.Person)
|
||||||
@@ -3759,7 +3773,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
def test_update_value_conversion(self):
|
def test_update_value_conversion(self):
|
||||||
"""Ensure that values used in updates are converted before use."""
|
"""Ensure that values used in updates are converted before use.
|
||||||
|
"""
|
||||||
|
|
||||||
class Group(Document):
|
class Group(Document):
|
||||||
members = ListField(ReferenceField(self.Person))
|
members = ListField(ReferenceField(self.Person))
|
||||||
@@ -3784,7 +3799,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Group.drop_collection()
|
Group.drop_collection()
|
||||||
|
|
||||||
def test_bulk(self):
|
def test_bulk(self):
|
||||||
"""Ensure bulk querying by object id returns a proper dict."""
|
"""Ensure bulk querying by object id returns a proper dict.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -3822,7 +3838,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
self.Person.drop_collection()
|
self.Person.drop_collection()
|
||||||
|
|
||||||
def test_custom_querysets(self):
|
def test_custom_querysets(self):
|
||||||
"""Ensure that custom QuerySet classes may be used."""
|
"""Ensure that custom QuerySet classes may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class CustomQuerySet(QuerySet):
|
class CustomQuerySet(QuerySet):
|
||||||
def not_empty(self):
|
def not_empty(self):
|
||||||
@@ -3842,7 +3859,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Post.drop_collection()
|
Post.drop_collection()
|
||||||
|
|
||||||
def test_custom_querysets_set_manager_directly(self):
|
def test_custom_querysets_set_manager_directly(self):
|
||||||
"""Ensure that custom QuerySet classes may be used."""
|
"""Ensure that custom QuerySet classes may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class CustomQuerySet(QuerySet):
|
class CustomQuerySet(QuerySet):
|
||||||
def not_empty(self):
|
def not_empty(self):
|
||||||
@@ -3865,7 +3883,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Post.drop_collection()
|
Post.drop_collection()
|
||||||
|
|
||||||
def test_custom_querysets_managers_directly(self):
|
def test_custom_querysets_managers_directly(self):
|
||||||
"""Ensure that custom QuerySet classes may be used."""
|
"""Ensure that custom QuerySet classes may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class CustomQuerySetManager(QuerySetManager):
|
class CustomQuerySetManager(QuerySetManager):
|
||||||
@staticmethod
|
@staticmethod
|
||||||
@@ -3886,7 +3905,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Post.drop_collection()
|
Post.drop_collection()
|
||||||
|
|
||||||
def test_custom_querysets_inherited(self):
|
def test_custom_querysets_inherited(self):
|
||||||
"""Ensure that custom QuerySet classes may be used."""
|
"""Ensure that custom QuerySet classes may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class CustomQuerySet(QuerySet):
|
class CustomQuerySet(QuerySet):
|
||||||
def not_empty(self):
|
def not_empty(self):
|
||||||
@@ -3908,7 +3928,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Post.drop_collection()
|
Post.drop_collection()
|
||||||
|
|
||||||
def test_custom_querysets_inherited_direct(self):
|
def test_custom_querysets_inherited_direct(self):
|
||||||
"""Ensure that custom QuerySet classes may be used."""
|
"""Ensure that custom QuerySet classes may be used.
|
||||||
|
"""
|
||||||
|
|
||||||
class CustomQuerySet(QuerySet):
|
class CustomQuerySet(QuerySet):
|
||||||
def not_empty(self):
|
def not_empty(self):
|
||||||
@@ -3969,7 +3990,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert A.objects(b=[{"c": "c"}]).count() == 0
|
assert A.objects(b=[{"c": "c"}]).count() == 0
|
||||||
|
|
||||||
def test_call_after_limits_set(self):
|
def test_call_after_limits_set(self):
|
||||||
"""Ensure that re-filtering after slicing works"""
|
"""Ensure that re-filtering after slicing works
|
||||||
|
"""
|
||||||
|
|
||||||
class Post(Document):
|
class Post(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -3985,7 +4007,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Post.drop_collection()
|
Post.drop_collection()
|
||||||
|
|
||||||
def test_order_then_filter(self):
|
def test_order_then_filter(self):
|
||||||
"""Ensure that ordering still works after filtering."""
|
"""Ensure that ordering still works after filtering.
|
||||||
|
"""
|
||||||
|
|
||||||
class Number(Document):
|
class Number(Document):
|
||||||
n = IntField()
|
n = IntField()
|
||||||
@@ -4002,7 +4025,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Number.drop_collection()
|
Number.drop_collection()
|
||||||
|
|
||||||
def test_clone(self):
|
def test_clone(self):
|
||||||
"""Ensure that cloning clones complex querysets"""
|
"""Ensure that cloning clones complex querysets
|
||||||
|
"""
|
||||||
|
|
||||||
class Number(Document):
|
class Number(Document):
|
||||||
n = IntField()
|
n = IntField()
|
||||||
@@ -4031,7 +4055,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Number.drop_collection()
|
Number.drop_collection()
|
||||||
|
|
||||||
def test_clone_retains_settings(self):
|
def test_clone_retains_settings(self):
|
||||||
"""Ensure that cloning retains the read_preference and read_concern"""
|
"""Ensure that cloning retains the read_preference and read_concern
|
||||||
|
"""
|
||||||
|
|
||||||
class Number(Document):
|
class Number(Document):
|
||||||
n = IntField()
|
n = IntField()
|
||||||
@@ -4056,7 +4081,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Number.drop_collection()
|
Number.drop_collection()
|
||||||
|
|
||||||
def test_using(self):
|
def test_using(self):
|
||||||
"""Ensure that switching databases for a queryset is possible"""
|
"""Ensure that switching databases for a queryset is possible
|
||||||
|
"""
|
||||||
|
|
||||||
class Number2(Document):
|
class Number2(Document):
|
||||||
n = IntField()
|
n = IntField()
|
||||||
@@ -4108,7 +4134,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Number.drop_collection()
|
Number.drop_collection()
|
||||||
|
|
||||||
def test_order_works_with_primary(self):
|
def test_order_works_with_primary(self):
|
||||||
"""Ensure that order_by and primary work."""
|
"""Ensure that order_by and primary work.
|
||||||
|
"""
|
||||||
|
|
||||||
class Number(Document):
|
class Number(Document):
|
||||||
n = IntField(primary_key=True)
|
n = IntField(primary_key=True)
|
||||||
@@ -4127,7 +4154,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Number.drop_collection()
|
Number.drop_collection()
|
||||||
|
|
||||||
def test_ensure_index(self):
|
def test_ensure_index(self):
|
||||||
"""Ensure that manual creation of indexes works."""
|
"""Ensure that manual creation of indexes works.
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(Document):
|
class Comment(Document):
|
||||||
message = StringField()
|
message = StringField()
|
||||||
@@ -4143,7 +4171,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert ([("_cls", 1), ("message", 1)], False, False) in info
|
assert ([("_cls", 1), ("message", 1)], False, False) in info
|
||||||
|
|
||||||
def test_where(self):
|
def test_where(self):
|
||||||
"""Ensure that where clauses work."""
|
"""Ensure that where clauses work.
|
||||||
|
"""
|
||||||
|
|
||||||
class IntPair(Document):
|
class IntPair(Document):
|
||||||
fielda = IntField()
|
fielda = IntField()
|
||||||
@@ -4244,15 +4273,15 @@ class TestQueryset(unittest.TestCase):
|
|||||||
ulist = list(UserDoc.objects.scalar("name", "age"))
|
ulist = list(UserDoc.objects.scalar("name", "age"))
|
||||||
|
|
||||||
assert ulist == [
|
assert ulist == [
|
||||||
("Wilson Jr", 19),
|
(u"Wilson Jr", 19),
|
||||||
("Wilson", 43),
|
(u"Wilson", 43),
|
||||||
("Eliana", 37),
|
(u"Eliana", 37),
|
||||||
("Tayza", 15),
|
(u"Tayza", 15),
|
||||||
]
|
]
|
||||||
|
|
||||||
ulist = list(UserDoc.objects.scalar("name").order_by("age"))
|
ulist = list(UserDoc.objects.scalar("name").order_by("age"))
|
||||||
|
|
||||||
assert ulist == [("Tayza"), ("Wilson Jr"), ("Eliana"), ("Wilson")]
|
assert ulist == [(u"Tayza"), (u"Wilson Jr"), (u"Eliana"), (u"Wilson")]
|
||||||
|
|
||||||
def test_scalar_embedded(self):
|
def test_scalar_embedded(self):
|
||||||
class Profile(EmbeddedDocument):
|
class Profile(EmbeddedDocument):
|
||||||
@@ -4291,7 +4320,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
|
|
||||||
assert list(
|
assert list(
|
||||||
Person.objects.order_by("profile__age").scalar("profile__name")
|
Person.objects.order_by("profile__age").scalar("profile__name")
|
||||||
) == ["Wilson Jr", "Gabriel Falcao", "Lincoln de souza", "Walter cruz"]
|
) == [u"Wilson Jr", u"Gabriel Falcao", u"Lincoln de souza", u"Walter cruz"]
|
||||||
|
|
||||||
ulist = list(
|
ulist = list(
|
||||||
Person.objects.order_by("locale.city").scalar(
|
Person.objects.order_by("locale.city").scalar(
|
||||||
@@ -4299,10 +4328,10 @@ class TestQueryset(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
)
|
)
|
||||||
assert ulist == [
|
assert ulist == [
|
||||||
("Lincoln de souza", 28, "Belo Horizonte"),
|
(u"Lincoln de souza", 28, u"Belo Horizonte"),
|
||||||
("Walter cruz", 30, "Brasilia"),
|
(u"Walter cruz", 30, u"Brasilia"),
|
||||||
("Wilson Jr", 19, "Corumba-GO"),
|
(u"Wilson Jr", 19, u"Corumba-GO"),
|
||||||
("Gabriel Falcao", 23, "New York"),
|
(u"Gabriel Falcao", 23, u"New York"),
|
||||||
]
|
]
|
||||||
|
|
||||||
def test_scalar_decimal(self):
|
def test_scalar_decimal(self):
|
||||||
@@ -4316,7 +4345,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Person(name="Wilson Jr", rating=Decimal("1.0")).save()
|
Person(name="Wilson Jr", rating=Decimal("1.0")).save()
|
||||||
|
|
||||||
ulist = list(Person.objects.scalar("name", "rating"))
|
ulist = list(Person.objects.scalar("name", "rating"))
|
||||||
assert ulist == [("Wilson Jr", Decimal("1.0"))]
|
assert ulist == [(u"Wilson Jr", Decimal("1.0"))]
|
||||||
|
|
||||||
def test_scalar_reference_field(self):
|
def test_scalar_reference_field(self):
|
||||||
class State(Document):
|
class State(Document):
|
||||||
@@ -4335,7 +4364,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Person(name="Wilson JR", state=s1).save()
|
Person(name="Wilson JR", state=s1).save()
|
||||||
|
|
||||||
plist = list(Person.objects.scalar("name", "state"))
|
plist = list(Person.objects.scalar("name", "state"))
|
||||||
assert plist == [("Wilson JR", s1)]
|
assert plist == [(u"Wilson JR", s1)]
|
||||||
|
|
||||||
def test_scalar_generic_reference_field(self):
|
def test_scalar_generic_reference_field(self):
|
||||||
class State(Document):
|
class State(Document):
|
||||||
@@ -4354,7 +4383,7 @@ class TestQueryset(unittest.TestCase):
|
|||||||
Person(name="Wilson JR", state=s1).save()
|
Person(name="Wilson JR", state=s1).save()
|
||||||
|
|
||||||
plist = list(Person.objects.scalar("name", "state"))
|
plist = list(Person.objects.scalar("name", "state"))
|
||||||
assert plist == [("Wilson JR", s1)]
|
assert plist == [(u"Wilson JR", s1)]
|
||||||
|
|
||||||
def test_generic_reference_field_with_only_and_as_pymongo(self):
|
def test_generic_reference_field_with_only_and_as_pymongo(self):
|
||||||
class TestPerson(Document):
|
class TestPerson(Document):
|
||||||
@@ -4422,7 +4451,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert list(val) == [("test", "test value")]
|
assert list(val) == [("test", "test value")]
|
||||||
|
|
||||||
def test_scalar_cursor_behaviour(self):
|
def test_scalar_cursor_behaviour(self):
|
||||||
"""Ensure that a query returns a valid set of results."""
|
"""Ensure that a query returns a valid set of results.
|
||||||
|
"""
|
||||||
person1 = self.Person(name="User A", age=20)
|
person1 = self.Person(name="User A", age=20)
|
||||||
person1.save()
|
person1.save()
|
||||||
person2 = self.Person(name="User B", age=30)
|
person2 = self.Person(name="User B", age=30)
|
||||||
@@ -4639,7 +4669,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert [b3] == ak
|
assert [b3] == ak
|
||||||
|
|
||||||
def test_upsert_includes_cls(self):
|
def test_upsert_includes_cls(self):
|
||||||
"""Upserts should include _cls information for inheritable classes"""
|
"""Upserts should include _cls information for inheritable classes
|
||||||
|
"""
|
||||||
|
|
||||||
class Test(Document):
|
class Test(Document):
|
||||||
test = StringField()
|
test = StringField()
|
||||||
@@ -4885,11 +4916,13 @@ class TestQueryset(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
results = User.objects.as_pymongo()
|
results = User.objects.as_pymongo()
|
||||||
assert set(results[0].keys()) == {"_id", "name", "age", "price"}
|
assert set(results[0].keys()) == set(["_id", "name", "age", "price"])
|
||||||
assert set(results[1].keys()) == {"_id", "name", "age", "price", "last_login"}
|
assert set(results[1].keys()) == set(
|
||||||
|
["_id", "name", "age", "price", "last_login"]
|
||||||
|
)
|
||||||
|
|
||||||
results = User.objects.only("id", "name").as_pymongo()
|
results = User.objects.only("id", "name").as_pymongo()
|
||||||
assert set(results[0].keys()) == {"_id", "name"}
|
assert set(results[0].keys()) == set(["_id", "name"])
|
||||||
|
|
||||||
users = User.objects.only("name", "price").as_pymongo()
|
users = User.objects.only("name", "price").as_pymongo()
|
||||||
results = list(users)
|
results = list(users)
|
||||||
@@ -5333,7 +5366,8 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert obj.__class__ == C
|
assert obj.__class__ == C
|
||||||
|
|
||||||
def test_query_generic_embedded_document(self):
|
def test_query_generic_embedded_document(self):
|
||||||
"""Ensure that querying sub field on generic_embedded_field works"""
|
"""Ensure that querying sub field on generic_embedded_field works
|
||||||
|
"""
|
||||||
|
|
||||||
class A(EmbeddedDocument):
|
class A(EmbeddedDocument):
|
||||||
a_name = StringField()
|
a_name = StringField()
|
||||||
@@ -5521,12 +5555,12 @@ class TestQueryset(unittest.TestCase):
|
|||||||
assert Person.objects._has_data(), "Cursor has data and returned False"
|
assert Person.objects._has_data(), "Cursor has data and returned False"
|
||||||
|
|
||||||
def test_delete_count(self):
|
def test_delete_count(self):
|
||||||
[self.Person(name=f"User {i}", age=i * 10).save() for i in range(1, 4)]
|
[self.Person(name="User {0}".format(i), age=i * 10).save() for i in range(1, 4)]
|
||||||
assert (
|
assert (
|
||||||
self.Person.objects().delete() == 3
|
self.Person.objects().delete() == 3
|
||||||
) # test ordinary QuerySey delete count
|
) # test ordinary QuerySey delete count
|
||||||
|
|
||||||
[self.Person(name=f"User {i}", age=i * 10).save() for i in range(1, 4)]
|
[self.Person(name="User {0}".format(i), age=i * 10).save() for i in range(1, 4)]
|
||||||
|
|
||||||
assert (
|
assert (
|
||||||
self.Person.objects().skip(1).delete() == 2
|
self.Person.objects().skip(1).delete() == 2
|
||||||
@@ -5676,31 +5710,6 @@ class TestQueryset(unittest.TestCase):
|
|||||||
qs = self.Person.objects().timeout(False)
|
qs = self.Person.objects().timeout(False)
|
||||||
assert qs._cursor_args == {"no_cursor_timeout": True}
|
assert qs._cursor_args == {"no_cursor_timeout": True}
|
||||||
|
|
||||||
@requires_mongodb_gte_44
|
|
||||||
def test_allow_disk_use(self):
|
|
||||||
qs = self.Person.objects()
|
|
||||||
assert qs._cursor_args == {}
|
|
||||||
|
|
||||||
qs = self.Person.objects().allow_disk_use(False)
|
|
||||||
assert qs._cursor_args == {}
|
|
||||||
|
|
||||||
qs = self.Person.objects().allow_disk_use(True)
|
|
||||||
assert qs._cursor_args == {"allow_disk_use": True}
|
|
||||||
|
|
||||||
# Test if allow_disk_use changes the results
|
|
||||||
self.Person.drop_collection()
|
|
||||||
self.Person.objects.create(name="Foo", age=12)
|
|
||||||
self.Person.objects.create(name="Baz", age=17)
|
|
||||||
self.Person.objects.create(name="Bar", age=13)
|
|
||||||
|
|
||||||
qs_disk = self.Person.objects().order_by("age").allow_disk_use(True)
|
|
||||||
qs = self.Person.objects().order_by("age")
|
|
||||||
|
|
||||||
assert qs_disk.count() == qs.count()
|
|
||||||
|
|
||||||
for index in range(qs_disk.count()):
|
|
||||||
assert qs_disk[index] == qs[index]
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -64,7 +64,7 @@ class TestQuerysetAggregate(MongoDBTestCase):
|
|||||||
|
|
||||||
pipeline = [{"$match": {"name": "Isabella Luanna"}}]
|
pipeline = [{"$match": {"name": "Isabella Luanna"}}]
|
||||||
data = Person.objects().aggregate(pipeline)
|
data = Person.objects().aggregate(pipeline)
|
||||||
assert list(data) == [{"_id": p1.pk, "age": 16, "name": "Isabella Luanna"}]
|
assert list(data) == [{u"_id": p1.pk, u"age": 16, u"name": u"Isabella Luanna"}]
|
||||||
|
|
||||||
def test_queryset_aggregation_with_skip(self):
|
def test_queryset_aggregation_with_skip(self):
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
@@ -248,34 +248,6 @@ class TestQuerysetAggregate(MongoDBTestCase):
|
|||||||
|
|
||||||
assert list(data) == [{"_id": p1.pk, "name": "ISABELLA LUANNA"}]
|
assert list(data) == [{"_id": p1.pk, "name": "ISABELLA LUANNA"}]
|
||||||
|
|
||||||
def test_queryset_aggregation_geonear_aggregation_on_pointfield(self):
|
|
||||||
"""test ensures that $geonear can be used as a 1-stage pipeline and that
|
|
||||||
MongoEngine does not interfer with such pipeline (#2473)
|
|
||||||
"""
|
|
||||||
|
|
||||||
class Aggr(Document):
|
|
||||||
name = StringField()
|
|
||||||
c = PointField()
|
|
||||||
|
|
||||||
Aggr.drop_collection()
|
|
||||||
|
|
||||||
agg1 = Aggr(name="X", c=[10.634584, 35.8245029]).save()
|
|
||||||
agg2 = Aggr(name="Y", c=[10.634584, 35.8245029]).save()
|
|
||||||
|
|
||||||
pipeline = [
|
|
||||||
{
|
|
||||||
"$geoNear": {
|
|
||||||
"near": {"type": "Point", "coordinates": [10.634584, 35.8245029]},
|
|
||||||
"distanceField": "c",
|
|
||||||
"spherical": True,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
]
|
|
||||||
assert list(Aggr.objects.aggregate(*pipeline)) == [
|
|
||||||
{"_id": agg1.id, "c": 0.0, "name": "X"},
|
|
||||||
{"_id": agg2.id, "c": 0.0, "name": "Y"},
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
@@ -1,7 +1,7 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
import pytest
|
|
||||||
from bson.son import SON
|
from bson.son import SON
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.queryset import Q, transform
|
from mongoengine.queryset import Q, transform
|
||||||
@@ -12,7 +12,8 @@ class TestTransform(unittest.TestCase):
|
|||||||
connect(db="mongoenginetest")
|
connect(db="mongoenginetest")
|
||||||
|
|
||||||
def test_transform_query(self):
|
def test_transform_query(self):
|
||||||
"""Ensure that the _transform_query function operates correctly."""
|
"""Ensure that the _transform_query function operates correctly.
|
||||||
|
"""
|
||||||
assert transform.query(name="test", age=30) == {"name": "test", "age": 30}
|
assert transform.query(name="test", age=30) == {"name": "test", "age": 30}
|
||||||
assert transform.query(age__lt=30) == {"age": {"$lt": 30}}
|
assert transform.query(age__lt=30) == {"age": {"$lt": 30}}
|
||||||
assert transform.query(age__gt=20, age__lt=50) == {
|
assert transform.query(age__gt=20, age__lt=50) == {
|
||||||
@@ -87,7 +88,8 @@ class TestTransform(unittest.TestCase):
|
|||||||
assert update == {"$set": {"tags": ["mongo", "db"]}}
|
assert update == {"$set": {"tags": ["mongo", "db"]}}
|
||||||
|
|
||||||
def test_query_field_name(self):
|
def test_query_field_name(self):
|
||||||
"""Ensure that the correct field name is used when querying."""
|
"""Ensure that the correct field name is used when querying.
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(EmbeddedDocument):
|
class Comment(EmbeddedDocument):
|
||||||
content = StringField(db_field="commentContent")
|
content = StringField(db_field="commentContent")
|
||||||
@@ -104,17 +106,18 @@ class TestTransform(unittest.TestCase):
|
|||||||
post = BlogPost(**data)
|
post = BlogPost(**data)
|
||||||
post.save()
|
post.save()
|
||||||
|
|
||||||
qs = BlogPost.objects(title=data["title"])
|
assert "postTitle" in BlogPost.objects(title=data["title"])._query
|
||||||
assert qs._query == {"postTitle": data["title"]}
|
assert not ("title" in BlogPost.objects(title=data["title"])._query)
|
||||||
assert qs.count() == 1
|
assert BlogPost.objects(title=data["title"]).count() == 1
|
||||||
|
|
||||||
qs = BlogPost.objects(pk=post.id)
|
assert "_id" in BlogPost.objects(pk=post.id)._query
|
||||||
assert qs._query == {"_id": post.id}
|
assert BlogPost.objects(pk=post.id).count() == 1
|
||||||
assert qs.count() == 1
|
|
||||||
|
|
||||||
qs = BlogPost.objects(comments__content="test")
|
assert (
|
||||||
assert qs._query == {"postComments.commentContent": "test"}
|
"postComments.commentContent"
|
||||||
assert qs.count() == 1
|
in BlogPost.objects(comments__content="test")._query
|
||||||
|
)
|
||||||
|
assert BlogPost.objects(comments__content="test").count() == 1
|
||||||
|
|
||||||
BlogPost.drop_collection()
|
BlogPost.drop_collection()
|
||||||
|
|
||||||
@@ -327,7 +330,7 @@ class TestTransform(unittest.TestCase):
|
|||||||
word = Word(word="abc", index=1)
|
word = Word(word="abc", index=1)
|
||||||
update = transform.update(MainDoc, pull__content__text=word)
|
update = transform.update(MainDoc, pull__content__text=word)
|
||||||
assert update == {
|
assert update == {
|
||||||
"$pull": {"content.text": SON([("word", "abc"), ("index", 1)])}
|
"$pull": {"content.text": SON([("word", u"abc"), ("index", 1)])}
|
||||||
}
|
}
|
||||||
|
|
||||||
update = transform.update(MainDoc, pull__content__heading="xyz")
|
update = transform.update(MainDoc, pull__content__heading="xyz")
|
||||||
|
@@ -2,8 +2,8 @@ import datetime
|
|||||||
import re
|
import re
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
import pytest
|
|
||||||
from bson import ObjectId
|
from bson import ObjectId
|
||||||
|
import pytest
|
||||||
|
|
||||||
from mongoengine import *
|
from mongoengine import *
|
||||||
from mongoengine.errors import InvalidQueryError
|
from mongoengine.errors import InvalidQueryError
|
||||||
@@ -23,7 +23,8 @@ class TestQ(unittest.TestCase):
|
|||||||
self.Person = Person
|
self.Person = Person
|
||||||
|
|
||||||
def test_empty_q(self):
|
def test_empty_q(self):
|
||||||
"""Ensure that empty Q objects won't hurt."""
|
"""Ensure that empty Q objects won't hurt.
|
||||||
|
"""
|
||||||
q1 = Q()
|
q1 = Q()
|
||||||
q2 = Q(age__gte=18)
|
q2 = Q(age__gte=18)
|
||||||
q3 = Q()
|
q3 = Q()
|
||||||
@@ -57,7 +58,8 @@ class TestQ(unittest.TestCase):
|
|||||||
assert Post.objects.filter(Q(created_user=user)).count() == 1
|
assert Post.objects.filter(Q(created_user=user)).count() == 1
|
||||||
|
|
||||||
def test_and_combination(self):
|
def test_and_combination(self):
|
||||||
"""Ensure that Q-objects correctly AND together."""
|
"""Ensure that Q-objects correctly AND together.
|
||||||
|
"""
|
||||||
|
|
||||||
class TestDoc(Document):
|
class TestDoc(Document):
|
||||||
x = IntField()
|
x = IntField()
|
||||||
@@ -87,7 +89,8 @@ class TestQ(unittest.TestCase):
|
|||||||
assert query.to_query(TestDoc) == mongo_query
|
assert query.to_query(TestDoc) == mongo_query
|
||||||
|
|
||||||
def test_or_combination(self):
|
def test_or_combination(self):
|
||||||
"""Ensure that Q-objects correctly OR together."""
|
"""Ensure that Q-objects correctly OR together.
|
||||||
|
"""
|
||||||
|
|
||||||
class TestDoc(Document):
|
class TestDoc(Document):
|
||||||
x = IntField()
|
x = IntField()
|
||||||
@@ -98,7 +101,8 @@ class TestQ(unittest.TestCase):
|
|||||||
assert query == {"$or": [{"x": {"$lt": 3}}, {"x": {"$gt": 7}}]}
|
assert query == {"$or": [{"x": {"$lt": 3}}, {"x": {"$gt": 7}}]}
|
||||||
|
|
||||||
def test_and_or_combination(self):
|
def test_and_or_combination(self):
|
||||||
"""Ensure that Q-objects handle ANDing ORed components."""
|
"""Ensure that Q-objects handle ANDing ORed components.
|
||||||
|
"""
|
||||||
|
|
||||||
class TestDoc(Document):
|
class TestDoc(Document):
|
||||||
x = IntField()
|
x = IntField()
|
||||||
@@ -132,7 +136,8 @@ class TestQ(unittest.TestCase):
|
|||||||
assert 2 == TestDoc.objects(q1 & q2).count()
|
assert 2 == TestDoc.objects(q1 & q2).count()
|
||||||
|
|
||||||
def test_or_and_or_combination(self):
|
def test_or_and_or_combination(self):
|
||||||
"""Ensure that Q-objects handle ORing ANDed ORed components. :)"""
|
"""Ensure that Q-objects handle ORing ANDed ORed components. :)
|
||||||
|
"""
|
||||||
|
|
||||||
class TestDoc(Document):
|
class TestDoc(Document):
|
||||||
x = IntField()
|
x = IntField()
|
||||||
@@ -203,7 +208,8 @@ class TestQ(unittest.TestCase):
|
|||||||
assert test.count() == 3
|
assert test.count() == 3
|
||||||
|
|
||||||
def test_q(self):
|
def test_q(self):
|
||||||
"""Ensure that Q objects may be used to query for documents."""
|
"""Ensure that Q objects may be used to query for documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
title = StringField()
|
title = StringField()
|
||||||
@@ -280,7 +286,8 @@ class TestQ(unittest.TestCase):
|
|||||||
self.Person.objects.filter("user1")
|
self.Person.objects.filter("user1")
|
||||||
|
|
||||||
def test_q_regex(self):
|
def test_q_regex(self):
|
||||||
"""Ensure that Q objects can be queried using regexes."""
|
"""Ensure that Q objects can be queried using regexes.
|
||||||
|
"""
|
||||||
person = self.Person(name="Guido van Rossum")
|
person = self.Person(name="Guido van Rossum")
|
||||||
person.save()
|
person.save()
|
||||||
|
|
||||||
@@ -313,7 +320,8 @@ class TestQ(unittest.TestCase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
def test_q_lists(self):
|
def test_q_lists(self):
|
||||||
"""Ensure that Q objects query ListFields correctly."""
|
"""Ensure that Q objects query ListFields correctly.
|
||||||
|
"""
|
||||||
|
|
||||||
class BlogPost(Document):
|
class BlogPost(Document):
|
||||||
tags = ListField(StringField())
|
tags = ListField(StringField())
|
||||||
|
@@ -1,3 +1,5 @@
|
|||||||
|
import unittest
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import Document
|
from mongoengine import Document
|
||||||
|
@@ -1,11 +1,16 @@
|
|||||||
import datetime
|
import datetime
|
||||||
import unittest
|
|
||||||
|
|
||||||
import pymongo
|
|
||||||
import pytest
|
|
||||||
from bson.tz_util import utc
|
from bson.tz_util import utc
|
||||||
|
import pymongo
|
||||||
|
|
||||||
from pymongo import MongoClient, ReadPreference
|
from pymongo import MongoClient, ReadPreference
|
||||||
from pymongo.errors import InvalidName, OperationFailure
|
from pymongo.errors import InvalidName, OperationFailure
|
||||||
|
import pytest
|
||||||
|
|
||||||
|
try:
|
||||||
|
import unittest2 as unittest
|
||||||
|
except ImportError:
|
||||||
|
import unittest
|
||||||
|
|
||||||
import mongoengine.connection
|
import mongoengine.connection
|
||||||
from mongoengine import (
|
from mongoengine import (
|
||||||
@@ -17,8 +22,8 @@ from mongoengine import (
|
|||||||
register_connection,
|
register_connection,
|
||||||
)
|
)
|
||||||
from mongoengine.connection import (
|
from mongoengine.connection import (
|
||||||
DEFAULT_DATABASE_NAME,
|
|
||||||
ConnectionFailure,
|
ConnectionFailure,
|
||||||
|
DEFAULT_DATABASE_NAME,
|
||||||
disconnect,
|
disconnect,
|
||||||
get_connection,
|
get_connection,
|
||||||
get_db,
|
get_db,
|
||||||
@@ -29,6 +34,18 @@ def get_tz_awareness(connection):
|
|||||||
return connection.codec_options.tz_aware
|
return connection.codec_options.tz_aware
|
||||||
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
import mongomock
|
||||||
|
|
||||||
|
MONGOMOCK_INSTALLED = True
|
||||||
|
except ImportError:
|
||||||
|
MONGOMOCK_INSTALLED = False
|
||||||
|
|
||||||
|
require_mongomock = pytest.mark.skipif(
|
||||||
|
not MONGOMOCK_INSTALLED, reason="you need mongomock installed to run this testcase"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
class ConnectionTest(unittest.TestCase):
|
class ConnectionTest(unittest.TestCase):
|
||||||
@classmethod
|
@classmethod
|
||||||
def setUpClass(cls):
|
def setUpClass(cls):
|
||||||
@@ -177,12 +194,14 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
assert len(mongoengine.connection._connections) == 3
|
assert len(mongoengine.connection._connections) == 3
|
||||||
|
|
||||||
def test_connect_with_invalid_db_name(self):
|
def test_connect_with_invalid_db_name(self):
|
||||||
"""Ensure that connect() method fails fast if db name is invalid"""
|
"""Ensure that connect() method fails fast if db name is invalid
|
||||||
|
"""
|
||||||
with pytest.raises(InvalidName):
|
with pytest.raises(InvalidName):
|
||||||
connect("mongodb://localhost")
|
connect("mongomock://localhost")
|
||||||
|
|
||||||
def test_connect_with_db_name_external(self):
|
def test_connect_with_db_name_external(self):
|
||||||
"""Ensure that connect() works if db name is $external"""
|
"""Ensure that connect() works if db name is $external
|
||||||
|
"""
|
||||||
"""Ensure that the connect() method works properly."""
|
"""Ensure that the connect() method works properly."""
|
||||||
connect("$external")
|
connect("$external")
|
||||||
|
|
||||||
@@ -198,11 +217,112 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
assert isinstance(conn, pymongo.mongo_client.MongoClient)
|
assert isinstance(conn, pymongo.mongo_client.MongoClient)
|
||||||
|
|
||||||
def test_connect_with_invalid_db_name_type(self):
|
def test_connect_with_invalid_db_name_type(self):
|
||||||
"""Ensure that connect() method fails fast if db name has invalid type"""
|
"""Ensure that connect() method fails fast if db name has invalid type
|
||||||
|
"""
|
||||||
with pytest.raises(TypeError):
|
with pytest.raises(TypeError):
|
||||||
non_string_db_name = ["e. g. list instead of a string"]
|
non_string_db_name = ["e. g. list instead of a string"]
|
||||||
connect(non_string_db_name)
|
connect(non_string_db_name)
|
||||||
|
|
||||||
|
@require_mongomock
|
||||||
|
def test_connect_in_mocking(self):
|
||||||
|
"""Ensure that the connect() method works properly in mocking.
|
||||||
|
"""
|
||||||
|
connect("mongoenginetest", host="mongomock://localhost")
|
||||||
|
conn = get_connection()
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect("mongoenginetest2", host="mongomock://localhost", alias="testdb2")
|
||||||
|
conn = get_connection("testdb2")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(
|
||||||
|
"mongoenginetest3",
|
||||||
|
host="mongodb://localhost",
|
||||||
|
is_mock=True,
|
||||||
|
alias="testdb3",
|
||||||
|
)
|
||||||
|
conn = get_connection("testdb3")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect("mongoenginetest4", is_mock=True, alias="testdb4")
|
||||||
|
conn = get_connection("testdb4")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(
|
||||||
|
host="mongodb://localhost:27017/mongoenginetest5",
|
||||||
|
is_mock=True,
|
||||||
|
alias="testdb5",
|
||||||
|
)
|
||||||
|
conn = get_connection("testdb5")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(host="mongomock://localhost:27017/mongoenginetest6", alias="testdb6")
|
||||||
|
conn = get_connection("testdb6")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(
|
||||||
|
host="mongomock://localhost:27017/mongoenginetest7",
|
||||||
|
is_mock=True,
|
||||||
|
alias="testdb7",
|
||||||
|
)
|
||||||
|
conn = get_connection("testdb7")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
@require_mongomock
|
||||||
|
def test_default_database_with_mocking(self):
|
||||||
|
"""Ensure that the default database is correctly set when using mongomock.
|
||||||
|
"""
|
||||||
|
disconnect_all()
|
||||||
|
|
||||||
|
class SomeDocument(Document):
|
||||||
|
pass
|
||||||
|
|
||||||
|
conn = connect(host="mongomock://localhost:27017/mongoenginetest")
|
||||||
|
some_document = SomeDocument()
|
||||||
|
# database won't exist until we save a document
|
||||||
|
some_document.save()
|
||||||
|
assert conn.get_default_database().name == "mongoenginetest"
|
||||||
|
assert conn.list_database_names()[0] == "mongoenginetest"
|
||||||
|
|
||||||
|
@require_mongomock
|
||||||
|
def test_connect_with_host_list(self):
|
||||||
|
"""Ensure that the connect() method works when host is a list
|
||||||
|
|
||||||
|
Uses mongomock to test w/o needing multiple mongod/mongos processes
|
||||||
|
"""
|
||||||
|
connect(host=["mongomock://localhost"])
|
||||||
|
conn = get_connection()
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(host=["mongodb://localhost"], is_mock=True, alias="testdb2")
|
||||||
|
conn = get_connection("testdb2")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(host=["localhost"], is_mock=True, alias="testdb3")
|
||||||
|
conn = get_connection("testdb3")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(
|
||||||
|
host=["mongomock://localhost:27017", "mongomock://localhost:27018"],
|
||||||
|
alias="testdb4",
|
||||||
|
)
|
||||||
|
conn = get_connection("testdb4")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(
|
||||||
|
host=["mongodb://localhost:27017", "mongodb://localhost:27018"],
|
||||||
|
is_mock=True,
|
||||||
|
alias="testdb5",
|
||||||
|
)
|
||||||
|
conn = get_connection("testdb5")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
|
connect(
|
||||||
|
host=["localhost:27017", "localhost:27018"], is_mock=True, alias="testdb6"
|
||||||
|
)
|
||||||
|
conn = get_connection("testdb6")
|
||||||
|
assert isinstance(conn, mongomock.MongoClient)
|
||||||
|
|
||||||
def test_disconnect_cleans_globals(self):
|
def test_disconnect_cleans_globals(self):
|
||||||
"""Ensure that the disconnect() method cleans the globals objects"""
|
"""Ensure that the disconnect() method cleans the globals objects"""
|
||||||
connections = mongoengine.connection._connections
|
connections = mongoengine.connection._connections
|
||||||
@@ -332,7 +452,8 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
disconnect_all()
|
disconnect_all()
|
||||||
|
|
||||||
def test_sharing_connections(self):
|
def test_sharing_connections(self):
|
||||||
"""Ensure that connections are shared when the connection settings are exactly the same"""
|
"""Ensure that connections are shared when the connection settings are exactly the same
|
||||||
|
"""
|
||||||
connect("mongoenginetests", alias="testdb1")
|
connect("mongoenginetests", alias="testdb1")
|
||||||
expected_connection = get_connection("testdb1")
|
expected_connection = get_connection("testdb1")
|
||||||
|
|
||||||
@@ -443,7 +564,8 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
authd_conn.admin.system.users.delete_many({})
|
authd_conn.admin.system.users.delete_many({})
|
||||||
|
|
||||||
def test_register_connection(self):
|
def test_register_connection(self):
|
||||||
"""Ensure that connections with different aliases may be registered."""
|
"""Ensure that connections with different aliases may be registered.
|
||||||
|
"""
|
||||||
register_connection("testdb", "mongoenginetest2")
|
register_connection("testdb", "mongoenginetest2")
|
||||||
|
|
||||||
with pytest.raises(ConnectionFailure):
|
with pytest.raises(ConnectionFailure):
|
||||||
@@ -456,7 +578,8 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
assert db.name == "mongoenginetest2"
|
assert db.name == "mongoenginetest2"
|
||||||
|
|
||||||
def test_register_connection_defaults(self):
|
def test_register_connection_defaults(self):
|
||||||
"""Ensure that defaults are used when the host and port are None."""
|
"""Ensure that defaults are used when the host and port are None.
|
||||||
|
"""
|
||||||
register_connection("testdb", "mongoenginetest", host=None, port=None)
|
register_connection("testdb", "mongoenginetest", host=None, port=None)
|
||||||
|
|
||||||
conn = get_connection("testdb")
|
conn = get_connection("testdb")
|
||||||
|
@@ -1,167 +0,0 @@
|
|||||||
import unittest
|
|
||||||
|
|
||||||
import pytest
|
|
||||||
|
|
||||||
import mongoengine.connection
|
|
||||||
from mongoengine import Document, StringField, connect, disconnect_all
|
|
||||||
from mongoengine.connection import get_connection
|
|
||||||
|
|
||||||
try:
|
|
||||||
import mongomock
|
|
||||||
|
|
||||||
MONGOMOCK_INSTALLED = True
|
|
||||||
except ImportError:
|
|
||||||
MONGOMOCK_INSTALLED = False
|
|
||||||
|
|
||||||
require_mongomock = pytest.mark.skipif(
|
|
||||||
not MONGOMOCK_INSTALLED, reason="you need mongomock installed to run this testcase"
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class MongoMockConnectionTest(unittest.TestCase):
|
|
||||||
@classmethod
|
|
||||||
def setUpClass(cls):
|
|
||||||
disconnect_all()
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def tearDownClass(cls):
|
|
||||||
disconnect_all()
|
|
||||||
|
|
||||||
def tearDown(self):
|
|
||||||
mongoengine.connection._connection_settings = {}
|
|
||||||
mongoengine.connection._connections = {}
|
|
||||||
mongoengine.connection._dbs = {}
|
|
||||||
|
|
||||||
@require_mongomock
|
|
||||||
def test_connect_in_mocking(self):
|
|
||||||
"""Ensure that the connect() method works properly in mocking."""
|
|
||||||
connect("mongoenginetest", host="mongomock://localhost")
|
|
||||||
conn = get_connection()
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect("mongoenginetest2", host="mongomock://localhost", alias="testdb2")
|
|
||||||
conn = get_connection("testdb2")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(
|
|
||||||
"mongoenginetest3",
|
|
||||||
host="mongodb://localhost",
|
|
||||||
is_mock=True,
|
|
||||||
alias="testdb3",
|
|
||||||
)
|
|
||||||
conn = get_connection("testdb3")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect("mongoenginetest4", is_mock=True, alias="testdb4")
|
|
||||||
conn = get_connection("testdb4")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(
|
|
||||||
host="mongodb://localhost:27017/mongoenginetest5",
|
|
||||||
is_mock=True,
|
|
||||||
alias="testdb5",
|
|
||||||
)
|
|
||||||
conn = get_connection("testdb5")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(host="mongomock://localhost:27017/mongoenginetest6", alias="testdb6")
|
|
||||||
conn = get_connection("testdb6")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(
|
|
||||||
host="mongomock://localhost:27017/mongoenginetest7",
|
|
||||||
is_mock=True,
|
|
||||||
alias="testdb7",
|
|
||||||
)
|
|
||||||
conn = get_connection("testdb7")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
@require_mongomock
|
|
||||||
def test_default_database_with_mocking(self):
|
|
||||||
"""Ensure that the default database is correctly set when using mongomock."""
|
|
||||||
disconnect_all()
|
|
||||||
|
|
||||||
class SomeDocument(Document):
|
|
||||||
pass
|
|
||||||
|
|
||||||
conn = connect(host="mongomock://localhost:27017/mongoenginetest")
|
|
||||||
some_document = SomeDocument()
|
|
||||||
# database won't exist until we save a document
|
|
||||||
some_document.save()
|
|
||||||
assert SomeDocument.objects.count() == 1
|
|
||||||
assert conn.get_default_database().name == "mongoenginetest"
|
|
||||||
assert conn.list_database_names()[0] == "mongoenginetest"
|
|
||||||
|
|
||||||
@require_mongomock
|
|
||||||
def test_basic_queries_against_mongomock(self):
|
|
||||||
disconnect_all()
|
|
||||||
|
|
||||||
connect(host="mongomock://localhost:27017/mongoenginetest")
|
|
||||||
|
|
||||||
class Person(Document):
|
|
||||||
name = StringField()
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
assert Person.objects.count() == 0
|
|
||||||
|
|
||||||
bob = Person(name="Bob").save()
|
|
||||||
john = Person(name="John").save()
|
|
||||||
assert Person.objects.count() == 2
|
|
||||||
|
|
||||||
qs = Person.objects(name="Bob")
|
|
||||||
assert qs.count() == 1
|
|
||||||
assert qs.first() == bob
|
|
||||||
assert list(qs.as_pymongo()) == [{"_id": bob.id, "name": "Bob"}]
|
|
||||||
|
|
||||||
pipeline = [{"$project": {"name": {"$toUpper": "$name"}}}]
|
|
||||||
data = Person.objects.order_by("name").aggregate(pipeline)
|
|
||||||
assert list(data) == [
|
|
||||||
{"_id": bob.id, "name": "BOB"},
|
|
||||||
{"_id": john.id, "name": "JOHN"},
|
|
||||||
]
|
|
||||||
|
|
||||||
Person.drop_collection()
|
|
||||||
assert Person.objects.count() == 0
|
|
||||||
|
|
||||||
@require_mongomock
|
|
||||||
def test_connect_with_host_list(self):
|
|
||||||
"""Ensure that the connect() method works when host is a list
|
|
||||||
|
|
||||||
Uses mongomock to test w/o needing multiple mongod/mongos processes
|
|
||||||
"""
|
|
||||||
connect(host=["mongomock://localhost"])
|
|
||||||
conn = get_connection()
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(host=["mongodb://localhost"], is_mock=True, alias="testdb2")
|
|
||||||
conn = get_connection("testdb2")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(host=["localhost"], is_mock=True, alias="testdb3")
|
|
||||||
conn = get_connection("testdb3")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(
|
|
||||||
host=["mongomock://localhost:27017", "mongomock://localhost:27018"],
|
|
||||||
alias="testdb4",
|
|
||||||
)
|
|
||||||
conn = get_connection("testdb4")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(
|
|
||||||
host=["mongodb://localhost:27017", "mongodb://localhost:27018"],
|
|
||||||
is_mock=True,
|
|
||||||
alias="testdb5",
|
|
||||||
)
|
|
||||||
conn = get_connection("testdb5")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
connect(
|
|
||||||
host=["localhost:27017", "localhost:27018"], is_mock=True, alias="testdb6"
|
|
||||||
)
|
|
||||||
conn = get_connection("testdb6")
|
|
||||||
assert isinstance(conn, mongomock.MongoClient)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
unittest.main()
|
|
@@ -117,7 +117,8 @@ class TestContextManagers:
|
|||||||
assert 1 == Group.objects.count()
|
assert 1 == Group.objects.count()
|
||||||
|
|
||||||
def test_no_dereference_context_manager_object_id(self):
|
def test_no_dereference_context_manager_object_id(self):
|
||||||
"""Ensure that DBRef items in ListFields aren't dereferenced."""
|
"""Ensure that DBRef items in ListFields aren't dereferenced.
|
||||||
|
"""
|
||||||
connect("mongoenginetest")
|
connect("mongoenginetest")
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
@@ -154,7 +155,8 @@ class TestContextManagers:
|
|||||||
assert isinstance(group.generic, User)
|
assert isinstance(group.generic, User)
|
||||||
|
|
||||||
def test_no_dereference_context_manager_dbref(self):
|
def test_no_dereference_context_manager_dbref(self):
|
||||||
"""Ensure that DBRef items in ListFields aren't dereferenced."""
|
"""Ensure that DBRef items in ListFields aren't dereferenced.
|
||||||
|
"""
|
||||||
connect("mongoenginetest")
|
connect("mongoenginetest")
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
@@ -180,11 +182,11 @@ class TestContextManagers:
|
|||||||
|
|
||||||
with no_dereference(Group) as Group:
|
with no_dereference(Group) as Group:
|
||||||
group = Group.objects.first()
|
group = Group.objects.first()
|
||||||
assert all(not isinstance(m, User) for m in group.members)
|
assert all([not isinstance(m, User) for m in group.members])
|
||||||
assert not isinstance(group.ref, User)
|
assert not isinstance(group.ref, User)
|
||||||
assert not isinstance(group.generic, User)
|
assert not isinstance(group.generic, User)
|
||||||
|
|
||||||
assert all(isinstance(m, User) for m in group.members)
|
assert all([isinstance(m, User) for m in group.members])
|
||||||
assert isinstance(group.ref, User)
|
assert isinstance(group.ref, User)
|
||||||
assert isinstance(group.generic, User)
|
assert isinstance(group.generic, User)
|
||||||
|
|
||||||
|
@@ -3,14 +3,10 @@ import unittest
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from mongoengine import Document
|
from mongoengine import Document
|
||||||
from mongoengine.base.datastructures import (
|
from mongoengine.base.datastructures import BaseDict, BaseList, StrictDict
|
||||||
BaseDict,
|
|
||||||
BaseList,
|
|
||||||
StrictDict,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class DocumentStub:
|
class DocumentStub(object):
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self._changed_fields = []
|
self._changed_fields = []
|
||||||
self._unset_fields = []
|
self._unset_fields = []
|
||||||
@@ -425,7 +421,7 @@ class TestStrictDict(unittest.TestCase):
|
|||||||
d.a = 1
|
d.a = 1
|
||||||
assert d.a == 1
|
assert d.a == 1
|
||||||
with pytest.raises(AttributeError):
|
with pytest.raises(AttributeError):
|
||||||
d.b
|
getattr(d, "b")
|
||||||
|
|
||||||
def test_setattr_raises_on_nonexisting_attr(self):
|
def test_setattr_raises_on_nonexisting_attr(self):
|
||||||
d = self.dtype()
|
d = self.dtype()
|
||||||
|
@@ -16,7 +16,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
cls.db.drop_database("mongoenginetest")
|
cls.db.drop_database("mongoenginetest")
|
||||||
|
|
||||||
def test_list_item_dereference(self):
|
def test_list_item_dereference(self):
|
||||||
"""Ensure that DBRef items in ListFields are dereferenced."""
|
"""Ensure that DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -49,7 +50,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
len(group_obj.members)
|
len(group_obj.members)
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@@ -58,7 +59,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
|
|
||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 2
|
assert q == 2
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@@ -67,14 +68,15 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_objs = Group.objects.select_related()
|
group_objs = Group.objects.select_related()
|
||||||
assert q == 2
|
assert q == 2
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
User.drop_collection()
|
User.drop_collection()
|
||||||
Group.drop_collection()
|
Group.drop_collection()
|
||||||
|
|
||||||
def test_list_item_dereference_dref_false(self):
|
def test_list_item_dereference_dref_false(self):
|
||||||
"""Ensure that DBRef items in ListFields are dereferenced."""
|
"""Ensure that DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -99,14 +101,14 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
assert group_obj._data["members"]._dereferenced
|
assert group_obj._data["members"]._dereferenced
|
||||||
|
|
||||||
# verifies that no additional queries gets executed
|
# verifies that no additional queries gets executed
|
||||||
# if we re-iterate over the ListField once it is
|
# if we re-iterate over the ListField once it is
|
||||||
# dereferenced
|
# dereferenced
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
assert group_obj._data["members"]._dereferenced
|
assert group_obj._data["members"]._dereferenced
|
||||||
|
|
||||||
@@ -117,7 +119,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
|
|
||||||
assert q == 2
|
assert q == 2
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@@ -126,11 +128,12 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_objs = Group.objects.select_related()
|
group_objs = Group.objects.select_related()
|
||||||
assert q == 2
|
assert q == 2
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
def test_list_item_dereference_orphan_dbref(self):
|
def test_list_item_dereference_orphan_dbref(self):
|
||||||
"""Ensure that orphan DBRef items in ListFields are dereferenced."""
|
"""Ensure that orphan DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -158,14 +161,14 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
assert group_obj._data["members"]._dereferenced
|
assert group_obj._data["members"]._dereferenced
|
||||||
|
|
||||||
# verifies that no additional queries gets executed
|
# verifies that no additional queries gets executed
|
||||||
# if we re-iterate over the ListField once it is
|
# if we re-iterate over the ListField once it is
|
||||||
# dereferenced
|
# dereferenced
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
assert group_obj._data["members"]._dereferenced
|
assert group_obj._data["members"]._dereferenced
|
||||||
|
|
||||||
@@ -173,7 +176,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
Group.drop_collection()
|
Group.drop_collection()
|
||||||
|
|
||||||
def test_list_item_dereference_dref_false_stores_as_type(self):
|
def test_list_item_dereference_dref_false_stores_as_type(self):
|
||||||
"""Ensure that DBRef items are stored as their type"""
|
"""Ensure that DBRef items are stored as their type
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
my_id = IntField(primary_key=True)
|
my_id = IntField(primary_key=True)
|
||||||
@@ -194,7 +198,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert group.members == [user]
|
assert group.members == [user]
|
||||||
|
|
||||||
def test_handle_old_style_references(self):
|
def test_handle_old_style_references(self):
|
||||||
"""Ensure that DBRef items in ListFields are dereferenced."""
|
"""Ensure that DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -227,7 +232,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert group.members[-1].name == "String!"
|
assert group.members[-1].name == "String!"
|
||||||
|
|
||||||
def test_migrate_references(self):
|
def test_migrate_references(self):
|
||||||
"""Example of migrating ReferenceField storage"""
|
"""Example of migrating ReferenceField storage
|
||||||
|
"""
|
||||||
|
|
||||||
# Create some sample data
|
# Create some sample data
|
||||||
class User(Document):
|
class User(Document):
|
||||||
@@ -272,7 +278,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert isinstance(raw_data["members"][0], ObjectId)
|
assert isinstance(raw_data["members"][0], ObjectId)
|
||||||
|
|
||||||
def test_recursive_reference(self):
|
def test_recursive_reference(self):
|
||||||
"""Ensure that ReferenceFields can reference their own documents."""
|
"""Ensure that ReferenceFields can reference their own documents.
|
||||||
|
"""
|
||||||
|
|
||||||
class Employee(Document):
|
class Employee(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -395,7 +402,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert "[<Person: Mother>, <Person: Daughter>]" == "%s" % Person.objects()
|
assert "[<Person: Mother>, <Person: Daughter>]" == "%s" % Person.objects()
|
||||||
|
|
||||||
def test_circular_reference_on_self(self):
|
def test_circular_reference_on_self(self):
|
||||||
"""Ensure you can handle circular references"""
|
"""Ensure you can handle circular references
|
||||||
|
"""
|
||||||
|
|
||||||
class Person(Document):
|
class Person(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -422,7 +430,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert "[<Person: Mother>, <Person: Daughter>]" == "%s" % Person.objects()
|
assert "[<Person: Mother>, <Person: Daughter>]" == "%s" % Person.objects()
|
||||||
|
|
||||||
def test_circular_tree_reference(self):
|
def test_circular_tree_reference(self):
|
||||||
"""Ensure you can handle circular references with more than one level"""
|
"""Ensure you can handle circular references with more than one level
|
||||||
|
"""
|
||||||
|
|
||||||
class Other(EmbeddedDocument):
|
class Other(EmbeddedDocument):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -505,10 +514,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for m in group_obj.members:
|
for m in group_obj.members:
|
||||||
@@ -521,10 +530,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for m in group_obj.members:
|
for m in group_obj.members:
|
||||||
@@ -538,17 +547,18 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for m in group_obj.members:
|
for m in group_obj.members:
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
def test_generic_reference_orphan_dbref(self):
|
def test_generic_reference_orphan_dbref(self):
|
||||||
"""Ensure that generic orphan DBRef items in ListFields are dereferenced."""
|
"""Ensure that generic orphan DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class UserA(Document):
|
class UserA(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -592,11 +602,11 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
assert group_obj._data["members"]._dereferenced
|
assert group_obj._data["members"]._dereferenced
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
assert group_obj._data["members"]._dereferenced
|
assert group_obj._data["members"]._dereferenced
|
||||||
|
|
||||||
@@ -648,10 +658,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for m in group_obj.members:
|
for m in group_obj.members:
|
||||||
@@ -664,10 +674,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for m in group_obj.members:
|
for m in group_obj.members:
|
||||||
@@ -681,10 +691,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for m in group_obj.members:
|
for m in group_obj.members:
|
||||||
@@ -723,10 +733,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for _, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, User)
|
assert isinstance(m, User)
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@@ -736,7 +746,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -750,7 +760,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -801,10 +811,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -817,10 +827,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -834,10 +844,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -852,7 +862,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 1
|
assert q == 1
|
||||||
assert group_obj.members == {}
|
assert group_obj.members == {}
|
||||||
|
|
||||||
@@ -891,10 +901,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -907,10 +917,10 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for k, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
@@ -924,13 +934,13 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
for _, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
assert isinstance(m, UserA)
|
assert isinstance(m, UserA)
|
||||||
|
|
||||||
UserA.drop_collection()
|
UserA.drop_collection()
|
||||||
@@ -978,13 +988,13 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for _, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
# Document select_related
|
# Document select_related
|
||||||
@@ -994,13 +1004,13 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first().select_related()
|
group_obj = Group.objects.first().select_related()
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for _, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
# Queryset select_related
|
# Queryset select_related
|
||||||
@@ -1011,13 +1021,13 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for group_obj in group_objs:
|
for group_obj in group_objs:
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 4
|
assert q == 4
|
||||||
|
|
||||||
for _, m in group_obj.members.items():
|
for k, m in group_obj.members.items():
|
||||||
assert "User" in m.__class__.__name__
|
assert "User" in m.__class__.__name__
|
||||||
|
|
||||||
Group.objects.delete()
|
Group.objects.delete()
|
||||||
@@ -1029,7 +1039,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
group_obj = Group.objects.first()
|
group_obj = Group.objects.first()
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
_ = [m for m in group_obj.members]
|
[m for m in group_obj.members]
|
||||||
assert q == 1
|
assert q == 1
|
||||||
|
|
||||||
UserA.drop_collection()
|
UserA.drop_collection()
|
||||||
@@ -1158,7 +1168,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert msg.author.name == "new-name"
|
assert msg.author.name == "new-name"
|
||||||
|
|
||||||
def test_list_lookup_not_checked_in_map(self):
|
def test_list_lookup_not_checked_in_map(self):
|
||||||
"""Ensure we dereference list data correctly"""
|
"""Ensure we dereference list data correctly
|
||||||
|
"""
|
||||||
|
|
||||||
class Comment(Document):
|
class Comment(Document):
|
||||||
id = IntField(primary_key=True)
|
id = IntField(primary_key=True)
|
||||||
@@ -1180,7 +1191,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert 1 == msg.comments[1].id
|
assert 1 == msg.comments[1].id
|
||||||
|
|
||||||
def test_list_item_dereference_dref_false_save_doesnt_cause_extra_queries(self):
|
def test_list_item_dereference_dref_false_save_doesnt_cause_extra_queries(self):
|
||||||
"""Ensure that DBRef items in ListFields are dereferenced."""
|
"""Ensure that DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1209,7 +1221,8 @@ class FieldTest(unittest.TestCase):
|
|||||||
assert q == 2
|
assert q == 2
|
||||||
|
|
||||||
def test_list_item_dereference_dref_true_save_doesnt_cause_extra_queries(self):
|
def test_list_item_dereference_dref_true_save_doesnt_cause_extra_queries(self):
|
||||||
"""Ensure that DBRef items in ListFields are dereferenced."""
|
"""Ensure that DBRef items in ListFields are dereferenced.
|
||||||
|
"""
|
||||||
|
|
||||||
class User(Document):
|
class User(Document):
|
||||||
name = StringField()
|
name = StringField()
|
||||||
@@ -1321,7 +1334,7 @@ class FieldTest(unittest.TestCase):
|
|||||||
BrandGroup.drop_collection()
|
BrandGroup.drop_collection()
|
||||||
|
|
||||||
brand1 = Brand(title="Moschino").save()
|
brand1 = Brand(title="Moschino").save()
|
||||||
brand2 = Brand(title="Денис Симачёв").save()
|
brand2 = Brand(title=u"Денис Симачёв").save()
|
||||||
|
|
||||||
BrandGroup(title="top_brands", brands=[brand1, brand2]).save()
|
BrandGroup(title="top_brands", brands=[brand1, brand2]).save()
|
||||||
brand_groups = BrandGroup.objects().all()
|
brand_groups = BrandGroup.objects().all()
|
||||||
|
@@ -5,6 +5,7 @@ from pymongo import MongoClient, ReadPreference
|
|||||||
import mongoengine
|
import mongoengine
|
||||||
from mongoengine.connection import ConnectionFailure
|
from mongoengine.connection import ConnectionFailure
|
||||||
|
|
||||||
|
|
||||||
CONN_CLASS = MongoClient
|
CONN_CLASS = MongoClient
|
||||||
READ_PREF = ReadPreference.SECONDARY
|
READ_PREF = ReadPreference.SECONDARY
|
||||||
|
|
||||||
@@ -21,7 +22,8 @@ class ConnectionTest(unittest.TestCase):
|
|||||||
mongoengine.connection._dbs = {}
|
mongoengine.connection._dbs = {}
|
||||||
|
|
||||||
def test_replicaset_uri_passes_read_preference(self):
|
def test_replicaset_uri_passes_read_preference(self):
|
||||||
"""Requires a replica set called "rs" on port 27017"""
|
"""Requires a replica set called "rs" on port 27017
|
||||||
|
"""
|
||||||
try:
|
try:
|
||||||
conn = mongoengine.connect(
|
conn = mongoengine.connect(
|
||||||
db="mongoenginetest",
|
db="mongoenginetest",
|
||||||
|
@@ -249,7 +249,7 @@ class TestSignal(unittest.TestCase):
|
|||||||
assert self.pre_signals == post_signals
|
assert self.pre_signals == post_signals
|
||||||
|
|
||||||
def test_model_signals(self):
|
def test_model_signals(self):
|
||||||
"""Model saves should throw some signals."""
|
""" Model saves should throw some signals. """
|
||||||
|
|
||||||
def create_author():
|
def create_author():
|
||||||
self.Author(name="Bill Shakespeare")
|
self.Author(name="Bill Shakespeare")
|
||||||
@@ -340,7 +340,7 @@ class TestSignal(unittest.TestCase):
|
|||||||
]
|
]
|
||||||
|
|
||||||
def test_signal_kwargs(self):
|
def test_signal_kwargs(self):
|
||||||
"""Make sure signal_kwargs is passed to signals calls."""
|
""" Make sure signal_kwargs is passed to signals calls. """
|
||||||
|
|
||||||
def live_and_let_die():
|
def live_and_let_die():
|
||||||
a = self.Author(name="Bill Shakespeare")
|
a = self.Author(name="Bill Shakespeare")
|
||||||
@@ -385,7 +385,7 @@ class TestSignal(unittest.TestCase):
|
|||||||
]
|
]
|
||||||
|
|
||||||
def test_queryset_delete_signals(self):
|
def test_queryset_delete_signals(self):
|
||||||
"""Queryset delete should throw some signals."""
|
""" Queryset delete should throw some signals. """
|
||||||
|
|
||||||
self.Another(name="Bill Shakespeare").save()
|
self.Another(name="Bill Shakespeare").save()
|
||||||
assert self.get_signal_output(self.Another.objects.delete) == [
|
assert self.get_signal_output(self.Another.objects.delete) == [
|
||||||
@@ -396,7 +396,7 @@ class TestSignal(unittest.TestCase):
|
|||||||
]
|
]
|
||||||
|
|
||||||
def test_signals_with_explicit_doc_ids(self):
|
def test_signals_with_explicit_doc_ids(self):
|
||||||
"""Model saves must have a created flag the first time."""
|
""" Model saves must have a created flag the first time."""
|
||||||
ei = self.ExplicitId(id=123)
|
ei = self.ExplicitId(id=123)
|
||||||
# post save must received the created flag, even if there's already
|
# post save must received the created flag, even if there's already
|
||||||
# an object id present
|
# an object id present
|
||||||
|
@@ -1,4 +1,5 @@
|
|||||||
import re
|
import re
|
||||||
|
import unittest
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
@@ -9,7 +10,7 @@ signal_output = []
|
|||||||
|
|
||||||
class TestLazyRegexCompiler:
|
class TestLazyRegexCompiler:
|
||||||
def test_lazy_regex_compiler_verify_laziness_of_descriptor(self):
|
def test_lazy_regex_compiler_verify_laziness_of_descriptor(self):
|
||||||
class UserEmail:
|
class UserEmail(object):
|
||||||
EMAIL_REGEX = LazyRegexCompiler("@", flags=32)
|
EMAIL_REGEX = LazyRegexCompiler("@", flags=32)
|
||||||
|
|
||||||
descriptor = UserEmail.__dict__["EMAIL_REGEX"]
|
descriptor = UserEmail.__dict__["EMAIL_REGEX"]
|
||||||
@@ -23,7 +24,7 @@ class TestLazyRegexCompiler:
|
|||||||
assert user_email.EMAIL_REGEX is UserEmail.EMAIL_REGEX
|
assert user_email.EMAIL_REGEX is UserEmail.EMAIL_REGEX
|
||||||
|
|
||||||
def test_lazy_regex_compiler_verify_cannot_set_descriptor_on_instance(self):
|
def test_lazy_regex_compiler_verify_cannot_set_descriptor_on_instance(self):
|
||||||
class UserEmail:
|
class UserEmail(object):
|
||||||
EMAIL_REGEX = LazyRegexCompiler("@")
|
EMAIL_REGEX = LazyRegexCompiler("@")
|
||||||
|
|
||||||
user_email = UserEmail()
|
user_email = UserEmail()
|
||||||
@@ -31,7 +32,7 @@ class TestLazyRegexCompiler:
|
|||||||
user_email.EMAIL_REGEX = re.compile("@")
|
user_email.EMAIL_REGEX = re.compile("@")
|
||||||
|
|
||||||
def test_lazy_regex_compiler_verify_can_override_class_attr(self):
|
def test_lazy_regex_compiler_verify_can_override_class_attr(self):
|
||||||
class UserEmail:
|
class UserEmail(object):
|
||||||
EMAIL_REGEX = LazyRegexCompiler("@")
|
EMAIL_REGEX = LazyRegexCompiler("@")
|
||||||
|
|
||||||
UserEmail.EMAIL_REGEX = re.compile("cookies")
|
UserEmail.EMAIL_REGEX = re.compile("cookies")
|
||||||
|
@@ -1,4 +1,3 @@
|
|||||||
import operator
|
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
@@ -7,6 +6,7 @@ from mongoengine import connect
|
|||||||
from mongoengine.connection import disconnect_all, get_db
|
from mongoengine.connection import disconnect_all, get_db
|
||||||
from mongoengine.mongodb_support import get_mongodb_version
|
from mongoengine.mongodb_support import get_mongodb_version
|
||||||
|
|
||||||
|
|
||||||
MONGO_TEST_DB = "mongoenginetest" # standard name for the test database
|
MONGO_TEST_DB = "mongoenginetest" # standard name for the test database
|
||||||
|
|
||||||
|
|
||||||
@@ -33,14 +33,6 @@ def get_as_pymongo(doc):
|
|||||||
return doc.__class__.objects.as_pymongo().get(id=doc.id)
|
return doc.__class__.objects.as_pymongo().get(id=doc.id)
|
||||||
|
|
||||||
|
|
||||||
def requires_mongodb_lt_42(func):
|
|
||||||
return _decorated_with_ver_requirement(func, (4, 2), oper=operator.lt)
|
|
||||||
|
|
||||||
|
|
||||||
def requires_mongodb_gte_44(func):
|
|
||||||
return _decorated_with_ver_requirement(func, (4, 4), oper=operator.ge)
|
|
||||||
|
|
||||||
|
|
||||||
def _decorated_with_ver_requirement(func, mongo_version_req, oper):
|
def _decorated_with_ver_requirement(func, mongo_version_req, oper):
|
||||||
"""Return a MongoDB version requirement decorator.
|
"""Return a MongoDB version requirement decorator.
|
||||||
|
|
||||||
@@ -67,7 +59,7 @@ def _decorated_with_ver_requirement(func, mongo_version_req, oper):
|
|||||||
return func(*args, **kwargs)
|
return func(*args, **kwargs)
|
||||||
|
|
||||||
pretty_version = ".".join(str(n) for n in mongo_version_req)
|
pretty_version = ".".join(str(n) for n in mongo_version_req)
|
||||||
pytest.skip(f"Needs MongoDB v{pretty_version}+")
|
pytest.skip("Needs MongoDB v{}+".format(pretty_version))
|
||||||
|
|
||||||
_inner.__name__ = func.__name__
|
_inner.__name__ = func.__name__
|
||||||
_inner.__doc__ = func.__doc__
|
_inner.__doc__ = func.__doc__
|
||||||
|
Reference in New Issue
Block a user