mirror of
https://github.com/paperless-ngx/paperless-ngx.git
synced 2025-04-02 13:45:10 -05:00
Merge branch 'dev' into feature-permissions
This commit is contained in:
commit
f021e7fcc3
24
.github/workflows/ci.yml
vendored
24
.github/workflows/ci.yml
vendored
@ -16,7 +16,7 @@ on:
|
|||||||
jobs:
|
jobs:
|
||||||
pre-commit:
|
pre-commit:
|
||||||
name: Linting Checks
|
name: Linting Checks
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
steps:
|
steps:
|
||||||
-
|
-
|
||||||
name: Checkout repository
|
name: Checkout repository
|
||||||
@ -34,7 +34,7 @@ jobs:
|
|||||||
|
|
||||||
documentation:
|
documentation:
|
||||||
name: "Build Documentation"
|
name: "Build Documentation"
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
needs:
|
needs:
|
||||||
- pre-commit
|
- pre-commit
|
||||||
steps:
|
steps:
|
||||||
@ -44,7 +44,7 @@ jobs:
|
|||||||
-
|
-
|
||||||
name: Install pipenv
|
name: Install pipenv
|
||||||
run: |
|
run: |
|
||||||
pipx install pipenv==2022.10.12
|
pipx install pipenv==2022.11.30
|
||||||
-
|
-
|
||||||
name: Set up Python
|
name: Set up Python
|
||||||
uses: actions/setup-python@v4
|
uses: actions/setup-python@v4
|
||||||
@ -73,7 +73,7 @@ jobs:
|
|||||||
|
|
||||||
documentation-deploy:
|
documentation-deploy:
|
||||||
name: "Deploy Documentation"
|
name: "Deploy Documentation"
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
if: github.event_name == 'push' && github.ref == 'refs/heads/main'
|
if: github.event_name == 'push' && github.ref == 'refs/heads/main'
|
||||||
needs:
|
needs:
|
||||||
- documentation
|
- documentation
|
||||||
@ -92,7 +92,7 @@ jobs:
|
|||||||
|
|
||||||
tests-backend:
|
tests-backend:
|
||||||
name: "Tests (${{ matrix.python-version }})"
|
name: "Tests (${{ matrix.python-version }})"
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
needs:
|
needs:
|
||||||
- pre-commit
|
- pre-commit
|
||||||
strategy:
|
strategy:
|
||||||
@ -124,7 +124,7 @@ jobs:
|
|||||||
-
|
-
|
||||||
name: Install pipenv
|
name: Install pipenv
|
||||||
run: |
|
run: |
|
||||||
pipx install pipenv==2022.10.12
|
pipx install pipenv==2022.11.30
|
||||||
-
|
-
|
||||||
name: Set up Python
|
name: Set up Python
|
||||||
uses: actions/setup-python@v4
|
uses: actions/setup-python@v4
|
||||||
@ -181,7 +181,7 @@ jobs:
|
|||||||
|
|
||||||
tests-frontend:
|
tests-frontend:
|
||||||
name: "Tests Frontend"
|
name: "Tests Frontend"
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
needs:
|
needs:
|
||||||
- pre-commit
|
- pre-commit
|
||||||
strategy:
|
strategy:
|
||||||
@ -201,7 +201,7 @@ jobs:
|
|||||||
prepare-docker-build:
|
prepare-docker-build:
|
||||||
name: Prepare Docker Pipeline Data
|
name: Prepare Docker Pipeline Data
|
||||||
if: github.event_name == 'push' && (startsWith(github.ref, 'refs/heads/feature-') || github.ref == 'refs/heads/dev' || github.ref == 'refs/heads/beta' || contains(github.ref, 'beta.rc') || startsWith(github.ref, 'refs/tags/v'))
|
if: github.event_name == 'push' && (startsWith(github.ref, 'refs/heads/feature-') || github.ref == 'refs/heads/dev' || github.ref == 'refs/heads/beta' || contains(github.ref, 'beta.rc') || startsWith(github.ref, 'refs/tags/v'))
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
# If the push triggered the installer library workflow, wait for it to
|
# If the push triggered the installer library workflow, wait for it to
|
||||||
# complete here. This ensures the required versions for the final
|
# complete here. This ensures the required versions for the final
|
||||||
# image have been built, while not waiting at all if the versions haven't changed
|
# image have been built, while not waiting at all if the versions haven't changed
|
||||||
@ -278,7 +278,7 @@ jobs:
|
|||||||
|
|
||||||
# build and push image to docker hub.
|
# build and push image to docker hub.
|
||||||
build-docker-image:
|
build-docker-image:
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
concurrency:
|
concurrency:
|
||||||
group: ${{ github.workflow }}-build-docker-image-${{ github.ref_name }}
|
group: ${{ github.workflow }}-build-docker-image-${{ github.ref_name }}
|
||||||
cancel-in-progress: true
|
cancel-in-progress: true
|
||||||
@ -383,7 +383,7 @@ jobs:
|
|||||||
build-release:
|
build-release:
|
||||||
needs:
|
needs:
|
||||||
- build-docker-image
|
- build-docker-image
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
steps:
|
steps:
|
||||||
-
|
-
|
||||||
name: Checkout
|
name: Checkout
|
||||||
@ -462,7 +462,7 @@ jobs:
|
|||||||
path: dist/paperless-ngx.tar.xz
|
path: dist/paperless-ngx.tar.xz
|
||||||
|
|
||||||
publish-release:
|
publish-release:
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
outputs:
|
outputs:
|
||||||
prerelease: ${{ steps.get_version.outputs.prerelease }}
|
prerelease: ${{ steps.get_version.outputs.prerelease }}
|
||||||
changelog: ${{ steps.create-release.outputs.body }}
|
changelog: ${{ steps.create-release.outputs.body }}
|
||||||
@ -511,7 +511,7 @@ jobs:
|
|||||||
asset_content_type: application/x-xz
|
asset_content_type: application/x-xz
|
||||||
|
|
||||||
append-changelog:
|
append-changelog:
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
needs:
|
needs:
|
||||||
- publish-release
|
- publish-release
|
||||||
if: needs.publish-release.outputs.prerelease == 'false'
|
if: needs.publish-release.outputs.prerelease == 'false'
|
||||||
|
7
.github/workflows/cleanup-tags.yml
vendored
7
.github/workflows/cleanup-tags.yml
vendored
@ -6,12 +6,7 @@
|
|||||||
name: Cleanup Image Tags
|
name: Cleanup Image Tags
|
||||||
|
|
||||||
on:
|
on:
|
||||||
schedule:
|
|
||||||
- cron: '0 0 * * SAT'
|
|
||||||
delete:
|
delete:
|
||||||
pull_request:
|
|
||||||
types:
|
|
||||||
- closed
|
|
||||||
push:
|
push:
|
||||||
paths:
|
paths:
|
||||||
- ".github/workflows/cleanup-tags.yml"
|
- ".github/workflows/cleanup-tags.yml"
|
||||||
@ -26,7 +21,7 @@ concurrency:
|
|||||||
jobs:
|
jobs:
|
||||||
cleanup-images:
|
cleanup-images:
|
||||||
name: Cleanup Image Tags for ${{ matrix.primary-name }}
|
name: Cleanup Image Tags for ${{ matrix.primary-name }}
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
include:
|
include:
|
||||||
|
2
.github/workflows/codeql-analysis.yml
vendored
2
.github/workflows/codeql-analysis.yml
vendored
@ -23,7 +23,7 @@ on:
|
|||||||
jobs:
|
jobs:
|
||||||
analyze:
|
analyze:
|
||||||
name: Analyze
|
name: Analyze
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
permissions:
|
permissions:
|
||||||
actions: read
|
actions: read
|
||||||
contents: read
|
contents: read
|
||||||
|
2
.github/workflows/installer-library.yml
vendored
2
.github/workflows/installer-library.yml
vendored
@ -34,7 +34,7 @@ concurrency:
|
|||||||
jobs:
|
jobs:
|
||||||
prepare-docker-build:
|
prepare-docker-build:
|
||||||
name: Prepare Docker Image Version Data
|
name: Prepare Docker Image Version Data
|
||||||
runs-on: ubuntu-20.04
|
runs-on: ubuntu-22.04
|
||||||
steps:
|
steps:
|
||||||
-
|
-
|
||||||
name: Set ghcr repository name
|
name: Set ghcr repository name
|
||||||
|
4
.github/workflows/project-actions.yml
vendored
4
.github/workflows/project-actions.yml
vendored
@ -24,7 +24,7 @@ env:
|
|||||||
jobs:
|
jobs:
|
||||||
issue_opened_or_reopened:
|
issue_opened_or_reopened:
|
||||||
name: issue_opened_or_reopened
|
name: issue_opened_or_reopened
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
if: github.event_name == 'issues' && (github.event.action == 'opened' || github.event.action == 'reopened')
|
if: github.event_name == 'issues' && (github.event.action == 'opened' || github.event.action == 'reopened')
|
||||||
steps:
|
steps:
|
||||||
- name: Add issue to project and set status to ${{ env.todo }}
|
- name: Add issue to project and set status to ${{ env.todo }}
|
||||||
@ -37,7 +37,7 @@ jobs:
|
|||||||
status_value: ${{ env.todo }} # Target status
|
status_value: ${{ env.todo }} # Target status
|
||||||
pr_opened_or_reopened:
|
pr_opened_or_reopened:
|
||||||
name: pr_opened_or_reopened
|
name: pr_opened_or_reopened
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
permissions:
|
permissions:
|
||||||
# write permission is required for autolabeler
|
# write permission is required for autolabeler
|
||||||
pull-requests: write
|
pull-requests: write
|
||||||
|
2
.github/workflows/release-chart.yml
vendored
2
.github/workflows/release-chart.yml
vendored
@ -9,7 +9,7 @@ on:
|
|||||||
jobs:
|
jobs:
|
||||||
release_chart:
|
release_chart:
|
||||||
name: "Release Chart"
|
name: "Release Chart"
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout
|
- name: Checkout
|
||||||
uses: actions/checkout@v3
|
uses: actions/checkout@v3
|
||||||
|
@ -21,7 +21,7 @@ concurrency:
|
|||||||
jobs:
|
jobs:
|
||||||
build-image:
|
build-image:
|
||||||
name: Build ${{ fromJSON(inputs.build-json).name }} @ ${{ fromJSON(inputs.build-json).version }}
|
name: Build ${{ fromJSON(inputs.build-json).name }} @ ${{ fromJSON(inputs.build-json).version }}
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-22.04
|
||||||
steps:
|
steps:
|
||||||
-
|
-
|
||||||
name: Checkout
|
name: Checkout
|
||||||
|
@ -45,7 +45,7 @@ COPY Pipfile* ./
|
|||||||
|
|
||||||
RUN set -eux \
|
RUN set -eux \
|
||||||
&& echo "Installing pipenv" \
|
&& echo "Installing pipenv" \
|
||||||
&& python3 -m pip install --no-cache-dir --upgrade pipenv==2022.10.12 \
|
&& python3 -m pip install --no-cache-dir --upgrade pipenv==2022.11.30 \
|
||||||
&& echo "Generating requirement.txt" \
|
&& echo "Generating requirement.txt" \
|
||||||
&& pipenv requirements > requirements.txt
|
&& pipenv requirements > requirements.txt
|
||||||
|
|
||||||
|
@ -0,0 +1,21 @@
|
|||||||
|
# Generated by Django 4.1.3 on 2022-11-22 17:50
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
("documents", "1027_remove_paperlesstask_attempted_task_and_more"),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.RemoveField(
|
||||||
|
model_name="paperlesstask",
|
||||||
|
name="task_args",
|
||||||
|
),
|
||||||
|
migrations.RemoveField(
|
||||||
|
model_name="paperlesstask",
|
||||||
|
name="task_kwargs",
|
||||||
|
),
|
||||||
|
]
|
@ -572,20 +572,6 @@ class PaperlessTask(models.Model):
|
|||||||
help_text=_("Name of the Task which was run"),
|
help_text=_("Name of the Task which was run"),
|
||||||
)
|
)
|
||||||
|
|
||||||
task_args = models.JSONField(
|
|
||||||
null=True,
|
|
||||||
verbose_name=_("Task Positional Arguments"),
|
|
||||||
help_text=_(
|
|
||||||
"JSON representation of the positional arguments used with the task",
|
|
||||||
),
|
|
||||||
)
|
|
||||||
task_kwargs = models.JSONField(
|
|
||||||
null=True,
|
|
||||||
verbose_name=_("Task Named Arguments"),
|
|
||||||
help_text=_(
|
|
||||||
"JSON representation of the named arguments used with the task",
|
|
||||||
),
|
|
||||||
)
|
|
||||||
status = models.CharField(
|
status = models.CharField(
|
||||||
max_length=30,
|
max_length=30,
|
||||||
default=states.PENDING,
|
default=states.PENDING,
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
import logging
|
import logging
|
||||||
import os
|
import os
|
||||||
import shutil
|
import shutil
|
||||||
from ast import literal_eval
|
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from celery import states
|
from celery import states
|
||||||
@ -521,25 +520,24 @@ def add_to_index(sender, document, **kwargs):
|
|||||||
def before_task_publish_handler(sender=None, headers=None, body=None, **kwargs):
|
def before_task_publish_handler(sender=None, headers=None, body=None, **kwargs):
|
||||||
"""
|
"""
|
||||||
Creates the PaperlessTask object in a pending state. This is sent before
|
Creates the PaperlessTask object in a pending state. This is sent before
|
||||||
the task reaches the broker, but
|
the task reaches the broker, but before it begins executing on a worker.
|
||||||
|
|
||||||
https://docs.celeryq.dev/en/stable/userguide/signals.html#before-task-publish
|
https://docs.celeryq.dev/en/stable/userguide/signals.html#before-task-publish
|
||||||
|
|
||||||
|
https://docs.celeryq.dev/en/stable/internals/protocol.html#version-2
|
||||||
|
|
||||||
"""
|
"""
|
||||||
if "task" not in headers or headers["task"] != "documents.tasks.consume_file":
|
if "task" not in headers or headers["task"] != "documents.tasks.consume_file":
|
||||||
# Assumption: this is only ever a v2 message
|
# Assumption: this is only ever a v2 message
|
||||||
return
|
return
|
||||||
|
|
||||||
try:
|
try:
|
||||||
task_file_name = ""
|
task_args = body[0]
|
||||||
if headers["kwargsrepr"] is not None:
|
task_kwargs = body[1]
|
||||||
task_kwargs = literal_eval(headers["kwargsrepr"])
|
|
||||||
if "override_filename" in task_kwargs:
|
|
||||||
task_file_name = task_kwargs["override_filename"]
|
|
||||||
else:
|
|
||||||
task_kwargs = None
|
|
||||||
|
|
||||||
task_args = literal_eval(headers["argsrepr"])
|
task_file_name = ""
|
||||||
|
if "override_filename" in task_kwargs:
|
||||||
|
task_file_name = task_kwargs["override_filename"]
|
||||||
|
|
||||||
# Nothing was found, report the task first argument
|
# Nothing was found, report the task first argument
|
||||||
if not len(task_file_name):
|
if not len(task_file_name):
|
||||||
@ -552,8 +550,6 @@ def before_task_publish_handler(sender=None, headers=None, body=None, **kwargs):
|
|||||||
status=states.PENDING,
|
status=states.PENDING,
|
||||||
task_file_name=task_file_name,
|
task_file_name=task_file_name,
|
||||||
task_name=headers["task"],
|
task_name=headers["task"],
|
||||||
task_args=task_args,
|
|
||||||
task_kwargs=task_kwargs,
|
|
||||||
result=None,
|
result=None,
|
||||||
date_created=timezone.now(),
|
date_created=timezone.now(),
|
||||||
date_started=None,
|
date_started=None,
|
||||||
@ -562,7 +558,7 @@ def before_task_publish_handler(sender=None, headers=None, body=None, **kwargs):
|
|||||||
except Exception as e: # pragma: no cover
|
except Exception as e: # pragma: no cover
|
||||||
# Don't let an exception in the signal handlers prevent
|
# Don't let an exception in the signal handlers prevent
|
||||||
# a document from being consumed.
|
# a document from being consumed.
|
||||||
logger.error(f"Creating PaperlessTask failed: {e}")
|
logger.error(f"Creating PaperlessTask failed: {e}", exc_info=True)
|
||||||
|
|
||||||
|
|
||||||
@task_prerun.connect
|
@task_prerun.connect
|
||||||
@ -584,7 +580,7 @@ def task_prerun_handler(sender=None, task_id=None, task=None, **kwargs):
|
|||||||
except Exception as e: # pragma: no cover
|
except Exception as e: # pragma: no cover
|
||||||
# Don't let an exception in the signal handlers prevent
|
# Don't let an exception in the signal handlers prevent
|
||||||
# a document from being consumed.
|
# a document from being consumed.
|
||||||
logger.error(f"Setting PaperlessTask started failed: {e}")
|
logger.error(f"Setting PaperlessTask started failed: {e}", exc_info=True)
|
||||||
|
|
||||||
|
|
||||||
@task_postrun.connect
|
@task_postrun.connect
|
||||||
@ -607,4 +603,4 @@ def task_postrun_handler(
|
|||||||
except Exception as e: # pragma: no cover
|
except Exception as e: # pragma: no cover
|
||||||
# Don't let an exception in the signal handlers prevent
|
# Don't let an exception in the signal handlers prevent
|
||||||
# a document from being consumed.
|
# a document from being consumed.
|
||||||
logger.error(f"Updating PaperlessTask failed: {e}")
|
logger.error(f"Updating PaperlessTask failed: {e}", exc_info=True)
|
||||||
|
@ -103,6 +103,8 @@ def consume_file(
|
|||||||
|
|
||||||
# Celery converts this to a string, but everything expects a datetime
|
# Celery converts this to a string, but everything expects a datetime
|
||||||
# Long term solution is to not use JSON for the serializer but pickle instead
|
# Long term solution is to not use JSON for the serializer but pickle instead
|
||||||
|
# TODO: This will be resolved in kombu 5.3, expected with celery 5.3
|
||||||
|
# More types will be retained through JSON encode/decode
|
||||||
if override_created is not None and isinstance(override_created, str):
|
if override_created is not None and isinstance(override_created, str):
|
||||||
try:
|
try:
|
||||||
override_created = datetime.fromisoformat(override_created)
|
override_created = datetime.fromisoformat(override_created)
|
||||||
|
@ -3125,16 +3125,6 @@ class TestTasks(APITestCase):
|
|||||||
task_file_name="test.pdf",
|
task_file_name="test.pdf",
|
||||||
task_name="documents.tasks.some_task",
|
task_name="documents.tasks.some_task",
|
||||||
status=celery.states.SUCCESS,
|
status=celery.states.SUCCESS,
|
||||||
task_args=("/tmp/paperless/paperless-upload-5iq7skzc",),
|
|
||||||
task_kwargs={
|
|
||||||
"override_filename": "test.pdf",
|
|
||||||
"override_title": None,
|
|
||||||
"override_correspondent_id": None,
|
|
||||||
"override_document_type_id": None,
|
|
||||||
"override_tag_ids": None,
|
|
||||||
"task_id": "466e8fe7-7193-4698-9fff-72f0340e2082",
|
|
||||||
"override_created": None,
|
|
||||||
},
|
|
||||||
)
|
)
|
||||||
|
|
||||||
response = self.client.get(self.ENDPOINT)
|
response = self.client.get(self.ENDPOINT)
|
||||||
@ -3161,8 +3151,6 @@ class TestTasks(APITestCase):
|
|||||||
task_file_name="anothertest.pdf",
|
task_file_name="anothertest.pdf",
|
||||||
task_name="documents.tasks.some_task",
|
task_name="documents.tasks.some_task",
|
||||||
status=celery.states.SUCCESS,
|
status=celery.states.SUCCESS,
|
||||||
task_args=("/consume/anothertest.pdf",),
|
|
||||||
task_kwargs={"override_tag_ids": None},
|
|
||||||
)
|
)
|
||||||
|
|
||||||
response = self.client.get(self.ENDPOINT)
|
response = self.client.get(self.ENDPOINT)
|
||||||
|
@ -28,6 +28,14 @@ class TestTaskSignalHandler(DirectoriesMixin, TestCase):
|
|||||||
"ignore_result": False,
|
"ignore_result": False,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
BODY_CONSUME = (
|
||||||
|
# args
|
||||||
|
("/consume/hello-999.pdf",),
|
||||||
|
# kwargs
|
||||||
|
{"override_tag_ids": None},
|
||||||
|
{"callbacks": None, "errbacks": None, "chain": None, "chord": None},
|
||||||
|
)
|
||||||
|
|
||||||
HEADERS_WEB_UI = {
|
HEADERS_WEB_UI = {
|
||||||
"lang": "py",
|
"lang": "py",
|
||||||
"task": "documents.tasks.consume_file",
|
"task": "documents.tasks.consume_file",
|
||||||
@ -47,64 +55,90 @@ class TestTaskSignalHandler(DirectoriesMixin, TestCase):
|
|||||||
"ignore_result": False,
|
"ignore_result": False,
|
||||||
}
|
}
|
||||||
|
|
||||||
def util_call_before_task_publish_handler(self, headers_to_use):
|
BODY_WEB_UI = (
|
||||||
|
# args
|
||||||
|
("/tmp/paperless/paperless-upload-st9lmbvx",),
|
||||||
|
# kwargs
|
||||||
|
{
|
||||||
|
"override_filename": "statement.pdf",
|
||||||
|
"override_title": None,
|
||||||
|
"override_correspondent_id": None,
|
||||||
|
"override_document_type_id": None,
|
||||||
|
"override_tag_ids": None,
|
||||||
|
"task_id": "f5622ca9-3707-4ed0-b418-9680b912572f",
|
||||||
|
"override_created": None,
|
||||||
|
},
|
||||||
|
{"callbacks": None, "errbacks": None, "chain": None, "chord": None},
|
||||||
|
)
|
||||||
|
|
||||||
|
def util_call_before_task_publish_handler(self, headers_to_use, body_to_use):
|
||||||
|
"""
|
||||||
|
Simple utility to call the pre-run handle and ensure it created a single task
|
||||||
|
instance
|
||||||
|
"""
|
||||||
self.assertEqual(PaperlessTask.objects.all().count(), 0)
|
self.assertEqual(PaperlessTask.objects.all().count(), 0)
|
||||||
|
|
||||||
before_task_publish_handler(headers=headers_to_use)
|
before_task_publish_handler(headers=headers_to_use, body=body_to_use)
|
||||||
|
|
||||||
self.assertEqual(PaperlessTask.objects.all().count(), 1)
|
self.assertEqual(PaperlessTask.objects.all().count(), 1)
|
||||||
|
|
||||||
def test_before_task_publish_handler_consume(self):
|
def test_before_task_publish_handler_consume(self):
|
||||||
"""
|
"""
|
||||||
GIVEN:
|
GIVEN:
|
||||||
- A celery task completed with an exception
|
- A celery task is started via the consume folder
|
||||||
WHEN:
|
WHEN:
|
||||||
- API call is made to get tasks
|
- Task before publish handler is called
|
||||||
THEN:
|
THEN:
|
||||||
- The returned result is the exception info
|
- The task is created and marked as pending
|
||||||
"""
|
"""
|
||||||
self.util_call_before_task_publish_handler(headers_to_use=self.HEADERS_CONSUME)
|
self.util_call_before_task_publish_handler(
|
||||||
|
headers_to_use=self.HEADERS_CONSUME,
|
||||||
|
body_to_use=self.BODY_CONSUME,
|
||||||
|
)
|
||||||
|
|
||||||
task = PaperlessTask.objects.get()
|
task = PaperlessTask.objects.get()
|
||||||
self.assertIsNotNone(task)
|
self.assertIsNotNone(task)
|
||||||
self.assertEqual(self.HEADERS_CONSUME["id"], task.task_id)
|
self.assertEqual(self.HEADERS_CONSUME["id"], task.task_id)
|
||||||
self.assertListEqual(["/consume/hello-999.pdf"], task.task_args)
|
|
||||||
self.assertDictEqual({"override_tag_ids": None}, task.task_kwargs)
|
|
||||||
self.assertEqual("hello-999.pdf", task.task_file_name)
|
self.assertEqual("hello-999.pdf", task.task_file_name)
|
||||||
self.assertEqual("documents.tasks.consume_file", task.task_name)
|
self.assertEqual("documents.tasks.consume_file", task.task_name)
|
||||||
self.assertEqual(celery.states.PENDING, task.status)
|
self.assertEqual(celery.states.PENDING, task.status)
|
||||||
|
|
||||||
def test_before_task_publish_handler_webui(self):
|
def test_before_task_publish_handler_webui(self):
|
||||||
|
"""
|
||||||
self.util_call_before_task_publish_handler(headers_to_use=self.HEADERS_WEB_UI)
|
GIVEN:
|
||||||
|
- A celery task is started via the web ui
|
||||||
|
WHEN:
|
||||||
|
- Task before publish handler is called
|
||||||
|
THEN:
|
||||||
|
- The task is created and marked as pending
|
||||||
|
"""
|
||||||
|
self.util_call_before_task_publish_handler(
|
||||||
|
headers_to_use=self.HEADERS_WEB_UI,
|
||||||
|
body_to_use=self.BODY_WEB_UI,
|
||||||
|
)
|
||||||
|
|
||||||
task = PaperlessTask.objects.get()
|
task = PaperlessTask.objects.get()
|
||||||
|
|
||||||
self.assertIsNotNone(task)
|
self.assertIsNotNone(task)
|
||||||
|
|
||||||
self.assertEqual(self.HEADERS_WEB_UI["id"], task.task_id)
|
self.assertEqual(self.HEADERS_WEB_UI["id"], task.task_id)
|
||||||
self.assertListEqual(
|
|
||||||
["/tmp/paperless/paperless-upload-st9lmbvx"],
|
|
||||||
task.task_args,
|
|
||||||
)
|
|
||||||
self.assertDictEqual(
|
|
||||||
{
|
|
||||||
"override_filename": "statement.pdf",
|
|
||||||
"override_title": None,
|
|
||||||
"override_correspondent_id": None,
|
|
||||||
"override_document_type_id": None,
|
|
||||||
"override_tag_ids": None,
|
|
||||||
"task_id": "f5622ca9-3707-4ed0-b418-9680b912572f",
|
|
||||||
"override_created": None,
|
|
||||||
},
|
|
||||||
task.task_kwargs,
|
|
||||||
)
|
|
||||||
self.assertEqual("statement.pdf", task.task_file_name)
|
self.assertEqual("statement.pdf", task.task_file_name)
|
||||||
self.assertEqual("documents.tasks.consume_file", task.task_name)
|
self.assertEqual("documents.tasks.consume_file", task.task_name)
|
||||||
self.assertEqual(celery.states.PENDING, task.status)
|
self.assertEqual(celery.states.PENDING, task.status)
|
||||||
|
|
||||||
def test_task_prerun_handler(self):
|
def test_task_prerun_handler(self):
|
||||||
self.util_call_before_task_publish_handler(headers_to_use=self.HEADERS_CONSUME)
|
"""
|
||||||
|
GIVEN:
|
||||||
|
- A celery task is started via the consume folder
|
||||||
|
WHEN:
|
||||||
|
- Task starts execution
|
||||||
|
THEN:
|
||||||
|
- The task is marked as started
|
||||||
|
"""
|
||||||
|
self.util_call_before_task_publish_handler(
|
||||||
|
headers_to_use=self.HEADERS_CONSUME,
|
||||||
|
body_to_use=self.BODY_CONSUME,
|
||||||
|
)
|
||||||
|
|
||||||
task_prerun_handler(task_id=self.HEADERS_CONSUME["id"])
|
task_prerun_handler(task_id=self.HEADERS_CONSUME["id"])
|
||||||
|
|
||||||
@ -113,7 +147,18 @@ class TestTaskSignalHandler(DirectoriesMixin, TestCase):
|
|||||||
self.assertEqual(celery.states.STARTED, task.status)
|
self.assertEqual(celery.states.STARTED, task.status)
|
||||||
|
|
||||||
def test_task_postrun_handler(self):
|
def test_task_postrun_handler(self):
|
||||||
self.util_call_before_task_publish_handler(headers_to_use=self.HEADERS_CONSUME)
|
"""
|
||||||
|
GIVEN:
|
||||||
|
- A celery task is started via the consume folder
|
||||||
|
WHEN:
|
||||||
|
- Task finished execution
|
||||||
|
THEN:
|
||||||
|
- The task is marked as started
|
||||||
|
"""
|
||||||
|
self.util_call_before_task_publish_handler(
|
||||||
|
headers_to_use=self.HEADERS_CONSUME,
|
||||||
|
body_to_use=self.BODY_CONSUME,
|
||||||
|
)
|
||||||
|
|
||||||
task_postrun_handler(
|
task_postrun_handler(
|
||||||
task_id=self.HEADERS_CONSUME["id"],
|
task_id=self.HEADERS_CONSUME["id"],
|
||||||
|
@ -529,6 +529,7 @@ CELERY_RESULT_EXTENDED = True
|
|||||||
CELERY_RESULT_BACKEND = "django-db"
|
CELERY_RESULT_BACKEND = "django-db"
|
||||||
CELERY_CACHE_BACKEND = "default"
|
CELERY_CACHE_BACKEND = "default"
|
||||||
|
|
||||||
|
|
||||||
CELERY_BEAT_SCHEDULE = {
|
CELERY_BEAT_SCHEDULE = {
|
||||||
# Every ten minutes
|
# Every ten minutes
|
||||||
"Check all e-mail accounts": {
|
"Check all e-mail accounts": {
|
||||||
|
Loading…
x
Reference in New Issue
Block a user