mirror of
https://github.com/paperless-ngx/paperless-ngx.git
synced 2025-04-02 13:45:10 -05:00
323 lines
10 KiB
Python
Executable File
323 lines
10 KiB
Python
Executable File
import logging
|
|
import os
|
|
from subprocess import Popen
|
|
|
|
from django.conf import settings
|
|
from django.contrib.admin.models import ADDITION, LogEntry
|
|
from django.contrib.auth.models import User
|
|
from django.contrib.contenttypes.models import ContentType
|
|
from django.db import models, DatabaseError
|
|
from django.dispatch import receiver
|
|
from django.utils import timezone
|
|
from rest_framework.reverse import reverse
|
|
|
|
from .. import index, matching
|
|
from ..file_handling import delete_empty_directories, generate_filename, \
|
|
create_source_path_directory, archive_name_from_filename
|
|
from ..models import Document, Tag
|
|
|
|
|
|
def logger(message, group):
|
|
logging.getLogger(__name__).debug(message, extra={"group": group})
|
|
|
|
|
|
def add_inbox_tags(sender, document=None, logging_group=None, **kwargs):
|
|
inbox_tags = Tag.objects.filter(is_inbox_tag=True)
|
|
document.tags.add(*inbox_tags)
|
|
|
|
|
|
def set_correspondent(sender,
|
|
document=None,
|
|
logging_group=None,
|
|
classifier=None,
|
|
replace=False,
|
|
use_first=True,
|
|
**kwargs):
|
|
if document.correspondent and not replace:
|
|
return
|
|
|
|
potential_correspondents = matching.match_correspondents(document.content,
|
|
classifier)
|
|
|
|
potential_count = len(potential_correspondents)
|
|
if potential_correspondents:
|
|
selected = potential_correspondents[0]
|
|
else:
|
|
selected = None
|
|
if potential_count > 1:
|
|
if use_first:
|
|
logger(
|
|
f"Detected {potential_count} potential correspondents, "
|
|
f"so we've opted for {selected}",
|
|
logging_group
|
|
)
|
|
else:
|
|
logger(
|
|
f"Detected {potential_count} potential correspondents, "
|
|
f"not assigning any correspondent",
|
|
logging_group
|
|
)
|
|
return
|
|
|
|
if selected or replace:
|
|
logger(
|
|
f"Assigning correspondent {selected} to {document}",
|
|
logging_group
|
|
)
|
|
|
|
document.correspondent = selected
|
|
document.save(update_fields=("correspondent",))
|
|
|
|
|
|
def set_document_type(sender,
|
|
document=None,
|
|
logging_group=None,
|
|
classifier=None,
|
|
replace=False,
|
|
use_first=True,
|
|
**kwargs):
|
|
if document.document_type and not replace:
|
|
return
|
|
|
|
potential_document_type = matching.match_document_types(document.content,
|
|
classifier)
|
|
|
|
potential_count = len(potential_document_type)
|
|
if potential_document_type:
|
|
selected = potential_document_type[0]
|
|
else:
|
|
selected = None
|
|
|
|
if potential_count > 1:
|
|
if use_first:
|
|
logger(
|
|
f"Detected {potential_count} potential document types, "
|
|
f"so we've opted for {selected}",
|
|
logging_group
|
|
)
|
|
else:
|
|
logger(
|
|
f"Detected {potential_count} potential document types, "
|
|
f"not assigning any document type",
|
|
logging_group
|
|
)
|
|
return
|
|
|
|
if selected or replace:
|
|
logger(
|
|
f"Assigning document type {selected} to {document}",
|
|
logging_group
|
|
)
|
|
|
|
document.document_type = selected
|
|
document.save(update_fields=("document_type",))
|
|
|
|
|
|
def set_tags(sender,
|
|
document=None,
|
|
logging_group=None,
|
|
classifier=None,
|
|
replace=False,
|
|
**kwargs):
|
|
if replace:
|
|
document.tags.clear()
|
|
current_tags = set([])
|
|
else:
|
|
current_tags = set(document.tags.all())
|
|
|
|
matched_tags = matching.match_tags(document.content, classifier)
|
|
|
|
relevant_tags = set(matched_tags) - current_tags
|
|
|
|
if not relevant_tags:
|
|
return
|
|
|
|
message = 'Tagging "{}" with "{}"'
|
|
logger(
|
|
message.format(document, ", ".join([t.slug for t in relevant_tags])),
|
|
logging_group
|
|
)
|
|
|
|
document.tags.add(*relevant_tags)
|
|
|
|
|
|
def run_pre_consume_script(sender, filename, **kwargs):
|
|
|
|
if not settings.PRE_CONSUME_SCRIPT:
|
|
return
|
|
|
|
Popen((settings.PRE_CONSUME_SCRIPT, filename)).wait()
|
|
|
|
|
|
def run_post_consume_script(sender, document, **kwargs):
|
|
|
|
if not settings.POST_CONSUME_SCRIPT:
|
|
return
|
|
|
|
Popen((
|
|
settings.POST_CONSUME_SCRIPT,
|
|
str(document.pk),
|
|
document.file_name,
|
|
os.path.normpath(document.source_path),
|
|
os.path.normpath(document.thumbnail_path),
|
|
reverse("document-download", kwargs={"pk": document.pk}),
|
|
reverse("document-thumb", kwargs={"pk": document.pk}),
|
|
str(document.correspondent),
|
|
str(",".join(document.tags.all().values_list("slug", flat=True)))
|
|
)).wait()
|
|
|
|
|
|
@receiver(models.signals.post_delete, sender=Document)
|
|
def cleanup_document_deletion(sender, instance, using, **kwargs):
|
|
for f in (instance.source_path,
|
|
instance.archive_path,
|
|
instance.thumbnail_path):
|
|
if os.path.isfile(f):
|
|
try:
|
|
os.unlink(f)
|
|
logging.getLogger(__name__).debug(
|
|
f"Deleted file {f}.")
|
|
except OSError as e:
|
|
logging.getLogger(__name__).warning(
|
|
f"While deleting document {instance.file_name}, the file "
|
|
f"{f} could not be deleted: {e}"
|
|
)
|
|
|
|
delete_empty_directories(
|
|
os.path.dirname(instance.source_path),
|
|
root=settings.ORIGINALS_DIR
|
|
)
|
|
|
|
delete_empty_directories(
|
|
os.path.dirname(instance.archive_path),
|
|
root=settings.ARCHIVE_DIR
|
|
)
|
|
|
|
|
|
def validate_move(instance, old_path, new_path):
|
|
if not os.path.isfile(old_path):
|
|
# Can't do anything if the old file does not exist anymore.
|
|
logging.getLogger(__name__).fatal(
|
|
f"Document {str(instance)}: File {old_path} has gone.")
|
|
return False
|
|
|
|
if os.path.isfile(new_path):
|
|
# Can't do anything if the new file already exists. Skip updating file.
|
|
logging.getLogger(__name__).warning(
|
|
f"Document {str(instance)}: Cannot rename file "
|
|
f"since target path {new_path} already exists.")
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
@receiver(models.signals.m2m_changed, sender=Document.tags.through)
|
|
@receiver(models.signals.post_save, sender=Document)
|
|
def update_filename_and_move_files(sender, instance, **kwargs):
|
|
|
|
if not instance.filename:
|
|
# Can't update the filename if there is no filename to begin with
|
|
# This happens when the consumer creates a new document.
|
|
# The document is modified and saved multiple times, and only after
|
|
# everything is done (i.e., the generated filename is final),
|
|
# filename will be set to the location where the consumer has put
|
|
# the file.
|
|
#
|
|
# This will in turn cause this logic to move the file where it belongs.
|
|
return
|
|
|
|
old_filename = instance.filename
|
|
new_filename = generate_filename(instance)
|
|
|
|
if new_filename == instance.filename:
|
|
# Don't do anything if its the same.
|
|
return
|
|
|
|
old_source_path = instance.source_path
|
|
new_source_path = os.path.join(settings.ORIGINALS_DIR, new_filename)
|
|
|
|
if not validate_move(instance, old_source_path, new_source_path):
|
|
return
|
|
|
|
# archive files are optional, archive checksum tells us if we have one,
|
|
# since this is None for documents without archived files.
|
|
if instance.archive_checksum:
|
|
new_archive_filename = archive_name_from_filename(new_filename)
|
|
old_archive_path = instance.archive_path
|
|
new_archive_path = os.path.join(settings.ARCHIVE_DIR,
|
|
new_archive_filename)
|
|
|
|
if not validate_move(instance, old_archive_path, new_archive_path):
|
|
return
|
|
|
|
create_source_path_directory(new_archive_path)
|
|
else:
|
|
old_archive_path = None
|
|
new_archive_path = None
|
|
|
|
create_source_path_directory(new_source_path)
|
|
|
|
try:
|
|
os.rename(old_source_path, new_source_path)
|
|
if instance.archive_checksum:
|
|
os.rename(old_archive_path, new_archive_path)
|
|
instance.filename = new_filename
|
|
# Don't save here to prevent infinite recursion.
|
|
Document.objects.filter(pk=instance.pk).update(filename=new_filename)
|
|
|
|
logging.getLogger(__name__).debug(
|
|
f"Moved file {old_source_path} to {new_source_path}.")
|
|
|
|
if instance.archive_checksum:
|
|
logging.getLogger(__name__).debug(
|
|
f"Moved file {old_archive_path} to {new_archive_path}.")
|
|
|
|
except OSError as e:
|
|
instance.filename = old_filename
|
|
# this happens when we can't move a file. If that's the case for the
|
|
# archive file, we try our best to revert the changes.
|
|
try:
|
|
os.rename(new_source_path, old_source_path)
|
|
os.rename(new_archive_path, old_archive_path)
|
|
except Exception as e:
|
|
# This is fine, since:
|
|
# A: if we managed to move source from A to B, we will also manage
|
|
# to move it from B to A. If not, we have a serious issue
|
|
# that's going to get caught by the santiy checker.
|
|
# all files remain in place and will never be overwritten,
|
|
# so this is not the end of the world.
|
|
# B: if moving the orignal file failed, nothing has changed anyway.
|
|
pass
|
|
except DatabaseError as e:
|
|
os.rename(new_source_path, old_source_path)
|
|
if instance.archive_checksum:
|
|
os.rename(new_archive_path, old_archive_path)
|
|
instance.filename = old_filename
|
|
|
|
if not os.path.isfile(old_source_path):
|
|
delete_empty_directories(os.path.dirname(old_source_path),
|
|
root=settings.ORIGINALS_DIR)
|
|
|
|
if old_archive_path and not os.path.isfile(old_archive_path):
|
|
delete_empty_directories(os.path.dirname(old_archive_path),
|
|
root=settings.ARCHIVE_DIR)
|
|
|
|
|
|
def set_log_entry(sender, document=None, logging_group=None, **kwargs):
|
|
|
|
ct = ContentType.objects.get(model="document")
|
|
user = User.objects.get(username="consumer")
|
|
|
|
LogEntry.objects.create(
|
|
action_flag=ADDITION,
|
|
action_time=timezone.now(),
|
|
content_type=ct,
|
|
object_id=document.pk,
|
|
user=user,
|
|
object_repr=document.__str__(),
|
|
)
|
|
|
|
|
|
def add_to_index(sender, document, **kwargs):
|
|
index.add_or_update_document(document)
|