mirror of
https://github.com/paperless-ngx/paperless-ngx.git
synced 2025-04-02 13:45:10 -05:00
better exception logging
This commit is contained in:
parent
4211153527
commit
ed0b1fe115
@ -43,9 +43,9 @@ def load_classifier():
|
|||||||
version=version, timeout=86400)
|
version=version, timeout=86400)
|
||||||
except (EOFError, IncompatibleClassifierVersionError) as e:
|
except (EOFError, IncompatibleClassifierVersionError) as e:
|
||||||
# there's something wrong with the model file.
|
# there's something wrong with the model file.
|
||||||
logger.error(
|
logger.exception(
|
||||||
f"Unrecoverable error while loading document "
|
f"Unrecoverable error while loading document "
|
||||||
f"classification model: {str(e)}, deleting model file."
|
f"classification model, deleting model file."
|
||||||
)
|
)
|
||||||
os.unlink(settings.MODEL_FILE)
|
os.unlink(settings.MODEL_FILE)
|
||||||
classifier = None
|
classifier = None
|
||||||
|
@ -78,8 +78,8 @@ def open_index(recreate=False):
|
|||||||
try:
|
try:
|
||||||
if exists_in(settings.INDEX_DIR) and not recreate:
|
if exists_in(settings.INDEX_DIR) and not recreate:
|
||||||
return open_dir(settings.INDEX_DIR, schema=get_schema())
|
return open_dir(settings.INDEX_DIR, schema=get_schema())
|
||||||
except Exception as e:
|
except Exception:
|
||||||
logger.error(f"Error while opening the index: {e}, recreating.")
|
logger.exception(f"Error while opening the index, recreating.")
|
||||||
|
|
||||||
if not os.path.isdir(settings.INDEX_DIR):
|
if not os.path.isdir(settings.INDEX_DIR):
|
||||||
os.makedirs(settings.INDEX_DIR, exist_ok=True)
|
os.makedirs(settings.INDEX_DIR, exist_ok=True)
|
||||||
|
@ -60,7 +60,7 @@ def handle_document(document_id):
|
|||||||
index.update_document(writer, document)
|
index.update_document(writer, document)
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Error while parsing document {document}: {str(e)}")
|
logger.exception(f"Error while parsing document {document}")
|
||||||
finally:
|
finally:
|
||||||
parser.cleanup()
|
parser.cleanup()
|
||||||
|
|
||||||
|
@ -54,8 +54,7 @@ def _consume(filepath):
|
|||||||
if settings.CONSUMER_SUBDIRS_AS_TAGS:
|
if settings.CONSUMER_SUBDIRS_AS_TAGS:
|
||||||
tag_ids = _tags_from_path(filepath)
|
tag_ids = _tags_from_path(filepath)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(
|
logger.exception("Error creating tags from path")
|
||||||
"Error creating tags from path: {}".format(e))
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
async_task("documents.tasks.consume_file",
|
async_task("documents.tasks.consume_file",
|
||||||
@ -66,8 +65,7 @@ def _consume(filepath):
|
|||||||
# Catch all so that the consumer won't crash.
|
# Catch all so that the consumer won't crash.
|
||||||
# This is also what the test case is listening for to check for
|
# This is also what the test case is listening for to check for
|
||||||
# errors.
|
# errors.
|
||||||
logger.error(
|
logger.exception("Error while consuming document")
|
||||||
"Error while consuming document: {}".format(e))
|
|
||||||
|
|
||||||
|
|
||||||
def _consume_wait_unmodified(file, num_tries=20, wait_time=1):
|
def _consume_wait_unmodified(file, num_tries=20, wait_time=1):
|
||||||
|
@ -13,11 +13,8 @@ def process_mail_accounts():
|
|||||||
try:
|
try:
|
||||||
total_new_documents += MailAccountHandler().handle_mail_account(
|
total_new_documents += MailAccountHandler().handle_mail_account(
|
||||||
account)
|
account)
|
||||||
except MailError as e:
|
except MailError:
|
||||||
logger.error(
|
logger.exception(f"Error while processing mail account {account}")
|
||||||
f"Error while processing mail account {account}: {e}",
|
|
||||||
exc_info=True
|
|
||||||
)
|
|
||||||
|
|
||||||
if total_new_documents > 0:
|
if total_new_documents > 0:
|
||||||
return f"Added {total_new_documents} document(s)."
|
return f"Added {total_new_documents} document(s)."
|
||||||
|
Loading…
x
Reference in New Issue
Block a user