Compare commits
211 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
1a5548f171 | ||
|
|
53c49b9121 | ||
|
|
bd50872043 | ||
|
|
d37c4fd551 | ||
|
|
d9ed7cc6d1 | ||
|
|
0a0808c0a8 | ||
|
|
16354e3528 | ||
|
|
1018e88ec1 | ||
|
|
2916351517 | ||
|
|
46d4b02c81 | ||
|
|
58eac88a10 | ||
|
|
2ef3d8dfbe | ||
|
|
7cf4c29340 | ||
|
|
cdd84563dd | ||
|
|
8138ea9a60 | ||
|
|
c171f9a875 | ||
|
|
3f97a8b8ff | ||
|
|
1147767ff3 | ||
|
|
6a304215c3 | ||
|
|
31714506bd | ||
|
|
403d80cf3b | ||
|
|
7cf20f2877 | ||
|
|
afff08b017 | ||
|
|
685abb1905 | ||
|
|
41990727a3 | ||
|
|
b113d8ab33 | ||
|
|
7bd0e9cc2f | ||
|
|
1e030a51d4 | ||
|
|
238a5de5cc | ||
|
|
53b24ffb3d | ||
|
|
a7404afbcb | ||
|
|
0e6d5f993c | ||
|
|
60637ff13a | ||
|
|
df8259e3fe | ||
|
|
aab3b1cc75 | ||
|
|
9492b39f2d | ||
|
|
b9ac866ea7 | ||
|
|
678517dd89 | ||
|
|
221774c6f8 | ||
|
|
a2aa345712 | ||
|
|
d998beaff3 | ||
|
|
a379bc4e7c | ||
|
|
b8e1894f3b | ||
|
|
0b5a44090f | ||
|
|
b40835eca2 | ||
|
|
b38562c7cc | ||
|
|
70e9fb415e | ||
|
|
64612099cd | ||
|
|
45c20f979d | ||
|
|
743c76f159 | ||
|
|
bc5dad75d3 | ||
|
|
4acf05cf28 | ||
|
|
f5f7264077 | ||
|
|
4364bebf38 | ||
|
|
15e61768f2 | ||
|
|
dd0a0f5c4d | ||
|
|
086f5ba43b | ||
|
|
e16dcf629e | ||
|
|
f16ae5110b | ||
|
|
ab5cbe5ec3 | ||
|
|
be5da92f16 | ||
|
|
95954a65a0 | ||
|
|
0b55e8fa77 | ||
|
|
6f159e8959 | ||
|
|
5c413b4b00 | ||
|
|
cca64d3ed1 | ||
|
|
e11eaf6ff8 | ||
|
|
85a4c1d855 | ||
|
|
0f5e72623b | ||
|
|
061e549ad7 | ||
|
|
49d1a5d248 | ||
|
|
03d6b00db8 | ||
|
|
a2eb0f7201 | ||
|
|
5c71a0ca07 | ||
|
|
489d16fad9 | ||
|
|
a458707777 | ||
|
|
a1d0a86305 | ||
|
|
29f1d21be7 | ||
|
|
34c5b526c3 | ||
|
|
3edb6578ba | ||
|
|
984c036b6e | ||
|
|
a403ecb5c9 | ||
|
|
a44713409c | ||
|
|
567467b8ed | ||
|
|
139edd6104 | ||
|
|
e756d55abf | ||
|
|
015c8d2a78 | ||
|
|
5af846c13d | ||
|
|
e69a9c7a54 | ||
|
|
23cb2a9a53 | ||
|
|
b5e3397b88 | ||
|
|
5685a4055a | ||
|
|
97b72ef767 | ||
|
|
31489848be | ||
|
|
e5988d447b | ||
|
|
adfed748ce | ||
|
|
2444ba87c0 | ||
|
|
534457e713 | ||
|
|
ea82a9730e | ||
|
|
f04b872e14 | ||
|
|
e61894e425 | ||
|
|
35fbdaadac | ||
|
|
8441bc7328 | ||
|
|
37c95ff97b | ||
|
|
5b1ca33c57 | ||
|
|
ec8e09b34e | ||
|
|
bcf97eccfa | ||
|
|
f53b5cc7a6 | ||
|
|
27adde7aa7 | ||
|
|
a9719833e0 | ||
|
|
7bf37090ca | ||
|
|
44fcef7123 | ||
|
|
69a44b10a1 | ||
|
|
b085e3d049 | ||
|
|
4b953bcddc | ||
|
|
315eaa7080 | ||
|
|
5a9bc967f1 | ||
|
|
a0d7236d0b | ||
|
|
21e9a31ffe | ||
|
|
4ee46c0a97 | ||
|
|
10d9a930f9 | ||
|
|
a21ccdb21b | ||
|
|
7cdd7a8f35 | ||
|
|
01e03a47e0 | ||
|
|
1b9a71551a | ||
|
|
911efd1db4 | ||
|
|
e61f0a68c4 | ||
|
|
cefbcffd59 | ||
|
|
622a98d6d5 | ||
|
|
7901f5d1dc | ||
|
|
555c34d7a8 | ||
|
|
dcc0d1de20 | ||
|
|
7bdf4d622b | ||
|
|
4a7207b0d0 | ||
|
|
ad9c2ca0e8 | ||
|
|
da998ddb60 | ||
|
|
07f68815df | ||
|
|
35645f9ade | ||
|
|
7c3d91301c | ||
|
|
5bcf7403ad | ||
|
|
2d6c092b65 | ||
|
|
6d0689cad6 | ||
|
|
3f80e93ee0 | ||
|
|
1b18abab1d | ||
|
|
03dd5af5ab | ||
|
|
dfba82b07c | ||
|
|
08ca02c87f | ||
|
|
b61f4ec095 | ||
|
|
9dbe6a494b | ||
|
|
44e70939d6 | ||
|
|
ab6066eafa | ||
|
|
42258cdd36 | ||
|
|
d3de9e6893 | ||
|
|
333beb94af | ||
|
|
f3c0942c49 | ||
|
|
02adf53ab9 | ||
|
|
3497b5cab4 | ||
|
|
9c17dca17c | ||
|
|
de342d3177 | ||
|
|
743b452daf | ||
|
|
c762f3c337 | ||
|
|
31803d41bc | ||
|
|
2ecd9779fc | ||
|
|
5a7047233e | ||
|
|
999e9c3725 | ||
|
|
f6940fe878 | ||
|
|
ce65ff8fc8 | ||
|
|
eed6784f25 | ||
|
|
3d9ae356ea | ||
|
|
f688d074b5 | ||
|
|
41457d8fbd | ||
|
|
2c416ecc0e | ||
|
|
3ab33c557b | ||
|
|
8db05f4950 | ||
|
|
3d5bc82a03 | ||
|
|
14e3a0d80b | ||
|
|
fcd9e2716a | ||
|
|
d62d633215 | ||
|
|
3b0c9273c1 | ||
|
|
e4c1c11845 | ||
|
|
ae508fe0e5 | ||
|
|
2cb9b5399c | ||
|
|
134c47f515 | ||
|
|
d328c9d128 | ||
|
|
77b62b0fef | ||
|
|
ad09b18f3f | ||
|
|
b5af9c03a5 | ||
|
|
176c7e7532 | ||
|
|
15976888a0 | ||
|
|
9bee502903 | ||
|
|
e3cb8dc3e6 | ||
|
|
385285e5de | ||
|
|
127771dac9 | ||
|
|
cc98befba5 | ||
|
|
567549cba4 | ||
|
|
76c4f49be8 | ||
|
|
d9d89dc250 | ||
|
|
9358db6cee | ||
|
|
08c9d7db03 | ||
|
|
390b0a6dc9 | ||
|
|
ed6b861a47 | ||
|
|
85c7ee980c | ||
|
|
7ea773e6a9 | ||
|
|
e3c4ff38fe | ||
|
|
8b1f87b1c0 | ||
|
|
c5e4cd9ab4 | ||
|
|
215e36e9ed | ||
|
|
e3ef622031 | ||
|
|
f16760c466 | ||
|
|
b36287e573 | ||
|
|
4df7aa284b |
20
.build/Dockerfile
Normal file
20
.build/Dockerfile
Normal file
@@ -0,0 +1,20 @@
|
||||
FROM python:3.9-slim-buster AS deps
|
||||
|
||||
RUN apt-get update && \
|
||||
apt-get install -y build-essential git curl python3-dev libatlas3-base libatlas-base-dev liblapack-dev libxml2 libxml2-dev libffi6 libffi-dev musl-dev libxslt-dev
|
||||
|
||||
FROM deps AS testable
|
||||
ARG REPORT_PATH
|
||||
|
||||
VOLUME ["/var/mailgun", "/etc/mailgun/ssl", ${REPORT_PATH}]
|
||||
|
||||
ADD . /app
|
||||
WORKDIR /app
|
||||
COPY wheel/* /wheel/
|
||||
|
||||
RUN mkdir -p ${REPORT_PATH}
|
||||
|
||||
RUN python ./setup.py build bdist_wheel -d /wheel && \
|
||||
pip install --no-deps /wheel/*
|
||||
|
||||
ENTRYPOINT ["/bin/sh", "/app/run_tests.sh"]
|
||||
10
.gitignore
vendored
10
.gitignore
vendored
@@ -39,6 +39,8 @@ nosetests.xml
|
||||
/.emacs.desktop
|
||||
/.emacs.desktop.lock
|
||||
.elc
|
||||
.idea
|
||||
.cache
|
||||
auto-save-list
|
||||
tramp
|
||||
.\#*
|
||||
@@ -48,4 +50,10 @@ tramp
|
||||
*_archive
|
||||
|
||||
# Trial temp
|
||||
_trial_temp
|
||||
_trial_temp
|
||||
|
||||
# OSX
|
||||
.DS_Store
|
||||
|
||||
# vim-backup
|
||||
*.bak
|
||||
|
||||
11
MANIFEST.in
11
MANIFEST.in
@@ -1,9 +1,14 @@
|
||||
recursive-include tests *
|
||||
recursive-include talon *
|
||||
recursive-exclude tests *.pyc *~
|
||||
recursive-exclude talon *.pyc *~
|
||||
include train.data
|
||||
include classifier
|
||||
include LICENSE
|
||||
include MANIFEST.in
|
||||
include README.rst
|
||||
include README.rst
|
||||
include talon/signature/data/train.data
|
||||
include talon/signature/data/classifier
|
||||
include talon/signature/data/classifier_01.npy
|
||||
include talon/signature/data/classifier_02.npy
|
||||
include talon/signature/data/classifier_03.npy
|
||||
include talon/signature/data/classifier_04.npy
|
||||
include talon/signature/data/classifier_05.npy
|
||||
|
||||
47
README.rst
47
README.rst
@@ -89,21 +89,62 @@ the power of machine learning algorithms:
|
||||
# text == "Thanks Sasha, I can't go any higher and is why I limited it to the\nhomepage."
|
||||
# signature == "John Doe\nvia mobile"
|
||||
|
||||
For machine learning talon currently uses `PyML`_ library to build SVM
|
||||
For machine learning talon currently uses the `scikit-learn`_ library to build SVM
|
||||
classifiers. The core of machine learning algorithm lays in
|
||||
``talon.signature.learning package``. It defines a set of features to
|
||||
apply to a message (``featurespace.py``), how data sets are built
|
||||
(``dataset.py``), classifier’s interface (``classifier.py``).
|
||||
|
||||
The data used for training is taken from our personal email
|
||||
Currently the data used for training is taken from our personal email
|
||||
conversations and from `ENRON`_ dataset. As a result of applying our set
|
||||
of features to the dataset we provide files ``classifier`` and
|
||||
``train.data`` that don’t have any personal information but could be
|
||||
used to load trained classifier. Those files should be regenerated every
|
||||
time the feature/data set is changed.
|
||||
|
||||
.. _PyML: http://pyml.sourceforge.net/
|
||||
To regenerate the model files, you can run
|
||||
|
||||
.. code:: sh
|
||||
|
||||
python train.py
|
||||
|
||||
or
|
||||
|
||||
.. code:: python
|
||||
|
||||
from talon.signature import EXTRACTOR_FILENAME, EXTRACTOR_DATA
|
||||
from talon.signature.learning.classifier import train, init
|
||||
train(init(), EXTRACTOR_DATA, EXTRACTOR_FILENAME)
|
||||
|
||||
Open-source Dataset
|
||||
-------------------
|
||||
|
||||
Recently we started a `forge`_ project to create an open-source, annotated dataset of raw emails. In the project we
|
||||
used a subset of `ENRON`_ data, cleansed of private, health and financial information by `EDRM`_. At the moment over 190
|
||||
emails are annotated. Any contribution and collaboration on the project are welcome. Once the dataset is ready we plan to
|
||||
start using it for talon.
|
||||
|
||||
.. _scikit-learn: http://scikit-learn.org
|
||||
.. _ENRON: https://www.cs.cmu.edu/~enron/
|
||||
.. _EDRM: http://www.edrm.net/resources/data-sets/edrm-enron-email-data-set
|
||||
.. _forge: https://github.com/mailgun/forge
|
||||
|
||||
Training on your dataset
|
||||
------------------------
|
||||
|
||||
talon comes with a pre-processed dataset and a pre-trained classifier. To retrain the classifier on your own dataset of raw emails, structure and annotate them in the same way the `forge`_ project does. Then do:
|
||||
|
||||
.. code:: python
|
||||
|
||||
from talon.signature.learning.dataset import build_extraction_dataset
|
||||
from talon.signature.learning import classifier as c
|
||||
|
||||
build_extraction_dataset("/path/to/your/P/folder", "/path/to/talon/signature/data/train.data")
|
||||
c.train(c.init(), "/path/to/talon/signature/data/train.data", "/path/to/talon/signature/data/classifier")
|
||||
|
||||
Note that for signature extraction you need just the folder with the positive samples with annotated signature lines (P folder).
|
||||
|
||||
.. _forge: https://github.com/mailgun/forge
|
||||
|
||||
Research
|
||||
--------
|
||||
|
||||
11
requirements.txt
Normal file
11
requirements.txt
Normal file
@@ -0,0 +1,11 @@
|
||||
chardet>=1.0.1
|
||||
cchardet>=0.3.5
|
||||
cssselect
|
||||
html5lib
|
||||
joblib
|
||||
lxml>=2.3.3
|
||||
numpy
|
||||
regex>=1
|
||||
scikit-learn>=1.0.0
|
||||
scipy
|
||||
six>=1.10.0
|
||||
4
run_tests.sh
Executable file
4
run_tests.sh
Executable file
@@ -0,0 +1,4 @@
|
||||
#!/usr/bin/env bash
|
||||
set -ex
|
||||
REPORT_PATH="${REPORT_PATH:-./}"
|
||||
nosetests --with-xunit --with-coverage --cover-xml --cover-xml-file $REPORT_PATH/coverage.xml --xunit-file=$REPORT_PATH/nosetests.xml --cover-package=talon .
|
||||
136
setup.py
Normal file → Executable file
136
setup.py
Normal file → Executable file
@@ -1,13 +1,35 @@
|
||||
import os
|
||||
import sys
|
||||
import contextlib
|
||||
|
||||
from distutils.spawn import find_executable
|
||||
from __future__ import absolute_import
|
||||
from setuptools import setup, find_packages
|
||||
from setuptools.command.install import install
|
||||
|
||||
|
||||
class InstallCommand(install):
|
||||
user_options = install.user_options + [
|
||||
('no-ml', None, "Don't install without Machine Learning modules."),
|
||||
]
|
||||
|
||||
boolean_options = install.boolean_options + ['no-ml']
|
||||
|
||||
def initialize_options(self):
|
||||
install.initialize_options(self)
|
||||
self.no_ml = None
|
||||
|
||||
def finalize_options(self):
|
||||
install.finalize_options(self)
|
||||
if self.no_ml:
|
||||
dist = self.distribution
|
||||
dist.packages=find_packages(exclude=[
|
||||
"tests",
|
||||
"tests.*",
|
||||
"talon.signature",
|
||||
"talon.signature.*",
|
||||
])
|
||||
for not_required in ["numpy", "scipy", "scikit-learn==0.24.1"]:
|
||||
dist.install_requires.remove(not_required)
|
||||
|
||||
|
||||
setup(name='talon',
|
||||
version='1.0.2',
|
||||
version='1.4.9',
|
||||
description=("Mailgun library "
|
||||
"to extract message quotations and signatures."),
|
||||
long_description=open("README.rst").read(),
|
||||
@@ -15,92 +37,28 @@ setup(name='talon',
|
||||
author_email='admin@mailgunhq.com',
|
||||
url='https://github.com/mailgun/talon',
|
||||
license='APACHE2',
|
||||
packages=find_packages(exclude=['tests']),
|
||||
cmdclass={
|
||||
'install': InstallCommand,
|
||||
},
|
||||
packages=find_packages(exclude=['tests', 'tests.*']),
|
||||
include_package_data=True,
|
||||
zip_safe=True,
|
||||
install_requires=[
|
||||
"lxml==2.3.3",
|
||||
"regex==0.1.20110315",
|
||||
"chardet==1.0.1",
|
||||
"dnspython==1.11.1",
|
||||
"html2text",
|
||||
"nose==1.2.1",
|
||||
"lxml",
|
||||
"regex",
|
||||
"numpy",
|
||||
"scipy",
|
||||
"scikit-learn>=1.0.0",
|
||||
"chardet",
|
||||
"cchardet",
|
||||
"cssselect",
|
||||
"six",
|
||||
"html5lib",
|
||||
"joblib",
|
||||
],
|
||||
tests_require=[
|
||||
"mock",
|
||||
"coverage",
|
||||
"flanker"
|
||||
"nose",
|
||||
"coverage"
|
||||
]
|
||||
)
|
||||
|
||||
|
||||
def install_pyml():
|
||||
'''
|
||||
Downloads and installs PyML
|
||||
'''
|
||||
try:
|
||||
import PyML
|
||||
except:
|
||||
pass
|
||||
else:
|
||||
return
|
||||
|
||||
# install numpy first
|
||||
pip('install numpy==1.6.1 --upgrade')
|
||||
|
||||
pyml_tarball = (
|
||||
'http://09cce49df173f6f6e61f-fd6930021b51685920a6fa76529ee321'
|
||||
'.r45.cf2.rackcdn.com/PyML-0.7.9.tar.gz')
|
||||
pyml_srcidr = 'PyML-0.7.9'
|
||||
|
||||
# see if PyML tarball needs to be fetched:
|
||||
if not dir_exists(pyml_srcidr):
|
||||
run("curl %s | tar -xz" % pyml_tarball)
|
||||
|
||||
# compile&install:
|
||||
with cd(pyml_srcidr):
|
||||
python('setup.py build')
|
||||
python('setup.py install')
|
||||
|
||||
|
||||
def run(command):
|
||||
if os.system(command) != 0:
|
||||
raise Exception("Failed '{}'".format(command))
|
||||
else:
|
||||
return 0
|
||||
|
||||
|
||||
def python(command):
|
||||
command = '{} {}'.format(sys.executable, command)
|
||||
run(command)
|
||||
|
||||
|
||||
def enforce_executable(name, install_info):
|
||||
if os.system("which {}".format(name)) != 0:
|
||||
raise Exception(
|
||||
'{} utility is missing.\nTo install, run:\n\n{}\n'.format(
|
||||
name, install_info))
|
||||
|
||||
|
||||
def pip(command):
|
||||
command = '{} {}'.format(find_executable('pip'), command)
|
||||
run(command)
|
||||
|
||||
|
||||
def dir_exists(path):
|
||||
return os.path.isdir(path)
|
||||
|
||||
|
||||
@contextlib.contextmanager
|
||||
def cd(directory):
|
||||
curdir = os.getcwd()
|
||||
try:
|
||||
os.chdir(directory)
|
||||
yield {}
|
||||
finally:
|
||||
os.chdir(curdir)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
if len(sys.argv) > 1 and sys.argv[1] in ['develop', 'install']:
|
||||
enforce_executable('curl', 'sudo aptitude install curl')
|
||||
|
||||
install_pyml()
|
||||
|
||||
@@ -1,7 +1,13 @@
|
||||
from __future__ import absolute_import
|
||||
from talon.quotations import register_xpath_extensions
|
||||
from talon import signature
|
||||
try:
|
||||
from talon import signature
|
||||
ML_ENABLED = True
|
||||
except ImportError:
|
||||
ML_ENABLED = False
|
||||
|
||||
|
||||
def init():
|
||||
register_xpath_extensions()
|
||||
signature.initialize()
|
||||
if ML_ENABLED:
|
||||
signature.initialize()
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
from __future__ import absolute_import
|
||||
import regex as re
|
||||
|
||||
|
||||
|
||||
@@ -3,8 +3,10 @@ The module's functions operate on message bodies trying to extract original
|
||||
messages (without quoted messages) from html
|
||||
"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
import regex as re
|
||||
|
||||
from talon.utils import cssselect
|
||||
|
||||
CHECKPOINT_PREFIX = '#!%!'
|
||||
CHECKPOINT_SUFFIX = '!%!#'
|
||||
@@ -12,6 +14,7 @@ CHECKPOINT_PATTERN = re.compile(CHECKPOINT_PREFIX + '\d+' + CHECKPOINT_SUFFIX)
|
||||
|
||||
# HTML quote indicators (tag ids)
|
||||
QUOTE_IDS = ['OLK_SRC_BODY_SECTION']
|
||||
RE_FWD = re.compile("^[-]+[ ]*Forwarded message[ ]*[-]+$", re.I | re.M)
|
||||
|
||||
|
||||
def add_checkpoint(html_note, counter):
|
||||
@@ -76,22 +79,32 @@ def delete_quotation_tags(html_note, counter, quotation_checkpoints):
|
||||
|
||||
def cut_gmail_quote(html_message):
|
||||
''' Cuts the outermost block element with class gmail_quote. '''
|
||||
gmail_quote = html_message.cssselect('.gmail_quote')
|
||||
if gmail_quote:
|
||||
gmail_quote = cssselect('div.gmail_quote', html_message)
|
||||
if gmail_quote and (gmail_quote[0].text is None or not RE_FWD.match(gmail_quote[0].text)):
|
||||
gmail_quote[0].getparent().remove(gmail_quote[0])
|
||||
return True
|
||||
|
||||
|
||||
def cut_microsoft_quote(html_message):
|
||||
''' Cuts splitter block and all following blocks. '''
|
||||
#use EXSLT extensions to have a regex match() function with lxml
|
||||
ns = {"re": "http://exslt.org/regular-expressions"}
|
||||
|
||||
#general pattern: @style='border:none;border-top:solid <color> 1.0pt;padding:3.0pt 0<unit> 0<unit> 0<unit>'
|
||||
#outlook 2007, 2010 (international) <color=#B5C4DF> <unit=cm>
|
||||
#outlook 2007, 2010 (american) <color=#B5C4DF> <unit=pt>
|
||||
#outlook 2013 (international) <color=#E1E1E1> <unit=cm>
|
||||
#outlook 2013 (american) <color=#E1E1E1> <unit=pt>
|
||||
#also handles a variant with a space after the semicolon
|
||||
splitter = html_message.xpath(
|
||||
#outlook 2007, 2010
|
||||
"//div[@style='border:none;border-top:solid #B5C4DF 1.0pt;"
|
||||
"padding:3.0pt 0cm 0cm 0cm']|"
|
||||
#outlook 2007, 2010, 2013 (international, american)
|
||||
"//div[@style[re:match(., 'border:none; ?border-top:solid #(E1E1E1|B5C4DF) 1.0pt; ?"
|
||||
"padding:3.0pt 0(in|cm) 0(in|cm) 0(in|cm)')]]|"
|
||||
#windows mail
|
||||
"//div[@style='padding-top: 5px; "
|
||||
"border-top-color: rgb(229, 229, 229); "
|
||||
"border-top-width: 1px; border-top-style: solid;']"
|
||||
, namespaces=ns
|
||||
)
|
||||
|
||||
if splitter:
|
||||
@@ -130,7 +143,7 @@ def cut_microsoft_quote(html_message):
|
||||
def cut_by_id(html_message):
|
||||
found = False
|
||||
for quote_id in QUOTE_IDS:
|
||||
quote = html_message.cssselect('#{}'.format(quote_id))
|
||||
quote = cssselect('#{}'.format(quote_id), html_message)
|
||||
if quote:
|
||||
found = True
|
||||
quote[0].getparent().remove(quote[0])
|
||||
@@ -138,9 +151,14 @@ def cut_by_id(html_message):
|
||||
|
||||
|
||||
def cut_blockquote(html_message):
|
||||
''' Cuts blockquote with wrapping elements. '''
|
||||
quote = html_message.find('.//blockquote')
|
||||
if quote is not None:
|
||||
''' Cuts the last non-nested blockquote with wrapping elements.'''
|
||||
quote = html_message.xpath(
|
||||
'(.//blockquote)'
|
||||
'[not(@class="gmail_quote") and not(ancestor::blockquote)]'
|
||||
'[last()]')
|
||||
|
||||
if quote:
|
||||
quote = quote[0]
|
||||
quote.getparent().remove(quote)
|
||||
return True
|
||||
|
||||
@@ -154,21 +172,58 @@ def cut_from_block(html_message):
|
||||
|
||||
if block:
|
||||
block = block[-1]
|
||||
parent_div = None
|
||||
while block.getparent() is not None:
|
||||
if block.tag == 'div':
|
||||
block.getparent().remove(block)
|
||||
parent_div = block
|
||||
break
|
||||
block = block.getparent()
|
||||
if parent_div is not None:
|
||||
maybe_body = parent_div.getparent()
|
||||
# In cases where removing this enclosing div will remove all
|
||||
# content, we should assume the quote is not enclosed in a tag.
|
||||
parent_div_is_all_content = (
|
||||
maybe_body is not None and maybe_body.tag == 'body' and
|
||||
len(maybe_body.getchildren()) == 1)
|
||||
|
||||
if not parent_div_is_all_content:
|
||||
parent = block.getparent()
|
||||
next_sibling = block.getnext()
|
||||
|
||||
# remove all tags after found From block
|
||||
# (From block and quoted message are in separate divs)
|
||||
while next_sibling is not None:
|
||||
parent.remove(block)
|
||||
block = next_sibling
|
||||
next_sibling = block.getnext()
|
||||
|
||||
# remove the last sibling (or the
|
||||
# From block if no siblings)
|
||||
if block is not None:
|
||||
parent.remove(block)
|
||||
|
||||
return True
|
||||
else:
|
||||
block = block.getparent()
|
||||
else:
|
||||
# handle the case when From: block goes right after e.g. <hr>
|
||||
# and not enclosed in some tag
|
||||
block = html_message.xpath(
|
||||
("//*[starts-with(mg:tail(), 'From:')]|"
|
||||
"//*[starts-with(mg:tail(), 'Date:')]"))
|
||||
if block:
|
||||
block = block[0]
|
||||
while(block.getnext() is not None):
|
||||
block.getparent().remove(block.getnext())
|
||||
block.getparent().remove(block)
|
||||
return True
|
||||
else:
|
||||
return False
|
||||
|
||||
# handle the case when From: block goes right after e.g. <hr>
|
||||
# and not enclosed in some tag
|
||||
block = html_message.xpath(
|
||||
("//*[starts-with(mg:tail(), 'From:')]|"
|
||||
"//*[starts-with(mg:tail(), 'Date:')]"))
|
||||
if block:
|
||||
block = block[0]
|
||||
|
||||
if RE_FWD.match(block.getparent().text or ''):
|
||||
return False
|
||||
|
||||
while(block.getnext() is not None):
|
||||
block.getparent().remove(block.getnext())
|
||||
block.getparent().remove(block)
|
||||
return True
|
||||
|
||||
def cut_zimbra_quote(html_message):
|
||||
zDivider = html_message.xpath('//hr[@data-marker="__DIVIDER__"]')
|
||||
if zDivider:
|
||||
zDivider[0].getparent().remove(zDivider[0])
|
||||
return True
|
||||
|
||||
@@ -5,25 +5,27 @@ The module's functions operate on message bodies trying to extract
|
||||
original messages (without quoted messages)
|
||||
"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
import regex as re
|
||||
import logging
|
||||
from copy import deepcopy
|
||||
|
||||
from lxml import html, etree
|
||||
import html2text
|
||||
|
||||
from talon.constants import RE_DELIMITER
|
||||
from talon.utils import random_token, get_delimiter
|
||||
from talon.utils import (get_delimiter, html_tree_to_text,
|
||||
html_document_fromstring)
|
||||
from talon import html_quotations
|
||||
from six.moves import range
|
||||
import six
|
||||
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
RE_FWD = re.compile("^[-]+[ ]*Forwarded message[ ]*[-]+$", re.I | re.M)
|
||||
RE_FWD = re.compile("^[-]+[ ]*Forwarded message[ ]*[-]+\s*$", re.I | re.M)
|
||||
|
||||
RE_ON_DATE_SMB_WROTE = re.compile(
|
||||
u'(-*[ ]?({0})[ ].*({1})(.*\n){{0,2}}.*({2}):?-*)'.format(
|
||||
u'(-*[>]?[ ]?({0})[ ].*({1})(.*\n){{0,2}}.*({2}):?-*)'.format(
|
||||
# Beginning of the line
|
||||
u'|'.join((
|
||||
# English
|
||||
@@ -33,7 +35,17 @@ RE_ON_DATE_SMB_WROTE = re.compile(
|
||||
# Polish
|
||||
'W dniu',
|
||||
# Dutch
|
||||
'Op'
|
||||
'Op',
|
||||
# German
|
||||
'Am',
|
||||
# Portuguese
|
||||
'Em',
|
||||
# Norwegian
|
||||
u'På',
|
||||
# Swedish, Danish
|
||||
'Den',
|
||||
# Vietnamese
|
||||
u'Vào',
|
||||
)),
|
||||
# Date and sender separator
|
||||
u'|'.join((
|
||||
@@ -51,18 +63,32 @@ RE_ON_DATE_SMB_WROTE = re.compile(
|
||||
# Polish
|
||||
u'napisał',
|
||||
# Dutch
|
||||
'schreef','verzond','geschreven'
|
||||
'schreef','verzond','geschreven',
|
||||
# German
|
||||
'schrieb',
|
||||
# Portuguese
|
||||
'escreveu',
|
||||
# Norwegian, Swedish
|
||||
'skrev',
|
||||
# Vietnamese
|
||||
u'đã viết',
|
||||
))
|
||||
))
|
||||
# Special case for languages where text is translated like this: 'on {date} wrote {somebody}:'
|
||||
RE_ON_DATE_WROTE_SMB = re.compile(
|
||||
u'(-*[ ]?({0})[ ].*(.*\n){{0,2}}.*({1})[ ].*:)'.format(
|
||||
u'(-*[>]?[ ]?({0})[ ].*(.*\n){{0,2}}.*({1})[ ]*.*:)'.format(
|
||||
# Beginning of the line
|
||||
u'|'.join((
|
||||
'Op',
|
||||
#German
|
||||
'Am'
|
||||
)),
|
||||
# Ending of the line
|
||||
u'|'.join((
|
||||
# Dutch
|
||||
'schreef','verzond','geschreven'
|
||||
'schreef','verzond','geschreven',
|
||||
# German
|
||||
'schrieb'
|
||||
))
|
||||
)
|
||||
)
|
||||
@@ -93,7 +119,7 @@ RE_EMPTY_QUOTATION = re.compile(
|
||||
(
|
||||
# quotation border: splitter line or a number of quotation marker lines
|
||||
(?:
|
||||
s
|
||||
(?:se*)+
|
||||
|
|
||||
(?:me*){2,}
|
||||
)
|
||||
@@ -113,37 +139,70 @@ RE_ORIGINAL_MESSAGE = re.compile(u'[\s]*[-]+[ ]*({})[ ]*[-]+'.format(
|
||||
'Oprindelig meddelelse',
|
||||
))), re.I)
|
||||
|
||||
RE_FROM_COLON_OR_DATE_COLON = re.compile(u'(_+\r?\n)?[\s]*(:?[*]?{})[\s]?:[*]? .*'.format(
|
||||
RE_FROM_COLON_OR_DATE_COLON = re.compile(u'((_+\r?\n)?[\s]*:?[*]?({})[\s]?:([^\n$]+\n){{1,2}}){{2,}}'.format(
|
||||
u'|'.join((
|
||||
# "From" in different languages.
|
||||
'From', 'Van', 'De', 'Von', 'Fra',
|
||||
'From', 'Van', 'De', 'Von', 'Fra', u'Från',
|
||||
# "Date" in different languages.
|
||||
'Date', 'Datum', u'Envoyé'
|
||||
'Date', '[S]ent', 'Datum', u'Envoyé', 'Skickat', 'Sendt', 'Gesendet',
|
||||
# "Subject" in different languages.
|
||||
'Subject', 'Betreff', 'Objet', 'Emne', u'Ämne',
|
||||
# "To" in different languages.
|
||||
'To', 'An', 'Til', u'À', 'Till'
|
||||
))), re.I | re.M)
|
||||
|
||||
# ---- John Smith wrote ----
|
||||
RE_ANDROID_WROTE = re.compile(u'[\s]*[-]+.*({})[ ]*[-]+'.format(
|
||||
u'|'.join((
|
||||
# English
|
||||
'wrote',
|
||||
))), re.I)
|
||||
|
||||
# Support polymail.io reply format
|
||||
# On Tue, Apr 11, 2017 at 10:07 PM John Smith
|
||||
#
|
||||
# <
|
||||
# mailto:John Smith <johnsmith@gmail.com>
|
||||
# > wrote:
|
||||
RE_POLYMAIL = re.compile('On.*\s{2}<\smailto:.*\s> wrote:', re.I)
|
||||
|
||||
SPLITTER_PATTERNS = [
|
||||
RE_ORIGINAL_MESSAGE,
|
||||
# <date> <person>
|
||||
re.compile("(\d+/\d+/\d+|\d+\.\d+\.\d+).*@", re.VERBOSE),
|
||||
RE_ON_DATE_SMB_WROTE,
|
||||
RE_ON_DATE_WROTE_SMB,
|
||||
RE_FROM_COLON_OR_DATE_COLON,
|
||||
# 02.04.2012 14:20 пользователь "bob@example.com" <
|
||||
# bob@xxx.mailgun.org> написал:
|
||||
re.compile("(\d+/\d+/\d+|\d+\.\d+\.\d+).*\s\S+@\S+", re.S),
|
||||
# 2014-10-17 11:28 GMT+03:00 Bob <
|
||||
# bob@example.com>:
|
||||
re.compile("\d{4}-\d{2}-\d{2}\s+\d{2}:\d{2}\s+GMT.*\s\S+@\S+", re.S),
|
||||
# Thu, 26 Jun 2014 14:00:51 +0400 Bob <bob@example.com>:
|
||||
re.compile('\S{3,10}, \d\d? \S{3,10} 20\d\d,? \d\d?:\d\d(:\d\d)?'
|
||||
'( \S+){3,6}@\S+:')
|
||||
'( \S+){3,6}@\S+:'),
|
||||
# Sent from Samsung MobileName <address@example.com> wrote:
|
||||
re.compile('Sent from Samsung.* \S+@\S+> wrote'),
|
||||
RE_ANDROID_WROTE,
|
||||
RE_POLYMAIL
|
||||
]
|
||||
|
||||
|
||||
RE_LINK = re.compile('<(http://[^>]*)>')
|
||||
RE_NORMALIZED_LINK = re.compile('@@(http://[^>@]*)@@')
|
||||
|
||||
RE_PARENTHESIS_LINK = re.compile("\(https?://")
|
||||
|
||||
SPLITTER_MAX_LINES = 4
|
||||
SPLITTER_MAX_LINES = 6
|
||||
MAX_LINES_COUNT = 1000
|
||||
# an extensive research shows that exceeding this limit
|
||||
# leads to excessive processing time
|
||||
MAX_HTML_LEN = 2794202
|
||||
|
||||
QUOT_PATTERN = re.compile('^>+ ?')
|
||||
NO_QUOT_LINE = re.compile('^[^>].*[\S].*')
|
||||
|
||||
# Regular expression to identify if a line is a header.
|
||||
RE_HEADER = re.compile(": ")
|
||||
|
||||
|
||||
def extract_from(msg_body, content_type='text/plain'):
|
||||
try:
|
||||
@@ -151,12 +210,25 @@ def extract_from(msg_body, content_type='text/plain'):
|
||||
return extract_from_plain(msg_body)
|
||||
elif content_type == 'text/html':
|
||||
return extract_from_html(msg_body)
|
||||
except Exception, e:
|
||||
except Exception:
|
||||
log.exception('ERROR extracting message')
|
||||
|
||||
return msg_body
|
||||
|
||||
|
||||
def remove_initial_spaces_and_mark_message_lines(lines):
|
||||
"""
|
||||
Removes the initial spaces in each line before marking message lines.
|
||||
|
||||
This ensures headers can be identified if they are indented with spaces.
|
||||
"""
|
||||
i = 0
|
||||
while i < len(lines):
|
||||
lines[i] = lines[i].lstrip(' ')
|
||||
i += 1
|
||||
return mark_message_lines(lines)
|
||||
|
||||
|
||||
def mark_message_lines(lines):
|
||||
"""Mark message lines with markers to distinguish quotation lines.
|
||||
|
||||
@@ -170,7 +242,7 @@ def mark_message_lines(lines):
|
||||
>>> mark_message_lines(['answer', 'From: foo@bar.com', '', '> question'])
|
||||
'tsem'
|
||||
"""
|
||||
markers = bytearray(len(lines))
|
||||
markers = ['e' for _ in lines]
|
||||
i = 0
|
||||
while i < len(lines):
|
||||
if not lines[i].strip():
|
||||
@@ -182,10 +254,11 @@ def mark_message_lines(lines):
|
||||
else:
|
||||
# in case splitter is spread across several lines
|
||||
splitter = is_splitter('\n'.join(lines[i:i + SPLITTER_MAX_LINES]))
|
||||
|
||||
if splitter:
|
||||
# append as many splitter markers as lines in splitter
|
||||
splitter_lines = splitter.group().splitlines()
|
||||
for j in xrange(len(splitter_lines)):
|
||||
for j in range(len(splitter_lines)):
|
||||
markers[i + j] = 's'
|
||||
|
||||
# skip splitter lines
|
||||
@@ -195,7 +268,7 @@ def mark_message_lines(lines):
|
||||
markers[i] = 't'
|
||||
i += 1
|
||||
|
||||
return markers
|
||||
return ''.join(markers)
|
||||
|
||||
|
||||
def process_marked_lines(lines, markers, return_flags=[False, -1, -1]):
|
||||
@@ -209,6 +282,7 @@ def process_marked_lines(lines, markers, return_flags=[False, -1, -1]):
|
||||
return_flags = [were_lines_deleted, first_deleted_line,
|
||||
last_deleted_line]
|
||||
"""
|
||||
markers = ''.join(markers)
|
||||
# if there are no splitter there should be no markers
|
||||
if 's' not in markers and not re.search('(me*){3}', markers):
|
||||
markers = markers.replace('m', 't')
|
||||
@@ -220,7 +294,7 @@ def process_marked_lines(lines, markers, return_flags=[False, -1, -1]):
|
||||
# inlined reply
|
||||
# use lookbehind assertions to find overlapping entries e.g. for 'mtmtm'
|
||||
# both 't' entries should be found
|
||||
for inline_reply in re.finditer('(?<=m)e*((?:t+e*)+)m', markers):
|
||||
for inline_reply in re.finditer('(?<=m)e*(t[te]*)m', markers):
|
||||
# long links could break sequence of quotation lines but they shouldn't
|
||||
# be considered an inline reply
|
||||
links = (
|
||||
@@ -254,10 +328,27 @@ def preprocess(msg_body, delimiter, content_type='text/plain'):
|
||||
Replaces link brackets so that they couldn't be taken for quotation marker.
|
||||
Splits line in two if splitter pattern preceded by some text on the same
|
||||
line (done only for 'On <date> <person> wrote:' pattern).
|
||||
|
||||
Converts msg_body into a unicode.
|
||||
"""
|
||||
# normalize links i.e. replace '<', '>' wrapping the link with some symbols
|
||||
# so that '>' closing the link couldn't be mistakenly taken for quotation
|
||||
# marker.
|
||||
msg_body = _replace_link_brackets(msg_body)
|
||||
|
||||
msg_body = _wrap_splitter_with_newline(msg_body, delimiter, content_type)
|
||||
|
||||
return msg_body
|
||||
|
||||
|
||||
def _replace_link_brackets(msg_body):
|
||||
"""
|
||||
Normalize links i.e. replace '<', '>' wrapping the link with some symbols
|
||||
so that '>' closing the link couldn't be mistakenly taken for quotation
|
||||
marker.
|
||||
|
||||
Converts msg_body into a unicode
|
||||
"""
|
||||
if isinstance(msg_body, bytes):
|
||||
msg_body = msg_body.decode('utf8')
|
||||
|
||||
def link_wrapper(link):
|
||||
newline_index = msg_body[:link.start()].rfind("\n")
|
||||
if msg_body[newline_index + 1] == ">":
|
||||
@@ -266,7 +357,14 @@ def preprocess(msg_body, delimiter, content_type='text/plain'):
|
||||
return "@@%s@@" % link.group(1)
|
||||
|
||||
msg_body = re.sub(RE_LINK, link_wrapper, msg_body)
|
||||
return msg_body
|
||||
|
||||
|
||||
def _wrap_splitter_with_newline(msg_body, delimiter, content_type='text/plain'):
|
||||
"""
|
||||
Splits line in two if splitter pattern preceded by some text on the same
|
||||
line (done only for 'On <date> <person> wrote:' pattern.
|
||||
"""
|
||||
def splitter_wrapper(splitter):
|
||||
"""Wraps splitter with new line"""
|
||||
if splitter.start() and msg_body[splitter.start() - 1] != '\n':
|
||||
@@ -294,12 +392,8 @@ def extract_from_plain(msg_body):
|
||||
|
||||
delimiter = get_delimiter(msg_body)
|
||||
msg_body = preprocess(msg_body, delimiter)
|
||||
lines = msg_body.splitlines()
|
||||
|
||||
# don't process too long messages
|
||||
if len(lines) > MAX_LINES_COUNT:
|
||||
return stripped_text
|
||||
|
||||
lines = msg_body.splitlines()[:MAX_LINES_COUNT]
|
||||
markers = mark_message_lines(lines)
|
||||
lines = process_marked_lines(lines, markers)
|
||||
|
||||
@@ -324,44 +418,65 @@ def extract_from_html(msg_body):
|
||||
then extracting quotations from text,
|
||||
then checking deleted checkpoints,
|
||||
then deleting necessary tags.
|
||||
"""
|
||||
|
||||
if msg_body.strip() == '':
|
||||
Returns a unicode string.
|
||||
"""
|
||||
if isinstance(msg_body, six.text_type):
|
||||
msg_body = msg_body.encode('utf8')
|
||||
elif not isinstance(msg_body, bytes):
|
||||
msg_body = msg_body.encode('ascii')
|
||||
|
||||
result = _extract_from_html(msg_body)
|
||||
if isinstance(result, bytes):
|
||||
result = result.decode('utf8')
|
||||
|
||||
return result
|
||||
|
||||
|
||||
def _extract_from_html(msg_body):
|
||||
"""
|
||||
Extract not quoted message from provided html message body
|
||||
using tags and plain text algorithm.
|
||||
|
||||
Cut out first some encoding html tags such as xml and doctype
|
||||
for avoiding conflict with unicode decoding
|
||||
|
||||
Cut out the 'blockquote', 'gmail_quote' tags.
|
||||
Cut Microsoft quotations.
|
||||
|
||||
Then use plain text algorithm to cut out splitter or
|
||||
leftover quotation.
|
||||
This works by adding checkpoint text to all html tags,
|
||||
then converting html to text,
|
||||
then extracting quotations from text,
|
||||
then checking deleted checkpoints,
|
||||
then deleting necessary tags.
|
||||
"""
|
||||
if msg_body.strip() == b'':
|
||||
return msg_body
|
||||
|
||||
html_tree = html.document_fromstring(
|
||||
msg_body,
|
||||
parser=html.HTMLParser(encoding="utf-8")
|
||||
)
|
||||
msg_body = msg_body.replace(b'\r\n', b'\n')
|
||||
|
||||
msg_body = re.sub(br"\<\?xml.+\?\>|\<\!DOCTYPE.+]\>", "", msg_body)
|
||||
|
||||
html_tree = html_document_fromstring(msg_body)
|
||||
|
||||
if html_tree is None:
|
||||
return msg_body
|
||||
|
||||
cut_quotations = (html_quotations.cut_gmail_quote(html_tree) or
|
||||
html_quotations.cut_zimbra_quote(html_tree) or
|
||||
html_quotations.cut_blockquote(html_tree) or
|
||||
html_quotations.cut_microsoft_quote(html_tree) or
|
||||
html_quotations.cut_by_id(html_tree) or
|
||||
html_quotations.cut_from_block(html_tree)
|
||||
)
|
||||
|
||||
html_tree_copy = deepcopy(html_tree)
|
||||
|
||||
number_of_checkpoints = html_quotations.add_checkpoint(html_tree, 0)
|
||||
quotation_checkpoints = [False for i in xrange(number_of_checkpoints)]
|
||||
msg_with_checkpoints = html.tostring(html_tree)
|
||||
|
||||
h = html2text.HTML2Text()
|
||||
h.body_width = 0 # generate plain text without wrap
|
||||
|
||||
# html2text adds unnecessary star symbols. Remove them.
|
||||
# Mask star symbols
|
||||
msg_with_checkpoints = msg_with_checkpoints.replace('*', '3423oorkg432')
|
||||
plain_text = h.handle(msg_with_checkpoints)
|
||||
# Remove created star symbols
|
||||
plain_text = plain_text.replace('*', '')
|
||||
# Unmask saved star symbols
|
||||
plain_text = plain_text.replace('3423oorkg432', '*')
|
||||
|
||||
delimiter = get_delimiter(plain_text)
|
||||
|
||||
plain_text = preprocess(plain_text, delimiter, content_type='text/html')
|
||||
quotation_checkpoints = [False] * number_of_checkpoints
|
||||
plain_text = html_tree_to_text(html_tree)
|
||||
plain_text = preprocess(plain_text, '\n', content_type='text/html')
|
||||
lines = plain_text.splitlines()
|
||||
|
||||
# Don't process too long messages
|
||||
@@ -384,25 +499,165 @@ def extract_from_html(msg_body):
|
||||
process_marked_lines(lines, markers, return_flags)
|
||||
lines_were_deleted, first_deleted, last_deleted = return_flags
|
||||
|
||||
if not lines_were_deleted and not cut_quotations:
|
||||
return msg_body
|
||||
|
||||
if lines_were_deleted:
|
||||
#collect checkpoints from deleted lines
|
||||
for i in xrange(first_deleted, last_deleted):
|
||||
for i in range(first_deleted, last_deleted):
|
||||
for checkpoint in line_checkpoints[i]:
|
||||
quotation_checkpoints[checkpoint] = True
|
||||
else:
|
||||
if cut_quotations:
|
||||
return html.tostring(html_tree_copy)
|
||||
else:
|
||||
return msg_body
|
||||
|
||||
# Remove tags with quotation checkpoints
|
||||
html_quotations.delete_quotation_tags(
|
||||
html_tree_copy, 0, quotation_checkpoints
|
||||
)
|
||||
# Remove tags with quotation checkpoints
|
||||
html_quotations.delete_quotation_tags(
|
||||
html_tree_copy, 0, quotation_checkpoints
|
||||
)
|
||||
|
||||
if _readable_text_empty(html_tree_copy):
|
||||
return msg_body
|
||||
|
||||
# NOTE: We remove_namespaces() because we are using an HTML5 Parser, HTML
|
||||
# parsers do not recognize namespaces in HTML tags. As such the rendered
|
||||
# HTML tags are no longer recognizable HTML tags. Example: <o:p> becomes
|
||||
# <oU0003Ap>. When we port this to golang we should look into using an
|
||||
# XML Parser NOT and HTML5 Parser since we do not know what input a
|
||||
# customer will send us. Switching to a common XML parser in python
|
||||
# opens us up to a host of vulnerabilities.
|
||||
# See https://docs.python.org/3/library/xml.html#xml-vulnerabilities
|
||||
#
|
||||
# The down sides to removing the namespaces is that customers might
|
||||
# judge the XML namespaces important. If that is the case then support
|
||||
# should encourage customers to preform XML parsing of the un-stripped
|
||||
# body to get the full unmodified XML payload.
|
||||
#
|
||||
# Alternatives to this approach are
|
||||
# 1. Ignore the U0003A in tag names and let the customer deal with it.
|
||||
# This is not ideal, as most customers use stripped-html for viewing
|
||||
# emails sent from a recipient, as such they cannot control the HTML
|
||||
# provided by a recipient.
|
||||
# 2. Preform a string replace of 'U0003A' to ':' on the rendered HTML
|
||||
# string. While this would solve the issue simply, it runs the risk
|
||||
# of replacing data outside the <tag> which might be essential to
|
||||
# the customer.
|
||||
remove_namespaces(html_tree_copy)
|
||||
return html.tostring(html_tree_copy)
|
||||
|
||||
|
||||
def remove_namespaces(root):
|
||||
"""
|
||||
Given the root of an HTML document iterate through all the elements
|
||||
and remove any namespaces that might have been provided and remove
|
||||
any attributes that contain a namespace
|
||||
|
||||
<html xmlns:o="urn:schemas-microsoft-com:office:office">
|
||||
becomes
|
||||
<html>
|
||||
|
||||
<o:p>Hi</o:p>
|
||||
becomes
|
||||
<p>Hi</p>
|
||||
|
||||
Start tags do NOT have a namespace; COLON characters have no special meaning.
|
||||
if we don't remove the namespace the parser translates the tag name into a
|
||||
unicode representation. For example <o:p> becomes <oU0003Ap>
|
||||
|
||||
See https://www.w3.org/TR/2011/WD-html5-20110525/syntax.html#start-tags
|
||||
|
||||
|
||||
"""
|
||||
for child in root.iter():
|
||||
for key, value in child.attrib.items():
|
||||
# If the attribute includes a colon
|
||||
if key.rfind("U0003A") != -1:
|
||||
child.attrib.pop(key)
|
||||
|
||||
# If the tag includes a colon
|
||||
idx = child.tag.rfind("U0003A")
|
||||
if idx != -1:
|
||||
child.tag = child.tag[idx+6:]
|
||||
|
||||
return root
|
||||
|
||||
|
||||
def split_emails(msg):
|
||||
"""
|
||||
Given a message (which may consist of an email conversation thread with
|
||||
multiple emails), mark the lines to identify split lines, content lines and
|
||||
empty lines.
|
||||
|
||||
Correct the split line markers inside header blocks. Header blocks are
|
||||
identified by the regular expression RE_HEADER.
|
||||
|
||||
Return the corrected markers
|
||||
"""
|
||||
msg_body = _replace_link_brackets(msg)
|
||||
|
||||
# don't process too long messages
|
||||
lines = msg_body.splitlines()[:MAX_LINES_COUNT]
|
||||
markers = remove_initial_spaces_and_mark_message_lines(lines)
|
||||
|
||||
markers = _mark_quoted_email_splitlines(markers, lines)
|
||||
|
||||
# we don't want splitlines in header blocks
|
||||
markers = _correct_splitlines_in_headers(markers, lines)
|
||||
|
||||
return markers
|
||||
|
||||
|
||||
def _mark_quoted_email_splitlines(markers, lines):
|
||||
"""
|
||||
When there are headers indented with '>' characters, this method will
|
||||
attempt to identify if the header is a splitline header. If it is, then we
|
||||
mark it with 's' instead of leaving it as 'm' and return the new markers.
|
||||
"""
|
||||
# Create a list of markers to easily alter specific characters
|
||||
markerlist = list(markers)
|
||||
for i, line in enumerate(lines):
|
||||
if markerlist[i] != 'm':
|
||||
continue
|
||||
for pattern in SPLITTER_PATTERNS:
|
||||
matcher = re.search(pattern, line)
|
||||
if matcher:
|
||||
markerlist[i] = 's'
|
||||
break
|
||||
|
||||
return "".join(markerlist)
|
||||
|
||||
|
||||
def _correct_splitlines_in_headers(markers, lines):
|
||||
"""
|
||||
Corrects markers by removing splitlines deemed to be inside header blocks.
|
||||
"""
|
||||
updated_markers = ""
|
||||
i = 0
|
||||
in_header_block = False
|
||||
for m in markers:
|
||||
# Only set in_header_block flag when we hit an 's' and line is a header
|
||||
if m == 's':
|
||||
if not in_header_block:
|
||||
if bool(re.search(RE_HEADER, lines[i])):
|
||||
in_header_block = True
|
||||
else:
|
||||
if QUOT_PATTERN.match(lines[i]):
|
||||
m = 'm'
|
||||
else:
|
||||
m = 't'
|
||||
|
||||
# If the line is not a header line, set in_header_block false.
|
||||
if not bool(re.search(RE_HEADER, lines[i])):
|
||||
in_header_block = False
|
||||
|
||||
# Add the marker to the new updated markers string.
|
||||
updated_markers += m
|
||||
i += 1
|
||||
|
||||
return updated_markers
|
||||
|
||||
|
||||
def _readable_text_empty(html_tree):
|
||||
return not bool(html_tree_to_text(html_tree).strip())
|
||||
|
||||
|
||||
def is_splitter(line):
|
||||
'''
|
||||
Returns Matcher object if provided string is a splitter and
|
||||
@@ -416,7 +671,7 @@ def is_splitter(line):
|
||||
|
||||
def text_content(context):
|
||||
'''XPath Extension function to return a node text content.'''
|
||||
return context.context_node.text_content().strip()
|
||||
return context.context_node.xpath("string()").strip()
|
||||
|
||||
|
||||
def tail(context):
|
||||
|
||||
@@ -20,12 +20,11 @@ trained against, don't forget to regenerate:
|
||||
* signature/data/classifier
|
||||
"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
import os
|
||||
import sys
|
||||
from cStringIO import StringIO
|
||||
|
||||
from . import extraction
|
||||
from . extraction import extract
|
||||
from . extraction import extract #noqa
|
||||
from . learning import classifier
|
||||
|
||||
|
||||
@@ -36,13 +35,5 @@ EXTRACTOR_DATA = os.path.join(DATA_DIR, 'train.data')
|
||||
|
||||
|
||||
def initialize():
|
||||
try:
|
||||
# redirect output
|
||||
so, sys.stdout = sys.stdout, StringIO()
|
||||
|
||||
extraction.EXTRACTOR = classifier.load(EXTRACTOR_FILENAME,
|
||||
EXTRACTOR_DATA)
|
||||
sys.stdout = so
|
||||
except Exception, e:
|
||||
raise Exception(
|
||||
"Failed initializing signature parsing with classifiers", e)
|
||||
extraction.EXTRACTOR = classifier.load(EXTRACTOR_FILENAME,
|
||||
EXTRACTOR_DATA)
|
||||
|
||||
@@ -1,14 +1,15 @@
|
||||
from __future__ import absolute_import
|
||||
|
||||
import logging
|
||||
|
||||
import regex as re
|
||||
|
||||
from talon.utils import get_delimiter
|
||||
from talon.signature.constants import (SIGNATURE_MAX_LINES,
|
||||
TOO_LONG_SIGNATURE_LINE)
|
||||
from talon.utils import get_delimiter
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
# regex to fetch signature based on common signature words
|
||||
RE_SIGNATURE = re.compile(r'''
|
||||
(
|
||||
@@ -27,7 +28,6 @@ RE_SIGNATURE = re.compile(r'''
|
||||
)
|
||||
''', re.I | re.X | re.M | re.S)
|
||||
|
||||
|
||||
# signatures appended by phone email clients
|
||||
RE_PHONE_SIGNATURE = re.compile(r'''
|
||||
(
|
||||
@@ -44,7 +44,6 @@ RE_PHONE_SIGNATURE = re.compile(r'''
|
||||
)
|
||||
''', re.I | re.X | re.M | re.S)
|
||||
|
||||
|
||||
# see _mark_candidate_indexes() for details
|
||||
# c - could be signature line
|
||||
# d - line starts with dashes (could be signature or list item)
|
||||
@@ -111,7 +110,7 @@ def extract_signature(msg_body):
|
||||
|
||||
return (stripped_body.strip(),
|
||||
signature.strip())
|
||||
except Exception, e:
|
||||
except Exception:
|
||||
log.exception('ERROR extracting signature')
|
||||
return (msg_body, None)
|
||||
|
||||
@@ -162,7 +161,7 @@ def _mark_candidate_indexes(lines, candidate):
|
||||
'cdc'
|
||||
"""
|
||||
# at first consider everything to be potential signature lines
|
||||
markers = bytearray('c'*len(candidate))
|
||||
markers = list('c' * len(candidate))
|
||||
|
||||
# mark lines starting from bottom up
|
||||
for i, line_idx in reversed(list(enumerate(candidate))):
|
||||
@@ -173,7 +172,7 @@ def _mark_candidate_indexes(lines, candidate):
|
||||
if line.startswith('-') and line.strip("-"):
|
||||
markers[i] = 'd'
|
||||
|
||||
return markers
|
||||
return "".join(markers)
|
||||
|
||||
|
||||
def _process_marked_candidate_indexes(candidate, markers):
|
||||
|
||||
1
talon/signature/data/__init__.py
Normal file
1
talon/signature/data/__init__.py
Normal file
@@ -0,0 +1 @@
|
||||
|
||||
Binary file not shown.
BIN
talon/signature/data/classifier_01.npy
Normal file
BIN
talon/signature/data/classifier_01.npy
Normal file
Binary file not shown.
BIN
talon/signature/data/classifier_02.npy
Normal file
BIN
talon/signature/data/classifier_02.npy
Normal file
Binary file not shown.
BIN
talon/signature/data/classifier_03.npy
Normal file
BIN
talon/signature/data/classifier_03.npy
Normal file
Binary file not shown.
BIN
talon/signature/data/classifier_04.npy
Normal file
BIN
talon/signature/data/classifier_04.npy
Normal file
Binary file not shown.
BIN
talon/signature/data/classifier_05.npy
Normal file
BIN
talon/signature/data/classifier_05.npy
Normal file
Binary file not shown.
@@ -1,19 +1,15 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
import os
|
||||
from __future__ import absolute_import
|
||||
|
||||
import logging
|
||||
|
||||
import numpy
|
||||
import regex as re
|
||||
from PyML import SparseDataSet
|
||||
|
||||
from talon.constants import RE_DELIMITER
|
||||
from talon.signature.constants import (SIGNATURE_MAX_LINES,
|
||||
TOO_LONG_SIGNATURE_LINE)
|
||||
from talon.signature.learning.featurespace import features, build_pattern
|
||||
from talon.utils import get_delimiter
|
||||
from talon.signature.bruteforce import get_signature_candidate
|
||||
from talon.signature.learning.featurespace import features, build_pattern
|
||||
from talon.signature.learning.helpers import has_signature
|
||||
|
||||
from talon.utils import get_delimiter
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
@@ -36,8 +32,8 @@ RE_REVERSE_SIGNATURE = re.compile(r'''
|
||||
|
||||
def is_signature_line(line, sender, classifier):
|
||||
'''Checks if the line belongs to signature. Returns True or False.'''
|
||||
data = SparseDataSet([build_pattern(line, features(sender))])
|
||||
return classifier.decisionFunc(data, 0) > 0
|
||||
data = numpy.array(build_pattern(line, features(sender))).reshape(1, -1)
|
||||
return classifier.predict(data) > 0
|
||||
|
||||
|
||||
def extract(body, sender):
|
||||
@@ -61,7 +57,7 @@ def extract(body, sender):
|
||||
text = delimiter.join(text)
|
||||
if text.strip():
|
||||
return (text, delimiter.join(signature))
|
||||
except Exception, e:
|
||||
except Exception as e:
|
||||
log.exception('ERROR when extracting signature with classifiers')
|
||||
|
||||
return (body, None)
|
||||
@@ -84,7 +80,7 @@ def _mark_lines(lines, sender):
|
||||
candidate = get_signature_candidate(lines)
|
||||
|
||||
# at first consider everything to be text no signature
|
||||
markers = bytearray('t'*len(lines))
|
||||
markers = list('t' * len(lines))
|
||||
|
||||
# mark lines starting from bottom up
|
||||
# mark only lines that belong to candidate
|
||||
@@ -99,7 +95,7 @@ def _mark_lines(lines, sender):
|
||||
elif is_signature_line(line, sender, EXTRACTOR):
|
||||
markers[j] = 's'
|
||||
|
||||
return markers
|
||||
return "".join(markers)
|
||||
|
||||
|
||||
def _process_marked_lines(lines, markers):
|
||||
@@ -114,3 +110,4 @@ def _process_marked_lines(lines, markers):
|
||||
return (lines[:-signature.end()], lines[-signature.end():])
|
||||
|
||||
return (lines, None)
|
||||
|
||||
|
||||
@@ -5,32 +5,65 @@ The classifier could be used to detect if a certain line of the message
|
||||
body belongs to the signature.
|
||||
"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
from __future__ import absolute_import
|
||||
|
||||
from PyML import SparseDataSet, SVM
|
||||
from numpy import genfromtxt
|
||||
import joblib
|
||||
from sklearn.svm import LinearSVC
|
||||
|
||||
|
||||
def init():
|
||||
'''Inits classifier with optimal options.'''
|
||||
return SVM(C=10, optimization='liblinear')
|
||||
"""Inits classifier with optimal options."""
|
||||
return LinearSVC(C=10.0)
|
||||
|
||||
|
||||
def train(classifier, train_data_filename, save_classifier_filename=None):
|
||||
'''Trains and saves classifier so that it could be easily loaded later.'''
|
||||
data = SparseDataSet(train_data_filename, labelsColumn=-1)
|
||||
classifier.train(data)
|
||||
"""Trains and saves classifier so that it could be easily loaded later."""
|
||||
file_data = genfromtxt(train_data_filename, delimiter=",")
|
||||
train_data, labels = file_data[:, :-1], file_data[:, -1]
|
||||
classifier.fit(train_data, labels)
|
||||
|
||||
if save_classifier_filename:
|
||||
classifier.save(save_classifier_filename)
|
||||
joblib.dump(classifier, save_classifier_filename)
|
||||
return classifier
|
||||
|
||||
|
||||
def load(saved_classifier_filename, train_data_filename):
|
||||
"""Loads saved classifier.
|
||||
"""Loads saved classifier. """
|
||||
try:
|
||||
return joblib.load(saved_classifier_filename)
|
||||
except Exception:
|
||||
import sys
|
||||
if sys.version_info > (3, 0):
|
||||
return load_compat(saved_classifier_filename)
|
||||
|
||||
Classifier should be loaded with the same data it was trained against
|
||||
"""
|
||||
train_data = SparseDataSet(train_data_filename, labelsColumn=-1)
|
||||
classifier = init()
|
||||
classifier.load(saved_classifier_filename, train_data)
|
||||
return classifier
|
||||
raise
|
||||
|
||||
|
||||
def load_compat(saved_classifier_filename):
|
||||
import os
|
||||
import pickle
|
||||
import tempfile
|
||||
|
||||
# we need to switch to the data path to properly load the related _xx.npy files
|
||||
cwd = os.getcwd()
|
||||
os.chdir(os.path.dirname(saved_classifier_filename))
|
||||
|
||||
# convert encoding using pick.load and write to temp file which we'll tell joblib to use
|
||||
pickle_file = open(saved_classifier_filename, 'rb')
|
||||
classifier = pickle.load(pickle_file, encoding='latin1')
|
||||
|
||||
try:
|
||||
# save our conversion if permissions allow
|
||||
joblib.dump(classifier, saved_classifier_filename)
|
||||
except Exception:
|
||||
# can't write to classifier, use a temp file
|
||||
tmp = tempfile.SpooledTemporaryFile()
|
||||
joblib.dump(classifier, tmp)
|
||||
saved_classifier_filename = tmp
|
||||
|
||||
# important, use joblib.load before switching back to original cwd
|
||||
jb_classifier = joblib.load(saved_classifier_filename)
|
||||
os.chdir(cwd)
|
||||
|
||||
return jb_classifier
|
||||
|
||||
@@ -16,13 +16,16 @@ suffix and the corresponding sender file has the same name except for the
|
||||
suffix which should be `_sender`.
|
||||
"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
|
||||
import os
|
||||
|
||||
import regex as re
|
||||
from six.moves import range
|
||||
|
||||
from talon.signature.constants import SIGNATURE_MAX_LINES
|
||||
from talon.signature.learning.featurespace import build_pattern, features
|
||||
|
||||
|
||||
SENDER_SUFFIX = '_sender'
|
||||
BODY_SUFFIX = '_body'
|
||||
|
||||
@@ -55,9 +58,14 @@ def parse_msg_sender(filename, sender_known=True):
|
||||
algorithm:
|
||||
>>> parse_msg_sender(filename, False)
|
||||
"""
|
||||
import sys
|
||||
kwargs = {}
|
||||
if sys.version_info > (3, 0):
|
||||
kwargs["encoding"] = "utf8"
|
||||
|
||||
sender, msg = None, None
|
||||
if os.path.isfile(filename) and not is_sender_filename(filename):
|
||||
with open(filename) as f:
|
||||
with open(filename, **kwargs) as f:
|
||||
msg = f.read()
|
||||
sender = u''
|
||||
if sender_known:
|
||||
@@ -144,8 +152,8 @@ def build_extraction_dataset(folder, dataset_filename,
|
||||
if not sender or not msg:
|
||||
continue
|
||||
lines = msg.splitlines()
|
||||
for i in xrange(1, min(SIGNATURE_MAX_LINES,
|
||||
len(lines)) + 1):
|
||||
for i in range(1, min(SIGNATURE_MAX_LINES,
|
||||
len(lines)) + 1):
|
||||
line = lines[-i]
|
||||
label = -1
|
||||
if line[:len(SIGNATURE_ANNOTATION)] == \
|
||||
|
||||
@@ -7,9 +7,12 @@ The body and the message sender string are converted into unicode before
|
||||
applying features to them.
|
||||
"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
from talon.signature.constants import (SIGNATURE_MAX_LINES,
|
||||
TOO_LONG_SIGNATURE_LINE)
|
||||
from talon.signature.learning.helpers import *
|
||||
from six.moves import zip
|
||||
from functools import reduce
|
||||
|
||||
|
||||
def features(sender=''):
|
||||
|
||||
@@ -6,6 +6,7 @@
|
||||
|
||||
"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
import unicodedata
|
||||
import regex as re
|
||||
|
||||
@@ -16,8 +17,8 @@ from talon.signature.constants import SIGNATURE_MAX_LINES
|
||||
|
||||
rc = re.compile
|
||||
|
||||
RE_EMAIL = rc('@')
|
||||
RE_RELAX_PHONE = rc('.*(\(? ?[\d]{2,3} ?\)?.{,3}){2,}')
|
||||
RE_EMAIL = rc('\S@\S')
|
||||
RE_RELAX_PHONE = rc('(\(? ?[\d]{2,3} ?\)?.{,3}?){2,}')
|
||||
RE_URL = rc(r'''https?://|www\.[\S]+\.[\S]''')
|
||||
|
||||
# Taken from:
|
||||
@@ -40,14 +41,6 @@ RE_SIGNATURE_WORDS = rc(('(T|t)hank.*,|(B|b)est|(R|r)egards|'
|
||||
# Line contains a pattern like Vitor R. Carvalho or William W. Cohen.
|
||||
RE_NAME = rc('[A-Z][a-z]+\s\s?[A-Z][\.]?\s\s?[A-Z][a-z]+')
|
||||
|
||||
# Pattern to match if e.g. 'Sender:' header field has sender names.
|
||||
SENDER_WITH_NAME_PATTERN = '([\s]*[\S]+,?)+[\s]*<.*>.*'
|
||||
RE_SENDER_WITH_NAME = rc(SENDER_WITH_NAME_PATTERN)
|
||||
|
||||
# Reply line clue line endings, as in regular expression:
|
||||
# " wrote:$" or " writes:$"
|
||||
RE_CLUE_LINE_END = rc('.*(W|w)rotes?:$')
|
||||
|
||||
INVALID_WORD_START = rc('\(|\+|[\d]')
|
||||
|
||||
BAD_SENDER_NAMES = [
|
||||
@@ -128,7 +121,7 @@ def contains_sender_names(sender):
|
||||
names = names or sender
|
||||
if names != '':
|
||||
return binary_regex_search(re.compile(names))
|
||||
return lambda s: False
|
||||
return lambda s: 0
|
||||
|
||||
|
||||
def extract_names(sender):
|
||||
@@ -142,7 +135,7 @@ def extract_names(sender):
|
||||
>>> extract_names('')
|
||||
[]
|
||||
"""
|
||||
sender = to_unicode(sender)
|
||||
sender = to_unicode(sender, precise=True)
|
||||
# Remove non-alphabetical characters
|
||||
sender = "".join([char if char.isalpha() else ' ' for char in sender])
|
||||
# Remove too short words and words from "black" list i.e.
|
||||
@@ -169,7 +162,7 @@ def categories_percent(s, categories):
|
||||
50.0
|
||||
'''
|
||||
count = 0
|
||||
s = to_unicode(s)
|
||||
s = to_unicode(s, precise=True)
|
||||
for c in s:
|
||||
if unicodedata.category(c) in categories:
|
||||
count += 1
|
||||
@@ -189,15 +182,16 @@ def punctuation_percent(s):
|
||||
|
||||
def capitalized_words_percent(s):
|
||||
'''Returns capitalized words percent.'''
|
||||
s = to_unicode(s)
|
||||
s = to_unicode(s, precise=True)
|
||||
words = re.split('\s', s)
|
||||
words = [w for w in words if w.strip()]
|
||||
words = [w for w in words if len(w) > 2]
|
||||
capitalized_words_counter = 0
|
||||
valid_words_counter = 0
|
||||
for word in words:
|
||||
if not INVALID_WORD_START.match(word):
|
||||
valid_words_counter += 1
|
||||
if word[0].isupper():
|
||||
if word[0].isupper() and not word[1].isupper():
|
||||
capitalized_words_counter += 1
|
||||
if valid_words_counter > 0 and len(words) > 1:
|
||||
return 100 * float(capitalized_words_counter) / valid_words_counter
|
||||
|
||||
206
talon/utils.py
206
talon/utils.py
@@ -1,14 +1,20 @@
|
||||
# coding:utf-8
|
||||
|
||||
import logging
|
||||
from __future__ import absolute_import
|
||||
|
||||
from random import shuffle
|
||||
|
||||
import cchardet
|
||||
import chardet
|
||||
import html5lib
|
||||
import regex as re
|
||||
import six
|
||||
from lxml.cssselect import CSSSelector
|
||||
from lxml.html import html5parser
|
||||
|
||||
from talon.constants import RE_DELIMITER
|
||||
|
||||
|
||||
log = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def safe_format(format_string, *args, **kwargs):
|
||||
"""
|
||||
Helper: formats string with any combination of bytestrings/unicode
|
||||
@@ -25,7 +31,7 @@ def safe_format(format_string, *args, **kwargs):
|
||||
except (UnicodeEncodeError, UnicodeDecodeError):
|
||||
format_string = to_utf8(format_string)
|
||||
args = [to_utf8(p) for p in args]
|
||||
kwargs = {k: to_utf8(v) for k, v in kwargs.iteritems()}
|
||||
kwargs = {k: to_utf8(v) for k, v in six.iteritems(kwargs)}
|
||||
return format_string.format(*args, **kwargs)
|
||||
|
||||
# ignore other errors
|
||||
@@ -42,19 +48,51 @@ def to_unicode(str_or_unicode, precise=False):
|
||||
u'привет'
|
||||
If `precise` flag is True, tries to guess the correct encoding first.
|
||||
"""
|
||||
encoding = detect_encoding(str_or_unicode) if precise else 'utf-8'
|
||||
if isinstance(str_or_unicode, str):
|
||||
return unicode(str_or_unicode, encoding, 'replace')
|
||||
if not isinstance(str_or_unicode, six.text_type):
|
||||
encoding = quick_detect_encoding(str_or_unicode) if precise else 'utf-8'
|
||||
return six.text_type(str_or_unicode, encoding, 'replace')
|
||||
return str_or_unicode
|
||||
|
||||
|
||||
def detect_encoding(string):
|
||||
"""
|
||||
Tries to detect the encoding of the passed string.
|
||||
|
||||
Defaults to UTF-8.
|
||||
"""
|
||||
assert isinstance(string, bytes)
|
||||
try:
|
||||
detected = chardet.detect(string)
|
||||
if detected:
|
||||
return detected.get('encoding') or 'utf-8'
|
||||
except Exception as e:
|
||||
pass
|
||||
return 'utf-8'
|
||||
|
||||
|
||||
def quick_detect_encoding(string):
|
||||
"""
|
||||
Tries to detect the encoding of the passed string.
|
||||
|
||||
Uses cchardet. Fallbacks to detect_encoding.
|
||||
"""
|
||||
assert isinstance(string, bytes)
|
||||
try:
|
||||
detected = cchardet.detect(string)
|
||||
if detected:
|
||||
return detected.get('encoding') or detect_encoding(string)
|
||||
except Exception as e:
|
||||
pass
|
||||
return detect_encoding(string)
|
||||
|
||||
|
||||
def to_utf8(str_or_unicode):
|
||||
"""
|
||||
Safely returns a UTF-8 version of a given string
|
||||
>>> utils.to_utf8(u'hi')
|
||||
'hi'
|
||||
"""
|
||||
if isinstance(str_or_unicode, unicode):
|
||||
if not isinstance(str_or_unicode, six.text_type):
|
||||
return str_or_unicode.encode("utf-8", "ignore")
|
||||
return str(str_or_unicode)
|
||||
|
||||
@@ -74,3 +112,153 @@ def get_delimiter(msg_body):
|
||||
delimiter = '\n'
|
||||
|
||||
return delimiter
|
||||
|
||||
|
||||
def html_tree_to_text(tree):
|
||||
for style in CSSSelector('style')(tree):
|
||||
style.getparent().remove(style)
|
||||
|
||||
for c in tree.xpath('//comment()'):
|
||||
parent = c.getparent()
|
||||
|
||||
# comment with no parent does not impact produced text
|
||||
if parent is None:
|
||||
continue
|
||||
|
||||
parent.remove(c)
|
||||
|
||||
text = ""
|
||||
for el in tree.iter():
|
||||
el_text = (el.text or '') + (el.tail or '')
|
||||
if len(el_text) > 1:
|
||||
if el.tag in _BLOCKTAGS + _HARDBREAKS:
|
||||
text += "\n"
|
||||
if el.tag == 'li':
|
||||
text += " * "
|
||||
text += el_text.strip() + " "
|
||||
|
||||
# add href to the output
|
||||
href = el.attrib.get('href')
|
||||
if href:
|
||||
text += "(%s) " % href
|
||||
|
||||
if (el.tag in _HARDBREAKS and text and
|
||||
not text.endswith("\n") and not el_text):
|
||||
text += "\n"
|
||||
|
||||
retval = _rm_excessive_newlines(text)
|
||||
return _encode_utf8(retval)
|
||||
|
||||
|
||||
def html_to_text(string):
|
||||
"""
|
||||
Dead-simple HTML-to-text converter:
|
||||
>>> html_to_text("one<br>two<br>three")
|
||||
>>> "one\ntwo\nthree"
|
||||
|
||||
NOTES:
|
||||
1. the string is expected to contain UTF-8 encoded HTML!
|
||||
2. returns utf-8 encoded str (not unicode)
|
||||
3. if html can't be parsed returns None
|
||||
"""
|
||||
if isinstance(string, six.text_type):
|
||||
string = string.encode('utf8')
|
||||
|
||||
s = _prepend_utf8_declaration(string)
|
||||
s = s.replace(b"\n", b"")
|
||||
tree = html_fromstring(s)
|
||||
|
||||
if tree is None:
|
||||
return None
|
||||
|
||||
return html_tree_to_text(tree)
|
||||
|
||||
|
||||
def html_fromstring(s):
|
||||
"""Parse html tree from string. Return None if the string can't be parsed.
|
||||
"""
|
||||
if isinstance(s, six.text_type):
|
||||
s = s.encode('utf8')
|
||||
try:
|
||||
if html_too_big(s):
|
||||
return None
|
||||
|
||||
return html5parser.fromstring(s, parser=_html5lib_parser())
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
|
||||
def html_document_fromstring(s):
|
||||
"""Parse html tree from string. Return None if the string can't be parsed.
|
||||
"""
|
||||
if isinstance(s, six.text_type):
|
||||
s = s.encode('utf8')
|
||||
try:
|
||||
if html_too_big(s):
|
||||
return None
|
||||
|
||||
return html5parser.document_fromstring(s, parser=_html5lib_parser())
|
||||
except Exception:
|
||||
pass
|
||||
|
||||
|
||||
def cssselect(expr, tree):
|
||||
return CSSSelector(expr)(tree)
|
||||
|
||||
|
||||
def html_too_big(s):
|
||||
if isinstance(s, six.text_type):
|
||||
s = s.encode('utf8')
|
||||
return s.count(b'<') > _MAX_TAGS_COUNT
|
||||
|
||||
|
||||
def _contains_charset_spec(s):
|
||||
"""Return True if the first 4KB contain charset spec
|
||||
"""
|
||||
return s.lower().find(b'html; charset=', 0, 4096) != -1
|
||||
|
||||
|
||||
def _prepend_utf8_declaration(s):
|
||||
"""Prepend 'utf-8' encoding declaration if the first 4KB don't have any
|
||||
"""
|
||||
return s if _contains_charset_spec(s) else _UTF8_DECLARATION + s
|
||||
|
||||
|
||||
def _rm_excessive_newlines(s):
|
||||
"""Remove excessive newlines that often happen due to tons of divs
|
||||
"""
|
||||
return _RE_EXCESSIVE_NEWLINES.sub("\n\n", s).strip()
|
||||
|
||||
|
||||
def _encode_utf8(s):
|
||||
"""Encode in 'utf-8' if unicode
|
||||
"""
|
||||
return s.encode('utf-8') if isinstance(s, six.text_type) else s
|
||||
|
||||
|
||||
def _html5lib_parser():
|
||||
"""
|
||||
html5lib is a pure-python library that conforms to the WHATWG HTML spec
|
||||
and is not vulnarable to certain attacks common for XML libraries
|
||||
"""
|
||||
return html5lib.HTMLParser(
|
||||
# build lxml tree
|
||||
html5lib.treebuilders.getTreeBuilder("lxml"),
|
||||
# remove namespace value from inside lxml.html.html5paser element tag
|
||||
# otherwise it yields something like "{http://www.w3.org/1999/xhtml}div"
|
||||
# instead of "div", throwing the algo off
|
||||
namespaceHTMLElements=False
|
||||
)
|
||||
|
||||
|
||||
_UTF8_DECLARATION = (b'<meta http-equiv="Content-Type" content="text/html;'
|
||||
b'charset=utf-8">')
|
||||
|
||||
_BLOCKTAGS = ['div', 'p', 'ul', 'li', 'h1', 'h2', 'h3']
|
||||
_HARDBREAKS = ['br', 'hr', 'tr']
|
||||
|
||||
_RE_EXCESSIVE_NEWLINES = re.compile("\n{2,10}")
|
||||
|
||||
# an extensive research shows that exceeding this limit
|
||||
# might lead to excessive processing time
|
||||
_MAX_TAGS_COUNT = 419
|
||||
|
||||
3
test-requirements.txt
Normal file
3
test-requirements.txt
Normal file
@@ -0,0 +1,3 @@
|
||||
coverage
|
||||
mock
|
||||
nose>=1.2.1
|
||||
@@ -1,3 +1,4 @@
|
||||
from __future__ import absolute_import
|
||||
from nose.tools import *
|
||||
from mock import *
|
||||
|
||||
|
||||
1
tests/fixtures/html_replies/hotmail.html
vendored
1
tests/fixtures/html_replies/hotmail.html
vendored
@@ -1,3 +1,4 @@
|
||||
<?xml version="1.0" encoding="UTF-8"?>
|
||||
<html>
|
||||
<head>
|
||||
<style><!--
|
||||
|
||||
87
tests/fixtures/html_replies/ms_outlook_2010.html
vendored
Normal file
87
tests/fixtures/html_replies/ms_outlook_2010.html
vendored
Normal file
@@ -0,0 +1,87 @@
|
||||
<html>
|
||||
<head>
|
||||
<meta http-equiv="Content-Type" content="text/html; charset=iso-2022-jp">
|
||||
<meta name="Generator" content="Microsoft Word 14 (filtered medium)">
|
||||
<style><!--
|
||||
/* Font Definitions */
|
||||
@font-face
|
||||
{font-family:Calibri;
|
||||
panose-1:2 15 5 2 2 2 4 3 2 4;}
|
||||
@font-face
|
||||
{font-family:Tahoma;
|
||||
panose-1:2 11 6 4 3 5 4 4 2 4;}
|
||||
/* Style Definitions */
|
||||
p.MsoNormal, li.MsoNormal, div.MsoNormal
|
||||
{margin:0in;
|
||||
margin-bottom:.0001pt;
|
||||
font-size:12.0pt;
|
||||
font-family:"Times New Roman","serif";}
|
||||
h3
|
||||
{mso-style-priority:9;
|
||||
mso-style-link:"Heading 3 Char";
|
||||
mso-margin-top-alt:auto;
|
||||
margin-right:0in;
|
||||
mso-margin-bottom-alt:auto;
|
||||
margin-left:0in;
|
||||
font-size:13.5pt;
|
||||
font-family:"Times New Roman","serif";
|
||||
font-weight:bold;}
|
||||
a:link, span.MsoHyperlink
|
||||
{mso-style-priority:99;
|
||||
color:blue;
|
||||
text-decoration:underline;}
|
||||
a:visited, span.MsoHyperlinkFollowed
|
||||
{mso-style-priority:99;
|
||||
color:purple;
|
||||
text-decoration:underline;}
|
||||
p
|
||||
{mso-style-priority:99;
|
||||
mso-margin-top-alt:auto;
|
||||
margin-right:0in;
|
||||
mso-margin-bottom-alt:auto;
|
||||
margin-left:0in;
|
||||
font-size:12.0pt;
|
||||
font-family:"Times New Roman","serif";}
|
||||
span.Heading3Char
|
||||
{mso-style-name:"Heading 3 Char";
|
||||
mso-style-priority:9;
|
||||
mso-style-link:"Heading 3";
|
||||
font-family:"Cambria","serif";
|
||||
color:#4F81BD;
|
||||
font-weight:bold;}
|
||||
span.EmailStyle19
|
||||
{mso-style-type:personal-reply;
|
||||
font-family:"Calibri","sans-serif";
|
||||
color:#1F497D;}
|
||||
.MsoChpDefault
|
||||
{mso-style-type:export-only;
|
||||
font-family:"Calibri","sans-serif";}
|
||||
@page WordSection1
|
||||
{size:8.5in 11.0in;
|
||||
margin:1.0in 1.0in 1.0in 1.0in;}
|
||||
div.WordSection1
|
||||
{page:WordSection1;}
|
||||
--></style><!--[if gte mso 9]><xml>
|
||||
<o:shapedefaults v:ext="edit" spidmax="1026" />
|
||||
</xml><![endif]--><!--[if gte mso 9]><xml>
|
||||
<o:shapelayout v:ext="edit">
|
||||
<o:idmap v:ext="edit" data="1" />
|
||||
</o:shapelayout></xml><![endif]-->
|
||||
</head>
|
||||
<body lang="EN-US" link="blue" vlink="purple">
|
||||
<div class="WordSection1">
|
||||
<p class="MsoNormal"><span style="font-size:11.0pt;font-family:"Calibri","sans-serif";color:#1F497D">Hi. I am fine.<o:p></o:p></span></p>
|
||||
<p class="MsoNormal"><span style="font-size:11.0pt;font-family:"Calibri","sans-serif";color:#1F497D">Thanks,<o:p></o:p></span></p>
|
||||
<p class="MsoNormal"><span style="font-size:11.0pt;font-family:"Calibri","sans-serif";color:#1F497D">Alex<o:p></o:p></span></p>
|
||||
<p class="MsoNormal"><b><span style="font-size:10.0pt;font-family:"Tahoma","sans-serif"">From:</span></b><span style="font-size:10.0pt;font-family:"Tahoma","sans-serif""> Foo [mailto:foo@bar.com]
|
||||
<b>On Behalf Of </b>baz@bar.com<br>
|
||||
<b>Sent:</b> Monday, January 01, 2000 12:00 AM<br>
|
||||
<b>To:</b> john@bar.com<br>
|
||||
<b>Cc:</b> jane@bar.io<br>
|
||||
<b>Subject:</b> Conversation<o:p></o:p></span></p>
|
||||
<p class="MsoNormal"><o:p> </o:p></p>
|
||||
<p>Hello! How are you?<o:p></o:p></p>
|
||||
<p class="MsoNormal"><o:p> </o:p></p>
|
||||
</div>
|
||||
</body>
|
||||
</html>
|
||||
19
tests/fixtures/standard_replies/apple_mail_2.eml
vendored
Normal file
19
tests/fixtures/standard_replies/apple_mail_2.eml
vendored
Normal file
@@ -0,0 +1,19 @@
|
||||
Content-Type: text/plain;
|
||||
charset=us-ascii
|
||||
Mime-Version: 1.0 (Mac OS X Mail 8.2 \(2104\))
|
||||
Subject: Re: Hello there
|
||||
X-Universally-Unique-Identifier: 85B1075D-5841-46A9-8565-FCB287A93AC4
|
||||
From: Adam Renberg <adam@tictail.com>
|
||||
In-Reply-To: <CABzQGhkMXDxUt_tSVQcg=43aniUhtsVfCZVzu-PG0kwS_uzqMw@mail.gmail.com>
|
||||
Date: Sat, 22 Aug 2015 19:22:20 +0200
|
||||
Content-Transfer-Encoding: 7bit
|
||||
X-Smtp-Server: smtp.gmail.com:adam@tictail.com
|
||||
Message-Id: <68001B29-8EA4-444C-A894-0537D2CA5208@tictail.com>
|
||||
References: <CABzQGhkMXDxUt_tSVQcg=43aniUhtsVfCZVzu-PG0kwS_uzqMw@mail.gmail.com>
|
||||
To: Adam Renberg <tgwizard@gmail.com>
|
||||
|
||||
Hello
|
||||
> On 22 Aug 2015, at 19:21, Adam Renberg <tgwizard@gmail.com> wrote:
|
||||
>
|
||||
> Hi there!
|
||||
|
||||
4
tests/fixtures/standard_replies/iphone.eml
vendored
4
tests/fixtures/standard_replies/iphone.eml
vendored
@@ -9,11 +9,11 @@ To: bob <bob@example.com>
|
||||
Content-Transfer-Encoding: quoted-printable
|
||||
Mime-Version: 1.0 (1.0)
|
||||
|
||||
hello
|
||||
Hello
|
||||
|
||||
Sent from my iPhone
|
||||
|
||||
On Apr 3, 2012, at 4:19 PM, bob <bob@example.com> wr=
|
||||
ote:
|
||||
|
||||
> Hi
|
||||
> Hi
|
||||
|
||||
3
tests/fixtures/standard_replies/iphone_reply_text
vendored
Normal file
3
tests/fixtures/standard_replies/iphone_reply_text
vendored
Normal file
@@ -0,0 +1,3 @@
|
||||
Hello
|
||||
|
||||
Sent from my iPhone
|
||||
@@ -1,15 +1,14 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
|
||||
# noinspection PyUnresolvedReferences
|
||||
import re
|
||||
|
||||
from talon import quotations, utils as u
|
||||
from . import *
|
||||
from . fixtures import *
|
||||
|
||||
import regex as re
|
||||
from flanker import mime
|
||||
|
||||
from talon import quotations
|
||||
|
||||
import html2text
|
||||
|
||||
from .fixtures import *
|
||||
from lxml import html
|
||||
|
||||
RE_WHITESPACE = re.compile("\s")
|
||||
RE_DOUBLE_WHITESPACE = re.compile("\s")
|
||||
@@ -29,7 +28,7 @@ def test_quotation_splitter_inside_blockquote():
|
||||
|
||||
</blockquote>"""
|
||||
|
||||
eq_("<html><body><p>Reply</p></body></html>",
|
||||
eq_("<html><head></head><body>Reply</body></html>",
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
@@ -46,7 +45,25 @@ def test_quotation_splitter_outside_blockquote():
|
||||
</div>
|
||||
</blockquote>
|
||||
"""
|
||||
eq_("<html><body><p>Reply</p><div></div></body></html>",
|
||||
eq_("<html><head></head><body>Reply</body></html>",
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
def test_regular_blockquote():
|
||||
msg_body = """Reply
|
||||
<blockquote>Regular</blockquote>
|
||||
|
||||
<div>
|
||||
On 11-Apr-2011, at 6:54 PM, Bob <bob@example.com> wrote:
|
||||
</div>
|
||||
|
||||
<blockquote>
|
||||
<div>
|
||||
<blockquote>Nested</blockquote>
|
||||
</div>
|
||||
</blockquote>
|
||||
"""
|
||||
eq_("<html><head></head><body>Reply<blockquote>Regular</blockquote></body></html>",
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
@@ -69,6 +86,7 @@ Reply
|
||||
|
||||
reply = """
|
||||
<html>
|
||||
<head></head>
|
||||
<body>
|
||||
Reply
|
||||
|
||||
@@ -112,7 +130,30 @@ def test_gmail_quote():
|
||||
</div>
|
||||
</div>
|
||||
</div>"""
|
||||
eq_("<html><body><p>Reply</p></body></html>",
|
||||
eq_("<html><head></head><body>Reply</body></html>",
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
def test_gmail_quote_compact():
|
||||
msg_body = 'Reply' \
|
||||
'<div class="gmail_quote">' \
|
||||
'<div class="gmail_quote">On 11-Apr-2011, at 6:54 PM, Bob <bob@example.com> wrote:' \
|
||||
'<div>Test</div>' \
|
||||
'</div>' \
|
||||
'</div>'
|
||||
eq_("<html><head></head><body>Reply</body></html>",
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
def test_gmail_quote_blockquote():
|
||||
msg_body = """Message
|
||||
<blockquote class="gmail_quote">
|
||||
<div class="gmail_default">
|
||||
My name is William Shakespeare.
|
||||
<br/>
|
||||
</div>
|
||||
</blockquote>"""
|
||||
eq_(RE_WHITESPACE.sub('', msg_body),
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
@@ -123,11 +164,11 @@ def test_unicode_in_reply():
|
||||
<br>
|
||||
</div>
|
||||
|
||||
<blockquote class="gmail_quote">
|
||||
<blockquote>
|
||||
Quote
|
||||
</blockquote>""".encode("utf-8")
|
||||
|
||||
eq_("<html><body><p>Reply  Text<br></p><div><br></div>"
|
||||
eq_("<html><head></head><body>Reply  Text<br><div><br></div>"
|
||||
"</body></html>",
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
@@ -153,6 +194,7 @@ def test_blockquote_disclaimer():
|
||||
|
||||
stripped_html = """
|
||||
<html>
|
||||
<head></head>
|
||||
<body>
|
||||
<div>
|
||||
<div>
|
||||
@@ -184,7 +226,7 @@ def test_date_block():
|
||||
</div>
|
||||
</div>
|
||||
"""
|
||||
eq_('<html><body><div>message<br></div></body></html>',
|
||||
eq_('<html><head></head><body><div>message<br></div></body></html>',
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
@@ -201,7 +243,7 @@ Subject: You Have New Mail From Mary!<br><br>
|
||||
text
|
||||
</div></div>
|
||||
"""
|
||||
eq_('<html><body><div>message<br></div></body></html>',
|
||||
eq_('<html><head></head><body><div>message<br></div></body></html>',
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
@@ -219,51 +261,63 @@ def test_reply_shares_div_with_from_block():
|
||||
|
||||
</div>
|
||||
</body>'''
|
||||
eq_('<html><body><div>Blah<br><br></div></body></html>',
|
||||
eq_('<html><head></head><body><div>Blah<br><br></div></body></html>',
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
def test_reply_quotations_share_block():
|
||||
msg = mime.from_string(REPLY_QUOTATIONS_SHARE_BLOCK)
|
||||
html_part = list(msg.walk())[1]
|
||||
assert html_part.content_type == 'text/html'
|
||||
stripped_html = quotations.extract_from_html(html_part.body)
|
||||
stripped_html = quotations.extract_from_plain(REPLY_QUOTATIONS_SHARE_BLOCK)
|
||||
ok_(stripped_html)
|
||||
ok_('From' not in stripped_html)
|
||||
|
||||
|
||||
def test_OLK_SRC_BODY_SECTION_stripped():
|
||||
eq_('<html><body><div>Reply</div></body></html>',
|
||||
eq_('<html><head></head><body><div>Reply</div></body></html>',
|
||||
RE_WHITESPACE.sub(
|
||||
'', quotations.extract_from_html(OLK_SRC_BODY_SECTION)))
|
||||
|
||||
|
||||
def test_reply_separated_by_hr():
|
||||
eq_('<html><body><div>Hi<div>there</div></div></body></html>',
|
||||
eq_('<html><head></head><body><div>Hi<div>there</div></div></body></html>',
|
||||
RE_WHITESPACE.sub(
|
||||
'', quotations.extract_from_html(REPLY_SEPARATED_BY_HR)))
|
||||
|
||||
|
||||
RE_REPLY = re.compile(r"^Hi\. I am fine\.\s*\n\s*Thanks,\s*\n\s*Alex\s*$")
|
||||
def test_from_block_and_quotations_in_separate_divs():
|
||||
msg_body = '''
|
||||
Reply
|
||||
<div>
|
||||
<hr/>
|
||||
<div>
|
||||
<font>
|
||||
<b>From: bob@example.com</b>
|
||||
<b>Date: Thu, 24 Mar 2016 08:07:12 -0700</b>
|
||||
</font>
|
||||
</div>
|
||||
<div>
|
||||
Quoted message
|
||||
</div>
|
||||
</div>
|
||||
'''
|
||||
eq_('<html><head></head><body>Reply<div><hr></div></body></html>',
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
def extract_reply_and_check(filename):
|
||||
f = open(filename)
|
||||
import sys
|
||||
kwargs = {}
|
||||
if sys.version_info > (3, 0):
|
||||
kwargs["encoding"] = "utf8"
|
||||
|
||||
msg_body = f.read().decode("utf-8")
|
||||
f = open(filename, **kwargs)
|
||||
|
||||
msg_body = f.read()
|
||||
reply = quotations.extract_from_html(msg_body)
|
||||
plain_reply = u.html_to_text(reply)
|
||||
plain_reply = plain_reply.decode('utf8')
|
||||
|
||||
h = html2text.HTML2Text()
|
||||
h.body_width = 0
|
||||
plain_reply = h.handle(reply)
|
||||
|
||||
#remove spaces
|
||||
plain_reply = plain_reply.replace(u'\xa0', u' ')
|
||||
|
||||
if RE_REPLY.match(plain_reply):
|
||||
eq_(1, 1)
|
||||
else:
|
||||
eq_("Hi. I am fine.\n\nThanks,\nAlex", plain_reply)
|
||||
eq_(RE_WHITESPACE.sub('', "Hi. I am fine.\n\nThanks,\nAlex"),
|
||||
RE_WHITESPACE.sub('', plain_reply))
|
||||
|
||||
|
||||
def test_gmail_reply():
|
||||
@@ -286,6 +340,10 @@ def test_ms_outlook_2007_reply():
|
||||
extract_reply_and_check("tests/fixtures/html_replies/ms_outlook_2007.html")
|
||||
|
||||
|
||||
def test_ms_outlook_2010_reply():
|
||||
extract_reply_and_check("tests/fixtures/html_replies/ms_outlook_2010.html")
|
||||
|
||||
|
||||
def test_thunderbird_reply():
|
||||
extract_reply_and_check("tests/fixtures/html_replies/thunderbird.html")
|
||||
|
||||
@@ -296,3 +354,94 @@ def test_windows_mail_reply():
|
||||
|
||||
def test_yandex_ru_reply():
|
||||
extract_reply_and_check("tests/fixtures/html_replies/yandex_ru.html")
|
||||
|
||||
|
||||
def test_CRLF():
|
||||
"""CR is not converted to ' '
|
||||
"""
|
||||
symbol = ' '
|
||||
extracted = quotations.extract_from_html('<html>\r\n</html>')
|
||||
assert_false(symbol in extracted)
|
||||
eq_('<html></html>', RE_WHITESPACE.sub('', extracted))
|
||||
|
||||
msg_body = """My
|
||||
reply
|
||||
<blockquote>
|
||||
|
||||
<div>
|
||||
On 11-Apr-2011, at 6:54 PM, Bob <bob@example.com> wrote:
|
||||
</div>
|
||||
|
||||
<div>
|
||||
Test
|
||||
</div>
|
||||
|
||||
</blockquote>"""
|
||||
msg_body = msg_body.replace('\n', '\r\n')
|
||||
extracted = quotations.extract_from_html(msg_body)
|
||||
assert_false(symbol in extracted)
|
||||
# Keep new lines otherwise "My reply" becomes one word - "Myreply"
|
||||
eq_("<html><head></head><body>My\nreply\n</body></html>", extracted)
|
||||
|
||||
|
||||
def test_gmail_forwarded_msg():
|
||||
msg_body = """<div dir="ltr"><br><div class="gmail_quote">---------- Forwarded message ----------<br>From: <b class="gmail_sendername">Bob</b> <span dir="ltr"><<a href="mailto:bob@example.com">bob@example.com</a>></span><br>Date: Fri, Feb 11, 2010 at 5:59 PM<br>Subject: Bob WFH today<br>To: Mary <<a href="mailto:mary@example.com">mary@example.com</a>><br><br><br><div dir="ltr">eom</div>
|
||||
</div><br></div>"""
|
||||
extracted = quotations.extract_from_html(msg_body)
|
||||
eq_(RE_WHITESPACE.sub('', msg_body), RE_WHITESPACE.sub('', extracted))
|
||||
|
||||
|
||||
@patch.object(u, '_MAX_TAGS_COUNT', 4)
|
||||
def test_too_large_html():
|
||||
msg_body = 'Reply' \
|
||||
'<div class="gmail_quote">' \
|
||||
'<div class="gmail_quote">On 11-Apr-2011, at 6:54 PM, Bob <bob@example.com> wrote:' \
|
||||
'<div>Test</div>' \
|
||||
'</div>' \
|
||||
'</div>'
|
||||
eq_(RE_WHITESPACE.sub('', msg_body),
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
def test_readable_html_empty():
|
||||
msg_body = """
|
||||
<blockquote>
|
||||
Reply
|
||||
<div>
|
||||
On 11-Apr-2011, at 6:54 PM, Bob <bob@example.com> wrote:
|
||||
</div>
|
||||
|
||||
<div>
|
||||
Test
|
||||
</div>
|
||||
|
||||
</blockquote>"""
|
||||
|
||||
eq_(RE_WHITESPACE.sub('', msg_body),
|
||||
RE_WHITESPACE.sub('', quotations.extract_from_html(msg_body)))
|
||||
|
||||
|
||||
@patch.object(quotations, 'html_document_fromstring', Mock(return_value=None))
|
||||
def test_bad_html():
|
||||
bad_html = "<html></html>"
|
||||
eq_(bad_html, quotations.extract_from_html(bad_html))
|
||||
|
||||
|
||||
def test_remove_namespaces():
|
||||
msg_body = """
|
||||
<html xmlns:o="urn:schemas-microsoft-com:office:office" xmlns="http://www.w3.org/TR/REC-html40">
|
||||
<body>
|
||||
<o:p>Dear Sir,</o:p>
|
||||
<o:p>Thank you for the email.</o:p>
|
||||
<blockquote>thing</blockquote>
|
||||
</body>
|
||||
</html>
|
||||
"""
|
||||
|
||||
rendered = quotations.extract_from_html(msg_body)
|
||||
|
||||
assert_true("<p>" in rendered)
|
||||
assert_true("xmlns" in rendered)
|
||||
|
||||
assert_true("<o:p>" not in rendered)
|
||||
assert_true("<xmlns:o>" not in rendered)
|
||||
|
||||
@@ -1,10 +1,9 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
from . import *
|
||||
from . fixtures import *
|
||||
|
||||
from flanker import mime
|
||||
|
||||
from talon import quotations
|
||||
|
||||
|
||||
|
||||
@@ -1,11 +1,8 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
from .. import *
|
||||
|
||||
import os
|
||||
|
||||
from flanker import mime
|
||||
|
||||
from talon.signature import bruteforce
|
||||
|
||||
|
||||
|
||||
@@ -1,15 +1,15 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from .. import *
|
||||
from __future__ import absolute_import
|
||||
|
||||
import os
|
||||
|
||||
from PyML import SparseDataSet
|
||||
from six.moves import range
|
||||
|
||||
from talon.signature.learning import dataset
|
||||
from talon import signature
|
||||
from talon.signature import bruteforce, extraction, extract
|
||||
from talon.signature import extraction as e
|
||||
from talon.signature import bruteforce
|
||||
from talon.signature.learning import dataset
|
||||
from .. import *
|
||||
|
||||
|
||||
def test_message_shorter_SIGNATURE_MAX_LINES():
|
||||
@@ -18,23 +18,28 @@ def test_message_shorter_SIGNATURE_MAX_LINES():
|
||||
|
||||
Thanks in advance,
|
||||
Bob"""
|
||||
text, extracted_signature = signature.extract(body, sender)
|
||||
text, extracted_signature = extract(body, sender)
|
||||
eq_('\n'.join(body.splitlines()[:2]), text)
|
||||
eq_('\n'.join(body.splitlines()[-2:]), extracted_signature)
|
||||
|
||||
|
||||
def test_messages_longer_SIGNATURE_MAX_LINES():
|
||||
import sys
|
||||
kwargs = {}
|
||||
if sys.version_info > (3, 0):
|
||||
kwargs["encoding"] = "utf8"
|
||||
|
||||
for filename in os.listdir(STRIPPED):
|
||||
filename = os.path.join(STRIPPED, filename)
|
||||
if not filename.endswith('_body'):
|
||||
continue
|
||||
sender, body = dataset.parse_msg_sender(filename)
|
||||
text, extracted_signature = signature.extract(body, sender)
|
||||
text, extracted_signature = extract(body, sender)
|
||||
extracted_signature = extracted_signature or ''
|
||||
with open(filename[:-len('body')] + 'signature') as ms:
|
||||
with open(filename[:-len('body')] + 'signature', **kwargs) as ms:
|
||||
msg_signature = ms.read()
|
||||
eq_(msg_signature.strip(), extracted_signature.strip())
|
||||
stripped_msg = body.strip()[:len(body.strip())-len(msg_signature)]
|
||||
stripped_msg = body.strip()[:len(body.strip()) - len(msg_signature)]
|
||||
eq_(stripped_msg.strip(), text.strip())
|
||||
|
||||
|
||||
@@ -47,7 +52,7 @@ Thanks in advance,
|
||||
some text which doesn't seem to be a signature at all
|
||||
Bob"""
|
||||
|
||||
text, extracted_signature = signature.extract(body, sender)
|
||||
text, extracted_signature = extract(body, sender)
|
||||
eq_('\n'.join(body.splitlines()[:2]), text)
|
||||
eq_('\n'.join(body.splitlines()[-3:]), extracted_signature)
|
||||
|
||||
@@ -60,7 +65,7 @@ Thanks in advance,
|
||||
some long text here which doesn't seem to be a signature at all
|
||||
Bob"""
|
||||
|
||||
text, extracted_signature = signature.extract(body, sender)
|
||||
text, extracted_signature = extract(body, sender)
|
||||
eq_('\n'.join(body.splitlines()[:-1]), text)
|
||||
eq_('Bob', extracted_signature)
|
||||
|
||||
@@ -68,13 +73,38 @@ Bob"""
|
||||
|
||||
some *long* text here which doesn't seem to be a signature at all
|
||||
"""
|
||||
((body, None), signature.extract(body, "david@example.com"))
|
||||
((body, None), extract(body, "david@example.com"))
|
||||
|
||||
|
||||
def test_basic():
|
||||
msg_body = 'Blah\r\n--\r\n\r\nSergey Obukhov'
|
||||
eq_(('Blah', '--\r\n\r\nSergey Obukhov'),
|
||||
signature.extract(msg_body, 'Sergey'))
|
||||
extract(msg_body, 'Sergey'))
|
||||
|
||||
|
||||
def test_capitalized():
|
||||
msg_body = """Hi Mary,
|
||||
|
||||
Do you still need a DJ for your wedding? I've included a video demo of one of our DJs available for your wedding date.
|
||||
|
||||
DJ Doe
|
||||
http://example.com
|
||||
Password: SUPERPASSWORD
|
||||
|
||||
Would you like to check out more?
|
||||
|
||||
|
||||
At your service,
|
||||
|
||||
John Smith
|
||||
Doe Inc
|
||||
555-531-7967"""
|
||||
|
||||
sig = """John Smith
|
||||
Doe Inc
|
||||
555-531-7967"""
|
||||
|
||||
eq_(sig, extract(msg_body, 'Doe')[1])
|
||||
|
||||
|
||||
def test_over_2_text_lines_after_signature():
|
||||
@@ -85,25 +115,25 @@ def test_over_2_text_lines_after_signature():
|
||||
2 non signature lines in the end
|
||||
It's not signature
|
||||
"""
|
||||
text, extracted_signature = signature.extract(body, "Bob")
|
||||
text, extracted_signature = extract(body, "Bob")
|
||||
eq_(extracted_signature, None)
|
||||
|
||||
|
||||
def test_no_signature():
|
||||
sender, body = "bob@foo.bar", "Hello"
|
||||
eq_((body, None), signature.extract(body, sender))
|
||||
eq_((body, None), extract(body, sender))
|
||||
|
||||
|
||||
def test_handles_unicode():
|
||||
sender, body = dataset.parse_msg_sender(UNICODE_MSG)
|
||||
text, extracted_signature = signature.extract(body, sender)
|
||||
text, extracted_signature = extract(body, sender)
|
||||
|
||||
|
||||
@patch.object(signature.extraction, 'has_signature')
|
||||
@patch.object(extraction, 'has_signature')
|
||||
def test_signature_extract_crash(has_signature):
|
||||
has_signature.side_effect = Exception('Bam!')
|
||||
msg_body = u'Blah\r\n--\r\n\r\nСергей'
|
||||
eq_((msg_body, None), signature.extract(msg_body, 'Сергей'))
|
||||
eq_((msg_body, None), extract(msg_body, 'Сергей'))
|
||||
|
||||
|
||||
def test_mark_lines():
|
||||
@@ -112,37 +142,37 @@ def test_mark_lines():
|
||||
# (starting from the bottom) because we don't count empty line
|
||||
eq_('ttset',
|
||||
e._mark_lines(['Bob Smith',
|
||||
'Bob Smith',
|
||||
'Bob Smith',
|
||||
'',
|
||||
'some text'], 'Bob Smith'))
|
||||
'Bob Smith',
|
||||
'Bob Smith',
|
||||
'',
|
||||
'some text'], 'Bob Smith'))
|
||||
|
||||
with patch.object(bruteforce, 'SIGNATURE_MAX_LINES', 3):
|
||||
# we don't analyse the 1st line because
|
||||
# signature cant start from the 1st line
|
||||
eq_('tset',
|
||||
e._mark_lines(['Bob Smith',
|
||||
'Bob Smith',
|
||||
'',
|
||||
'some text'], 'Bob Smith'))
|
||||
'Bob Smith',
|
||||
'',
|
||||
'some text'], 'Bob Smith'))
|
||||
|
||||
|
||||
def test_process_marked_lines():
|
||||
# no signature found
|
||||
eq_((range(5), None), e._process_marked_lines(range(5), 'telt'))
|
||||
eq_((list(range(5)), None), e._process_marked_lines(list(range(5)), 'telt'))
|
||||
|
||||
# signature in the middle of the text
|
||||
eq_((range(9), None), e._process_marked_lines(range(9), 'tesestelt'))
|
||||
eq_((list(range(9)), None), e._process_marked_lines(list(range(9)), 'tesestelt'))
|
||||
|
||||
# long line splits signature
|
||||
eq_((range(7), [7, 8]),
|
||||
e._process_marked_lines(range(9), 'tsslsless'))
|
||||
eq_((list(range(7)), [7, 8]),
|
||||
e._process_marked_lines(list(range(9)), 'tsslsless'))
|
||||
|
||||
eq_((range(20), [20]),
|
||||
e._process_marked_lines(range(21), 'ttttttstttesllelelets'))
|
||||
eq_((list(range(20)), [20]),
|
||||
e._process_marked_lines(list(range(21)), 'ttttttstttesllelelets'))
|
||||
|
||||
# some signature lines could be identified as text
|
||||
eq_(([0], range(1, 9)), e._process_marked_lines(range(9), 'tsetetest'))
|
||||
eq_(([0], list(range(1, 9))), e._process_marked_lines(list(range(9)), 'tsetetest'))
|
||||
|
||||
eq_(([], range(5)),
|
||||
e._process_marked_lines(range(5), "ststt"))
|
||||
eq_(([], list(range(5))),
|
||||
e._process_marked_lines(list(range(5)), "ststt"))
|
||||
|
||||
@@ -1,11 +1,11 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
from ... import *
|
||||
import os
|
||||
|
||||
from PyML import SparseDataSet
|
||||
from numpy import genfromtxt
|
||||
|
||||
from talon.utils import to_unicode
|
||||
from talon.signature.learning import dataset as d
|
||||
|
||||
from talon.signature.learning.featurespace import features
|
||||
@@ -42,10 +42,13 @@ def test_build_extraction_dataset():
|
||||
d.build_extraction_dataset(os.path.join(EMAILS_DIR, 'P'),
|
||||
os.path.join(TMP_DIR,
|
||||
'extraction.data'), 1)
|
||||
test_data = SparseDataSet(os.path.join(TMP_DIR, 'extraction.data'),
|
||||
labelsColumn=-1)
|
||||
|
||||
filename = os.path.join(TMP_DIR, 'extraction.data')
|
||||
file_data = genfromtxt(filename, delimiter=",")
|
||||
test_data = file_data[:, :-1]
|
||||
|
||||
# the result is a loadable signature extraction dataset
|
||||
# 32 comes from 3 emails in emails/P folder, 11 lines checked to be
|
||||
# a signature, one email has only 10 lines
|
||||
eq_(test_data.size(), 32)
|
||||
eq_(len(features('')), test_data.numFeatures)
|
||||
eq_(test_data.shape[0], 32)
|
||||
eq_(len(features('')), test_data.shape[1])
|
||||
|
||||
@@ -1,12 +1,15 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
from ... import *
|
||||
|
||||
from talon.signature.learning import featurespace as fs
|
||||
|
||||
|
||||
def test_apply_features():
|
||||
s = '''John Doe
|
||||
s = '''This is John Doe
|
||||
|
||||
Tuesday @3pm suits. I'll chat to you then.
|
||||
|
||||
VP Research and Development, Xxxx Xxxx Xxxxx
|
||||
|
||||
@@ -19,11 +22,12 @@ john@example.com'''
|
||||
# note that we don't consider the first line because signatures don't
|
||||
# usually take all the text, empty lines are not considered
|
||||
eq_(result, [[1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1],
|
||||
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
|
||||
[1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
|
||||
[0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0],
|
||||
[0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0]])
|
||||
|
||||
with patch.object(fs, 'SIGNATURE_MAX_LINES', 4):
|
||||
with patch.object(fs, 'SIGNATURE_MAX_LINES', 5):
|
||||
features = fs.features(sender)
|
||||
new_result = fs.apply_features(s, features)
|
||||
# result remains the same because we don't consider empty lines
|
||||
|
||||
@@ -1,11 +1,13 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
from ... import *
|
||||
|
||||
import regex as re
|
||||
|
||||
from talon.signature.learning import helpers as h
|
||||
from talon.signature.learning.helpers import *
|
||||
from six.moves import range
|
||||
|
||||
# First testing regex constants.
|
||||
VALID = '''
|
||||
@@ -43,7 +45,7 @@ VALID_PHONE_NUMBERS = [e.strip() for e in VALID.splitlines() if e.strip()]
|
||||
|
||||
def test_match_phone_numbers():
|
||||
for phone in VALID_PHONE_NUMBERS:
|
||||
ok_(RE_RELAX_PHONE.match(phone), "{} should be matched".format(phone))
|
||||
ok_(RE_RELAX_PHONE.search(phone), "{} should be matched".format(phone))
|
||||
|
||||
|
||||
def test_match_names():
|
||||
@@ -52,29 +54,6 @@ def test_match_names():
|
||||
ok_(RE_NAME.match(name), "{} should be matched".format(name))
|
||||
|
||||
|
||||
def test_sender_with_name():
|
||||
ok_lines = ['Sergey Obukhov <serobnic@example.com>',
|
||||
'\tSergey <serobnic@example.com>',
|
||||
('"Doe, John (TX)"'
|
||||
'<DowJ@example.com>@EXAMPLE'
|
||||
'<IMCEANOTES-+22Doe+2C+20John+20'
|
||||
'+28TX+29+22+20+3CDoeJ+40example+2Ecom+3E'
|
||||
'+40EXAMPLE@EXAMPLE.com>'),
|
||||
('Company Sleuth <csleuth@email.xxx.com>'
|
||||
'@EXAMPLE <XXX-Company+20Sleuth+20+3Ccsleuth'
|
||||
'+40email+2Exxx+2Ecom+3E+40EXAMPLE@EXAMPLE.com>'),
|
||||
('Doe III, John '
|
||||
'</O=EXAMPLE/OU=NA/CN=RECIPIENTS/CN=jDOE5>')]
|
||||
for line in ok_lines:
|
||||
ok_(RE_SENDER_WITH_NAME.match(line),
|
||||
'{} should be matched'.format(line))
|
||||
|
||||
nok_lines = ['', '<serobnic@xxx.ru>', 'Sergey serobnic@xxx.ru']
|
||||
for line in nok_lines:
|
||||
assert_false(RE_SENDER_WITH_NAME.match(line),
|
||||
'{} should not be matched'.format(line))
|
||||
|
||||
|
||||
# Now test helpers functions
|
||||
def test_binary_regex_search():
|
||||
eq_(1, h.binary_regex_search(re.compile("12"))("12"))
|
||||
@@ -177,7 +156,7 @@ def test_extract_names():
|
||||
# check that extracted names could be compiled
|
||||
try:
|
||||
re.compile("|".join(extracted_names))
|
||||
except Exception, e:
|
||||
except Exception as e:
|
||||
ok_(False, ("Failed to compile extracted names {}"
|
||||
"\n\nReason: {}").format(extracted_names, e))
|
||||
if expected_names:
|
||||
@@ -213,10 +192,11 @@ def test_punctuation_percent(categories_percent):
|
||||
def test_capitalized_words_percent():
|
||||
eq_(0.0, h.capitalized_words_percent(''))
|
||||
eq_(100.0, h.capitalized_words_percent('Example Corp'))
|
||||
eq_(50.0, h.capitalized_words_percent('Qqq qqq QQQ 123 sss'))
|
||||
eq_(50.0, h.capitalized_words_percent('Qqq qqq Aqs 123 sss'))
|
||||
eq_(100.0, h.capitalized_words_percent('Cell 713-444-7368'))
|
||||
eq_(100.0, h.capitalized_words_percent('8th Floor'))
|
||||
eq_(0.0, h.capitalized_words_percent('(212) 230-9276'))
|
||||
eq_(50.0, h.capitalized_words_percent('Password: REMARKABLE'))
|
||||
|
||||
|
||||
def test_has_signature():
|
||||
@@ -227,7 +207,7 @@ def test_has_signature():
|
||||
'sender@example.com'))
|
||||
assert_false(h.has_signature('http://www.example.com/555-555-5555',
|
||||
'sender@example.com'))
|
||||
long_line = ''.join(['q' for e in xrange(28)])
|
||||
long_line = ''.join(['q' for e in range(28)])
|
||||
assert_false(h.has_signature(long_line + ' sender', 'sender@example.com'))
|
||||
# wont crash on an empty string
|
||||
assert_false(h.has_signature('', ''))
|
||||
|
||||
@@ -1,23 +1,26 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from __future__ import absolute_import
|
||||
from . import *
|
||||
from . fixtures import *
|
||||
|
||||
import os
|
||||
|
||||
from flanker import mime
|
||||
|
||||
import email.iterators
|
||||
from talon import quotations
|
||||
import six
|
||||
from six.moves import range
|
||||
from six import StringIO
|
||||
|
||||
|
||||
@patch.object(quotations, 'MAX_LINES_COUNT', 1)
|
||||
def test_too_many_lines():
|
||||
msg_body = """Test reply
|
||||
|
||||
Hi
|
||||
-----Original Message-----
|
||||
|
||||
Test"""
|
||||
eq_(msg_body, quotations.extract_from_plain(msg_body))
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_pattern_on_date_somebody_wrote():
|
||||
@@ -25,6 +28,32 @@ def test_pattern_on_date_somebody_wrote():
|
||||
|
||||
On 11-Apr-2011, at 6:54 PM, Roman Tkachenko <romant@example.com> wrote:
|
||||
|
||||
>
|
||||
> Test
|
||||
>
|
||||
> Roman"""
|
||||
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
def test_pattern_on_date_polymail():
|
||||
msg_body = """Test reply
|
||||
|
||||
On Tue, Apr 11, 2017 at 10:07 PM John Smith
|
||||
|
||||
<
|
||||
mailto:John Smith <johnsmith@gmail.com>
|
||||
> wrote:
|
||||
Test quoted data
|
||||
"""
|
||||
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_pattern_sent_from_samsung_smb_wrote():
|
||||
msg_body = """Test reply
|
||||
|
||||
Sent from Samsung MobileName <address@example.com> wrote:
|
||||
|
||||
>
|
||||
> Test
|
||||
>
|
||||
@@ -38,7 +67,7 @@ def test_pattern_on_date_wrote_somebody():
|
||||
"""Lorem
|
||||
|
||||
Op 13-02-2014 3:18 schreef Julius Caesar <pantheon@rome.com>:
|
||||
|
||||
|
||||
Veniam laborum mlkshk kale chips authentic. Normcore mumblecore laboris, fanny pack readymade eu blog chia pop-up freegan enim master cleanse.
|
||||
"""))
|
||||
|
||||
@@ -55,6 +84,18 @@ On 04/19/2011 07:10 AM, Roman Tkachenko wrote:
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_date_time_email_splitter():
|
||||
msg_body = """Test reply
|
||||
|
||||
2014-10-17 11:28 GMT+03:00 Postmaster <
|
||||
postmaster@sandboxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx.mailgun.org>:
|
||||
|
||||
> First from site
|
||||
>
|
||||
"""
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_pattern_on_date_somebody_wrote_allows_space_in_front():
|
||||
msg_body = """Thanks Thanmai
|
||||
On Mar 8, 2012 9:59 AM, "Example.com" <
|
||||
@@ -78,6 +119,38 @@ On 11-Apr-2011, at 6:54 PM, Roman Tkachenko <romant@example.com> sent:
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_appointment():
|
||||
msg_body = """Response
|
||||
|
||||
10/19/2017 @ 9:30 am for physical therapy
|
||||
Bla
|
||||
1517 4th Avenue Ste 300
|
||||
London CA 19129, 555-421-6780
|
||||
|
||||
John Doe, FCLS
|
||||
Mailgun Inc
|
||||
555-941-0697
|
||||
|
||||
From: from@example.com [mailto:from@example.com]
|
||||
Sent: Wednesday, October 18, 2017 2:05 PM
|
||||
To: John Doer - SIU <jd@example.com>
|
||||
Subject: RE: Claim # 5551188-1
|
||||
|
||||
Text"""
|
||||
|
||||
expected = """Response
|
||||
|
||||
10/19/2017 @ 9:30 am for physical therapy
|
||||
Bla
|
||||
1517 4th Avenue Ste 300
|
||||
London CA 19129, 555-421-6780
|
||||
|
||||
John Doe, FCLS
|
||||
Mailgun Inc
|
||||
555-941-0697"""
|
||||
eq_(expected, quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_line_starts_with_on():
|
||||
msg_body = """Blah-blah-blah
|
||||
On blah-blah-blah"""
|
||||
@@ -114,7 +187,8 @@ def _check_pattern_original_message(original_message_indicator):
|
||||
-----{}-----
|
||||
|
||||
Test"""
|
||||
eq_('Test reply', quotations.extract_from_plain(msg_body.format(unicode(original_message_indicator))))
|
||||
eq_('Test reply', quotations.extract_from_plain(
|
||||
msg_body.format(six.text_type(original_message_indicator))))
|
||||
|
||||
def test_english_original_message():
|
||||
_check_pattern_original_message('Original Message')
|
||||
@@ -137,6 +211,17 @@ Test reply"""
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_android_wrote():
|
||||
msg_body = """Test reply
|
||||
|
||||
---- John Smith wrote ----
|
||||
|
||||
> quoted
|
||||
> text
|
||||
"""
|
||||
eq_("Test reply", quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def test_reply_wraps_quotations():
|
||||
msg_body = """Test reply
|
||||
|
||||
@@ -216,7 +301,7 @@ def test_with_indent():
|
||||
|
||||
------On 12/29/1987 17:32 PM, Julius Caesar wrote-----
|
||||
|
||||
Brunch mumblecore pug Marfa tofu, irure taxidermy hoodie readymade pariatur.
|
||||
Brunch mumblecore pug Marfa tofu, irure taxidermy hoodie readymade pariatur.
|
||||
"""
|
||||
eq_("YOLO salvia cillum kogi typewriter mumblecore cardigan skateboard Austin.", quotations.extract_from_plain(msg_body))
|
||||
|
||||
@@ -312,15 +397,50 @@ Emne: The manager has commented on your Loop
|
||||
Blah-blah-blah
|
||||
"""))
|
||||
|
||||
def test_dutch_from_block():
|
||||
eq_('Gluten-free culpa lo-fi et nesciunt nostrud.', quotations.extract_from_plain(
|
||||
"""Gluten-free culpa lo-fi et nesciunt nostrud.
|
||||
def test_swedish_from_block():
|
||||
eq_('Allo! Follow up MIME!', quotations.extract_from_plain(
|
||||
u"""Allo! Follow up MIME!
|
||||
Från: Anno Sportel [mailto:anno.spoel@hsbcssad.com]
|
||||
Skickat: den 26 augusti 2015 14:45
|
||||
Till: Isacson Leiff
|
||||
Ämne: RE: Week 36
|
||||
|
||||
Op 17-feb.-2015, om 13:18 heeft Julius Caesar <pantheon@rome.com> het volgende geschreven:
|
||||
|
||||
Small batch beard laboris tempor, non listicle hella Tumblr heirloom.
|
||||
Blah-blah-blah
|
||||
"""))
|
||||
|
||||
def test_swedish_from_line():
|
||||
eq_('Lorem', quotations.extract_from_plain(
|
||||
"""Lorem
|
||||
Den 14 september, 2015 02:23:18, Valentino Rudy (valentino@rudy.be) skrev:
|
||||
|
||||
Veniam laborum mlkshk kale chips authentic. Normcore mumblecore laboris, fanny pack readymade eu blog chia pop-up freegan enim master cleanse.
|
||||
"""))
|
||||
|
||||
def test_norwegian_from_line():
|
||||
eq_('Lorem', quotations.extract_from_plain(
|
||||
u"""Lorem
|
||||
På 14 september 2015 på 02:23:18, Valentino Rudy (valentino@rudy.be) skrev:
|
||||
|
||||
Veniam laborum mlkshk kale chips authentic. Normcore mumblecore laboris, fanny pack readymade eu blog chia pop-up freegan enim master cleanse.
|
||||
"""))
|
||||
|
||||
def test_dutch_from_block():
|
||||
eq_('Gluten-free culpa lo-fi et nesciunt nostrud.', quotations.extract_from_plain(
|
||||
"""Gluten-free culpa lo-fi et nesciunt nostrud.
|
||||
|
||||
Op 17-feb.-2015, om 13:18 heeft Julius Caesar <pantheon@rome.com> het volgende geschreven:
|
||||
|
||||
Small batch beard laboris tempor, non listicle hella Tumblr heirloom.
|
||||
"""))
|
||||
|
||||
def test_vietnamese_from_block():
|
||||
eq_('Hello', quotations.extract_from_plain(
|
||||
u"""Hello
|
||||
|
||||
Vào 14:24 8 tháng 6, 2017, Hùng Nguyễn <hungnguyen@xxx.com> đã viết:
|
||||
|
||||
> Xin chào
|
||||
"""))
|
||||
|
||||
def test_quotation_marker_false_positive():
|
||||
msg_body = """Visit us now for assistance...
|
||||
@@ -333,7 +453,8 @@ def test_link_closed_with_quotation_marker_on_new_line():
|
||||
msg_body = '''8.45am-1pm
|
||||
|
||||
From: somebody@example.com
|
||||
|
||||
Date: Wed, 16 May 2012 00:15:02 -0600
|
||||
|
||||
<http://email.example.com/c/dHJhY2tpbmdfY29kZT1mMDdjYzBmNzM1ZjYzMGIxNT
|
||||
> <bob@example.com <mailto:bob@example.com> >
|
||||
|
||||
@@ -374,7 +495,9 @@ def test_from_block_starts_with_date():
|
||||
msg_body = """Blah
|
||||
|
||||
Date: Wed, 16 May 2012 00:15:02 -0600
|
||||
To: klizhentas@example.com"""
|
||||
To: klizhentas@example.com
|
||||
|
||||
"""
|
||||
eq_('Blah', quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
@@ -444,11 +567,12 @@ def test_mark_message_lines():
|
||||
# next line should be marked as splitter
|
||||
'_____________',
|
||||
'From: foo@bar.com',
|
||||
'Date: Wed, 16 May 2012 00:15:02 -0600',
|
||||
'',
|
||||
'> Hi',
|
||||
'',
|
||||
'Signature']
|
||||
eq_('tessemet', quotations.mark_message_lines(lines))
|
||||
eq_('tesssemet', quotations.mark_message_lines(lines))
|
||||
|
||||
lines = ['Just testing the email reply',
|
||||
'',
|
||||
@@ -611,25 +735,107 @@ def test_preprocess_postprocess_2_links():
|
||||
eq_(msg_body, quotations.extract_from_plain(msg_body))
|
||||
|
||||
|
||||
def body_iterator(msg, decode=False):
|
||||
for subpart in msg.walk():
|
||||
payload = subpart.get_payload(decode=decode)
|
||||
if isinstance(payload, six.text_type):
|
||||
yield payload
|
||||
else:
|
||||
yield payload.decode('utf8')
|
||||
|
||||
|
||||
def test_standard_replies():
|
||||
for filename in os.listdir(STANDARD_REPLIES):
|
||||
filename = os.path.join(STANDARD_REPLIES, filename)
|
||||
if os.path.isdir(filename):
|
||||
if not filename.endswith('.eml') or os.path.isdir(filename):
|
||||
continue
|
||||
with open(filename) as f:
|
||||
msg = f.read()
|
||||
m = mime.from_string(msg)
|
||||
for part in m.walk():
|
||||
if part.content_type == 'text/plain':
|
||||
text = part.body
|
||||
stripped_text = quotations.extract_from_plain(text)
|
||||
reply_text_fn = filename[:-4] + '_reply_text'
|
||||
if os.path.isfile(reply_text_fn):
|
||||
with open(reply_text_fn) as f:
|
||||
reply_text = f.read()
|
||||
else:
|
||||
reply_text = 'Hello'
|
||||
eq_(reply_text, stripped_text,
|
||||
"'%(reply)s' != %(stripped)s for %(fn)s" %
|
||||
{'reply': reply_text, 'stripped': stripped_text,
|
||||
'fn': filename})
|
||||
message = email.message_from_file(f)
|
||||
body = next(email.iterators.typed_subpart_iterator(message, subtype='plain'))
|
||||
text = ''.join(body_iterator(body, True))
|
||||
|
||||
stripped_text = quotations.extract_from_plain(text)
|
||||
reply_text_fn = filename[:-4] + '_reply_text'
|
||||
if os.path.isfile(reply_text_fn):
|
||||
with open(reply_text_fn) as f:
|
||||
reply_text = f.read().strip()
|
||||
else:
|
||||
reply_text = 'Hello'
|
||||
yield eq_, reply_text, stripped_text, \
|
||||
"'%(reply)s' != %(stripped)s for %(fn)s" % \
|
||||
{'reply': reply_text, 'stripped': stripped_text,
|
||||
'fn': filename}
|
||||
|
||||
|
||||
def test_split_email():
|
||||
msg = """From: Mr. X
|
||||
Date: 24 February 2016
|
||||
To: Mr. Y
|
||||
Subject: Hi
|
||||
Attachments: none
|
||||
Goodbye.
|
||||
From: Mr. Y
|
||||
To: Mr. X
|
||||
Date: 24 February 2016
|
||||
Subject: Hi
|
||||
Attachments: none
|
||||
|
||||
Hello.
|
||||
|
||||
On 24th February 2016 at 09.32am, Conal wrote:
|
||||
|
||||
Hey!
|
||||
|
||||
On Mon, 2016-10-03 at 09:45 -0600, Stangel, Dan wrote:
|
||||
> Mohan,
|
||||
>
|
||||
> We have not yet migrated the systems.
|
||||
>
|
||||
> Dan
|
||||
>
|
||||
> > -----Original Message-----
|
||||
> > Date: Mon, 2 Apr 2012 17:44:22 +0400
|
||||
> > Subject: Test
|
||||
> > From: bob@xxx.mailgun.org
|
||||
> > To: xxx@gmail.com; xxx@hotmail.com; xxx@yahoo.com; xxx@aol.com; xxx@comcast.net; xxx@nyc.rr.com
|
||||
> >
|
||||
> > Hi
|
||||
> >
|
||||
> > > From: bob@xxx.mailgun.org
|
||||
> > > To: xxx@gmail.com; xxx@hotmail.com; xxx@yahoo.com; xxx@aol.com; xxx@comcast.net; xxx@nyc.rr.com
|
||||
> > > Date: Mon, 2 Apr 2012 17:44:22 +0400
|
||||
> > > Subject: Test
|
||||
> > > Hi
|
||||
> > >
|
||||
> >
|
||||
>
|
||||
>
|
||||
"""
|
||||
expected_markers = "stttttsttttetesetesmmmmmmsmmmmmmmmmmmmmmmm"
|
||||
markers = quotations.split_emails(msg)
|
||||
eq_(markers, expected_markers)
|
||||
|
||||
|
||||
|
||||
def test_feedback_below_left_unparsed():
|
||||
msg_body = """Please enter your feedback below. Thank you.
|
||||
|
||||
------------------------------------- Enter Feedback Below -------------------------------------
|
||||
|
||||
The user experience was unparallelled. Please continue production. I'm sending payment to ensure
|
||||
that this line is intact."""
|
||||
|
||||
parsed = quotations.extract_from_plain(msg_body)
|
||||
eq_(msg_body, parsed)
|
||||
|
||||
|
||||
def test_appointment_2():
|
||||
msg_body = """Invitation for an interview:
|
||||
|
||||
Date: Wednesday 3, October 2011
|
||||
Time: 7 : 00am
|
||||
Address: 130 Fox St
|
||||
|
||||
Please bring in your ID."""
|
||||
parsed = quotations.extract_from_plain(msg_body)
|
||||
eq_(msg_body, parsed)
|
||||
|
||||
@@ -1,9 +1,163 @@
|
||||
from . import *
|
||||
# coding:utf-8
|
||||
|
||||
from talon import utils
|
||||
from __future__ import absolute_import
|
||||
|
||||
import cchardet
|
||||
import six
|
||||
|
||||
from talon import utils as u
|
||||
from . import *
|
||||
|
||||
|
||||
def test_get_delimiter():
|
||||
eq_('\r\n', utils.get_delimiter('abc\r\n123'))
|
||||
eq_('\n', utils.get_delimiter('abc\n123'))
|
||||
eq_('\n', utils.get_delimiter('abc'))
|
||||
eq_('\r\n', u.get_delimiter('abc\r\n123'))
|
||||
eq_('\n', u.get_delimiter('abc\n123'))
|
||||
eq_('\n', u.get_delimiter('abc'))
|
||||
|
||||
|
||||
def test_unicode():
|
||||
eq_(u'hi', u.to_unicode('hi'))
|
||||
eq_(type(u.to_unicode('hi')), six.text_type)
|
||||
eq_(type(u.to_unicode(u'hi')), six.text_type)
|
||||
eq_(type(u.to_unicode('привет')), six.text_type)
|
||||
eq_(type(u.to_unicode(u'привет')), six.text_type)
|
||||
eq_(u"привет", u.to_unicode('привет'))
|
||||
eq_(u"привет", u.to_unicode(u'привет'))
|
||||
# some latin1 stuff
|
||||
eq_(u"Versión", u.to_unicode(u'Versi\xf3n'.encode('iso-8859-2'), precise=True))
|
||||
|
||||
|
||||
def test_detect_encoding():
|
||||
eq_('ascii', u.detect_encoding(b'qwe').lower())
|
||||
ok_(u.detect_encoding(
|
||||
u'Versi\xf3n'.encode('iso-8859-2')).lower() in [
|
||||
'iso-8859-1', 'iso-8859-2'])
|
||||
eq_('utf-8', u.detect_encoding(u'привет'.encode('utf8')).lower())
|
||||
# fallback to utf-8
|
||||
with patch.object(u.chardet, 'detect') as detect:
|
||||
detect.side_effect = Exception
|
||||
eq_('utf-8', u.detect_encoding('qwe'.encode('utf8')).lower())
|
||||
|
||||
|
||||
def test_quick_detect_encoding():
|
||||
eq_('ascii', u.quick_detect_encoding(b'qwe').lower())
|
||||
ok_(u.quick_detect_encoding(
|
||||
u'Versi\xf3n'.encode('windows-1252')).lower() in [
|
||||
'windows-1252', 'windows-1250'])
|
||||
eq_('utf-8', u.quick_detect_encoding(u'привет'.encode('utf8')).lower())
|
||||
|
||||
|
||||
@patch.object(cchardet, 'detect')
|
||||
@patch.object(u, 'detect_encoding')
|
||||
def test_quick_detect_encoding_edge_cases(detect_encoding, cchardet_detect):
|
||||
cchardet_detect.return_value = {'encoding': 'ascii'}
|
||||
eq_('ascii', u.quick_detect_encoding(b"qwe"))
|
||||
cchardet_detect.assert_called_once_with(b"qwe")
|
||||
|
||||
# fallback to detect_encoding
|
||||
cchardet_detect.return_value = {}
|
||||
detect_encoding.return_value = 'utf-8'
|
||||
eq_('utf-8', u.quick_detect_encoding(b"qwe"))
|
||||
|
||||
# exception
|
||||
detect_encoding.reset_mock()
|
||||
cchardet_detect.side_effect = Exception()
|
||||
detect_encoding.return_value = 'utf-8'
|
||||
eq_('utf-8', u.quick_detect_encoding(b"qwe"))
|
||||
ok_(detect_encoding.called)
|
||||
|
||||
|
||||
def test_html_to_text():
|
||||
html = """<body>
|
||||
<p>Hello world!</p>
|
||||
<br>
|
||||
<ul>
|
||||
<li>One!</li>
|
||||
<li>Two</li>
|
||||
</ul>
|
||||
<p>
|
||||
Haha
|
||||
</p>
|
||||
</body>"""
|
||||
text = u.html_to_text(html)
|
||||
eq_(b"Hello world! \n\n * One! \n * Two \nHaha", text)
|
||||
eq_(u"привет!", u.html_to_text("<b>привет!</b>").decode('utf8'))
|
||||
|
||||
html = '<body><br/><br/>Hi</body>'
|
||||
eq_(b'Hi', u.html_to_text(html))
|
||||
|
||||
html = """Hi
|
||||
<style type="text/css">
|
||||
|
||||
div, p, li {
|
||||
|
||||
font: 13px 'Lucida Grande', Arial, sans-serif;
|
||||
|
||||
}
|
||||
</style>
|
||||
|
||||
<style type="text/css">
|
||||
|
||||
h1 {
|
||||
|
||||
font: 13px 'Lucida Grande', Arial, sans-serif;
|
||||
|
||||
}
|
||||
</style>"""
|
||||
eq_(b'Hi', u.html_to_text(html))
|
||||
|
||||
html = """<div>
|
||||
<!-- COMMENT 1 -->
|
||||
<span>TEXT 1</span>
|
||||
<p>TEXT 2 <!-- COMMENT 2 --></p>
|
||||
</div>"""
|
||||
eq_(b'TEXT 1 \nTEXT 2', u.html_to_text(html))
|
||||
|
||||
|
||||
def test_comment_no_parent():
|
||||
s = b'<!-- COMMENT 1 --> no comment'
|
||||
d = u.html_document_fromstring(s)
|
||||
eq_(b"no comment", u.html_tree_to_text(d))
|
||||
|
||||
|
||||
@patch.object(u.html5parser, 'fromstring', Mock(side_effect=Exception()))
|
||||
def test_html_fromstring_exception():
|
||||
eq_(None, u.html_fromstring("<html></html>"))
|
||||
|
||||
|
||||
@patch.object(u, 'html_too_big', Mock())
|
||||
@patch.object(u.html5parser, 'fromstring')
|
||||
def test_html_fromstring_too_big(fromstring):
|
||||
eq_(None, u.html_fromstring("<html></html>"))
|
||||
assert_false(fromstring.called)
|
||||
|
||||
|
||||
@patch.object(u.html5parser, 'document_fromstring')
|
||||
def test_html_document_fromstring_exception(document_fromstring):
|
||||
document_fromstring.side_effect = Exception()
|
||||
eq_(None, u.html_document_fromstring("<html></html>"))
|
||||
|
||||
|
||||
@patch.object(u, 'html_too_big', Mock())
|
||||
@patch.object(u.html5parser, 'document_fromstring')
|
||||
def test_html_document_fromstring_too_big(document_fromstring):
|
||||
eq_(None, u.html_document_fromstring("<html></html>"))
|
||||
assert_false(document_fromstring.called)
|
||||
|
||||
|
||||
@patch.object(u, 'html_fromstring', Mock(return_value=None))
|
||||
def test_bad_html_to_text():
|
||||
bad_html = "one<br>two<br>three"
|
||||
eq_(None, u.html_to_text(bad_html))
|
||||
|
||||
|
||||
@patch.object(u, '_MAX_TAGS_COUNT', 3)
|
||||
def test_html_too_big():
|
||||
eq_(False, u.html_too_big("<div></div>"))
|
||||
eq_(True, u.html_too_big("<div><span>Hi</span></div>"))
|
||||
|
||||
|
||||
@patch.object(u, '_MAX_TAGS_COUNT', 3)
|
||||
def test_html_to_text():
|
||||
eq_(b"Hello", u.html_to_text("<div>Hello</div>"))
|
||||
eq_(None, u.html_to_text("<div><span>Hi</span></div>"))
|
||||
|
||||
11
train.py
Normal file
11
train.py
Normal file
@@ -0,0 +1,11 @@
|
||||
from __future__ import absolute_import
|
||||
from talon.signature import EXTRACTOR_FILENAME, EXTRACTOR_DATA
|
||||
from talon.signature.learning.classifier import train, init
|
||||
|
||||
|
||||
def train_model():
|
||||
""" retrain model and persist """
|
||||
train(init(), EXTRACTOR_DATA, EXTRACTOR_FILENAME)
|
||||
|
||||
if __name__ == "__main__":
|
||||
train_model()
|
||||
Reference in New Issue
Block a user