|
import re |
|
import math |
|
import unidecode |
|
import copy |
|
from pprint import pp |
|
|
|
from lexical_diversity import lex_div as ld |
|
from dtaidistance import dtw |
|
from fastdtw import fastdtw |
|
import numpy as np |
|
|
|
import encode_lines as encode |
|
|
|
|
|
def measure_lex_div(p1, mode="mtld"): |
|
""" |
|
Measure of lexical diversity of a set of lines |
|
|
|
Parameters |
|
---------- |
|
p1 : list(str) |
|
list of lines of paragraph or song as strings |
|
mode : str, optional |
|
lexical diversity metric, by default "mtld" |
|
|
|
Returns |
|
------- |
|
float |
|
lexical diversity score |
|
""" |
|
lines = " ".join(p1) |
|
flem_tokens = ld.flemmatize(unidecode.unidecode(lines)) |
|
if mode == "avg_ttr": |
|
lex_div = ld.ttr(flem_tokens) / len(flem_tokens) |
|
elif mode == "mtld": |
|
lex_div = ld.mtld(flem_tokens) |
|
elif mode == "avg_mtld": |
|
lex_div = ld.mtld_ma_wrap(flem_tokens) |
|
return lex_div |
|
|
|
|
|
def measure_meter(p1, p2): |
|
""" |
|
Measure: meter consistency |
|
|
|
Meter consistency between two lines or paragraphs is defined as the negative |
|
exponential of the edit distance between their stress encodings. |
|
|
|
Note that this mapping limits the "maximal" error, as extremely high syllable differences |
|
will incur exponentially lower loss. The score is scaled between 0 and 100 |
|
|
|
Parameters |
|
---------- |
|
p1 : list |
|
paragraph as a list of line strings |
|
p2 : list |
|
comparison paragraph as a list of line strings |
|
|
|
Returns |
|
------- |
|
float |
|
score of meter consistency |
|
""" |
|
|
|
p1 = [encode.encode_line_meter_count(line) for line in p1] |
|
p2 = [encode.encode_line_meter_count(line) for line in p2] |
|
|
|
p1_string = "".join(line_stress for line_stress in p1).replace(" ", "").strip() |
|
p2_string = "".join(line_stress for line_stress in p2).replace(" ", "").strip() |
|
|
|
p1_string = re.sub(r"\s+", "", p1_string, flags=re.UNICODE) |
|
p2_string = re.sub(r"\s+", "", p2_string, flags=re.UNICODE) |
|
|
|
edit_dist = levenshteinDistance(p1_string, p2_string) |
|
return edit_dist |
|
|
|
|
|
def measure_syllable(p1: list, p2: list): |
|
""" |
|
Measure: syllable consistency |
|
|
|
We perform an altered version of the syllable DTW described in the methods |
|
section of https://staff.aist.go.jp/m.goto/PAPER/TIEICE202309watanabe.pdf. |
|
By recognizing that sometimes entire lines are out of "sync" between paragraphs, |
|
we first minize the DTW via sliding window between the two paragraphs, then compute |
|
the DTW line-by-line at the minimized offset. |
|
|
|
Syllable consistency between two lines or paragraphs is defined as the |
|
negative exponential of the DTW distance between two paragraph syllables. |
|
|
|
Note that this mapping limits the "maximal" error, as extremely high syllable differences |
|
will incur exponentially lower loss. The score is scaled between 0 and 100 |
|
|
|
Parameters |
|
---------- |
|
p1 : list |
|
paragraph as a list of line strings |
|
p2 : list |
|
comparison paragraph as a list of line strings |
|
|
|
Returns |
|
------- |
|
float |
|
length-normalized syllabic consistency score |
|
""" |
|
score = 0 |
|
|
|
|
|
enc_fn = encode.encode_line_syllable_count |
|
|
|
def encode_paragraph(par): |
|
out = [] |
|
for line in par: |
|
encoded_line = enc_fn(line).split(" ") |
|
encoded_line = [word for word in encoded_line if word] |
|
if len(encoded_line) == 0: |
|
out.append([0]) |
|
else: |
|
out.append(list(map(int, encoded_line))) |
|
return out |
|
|
|
p1, p2 = encode_paragraph(p1), encode_paragraph(p2) |
|
|
|
|
|
_, p1_c, p2_c = min_dtw_offset(p1, p2, return_cropped=True, use_short_window=True) |
|
|
|
for i in range(len(p1_c)): |
|
score += abs(sum(p1_c[i]) - sum(p2_c[i])) |
|
|
|
return score |
|
|
|
|
|
def min_dtw_offset(p1, p2, return_cropped=True, use_short_window=True): |
|
""" |
|
Use a sliding window (of lines) to find the line index offset which minimizes |
|
the syllabic DTW between two paragraphs. |
|
|
|
In addition to outside sliding window, setting inner_window > 0 allows for |
|
an moving inner window across the smaller signal to find a cropping |
|
|
|
Parameters |
|
---------- |
|
p1 : list |
|
list of lists of song lines encoded as syllable counts |
|
p2 : list |
|
list of song lines encoded as syllable counts |
|
return_cropped : bool, optional |
|
whether to return the cropped min window for p1 and p2 ro the offset, by |
|
default True |
|
use_short_window : bool, optional |
|
whether to compare at smaller or larger paragraphs length, by default true |
|
Note: if False, p1 and p2 will not be the same length |
|
|
|
Returns |
|
------- |
|
int |
|
dtw minimizing offset value of smaller signal in larger signal |
|
list |
|
(p1) cropped or uncropped p1, squared with zeros |
|
list |
|
(p2) cropped or uncropped p2, squared with zeros |
|
""" |
|
switched = False |
|
if len(p1) < len(p2): |
|
switched = True |
|
p1, p2 = p2, p1 |
|
|
|
|
|
win_length = len(p2) if use_short_window else len(p1) |
|
|
|
|
|
sig1, sig2 = copy.deepcopy(p1), copy.deepcopy(p2) |
|
max_val = max(max(len(l) for l in sig1), max(len(l) for l in sig2)) |
|
for i in range(len(p1)): |
|
sig1[i] += [0] * (max_val - len(sig1[i])) |
|
for i in range(len(p2)): |
|
sig2[i] += [0] * (max_val - len(sig2[i])) |
|
sig1, sig2 = np.sum(np.array(sig1), axis=1), np.sum(np.array(sig2), axis=1) |
|
|
|
|
|
sig1, sig2 = np.pad(sig1, (win_length, win_length)), sig2.astype(np.double) |
|
min_idx, min_error = 0, np.inf |
|
for j in range(sig1.shape[0] - win_length): |
|
sig1_win = sig1[j : j + sig1.shape[0]].astype(np.double) |
|
|
|
error = dtw.distance_fast(sig1_win, sig2, inner_dist="euclidean") |
|
|
|
if error < min_error: |
|
min_error = error |
|
min_idx = j |
|
|
|
|
|
if return_cropped: |
|
p1 = [[0]] * win_length + p1 + [[0]] * win_length |
|
p1 = p1[min_idx : min_idx + win_length] |
|
|
|
if switched: |
|
p1, p2 = p2, p1 |
|
|
|
return min_idx - win_length, p1, p2 |
|
|
|
|
|
def levenshteinDistance(s1, s2): |
|
""" |
|
DP edit distance implementation |
|
credit to https://stackoverflow.com/questions/2460177 |
|
|
|
Parameters |
|
---------- |
|
s1 : str |
|
first string to compare |
|
s2 : str |
|
second string to compare |
|
|
|
Returns |
|
------- |
|
int |
|
edit distance between two strings (absolute) |
|
""" |
|
if len(s1) > len(s2): |
|
s1, s2 = s2, s1 |
|
|
|
distances = range(len(s1) + 1) |
|
for i2, c2 in enumerate(s2): |
|
distances_ = [i2 + 1] |
|
for i1, c1 in enumerate(s1): |
|
if c1 == c2: |
|
distances_.append(distances[i1]) |
|
else: |
|
distances_.append( |
|
1 + min((distances[i1], distances[i1 + 1], distances_[-1])) |
|
) |
|
distances = distances_ |
|
return distances[-1] |
|
|