Spaces:
Running
Running
File size: 1,644 Bytes
91eaff6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
unit tests:
* serialization and deserialization
see copyright/license https://huggingface.co/spaces/DerwenAI/textgraphs/blob/main/README.md
"""
from os.path import abspath, dirname
import json
import pathlib
import sys
import deepdiff # pylint: disable=E0401
sys.path.insert(0, str(pathlib.Path(dirname(dirname(abspath(__file__))))))
import textgraphs # pylint: disable=C0413
def test_load_minimal (
*,
debug: bool = False,
) -> None:
"""
Construct a _lemma graph_ from a minimal example, then compare
serialized and deserialized data to ensure no fields get corrupted
in the conversions.
"""
text: str = """
See Spot run.
"""
tg: textgraphs.TextGraphs = textgraphs.TextGraphs() # pylint: disable=C0103
pipe: textgraphs.Pipeline = tg.create_pipeline(text.strip())
# serialize into node-link format
tg.collect_graph_elements(pipe)
tg.construct_lemma_graph()
tg.calc_phrase_ranks()
json_str: str = tg.dump_lemma_graph()
exp_graph = json.loads(json_str)
# deserialize from node-link format
tg = textgraphs.TextGraphs() # pylint: disable=C0103
tg.load_lemma_graph(json_str)
tg.construct_lemma_graph()
obs_graph: dict = json.loads(tg.dump_lemma_graph())
if debug:
print(obs_graph)
# compare
diff: deepdiff.diff.DeepDiff = deepdiff.DeepDiff(exp_graph, obs_graph)
if debug:
print(diff)
if len(diff) > 0:
print(json.dumps(json.loads(diff.to_json()), indent = 2))
assert len(diff) == 0
if __name__ == "__main__":
test_load_minimal(debug = True)
|