delta_m / README.md
angelosps's picture
Update README.md
1fed022 verified
|
raw
history blame
3.5 kB
metadata
license: mit
tags:
  - textual-entailment
  - logical-reasoning
  - deberta
language:
  - en
metrics:
  - accuracy
pipeline_tag: text-classification

DELTA: Description Logics with Transformers

Fine-tuning a transformer model for textual entailment over expressive contexts generated from description logic knowledge bases. Specifically, the model is given a context (a set of facts and rules) and a question. The model should answer with "True" if the question is logically implied from the context, "False" if it contradicts the context, and "Unknown" if none of the two.

For more info please see our paper.

Model Details

Model Description

DELTAM is a DeBERTaV3 large model fine-tuned on the DELTAD dataset.

  • License: MIT
  • Finetuned from model: microsoft/deberta-v3-large

Model Sources

Citation

BibTeX:

@misc{poulis2024transformers,
      title={Transformers in the Service of Description Logic-based Contexts}, 
      author={Angelos Poulis and Eleni Tsalapati and Manolis Koubarakis},
      year={2024},
      eprint={2311.08941},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}