Conditional Utilitarian Roberta 01

Model description

This is a Roberta-based model. It was first fine-tuned on for computing utility estimates of experiences (see utilitarian-roberta-01. It was then further fine-tuned on 160 examples of pairwise comparisons of conditional utilities.

Intended use

The main use case is the computation of utility estimates of first-person text scenarios, under extra contextual information.

Limitations

The model was fine-tuned on only 160 examples, so it should be expected to have limited performance.

Further, while the base model was trained on ~10000 examples, they are still restricted, and only on first-person sentences. It does not have the capability of interpreting highly complex or unusual scenarios, and it does not have hard guarantees on its domain of accuracy.

How to use

Given a scenario S under a context C, and the model U, one computes the estimated conditional utility with U(f'{C} {S}') - U(C).

Training data

The first training data is the train split from the Utilitarianism part of the ETHICS dataset.

The second training data consists of 160 crowdsourced examples of triples (S, C0, C1) consisting of one scenario and two possible contexts, where U(S | C0) > U(S | C1).

Training procedure

Starting from utilitarian-roberta-01, we fine-tune the model over the training data of 160 examples, with a learning rate of 1e-5, a batch size of 8, and for 2 epochs.

Evaluation results

The model achieves ~70% accuracy over 40 crowdsourced examples, from the same distribution as the training data.

Downloads last month
6
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.