Update README.md
Browse files
README.md
CHANGED
@@ -11,9 +11,9 @@ widget:
|
|
11 |
DistilCamemBERT-Sentiment
|
12 |
=========================
|
13 |
|
14 |
-
We present DistilCamemBERT-Sentiment which is [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base) fine tuned for the sentiment analysis task for the French language. This model is
|
15 |
|
16 |
-
This
|
17 |
|
18 |
Dataset
|
19 |
-------
|
|
|
11 |
DistilCamemBERT-Sentiment
|
12 |
=========================
|
13 |
|
14 |
+
We present DistilCamemBERT-Sentiment which is [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base) fine tuned for the sentiment analysis task for the French language. This model is constructed over 2 datasets: [Amazon Reviews](https://huggingface.co/datasets/amazon_reviews_multi) and [Allociné.fr](https://huggingface.co/datasets/allocine) in order to minimize the bias. Indeed, Amazon reviews are very similar in the messages and relatively shorts, contrary to Allociné critics which are long and rich texts.
|
15 |
|
16 |
+
This modelization is close to [tblard/tf-allocine](https://huggingface.co/tblard/tf-allocine) based on [CamemBERT](https://huggingface.co/camembert-base) model. The problem of the modelizations based on CamemBERT is at the scaling moment, for the production phase for example. Indeed, inference cost can be a technological issue. To counteract this effect, we propose this modelization which **divides the inference time by 2** with the same consumption power thanks to [DistilCamemBERT](https://huggingface.co/cmarkea/distilcamembert-base).
|
17 |
|
18 |
Dataset
|
19 |
-------
|