File size: 1,948 Bytes
3fd7099
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
license: mit
language:
- en
library_name: transformers
---
# Milenium AI
This is a custom transformer-based model designed to answer questions based on a given context. It was trained on the SQuAD dataset and achieves a high accuracy on the validation set.

#### Model Architecture

The model consists of an encoder and a decoder. The encoder takes in the context and question as input and generates a encoded representation of the input. The decoder takes this encoded representation and generates the answer.

#### Training

The model was trained on the SQuAD dataset with a batch size of 32 and a maximum sequence length of 100. It was trained for 1 epoch with the Adam optimizer and sparse categorical crossentropy loss.

#### Evaluation

The model achieves an accuracy of 85% on the validation set.

#### Usage

You can use this model to answer questions based on a given context. Simply tokenize the context and question, and pass them as input to the model.

#### Limitations

This model is limited to answering questions based on the SQuAD dataset. It may not generalize well to other datasets or tasks.

#### Authors

Caeden Rajoo

#### How to use

You can use this model by loading it with the `transformers` library and passing in the context and question as input. For example:
python
```
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer

model = AutoModelForSeq2SeqLM.from_pretrained("milenium_model")
tokenizer = AutoTokenizer.from_pretrained("milenium_model")

context = "This is some context."
question = "What is the meaning of life?"

input_ids = tokenizer.encode(context, return_tensors="pt")
attention_mask = tokenizer.encode(context, return_tensors="pt", max_length=100, padding="max_length", truncation=True)
labels = tokenizer.encode(question, return_tensors="pt")

outputs = model(input_ids, attention_mask=attention_mask, labels=labels)

answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
```