🚩 Report

#137
by hayleecs - opened

No information about the training data. How are we supposed to compare it with LLaMa 2 models and how can we justify the change in perplexity on common datasets such as wikitext?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment