colonelwatch's picture
Update documentation
cd88b4f
metadata
language:
  - en
license: cc0-1.0
size_categories:
  - 10M<n<100m
task_categories:
  - text-retrieval
task_ids:
  - document-retrieval

abstracts-embeddings

This is the embeddings of the titles and abstracts of 110 million academic publications taken from the OpenAlex dataset as of January 1, 2025. The embeddings are generated with a Unix pipeline, chaining together the AWS CLI, gzip, oa_jsonl (a C parser tailored to the JSON Lines structure of the OpenAlex snapshot), and a Python embedding script. The source code of oa_jsonl and the Makefile which sets up the pipeline is available on Github, but the general process is as follows:

  1. Decode the JSON entry of an individual work
  2. From the language field, determine if the abstract will be in English, and if not, go back to step 1
  3. From the abstract inverted index field, reconstruct the text of the abstract
  4. If there is a title field, construct a single document in the format title + ' ' + abstract, or if not, just use the abstract
  5. Compute an embedding with the stella_en_1.5B_v5 model
  6. Write it to a local SQLite3 database

Said database is then exported in parquet format as pairs of OpenAlex IDs and length-1024 float32 vectors. The model was run with bfloat16 quantization, yielding bfloat16 vectors, but the conversion from bfloat16 to float32 leaves the lower two bytes as all-zero. This was exploited with byte-stream compression to store the vectors in a parquet with full precision but no wasted space. This does however mean that opening the parquets in the Hugging Face datasets library will lead to the cache using twice the space.

Though the OpenAlex dataset records 240 million works, not all of these works have abstracts or are in English. Besides the works without abstracts, the stella_en_1.5B_v5 model was only trained on English texts, hence the filtering.