from transformers import AutoTokenizer from datasets import load_dataset def main(): # dataset = load_dataset("tau/fs",name="summ_screen_fd", max_source_length=512, tokenizer=tokenizer, prompt="Summary:") ssfd_debug = load_dataset("/Users/yuvalkirstain/repos/fs", name="summ_screen_fd") x = 5 # arxiv_debug = load_dataset("/Users/yuvalkirstain/repos/fs", name="arxiv_debug", max_source_length=512, # tokenizer=tokenizer, prompt="Summarize the above:") if __name__ == '__main__': main()