|
{ |
|
"config": { |
|
"query_token_id": "[unused0]", |
|
"doc_token_id": "[unused1]", |
|
"query_token": "[Q]", |
|
"doc_token": "[D]", |
|
"ncells": null, |
|
"centroid_score_threshold": null, |
|
"ndocs": null, |
|
"load_index_with_mmap": false, |
|
"index_path": null, |
|
"index_bsize": 32, |
|
"nbits": 8, |
|
"kmeans_niters": 20, |
|
"resume": false, |
|
"similarity": "cosine", |
|
"bsize": 64, |
|
"accumsteps": 1, |
|
"lr": 1e-5, |
|
"maxsteps": 400000, |
|
"save_every": null, |
|
"warmup": 20000, |
|
"warmup_bert": null, |
|
"relu": false, |
|
"nway": 64, |
|
"use_ib_negatives": true, |
|
"reranker": false, |
|
"distillation_alpha": 1.0, |
|
"ignore_scores": false, |
|
"model_name": null, |
|
"query_maxlen": 32, |
|
"attend_to_mask_tokens": false, |
|
"interaction": "colbert", |
|
"dim": 128, |
|
"doc_maxlen": 256, |
|
"mask_punctuation": true, |
|
"checkpoint": "colbert-ir\/colbertv2.0", |
|
"triples": "\/future\/u\/okhattab\/root\/unit\/experiments\/2021.10\/downstream.distillation.round2.2_score\/round2.nway6.cosine.ib\/examples.64.json", |
|
"collection": [ |
|
"list with 4770 elements starting with...", |
|
[ |
|
"Deep neural networks have demonstrated remarkable performance in supervised learning tasks but require large amounts of labeled data. Self-supervised learning offers an alternative paradigm, enabling the model to learn from data without explicit labels. Information theory has been instrumental in understanding and optimizing deep neural networks. Specifically, the information bottleneck principle has been applied to optimize the trade-off between compression and relevant information preservation in supervised settings. However, the optimal information objective in self-supervised learning remains unclear. In this paper, we review various approaches to self-supervised learning from an information-theoretic standpoint and present a unified framework that formalizes the self-supervised information-theoretic learning problem. We integrate existing research into a coherent framework, examine recent self-supervised methods, and identify research opportunities and challenges. Moreover, we discuss empirical measurement of information-theoretic quantities and their estimators. This paper offers a comprehensive review of the intersection between information theory, self-supervised learning, and deep neural networks.", |
|
"Pre-trained large language models (LLMs) capture procedural knowledge about the world. Recent work has leveraged LLM's ability to generate abstract plans to simplify challenging control tasks, either by action scoring, or action modeling (fine-tuning). However, the transformer architecture inherits several constraints that make it difficult for the LLM to directly serve as the agent: e.g. limited input lengths, fine-tuning inefficiency, bias from pre-training, and incompatibility with non-text environments. To maintain compatibility with a low-level trainable actor, we propose to instead use the knowledge in LLMs to simplify the control problem, rather than solving it. We propose the Plan, Eliminate, and Track (PET) framework. The Plan module translates a task description into a list of high-level sub-tasks. The Eliminate module masks out irrelevant objects and receptacles from the observation for the current sub-task. Finally, the Track module determines whether the agent has accomplished each sub-task. On the AlfWorld instruction following benchmark, the PET framework leads to a significant 15% improvement over SOTA for generalization to human goal specifications.", |
|
"Large language models (LLMs) power many state-of-the-art systems in natural language processing. However, these models are extremely computationally expensive, even at inference time, raising the natural question: when is the extra cost of deploying a larger model worth the anticipated boost in capabilities? Better understanding this tradeoff fundamentally could benefit from an inference efficiency metric that is both (i) easily comparable across models from different providers, and (ii) representative of the true cost of running queries in an isolated performance environment. Unfortunately, access to LLMs today is largely restricted to black-box text generation APIs and raw runtimes measured through this interface do not satisfy these desiderata: model providers can apply various software and hardware optimizations orthogonal to the model, and models served on shared infrastructure are susceptible to performance contention. To circumvent these problems, we propose a new metric for comparing inference efficiency across models. This metric puts models on equal footing as though they were served (i) on uniform hardware and software, and (ii) without performance contention. We call this metric the idealized runtime, and we propose a methodology to efficiently estimate this metric for autoregressive Transformer models." |
|
] |
|
], |
|
"queries": "\/future\/u\/okhattab\/data\/MSMARCO\/queries.train.tsv", |
|
"index_name": "daily-papers-abstract-index", |
|
"overwrite": false, |
|
"root": ".ragatouille\/", |
|
"experiment": "colbert", |
|
"index_root": null, |
|
"name": "2024-08\/16\/01.54.53", |
|
"rank": 0, |
|
"nranks": 1, |
|
"amp": true, |
|
"gpus": 1, |
|
"avoid_fork_if_possible": false |
|
}, |
|
"num_chunks": 1, |
|
"num_partitions": 8192, |
|
"num_embeddings_est": 816999.9897766113, |
|
"avg_doclen_est": 171.27882385253906 |
|
} |
|
|