LMUnit Models & Datasets for LMunit evaluation ContextualAI/LFQA Viewer • Updated 13 days ago • 260 • 83 ContextualAI/reward-bench-filtered Viewer • Updated 13 days ago • 2.99k • 54 ContextualAI/BiGGenBench Viewer • Updated 12 days ago • 3.2k • 59 ContextualAI/Flask Viewer • Updated 14 days ago • 2k • 64
Archangel Archangel is a suite of human feedback-aligned LLMs, released as part of the Human-Aware Loss Functions (HALOs) project by Ethayarajh et al. (2024). ContextualAI/archangel_sft_llama7b Text Generation • 7B • Updated Jan 11, 2024 • 5 • 1 ContextualAI/archangel_kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 5 • 2 ContextualAI/archangel_sft-kto_llama13b Text Generation • 13B • Updated Jan 11, 2024 • 1.45k • 3 ContextualAI/archangel_sft-kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 11 • 2
CLAIR and APO Data and Models for the paper "Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment" ContextualAI/ultrafeedback_clair_32k Viewer • Updated Sep 17, 2024 • 29.1k • 31 • 4 ContextualAI/ultrafeedback_rlaif_32k Viewer • Updated Aug 13, 2024 • 31.8k • 17 • 1 ContextualAI/ultrafeedback_rlaif-offpolicy_32k Viewer • Updated Aug 13, 2024 • 29.1k • 12 • 1 ContextualAI/ultrafeedback_stronger-preferred_32k Viewer • Updated Aug 13, 2024 • 29.1k • 8 • 1
Contextual-SQL Text-to-SQL models ContextualAI/ctx-bird-reward-250121 Text Generation • 33B • Updated 29 days ago • 7 • 1
Zephyr KTO Aligned models based on Zephyr-SFT from Table 2 and 3 in the KTO paper by Ethayarajh et al. (2024) (https://arxiv.org/pdf/2402.01306). ContextualAI/zephyr_sft_kto Text Generation • Updated May 5, 2024 • 2 • 1 ContextualAI/zephyr_sft_kto_unary Text Generation • Updated May 5, 2024 • 2 ContextualAI/zephyr_sft_dpo Text Generation • Updated May 5, 2024 • 2
LMUnit Models & Datasets for LMunit evaluation ContextualAI/LFQA Viewer • Updated 13 days ago • 260 • 83 ContextualAI/reward-bench-filtered Viewer • Updated 13 days ago • 2.99k • 54 ContextualAI/BiGGenBench Viewer • Updated 12 days ago • 3.2k • 59 ContextualAI/Flask Viewer • Updated 14 days ago • 2k • 64
Contextual-SQL Text-to-SQL models ContextualAI/ctx-bird-reward-250121 Text Generation • 33B • Updated 29 days ago • 7 • 1
Archangel Archangel is a suite of human feedback-aligned LLMs, released as part of the Human-Aware Loss Functions (HALOs) project by Ethayarajh et al. (2024). ContextualAI/archangel_sft_llama7b Text Generation • 7B • Updated Jan 11, 2024 • 5 • 1 ContextualAI/archangel_kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 5 • 2 ContextualAI/archangel_sft-kto_llama13b Text Generation • 13B • Updated Jan 11, 2024 • 1.45k • 3 ContextualAI/archangel_sft-kto_llama30b Text Generation • 33B • Updated Jan 11, 2024 • 11 • 2
Zephyr KTO Aligned models based on Zephyr-SFT from Table 2 and 3 in the KTO paper by Ethayarajh et al. (2024) (https://arxiv.org/pdf/2402.01306). ContextualAI/zephyr_sft_kto Text Generation • Updated May 5, 2024 • 2 • 1 ContextualAI/zephyr_sft_kto_unary Text Generation • Updated May 5, 2024 • 2 ContextualAI/zephyr_sft_dpo Text Generation • Updated May 5, 2024 • 2
CLAIR and APO Data and Models for the paper "Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment" ContextualAI/ultrafeedback_clair_32k Viewer • Updated Sep 17, 2024 • 29.1k • 31 • 4 ContextualAI/ultrafeedback_rlaif_32k Viewer • Updated Aug 13, 2024 • 31.8k • 17 • 1 ContextualAI/ultrafeedback_rlaif-offpolicy_32k Viewer • Updated Aug 13, 2024 • 29.1k • 12 • 1 ContextualAI/ultrafeedback_stronger-preferred_32k Viewer • Updated Aug 13, 2024 • 29.1k • 8 • 1