{ "embed_dim": 1024, "init_logit_bias": -10, "quick_gelu": true, "text_cfg": { "context_length": 77, "embed_cls": true, "embed_unc": true, "heads": 16, "init_unc_bias": -10, "layers": 24, "no_causal_mask": true, "pool_type": "last", "proj_bias": true, "vocab_size": 49408, "width": 1024 }, "vision_cfg": { "embed_unc": true, "head_width": 80, "image_size": 224, "init_unc_bias": -10, "layers": 32, "patch_size": 14, "width": 1280 } }