Edit model card

final_merge_medical2

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536 as a base.

Models Merged

The following models were included in the merge:

  • ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
  • ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785

Configuration

The following YAML configuration was used to produce this model:

base_model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
dtype: bfloat16
merge_method: dare_ties
parameters:
  int8_mask: 1.0
  normalize: 1.0
slices:
- sources:
  - layer_range: [0, 4]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 1.0
      weight: 0.17884011663118993
  - layer_range: [0, 4]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 0.8890748672494518
      weight: 0.8456004469240137
  - layer_range: [0, 4]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 0.8353965535147524
      weight: 0.39636256675038484
- sources:
  - layer_range: [4, 8]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 1.0
      weight: 1.020532319785238
  - layer_range: [4, 8]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.40553667157598416
  - layer_range: [4, 8]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 0.8955950669638045
      weight: 0.1256681933865999
- sources:
  - layer_range: [8, 12]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 0.8698472775356287
      weight: 0.712509376634506
  - layer_range: [8, 12]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.09302230163177483
  - layer_range: [8, 12]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 1.0
      weight: -0.05614416786806742
- sources:
  - layer_range: [12, 16]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 1.0
      weight: 0.39694054142633234
  - layer_range: [12, 16]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.2626803763724651
  - layer_range: [12, 16]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 0.9657199816167481
      weight: 0.2153132357599829
- sources:
  - layer_range: [16, 20]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 0.8289285693717532
      weight: 0.34481743191831266
  - layer_range: [16, 20]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 0.7115155293484904
      weight: 0.03210636463921337
  - layer_range: [16, 20]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 1.0
      weight: 0.23727760414855137
- sources:
  - layer_range: [20, 24]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 0.9992577736634561
      weight: 0.33608731549782594
  - layer_range: [20, 24]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.2672857028656318
  - layer_range: [20, 24]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 1.0
      weight: 0.4567325333292645
- sources:
  - layer_range: [24, 28]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 1.0
      weight: -0.061310476755757834
  - layer_range: [24, 28]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.23902257193439966
  - layer_range: [24, 28]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 0.5322311277242968
      weight: 0.37656791069123563
- sources:
  - layer_range: [28, 32]
    model: ./evol_merge_storage_medical2/input_models/BioMedGPT-LM-7B_1321966785
    parameters:
      density: 1.0
      weight: -0.018284769401157808
  - layer_range: [28, 32]
    model: ./evol_merge_storage_medical2/input_models/Llama-2-7b-hf_1207016536
    parameters:
      density: 1.0
      weight: 0.5831277146552916
  - layer_range: [28, 32]
    model: ./evol_merge_storage_medical2/input_models/meditron-7b_535713315
    parameters:
      density: 0.8437643378550103
      weight: 0.1599917740219515
Downloads last month
11
Safetensors
Model size
6.74B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ssktora/llama2_biomed_meditron

Adapters
1 model