This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the TIES (TrIm, Elect, and Merge) merging method, with medalpaca-7b as a base.

  • TIES - TrIm, Elect, and Merge (TIES) is a three-step method for merging models. First, redundant parameters are trimmed, then conflicting signs are resolved into an aggregated vector, and finally the parameters whose signs are the same as the aggregate sign are averaged. This method takes into account that some values (redundant and sign disagreement) can degrade performance in the merged model.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

base_model: medalpaca-7b
dtype: bfloat16
merge_method: ties
modules:
  default:
    slices:
    - sources:
      - layer_range: [0, 32]
        model: medalpaca-sft
        parameters:
          density: 0.6
          weight: 0.3
      - layer_range: [0, 32]
        model: medalpaca-kd
        parameters:
          density: 0.6
          weight: 0.7
      - layer_range: [0, 32]
        model: medalpaca-7b

Review all model metrics benchmark via Benchmark Document Preview.### Configuration

Downloads last month
23
Safetensors
Model size
7B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MedSwin/MedSwin-Merged-TIES-KD-0.7

Datasets used to train MedSwin/MedSwin-Merged-TIES-KD-0.7

Space using MedSwin/MedSwin-Merged-TIES-KD-0.7 1

Collection including MedSwin/MedSwin-Merged-TIES-KD-0.7