YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
Combined Task Vector Model
This model was created by combining task vectors from multiple fine-tuned models.
Task Vector Computation
t_1 = TaskVector("meta-llama/Llama-2-7b-chat-hf", "coastalcph/Llama-2-7b-chat-gsm8k_bs8_2e-4")
t_2 = TaskVector("meta-llama/Llama-2-7b-chat-hf", "coastalcph/Llama-2-7b-chat-pv-prompts-non-evil_5e-5")
t_combined = 1.0 * t_1 + 8.0 * t_2 - 8.0 * t_3
new_model = t_combined.apply_to("meta-llama/Llama-2-7b-chat-hf", scaling_coef=1.0)
Models Used
- Base Model: https://huggingface.co/meta-llama/Llama-2-7b-chat-hf
- Fine-tuned Model 1: https://huggingface.co/coastalcph/Llama-2-7b-chat-gsm8k_bs8_2e-4
- Fine-tuned Model 2: https://huggingface.co/coastalcph/Llama-2-7b-chat-pv-prompts-non-evil_5e-5
Technical Details
- Creation Script Git Hash: d0db42d73be516ec04f0ecdc8003189e98b5f722
- Task Vector Method: Additive combination
- Args: { "pretrained_model": "meta-llama/Llama-2-7b-chat-hf", "finetuned_model1": "coastalcph/Llama-2-7b-chat-gsm8k_bs8_2e-4", "finetuned_model2": "coastalcph/Llama-2-7b-chat-pv-prompts-non-evil_5e-5", "finetuned_model3": "coastalcph/Llama-2-7b-chat-pv-prompts-evil_5e-5", "output_model_name": "coastalcph/Llama-2-7b-chat-1t_gsm8k-8t_diff_pv_evil_5e-5", "output_dir": "/projects/nlp/data/constanzam/weight-interp/task-vectors/math_non_sycophant_12Aug", "scaling_coef": 1.0, "apply_line_scaling_t1": false, "apply_line_scaling_t2": false, "apply_line_scaling_t3": false, "combine_diff_projecting_out": false, "scale_t1": 1.0, "scale_t2": 8.0, "scale_t3": 8.0 }
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support