qwen-llama-pairs
Collection
12 items
•
Updated
This model is a result of parameter averaging (Model Soup) across 2 models.
The following models were included in the merge:
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("moe-dtoks/llama_onto_qwen_lambda-0")
tokenizer = AutoTokenizer.from_pretrained("moe-dtoks/llama_onto_qwen_lambda-0")