# MergeKit Configuration for merging alignment-DPO with roleplay capabilities merge_method: task_arithmetic # Define the base model (your alignment model) base_model: model: AlexCuadron/chai-ddpo # Path to your DPO-trained alignment model # If your model is in Hugging Face format, you could use: model_type: huggingface # If it's a local model, just use the path: # model: /path/to/alex_dpo # Models to merge with the base model models: - model: Nitral-AI/Community_Request-02-12B # Modify if needed: model_type: huggingface parameters: weight: 0.3 # Adjust between 0.2-0.4 based on how much roleplay vs alignment you want # Output settings dtype: bfloat16 # Or bfloat16 if your GPU supports it output_dir: ./aligned_roleplay_model # Optional: include parameters you want to skip or include specifically # parameters_settings: # include: [".*"] # Include all parameters, or specify patterns to include # exclude: [] # No exclusions, or specify patterns to exclude # Add this if you want to save disk space output_optimizations: safetensors: true # Save in safetensors format (more secure and efficient) # set to false if you want to keep intermediate models skip_saving_intermediate_models: true