--- base_model: - Nitral-AI/Community_Request-02-12B - AlexCuadron/chai-ddpo library_name: transformers tags: - mergekit - merge --- # output This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Task Arithmetic](https://arxiv.org/abs/2212.04089) merge method using [AlexCuadron/chai-ddpo](https://huggingface.co/AlexCuadron/chai-ddpo) as a base. ### Models Merged The following models were included in the merge: * [Nitral-AI/Community_Request-02-12B](https://huggingface.co/Nitral-AI/Community_Request-02-12B) ### Configuration The following YAML configuration was used to produce this model: ```yaml # MergeKit Configuration for merging alignment-DPO with roleplay capabilities merge_method: task_arithmetic # Define the base model (your alignment model) base_model: model: AlexCuadron/chai-ddpo # Path to your DPO-trained alignment model # If your model is in Hugging Face format, you could use: model_type: huggingface # If it's a local model, just use the path: # model: /path/to/alex_dpo # Models to merge with the base model models: - model: Nitral-AI/Community_Request-02-12B # Modify if needed: model_type: huggingface parameters: weight: 0.3 # Adjust between 0.2-0.4 based on how much roleplay vs alignment you want # Output settings dtype: bfloat16 # Or bfloat16 if your GPU supports it output_dir: ./aligned_roleplay_model # Optional: include parameters you want to skip or include specifically # parameters_settings: # include: [".*"] # Include all parameters, or specify patterns to include # exclude: [] # No exclusions, or specify patterns to exclude # Add this if you want to save disk space output_optimizations: safetensors: true # Save in safetensors format (more secure and efficient) # set to false if you want to keep intermediate models skip_saving_intermediate_models: true ```