---
license: apache-2.0
language:
- en
pipeline_tag: text-generation
base_model: SVECTOR/Theta-35
tags:
- chat
- reasoning
library_name: transformers
---
# Theta-35
## Introduction
Theta-35 is the advanced reasoning model in the Theta series by SVECTOR. Compared with conventional instruction-tuned models, Theta-35, which specializes in complex thinking and reasoning, achieves significantly enhanced performance in downstream tasks, particularly for challenging problems requiring deep logical analysis and multistep reasoning.
**This repo contains the Theta-35 model**, which has the following features:
- Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
- Architecture: Transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
- Number of Parameters: 33B
- Number of Parameters (Non-Embedding): 33B
- Number of Layers: 64
- Number of Attention Heads (GQA): 40 for Q and 8 for KV
- Context Length: Full 131,072 tokens
- Sliding Window: 32,768 tokens
**Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying Theta models.
For more details, please refer to our [documentation](https://www.svector.co.in/models/theta-35).
## Requirements
Theta-35 requires the latest version of Hugging Face `transformers`. We advise you to use version 4.43.1 or newer.
With older versions of transformers, you may encounter the following error:
```
KeyError: 'theta'
```
## Quickstart
Here is a code snippet showing how to load the tokenizer and model, and how to generate content:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer directly
model_name = "SVECTOR-CORPORATION/Theta-35"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
# Prepare prompt
prompt = "How many planets are in our solar system? Explain your reasoning."
messages = [
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True # This will automatically add "" tag
)
# Generate response
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=32768,
temperature=0.6,
top_p=0.95,
top_k=30
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
# Decode and print response
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)
```
### Usage Guidelines
To achieve optimal performance with Theta-35, we recommend the following settings:
1. **Enforce Thoughtful Output**: Ensure the model starts with "\\n" to promote step-by-step thinking, which enhances output quality. If you use `apply_chat_template` and set `add_generation_prompt=True`, this is automatically implemented.
2. **Sampling Parameters**:
- Use Temperature=0.6 and TopP=0.95 instead of Greedy decoding to avoid repetitions.
- Use TopK between 20 and 40 to filter out rare token occurrences while maintaining diversity.
3. **Standardize Output Format**: We recommend using prompts to standardize model outputs when benchmarking.
- **Math Problems**: Include "Please reason step by step, and put your final answer within \boxed{}." in the prompt.
- **Multiple-Choice Questions**: Add "Please show your choice in the `answer` field with only the choice letter, e.g.,`\"answer\": \"C\"`." to the prompt.
4. **Handle Long Inputs**: For inputs exceeding 32,768 tokens, enable sliding window attention to improve the model's ability to process long sequences efficiently.
For supported frameworks, you could add the following to `config.json` to enable extended context handling:
```json
{
...,
"use_sliding_window": true,
"sliding_window": 32768
}
```
## Evaluation & Performance
Theta-35 demonstrates exceptional performance across various reasoning tasks, including:
- Mathematical reasoning
- Logical deduction
- Multi-step problem solving
- Code understanding and generation
- Scientific reasoning
Detailed evaluation results are reported in our [documentation](https://www.svector.co.in/models/theta-35).
## Citation
If you find our work helpful, feel free to give us a cite.
```
@misc{theta35,
title = {Theta-35: Advanced Reasoning in Large Language Models},
url = {https://www.svector.co.in/models/theta-35},
author = {SVECTOR Team},
month = {March},
year = {2025}
}
@article{theta,
title={Theta Technical Report},
author={SVECTOR Research Team},
year={2025}
}
```