CRCA 1.5B Full Finetune

Overview

CR-CA (Causal Reasoning and Counterfactual Analysis) is a reasoning-focused stack that targets structured causal analysis, counterfactuals, and multi-step reasoning. This 1.5B model is a CR-CA reasoning-optimized causal language model based on the Qwen2 architecture (Qwen2ForCausalLM).

Model Details

  • Model type: qwen2
  • Architecture: Qwen2ForCausalLM
  • Hidden size: 1536
  • Layers: 28
  • Attention heads: 12 (KV heads: 2)
  • Max position embeddings: 32768
  • Vocab size: 151936
  • Dtype: float16

Training Summary

This model was produced via full finetuning for CR-CA reasoning. Training metadata is stored in training_args.bin.

Key training parameters:

  • Per-device batch size: 8
  • Gradient accumulation: 16
  • Epochs: 2
  • Learning rate: 5e-4
  • Precision: FP16
  • DeepSpeed config: training/deepspeed_zero2_1_5b.json
  • Scheduler: cosine
  • Warmup steps: 100
  • Save steps: 200

Training Data

The training data uses a prompt/response JSONL format:

{"prompt": "...", "response": "..."}

The dataset includes public reasoning data (e.g., GSM8K-style math word problems). This is used to strengthen multi-step reasoning, structured derivations, and final answer formatting.

Evaluation Report (Real-World Causal Tasks)

Evaluation was run on 2026-02-01 using GPT-4o-mini over 6 real-world causal tasks. Overall score: 48.3%.

Per-task scores:

  • Monetary Policy Counterfactual (US Macro 2025): 55/100
  • Tariff Pass-Through and Pricing (Beige Book + Firm Data): 55/100
  • Supply Chain Reroute Counterfactual (Port Disruption): 45/100
  • Inventory & Stockout Causal Impact (Retail): 25/100
  • Inflation Drivers (World Bank CPI Data): 65/100
  • Workforce Training Program (Labor Market Causal Impact): 45/100

Key strengths observed:

  • Clear task framing and attempt at counterfactual reasoning.
  • Some identification of confounders and causal factors.

Key limitations observed:

  • Inconsistent causal graphs and directional effects.
  • Weak counterfactual grounding and numerical reasoning errors.
  • Limited depth and rigor on confounder adjustment strategies.

Intended Use

For causal reasoning, counterfactual analysis, structured CR-CA reasoning prompts, and multi-step reasoning tasks.

Generation Settings

Default generation parameters are stored in generation_config.json:

  • do_sample: true
  • temperature: 0.7
  • top_p: 0.8
  • top_k: 20
  • repetition_penalty: 1.1

Limitations

  • Outputs should be validated for factual correctness.
  • The model may hallucinate causal claims without evidence.

License

Follow the base model and dataset licenses used for training. Add your explicit license here if required.

Downloads last month
162
Safetensors
Model size
2B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support

Model tree for Euroswarms/CR-CA

Base model

Qwen/Qwen2.5-1.5B
Finetuned
(1415)
this model