Training, evaluation datasets and model outputs for the arXiv 2026 preprint Controllable Reasoning Models are Private Thinkers
Haritz Puerto
haritzpuerto
AI & ML interests
Reasoning in LLMs, trustworthy AI, agents
Recent Activity
updated
a model 2 days ago
haritzpuerto/microsoft-Phi-4-14B-IF-RT updated
a model 2 days ago
haritzpuerto/microsoft-Phi-4-14B-IF-FA updated
a model 2 days ago
haritzpuerto/microsoft-Phi-4-14B-IF-Avg Organizations
DCoT
Models from the ACL 2025 paper "Fine-Tuning on Diverse Reasoning Chains Drives Within-Inference CoT Refinement in LLMs"
"
-
Fine-Tuning with Divergent Chains of Thought Boosts Reasoning Through Self-Correction in Language Models
Paper ⢠2407.03181 ⢠Published ⢠1 -
haritzpuerto/LLaMA2-7B-dcot
Text Generation ⢠Updated ⢠2 ⢠2 -
haritzpuerto/LLaMA2-13B-dcot
Text Generation ⢠Updated ⢠1 -
haritzpuerto/LLaMA2-70B-dcot
Text Generation ⢠Updated ⢠1
āļøš§ š Controllable Reasoning Models - Checkpoints
Training dataset and LoRA checkpoints for the arXiv 2026 preprint Controllable Reasoning Models are Private Thinkers
MIA-Pile
Samples used for the NAACL 2025 Findings paper: "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models."
-
haritzpuerto/the_pile_arxiv_1k_sample
Viewer ⢠Updated ⢠5.84k ⢠18 -
haritzpuerto/the_pile_arxiv_50k_sample
Viewer ⢠Updated ⢠54.8k ⢠63 -
parameterlab/scaling_mia_the_pile_00_arxiv
Viewer ⢠Updated ⢠83.9k ⢠61 -
parameterlab/scaling_mia_the_pile_00_wiki
Viewer ⢠Updated ⢠599k ⢠54
āļøš§ š Controllable Reasoning Models - Datasets
Training, evaluation datasets and model outputs for the arXiv 2026 preprint Controllable Reasoning Models are Private Thinkers
āļøš§ š Controllable Reasoning Models - Checkpoints
Training dataset and LoRA checkpoints for the arXiv 2026 preprint Controllable Reasoning Models are Private Thinkers
DCoT
Models from the ACL 2025 paper "Fine-Tuning on Diverse Reasoning Chains Drives Within-Inference CoT Refinement in LLMs"
"
-
Fine-Tuning with Divergent Chains of Thought Boosts Reasoning Through Self-Correction in Language Models
Paper ⢠2407.03181 ⢠Published ⢠1 -
haritzpuerto/LLaMA2-7B-dcot
Text Generation ⢠Updated ⢠2 ⢠2 -
haritzpuerto/LLaMA2-13B-dcot
Text Generation ⢠Updated ⢠1 -
haritzpuerto/LLaMA2-70B-dcot
Text Generation ⢠Updated ⢠1
MIA-Pile
Samples used for the NAACL 2025 Findings paper: "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models."
-
haritzpuerto/the_pile_arxiv_1k_sample
Viewer ⢠Updated ⢠5.84k ⢠18 -
haritzpuerto/the_pile_arxiv_50k_sample
Viewer ⢠Updated ⢠54.8k ⢠63 -
parameterlab/scaling_mia_the_pile_00_arxiv
Viewer ⢠Updated ⢠83.9k ⢠61 -
parameterlab/scaling_mia_the_pile_00_wiki
Viewer ⢠Updated ⢠599k ⢠54