FinChain: A Symbolic Benchmark for Verifiable Chain-of-Thought Financial Reasoning
Abstract
A new benchmark called FinChain evaluates multi-step symbolic reasoning in financial tasks with a focus on intermediate reasoning steps, introducing ChainEval as a metric for assessing both final answers and reasoning processes.
Multi-step symbolic reasoning is critical for advancing downstream performance on financial tasks. Yet, benchmarks for systematically evaluating this capability are lacking. Existing datasets like FinQA and ConvFinQA supervise only final numerical answers, without assessing intermediate reasoning steps. To address this, we introduce FinChain, the first symbolic benchmark designed for verifiable Chain-of- Thought (CoT) financial reasoning. Spanning 54 topics across 12 financial domains, Fin- Chain offers five parameterized templates per topic, each varying in reasoning complexity and domain expertise required. Each dataset instance includes an executable Python trace, enabling automatic generation of extensive training data and easy adaptation to other domains. We also introduce ChainEval, a new metric for automatic evaluation of both final answers and intermediate reasoning. Benchmarking 30 LLMs on our dataset, we find that even state-of-the-art models have considerable room for improvement in multi-step financial reasoning. All templates and evaluation metrics for FinChain are available at https: //github.com/mbzuai-nlp/finchain.
Community
๐ FinChain: A Symbolic Benchmark for Verifiable Chain-of-Thought Financial Reasoning
We introduce FinChain, a benchmark designed to evaluate and improve the reasoning abilities of LLMs in financial tasks. Unlike prior work that relies on final-answer supervision, FinChain provides symbolic, executable chain-of-thought traces across 54 financial topics.
It enables fine-grained, verifiable reasoning supervision and allows for the development of interpretable, trustworthy financial agents.
๐ Dataset & code: github.com/mbzuai-nlp/finchain
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Socratic-PRMBench: Benchmarking Process Reward Models with Systematic Reasoning Patterns (2025)
- FlashThink: An Early Exit Method For Efficient Reasoning (2025)
- BizFinBench: A Business-Driven Real-World Financial Benchmark for Evaluating LLMs (2025)
- RV-Syn: Rational and Verifiable Mathematical Reasoning Data Synthesis based on Structured Function Library (2025)
- One Missing Piece for Open-Source Reasoning Models: A Dataset to Mitigate Cold-Starting Short CoT LLMs in RL (2025)
- Learning to Reason Over Time: Timeline Self-Reflection for Improved Temporal Reasoning in Language Models (2025)
- RealSafe-R1: Safety-Aligned DeepSeek-R1 without Compromising Reasoning Capability (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper