Evaluating the Arithmetic Capabilities of Neural Networks
An in-depth investigation into why neural networks struggle with simple arithmetic expressions. This project builds and evaluates various neural network architecturesโincluding FNNs, RNNs, attention-based RNNs, and Transformersโto assess their performance on arithmetic tasks.
Understanding Neural Networks' Arithmetic Limitations
Investigate why neural networks often struggle with simple arithmetic expressions and evaluate the capabilities of different architectures.
Build and test various neural network models on arithmetic problems, comparing their performance and understanding their limitations.
Comprehensive evaluation with statistical analysis, p-values, and detailed findings documented in research papers.
Exploring Different Approaches to Arithmetic Tasks
Basic neural network architecture with multiple layers. FNN2 demonstrated the best performance among all tested models.
Networks with loops that allow information persistence, suitable for sequence-based arithmetic operations.
LSTM-based RNN with attention mechanism to focus on relevant parts of the input sequence.
Modern transformer architecture with self-attention mechanisms for parallel processing.
Fine-tuned pre-trained models including Gemma (270M & 1B parameters) and Gemini via Vertex AI.
Experience the Neural Predictive Calculator
Try the FNN2 model right here! Enter an arithmetic expression and see how the neural network computes it.
Download and run the model on your own machine:
# Install dependencies
pip install -r requirements.txt
# Run the demo
python FNN/NPC-final.py
All required Python packages are listed in requirements.txt. The models will run on CPU.
The demonstration script allows you to input arithmetic expressions and see how the neural network computes the results.
If you have an NVIDIA Jetson device, use requirements_GPU.txt for GPU-accelerated inference.
Comprehensive Research Materials
๐ Post-Submission Updates: A new branch with minor adjustments, MIT license, and documentation corrections is available at github.com/AntonStantan/matura/tree/post-submission
Detailed results and analysis of the research, including performance comparisons and insights.
View PDF โComplete description of research methods, experimental setup, and evaluation criteria.
View PDF โTransparent documentation of AI assistance used throughout the project.
View PDF โComprehensive review of related work and research (Zwischenprodukt).
View PDF โRaw evaluation data and findings table with statistical analysis.
View Excel โDetailed documentation of the work process (Arbeitsprozess).
View Folder โOrganized and well-documented codebase