Replies: 1 comment
-
There could be several reasons:
Yes, when we use a fully transparent benchmarking process like srbench, we often see differences from what authors report in their papers that introduce new methodologies. often it comes down to differences in experiment design that may favor one method or another. part of the motivation for this benchmark is to alleviate those issues. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello guys,
I have a question regarding the performance of the AI-Feynman algorithm as shown in the SR Benchmark results, particularly in comparison with the original AI-Feynman paper. In the results presented on the SR Benchmark website (link to results), the AI-Feynman algorithm finds a model symbolically equivalent to the ground-truth process on the Feynman dataset with a frequency of around 55%.
However, in the original AI-Feynman paper (AI Feynman: A physics-inspired method for symbolic regression, link to arXiv), the authors report being able to discover 118 out of 120 equations, which is close to 100%.
The difference between these two figures—55% in the benchmark versus nearly 100% in the original paper—seems quite significant, and I am unsure what might be causing this discrepancy.
Potential Questions:
Beta Was this translation helpful? Give feedback.
All reactions