TY - GEN
T1 - Comparing optimistic and pessimistic constraint evaluation in shape-constrained symbolic regression
AU - Haider, Christian
AU - De França, Fabrício Olivetti
AU - Kronberger, Gabriel
AU - Burlacu, Bogdan
N1 - Funding Information:
The authors gratefully acknowledge support by the Christian Doppler Research Association and the Federal Ministry of Digital and Economic Affairs within the Josef Ressel Centre for Symbolic Regression.
Publisher Copyright:
© 2022 ACM.
PY - 2022/7/8
Y1 - 2022/7/8
N2 - Shape-constrained Symbolic Regression integrates prior knowledge about the function shape into the symbolic regression model. This can be used to enforce that the model has desired properties such as monotonicity, or convexity, among others. Shape-constrained Symbolic Regression can also help to create models with better extrapolation behavior and reduced sensitivity to noise. The constraint evaluation can be challenging because exact evaluation of constraints may require a search for the extrema of non-convex functions. Approximations via interval arithmetic allow to efficiently find bounds for the extrema of functions. However, interval arithmetic can lead to overly wide bounds and therefore produces a pessimistic estimation. Another possibility is to use sampling which underestimates the true range. Sampling therefore produces an optimistic estimation. In this paper we evaluate both methods and compare them on different problem instances. In particular we evaluate the sensitivity to noise and the extrapolation capabilities in combination with noise data. The results indicate that the optimistic approach works better for predicting out-of-domain points (extrapolation) and the pessimistic approach works better for high noise levels.
AB - Shape-constrained Symbolic Regression integrates prior knowledge about the function shape into the symbolic regression model. This can be used to enforce that the model has desired properties such as monotonicity, or convexity, among others. Shape-constrained Symbolic Regression can also help to create models with better extrapolation behavior and reduced sensitivity to noise. The constraint evaluation can be challenging because exact evaluation of constraints may require a search for the extrema of non-convex functions. Approximations via interval arithmetic allow to efficiently find bounds for the extrema of functions. However, interval arithmetic can lead to overly wide bounds and therefore produces a pessimistic estimation. Another possibility is to use sampling which underestimates the true range. Sampling therefore produces an optimistic estimation. In this paper we evaluate both methods and compare them on different problem instances. In particular we evaluate the sensitivity to noise and the extrapolation capabilities in combination with noise data. The results indicate that the optimistic approach works better for predicting out-of-domain points (extrapolation) and the pessimistic approach works better for high noise levels.
KW - prior knowledge
KW - shape constraints
KW - symbolic regression
UR - http://www.scopus.com/inward/record.url?scp=85135212080&partnerID=8YFLogxK
U2 - 10.1145/3512290.3528714
DO - 10.1145/3512290.3528714
M3 - Conference contribution
AN - SCOPUS:85135212080
T3 - GECCO 2022 - Proceedings of the 2022 Genetic and Evolutionary Computation Conference
SP - 938
EP - 945
BT - GECCO 2022 - Proceedings of the 2022 Genetic and Evolutionary Computation Conference
PB - Association for Computing Machinery, Inc
T2 - 2022 Genetic and Evolutionary Computation Conference, GECCO 2022
Y2 - 9 July 2022 through 13 July 2022
ER -