Penalty Functions for Genetic Programming Algorithms

0 downloads 167 Views 367KB Size Report
Talk Layout. 1 The Symbolic Regression Problem. 2 Model Selection. 3 Experimental Results. C.E. Borges. Penalty Function
Penalty Functions for Genetic Programming Algorithms José L. Montaña,1 César L. Alonso,2 Cruz E. Borges,1 and Javier de la Dehesa.1 [email protected], [email protected], [email protected], [email protected]. Dept. Matemáticas, Estadística y Computación1

Centro de Inteligencia Artificial2

11th ICCSA 2011, 20-23 June 2011, Santander, Spain.

This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 License.

Talk Layout

C.E. Borges

1

The Symbolic Regression Problem

2

Model Selection

3

Experimental Results

Penalty Functions for Genetic Programming Algorithms

2/12

Talk Layout

C.E. Borges

1

The Symbolic Regression Problem

2

Model Selection

3

Experimental Results

Penalty Functions for Genetic Programming Algorithms

2/12

Talk Layout

C.E. Borges

1

The Symbolic Regression Problem

2

Model Selection

3

Experimental Results

Penalty Functions for Genetic Programming Algorithms

2/12

Regression Problems. Regressions Problems. Given a set of points and a space of functions, find the best function that fits the set of points. f (x)

f (x)

x x1 x2 x3 4

x0

x5

x7

x8 x x10 9

x6

x C.E. Borges

Penalty Functions for Genetic Programming Algorithms

3/12

Regression Problems. Regressions Problems. Given a set of points and a space of functions, find the best function that fits the set of points. f (x)

g(x)

x C.E. Borges

Penalty Functions for Genetic Programming Algorithms

3/12

How we solve it?

When the Function Set. . . is Known: Interpolation, Least Squares. is Unknown: Symbolic Regression.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

4/12

How we solve it?

When the Function Set. . . is Known: Interpolation, Least Squares. is Unknown: Symbolic Regression.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

4/12

How we solve it?

When the Function Set. . . is Known: Interpolation, Least Squares. is Unknown: Symbolic Regression.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

4/12

Symbolic Regression Problems

Symbolic Regression. Looks for a symbolic expression of the regressor. Looks for a computer program that evaluate the regressor. Genetic Programming.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

5/12

Symbolic Regression Problems

Symbolic Regression. Looks for a symbolic expression of the regressor. Looks for a computer program that evaluate the regressor. Genetic Programming.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

5/12

Symbolic Regression Problems

Symbolic Regression. Looks for a symbolic expression of the regressor. Looks for a computer program that evaluate the regressor. Genetic Programming.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

5/12

Genetic Programming with Trees.

Figure: Example of a computer program, encoded as a tree, that evaluate the function f := x 4 + x 3 + x 2 + x.

f ≡

+ +

+ ∗ ∗

x x

∗ x



x x

C.E. Borges

∗ x

x

∗ x

Penalty Functions for Genetic Programming Algorithms

x

x

6/12

Model Selection: Vapnik–Chervonenkis Dimension.

Ockham’s Razor. The simplest explanation is most likely the correct one.

Idea. Use a model fitness according to their empirical error plus a penalization based on their complexity.

Classification capacity. Cardinality of the largest set of points that the model can shatter.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

7/12

Model Selection: Vapnik–Chervonenkis Dimension.

Ockham’s Razor. The simplest explanation is most likely the correct one.

Idea. Use a model fitness according to their empirical error plus a penalization based on their complexity.

Classification capacity. Cardinality of the largest set of points that the model can shatter.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

7/12

Model Selection: Vapnik–Chervonenkis Dimension.

Ockham’s Razor. The simplest explanation is most likely the correct one.

Idea. Use a model fitness according to their empirical error plus a penalization based on their complexity.

Classification capacity. Cardinality of the largest set of points that the model can shatter.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

7/12

Model Selection: Vapnik–Chervonenkis Dimension.

Ockham’s Razor. The simplest explanation is most likely the correct one.

Idea. Use a model fitness according to their empirical error plus a penalization based on their complexity.

Classification capacity. Cardinality of the largest set of points that the model can shatter.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

7/12

Model Selection: Vapnik–Chervonenkis Dimension.

Ockham’s Razor. The simplest explanation is most likely the correct one.

Idea. Use a model fitness according to their empirical error plus a penalization based on their complexity.

Classification capacity. Cardinality of the largest set of points that the model can shatter.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

7/12

Model Selection: Vapnik–Chervonenkis Dimension.

Ockham’s Razor. The simplest explanation is most likely the correct one.

Idea. Use a model fitness according to their empirical error plus a penalization based on their complexity.

Classification capacity. Cardinality of the largest set of points that the model can shatter.

Theorem. The Vapnik–Chervonenkis Dimension of a GP-tree is polynomial in the non-scalar height, the number of analytic algebraic nodes and the number of constants and variables. C.E. Borges

Penalty Functions for Genetic Programming Algorithms

7/12

Experimental Results i.

Experimental Setup. Random polynomials of bounded degree. 300 repetitions: 30 learning points with noise. 200 validation points without noise.

Boxplots of the MSE over the validation set of points. Hypothesis Testing.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

8/12

Experimental Results i.

Experimental Setup. Random polynomials of bounded degree. 300 repetitions: 30 learning points with noise. 200 validation points without noise.

Boxplots of the MSE over the validation set of points. Hypothesis Testing.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

8/12

Experimental Results i.

Experimental Setup. Random polynomials of bounded degree. 300 repetitions: 30 learning points with noise. 200 validation points without noise.

Boxplots of the MSE over the validation set of points. Hypothesis Testing.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

8/12

Experimental Results i.

Experimental Setup. Random polynomials of bounded degree. 300 repetitions: 30 learning points with noise. 200 validation points without noise.

Boxplots of the MSE over the validation set of points. Hypothesis Testing.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

8/12

Experimental Results i.

Experimental Setup. Random polynomials of bounded degree. 300 repetitions: 30 learning points with noise. 200 validation points without noise.

Boxplots of the MSE over the validation set of points. Hypothesis Testing.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

8/12

Experimental Results i.

Experimental Setup. Random polynomials of bounded degree. 300 repetitions: 30 learning points with noise. 200 validation points without noise.

Boxplots of the MSE over the validation set of points. Hypothesis Testing.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

8/12

Experimental Results ii.

Regularization Techniques Used MSE (no regularization). AIC. BIC. SRM.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

9/12

Experimental Results ii.

Regularization Techniques Used MSE (no regularization). AIC. BIC. SRM.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

9/12

Experimental Results ii.

Regularization Techniques Used MSE (no regularization). AIC. BIC. SRM.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

9/12

Experimental Results ii.

Regularization Techniques Used MSE (no regularization). AIC. BIC. SRM.

C.E. Borges

Penalty Functions for Genetic Programming Algorithms

9/12

Experimental Results iii.

0.0

0.1

0.2

0.3

0.4

0.5

PR n [X]

MSE C.E. Borges

AIC

BIC

SRM

Penalty Functions for Genetic Programming Algorithms

10/12

Experimental Results iv.

Hypothesis Testing PR n [X ] MSE AIC BIC SRM

C.E. Borges

MSE 1 8,81 · 10−3 5,82 · 10−3 3,44 · 10−3

AIC 0,94 1 0,91 7,02 · 10−2

BIC 0,93 0,73 1 7,16 · 10−2

Penalty Functions for Genetic Programming Algorithms

SRM 1 0,8 0,94 1

11/12

Penalty Functions for Genetic Programming Algorithms José L. Montaña,1 César L. Alonso,2 Cruz E. Borges,1 and Javier de la Dehesa.1 [email protected], [email protected], [email protected], [email protected]. Dept. Matemáticas, Estadística y Computación1

Centro de Inteligencia Artificial2

11th ICCSA 2011, 20-23 June 2011, Santander, Spain.

This work is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 License.