源码:https://github.com/cheesezhe/Coursera-Machine-Learning-Exercise/tree/master/ex5

Introduction:

In this exercise, you will implement regularized linear regression and use it to study models with different bias-variance properties.

1. Regularized Linear Regression

(1). Implement regularized linear regression to predict the amount of water flowing out of a dam using the change of water level in a reservoir;

(2).Go through some diagnostics of debugging learning algorithms and examine the effects of bias v.s. variance;

1.1 Visualizing the dataset

The dataset contains historical records on the change in the water level, x, and the amount of water flowing out of the dam,y.

This dataset is divided into three parts:

  • A training set that your model will learn on:X,y
  • A cross validation set for determining the regularization parameter: Xval, yval
  • A test set for evaluating performance:Xtest, ytest

(1) plot the training data:

Next:

(1) Implement linear regression and use that to fit a straight line to the data and plot learning curves;

(2) Implement polynomial regression to find a better fit to the data;

1.2 Regularized linear regression cost function

Recall that the regularized linear regression has the following cost function:

where λ  is a regularization parameter which controls the degree of regularization(thus, help preventing overfitting). The regularization term puts a penalty on the overal cost J. As the magnitudes of the model paprameters θj increase, the penalty increase as well. Note that you should not regularize the θ0 term.

1.3 Regularized linear regression gradient

Correspondingly, the partial derivative of regularized linear regression's cost for θj is defined as:

the file linearRegCostFunction.m:

function [J, grad] = linearRegCostFunction(X, y, theta, lambda)
%LINEARREGCOSTFUNCTION Compute cost and gradient for regularized linear
%regression with multiple variables
% [J, grad] = LINEARREGCOSTFUNCTION(X, y, theta, lambda) computes the
% cost of using theta as the parameter for linear regression to fit the
% data points in X and y. Returns the cost in J and the gradient in grad % Initialize some useful values
m = length(y); % number of training examples % You need to return the following variables correctly
J = 0;
grad = zeros(size(theta)); % ====================== YOUR CODE HERE ======================
% Instructions: Compute the cost and gradient of regularized linear
% regression for a particular choice of theta.
%
% You should set J to the cost and grad to the gradient.
% %cost function
hx = X*theta;
J = (1/(2*m))*sumsq(hx-y);
reg = lambda/(2*m)*sumsq(theta(2:end));
J = J + reg;
%gradient
% Part 1.3
grad = (X' * (hx-y)) / m;
grad(2:end) += lambda * theta(2:end) / m;
% ========================================================================= grad = grad(:); end

1.4 Fitting linear regression

Once your cost function and gradient are working correctly, the next part is to run the code in trainLinearReg.m to compute the optimal values of θ. This training function uses fmincg to optimize the cost function:

function [theta] = trainLinearReg(X, y, lambda)
%TRAINLINEARREG Trains linear regression given a dataset (X, y) and a
%regularization parameter lambda
% [theta] = TRAINLINEARREG (X, y, lambda) trains linear regression using
% the dataset (X, y) and regularization parameter lambda. Returns the
% trained parameters theta.
% % Initialize Theta
initial_theta = zeros(size(X, 2), 1); % Create "short hand" for the cost function to be minimized
costFunction = @(t) linearRegCostFunction(X, y, t, lambda); % Now, costFunction is a function that takes in only one argument
options = optimset('MaxIter', 200, 'GradObj', 'on'); % Minimize using fmincg
theta = fmincg(costFunction, initial_theta, options); end

In this part, we set regularization parameter λ to zero. Because our current implementation of linear regression is trying to fit a 2-dimensional θ, regularization will not be incredibly helpful for a θ of such low dimension. In the later parts of the exercise, you will be using polynomial regression with regularization.

Finally we will plot the best fit line, resulting in an image similar to Figure 2.

While visualizing the best fit as shown in one possible way to debug your learing algorithm, it is not always easy to visualize the data and model. In the next section, you will implement a function to generate learing curves that can help you debug your learining algorithm even if it is not easy to visualize the data.

2 Bias-variance

An important concept in machine learning is the bias-vairance tradeoff.Models with high bias are not complex enough for the data and tend to underfit,  while models with high variance overfit to the training data.

In this part, we will plot training and test errors on a learning curve to diagnose bias-variance problems.

2.1 Learning curves

A learning curve plots training and cross validation error as a function of training set size. learningCurve.m returns a vector of errors for the training set and cross validation set.

To plot the learning curve, we need a training and cross validation set error for different training set sizes. To obtain different training set sizes, we should use different subsets of the original training set X.

We can use the trainLinearReg function to find  the θ parameters. Note that the lambda is passed as a parameter to the learningCurve function. After learning the θ parameters, we should compute the error on the training and cross validation sets. Recall that the training error for a dataset is defined as:

In particular, note that the training error does not include the regularization term. One way to compute the training error is to use exsiting cost function and set λ to 0 only when using it to compute the training error and cross validation error. When we are computing the training set error, make sure you compute it on the training subset instead of the entire training set. However, for the cross validation error, we should compute it over the entire validation set.

When we are finished, we will see a plot similar to Figure 3.

In Figure 3, we can observe that both the training error and cross validation error are high when the number of training examples is increased. This reflects a high bias problem in the model -- the linear regression is too simple and is unable to fit our dataset well. In the next section, you will implement polynomial regression to fit a better  model for this dataset.

function [error_train, error_val] = ...
learningCurve(X, y, Xval, yval, lambda)
%LEARNINGCURVE Generates the train and cross validation set errors needed
%to plot a learning curve
% [error_train, error_val] = ...
% LEARNINGCURVE(X, y, Xval, yval, lambda) returns the train and
% cross validation set errors for a learning curve. In particular,
% it returns two vectors of the same length - error_train and
% error_val. Then, error_train(i) contains the training error for
% i examples (and similarly for error_val(i)).
%
% In this function, you will compute the train and test errors for
% dataset sizes from 1 up to m. In practice, when working with larger
% datasets, you might want to do this in larger intervals.
% % Number of training examples
m = size(X, 1); % You need to return these values correctly
error_train = zeros(m, 1);
error_val = zeros(m, 1); % ====================== YOUR CODE HERE ======================
% Instructions: Fill in this function to return training errors in
% error_train and the cross validation errors in error_val.
% i.e., error_train(i) and
% error_val(i) should give you the errors
% obtained after training on i examples.
%
% Note: You should evaluate the training error on the first i training
% examples (i.e., X(1:i, :) and y(1:i)).
%
% For the cross-validation error, you should instead evaluate on
% the _entire_ cross validation set (Xval and yval).
%
% Note: If you are using your cost function (linearRegCostFunction)
% to compute the training and cross validation error, you should
% call the function with the lambda argument set to 0.
% Do note that you will still need to use lambda when running
% the training to obtain the theta parameters.
%
% Hint: You can loop over the examples with the following:
%
% for i = 1:m
% % Compute train/cross validation errors using training examples
% % X(1:i, :) and y(1:i), storing the result in
% % error_train(i) and error_val(i)
% ....
%
% end
% % ---------------------- Sample Solution ----------------------
for i = 1:m,
[theta] = trainLinearReg(X(1:i,:),y(1:i),lambda);
error_train(i) = linearRegCostFunction(X(1:i,:),y(1:i),theta,0);
error_val(i) = linearRegCostFunction(Xval,yval,theta,0);
end;
% ------------------------------------------------------------- % ========================================================================= end

3 Polynomial regression

The problem with our linear model was that it was too simple for the data and resulted in underfitting(high bias). In this part of the exercise, we will address this problem by adding more features.

Our hypothesis has the form:

aaarticlea/png;base64,iVBORw0KGgoAAAANSUhEUgAAA10AAABSCAIAAAAgpylAAAAgAElEQVR4nO2dd0ATyfv/hybnKYqKDVE80cNCsYFgV1AEFRTbnZ6CZ8E7PUURgY8NTgSxoODZGyqiKIIioKKAqKCgINIUpAUhlAQSCOllf3/M77Pf/SQhJCGAdzevv7KTzezsZnf2Pc/zzDMAQyAQCAQCgUAgMAx0dQMQCAQCgUAgEN8ESBciEAgEAoFAIDAM6UIEAoFAIBAIBATpQgQCgUAgEAgEhiFdiEAgEAgEAoGAIF2IQCAQCAQCgcAwpAsRCAQCgUAgEBCkCxEIBAKBQCAQGIZ0IQKBQCC6kJycnIyMjK5uBQKB+P8gXYhAIBCILiAtLe3q1avJyckPHjxwcnLatm1bS0tLVzcKgfi3g3QhAoFAIDobMpm8YcMGgUAAN9+/f6+pqXnr1q2ubRUCgfjWdeHz589zc3MV+gmbzY6Kiuqg9iA6goKCgtevX3d1K/753L9/v6ub8K3A4/GuXLlSVVXV1Q35h5OXlxcXFyf1q5SUlO+++y49PR1ustlsbW3tDRs2dGLrEO2FRCIlJSV1dSv++cTExPB4vE47nCp1IZ1OLy0tzc/PV1WwSFZWVmBgoBI/fPLkybFjx7hcrkqa0U5YLFZmZuatW7cePXrE4XC6ujn/B4/HKyoqioqKun//fm1tbVc148uXLzt27Pib+o9EIlFVVVWpBI2NjV3YKjKZXFpampGR8enTJ2L5mzdv/vOf/7BYrK5qGIZhzc3Nr1+/joiISEhI6MIn9ObNm39r0xSFQpG860gkUhc2iUajlZaWZmdn5+Tk8Pl8vDw4OPjly5eS+9fV1QUEBJDJZLjZ0NCgra29efPmTmru3xkKhZKYmHj79u3k5GTc4Nr5kEikDRs21NXVdVUD2gnsJ8Wor6/vwibV1dXBrrugoIBYXlBQ4OHhwWAwJH9CIpGuXr2amprK4XDevXt348aN+Pj4dnatqtSFL1688PX17dmz54gRI9pfW3Fx8ZAhQ5T7kwQCwenTp8PDw9vfjHZSV1c3adKke/fuUanUEydO/Pzzz8Qes2vZs2fPwYMHa2pqMjIypk6d2iXKjE6nz5o1q6SkpPMPrRKEQmFYWJi/v7+ZmVn//v379++/Z8+eoKAg3ArSJdy6dcvd3R0AYGNjQywXCASRkZHHjh3rqoZVVFRMnTo1OTmZQqF4enru3btXJBJ1fjOSk5O/qSdRCRISEoKCghwdHeFd9/vvvwcFBZ0/f74Lm5Senu7r6ztgwABDQ8OGhga8vKmpydHRUWyIIklaWpqGhkZrxkUETm5u7qJFiz5//lxfX79mzZqucgLweDxnZ+fCwsIuObpKuHPnTlBQkLW1NXyIdu3aFRQUlJCQ0IVNiomJ8fT0BABMmTKFWC4UCp88eeLp6Sn5k6tXr9JotPnz53t7e+fn59Pp9K1bt+7cuVMoFCrdDBX7kTkcDgDgp59+an9VmzZtyszMVPrnjY2No0eP/vz5c/tbojRMJnPp0qWnT5+GmyUlJUOHDv0WXFc8Hi84ONjZ2RluMhgMExOTmJiYzm/JyZMnL1261JlHrK6u7ohq/f39AQB79uzpiMqVw9raWkwXYhjG5XLnzJmTk5PT+e2h0+lWVla3b9+Gm3l5eRYWFp3pHIFQKJQJEyaUlpZ22hGZTGZ5eXlH1PzmzZvvv/9+1qxZHVG5cuzYsUNMF2IYlpmZOX78eBkjz+bm5oULFx44cOCb8qh8gxQUFOjp6eH21ydPnqjkbasEN27cOHToUKcdjk6nd5BT69KlSwCAbyqAYe7cuWK6EPLTTz+9ffuWWNLc3Ax7VAMDgw8fPsDCyMhIa2trNputdANUrAs/ffoEAGj/7ZKSkrJ8+fL21CASicLCwpYuXdqFVgE/Pz8rKyv8zScQCNTV1VWYkSEnJ0c5lZOamqquro7btEQikaWl5dq1a1XVMDnJzc0dM2ZMJ9vtr1271hHVnjhxAgCwf//+jqhcOZYsWSKpCzEMe/jwoYWFBZPJVK7ad+/eKTcS3bJly5IlS/Df8vl8DQ2NzrdSBwcHHz9+vDOPmJOT00F33bt373r06CH1X+4q/P39JXWhUCjcsmXLgwcPpP6Ey+VOmTIlJCREub761atXNBpNmbZ2EVwuV+mB2aJFi3bt2oVb2b9+/dqtWzfVNU1eSkpK9PT0Omi0I5WnT58mJiZ2RM3Xr18HAPz2228dUbly/PLLL1J1YWpq6pAhQ+h0Ol7C5/M5HE52draZmRle6O/vP3nyZKV7eEzlujAiIkJdXT0lJaU9lbBYrJkzZ0ZERLSzMU1NTVOmTCkrK2tnPcrx9u1bHR2d4OBgvAQaU8PCwlR1CCcnJyUeFRaLNWXKFBsbGzwEQSgUWlhYzJ8/vzONN0Kh0N3d/T//+U+nHRHi5+fXEdX+jXQhg8GwsLBIS0tTrtrFixcrcZ9ER0erq6sTfV5sNhsA0MnzjRobGy0tLTu5T3j06NG/XBdiGJacnDx79mzJdxWHwzl79ix+Y1RUVCh6xJEjR/69vJn19fXbtm1T9FdCoTAwMFBPT6+4uBgvLCsrAwB0vlssKCho06ZNnXnE4OBgpAtZLNakSZMkPXsHDx5ct24d/MzhcMaPH79w4cL2WMRUrAs9PDw0NDQqKyvbU0lubq6Ojo5KYqi3bdsWEhLS/nqUYMuWLWpqalQqFS8pLi4GAKjQbWplZaXEo3Lz5k01NTVi59Lc3Dxu3DhbW9vO1IUsFsvMzEzRyebt5N27d/Pnz++Imv9GuhDDsEOHDvn4+ChX7bRp05S4TxwdHbt3797U1ISX5OTkAACSk5OVa4ZyhIeHy3ZodgTr1q1DurC2ttbQ0FBsGMDn8729vdPT00X/5ebNm4oe8bvvvvt76cKamppff/1V0V9RqdTRo0fPnTuX+PTFxcUBAPLy8lTawLaZMWPGu3fvOu1w8A2FdCGGYefPn5dU5JqamrgPpKKiYsiQIc+ePWtPA1SsCxcsWGBubs7n84VCYXV1NVEVyc9ff/01fvx4GTvU1dURJ+awWCyiZZVIWFjYsGHDOj9mpbS0VFdX98CBA8TC1NRU1doLldCFTU1NEyZMmDlzJtEVWF9fb2RkZGdn1x5dKBAITp48Kb/5JycnRyXzk4iwWKwvX76UlpZKffFXV1dDs6iMGoRCIZxTLzU4o7m5GRo8WCyWmAdfCV1YWVmZm5vbTjd6VVVVXl4ePqmTiAxd+PLly8GDBys3MVkJXVhQUNCzZ89z584RCx88eNB+eyGVSvXx8ZE6TU8SPp9vZ2en8hmvFAolNze3tYiOhISEfv36ydaFDQ0NhYWFrc2+amhogE8rmUwWO1MldGFdXV1BQUE7LaZUKrWwsFBqjGZrupDP55uZmW3fvh33gbLZ7P3790+ZMmUFgTNnzijamH+JLnzw4EG3bt0ePXpELAwNDVWJvTAsLEx+JUEikYYMGdLOIxIRiURkMjk/P7+1CMIzZ850795d9suupqYmPz+/NXMSHmlQXV0tNldXCV0oo9eVH3jKUo1oMnQhiUTS09Mj2t2/fPmioaERFBSEYRibzd65c2f742RUqQt5PB4AICQkJCsra+PGjSdOnHB0dPT09FR0yrSrq2trUyarq6t37959/PhxJyenqKgogUAQHR29e/du6JGUDHt69+7d999/35kx5hiGiUSiI0eOaGhovHr1ilh+/Phx1cYXKqEL379/r6OjIzbg+Pz5c58+fX755Zf2NIbL5drY2Bw+fFjO/QMDA9evXy9WWFhY6Obm5uzs7Ozs/OuvvyYnJ79//x4v2bx585cvXzAMS0hIWLFihbOz8/r16+H1FAgEiYmJ3t7e4eHhFy9etLKyioqKwuULh8O5ceOGvr6+urr6sGHDtv2X/Px8/NACgeDp06f29vahoaHXrl2zsLC4fv06futOmzatR48eAIDc3NyUlJQdO3bY29v7+/vjP5dfF0LpuXbt2oMHD4aHh2/YsMHb2xsa0shk8tatW1etWuXs7Lx69eqAgACYhILNZu/atcvZ2fnnn38+cOAAk8kUiUS1tbWbNm3y8fEJDw/fuXPntm3bxKKsZOhCEok0ePBg4unLj6K6UCQSbd26VUtL6+vXr8Ty7du3tz++8NOnT0OGDJGzg25ubjYzMxN7swqFwrNnz65fv97Z2XnlypVubm5kMnnXrl2//PKLs7PzmjVrfH19mUxmc3Pzzp074X24b98++NvGxsbQ0NDDhw+Hh4evXbvWxcWF+Iauqqpav369vr4+nBiO33XEo9PpdH9//02bNoWHh4eEhMycObO8vBz2YxEREf369dPQ0LC2tqbRaCEhIYcPH7axsSGa2BXShQ0NDd7e3tu2bQsPDz9x4oSTk1NdXZ1IJGKxWIcPH3ZxccGvAB43dvjw4eXLl69YsWLTpk1FRUWwwX5+fm5ubuHh4adOnbKxsamqqiJOKm9NF2IYdujQIeLg882bNyMkUCIZ3r9BFwqFQkdHR21tbbFhpIODg5aWVvub9Msvv+zcuVPOnc+dO7d48WKxwi9fvmzZsgU+IOvXr3/w4EF5eTlesmnTJvgfpaenw/7NxcUFKlESieTv7x8SEnLjxo2ZM2fu3r2beI6fPn2ysbHp378/AMDR0RE+QTt27MB3EIlEFApl165dO3fuDA8PDwgIWL16NYVCgffkwYMH+/TpAwDYunUrjUbbv3+/l5eXjY0NMTJSfl0Ie93NmzfDF82uXbu2bt0KU5JRqdRt27atXr0a9tJ79+5tbm6Gv/L29nZ2dv7pp592795NpVJFIlF9ff2OHTs8PDzCw8N9fX3XrVsHy4l/R2u6kMlk/vjjj+/fv8dL4uPj1dTUzp8/f/r06c2bNz99+rTNE2kTVerCp0+fAgD27t2bmpoKX6hVVVUGBgaKek6nTp0qdW4slUoNCQmBb1AajWZubn7p0qW0tDQOh7NkyZI+ffpICtCysjJdXd3Hjx8re07K0NLSMnHixJEjRxLNpQKBYNGiRXp6elDZqAQldGFAQAAAQCwAPDo6GgDw119/KdGGL1++1NTUYBgmFAqXLFmCzzbNysqSncNv6dKlksGFHA7n69evixYtAgC8e/eupaWFy+V+/fp19uzZWlpaxcXF8C+m0+khISHm5ua5ubnQ6BUbGxscHIxHVOTn56urq588eRJuCoVCOp3e2NjYv3//uXPnNv4Xor45efLk7Nmz8QErlUqdO3cufk2qqqoiIyMBAFFRUQ8fPqRQKP369Zs0aRL+c/l1YXJy8g8//IC//1gs1q+//rpmzRoMw/h8PoVCOXv2LADA29sbvrYxDBOJRCUlJb169YqMjKyurhaJRO/fvx88eHB0dDSshMvlenh4ODg4EM2cMnQhlUo1NjZWwmeHKa4L6+rqdHV1raysiIUsFmvUqFGjRo1SwpbP4XBevHgB/3cGg2FsbIyrENlKl0KhGBgYSIYuNDc3p6WlDRo0aMqUKSQSic/nNzY2RkVFwT8UmuuEQmFZWdmCBQt8fHxwgfvnn3/iioTH4924cWP48OG4KY7L5cLbDABw9uxZ/K4jHtre3t7Lywv/12JjY4cOHfrx40fYqqKiIgsLC0tLy717937+/DkvLw8AQAwAkF8XUqnUKVOmHD9+HH9GLl68OHLkyIqKCpFIRKfTk5OTe/XqNXv27MrKSvz/rampGTNmjI+PT2lpKY/H43K5tra2+/fvx/+1e/fu6evrE3OtydCFDx8+NDU1bU84vFT+DbqwqKgIALBq1SpiIby1pk+frlwzSCQSbqzavn37hQsX4Of8/HzZToz169dLDum5XC6ZTP7pp58AAOnp6c3NzXw+v7q6esmSJQCAgoICeM+0tLRcvnzZ0NAwLS0NjgkPHjyID2jZbPaePXvMzMzwfpjNZjc2NiYmJgIAoqOjJR+iiooKExOT8+fP413l8ePHJ0yYAGtoaGj4+PGjpqamm5ubl5cXmUy+efMmAIBov5dfF2ZlZenr6+OrZnC5XE9PT3t7exaLJRAIGhoaYOWbN2+uqanBrVQVFRVaWloXLlz4+vWrQCD48uWLsbHxlStX4LcCgSAgIMDKyoqoFmToQi6XO23atNDQULxk3759s2fPxjCsqalJVRlhVakLN27c2L1799jYWLyksbHRyMjI1dVV6v4MBkOqtw4AkJWVJVbI5/N9fHxwZw2Hw5k8ebKVlRWHw2Gz2UuXLt20aZOkvZBKperp6d24cUN2yyMiIiYpguzXP1TDffv2nT179pz/MmPGjO7du//4449Se0zlUEIX2tvbAwBmzpw5hwAcjSmXFcjDwwMAkJiYyOPxVq5cGR8f39LSEhwcrKenJztvMADg1KlTUr96+/atlpYWMZNZSEiIpqYmMQXa69evIyMj4WeRSLRmzZoePXpkZ2fjO2zfvt3ExESs5oEDB0r1I79+/bpbt254RiHIzZs31dXVccfE169fAQC7d++GA8GMjAzi4eTUhVwud+LEiWJdeW1trba2NvGeHzx4sFi4Z0tLy7x58+Bjz+PxZs2aJeaFp1AoYq4WGbqwpaVlwoQJ0PWgKIrqwszMzO7duxsYGBBvOUtLSy0trXnz5imRlffLly9DhgxZtmxZdXW1UCg0MTGh0+mFhYUrVqyYPHmyjB9WVFQAAKQGt/B4vAULFpiYmBBtrn379hWzrOPvTgzDuFwuAIA4DZDL5Y4bN+6PP/4Qq1zsPQQRCoU7duzQ09MjFrJYLENDQ6JosLOz09fXx22ciYmJxJgZOXWhQCBYt27d4MGDiYXNzc06OjrE3BHTp08fOXIkMbqAz+cvWLAAXhOBQLB169b+/fsTK2Eymf3799+6dSteIkMX5ubmGhoa4nYUVfFv0IVQcIwbN474EJmamgIAdu3apVwzAgMDNTU1Y2JiuFyul5dXZGQki8W6ePHi0KFDW+uZIbq6uq0dtKysrFu3bnfu3MFLHjx4oK6uTny5ZGVl4SElcPLZvHnz8G/z8vJ69+598eJFYrVv3ryBbxmxw/H5/GXLlg0dOpRYSKPRevToQYxa0dLSmjBhAvQsNTc3w5cU/q2cupDH482ZM0dqr0s0PA0bNsza2pqoz1gslomJCRwO8fl8R0fHYcOGEStpbGzs0aMHsYuQoQv5fP7ChQs9PDzwkoULF6owOA2iMl3IYDAGDhy4cOFCYmFlZaW+vv7evXsl979///7JkycPHz4sNjSBb1/JJH9kMpl48k1NTdra2rg5h8fjSX1XQV3Y5qRXHo/XogiyjRx5eXkaGhpeXl5Ey3BNTY22tvZPP/2Ei9fm5uaPHz+2Z46OErpQW1vb0tJSTEDr6uqOGDFCbKgBr0mbFTY1NcXFxU2bNm3Pnj379u1LSUmZO3euo6Njbm6ujKtEpVKlvizxOkePHk2cMe3r66umpkbUbZs2baJQKPjmjRs3pk6dSoyaOnfuHADit3drunDnzp0AADFrE3zp4uoT3pmtZSiUUxfCgS9x7AQxMzMj6okzZ84AAN68eYOXREVF4VM0Xr16BQDATbM4VlZWxDFYm7pQzPwgJ4rqwjt37kA3B7EwOTlZQ0ND7HIVFRVBU5lshEJhZWWlv7+/sbHxs2fPtm/fHh0dPXr06KNHj8p2KL948ULylsC5du1at27d8MxwTU1NOjo6Ojo6xLkyRCXN5/NdXFx+//13vITH402fPn3MmDFiNUu91aFIxWcR4mzcuJFoh7azsxswYICYCx5HTl1YUFAAAJDsBp2cnIi/hfZI4nIA5eXl+OgOTn2VjM4Ue4fJ1oWampoqHBhDlNOFDAYjMDBQRuqMr1+/+vn5SVoocPLz8/fv30+cwCcnSujC3377rUePHmIxSH5+fpqamg8fPlS0AZCWlpbExMR58+Zt2bIFrkmzbNmyOXPmZGRkyEh919TU1KtXL7HQeSJWVlZjxozBrcKXL19WU1Pz8vLCd9i/fz90MWEYxuVyHR0d//zzT/zb8vJysZEG1rou/PDhAwDg6NGjYuV2dnaOjo74ppaWluRTiSOnLoS9rqSxw9ramvgUw7BpYrBmUlISbuN4//49jLUTq8TGxgZPJ4zJoQvxb//888/vvvtuyZIlqk3soDJdWFRU1Lt3b7GOPjU1tXv37mIObz6fHxwcvH37dj6fn5aW5ujoSJxQDfumNpM/x8fHQ+u07N0YDMaAAQOkZgnvOODYTixt+tWrV/Hbhc/nnzlz5uDBg0lJSWfOnPHx8ZE9+bqiomLz5s3rJOjfv7+dnZ1k+X/+8x+pzhr4eiA+hBiGvXv3DgCAh8o1NTW9evUqLi5u7ty58oevMhiM0NDQnj17Dho06MmTJ23uDzVWa7pQKBSuX79+2LBhuHn4zJkzzs7OixYtgptUKpWYAAiHxWIVFhampKQ8e/YMGjLFdmhNF9ra2gIAYmJiXhG4e/du7969cWcrbDNxKExEti6kUqnwXKAAvXr16qv/Zfr06cTEql++fOnfvz8epM/n8319ffFvDx06BAC4cOGCWCUODg4rV67Ed5OhCzkcjqWl5ZIlS6R+C7l+/brkrbVu3bqBAweuXbtWsry1IYqvry8AQOyVdvDgQRgqADc/f/4cEhJy+/bt8+fPz5s37+jRo/LMiSkrK5s/fz4AYM2aNfKkL4iJiZGhC+l0OrzmcDMhISEuLm7YsGF4GMyLFy+kxubX19dnZmY+ffr0xYsXZmZmcupCKFJXrlwp9if6+PgQbZB2dnajR49uTUu1qQuhYIIhEO7u7mLH+u2332bMmEHcf968eUZGRviAMDIyEo/EgkOatWvXilXi6ek5ceJEvAYZupBEIrVmr5WHtLQ0qTekhobG0qVLJcvPnj0rozZ4W44bN661OYtbtmwBANjZ2bWWrXPp0qVt5kMODg6WbNiKFSuMjY2lnguempiIQCBYvHjxkCFDxIYHzs7Ourq67UzcwWKxrl+/rqur26NHj3v37rW5f319fc+ePWXowr179+ro6OBj7GPHjrm4uOAeEiaTKRmwJBKJqqur37x58/jx43v37unq6sqpC2FK6gMHDojdk+vWrSP281paWvi7QxLZurClpQXmToIrF5w7d07sWAsXLlyxYgW+P5VKHTVq1IoVK3BVQ3xVwQH/4cOHxSr5+eefHRwc8N1k6EKhUOjs7Gxubt7a6agElenCqKgoNTW158+fEwv37dsHABALh3/16tXo0aPhVcvPz9fU1CSO8uXUhb///ruJiUmbi2g1NDTo6el1si48cOAAAEBs3ScLC4upU6fCLgbOhoFGF5FI9PjxY/xtJBWBQECj0RolmDx5MjHkAqepqUnqlXn06BEA4O7du8TC/fv36+npwaByDMMaGxtTU1NbWloAAPLrwvv37w8ePHj58uWbN2+2srIiOlilIlsXYhiWlJSkpqYGh8JFRUUxMTFv377V0NCAN0ZOTo7kXKJXr16NGTMGxv4zGIyQkBD5daG1tTUAoLi4mGgVZjAYNBoNt1nCNuMhfWLI1oVZWVlwiOzi4gIAgLE1ROh0OlHKi0SisWPHmpqaQptrU1MT8VpByZuUlCRZCdHEK9teOHHiRNm6kMlkSt5ajY2NU6ZMqa+vlyxvzcawYcMGAIDY22vcuHH29vbwLqXRaC4uLng3+uHDh169erXpGaFSqXAiyKpVq4YPH+7u7t6mS1q2LsQwzN3dXVNTEwqXK1eu8Hg8BweHCRMmwFPbvHmz2CGYTGZQUJChoeGNGzdqa2vpdLq1tbWcuvDJkycwHF7sT2xubiZaKO3s7MaNG9da3uY2dSF82OGg9NixY2LHampqEvPqHj16FACAG003bNiA/60PHz6EcRSyGyxDF+bn57dHF+LxmmJoa2u/fftWsly2uyM2NhYAYG1t3Zrx+9SpU1AHt/aWgUtNig2zxWAwGJIN+/z585o1a6Sei9TGwJAyIyMjMQlrYGDg5ubWzpUknz592qdPn59//tnT03PChAltmp3a1IWZmZl4tHpZWdmlS5dyc3PV1dWhuqqpqSG6QWCFO3fuHDduXFxcHJyYr6enJ6cuDA4OBgBcuXJF8sYmztzX0tJaunRpaw2WrQtJJBJMA7R7925o2ZHd64pEIgcHh2HDhuGqhjiJNjAwEBodZTe4TXvh30YXQglIVHgMBmPo0KFERziGYUwmc9KkSXjQ5cePHzU0NMRkRJu6EKqWjRs3ttkq6EeWalsi8unTpweKIDsUD45EiSVZWVlaWlq4CdrLy4soVaurqwcMGED0isqJon5kaGQlpsNgMpmWlpYrV66UjFeVUxfGxMTMmjXLwcGhoKDA3d395cuXcXFxU6dOXbZsmeyc/rJ1IYPBGDlyJHyYd+zYUVVVxefzTUxMYDjUpUuXGgmhx0Kh8OjRo7179ybOp2nTj8zn83Ev0ty5cwEArXnrIO3RhXfu3IEBA3/88YfU8FlJUlJSAABQGV+/fp04zIDGtjYTU7fpRyZO65MfRf3IGzduFPsjnj9/rqOjg7c/IyNDS0sLH1KyWKwhQ4bIWO6IRqNt3759/Pjx3t7ezc3Njo6O1dXVvr6+Q4cOlf2Shq5bGTs8efJEW1sbWoih4zssLGzAgAElJSXl5eVi8+dqa2ttbGxmz56NS155/Mj19fXQdJqcnAxteDLag7VPF9bX10PHd0REhFTvlSSlpaX4OzIlJYV4t0Mh6+3tLbsG2X7kPn36qHxtEuX8yEKhkEwmy3CY8ni8qqoqGcEwbDb769evSuT2UtSPDO8rU1NTYmF4ePjgwYPbs6pnUlLS/Pnz58yZk5WVdezYscTERJh7fOHChbL7FskUbERYLNbYsWPhPenv7w8H8FOnToW3+qVLl4hDRDhrZNWqVfgNI48fubKyEoYMnT9/HgDQ5vy59ujCV69eQa0MX+tiOUakkpGRAf05GIbdvn2bqBZgXiF85kprtKkLiRbKjkBlunDevHkLFiwgljx79kxdXf3FiyzxhogAACAASURBVBcYhgkEAjgBMy4ubsCAAbixJzExUV1dXcwdrKGhIRmARQRmxCVOySkrK5N6rWtra/v164eHiMmo8JIiyF5aOzw8XOz1s3HjRgMDA3y0N3PmzMDAQGIjDQwMWlskSgaK6sLCwkIxXZiamqqlpSU1mZmcutDHx+fEiRPw84oVK+CLpKGhYfHixfHx8TJ+qKmpief7kAo0XWRnZ8O4KJFItG7dOkNDw8bGRjFBQ6PR+vXr5+LiQiwk6kLcCUvUhWw2G28AXKpcUsgKhUK832+PLgwKCoJxtCkpKWpqamImW3h2kq+o6dOnT5gwgU6n29vbE8vfvHmjrq5+9epV2ZXI0IVNTU1mZmbEJ0h+FNWFfn5+xMeBy+XOmDFj9uzZ+Bu3rq7ut99+w1/tDAZDX19/2bJlrVVIJpP79u0LpxVDwyocJLx//x5O624NEomkrq4uY85yXV2dgYGBs7NzdnY2dBnX19draGhcuXIlOTkZD4qC7N69u1evXsTENDweb9q0aVAX5uTk4INboi7Mz8+Hl72yslJDQ2P16tWSzSBe3vbowvDwcNgrfv78WV1dfffu3ZL7SA4If/3116FDh3K5XDHNWlFRoaGhIXUSIbHBMnRhenq6kZGRnMkm5ecfP+8E+pGJupBGo3Xr1o1oXBCJRKmpqQ8fPuTxeGw2OyYmJi4uTvalDg0NxTvAPXv2QAt9S0uLi4uLbKWlp6cn23sOlVZ2dnZAQAAs2bt3r6amZm1tLVHQiESi2bNni80aIepCfM0zMV2YlpYGH09oiZS67i5R0LdHF4aEhEB9Anvdy5cvi+0gtetesGDB2LFjGQwGcUoNhmEfPnxQV1eXDIgUa7AMXcjj8ebNm9fRCyioRhfW19cDAIjxHEKhcN26dWPGjIFd9unTp4uLi2Hc2JAhQ+Lj41NSUlJSUrZu3SrpaF6yZInkO+/ly5cbNmyAXo+TJ09qaWnhxh4+n//HH39IHTkVFhb26tWLmOynE4ABqvhmaWmpubk5secaNGgQURc2NjaOHj1a9hQwqSiqC5ubm4m6kEqljh8/vrUsQgr5kTEM4/F4dnZ2YlN6ZeDu7i7be56dnd2rV6/ly5fjprL4+HgdHZ19+/aJBatB04vYwombN2/G/wXc+GRhYTFr1iz4mc1m44lscnJytLW1Jc/32bNn+FwzpXXhhw8fLC0t8ZtcTBVBiouLidkQIUFBQQAAT09PfCVrnMWLF48dO1YsCK+qqoqYP1KGLqypqfnhhx+UW2tEUV144cIF4uPw8uVLc3NzGXmVCwsLe/fuTZz5K4OKiorhw4fX1dXJszONRjM2NpYx4hcKha6uriNGjFi6dCnuYHVzc5s7d+6FCxeIV5vL5c6ZM8fIyIh4aA6HM3r0aKgLU1NT8YxUurq6+J2Wn5+PT+wIDAwcNGiQpIQiTsZUWhfW1tYS/+KdO3caGhqKOSI5HI7k2KC8vBwAEBAQICkO/Pz89PX1GyXyTxEnFcnQhVeuXJk2bZrKVxn4x+tCkUj0yy+/EHXhmTNnzM3NiVM2ExISEhIS9u/f7+zsfO/evdLS0tTUVAcHBzmzAm3YsEH+BUL9/PxkT1mrrq4eOHDgihUrcK90Wlqaurq6p6cntBNBKBSKjo6OmDkpJyenV69eUBfib8ns7GxtbW08e11aWhreJW7cuNHIyEgsHKKpqQm3VmDt0IWfPn0aM2YM3lk5OjoSp9RAxHpdyK1btwAAe/fuFYuswzBs3bp1I0eOFJPsjY2NRHezDF3IZrMnTpwo23DWflSjCx88eKCpqUk0+3G53NmzZ0+fPl0gEJBIJOh9gOujr1y58s5/sba2lrzDfHx83NzcxAqtra11dHQqKys5HM6hQ4f69OmDT2dJT09vLVf7q1evvvvuOyVctO2hvr5+1KhRMF6bQqEsWrRILE03AICoC+l0uqmp6ZEjRxQ9kBLzkdevXw/nRvH5/CNHjlhZWbUW4K+ELly2bBn+/muTe/fuifUIYsC3uLm5OR690djYqKurK7kWTmFhYY8ePbZt24aHfxUXFwcGBmpra79//762thY/kbNnz/bv3x9WWFdXR5xEcu7cuUmTJtXU1ODxOpWVlevWrYO9AJPJhLN2rl27xmQyiQNENpvNZDJhUKmnpyeTAJ1Oz8zMNDQ0HD9+PN5zvXr1atiwYXFxcXhrWSxWYGCg5Kv006dP2trauJAlkpOTM2jQoBs3buCVsNns06dPw7uOzWa3tLTMnTvXysqqpaVFcjhbUlLSr18/5ZxQiupC+O9AY1tVVZWNjQ1xlWQxuFyuq6vr2rVr5XyfffnyhZjtTDYsFmvatGmy83S+fv0aAEB8Rz59+lRTU5M47wfDMIFAsHbt2n79+uETqHk83qNHj2bNmjV8+HAajfb8+XPcvmhjYzNjxgw+ny8SiS5fvkw0Sy9evNjV1RW/w/l8/vPnz6FNgsvlwj/xxx9/rKysZDKZxPl5HA6HyWQmJyd379595syZxLuOwWBUVFQ4OTkRZ/Y0NDRYW1v7+PjgsozH40VHR0s16s+dO7d3796SYgtm83Fzc8P/HT6f//TpU6ggeTwek8n09PTU19cnkUiSHYu7u/uSJUuUyEwkm3+8LsQw7MmTJ6NGjWIwGCKRqKioSCz3glAohCPS9evX4zPP4GuFqMNksHXrVtneGyLPnj2ztraWsQOHw4GGQNxk3tzc3LdvX2NjY+JuLS0tY8eONTY2xvtGDodz+/btH374YdmyZSKRCLeyk8lkQ0NDd3d3kUgkFAp9fX1xZ3RlZaWpqWlgYCDeKXG53Fu3bkFXOJvNZjAYWlpaCxYsaGlpYTKZxNsPdt2nT5+G84eID1FTU1NhYaG5ubmBgQHeT378+FFfXz8sLIzY6/7111+So9yqqqohQ4aYm5tL9r1lZWWjR48mZtvlcrlhYWHwUeVwOC0tLYsWLTIzM5PadcPwPJWsEiwD1ejCgIAAExMTMc1++/btcePGPXz48MSJEzAwmcVi/fjjj3hP1NTU1KdPH8lQhoSEhEGDBhFjmTEMO3bs2Pbt22NiYuA98fjxYxcXl9jY2GPHjsXFxbW2RHRwcLCMgUIHIRQKb9265eTk9PDhwz/++OPx48dizhp9fX2iLmxoaDA2Nu4EeyGGYQUFBZaWlvfv33dzc7t06ZKM0GxFdSGcUyb2r8mgrq5uxIgRMiS7SCTau3evs7MzcT5gQEAA8dLhe96/f3/KlCkeHh6xsbHnz5+PiIgQiURubm6jR4/28vLCQ93ZbPbGjRstLS2jo6MPHjwoFlD45MkTV1fXvXv3xsbGHj169PLly3jzLC0tjY2NAQADBw60sLDYtm0bvOUEAsGmTZssLCx69+4NAOjfv78lAVNT0549ewIApk+fTnzCCwsLd+3aBd3uYWFhwcHBUp2bfD5/8eLFrf0L5eXlXl5eDg4OUVFRERERx48fz8rKgqJ27969kydPBgAAACZNmiS5kkFUVJStrW1rV142iupCgUAQGhq6efPmBw8ewAjU1pQBvJgeHh7yp7jj8/kVFRVySg2RSOTp6Sl7vUcej2dpaUm8MeAEQ8klaOl0OgzVP3/+fGxsbGhoaHp6+sePH42NjadPn37nzh28U0pLSxs9evSBAwciIiKuX79O7KwaGhqOHz/u4OBw8+bN2NhYf3//hIQE2F0EBARMnjy5Z8+empqapqamlpaWxBVpQ0NDLS0tR4wYAQDo0aMH8a6bOHGinp4e/PeJ6zrU1tb6+/vb2trevn07NjY2MDAwOTlZas8ZEhIye/ZsqTlyqVRqUFDQggULwsPDYYPx/u3evXsWFhbwhjc1NSVOjccwjMlkjhkzRvYcYeX4N+hCLpfr7e0dGBh4+fJlT09PsXROAoGATCY3NTUNHDgQN7RTKJThw4fLmcWmrq5O0gzcGiwWy9jYWMZgTCQShYSETJ48mXh3Xbx4UdJBRKFQXF1dJ0+efPPmzfv370OnbVJS0o8//mhra4tbu4VCYXR09Lhx465cuXLu3LkHDx4QH3kYXjxnzpw7d+7ExMQcOXLk1atXcAdnZ+cJEyaoqal99913kyZNsrS0JK4rsXv3bktLS5i+t2/fvsSHyNzcvFevXgAAIyMjorW+oqLC29vb3t4e9rrHjh3De10iQqFw48aNHh4eUmcFVVVV7d+/f9asWZGRkffv3w8KCkpPT4cNPnr0KLHr3rJli9hv3759K7ZGQEegGl1IpVKJHRAOiUQqLCzEB5csFmvkyJG4uI6IiLCwsJDs08lkcr9+/STXiysvL8/NzcVVAoVCycnJke1CWrJkiTyBoh1BbW1tTk6OVJ1ka2srGV+oxKIsSuhCDMOam5vz8/PFgqUkUVQXKgqPx7OxsWnNLQthMBhiNi0Oh9Oa9OTxeF++fHn//j3xJyQSSVIElJeX5+fnSxUfcJG6zMxMOV2T7eHr16+ZmZlwwYnW9qmtrZUdp19dXf3+/fvS0lKFzDDr1q0Tc7vLjxLrI2MYVlVVlZOTI8MKyOfz7927hxvzOsjG//79+1GjRsmunEwmi6UmkZGjjkqlfvz4sbCwEL+deDyeZP0CgaCgoKCkpESqDqPRaHl5efn5+SpfC0QS2G1++vRJhj+3qalJ9hysxsbGvLy8goIC+ZfYLikpGTFiRDuXlJXKv0EXYhgmEAjKysoKCwtbe9Jh0BQ+pw0mIFPh8lpEVq9eff36dRk7MJlMOAEZBy5YJ3VnMpmclZWFL2eFYVhzc7Nkv9fS0iIj429tbe2HDx+KiopUteaHDOTpdSkUiux+pqamJjs7u7i4WKHu9Pfffye6yDsIVa530iZsNhufrNfU1DR27FipCZOEQqG3t3ebM/XaJDc3d+rUqZ3Q1SqKp6cncZp2VVVV3759lbAM+/r6pqamqrRp/0dH60IMw65duwZXrOnQoyDEKC4uHjVqVJsDg9bw8vJqzTyvNCKR6OzZsw8fPsT7WdlvHaURCATLly+XJ8UmQrWcO3dOzBevKlxdXZVILt2F0Ol05ZYakk10dLSamhouMnbt2vXzzz+r/CiQ2NhYAwMD2ZmAECqnsrKyb9++nXC3d6ou5PP5dnZ20CNz/fr1AwcOtJYmoKGhwcHBQf7BqCQ8Hs/FxeXChQvtTO/UEWRlZRkbG+NRw2fPnpVcRaBr4fF4AAAlQh4Vorm52cbGps1khwgVAgddgYGBKg/zUhoej3ft2jV7e3tPAvJHOykKXNehE4wKCBwajWZra0ucu41QOV5eXmPGjIF2azi7i5h6QrUwmcxZs2Yp4a1CtIeTJ096enoq4a5RlE7VhRiGZWZmOjg47Ny5MyQkRPbpXbt2rT320qKioo0bN36DohDDMJFIFBsba29vn5GRcfbs2UWLFqk8p5fSUCiU9evXGxgYAAA0NTUNDQ3bzA7VHjIyMqQuk4joIJqbm9evX9/aEg5dQmZmprq6OvhfiKuxqRYul/v777/LTqKEUC2BgYFtJpFFtJNZs2bdvHnT09MTxuKr3KgvRllZmVhyYkRHs3bt2s7pujtbF2IY1tzcLE/KeyaTuXbtWiWy+mEYRqPRVqxYIbkkxjdFbW1tdnb2v3wMzefz/fz8jh49+k0plX8qDAZj4cKF8mTV/mfT2Ni4du3ajp7Th4Dk5ubu3r1bRgZpRPupra3t06dPXV0dk8lsT7Jr+RGJRKdPn/by8uoE8xWCxWI5OTnJWM5btXSBLpQfkUh08OBBRUOVeTyeh4eH1HkwiG+TCxcuyF4cBaESAgIC8AUP/+VUV1e7uLi0tjYuQlVUVlauWrVK/gnmCOW4c+fO5MmTOz/gLyIiQrk0qAiFOH/+fGcGXH3TuhCTY2acJHw+X858ZohvBA6HIzZ5DdERdI4h4e+CQrk5EMpRX18vf+4qhHKQyeTJkycbGRnJs9ShauHz+cju3glUVlZ2ZlDct64LEQgEAoFAIBCdA9KFCAQCgUAgEAgMQ7oQgUAgEAgEAgFBuhCBQCAQCAQCgWFIFyIQCAQCgUAgIEgXIhAIBAKBQCAwDOlCBAKBQCAQCAQE6UIEAoFAIBAIBIYhXYhAIBAIBAKBgCBdiEAgEAgEAoHAMKQLEQgEAoFAIBAQpAs7AxaLVVpaSqfTu7oh4giFwsrKSjKZLBQKu7otHUVLS0tZWVlzc3NXN+RfR319fUVFBY/H6+qGIBAIBEJekC7scAoLC7ds2RIVFTVz5syysrKubs7/UVtbu2XLlrNnz/r4+Ny7d6+rm9MhZGRkuLm5RUZGmpiYVFVVdXVz/i3weLzjx497eXldvnx55cqVAoGgq1uEQCAQCLlAurBjiYmJsbS0rKiowDAsNTV13LhxDAajqxuFYRhWV1dnb2//6tUrDMPq6+tHjBhRXFzc1Y1SMVeuXLG1ta2rq8MwLC4ubt68ech21QkwGIzVq1efOnUKbrq5uSUkJHRtkxAIBAIhJ0gXdiAlJSUaGhpkMhlufv36VVdX99OnT6qqXyAQKOf/bWxs1NDQOHToENxkMplmZmZHjhxRVcNUhdIniGHYp0+fNDU1qVQq3CwtLR08eDCNRlNd61TDt2lLE4lESl95V1fXESNGcDgcuBkUFGRjY6O6piEQCASiA0G6sKOg0+kzZsxYsWKFSCTCCwEAsbGxqjrEsWPHMjIylPhhQECAmpoaLlgxDJsyZYqjo+O3plGCgoIyMzOV+GF9ff2PP/64fft2YiEA4MOHDypqmsrw8/PDJdS3Q2Zm5u3bt5X4YXZ2tq6uLvE6h4eHAwCamppU1zoEAoFAdBRIF3YUFy5cAAAQdVtDQwMA4NKlS6o6hJWVVWJioqK/qqioAACEhYXhJVwud8KECba2tt+am9XS0lKJE8QwbP/+/QCAgoICvKSurg4AEB8fr7rWqQYLCwsWi9XVrRAnPDw8KChI0V/RaDQ1NbXVq1cTC0NDQwEAlZWVqmsdAoFAIDoKpAs7CisrKzGllZ2d/S3owi1btowYMYIY5kij0YyNjf9JutDMzMzJyYnP5+MlaWlpSBfKj3K6MDY2FgAQFRVFLPT09ES6EIFAIP4uIF3YISQkJAAAbty4QSyMiYkBAERHR6vqKEroQhKJ1L9/fycnJ6J3u6qqaujQoYsXL5bTj0ylUt+8eRMfH5+Xlwe1V0FBQUJCwps3b5hMpkLtkY1yujAyMhIAIPbDa9euAQDevXsnTw0CgeDz588vXryIj4+H3vbGxsa3b98mJSXl5+er1tv+T9KFP/30EwCARCIRC1euXAkAaGxslLOSmpqa9PT0+Pj4z58/CwQCPp+fm5v75MmTjIwMNputaJMQCAQCoRBIF6oegUDg4OCgp6cHpyHjuLq6fv/99x8/flTVgZTQhZGRkerq6idPniQWZmRkaGho7NmzR54aampqfH19SSRSVVWViYlJRETEqVOn0tPTyWTy2rVrPT09FWqPbJTQhVwu19TU1MDAQCxob86cObq6uvg0FNm8fv360aNHFAolNzfXwsJCKBSeP3+eTCY/e/Zs5syZRUVFCjVJNv8YXUij0bp37z5nzhyxCSvdu3cfO3asnJVUVlYeP36cTCZXVFT069fv2bNnISEhHz58KCkpcXJyOnHihEJNQiAQCISiIF2IYRiWm5t7VhGSkpJk1NbQ0DB69OhBgwadOHEC/8lff/01duzYoUOHqjCLnhK6cOvWrQCA7du3E0/H1dUVAHD//v02f87lck+ePNnS0gI3z507Z2hoePHiRQzD0tPTAQCbNm1S4kRaQwldWFFRoaenZ2RkRDzBU6dO6evrm5qa4i2XQUZGxpkzZ3Cj4KJFixYsWJCbm8vn8+3t7XV0dIhhi+3nH6MLIyIiAABLliwhXnk/Pz8AgLe3tzw1NDY2BgUF4VcjMDBwzJgxcPrLnTt3AAC+vr6KnggCgUAgFALpQgzDMBKJ9FIRZOea+fLlS+/evW1tbZlMJue/1NfXjxw50szMTIWeViV0obGxMQCASqVyCDg7O0u6/6SSmpr68uVLfNPT03P48OHQlczj8XJyclTrY1VCF758+VJbW3v9+vXEE/z06dOAAQPs7e2JEYdS4fP5AQEBRCe7paWlg4MD/FxVVfXp0yfit+3nH6MLly5dCuMIiVc+MDAQAPDixQt5anj06BHxyXJ1dR07diy82kwmMzs7+x+8Kg8CgUB8IyBdqHqePn0KALhw4QKxMDMzEwDg7+8PN2k02o0bN+7fv3/hwoXU1FTZeqWlpSUzM/ONBCYmJqdPn5Ys//jxo1R9Bufkrlu3jlhIoVAAAIsXLxY7YlRU1PPnz8Vq+PDhA55whMPhDB8+fP78+TJ0EoVCCQgIICbEUegEx44d+9dff8l/ghiGXb58WXLqw/3794n/SFFR0dWrV69fv75v376MjAyi2hAIBERzYH19vWScKE5jY2NYWNiNGzf8/f0jIiJky7uKigrJE4Hn+OLFC8nympoa2RdNVeTm5koe3dfXd/v27VIbLNXmymKxDAwMxo8fT7wZOByOmZmZkZGRnIsQ5uTk4NeQyWSqqamtWbNGJeeIQCAQCDlBulD1hIWFiWWowTBs9+7d3bp1gy97Op2up6eXnJyMYRiDwdi4ceOjR49kVJiXl2dkZDRQAi0trT59+kiWz5gxQ+pazLm5uZKCNSoqSl1dHc+qWFJS4ubmtnnzZgDA8ePHZbQKTmHZsmWL1G8TExM3bNjw22+/yZM1MDc3d8SIEfKf4KxZs1pbbBoaqPLy8oiFGzdu1NbWplAoGIZlZ2cfPXoUzmB49+7dd999Fx4e3lrD7t69q6GhUVtbK/mVQCAICQmBAYstLS0zZsywsLCQYQw+ePCg5InAcxwwYIBkeWhoqOyLpirs7Owkj967d++ePXtKbfD79+8lK6FSqYMGDRIbXZBIJH19/RUrVihhRc7Pz1dTUwsICFD+xBAIBAKhOEgXqp69e/eKJfKlUCh9+vTZunUr3AwLCyOuAFFYWGhubq7EgRT1Iz98+FBsTi6fz1+5cuXIkSOhZsIwjMvlQq3Tpi5MSkrS0tKKiYmR+m1LSwt0o7cnm7QSfuQtW7YAAIg2rZqaml69euGhab/99ltcXBz+7YEDBwYMGNDQ0CC1trlz506cOFGq+zIhIeHXX3/FN1+/fq2trS2nw5TIP8OPXFZW1q9fPzc3N2LhvXv3NDU14VqLigIHV8SgBQQCgUB0AkgXYhiGxcXF/a4IcKZFaxw7dgyA/7mwt2/f7tmzJz4Tec2aNYGBgfi3tbW1BgYG2dnZijZbUV344sULAEBJSQnx0D/88MOZM2ckd25TF8LTLCsrg5sikUjSrtb5uhBmy8M3RSLR6dOnhwwZgufPc3V1NTIywneIj48HAOTk5EhW1djYCM1duG+Ux+Ph7t179+4NGjQI35nH4wEA/vrrL4Vai/1TdGFVVVX//v3d3d3xEh6P5+DgIBa0QKPRYOJMgUBAoVBkrBXu7u4OAKiuroabQqEQ/ywQCOrq6rhcLoZhMHL3G1wwBoFAIP6mIF2IYRjW0tJSowitmZcgd+/eFdOFNjY23t7eeBDhzJkzibqwvr5++PDhrQWxyUBRXVhSUiKmC//888/58+dLTQvXpi5cvnz5pEmTcBchk8m8du2a2D6drwtPnjxJvPgMBmP8+PFBQUG4za+2thZXGBiGHTlyhKg/iBQWFurq6hJzoyQlJSUkJMDPLBarvLwc/womLX/48KFCrcX+KbqQTqfr6+sTdWF8fPzgwYM/f/6Ml5BIpMuXL//yyy85OTknT56Mi4vz8PBobXK3ra3ttGnT8H+tpqYGv7sePXoUExPj7u5eXl5+8uTJR48ezZ8/v76+XtFzRCAQCIQkSBeqnqKioh49ekCxJRAIxLzGGIYBAIi6kE6nm5qaHjlyRNEDKTEf2cDAIDc3F35+8+aNjo6OWCgesZFiuhCGJ8KQr+Li4gEDBixduhT/NiEhITU1VaySzteF0LsNjUl8Pv/EiRPLli1rbee6ujoLC4tTp07hJRcvXsSNoOfPn9fU1Lx58yb8isPh+Pn5SZ11wePx/vjjj9mzZyuxCrASurCqqsrJyWnfvn3EuD0ymbx06VLi8IPH4/n5+S1ZsuTr16+KtkqJ+ci2trZ4pERDQ8OkSZMuX75M3MHd3Z1Op+/YscPJyamurg7DsPT0dC0tLRjDAPMcwdFRXl5e3759ibbGyMhI/EYNCwvj8XgWFha+vr7wKQsMDFyxYgUuIikUysqVK7dv3/6trd+DQCAQ3z5IF6oeNps9ZcqUZ8+eYRj2/PnzqVOn4s5WSBfqwtDQUHggKpW6atWqly9ftjabWFIX3rp1q1u3bs+ePYMKKTExsW/fvlAnpaSkXL58WXJ6QefrwqamJgMDg7dv32IYFh0dbWdn19psaC6Xu3Dhwm3bthFl2YIFCywsLHg8HolECg4OXr16Na51jh49WlhYKLWqp0+fWlpaKrfUmxK68NatWwAAAABx8g2cc01cWYRGo40bNw4A8PTpU0VbpYQu/Pz587x58zAMa2lp8fDwCAkJId4PVCoVzqTp3r07blX9+PFjt27dsrKyMAwLDg7u06fPx48fm5ubd+3aFR4e/v3338MrExMTc/fuXSj7iouLX79+zWQyNTU1cffxlStXZsyYgQvipKQkeCmg+kQgEAiE/CBd2CF8/fp17969ERER586dk0zSYW5uTtSFFArlhx9+uH79uqJHUUIXtrS0+Pj4RERE/Pnnn7KzMErqQgaDERoaGhERcevWrYqKCj6fHxkZefbs2aioKNy7Kkbn60IMw8rLyw8dOnT79u1r1661lsiay+WGhoaeOHFCTJNVVFRcvXo1IiIiIiKCzWYzmcwzZ85ERERcu3aN6BIl8vr1awcHB6JPWSGU0IVMJjMkJERsTT/Y1LS0NLxEKBQmJSUdP35cCT+1cuvgRUZG3rp1U4DEXgAAAypJREFUKzQ0NDk5WWrqpZycnH79+uELxkRHR6urq+fn52MYRqPR4KWOiIior6/n8XhXr169ePFiZGRkSkqKWD137961s7PDN5cvXz537lz8iGw2+8qVK48fP1a0/QgEAoFAurALcHZ2lpx38vr1a0Xrsba2VkI2yUmb8YXy0E5diJtdVYtIJDpx4gSenfH58+elpaXKVVVZWYkbeuvq6pSYe2tpafkNxhfeunVLCV3YJlevXjU0NKTRaHDT3d1dXV1diXp+/vlnLy8v+JlKpQ4bNkxsKjQCgUAglAPpwi4gLCzM0dER38zOzjYxMZFniTYxsrOzpU6YUAnfgi7MyspS+QnCpfzCwsLK/4uzs3Nr2RBl8/r1ax8fn48fP8J6Hj58KJa0Uh7evn3b5iosnU9dXZ1ql/vDMEwkEq1evRrPcVhfXz9hwgQxq6c80Ol0dXX1u3fvws24uDhTU1MVLi+JQCAQ/2aQLuwCKBSKsbExDNRrampas2ZNREREVzfq/wOTgBQUFAAAdu3aVVFRIediFWLAlC5paWkAgHv37pHJZKmznjufM2fOgP9lwoQJStRDIpGMjIyI9ejq6ooFkiKI8Pl8a2vriRMnVlRU1NTUbN269dKlS0osKpiRkaGmpubj40Oj0YqLi9esWdNa3CcCgUAgFAXpwq6hrq4uPDz81q1b4eHh35TFiMvlZmVlPX78+MGDBw8ePHjy5Ily0ymam5vfvn374L+kpaXJTu7TaWRkZDz4X+AkFUUhkUhi9Tx58uQb0b7fJmw2u2fPnmQy+cWLF48fPy4uLlauntOnT3fv3r2oqCgxMTEpKUnqajQIBAKBUA6kCxEIRGcQFxc3Z86c9tfz66+/7tq1q/31IBAIBEISpAsRCERnsGrVqp07d7a/HnNzcyXWG0QgEAiEPCBdiEAgOpzMzEwXFxcPDw98vohynD9/3sXF5c8//0QxhQgEAtERIF2IQCAQCAQCgcAwpAsRCAQCgUAgEBCkCxEIBAKBQCAQGIZ0IQKBQCAQCAQCgnQhAoFAIBAIBALDkC5EIBAIBAKBQECQLkQgEAgEAoFAYBiG/T+Ngk7p32471AAAAABJRU5ErkJggg==" alt="" />

We obtain a linear regression model where the features are the various powers of the original value(waterLevel). Complete the code in polyFeatures.m so that the function maps the original training set X of size m*1 into its higher powers. Specifically, when a training set X of size m*1 is passed into the function, the function should reture a m*p matrix X_poly, where column 1 holds the original values of X, column 2 holds the values of X.^2, column 3 holds the values of X.^3, and so on. Now we have a function that will map features to a higher dimension.

function [X_poly] = polyFeatures(X, p)
%POLYFEATURES Maps X (1D vector) into the p-th power
% [X_poly] = POLYFEATURES(X, p) takes a data matrix X (size m x 1) and
% maps each example into its polynomial features where
% X_poly(i, :) = [X(i) X(i).^2 X(i).^3 ... X(i).^p];
%
% You need to return the following variables correctly.
X_poly = zeros(numel(X), p); % ====================== YOUR CODE HERE ======================
% Instructions: Given a vector X, return a matrix X_poly where the p-th
% column of X contains the values of X to the p-th power.
%
% for i=1:p
X_poly(:,i) = X.^i;
end;
% ========================================================================= end

3.1 Learning Polynomial Regression

Before we are still solving a linear regressino optimization problem.

In this part, we will be using a polynomial of degree 8. It turns out that if we run the training directly on the projected data, will not work well as the features would be badly scaled(e.g., an example with x=40 will now have a feature x8=408=6.5*1012). There fore, we will need to use feature normalization. Call featureNormalize and normalize the features of the training set, storing the mu, sigma parameters separately.

function [X_norm, mu, sigma] = featureNormalize(X)
%FEATURENORMALIZE Normalizes the features in X
% FEATURENORMALIZE(X) returns a normalized version of X where
% the mean value of each feature is 0 and the standard deviation
% is 1. This is often a good preprocessing step to do when
% working with learning algorithms. mu = mean(X);
X_norm = bsxfun(@minus, X, mu); sigma = std(X_norm);
X_norm = bsxfun(@rdivide, X_norm, sigma); % ============================================================ end

After learning the parameters θ, we we see two plots(Figure 4,5) generated for polynomial regression with λ=0.

From Figure 4, we should see that the poly nomial fit is able to follow the datapoints very well - thus, obtaining a low training error. However, the polynomial fit is very complex and even drops off at the extremes. This is an indicator that the polynomial regression model is overfitting the training data and will not generalize well.

Figure 5 shows the same effect where the low training error is low, but the cross validation error is high. There is a gap between the training and cross validation errors, indicating a high variance problem.

One way to combat the overfitting(hight-variance) problem is to add regularization to the model.

3.2 Adjusting the regularization parameter

In this section, we will get to observe how the regularization parameter affects the bias-vairance of regularized polynomial regression. Modify the lambda parameter and try λ=1,100.

For λ=1, we should see a polynomial fit that follows the data trend well(Figure 6) and a learning curve(Figure 7) showing that both the cross validation and training error converge to a relatively low value. This shows the λ=1 regularized polynomial regression model does not have the high bias or high variance problems. In effect, it achieves a good trade-off between bias and variance.

For λ=100, we should see a polynomial fit(Figure 8) that does not follow the data well. In this case, there is too much regularization and the model is unable to fit the training data.

3.3 Selecting λ using a cross validation set

A model without regularization(λ=0) fits the training set well, but does not generalize. Conversely, a model with too much regularization(λ=100) does not fit the training set and testing set well.

We will use a validation set to evaluate how good each λ value is. After selecting the best λ value using the cross validation set, we an then evaluate the model on teh test set to estimate how well the model will perform on actual unseen data.

We should use the trainLinearReg function to train the model using different values of λ and compute the train error and cross validation error. Try λ in the following range:{0,0.001,0.003,0.01,0.03,0.1,0.3,1,3,10}.

validationCurve.m:

function [lambda_vec, error_train, error_val] = ...
validationCurve(X, y, Xval, yval)
%VALIDATIONCURVE Generate the train and validation errors needed to
%plot a validation curve that we can use to select lambda
% [lambda_vec, error_train, error_val] = ...
% VALIDATIONCURVE(X, y, Xval, yval) returns the train
% and validation errors (in error_train, error_val)
% for different values of lambda. You are given the training set (X,
% y) and validation set (Xval, yval).
% % Selected values of lambda (you should not change this)
lambda_vec = [0 0.001 0.003 0.01 0.03 0.1 0.3 1 3 10]'; % You need to return these variables correctly.
error_train = zeros(length(lambda_vec), 1);
error_val = zeros(length(lambda_vec), 1); % ====================== YOUR CODE HERE ======================
% Instructions: Fill in this function to return training errors in
% error_train and the validation errors in error_val. The
% vector lambda_vec contains the different lambda parameters
% to use for each calculation of the errors, i.e,
% error_train(i), and error_val(i) should give
% you the errors obtained after training with
% lambda = lambda_vec(i)
%
% Note: You can loop over lambda_vec with the following:
%
% for i = 1:length(lambda_vec)
% lambda = lambda_vec(i);
% % Compute train / val errors when training linear
% % regression with regularization parameter lambda
% % You should store the result in error_train(i)
% % and error_val(i)
% ....
%
% end
%
% for i = 1:length(lambda_vec),
lambda = lambda_vec(i);
theta = trainLinearReg(X, y, lambda);
error_train(i) = linearRegCostFunction(X, y, theta, 0);
error_val(i) = linearRegCostFunction(Xval, yval, theta, 0);
end; % ========================================================================= end

We should see a plot similar to Figure 9. In this figure, we can see that the best value of λ is around 3. Due to randomness in the training and validation splits of the dataset, the cross validation error can sometimes be lower thatn the training error.

3.4 Computing test set error

To get a better indication of the model's performance in the real world, it is important to evaluate the "final" model on a test set that was not used in any part of training.

3.5 Plotting learning curves with randomly selected examples

In practivce, especially for small training sets, when you polt learning curves to debug your algorithms, it is often helpful to average across multiple sets of randomly selected examples to determin the training error and cross validation error.

Concretely, to determine the training error and cross validation error for i examples, you should first randomly select i examples from the training set and i examples from the cross validation set. We will then learn the parameters θ on the randomly chosen training set and evaluate the parameters θ using the randomly chosen training set and cross validation set. The above steps should then be repeated multiple times and averaged error should be used to determine the training error and cross validation error for i examples.

Figure 10 shows the learning curve we obtained for polynomial regression with λ=0.01.

CheeseZH: Stanford University: Machine Learning Ex5:Regularized Linear Regression and Bias v.s. Variance的更多相关文章

  1. Andrew Ng机器学习 五:Regularized Linear Regression and Bias v.s. Variance

    背景:实现一个线性回归模型,根据这个模型去预测一个水库的水位变化而流出的水量. 加载数据集ex5.data1后,数据集分为三部分: 1,训练集(training set)X与y: 2,交叉验证集(cr ...

  2. 第五次编程作业-Regularized Linear Regression and Bias v.s. Variance

    1.正规化的线性回归 (1)代价函数 (2)梯度 linearRegCostFunction.m function [J, grad] = linearRegCostFunction(X, y, th ...

  3. CheeseZH: Stanford University: Machine Learning Ex3: Multiclass Logistic Regression and Neural Network Prediction

    Handwritten digits recognition (0-9) Multi-class Logistic Regression 1. Vectorizing Logistic Regress ...

  4. CheeseZH: Stanford University: Machine Learning Ex1:Linear Regression

    (1) How to comput the Cost function in Univirate/Multivariate Linear Regression; (2) How to comput t ...

  5. CheeseZH: Stanford University: Machine Learning Ex2:Logistic Regression

    1. Sigmoid Function In Logisttic Regression, the hypothesis is defined as: where function g is the s ...

  6. CheeseZH: Stanford University: Machine Learning Ex4:Training Neural Network(Backpropagation Algorithm)

    1. Feedforward and cost function; 2.Regularized cost function: 3.Sigmoid gradient The gradient for t ...

  7. 机器学习---最小二乘线性回归模型的5个基本假设(Machine Learning Least Squares Linear Regression Assumptions)

    在之前的文章<机器学习---线性回归(Machine Learning Linear Regression)>中说到,使用最小二乘回归模型需要满足一些假设条件.但是这些假设条件却往往是人们 ...

  8. Andrew Ng Machine Learning 专题【Linear Regression】

    此文是斯坦福大学,机器学习界 superstar - Andrew Ng 所开设的 Coursera 课程:Machine Learning 的课程笔记. 力求简洁,仅代表本人观点,不足之处希望大家探 ...

  9. 机器学习---用python实现最小二乘线性回归算法并用随机梯度下降法求解 (Machine Learning Least Squares Linear Regression Application SGD)

    在<机器学习---线性回归(Machine Learning Linear Regression)>一文中,我们主要介绍了最小二乘线性回归算法以及简单地介绍了梯度下降法.现在,让我们来实践 ...

随机推荐

  1. GIS中的数据库.gdb与.mdb的区别

    gdb是文件地理数据库,mdb是个人地理数据库,都是数据库文件类型. 个人地理数据库,是以access数据库为基础的个人将数据库格式mdb,可以存储不超过2G的文件,只适合Windows系统下: 文件 ...

  2. iOS 字符串 中包含 % 百分号的方法

    百分号的转换,NSString中需要格式化的字符串中百分号使用%%表示,而char*中百分号也是使用%%表示. 例如:NSLog(@"%%%@%%",@"hello&qu ...

  3. 【elasticsearch】关于elasticSearch的基础概念了解【转载】

    转载原文:https://www.cnblogs.com/chenmc/p/9516100.html 该作者本系列文章,写的很详尽 ================================== ...

  4. gzip格式解压缩

    gzip格式解压缩 有时候网络请求中会出现gzip格式的数据,而我们无法通过常规办法进行解析: 这时候可以使用下面的这个工具来解决这个问题: https://github.com/mattt/Godz ...

  5. Spring框架中IoC(控制反转)的原理(转)

    原文链接:Spring框架中IoC(控制反转)的原理 一.IoC的基础知识以及原理: 1.IoC理论的背景:在采用面向对象方法设计的软件系统中,底层实现都是由N个对象组成的,所有的对象通过彼此的合作, ...

  6. 第二十五章 springboot + hystrixdashboard

    注意: hystrix基本使用:第十九章 springboot + hystrix(1) hystrix计数原理:附6 hystrix metrics and monitor 一.hystrixdas ...

  7. Go语言之高级篇beego框架之参数配置与路由配置

    一.参数配置 beego默认会解析当前应用下的conf/app.conf文件 1.1.beego的参数配置 appname = WEB httpport = 8080 runmode = dev 几种 ...

  8. log4net修改数据库连接字符串和写自定义信息

    最近项目需要用log4net来写日志,因为整个平台式在sharepoint上,我们需要记录具体是哪个子站点发生的日志,因此需要再原来的log表里面添加一个自定义信息列.由于平台的安全性要求,我们需要对 ...

  9. Android -- 状态栏高度

    干货 Class<?> c = null; Object obj = null; Field field = null; int x = 0, sbar = 0; try { c = Cl ...

  10. 双数组Trie树(DoubleArrayTrie)Java实现

    http://www.hankcs.com/program/java/%E5%8F%8C%E6%95%B0%E7%BB%84trie%E6%A0%91doublearraytriejava%E5%AE ...