You can change the following settings in the Options pane of the pipeline. For more information, see Options Pane.
Specify the task to run for this modeling node. These tasks must be run sequentially. For example, the Diagnose task can be run independently, but the Fit task requires that you run the Diagnose task first.
To get a good model, enable some of the Feature Generation settings, which includes setting the number of dependent and independent variable lags. The generated features are included as independent variables when training the model.
Specify a positive integer for the number of dependent variable lags to generate and include as independent variables. Dependent variable lags are required for the neural network to learn the order of the time series.
For lags of missing values in the horizonthe number of intervals into the future, beyond a base date, for which analyses and predictions are made., the previous forecasted values are used to generate new lag values and extend the forecasta numerical prediction of a future value for a specified time period for each unique combination of BY variable values recursively. For example, setting the value to 3 computes three independent variables with lagged values of the dependent variable.
The default value is 4.
Specify a positive integer for the number of independent variable lags to generate and include as independent variables. For example, setting the value to 3 computes three variables with lagged values for each independent variable defined in the project.
The default value is 4.
Select this setting to generate seasonal dummy variables during feature extraction. The number of seasonal dummy variables corresponds to the Seasonal cycle length specified for the time variable.
This setting is enabled by default.
Enter a valid time interval value for creating seasonal dummy variables. If this value is left blank, the time interval specified for the time variable is used. For example, if the time variable for the project uses Week for the time interval, 52 seasonal dummy variables are generated. If you specify Month , then only 12 seasonal dummy variables are generated.
Enter one of the following specification values. Each specification shows the corresponding setting for the Time variable on the Data tab.
|
Time interval specification |
Time variable setting |
|---|---|
|
year |
Year |
|
yearv |
ISO 8601 |
|
r445yr |
Retail 4-4-5 year |
|
r454yr |
Retail 4-5-4 year |
|
r544yr |
Retail 5-4-4 year |
|
semiyear |
Semiyear |
|
r445qtr |
Retail 4-4-5 quarter |
|
r454qtr |
Retail 4-5-4 quarter |
|
r544qtr |
Retail 5-4-4 quarter |
|
quarter |
Quarter |
|
month |
Month |
|
r445mon |
Retail 4-4-5 month |
|
r454mon |
Retail 4-5-4 month |
|
r544mon |
Retail 5-4-4 month |
|
semimonth |
Semimonth |
|
tenday |
Ten-day |
|
week |
Week |
|
weekv |
ISO 8601 week |
|
weekday |
Weekday |
|
day |
Day |
|
hour |
Hour |
|
minute |
Minute |
|
second |
Second |
Specify Yes to use an ESM forecast of the dependent variable as an independent variable.
Specify the method to create a dependent variable trend as an independent variable. You can choose a Linear trend or Damped trend . The default is none None , in which no trend variable is created.
Specify the method that is used to standardize the interval input variables. Select from these options:
Specify 0, 1, or 2 hidden layers to include in the neural network model. If the number of hidden layers is 0, a GLIM model is trained.
You are required to specify the number of neurons and the activation function for each hidden layer.
Specify an integer between 0 and 100 for the number of neurons in the first hidden layer. This is required if Number of hidden layers is greater than 0.
Specify the activation function for the first hidden layer. This is required if Number of hidden layers is greater than 0.
Select from these options:
Specify an integer between 0 and 100 for the number of neurons in first second layer. This is required if Number of hidden layers is 2.
Specify the activation function for the second hidden layer. This is required if Number of hidden layers is 2.
Select from these options:
Specify whether direct connections from nodes in the input layer to nodes in the output layer should be included in the neural network. By default, this is not selected. This setting is disabled if Number of hidden layers is 0.
Specify Log for a logarithmic transformation for the dependent variable or None for no transformation.
Specify the method that is used to standardize the dependent variable. Select from these options:
Specify the error function for the dependent variable output layer. Select one of the following options:
Specify the activation function to use on the output layer of the network. If Error function is not set to Normal , this setting is disabled and the Exponential function is used.
Specify the target layer activation function for interval targets. Select from these options:
Specify the distribution of randomly generated initial neuron connection weights. Select from these options:
Specify a positive integer to use for generating random numbers to initialize the network.
Specify the optimization method used to train the neural network. Select one of the following options:
If you select the SGD algorithm, the following settings apply.
Specify the learning rate parameter for SGD optimization. The default value is 0.001.
Specify the annealing rate parameter for SGD optimization. The default value is 0.000001.
Specify the random seed to use for the SGD algorithm.
Specify the dropout ratio for the input layer when SGD optimization is used. The default value is 0. Enter a nonnegative number that is less than 1.
Specify the dropout ratio for the hidden layers when SGD optimization is used. The default value is 0. Enter a nonnegative number that is less than 1.
Specifies whether to create deterministic (reproducible) results using the specified SGD seed . Checking this can significantly increase run time.
Specify the number of times to train the network with different initial estimates for connection weights. The network with the smallest error is chosen as the optimal network.
Specify the maximum number of training iterations within each try.
Specify in minutes the maximum time allowed for each try. Training continues until the all tries have completed. Zero indicates not to use time to limit the training.
Specify a nonnegative multiplier for the L1 norm of the weights that is used in the neural network loss function. The default is 0.
Specify a nonnegative multiplier for the L2 norm of the weights that is used in the neural network loss function. The default is 0.
Select this option to stop training when the model begins to overfit. The training stops after a number of consecutive iterations without improvement in the holdout region. Set the number of consecutive iterations in Stagnation limit for early stopping .
Specify the number of consecutive iterations without improvement in validation error before stopping the optimization. Specifying 0 has the same effect as deselecting Enable early stopping . The default setting is 10.
This field is available only when Enable early stopping is selected.
Enable Autotune — Turn this setting on to enable autotuning of neural network parameters. Autotuning is applied only when running the Diagnose task.
Specify whether to autotune the number of hidden layers. If you enable this setting for autotuning, specify the initial value and the lower and upper bounds for the number of hidden layers. This setting overrides Number of hidden layers in the Model Initialization settings.
Specify whether to autotune the number of neurons in each hidden layer. Specify the initial value and the lower and upper bounds for the number of neurons. This setting overrides Number of hidden layers in the Model Initialization settings.
This setting is enabled if Hidden layer tuning is enabled.
Specify whether to autotune the nonnegative multiplier of the L1-norm of the weights in the loss function. This setting overrides L1 regularization in Model Training settings. This setting overrides L1 regularization in Model Training settings.
If you enable this setting for autotuning, specify the initial value and the lower and upper bounds for the multiplier. By default, the initial value and lower bound are 0 and the upper bound is 0.1.
Specify whether to autotune the nonnegative multiplier of the L2-norm of the weights in the loss function. This setting overrides L2 regularization in Model Training settings.
If you enable this setting for autotuning, specify the initial value and the lower and upper bounds for the multiplier. By default, the initial value and lower bound are 0 and the upper bound is 0.1.
Specify whether to autotune the learning rate for the hidden layers. If you enable this setting for autotuning, specify the initial value and the lower and upper bounds for the learning rate. This setting overrides Learning Rate in the Model Training settings.
This setting is disabled if the Algorithm under Model Training is set to LBFGS.
Specify whether to autotune the annealing rate for the hidden layers. If you enable this setting for autotuning, specify the initial value and the lower and upper bounds for the learning rate. This setting overrides Annealing Rate in the Model Training settings.
This setting is disabled if the Algorithm under Model Training is set to LBFGS.
Specify the maximum number of iterations for autotuning. Autotuning continues until all iterations are completed or the Maximum autotuning time (minutes) is reached, whichever is first.
Specify the maximum time allowed for autotuning in minutes. Autotuning continues until this time is reached or the Maximum autotuning iterations are completed, whichever is first.
Specify the random seed for the autotuning.
Specify the objective function to optimize when tuning parameters. You can select one of these options:
The holdout samplethe number of periods of the most recent data that should be excluded from the parameter estimation. The holdout sample can be used to evaluate the forecasting performance of a candidate model. is the minimum calculated value between these two settings.
Enter a positive integer to be used as the size of the holdout sample. The actual holdout sample is the minimum between this value and the Percentage of data points used in the holdout sample . The default value is zero, which means no holdout sample is used.
Enter a value between 0 and 100 to specify the percentage of the sample that is used for the holdout sample. The actual holdout sample is the minimum between this value and the Number of data points used in the holdout sample . This option is displayed only if Number of data points used in the holdout sample is greater than zero.
The following tables are automatically generated when running this modeling node. The tables can be saved to another caslib by using the Save Data Node.
Select any of the following optional tables that you want generated when this node is run. After the node is run, you can use the Save Data Node to save the tables.
The following tables can be selected only if Enable Autotune is turned on. If you select any of the autotune tables, you must set the task to Diagnose, Fit, or Update. The node fails if the task is set to Forecast.