Tools Cross-Validation: Difference between revisions

From Eigenvector Research Documentation Wiki
Jump to navigation Jump to search
imported>Jeremy
No edit summary
 
(23 intermediate revisions by 2 users not shown)
Line 1: Line 1:
__TOC__
__TOC__
[[TableOfContents|Table of Contents]] | [[ModelApplication_ValidationPhase|Previous]] | [[Tools_ModelRobustness|Next]]
[[TableOfContents|Table of Contents]] | [[ModelApplication_ValidationPhase|Previous]] | [[Tools_ModelRobustness|Next]]


Line 6: Line 7:
You use the Cross-Validation tool to:
You use the Cross-Validation tool to:


{| style="
{|  


|-
|- valign="top"


|
|
Line 15: Line 16:
|}
|}


{| style="
{|  


|-
|- valign="top"


|
|
* Estimate the performance of a model when you apply the model to unknown data.  
* Estimate the expected performance of a model when you apply the model to unknown data.  


|}
|}
Line 26: Line 27:
For a given set of data, cross-validation involves a series of steps called subvalidation steps in which you remove a subset of objects from a set of data (the test set), build of a model using the remaining objects in the set of data (the model building set), and then apply the resulting model to the removed objects. You note how the errors accumulate as you leave out samples to determine the number of principal components/latent variables/factors to retain in the model. Cross-validation typically involves more than one subvalidation step, each of which in turn involves the selection of different subsets of samples for model building and model testing. In Solo, five different cross-validation methods are available, and these methods vary with respect to how the different sample subsets are selected for these subvalidation steps.  
For a given set of data, cross-validation involves a series of steps called subvalidation steps in which you remove a subset of objects from a set of data (the test set), build of a model using the remaining objects in the set of data (the model building set), and then apply the resulting model to the removed objects. You note how the errors accumulate as you leave out samples to determine the number of principal components/latent variables/factors to retain in the model. Cross-validation typically involves more than one subvalidation step, each of which in turn involves the selection of different subsets of samples for model building and model testing. In Solo, five different cross-validation methods are available, and these methods vary with respect to how the different sample subsets are selected for these subvalidation steps.  


{| style="
{|  


|-
|- valign="top"


|1.
|1.
Line 36: Line 37:
|}
|}


{| style="margin-left:18pt"  style="
{| style="margin-left:18pt"   


|-
|- valign="top"


|
|
Line 45: Line 46:
|}
|}


{| style="margin-left:18pt"  style="
{| style="margin-left:18pt"   


|-
|- valign="top"


|
|
* Click the Cross-Validation icon in the Analysis window.
* Click the Cross-Validation icon in the Analysis window.
 
'''Note:''' You must load data into the Analysis window before the Cross-Validation icon is available.
|}
|}


Note: You must load data into the Analysis window before the Cross-Validation icon is available.
:''Cross-validation icon in the Analysis window''
 
::''Cross-validation icon in the Analysis window''
 
::[[Image:Cross_validation_icon_Analysis_window.png|406x83px]]
::[[Image:Cross_validation_icon_Analysis_window.png|406x83px]]
::
::


{| style="margin-left:18pt"  style="
{| style="margin-left:18pt"   


|-
|- valign="top"


|
|
Line 70: Line 68:
|}
|}


{| style="
{|  


|-
|- valign="top"


|2.
|2.


|In the Cross-Validation dialog box, select the method of cross-validation that you want to use.
|In the Cross-Validation dialog box, select the method of cross-validation that you want to use. You will notice a visual representation of the splits at the top of the window. Other relevant information and warnings will appear at the bottom of the window.


|}
|}


::''Cross-Validation dialog box''
:''Cross-Validation dialog box''


::[[Image:Cross_validation_icon_dialog_box.png|288x138px]]
::[[Image:CrossValGUI.png|500px||CrossVal Window]]
::
::


{| style="
{|  


|-
|- valign="top"


|3.
|3.
Line 95: Line 93:
|}
|}


Note: Not all parameters are relevant for all cross-validation methods. The initial values that are specified for the available parameters are default values that are based on the dimensionality of the data. You can click Reset at any time to reset the parameters to their default settings. For the following descriptions:
:'''Note:''' Not all parameters are relevant for all cross-validation methods. The initial values that are specified for the available parameters are default values that are based on the dimensionality of the data. You can click Reset at any time to reset the parameters to their default settings. See the [[Using_Cross-Validation]] page for details on how to use the different methods and settings.
 
{| style="
 
|-
 
|
* n is the total number of objects in the set of data.
 
|}
 
{| style="
 
|-
 
|
* s is the number of data splits specified for the cross-validation procedure, which must be less than n/2.
 
|}
 
{| style="
 
|-
 
|
* r is the number of iterations.
 
|}
 
::''Cross-validation methods compared''
 
::
::
{| style="
 
|-
 
|
 
''''''
 
|
 
'''Leave One Out'''
 
|
 
'''Venetian Blinds'''
 
|
 
'''Contiguous Block'''
 
|
 
'''Random Subsets'''
 
|
 
'''Custom'''
 
|-
 
|
Cross-validation method
 
|
[[Image:CV_Leave_One_Out.jpg|86x126px]]
 
|
[[Image:CVB_VenetianBlinds.jpg|84x126px]]
 
|
[[Image:CV_ContinguousBlocks.jpg|85x126px]]
 
|
[[Image:CV_RandomSubsets.jpg|85x127px]]
 
|
 
|-
 
|
Description
 
|
The default value. All samples in the set of data are used to build the model.
 
|
Each test set is determined by selecting every sth object in the set of data, starting at objects numbered 1 through s.
 
|
An alternative to Venetian Blinds. Each test set is determined by selecting contiguous blocks of n/s objects in the set of data, starting at object number 1.
 
|
"s" different test sets are determined through random selection of n/s objects in the set of data, such that no single object is in more than one test set. This procedure is repeated "r" times, where "r" is the number of iterations.
 
|
You manually define each of the test sets. You can assign specific objects in your set of data in one of three ways:
 
{| style="
 
|-
 
|
* To be in every test set.
 
|}
 
{| style="
 
|-
 
|
* To never be in a test set.
 
|}
 
{| style="
 
|-
 
|
* To not be used in the cross- validation procedure at all.
 
|}
 
|-
 
|
Available Parameters
 
|
Maximum Number of LVs
 
|
 
{| style="
 
|-
 
|
* Maximum Number of LVs
 
|}
 
{| style="
 
|-
 
|
* Number of Data Splits
 
|}
 
|
 
{| style="
 
|-
 
|
* Maximum Number of LVs
 
|}
 
{| style="
 
|-
 
|
* Number of Data Splits
 
|}
 
|
 
{| style="
 
|-
 
|
* Maximum Number of LVs
 
|}
 
{| style="
 
|-
 
|
* Number of Data Splits
 
|}
 
{| style="
 
|-
 
|
* Number of Iterations
 
|}
 
|
 
{| style="
 
|-
 
|
* Number of data splits
 
|}
 
{| style="
 
|-
 
|
* Object membership for each split
 
|}
 
{| style="
 
|-
 
|
* Total number of objects
 
|}
 
|-
 
|
<nowiki>#</nowiki> of Subvalidation Steps
 
|
n
 
|
s
 
|
s
 
|
(s*r)
 
|
s
 
|-
 
|
<nowiki>#</nowiki> of Test Samples per Subvalidation
 
|
1
 
|
n/s
 
|
n/s
 
|
n/s
 
|
Varies. User-defined.
 
|}


{| style="
{|  


|-
|- valign="top"


|4.
|4.
Line 380: Line 105:
|}
|}


{| style="margin-left:18pt"  style="
{| style="margin-left:18pt"   


|-
|- valign="top"


|
|
Line 389: Line 114:
|}
|}


{| style="margin-left:18pt"  style="
{| style="margin-left:18pt"   


|-
|- valign="top"


|
|
Line 397: Line 122:


|}
|}
===Technical Details===
More technical details can be found on the [[Using_Cross-Validation]] page.

Latest revision as of 12:16, 11 August 2020

Table of Contents | Previous | Next

Cross-Validation Tool

You use the Cross-Validation tool to:

  • Assess the optimal complexity of a model (for example, the number of principal components in a PCA or PCR model, or the number of latent variables in a PLS model).
  • Estimate the expected performance of a model when you apply the model to unknown data.

For a given set of data, cross-validation involves a series of steps called subvalidation steps in which you remove a subset of objects from a set of data (the test set), build of a model using the remaining objects in the set of data (the model building set), and then apply the resulting model to the removed objects. You note how the errors accumulate as you leave out samples to determine the number of principal components/latent variables/factors to retain in the model. Cross-validation typically involves more than one subvalidation step, each of which in turn involves the selection of different subsets of samples for model building and model testing. In Solo, five different cross-validation methods are available, and these methods vary with respect to how the different sample subsets are selected for these subvalidation steps.

1. To open the Cross-Validation tool, do one of the following:
  • On the Analysis window, click Tools > Cross-Validation.
  • Click the Cross-Validation icon in the Analysis window.

Note: You must load data into the Analysis window before the Cross-Validation icon is available.

Cross-validation icon in the Analysis window
Cross validation icon Analysis window.png
  • In the Analysis window Flowchart pane, click Choose Cross-Validation.
2. In the Cross-Validation dialog box, select the method of cross-validation that you want to use. You will notice a visual representation of the splits at the top of the window. Other relevant information and warnings will appear at the bottom of the window.
Cross-Validation dialog box
CrossVal Window
3. Use the slider bars to change the default values for the available parameters.
Note: Not all parameters are relevant for all cross-validation methods. The initial values that are specified for the available parameters are default values that are based on the dimensionality of the data. You can click Reset at any time to reset the parameters to their default settings. See the Using_Cross-Validation page for details on how to use the different methods and settings.
4. Do one of the following:
  • Click Apply button to apply these settings and keep the Cross Validation dialog box open.
  • Click OK to apply these settings and close the Cross Validation dialog box.

Technical Details

More technical details can be found on the Using_Cross-Validation page.