Overview of Model Comparison, Validation, and Summary Reports

What Are Model Comparison, Validation, and Summary Reports?

The SAS Model Manager model comparison, validation, and summary reports are tools that you can use to evaluate and compare the candidate models in a version or across versions to help you select and approve the champion model that moves to production status. The SAS Model Manager model comparison reports are analytical tools that project managers, statisticians, and analysts can use to assess the structure, performance, and resilience of candidate models. The model validation reports use statistical measures to validate the stability, performance, and calibration of Basel II risk models and parameters. The training summary data set report creates frequency and distribution charts that summarize the train table variables.
The reports present information about a number of attributes that can affect model performance. Together, the reports provide qualified information that can serve as the analytical basis for choosing and monitoring a champion model.
Here is a description of the comparison reports:
For a single model, this report displays the profile data that is associated with input, output, and target variables. Profile data includes the variable name, type, length, label, SAS format, measurement level, and role.
This report compares the profile data for two models and notes the differences.
The Dynamic Lift report provides visual summaries of the performance of one or more models for predicting a binary outcome variable.
The Interval Target Variable report creates two plots for you to view the actual versus predicted values for a model and the actual versus residual values for a model. Interval Target Variable report can be created only for prediction models.
These are the Basel II model validation reports:
The Loss Given Default (LGD) report calculates the amount that might be lost in an investment and calculates the economic or regulatory capital for Basel II compliance.
The Probability of Default (PD) Validation report estimates the probability of defaulting on a debt that is owed. Probability of default is used to calculate economic or regulatory capital for Basel II compliance.
The model validation reports use statistical measures that report on these model validation measures:
  • The model stability measures track the change in distribution for the modeling data and scoring data.
  • The model performance measures check the model’s ability to distinguish between accounts that have not defaulted and accounts that have defaulted, as well as report on the relationship between actual default probability and predicted default probability.
  • The model calibration measures check the accuracy of the selected models for the LGD and the PD reports by comparing the correct quantification of the risk components with the available standards.
This is the train table data set summary report:
The Training Summary Data Set report creates frequency and distribution charts to validate data set variables.
You create the model comparison, validation, and summary reports using the New Report window that you start from a version's Reports node:
New Report Window

The Model Comparison, Validation, and Summary Report Input Files

SAS Model Manager uses a test table as the input table for the Dynamic Lift report and the Interval Target Variable report.
Before you can create a Dynamic Lift report or the Interval Target Variable report, make sure that a test table has been added to the SAS Metadata Repository using SAS Management Console or that a libref has been defined in the Edit Start-up Code window for the SAS library where the test table resides. The test table can be viewed in the Data Sources category view. Then, specify the test table in the project property Default Test Table.
You specify the input table for Basel II validation reports in the New Report window. The input file for the validation reports can contain only input variables or it can contain input and output variables. If the input table contains input and output variables, the report generation does not need to run a scoring task to obtain the output variables.
To create the input table for the training summary data set report, make sure that the train table has been added to the SAS Metadata Repository or that a libref has been defined in the Edit Start-up Code window for the train table library. The train table must be specified in the project or version property for Default Train Table.

The Model Comparison, Validation, and Summary Report Output Files

The New Reports window stores the model comparison, validation, and summary report output files in a report node under the Reports node. The name of the report node is the value of the Name box that you specified in the New Report window Report Properties table.
Each time you create a report, these files are generated:
  • the report in either HTML, PDF, RTF, or EXCEL format
    Note: The Loss Given Default and Probability of Default reports can be created only in PDF and RTF formats.
  • taskCode.log
  • taskCode.sas
CAUTION:
The wizard overwrites the output files if output files of the same name already exist.
Here is a description of the model comparison output files:
Report File
Description
report-name.html
This file is the report output in HTML format.
report-name.pdf
This file is the report output in PDF format.
report-name.rtf
This file is the report output in RTF format.
report-name.xls
This file is the report output in Excel format.
taskCode.log
This file is the log file that contains messages from running the SAS code to create the report.
taskCode.sas
This file is the SAS code that is used to create the report.
After you create a report, you can view the report from the Reports node:
Report Folder with Files
Note: If you save a report to a local drive, images in the reports, such as graphs, do not appear. The report images are separate files and are stored in the SAS Content Server. Always view reports from the Reports node.