How To: Improve Packager Performance


Overview


Below are suggestions that can improve packager performance.  Which suggestions may be most relevant can depend on which stages of packager are showing longer times spent in the packagerReport.html and/or which packaging folders and script types are showing longer times spent.  The first step is to look in your packager reports to see which stage(s) of packaging are taking the most time.

Note that although turning off certain features and settings can improve performance time, it can be a trade off since you may lose functionality.  Which features and settings to narrow down the scope of or disable may be different for each customer.

Please contact our Technical Support team if you would like assistance analyzing the timing shown in your packager reports and adjusting the settings that are relevant for your environment.

Timing in packagerReport.html


First run deployPackager and obtain timing data on the complete workflow in the packagerReport.html. You may want to run packager multiple times, with different types of scripts committed in different folders.  The timing will likely be different for some types of scripts compared to others because packager uses different processes.  Such as timing for ddl folder (when using the default ddl packageMethod=convert) versus timing for sql_direct folder or ddl_direct folder (or any folder with any name that uses either packageMethod=direct or packageMethod=ddl_direct) versus timing for function/procedure/etc stored logic folders (uses packageMethod=storedlogic).

Each packagerReport.html has six main sections:

  • Overview - Scripts Processed
  • Additional Logs & Reports
  • Deploy Packager Phases
  • Runtime Environment Properties
  • Packager Configuration Properties
  • All Properties

Here is an example of the whole report (click image to enlarge):

For analyzing performance timing look at the third section called Deploy Packager Phases.  The Deploy Packager Phases section breaks down the timing for each stage of packager:

  • Initialization
  • Detect Files to Package
  • Verify Metadata Properties
  • Run SQL Rules
  • Back up database
  • Package Files
  • Restore Database
  • Final Forecast
  • Final Deploy
  • Archiving DDL and DML scripts
  • Final checkin

The Deploy Packager Phases section shows the timing for each stage (click image to enlarge):

Items that can affect Packager Performance


1. When managing multiple schema in a multischema project, use the optional "schemaName" property in the metadata.properties file.  Here are details about it:

  • When configuring the repository for your SQL scripts, you may now specify which schema a script applies to in a multi-schema project. This will allow the Deployment Packager to limit it's change detection & recovery capabilities to only those schema that are impacted by the scripts being processed during that packager run. 
  • For example: If there are 30 schema in your project, but there are only scripts for 4 of the schema in this packaging run, then packager will be limited to the 4 relevant schema for that packaging run (instead of all 30 schema).  This can shorten the packager run time.  This only works if all of the scripts for that packaging run have the schemaName property set in the metadata.properties files in their folders.
  • To configure, set the value of a 'schemaName' property in the metadata.properties file for a specific directory. The value can be one schema name or a comma separated list of schema names. Do not put quotes around the schema values with Datical versions 6.12 or lower.  It is important to use casing for the value of the schemaName property that matches the casing of how your schema is listed in your datical.project file.  For example: schemaName=SCHEMA1.  Please see this page → Using the metadata.properties file
  • Setting this property does not affect script execution in any way.  It is used internally by the Deployment Packager to limit the scope of operations by schema.
  • We recommend using either schemaName property OR backup on demand:
    • If you are using backup mode "always", then the schemaName property will limit the schema list to the relevant schemas for three aspects of packager: backup, restore, and snapshot. 
    • However, if you are using backup mode "on demand" then the schemaName property will limit the schema list to the relevant schemas for only one aspect of packager: snapshot.   When you set backup to "on_demand" all of the schemas in the project will be backed up and restored, the schemas for backup/restore will NOT be limited based on the schemaName property.


2.  If packaging scripts in the "ddl" folder (when using Fixed Folders) or packageMethod=convert (when using Flexible Folders) is slow, consider using the optional "ddlExcludeList" property in the deployPackager.properties file to exclude certain object types from the snapshots that are used for before and after comparisons with ddl scripts.  If you are packaging stored logic objects in their recommended corresponding packaging folders (such as packaging function scripts in the "function" folder, and packaging procedure scripts in the "procedure" folder) then you do NOT need to have those object types in the snapshots used by the convert method when packaging ddl scripts in the "ddl" folder.  Excluding stored logic object types from the snapshots used in the "ddl" folder can improve performance, especially if you have a lot of stored logic objects in your database. 

Please see "ddlExcludeList" in this document → Using the required deployPackager.properties file


3.  Use backup "on_demand" to create static backups, such as a nightly backup that can be used by packaging jobs as needed.  You would have two different automation jobs.  One job is your existing packager job that processes your new sql scripts.  The other job will be a job that runs packager in preview mode each night to only create the backup file.  Please see this document for the appropriate settings for each job → Managing Database Backup and Restore

  • If you want to use static on demand backups with Oracle or SQL Server, we recommend running Datical DB versions 5.6 or higher.  
  • If you want to use static on demand backups with Postgres, we recommend running Datical DB versions 6.12 or higher.
  • Note that if you use backup on_demand, you will NOT get the full benefits of using the schemaName property in the metadata.properties file:
    • If you are using backup mode "always", then the schemaName property will limit the schema list to the relevant schemas for three aspects of packager: backup, restore, and snapshot. 
    • However, if you are using backup mode "on demand" then the schemaName property will limit the schema list to the relevant schemas for only one aspect of packager: snapshot.   When you set backup to "on_demand" all of the schemas in the project will be backed up and restored, the schemas for backup/restore will NOT be limited based on the schemaName property.
  • Troubleshooting performance time of on_demand static backups:  If you are implementing static/on_demand backups but have not seen a significant improvement in how long your packager jobs take to run, check your configuration.  If you have set databaseBackupMode=on_demand but are still using createDatabaseBackup=true in your main packager jobs that process scripts, that is an unusual configuration.  Packager will work with that configuration but it will still create a new backup file for each packager job that processes scripts, and therefore you would NOT be getting the possible performance benefit that you would have in the more typical configuration of creating a nightly backup separately and re-using that backup file when processing scripts (to avoid running backup each time).  To optimize the on_demand backup, do NOT use createDatabaseBackup=true with your main packaging job that processes scripts (assuming that the backup file was already created and is in place). 


4.  For Oracle, you can set the "parallel" property in the deployPackager.properties file for Oracle backup/restore as described in these pages:


5.  Change your Row Count setting to "approximate" or "disabled" (because "approximate" is significantly faster than "exact") → Settings for Collecting Row Counts

  • If you use rules that relate to row count, set this to "approximate".  If you do not use rules relating to row count, set this to "disabled".

There are different ways you can set the row count option, use the method you prefer:


6.  Use the optional Limited Forecast which is typically faster than Full Forecast, please see this page →  Limited Forecast

  • Caveat: This could have an impact on Rules enforcement if using Rules.


7.  Run packager as a different user than the schema owner (so packager drops the schema and re-creates it, instead of clearing out each object).


8.  For Oracle/SQL Server/DB2, change the Stored Logic Validity Check to "local" (the default) or "limited" or "disabled", depending on which features you actively use.

  • Please see the notes in these pages:
  • Performance recommendations:
    • If you do not review or use the information in the Stored Logic Validity Check section of your deploy reports, then set storedLogicValidityCheck="disabled" to avoid possible performance slow down for a feature you aren't actively using.
    • If you review and use the Stored Logic Validity Check information in your deploy reports but you do not use the storedLogicValidityAction=FAIL option, then we recommend setting storedLogicValidityCheck="limited".
    • If you review and use the Stored Logic Validity Check information in your deploy reports and you also have enabled the storedLogicValidityAction=FAIL option, then we recommend setting storedLogicValidityCheck="local".
    • Although storedLogicValidityCheck="global" is an available setting and is the most comprehensive, if performance timing is an important consideration then it may be better to use a smaller scope such as "local" or "limited".
  • There are different ways you can set the stored logic validity check level, use the method you prefer to set the value:


9.  Packaging ddl scripts from the sql_direct folder (packageMethod=direct) or from the ddl_direct folder (packageMethod=ddl_direct) is typically faster than from the ddl folder (packageMethod=convert).  You could also opt to set packageMethod=ddl_direct for your ddl folder for better performance.

  • Caveat: If you are not using SQL Parser for Oracle, then only sqlrules would apply in sql_direct and ddl_direct folder, or any folder with any name that uses packageMethod=ddl_direct or packageMethod=direct.  (Other types of rules and forecast modeling do NOT apply for ddl_direct/direct/sql_direct changes if you are not using SQL Parser for Oracle.)
  • If you are using Oracle with a recent version of Liquibase Enterprise/Datical DB 7.x, you could consider using SQL Parser for Oracle to add forecast modeling and forecast rules.
    • When you enable the SQL Parser for Oracle project setting, SQL Parser is applicable by default to the ddl_direct folder (packageMethod=ddl_direct), the sql_direct folder (packageMethod=direct), and sql folder (packageMethod=sqlfile).  
    • You could also opt to set packageMethod=ddl_direct for your ddl folder using flexible folder configuration so that folder would also use SQL Parser.  Using SQL Parser with packageMethod=ddl_direct/packageMethod=direct for ddl would be faster than using the packageMethod=convert for ddl (convert is the default for ddl).  You can change the packageMethod for the ddl folder in the metadata.properties file for the ddl folder.  Using packageMethod=ddl_direct would make that folder be processed near the beginning of the folder order.  Or using packageMethod=direct would make that folder be processed near the end of the folder order.
    • If your DML scripts are quite large, for performance reasons you could disable SQL Parser for any folder where large DML scripts may be packaged.  You can disable SQL Parser at the folder level by setting disableSqlParser=true in the metadata.properties file for that folder.  Note that disableSqlParser=true is already set by default for the DATA_DML folder in newer versions 7.6 and higher.
    • There were improvements to SQL Parser for Oracle in versions 7.8 and 7.12.  We recommend upgrading to a recent 7.x version if you are using SQL Parser for Oracle.
    • Please see these pages:


10.  Having the build agent and the database in close proximity can help performance.


11.  If you have enabled the Automatically Generate SQL setting you may see performance improvements by turning off that project setting.  

There are different ways you can disable this setting, use the method you prefer:


12.  Certain types of script errors (such as missing end delimiter) can cause a packager build job to hang indefinitely during the deploy section of packager. To avoid having this happen, you can set a Script Execution Timeout for your REF environments. 

  • It is recommended to set it on your REF dbDef step only, not for your higher environments. 
  • You can configure the number of seconds packager should wait before timing out when deploying a script, for example 600 seconds (10 minutes).  You can use the number of seconds that seems appropriate for the types of scripts you typically package.  The timeout period is per each individual script. 
  • If a script is taking longer to deploy than the timeout value, packager build will fail with a timeout error and indicate which script timed out.  Having packager fail with a timeout error allows packager to finish all of its steps normally, including restoring/reverting REF DB to its previous state from prior to the error.  This is better than having packager hang indefinitely then terminating the job manually, which could leave REF in an unexpected state with an incomplete deployment. 

Note that the Script Execution Timeout only applies to changes executed using SQLPlus/SQLCMD/CLPPlus/EDBPlus.  It does NOT apply to scripts packaged in the function/procedure/package/view/trigger/sql folders (when using fixed folders) or to folders using packageMethod=STOREDLOGIC or packageMethod=SQLFILE (when using flexible folders).

There are different ways you can set the timeout, use the method you prefer:


13.  If you don't need to use the Forecast DML project setting, disable it.

There are different ways you can disable this setting, use the method you prefer:

  • In the desktop client/Eclipse GUI, go to the Settings tab.  Expand the Deployment Settings section.  Disable the "Forecast Data Modification Changes (DML)" checkbox.
  • Disable it using the hammer CLI command:  "hammer set forecastDML false"
  • Recommendations for Working with Large DML Scripts


14.  Consider increasing the amount of RAM used by Datical using the Xmx setting.  See the instructions here: Increase the amount of RAM used by Liquibase Enterprise


15.  Upgrade to a current version of Liquibase Enterprise/Datical:

  • There were performance improvements for those who use the Stored Logic Validity Check project setting in Datical DB version 6.12 (and higher).
  • There were performance improvements for Status, statusDetails, and Pipeline Status in Datical DB version 6.14 (and higher). This also improved operations which run status implicitly such as deploy, rollback, deployPackager, convert SQL and changeLogSync.  
  • There were performance improvements specifically for multi-database/multi-catalog configurations of SQL Server projects in Datical DB version 6.16 (and higher).
  • There is a fix for the DATAPUMP API Oracle backup and restore in Datical DB version 6.16 (and higher) to better handle running multiple packager jobs concurrently.
  • There is a new cleanup command for packager in Datical DB versions 7.3 (and higher).  The cleanup command can be run after any time that you might need to manually interrupt a packager build job midway.  The cleanup command is to unblock subsequent packager jobs after a manual interruption by clearing the locks on DATABASECHANGELOGLOCK and DATICAL_SPERRORLOG tables and also restoring REF.  Please see this page for more details: How To: Use ReleaseLocks Command and Packager with Cleanup Option
  • With Datical DB version 7.6 (and higher), there is a new feature that prevents continuing to run packager jobs after a backup error or restore error.  Please see the "Recovering from a Backup or Restore Failure" section here for more details: Recovering from a Backup or Restore Failure
  • There were improvements for memory utilization of SQL scripts that produce a high-volume output in 7.11 (and higher).
  • There were improvements for SQL Parser for Oracle in version 7.12 (and higher). 
  • There were improvements for Limited Forecast in version 7.13 (and higher):
    • Limited Forecast will only profile tables impacted by the changesets to be forecasted or deployed
    • Limited Forecast will only profile the schema impacted by the changesets to be forecasted or deployed in multi-schema projects
  • Significant performance improvements for Forecast profiling in version 7.14 (and higher):
    • Faster forecasting of Views and Materialized Views
    • Faster profiling for tables, columns, and views in multi-schema projects
    • Use multiple connections (maximum of 10 connections) to profile schemas simultaneously in Oracle multi-schema projects.  Note that with 7.14 (and higher) in Oracle projects with multiple schemas, you may notice higher CPU utilization due to multiple connections being used for Oracle forecast profiling.


16.  Although not specifically about packager, it may also be useful to check the items on these pages that may improve Deploy performance:




Copyright © Datical 2012-2020 - Proprietary and Confidential