The Patch Management Architecture consists of the following components:
Fig: Patch Management Architecture
The External Patch Crawler at Zoho Corp. is a critical component in maintaining cybersecurity. It continuously probes the internet and vendor websites, such as those of Microsoft, Apple, Linux distributions, and other third-party applications, to gather vulnerability information. The process of managing this data involves several key steps such as:
Only Windows and Mac patches will be tested before release to the central repository. It is recommended that Linux patches be tested in an environment before deployment.
The entire process, from information gathering to patch analysis and publishing, is conducted periodically, ensuring that updates are incorporated into the Central Patch Repository promptly.
Once the updates are supported the corresponding patch details will be available in the Central Patch Repository.
Once supported, the updates will be available in the product after the sync between the Central Patch Repository and the Patch Manager Plus server has occurred. as per the configured time (once every day at the time you have configured in the Patch Database settings). You can also manually initiate this sync for the patches to show up after they have been supported.
The Central Patch Repository is a portal in the Zoho Corp. site, which hosts the latest vulnerability database that has been published after a thorough analysis. This database is exposed for download by the Patch Manager Plus server situated in the customer site, and provides information required for patch scanning and installation. The data provided is encrypted and transferred over HTTPS.
The Patch Manager Plus Server is located at the enterprise (customer site) and subscribes to the Central Patch Repository, to periodically download the vulnerability database. The server:
Patch Management is a two-stage process:
Patch Assessment or Scanning
The systems in the network are periodically scanned to assess the patch needs. Using a comprehensive database consolidated from Microsoft's and other bulletins, the scanning mechanism checks for the existence and state of the patches by performing file version checks, registry checks and checksums. The vulnerability database is periodically updated with the latest information on patches, from the Central Patch Repository. The scanning logic automatically determines which updates are needed on each client system, taking into account the operating system, application, and update dependencies.
On successful completion of an assessment, the results of each assessment are returned and stored in the server database. The scan results can be viewed from the web console.
Patch Download and Deployment
On selecting the patches to be deployed, the server downloads the patches from the vendor website and verifies their integrity using checksum. Following this, the agent downloads the patches from the server. The URL of the patches downloaded from the server will be validated with the checksum. Patch binaries will be validated with checksum during the download and each time installation is initiated.