Skip to main content

Auto Scaling Group

This post is part of a bigger topic Autoscaling Publishers in AWS.

Now that we have a Launch Configuration, based on it, we can create an Auto Scaling Group. This will be in charge of creating/terminating publisher instances based on some rules (Scaling Policies).

We base our Auto Scaling Group on the Launch Configuration created earlier (sdl_publisher_lc). The very important properties are the Desired, Min and Max number of instances in the group. The Desired is set usually by the Scaling Policies, but more about those later. The Min and Max are the limits of this group. In my case I use Min=1 and Max=3, meaning that I want to have 1 publisher running at all times and when needed, based on load, 2 additional publishers can be added to the group by an 'Increasing size' Scaling Policy.

Once load passes, a 'Decreasing size' Scaling Policy reduces the number of instances in the group.

Scaling Policies

These policies represent the rules for adding/removing instances to the group. They can monitor metrics on the instance itself (e.g. CPU), CloudWatch metrics, or even CloudWatch alarms (e.g. Publish Alarm defined earlier) in order to increase and decrease the number of instances.

We define an 'increase_group_size' as a Scaling Policy with Steps in order to add more publisher instances as the size of the Publish Queue increases.

We also define a 'decrease_group_size' as Simple Scaling Policy that reduces the size of the group. But more details about these policies in a followup post.

Lifecycle Hooks

We are going to use a lifecycle hook when scaling-in (decreasing) the size of out group, when publishing load has passed.

More details about the termination hook in a later post. For now, we create one hook that is going to be raised when the group attempts to terminate an instance. This hook can be intercepted in a CloudWatch event that can then trigger a Lambda Function that will instruct the publisher to shutdown gracefully. Once that happens, the termination hook is released and termination occurs normally.

The Lifecycle Hook Name is important, because in our Lambda Function we will instruct this particular hook to continue termination.

Heartbeat Timeout specifies the time needed to this hook to expire. This means that in the case the termination Lambda did not release the hook in the meantime, the hook will be automatically released once this timeout expires.


Popular posts from this blog

Running sp_updatestats on AWS RDS database

Part of the maintenance tasks that I perform on a MSSQL Content Manager database is to run stored procedure sp_updatestats . exec sp_updatestats However, that is not supported on an AWS RDS instance. The error message below indicates that only the sa  account can perform this: Msg 15247 , Level 16 , State 1 , Procedure sp_updatestats, Line 15 [Batch Start Line 0 ] User does not have permission to perform this action. Instead there are several posts that suggest using UPDATE STATISTICS instead: I stumbled upon the following post from 2008 (!!!), , which describes a way to wrap the call to sp_updatestats and execute it under a different user: create procedure dbo.sp_updstats with execute as 'dbo' as

Content Delivery Monitoring in AWS with CloudWatch

This post describes a way of monitoring a Tridion 9 combined Deployer by sending the health checks into a custom metric in CloudWatch in AWS. The same approach can also be used for other Content Delivery services. Once the metric is available in CloudWatch, we can create alarms in case the service errors out or becomes unresponsive. The overall architecture is as follows: Content Delivery service sends heartbeat (or exposes HTTP endpoint) for monitoring Monitoring Agent checks heartbeat (or HTTP health check) regularly and stores health state AWS lambda function: runs regularly reads the health state from Monitoring Agent pushes custom metrics into CloudWatch I am running the Deployer ( installation docs ) and Monitoring Agent ( installation docs ) on a t2.medium EC2 instance running CentOS on which I also installed the Systems Manager Agent (SSM Agent) ( installation docs ). In my case I have a combined Deployer that I want to monitor. This consists of an Endpoint and a

Debugging a Tridion 2011 Event System

OK, so you wrote your Tridion Event System. Now it's time to debug it. I know this is a hypothetical situtation -- your code never needs any kind of debugging ;) but indulge me... Recently, Alvin Reyes ( @nivlong ) blogged about being difficult to know how exactly to debug a Tridion Event System. More exactly, the question was " What process do I attach to for debugging even system code? ". Unfortunately, there is no simple or generic answer for it. Different events are fired by different Tridion CM modules. These modules run as different programs (or services) or run inside other programs (e.g. IIS). This means that you will need to monitor (or debug) different processes, based on which events your code handles. So the usual suspects are: dllhost.exe (or dllhost3g.exe ) - running as the MTSUser is the SDL Tridion Content Manager COM+ application and it fires events on generic TOM objects (e.g. events based on Tridion.ContentManager.Extensibility.Events.CrudEven