Vembu Community and Free Editions Updates

There are lots of resources available which help technical practitioners and decision makers evaluate options. One of the most valuable things that I’m enjoying with partner solutions that I work with is the community and free editions that are available. I run a reasonable sized lab environment which is ideal for using these platforms.

Vembu has been doing a lot around opening up more freely available tools to the market which helps get more environments protected and more of us involved in active testing of their products. I’m very pleased with the product experience and will be sharing those experiences in my upcoming data protection series.

Before you read those posts, I highly recommend taking some time and seeing what solutions may map out for you. With the timing of VMworld 2019 very near, it’s also super important for you to know what works and what may have any limitations based on your requirements. This means that you can have a more data-backed conversation around data protection as you evaluate potential solutions.

Free VMware backup software

VMware is the clear leader in virtualization but lacks an out-of-the-box data protection capability.  Despite years of taking a run at an internal offering, the rise in the data protection market shows that incumbent vendors generally can not get the same results on management features within their own products.  It’s ideal to look at the best-of-breed options by dedicated platform specialists.  VMware backup for free is part my own lab environment and I even use it on some cloud-hosted vSphere resources and store both on-premises and out in AWS using S3.

Product Page:

Data Sheet Page:  Product Datasheet

Free Backup for Hyper-V

I am seeing more and more Hyper-V implementations and the ease of building out a Hyper-V (also free with the Windows license) environment for labs is a great exercise to make sure to keep in tune with the latest happenings from the Microsoft virtualization camp.  This is also where having a free backup alternative is super helpful and I’ve been using Vembu as part of my lab infrastructure for Hyper-V as well.

Product Page:

Data Sheet page: Product Datasheet

Free Backup for Windows Server

Hey, don’t forget that you may have some standalone servers which need protection as well!  Despite everything being virtualized (which is not 100% true) and everything going to the cloud (also, not 100% true), the rest of us have a bunch of servers which need a little care and feeding, including data protection.

Product Page:

Data Sheet Page:  Product Datasheet

The differences you will find in the free versus the paid versions of any of the suite of tools is available in the online feature chart here:

Most notably different is the full support for CBT (Change Block Tracking) and CDP (Continuous Data Protection) which will be limited to a small number of virtual machines in the free editions.  Despite the limited count, you can run some full-featured testing using a subset of the environment which will give you an idea on the performance and capabilities.

Please let me know if there are any specific features you want to have explored and shared in the series.  Big thanks go out to Vembu for their support of the community as well which includes things like blog sponsorship and tons of information sharing with the blogging community.

Very cool! VeeamON Virtual Conference

One of the greatest things about conferences is the ability to get very up-close access to engineers and other technology professionals.  What makes the upcoming VeeamON virtual conference on December 5th a double-win is that you get that access but you don’t need to book travel and get approvals for budget!

Virtual Event, Real Value

Sign up today for the event and this will also get you access to great live content, post-event recordings and materials, plus there are prizes given away throughout the event day which is always a nice bonus.

There are three tracks (Business, Technical, Cloud) that can satisfy lots of different attendees and audiences.  I’m honored to know most of the folks presenting, so I can attest to their skill and presentation abilities.  This will be a must-attend IMHO!  There are not many times where you can get to interact with people at the event from the comfort of a chair and switch tracks without having to run from one building to another.

Did I mention prizes?!  This is pretty awesome…

There’s definitely nothing to lose in joining the event.  Register for free at the event site ( and please do let me know what you thought of the content as I’ll be watching it closely myself.  Always great to share notes!


Two of Your Nines Don’t Need Five Nines

The99,9 % uptimere’s a pretty good chance that you have a number of environments that are tagged as having a five nines requirement for availability That don’t actually need it. We often find ourselves getting confused with availability and the true cost of keeping application environments up on the ever elusive 99.999% uptime.

Breaking Down the Nines

It is eye opening when we break down the actual downtime allowed in the sense of minutes per year when we look at the level of nines as listed on Wikipedia (


So, when we think about the cost of maintaining an uptime level for the application, it becomes important to see the real numbers in relation to availability. The cost of achieving 99.999 versus 99 is significant.

Fine Nines, Nine to Five

This is the real situation that many of us need to deal with. While we talk about the criticality of an application environment, it’s often thought about as critical only when it’s in active use. Most folks would obviously look at a straight 99 percent uptime with 87 hours and balk at that as a suggested availability. Here’s the catch, though. What we are really looking for is a five nines availability, but only during access hours. Many, if not most, of our internal business applications are only accessed during the day inside office hours.

Even if we span across time zones, the reality is that we aren’t using the applications during a decent amount of time in the day. Assuming that your application needs to cover time zones that span a continent, you are probably needing to cover a 10 hour day with a maximum of a 5 hour variance, totaling 9 hours a day that it is not needed for primary use. That means that you can effectively sustain 2964 hours…yes hours…of downtime. That means 177,840 minutes.

Does this mean we can shut them off? Well, not quite. Let’s talk about why.

The Highly Used Unused Application

Applications are considered active during a certain window which I refer to as primary use. There is a non-primary use set of processes which happen as well.

Backing up the environment is a good example of this. Backups tend to run off hours so as not to collide with primary use inside business hours. Security scans and other safety practices also take place outside of core hours to help with keeping application performance more stable during primary use hours.

Because of this we do still have requirements to keep the application platforms available to run these other operational processes.

Scale-back Versus Shut Down

As your application environments are being architected or refactored, it is good to think about the importance of a microservices approach and why it can help with this issue.  I know that there are assumptions around the fact that we are choosing when system availability occurs, but the important part of this discussion is that you may be paying for a surprising amount of warranty on systems that don’t need it.

We can see that we can’t really just power off servers a lot of the time because of the backups, security scans, and other non-primary use access. What we can do is to use a scale-out and scale-back approach.

Web applications may need the back-end to be continuously available but at different levels of usage. During off hours, why not have less front-end servers? Data layers can stay up, but can also be scaled down.

Some applications like file servers and less variable use applications will not do well in scale-up/scale-down scenarios. That’s ok. We have to accept that hybrid approaches are needed across all areas of IT.

Why is This Important?

Think ahead. When the architecture is being evaluated for production and disaster recovery, we should be thinking about primary use and availability as well as the non-primary use functions like data protection.

All of a sudden, those buzzwords like microservices and containers with infrastructure as code seem to make some sense. Should you be racing to refactor all of your apps? No. Should you be continuously evaluating the environment? Yes.

Most importantly, be aware of the true cost of the five nines and whether you really need it for all of your applications.

Interesting results from a Vision Solutions survey

I’m always watching for stats and numbers in the industry. We are continuously presented with the “fastest growing” and “x% better than the competitor” based on a number of sometimes skewed statistics. While I love what information I can gather from statistics for hardware and software, it is almost always based on sales.

When I was given some statistics recently by my friends at Vision Solutions, I really dug in because these numbers presented some interesting views on what is happening in the technology. Of course, I know that even these numbers may be open to a certain amount of interpretation, but hopefully you can read some of the same information that I have from it.

Consistency across the results

This survey was done using 985 respondents who ranged in company size. Here is the description of the survey participants:


The interesting thing is that we see information which is contrary to much of what is trending in the world of marketing IT solutions. It isn’t that all-flash, or public cloud, or any of the specific big trends aren’t real. What it does show is that there is a big representation of the “middle class” of technology consumers.

  • 51% indicated that storage growth was between 10-30% per year
  • Replication features was nearly even at 39% for hardware and 35% for software
  • Tape isn’t dead – 81% of respondents use tape as part of their data protection solution

There are more details throughout, but these jumped out at me in particular.

Test the plan, don’t plan the test

A frightening number that came from the survey was that 83% of respondents had no plan, or were less than 100% confident that their current plan was complete, tested and ready to execute. One word reaction to this: Wow!


As Mike Tyson said: “Everyone has a plan until they get punched in the mouth”. Many people that I have spoken to have a long lead up to their BCP/DR test and a significant portion of the planning is one-time activities to ensure the test goes well, but just to satisfy the test, not necessarily to build a self-healing infrastructure which is where we should be working towards.

To date, one of my most popular blog post series is my BCP/DR Primer:

BCP/DR Primer – An introduction to planning your IT recovery strategy Part 12345

This is a clear sign in my mind that people are continually looking for resources and tools to build, or re-evaluate their BCP plan. Since I’m a Double-Take user, the combination of all this hits pretty close to home. I’ve been using products with online testing capability for a number of years which helps to increase the confidence for me and my team that we are protected in the event of a significant business disruption at the primary data center.

Enterprises love their pets

keep-calm-and-love-petsWith years of work in Financial Services and enterprise business environments, I get to see the other side of the “pets versus cattle” debate which is the abundance of pets in a corporate data center. Sometimes I even think the cattle have names too.

Legacy application environments are a reality. Not every existing application has been or will be developed as a fully distributed n-tier application. There are a significant number of current and future applications that are still deployed in the traditional model with co-located servers, single instances, and other architectural challenges that don’t allow for the “cloudy” style of failure.

There is nothing wrong with this environment for the people who are mapped to this model today. Most organizations are actively redesigning applications and rethinking their development practices, but the existence of legacy products and servers is a reality for some time to come.

Evolving application and data protection

I’m a fan of Double-Take, so I’m a little biased when I see great content from Vision Solutions 🙂 What I take away from this is that there are a lot of us who may not have the ideal plan in place, or may not have an effective plan in place at all for a BCP situation. The content of seeing people’s preparation is only half of the story.


Having a plan is one thing, but seeing what the results of real data loss and the reason behind it is particularly important. Using manual processes is definitely a fast track to issues.

Beyond orchestration, the next step I recommend is using CDP (Continuous Data Protection) where possible. My protected content (servers, volumes and folders) are asynchronously replicated, plus I take daily snapshots for full servers and semi-hourly snapshots of file data. This ensures that multiple RPOs (Recovery Point Objectives).

In the event of a data corruption, the corruption would be immediately replicated…by design of the protection tool. Using a previous RPO snapshot prevents the risk of a total data loss by using an automated snapshot. Phew!

Ultimately, the onus is on us to enhance the plan, build the process, and evaluate the tools. If you want to find out more on how I’ve done data and server protection, please feel free to reach out to me (eric at discoposse dot com) and if you want to find out more on the Vision Solutions portfolio and Double-Take, you can go right to the source at and there are some great whitepapers and resources there to help out.