Without Desktop Virtualisation, you can’t attain True Business Continuity

Even if you’ve invested on virtualisation, off-site backup, redundancy, data replication, and other related technologies, I?m willing to bet your BC/DR program still lacks an important ingredient. I bet you’ve forgotten about your end users and their desktops.

Picture this. A major disaster strikes your city and brings your entire main site down. No problem. You’ve got all your data backed up on another site. You just need to connect to it and voila! you’ll be back up and running in no time.

Really?

Do you have PCs ready for your employees to use? Do those machines already have the necessary applications for working on your data? If you still have to install them, then that’s going to take a lot of precious time. When your users get a hold of those machines, will they be facing exactly the same interface that they’ve been used to?

If not, more time will be wasted as they try to familiarise themselves. By the time you’re able to declare ?business as usual?, you’ll have lost customer confidence (or even customers themselves), missed business opportunities, and dropped potential earnings.

That’s not going to happen with desktop virtualisation.

The beauty of?virtualisation

Virtualisation in general is a vital component in modern Business Continuity/Disaster Recovery strategies. For instance, by creating multiple copies of virtualised disks and implementing disk redundancy, your operations can continue even if a disk breaks down. Better yet, if you put copies on separate physical servers, then you can likewise continue even if a physical server breaks down.

You can take an even greater step by placing copies of those disks on an entirely separate geographical location so that if a disaster brings your entire main site down, you can still gain access to your data from the other site.

Because you’re essentially just dealing with files and not physical hardware, virtualisation makes the implementation of redundancy less costly, less tedious, greener, and more effective.

But virtualisation, when used for BC/DR, is mostly focused on the server side. As we’ve pointed out earlier in the article, server side BC/DR efforts are not enough. A significant share of business operations are also dependent on the client side.

Desktop virtualisation (DV) is very similar to server virtualisation. It comes with nearly the same kind of benefits too. That means, a virtualised desktop can be copied just like ordinary files. If you have a copy of a desktop, then you can easily use that if the active copy is destroyed.

In fact, if the PC on which the desktop is running becomes incapacitated, you can simply move to another machine, stream or install a copy of the virtualised desktop there, and get back into the action right away. If all your PCs are incapacitated after a disaster, rapid provisioning of your desktops will keep customers and stakeholders from waiting.

In addition to that, DV will enable your user interface to look like the one you had on your previous PC. This particular feature is actually very important to end users. You see, users normally have their own way of organising things on their desktops. The moment you put them in front of a desktop not their own, even if it has the same OS and the same set of applications, they?ll feel disoriented and won’t be able to perform optimally.

Contact Us

  • (+353)(0)1-443-3807 – IRL
  • (+44)(0)20-7193-9751 – UK

Check our similar posts

Will UK Retailers Skim the Cream with ESOS?

The British Retail Consortium (BRC) was quick out on the starting blocks with an ambitious plan to cut energy costs by 25% in 5 years. Their ?25-in-5? initiative is chasing a target of ?4.4 billion savings during the duration. Part of this program involves ?cutting a path through a complex and inaccessible policy landscape?. BRC believes this drawback is making its members think twice about making energy efficiency investments.

The UK?s sprawling network of grocers, department stores and malls is the nation?s second most hungry energy customer, having spent ?3.3 billion on it in 2013 when it accounted for almost 20% of carbon released. If you think that sounds bad, it purchased double that amount in 2005. However the consortium believes there is still more to come.

It bases this assumption on the push effect of UK energy rates increasing by a quarter during the duration of the project. ?So it makes sense to be investing in energy efficiency rather than paying bills,? Andrew Bolitho (property, energy, and transport policy adviser) told Business Green. The numbers mentioned exclude third party transport and distribution networks not under the British Retail Consortium umbrella.

The ?complex and inaccessible policy landscape? is the reflection of UK legislators not tidying up as they go along. BRC cites a ?vast number of policies ? spreading confusion, undermining investment and making it harder to raise capital?. The prime culprits are Britain?s CRC Energy Efficient Scheme (previously Carbon Reduction Commitment) which publishes league tables and ESOS. Andrew Bolitho believes this duality is driving confused investors away.

The British Retail Consortium is at pains to point out that this is not about watering things down, but making it simpler for participating companies to report on energy matters at a single point. It will soon go live with its own information hub providing information for retailers wishing to measure consumption at critical points, assemble the bigger picture and implement best practice.

Ecovaro agrees with Andrew Bolitho that lowering energy demand and cutting carbon is not just about technology. We can do much in terms of changing attitudes and providing refresher training and this does not have to cost that much. Studies have shown repeatedly that there is huge benefit in inviting employees to cross over to our side. In fact, they may already be on board to an extent that may surprise.

What Sub-Metering did for Nissan in Tennessee

When Nissan built its motor manufacturing plant in Smyrna 30 years ago, the 5.9 million square-foot factory employing over 8,000 people was state of art. After the 2005 hurricane season sky-rocketed energy prices, the energy team looked beyond efficient lighting at the more important aspect of utility usage in the plant itself. Let’s examine how they went about sub-metering and what it gained for them.

The Nissan energy team faced three challenges as they began their study. They had a rudimentary high-level data collection system (NEMAC) that was so primitive they had to transfer the data to spread-sheets to analyse it. To compound this, the engineering staff were focused on the priority of getting cars faster through the line. Finally, they faced the daunting task of making modifications to reticulation systems without affecting manufacturing throughput. But where to start?

The energy team chose the route of collaboration with assembly and maintenance people as they began the initial phase of tracking down existing meters and detecting gaps. They installed most additional equipment during normal service outages. Exceptions were treated as minor jobs to be done when convenient. Their next step was to connect the additional meters to their ageing NEMAC, and learn how to use it properly for the first time.

Although this was a cranky solution, it had the advantage of not calling for additional funding which would have caused delays. However operations personnel were concerned that energy-saving shutdowns between shifts and over weekends could cause false starts. ?We’ve already squeezed the lemon dry,? they seemed to say. ?What makes you think there?s more to come??

The energy team had a lucky break when they stumbled into an opportunity to prove their point early into implementation. They spotted a four-hourly power consumption spike they knew was worth examining. They traced this to an air dryer that was set to cyclical operation because it lacked a dew-point sensor. The company recovered the $1,500 this cost to fix, in an amazing 6 weeks.

Suitably encouraged and now supported by the operating and maintenance departments, the Smyrna energy team expanded their project to empower operating staff to adjust production schedules to optimise energy use, and maintenance staff to detect machines that were running without output value. The ongoing savings are significant and levels of shop floor staff motivation are higher.

Let’s leave the final word to the energy team facilitator who says, ?The only disadvantage of sub-metering is that now we can’t imagine doing without it.?

Contact Us

  • (+353)(0)1-443-3807 – IRL
  • (+44)(0)20-7193-9751 – UK
8 Reasons why you Need to Undertake Technical and Application Assessments

Are your information assets enabling you to operate more cost-effectively or are they just drawing in more risks than you are actually aware of? Obviously, you now need to get a better picture of those assets to see if your IT investments are giving you the benefits you were expecting and to help you identify areas where improvements should be made.

The best way to get the answers to those questions is through technical and application assessments. In this post, we?ll identify 8 good reasons why it is now imperative to undertake such assessments.

1. Address known issues – Perhaps the most common reason that drives companies to undertake a technology/application assessment is to identify the causes of existing issues such as those related to data accessibility, hardware and software scalability, and performance.

2. Cut down liabilities and risks – Unless you know what and where the risks are, there is no way you can implement an appropriate risk mitigation strategy. A technology and application assessment will enable you to thoroughly test and examine your information systems to see where your business-critical areas and points of failure are and subsequently allow you to act on them.

3. Discover emerging risks – Some risks may not yet be as threatening as others. But it would certainly be reassuring to be aware if any exist. That way, you can either nip them in the bud or keep them monitored.

4. Comply with regulations – Regulations like SOX require you to establish adequate internal controls to achieve compliance. Other regulations call for the protection of personally identifiable information. Assessments will help you pinpoint processes that lack controls, identify data that need protection, and areas that don’t meet regulatory requirements. This will enable you to act accordingly and keep your company away from tedious, time-consuming and costly sanctions.

5. Enhance performance – Poor performance is not always caused by an ageing hardware or an overloaded infrastructure. Sometimes, the culprits are: unsuitable configuration settings, inappropriate security policies, or misplaced business logic. A well-executed assessment can provide enough information that would lead to a more cost-effective action plan and help you avoid an expensive but useless purchase.

6. Improve interoperability – Disparate technologies working completely separate from each other may be preventing you from realising the maximum potential of your entire IT ecosystem. If you can examine your IT systems, you may be able to discover ways to make them interoperate and in turn harness untapped capabilities of already existing assets.

7. Ensure alignment of IT with business goals – An important factor in achieving IT governance is the proper alignment of IT with business goals. IT processes need to be assessed regularly to ensure that this alignment continues to exist. If it does not, then necessary adjustments can be made.

8. Provide assurance to customers and investors – Escalating cases of data breaches and identity theft are making customers and investors more conscious with a company?s capability of preserving the confidentiality of sensitive information. By conducting regular assessments, you can show your customers and investors concrete steps for keeping sensitive information confidential.

Ready to work with Denizon?