Tag Archive: SolarWinds


This is a blog post that I’ve had at the back of my mind for a good 6 months or so. The pieces of the puzzle have come together after the Gestalt IT Tech Field Day event in Boston. After spending the best part of a week with some very very clever virtualisation pro’s I think I’ve managed to marshal the ideas that have been trying to make the cerebral cortex to wordpress migration for some time !

Managing an environment , be it physical or virtual for capacity & performance requires tools that can provide you with a view along the timeline. Often the key difference between dedicated “capacity management” offerings and performance management tools is the very scale of that timeline.

clip_image002

Short Term : Performance & Availability

There we are looking at timings within a few seconds / minutes ( or less ) this is where a toolset is going to be focused for current performance on any particular metric , be it the response time to load a web application , Utilisation of a processor core or command operations rate on a disk array. The tools that are best placed to give us that information need to be capable of processing a large volume of data very quickly due to the requirement to pull in a given metric on a very frequent interval. The more frequently you can sample the data , the better quality output the tool can give. This can present a problem in large scale deployments due to a requirement that many tools have to write this data out to a table in a database – this potentially tethers the performance of a monitoring tool to the underlying storage available for that tools , which of course can be increased but sometimes at quite a significant cost. As a result you many want to scope the use of such tools only to the workloads that require that short term , high resolution monitoring. In a production environment with a known baseline workload , tools that use a dynamic threshold / profile for alerting on a metric can be very useful here ( for example Xangati or vCenter Operations ) If you don’t have a workload that can be suitably base lined ( and note that the baseline can vary on your business cycle , so may well take 12 months to establish ! ) then the dynamic thresholds are not of as much use.

Availability tools have less of a reliance on a high performance data layer as they are essentially storing a single bit of data on a given metric. This means the toolset can scale pretty well. The key part of availability monitoring is the visualisation and reporting layer. There is no point only displaying that data to a beautiful and elegant dashboard if no-one is there to see that dashboard ( and according to the Zen theory of network operations , would it change if there was no one there to watch it ! ) The data needs to be fed into a system that best allow an action to be made – even if it’s an SMS / Page to someone who is asleep. In this kind of case , having suitable thresholds are important – you don’t want to be setting fire alarms off for a blip in a system that does not affect the end service. Know the dependencies on the service and try to ensure that the root cause alert is the first one sent out. You do need to know that the router that affects 10,000 websites is out long before you have alerts for those individual websites.

Medium Term : Trending & Optimisation

Where the timeline goes beyond “what’s wrong now” , you can start to look at what’s going to go wrong soon. This is edge of the crystal ball stuff , where predictions are looking to be made in the order of days / weeks. Based on collected utilisation data in a given period , we can assess if we have sufficient capacity to be able to provide an acceptable service level in the near future. At this stage , adjustments can be made to the infrastructure in the form of resource balancing ( by storage or traditional load ) – tweaks can also be made to virtual machine configuration to “rightsize” an environment. By using these techniques it is possible to reclaim over allocated space and delay potential hardware expansions. This is especially valid where there may be a long lead time on a hardware order. The types of recommendations generated by the capacity optimisation components of VKernel , NetApp ( Akorri ) and Solarwinds products are great examples of rightsizing calculations.  As the environment scales up , not only are we looking for optimisations , but potential automated remediation ( within the bounds of a change controlled environment ) would save time and therefore money.

Long Term capacity analysis : When do we need to migrate Data centers ?

Trying to predict what is going to happen to an IT infrastructure in the long term is a little like trying to predict the weather in 5 years time , you know roughly what might happen but you don’t really know when. Taking a tangent away from the technology side of things , this is where the IT strategy comes in – knowing what applications are likely to come into the pipeline. Without this knowledge you can only guess how much capacity you will need in the long term. The process can be bidirectional though , with the information from a capacity management function being fed back into the wider picture for architectural strategy for example should a lack of physical space be discovered , this may combine with a strategy to refresh existing servers with blades. Larger Enterprises will often deploy dedicated capacity management software to do this ( for example Metron’s Athene product which will model capacity for not only the virtual but the physical environment )  Long term trending is a key part of a capacity management strategy but this will need to be blended with a solution to allow environmental modeling and what if scenarios. Within the virtual environment the scheduled modeling feature of VKernel’s vOperations Suite is possibly the best example of this that I’ve come across so far – all that is missing is an API to link to any particular enterprise architecture applications. When planning for growth not only must the growth of the application set be considered but the expansion in the management framework around it , including but not limited to backup and the short-medium term monitoring solutions.  Unless you are consuming your it infrastructure as a service , you will not be able to get away with a suite that only looks at the Virtual Piece of the puzzle – Power / Cooling & Available space need to be considered – look far enough into the future and you may want to look at some new premises !

We’re going to need a bigger house to fit the one pane of glass into…

“one pane of glass” – is a phrase I hear very often but not something I’ve really seen so far. Given the many facets of a management solution I have touched on above , that single pane of glass is going to need to display a lot ! So many metrics and visualisations to put together , you’d have a very cluttered single pane. Consolidating data from many systems into a mash-up portal is about the best that can occur , but yet there isn’t a single framework to date that can really tick all the boxes. Given the lack of a “savior” product you may feel disheartened , but have faith!. As the ecosystem begins to realise that no single vendor can give you everything and that an integrated management platform that can not only display consolidated data , but act as a databus to facilitate sharing between those discrete facets is very high on the enterprise wishlist , we may see something yet.

I’d like to leave you with some of the inspiration for this post – as seen on a recent “Demotivational Poster” –a quick reminder of perfection being in the eye of the beholder.

“No matter how good she looks, some other guy is sick and tired of putting up with her s***”

I’ve been lucky enough to be selected again to attend one of Gestalt IT’s Tech Field Day events. These place a selection of IT community members with a selection of Vendors for a series of sessions that go beyond the usual sales pitch you might get a user group event. The are also a lot more interactive , with a roundtable discussion before , after & sometimes during a session. The events are recorded and streamed live , you can also keep with with what the kids at the back of the class are whispering to each other by following the #TechFieldDay hashtag on twitter.

 

This Event is to be held in Boston in just over 2 weeks time and has a particular focus on Virtualisation technology. Other events have been based around Networking & Wireless technology, or just general datacenter technologies. The delegates have been selected for their work within the Virtualisation community , featuring more than its fair share of VMware vExperts and of course the whole vSoup Podcast crew! We are aiming to be able to record & publish an episode of the show live from the event.

 

The Presenters

Solarwinds :

I have seen Solarwinds present before and I’m looking forward to their deep dive style – as veteran TFD Sponsors they know that talking geeky is going to get a good response from us. I would imagine there will be some good detail on the product that is the fruit of the Hyper9 acquisition.

Vkernel:

I’ve enjoyed a good relationship with Vkernel over the last couple of years , both as an end user and as a blogger. Its not their first appearance at a Tech Field Day event so I’m sure that we’ll see something new around their infrastructure optimisation product set.

VMware:

I’ve heard good things about this little start-up , they have something called a Hypervisor , which could go far Smile Is what I’d have said man years ago , but like an ageing relative I’m going to have to say “look how they’ve grown!” I shall be looking forward to meeting up with the Wookie of Virtualisation , John Troyer and seeing what VMware have to show us beyond the press release!

Symantec:

Tech Field Day usually attracts a mix of sponsors , from the very fresh start-up ( in fact there will be a start-up coming out of “stealth mode” at the event ) to the established company. Symantec will sit firmly in the latter of those two and In my opinion have a harder task at these events because they have a PR/Marketing/Community machine that is more used to higher level , PowerPoint rich communication ; which is something that Tech Field Day just isn’t about. I’d love to see a “big” sponsor present with the passion and in depth knowledge of a start-up.

Embotics:

I was lucky enough to meet up with a few of the Embotics guys in the last year and while I like their policy based Virtualisation management product its been something that’s been quite a hard sell back to management. I’ve heard they might have something in the pipeline that will really emphasise its value. Watch this space for more details….

 

There is one extra vendor to be announced in addition to the “stealth mode” start-up launching itself , which I’m particularly looking forward to.  I think its going to be the perfect mixture of catching up with friends within the community , meeting some new ones and submersing myself in some seriously good technology. For more details, check out www.techfieldday.com

 

image

SolarWinds have just launched their latest product in their Admin focussed line of tools and while the underlying focus and data is bread and butter to the network admin , the usage seems focussed towards  a more security admin / HR role.

So , why the departure from the more conventional monitoring and management toolset? The first point Josh Stevens , SolarWinds “Head Geek” made was that the Community asked for it. One of the things that impressed me about SolarWinds when I saw them as a presenter at Tech Field Day was their attention to the IT community – through programs like thwack.com they keep a finger close to the pulse of what admins and other IT Pro’s are actually using their toolsets for and more importantly , what they would like to use it for. I’m a big fan of end user driven development – give the community some option / say in how you evolve after all without users/customers where will you be ? The best example of this would be the user driven feature set of dropbox – see www.dropbox.com/votebox for more details ! Much of the work of a system admin is driven by compliance requirements , be it within a security or regulatory context. The Device Tracker fills a niche currently only occupied by CiscoWorks in terms of tracking connected devices , but adds historical data tracking.

UDT Constantly scans your network devices for connected MAC and IP addresses, keeping a record of what is connected to what – this allows the tool to do a few neat things such as.

– Tracking down problem users / machines : you know the IP of the machine that violated your AUP ,  but not sure where in the building they were at the time ( handy if you hotdesk ) , if you know where your switch ports go to , then this tool could help you match up who was sat at what desk.

– Capacity Planning : So you’ve cabled all of your switch ports to your structured cabling , but how many are you really using over a given period ? Identify unused ports and spare capacity at the access layer.

 

The product seems pretty well integrated with the Orion NPM toolset – using the same interface and familiar look & feel so Admins will be able to pick it up and play right away. The only missing piece of the puzzle is a closer look at clients connected via a wireless network but I’m assured that will be in future version.

Image[3]

SolarWinds are offering some hefty pre-release discounts on User Device Tracker – which starts at $1700. For more details , head over to the SolarWinds Website