BLOG - Performance Troubleshooting



The Real Cost of Reactive Upgrades

IT Budget Savings   •   May 7, 2019

“Reduce cost, increase functionality and improve business operations” is the battle cry of many businesses. IT is continuously battling a world of infrastructure components and applications while also dealing with a myriad of old and new technologies. Adding complication, everyone's an expert and has not only an opinion, but “the” solution. Obsolete hardware just adds another hazard to the mix.

Continue Reading →

What's the deal with cDOT?

Industry News & Trends   •   April 3, 2019

If you run NetApp equipment in your data center, you’re almost certainly familiar with 7-Mode (also called ONTAP Traditional or 7G), the basic operating system for NetApp enterprise storage solutions for more than 20 years. NetApp’s newer and more complex operating system, Clustered Data ONTAP (cDOT), has gone through several versions since its’ rollout 10 years ago in an attempt to achieve more feature parity between 7-Mode and cDOT. As it stands now, the features that truly separate cDOT from 7-Mode are mostly for scale-out solutions as opposed to scale up. Despite offering minimal downtime during hardware upgrades or refreshes, the market has not quickly taken to cDOT. The main reason? It can be quite pricey to convert, which is why literally thousands of companies are still using 7-Mode.

Continue Reading →

Mistakes That Could Be Hurting Your VNX Performance: Part 2

How-To, Guides, & Tips    •   March 2, 2017

vnx performance mistakesWelcome to part 2 of our series about some mistakes that could be hurting your VNX’s performance. If you missed the first installment of the series, you can read up on it here: Part 1. The aim of this series is to help you understand some of the dos and don’ts when it comes to VNX systems, disk array enclosures (DAE), and drives. We asked our engineering experts to give us some common mistakes people make with VNX and how to avoid them. So, what mistake could be hurting your VNX’s performance?

Continue Reading →

Mistakes That Could Be Hurting Your VNX: Part 1

EMC   •   January 19, 2017

mistakes that could be hurting your vnx What are the biggest mistakes you can make with a VNX? There are a ton of things that range from not checking performance benchmarks to leaving it sitting in your data center collecting dust. No matter your role in a storage environment, making sure all systems work at peak performance should be a top

Continue Reading →


Storage System Failure

Your data is your business’ backbone. Without the structure of your finances, customer database, inventory and administrative information, your business is like a jellyfish floundering in the wide-open sea. Avoiding this catastrophe is a top priority. But what if you have committed to too little, too late in terms of backup, and your storage system does a belly flop into the deep end? After all, you never know what the future holds, whether that be a natural disaster, hackers, or social anarchy. Prepare for anything, even the worst-case scenario, with this recovery advice.

Storage System Flops

If you are experiencing the sudden death of your storage system, there are a couple of things to take into consideration...

Continue Reading →

The More You Know: Measure Storage Performance with IOPS, Latency, Throughput

How-To, Guides, & Tips    •   September 20, 2016

Storage Performance Metrics: IOPS, Throughput, Latency

Applications are the core of your data center. Whether simple or complex, clients and employees expect quick response times from every application they use. This is why optimizing storage performance remains a top priority for CIOs. Yet, with the advent of flash, the storage industry tends to mislead buyers when it comes to measuring performance.

Continue Reading →

Thin Provisioning vs Thick Provisioning: Which One is Best For You?

How-To, Guides, & Tips    •   August 25, 2016

data center picture

No matter your role in a data center, enhancing storage performance is usually a top priority. This is especially true for enterprise applications that are accessed the most. One common practice when it comes to addressing performance and storage needs is provisioning. According to TechTarget, storage provisioning is the process of assigning storage to optimize performance. When it comes to provisioning, thin or thick is usually the question at hand. There are benefits and downsides to both. Today we will go over the pros and cons as well as some best use cases for thin provisioning and thick provisioning.

Continue Reading →

The Best Metric For Measuring Storage Performance

Performance Troubleshooting   •   January 14, 2016

Reliant Technology - Storage Performance MetricsWhen measuring the performance of flash storage, the most commonly used performance benchmarks are throughput and IOPS. But latency may be an even more important metric to consider.

Reducing latency is a major concern for users, and manufacturers have responded by moving flash storage into servers, onto faster buses like PCIe, and even onto memory buses. So, what exactly is the difference between IOPS and latency in terms of performance benchmarking and why may latency be a better performance metric for your storage environment?

Continue Reading →

The Cause of Data Center Downtime: Is Human Error Always to Blame?

Performance Troubleshooting   •   October 19, 2015

Data Center Downtime CausesMurphy's Law dictates, "What can go wrong, will go wrong." This is true in all walks of life, but it is especially prescient in the world of the data center. Data centers are huge enterprises with a wide range of moving parts, any of which can effectively "break down" at a moment's notice. After you experience several instances of something that can go wrong actually going wrong, however, you must begin to ask yourself "why can these things actually go wrong in the first place?" Human error may just be the culprit that you've been looking for.

The Error of Assuming All Data Center Mistakes Are Human Error.

To get to the root of data center downtime, it's important to define what "human error" means in this context. It doesn't mean that employees are making careless mistakes, unless, of course, you made some mistakes in the hiring process. Instead,  hiring the right people for the right positions takes this particular variable out of the equation.

Rather, human error in relation to a data center is typically

Continue Reading →

Performance Metrics for SAN Environments

Performance Troubleshooting   •   March 19, 2015

Reliant Technology - Performance Metrics for your SAN EnvironmentPerformance metrics for SAN environments help to give those within your organization the powerful, real-time and actionable information that they need to make the most informed decisions possible. It's important to understand that performance analysis is not an exact science and is much more of an art form, however. The numbers themselves only tell one part of a much larger story. How you choose to interpret those numbers and the steps that you take thereafter are what will ultimately make the difference between success and failure.

The Percentage of SP Cache Dirty Pages

SP Cache Dirty Pages exist in a write cache and are pages that have already received new data from hosts, but have not yet flushed that data to disk. This percentage should always be relatively high as it increases the chance of additional writes to the same block of data that will ultimately be absorbed by the cache.

Continue Reading →
Memory usage: real: 13893632, emalloc: 12935344
Code ProfilerTimeCntEmallocRealMem