Violin Memory Blog

Hear what the experts at Violin Memory have to say about a variety of topics on all things Flash.

Larry Ellison, Harbinger of Change

by on September 24, 2014

This year you could say that #transformation is the best hashtag for Oracle OpenWorld. News of Larry Ellison’s stepping down as CEO of Oracle signals a change of season.

Oracle-relational-databaseThere is a fundamental shift in the data center today, just like in the 1970s, when “Ellison saw IBM’s relational-database research and realized he could make a business out of it before IBM knew what a good thing it had.” Ellison was a harbinger of change then – and a harbinger of change now.

At Violin Memory, we see major leaps in the structure favored by today’s enterprise data center. And just like Larry Ellison, we try to recognize fundamental shifts in data architecture and capitalize on our insight into data center trends.

database-cloudWhat was the harbinger for change this time?  Was it the conquest of the Cloud?  Or can we pin it on Software as a Service?  If we are experiencing a shift in the data center, then definitely data center consolidation driven by faster I/O point toward an evolution in database execution.

Although change is in the air, Oracle remains the grand daddy – and critical backbone – of many a serious datacenter. Software types who have been at this game for a while are keeping their glass full as they toast to Oracle this week. At the same time, those in the know are keeping their eyes open to options that might enable a positive change in their use of Oracle.

Do you need Oracle to run 20,000 IOPS in under 0.5ms?


Chat with Violin Memory to find out how we enable positive change for every business customer. Stop by our booth to hear for yourself how we transform business lines, consolidate Oracle database infrastructures and lower costs in the data center. We’re not just talking about a fundamental change in how data is stored and served – we are living it!  Are you?

Capitalize on your ability to spot fundamental changes in the data center. Get the full scoop from Violin Memory booth #307 at Oracle OpenWorld where we feature these cutting-edge presentations:

Eric Herzog, CMO and SVP of Business Development, will share how Violin Memory revolutionizes the economics of a datacenter.

Gilbert Standen, Consulting Engineer, will share how you can build Oracle on all flash arrays.

Matt Henderson, Director of Business Development Solutions, will share how to build a modern infrastructure for your monolithic database.

Nathan Fuzi, Consulting Engineer, will share how to analyze your own database performance.  Nitin will also share his recent publication, Database Cloud Storage: The Essential Guide to Oracle Automatic Storage Management, and be available to sign books after his presentation.

Beyond the transformative changes that Violin Memory will present, there are plenty of other cutting-edge players at Oracle OpenWorld. See five bands over three epic nights.

This year at Oracle OpenWorld we listen to the sounds of change. We think about Larry Ellison and recognize that this is not Ellison’s swan song. Not a chance. For those of us who admire this entrepreneur and grand daddy of relational databases, we recognize that Larry Ellison is not leaving the race, he’s just changing his heading. Larry is sailing into a rosy sunset. Whatever direction Larry takes, we wish him fair winds and following seas.


See you at the Violin Memory booth at Oracle OpenWorld.


With thanks to Tiernan Ray at Barron’s for source material:

With thanks to Oracle for image of relational database:

Migration Do’s and Dont’s

by on September 15, 2014


Any IT project will succeed or fail based on the quality of planning, communication and execution: A drama-free data migration is no exception.

Moving line-of-business data sets to new platforms is complex. A botched migration could be potentially devastating to the enterprise. Bringing in all the relevant stakeholders early is critical to a “no surprises” migration. Buy-in and domain expertise on the business side will minimize stress and sleepless nights during the execution phase. It is also critical to establish business ownership of the data being migrated. Knowing who can make the decisions will facilitate a smooth transition.

A simple migration tends to expose a lot of tangentially related issues. Preparation for moving a data set can often expose data quality problems. The data might be incomplete. Organizational disciplines tend to vary over time, and data sets can reflect that.

A data set is a reflection of business practices at a point in time. What gets collected and saved is a function of someone’s determination as to what information is useful, but if a business needs to change, and what business doesn’t, then discontinuous or irrelevant data sets may be part of the migration discussion. Is it time to archive subsets of data? In other words, the data migration may not be from like-to-like. The migration might have an archive component to it as well, again, adding to the complexity.

Data Migration is not an exercise that turns out well when the execution follows the “ready, Fire, Aim” method. Careful planning, “paper migration” is essential. Does anyone know what the “six P’s” stand for?

After planning and executing on paper, testing the migration plan with test data sets is prudent. Risking live data on untested processes is a formula for disaster. Don’t do it.

Just as the initial migration path should be “piloted” with test data sets, the load at the other end of the migration should also be tested. More than one IT professional has a horror story about choking the system by trying to load a full production data set all at once and without testing.

So at this point in a migration project, the planning is done, communications are ongoing, ownership and business practices are established and test data sets have proven that the plumbing works.

Before going any further it is worth asking whether the varsity team is handling this mission critical assignment. Are the skills and resources of the team assigned to the project the best available? The answer needs to be yes to prevent all reasonable problems and deal with the unforeseen as the need arises. And are they using the best available tools? An inexperienced migration team may not have the experience to pick the best tools or best practices.

One example of a best practice is the decision to base the migration on the Violin Memory 2510 data migration appliance.

Identified by top money center banks and Fortune 100 enterprises around the world as a critical leverage point in their data migration procedures, the Maestro 2510 installs directly in a Fiber channel SAN Fabric and seamlessly, migrates data as a background process to a Violin Memory 6000 Series All Flash Array or to any block based storage device currently in place.

The Maestro 2510 is a heterogeneous, non-disruptive solution that simply plugs into an existing Fibre Channel SAN and delivers immediate performance improvement. It requires no configuration changes to either the servers or storage and requires no application downtime.

Since the Maestro 2510 passes through all of the write I/Os and status commands, it is completely transparent to all server and storage operations and is able to perform all of its functions without disrupting existing Disaster Recovery and Backup processes.

The Maestro 2510 gives you the power to migrate your data from legacy-based fibre channel SANs to a new storage array – completely in the background – with no disruption to applications. The Maestro 2510 appliance is able to do this through its highly optimized Memory Services Engine. This engine enables data to be transparently migrated from disk to flash while applications continue to operate at accelerated speeds. Once migrated, the legacy arrays can be retired or repurposed for less performance sensitive applications.

The Maestro 2510 fully supports any-to-any data migration. This means that the appliance can be used to seamlessly migrate data from any vendors SAN storage array to any other vendors SAN storage array.

Webinar: SQL Server in a Flash

by on September 11, 2014


Increasing the performance of SQL Server workloads remains a hot topic. Given the importance that tabular data plays not only in your organization, but also that of your trading partners, it’s not at all surprising that this remains of considerable interest. Recently, Sumeet and I took the time to articulate the value proposition for using flash storage in an SQL Server environment. If you’d like to learn a bit more, I invite you to watch this webinar:

Webinar: SQL Server in a Flash


Redefine what is possible with your SQL database decisions today and into the future with purpose built Windows all-flash arrays that will optimize your SQL Server infrastructure, provide the best of physical, cloud and hybrid cloud environments and help you gain control of your windows business applications, SQL databases and workloads driving today’s big data and business SLAs.

In this webinar, you will learn about the benefits of a seamlessly integrated approach to all-flash storage for your Microsoft SQL Server environment. We will discuss the challenges and pitfalls to avoid when selecting performance-based storage with enterprise data services. We will also discuss how to consolidate and optimize your SQL Server instances, optimizing your Hyper-V infrastructure, and planning for SQL Server growth.


Sumeet Bansal, Principal Solutions Architect
Advising and consulting on integrating Flash towards high performance high efficiency database (MS SQL Server, Oracle) platforms. Over 15 years of experience with Database Systems and over 5 years dedicated to deploying Flash for Databases. Prior to joining Violin Memory, Sumeet was Principal Solutions Architect at Fusion-io and VP of IT at (the largest online wine retailer in the US).

Clay Ryder, Sr. Product Marketing Manager for the Violin Windows Flash Array
Before assuming his current role at Violin Memory, Clay was a marketing consultant at NetApp specializing in Microsoft-based solutions and storage efficiency. Prior to his marketing roles, Clay held executive positions in multiple industry analyst firms including The Sageza Group, which he co-founded in 2001, Zona Research, and The X Business Group.

Learn how Violin and Microsoft collaboratively optimized this solution to leverage WFA’s unique performance profile so you can run your SQL Server environment in a Flash.

For more information on the WFA and SQL Server, go to

SQL Server and the WFA: Part 2 – Latency Matters

by on August 28, 2014


The Windows Flash Array (WFA) is all about high throughput, low latency, scalability, and balanced performance. We previously discussed its high throughput, and in this blog, we are going to focus on how latency really matters, especially if you want your databases to hum along providing copious value to your enterprise.


Having a high level of IOPS may seem impressive (and it is), but that’s only part of the high performance equation. There are two other factors in SQL Server environments that are the greatest concern with respect to achieving the highest performance.

Latency is a measure of the time it takes for data to come back from a request. How much you can lower this measure is indicative of how much of an increase in CPU utilization and therefore decrease in application processing time (duration of reports, etc.) you can achieve.

The other factor is throughput. Each query has a finite amount of data to process. Faster storage doesn’t change the amount of data in a report, rather only the time it takes to deliver it. So a disk array that caps out at 200MBps will deliver a 4000MB a report in 20 seconds. In contrast, the WFA, which achieves 4000MBps, can complete the task in 1 second.

Poor storage performance leads to poor CPU utilization. CPUs spend more time waiting than executing, which means servers are expensive and mostly idle investments. This reality has not escaped the attention of CIOs, CFOs, and other senior management. The performance challenge grows more pronounced the greater the number of server cores you have in a system. Of course, today multi-core is par for the course, which means you are facing the aforementioned challenge.

When an application (or in our case, a database server) is active, it will reside in one of three queues:

  1. The Running Queue is where the work gets done. It is a currently executing process, and the point at which your servers are earning their keep, and perhaps a bit more. You are on the express train, sitting in First Class.
  2. The Waiting Queue is what its name implies; the process is waiting for some resource (I/O, network, locks, latches, etc.) to become available. You ran out of things you can do, so you’re stuck waiting at the train station for the rest of your team to get it together.
  3. The Runnable Queue holds processes that have their resources ready, and are waiting their turn for some CPU time. You’re ready to go, but the train already left the station, so you’ll hop on the next one as soon as it arrives and the conductor will show you your seat.

It’s not bad when 1 user is vying for the system resources.

It gets a whole lot worse with multiple users vying for the same resources.

Typically, disk arrays have good throughput until the number of users increases at which point logically sequential requests become physically random. Keep in mind that the concept of being logically sequential at the database layer is entirely different than it actually being physically sequential on the storage medium. Defragging, index rebuilds, clustered indices, read ahead processes, large batch requests, etc. all may cause the DBA to believe that storage is being accessed sequentially; however, a SAN layout with striped LUNs, multiple components, and support of multiple workloads means that it is almost certainly not sequential on disk or the access of it is being randomized due to concurrency of requests with other users.

Our patented Flash Fabric Architecture™ overcomes this resource contention for storage. Put on your geek hats, here’s how it works:

  1. One 64K inbound write is split into 16 separate 4K packets
  2. Each 4K packet has parity calculated creating a total of 5K of data to store
  3. The 5K final packet is split into five 1K packets and one is sent to each of the five VIMMS in a RAID group

As you can see, the packet is broken into small batches and processed in parallel; this is what helps drive down latency. In our large capacity arrays there are 64 high-speed, custom-built, flash controllers processing 1K vs. commodity SSD controllers processing 64K (RAID 1 or RAID 10 will send the full packet to both SSDs that host a log or tempdb device). The result is that data is spread across the array so there is no clumping of data, data locality or ultimately, hot spot issues. This means the WFA can scale to support massive user concurrency for both random and sequential focused workloads. Connecting through SMB Direct can free up to 30% of CPU usage; these cycles could be put towards BI initiatives or other value added activities.

For example, you can hammer away on OLTP workloads while also executing highly parallel BI with many simultaneous related queries, without dragging down your overall performance. For example, let’s say your organization has a 24-64 core DW/BI solution. Thus, you have the capacity to run in a parallelization factor of 24-64. In order to justify a dedicated device for parallelization you would need to data in a factor over 64.

Running a mix of multiple database workloads against a single array is now not only possible, but you can do so with incredible performance, consistent latency, and overall higher server and storage efficiency. Of course, if consolidation or virtualization of SQL Server instances is your preferred approach, the WFA can enable you to increase virtualization density as well.

With the low latency of the WFA, you can achieve these levels of performance enhancement to your SQL Server without any storage tuning, upgrading or modification:

  • 3-10x faster reports
  • 3-6x faster transactions
  • 2-5x higher consolidation of SQL Server VM farms without performance penalty
  • Streamlined re-index process drops from 20+ milliseconds to sub-millisecond
  • Index and Stats defrag locking process can be released in a sub-millisecond
  • Grow your database size without taking a performance hit
  • 24×7 maintenance (do backups, index maintenance, etc., during the day) so if there’s an issue, it’s when you’re at work, not when you’re asleep

As you can see from this and the previous blog, we can do a lot in 3U to not only boost your SQL Server performance, but transform the economics of SQL Server as well. Finally, you can have storage that is so fast that it doesn’t require you to change your behavior; you can run tasks during the day while you’re at work, so why not buy the storage solution that allows you to do so?

Nonetheless, performance alone will not meet the needs of a 21st century data center. Transforming the efficiency and economics of the data center, which ultimately yields reduced OPEX and CAPEX, is just as essential for overall corporate success. We’ll continue on this theme the next time.


Learn how Violin and Microsoft collaboratively optimized this solution to leverage WFA’s unique performance profile so you can run your SQL Server environment in a Flash.

For more information on the WFA and SQL Server, go to

> PREVIOUS: Part 1 of SQL Server and the WFA — Lots of I/O

SQL Server and the WFA: Part 1 – Lots of I/O

by on July 31, 2014


As you probably already know, Violin Memory has announced all-flash arrays integrated with Windows Storage Server 2012 R2, dubbed the Windows Flash Array. If you’ve had a chance to give them the once over, it’s obvious that they are a far cry from a white box server with JBOD attached running Windows Storage Server. The WFA is all about high throughput, consistent low latency, scalability, and balanced performance. It’s a true enterprise solution. Of course, this does not happen by generic technology combinations; rather it requires differentiated hardware, the tightest possible software integration, and solution-focused engineering with a vision to excel, or is that accel? Well OK, it’s both.

The unique performance of WFA is compelling for SQL Server workloads. For example, a common database application such as OLTP generates lots of random I/O, which makes sense as it’s driven by copious ad hoc inquiries from all across the enterprise. However, such a workload is a mismatch with legacy storage’s linear read/write focus. This results in high latency and poor application performance. Yet these queries are often associated with high-value real-time applications such as customer support, sales, inventory control, etc. As a result, storage professionals have tried their best to coax more performance from their cylinders of spinning rust but their success has been limited at best.

The Violin Windows Flash Array (WFA) combines Violin Memory’s patented Flash Fabric Architecture™ (FFA), Microsoft’s fast SMB Direct protocol, and Microsoft Windows Storage Server 2012 R2. That sounds very nice, but what does it mean? Look inside of a WFA, and you’ll note that there are no disks of any type (HDD or SSD). As a result, there is no need for striping techniques and optimization algorithms, nor is there a need to over provision storage or add cache to controllers in an attempt to garner enough I/O capacity. The FFA delivers fast consistent access to our all-flash storage so you don’t have to continuously monitor data hot spots and tune accordingly. In other words, it’s fast without your constant worry and attention.


How is fast is fast? Pretty darn fast. We did some SQLIO Load testing on a WFA-64 equipped with 56GBb FDR InfiniBand connectivity at a customer site; here’s the performance we observed:

The WFA delivered over 1.2 million IOPS on SQL reads. That’s up to 50% higher performance compared with an industry-standard all-flash array connected through Fibre Channel.

The WFA delivered 1.2 million IOPS on SQL writes, which was proportionally even more impressive. That’s up to twice the throughput compared with an industry standard all-flash array with Fibre Channel.

From these numbers you can see that the WFA is up to the task of delivering the sustained performance for short block random read/write workloads. And, there is no ongoing tuning required. But did you know that this performance was achieved over a file-based solution, not block?

The WFA delivers DAS-like performance due to its support for SMB Direct, also known as SMB over RDMA. You can have simplicity of a file-based environment but without sacrificing the performance you’re accustomed to with block-based solutions. The advantages of file over block can fundamentally change your ease of use and efficiency. So much so, this is worthy of a discussion of its own; we’ll dive in depth on this in a future blog.

From an I/O perspective, we can do a lot in 3U to boost your SQL Server performance. But just attaining a high IOPS number does not guarantee balanced performance for a mix of multiple workloads. Achieving consistent low latency is essential, as this will transform how your SQL database and related apps perform, and your expectations! We’ll continue on this theme the next time.


Learn how Violin and Microsoft collaboratively optimized this solution to leverage WFA’s unique performance profile so you can run your SQL Server environment in a Flash.

For more information on the WFA and SQL Server, go to

> NEXT: Part 2 of SQL Server and the WFA — Latency Matters

More Posts ...

Featured Posts