Data Driver

Blog archive

More Details on SQL Server 2014 In-Memory Capabilities

More details are emerging about in-memory capabilities in the new SQL Server 2014, announced at the recent TechEd 2013 conference.

The first Community Technology Preview is expected to be released soon, possibly this month, and you can register with Microsoft to be notified of its availability.

Highlights of the new release are data warehousing and business intelligence (BI) enhancements made possible through new in-memory capabilities built in to the core Relational Database Management System (RDBMS). As memory prices have fallen dramatically, 64-bit architectures have become more common and usage of multicore servers has increased, Microsoft has sought to tailor SQL Server to take advantage of these trends.

The in-memory Online Transaction Processing (OLTP) capability--formerly known by the codename Hekaton--lets developers boost performance and reduce processing time by declaring tables as "memory optimized," according to a whitepaper (PDF download) titled "SQL Server In-Memory OLTP Internals Overview for CTP1."

"Memory-optimized tables are stored completely differently than disk-based tables and these new data structures allow the data to be accessed and processed much more efficiently," Kalen Delaney wrote in the whitepaper. "It is not unreasonable to think that most, if not all, OLTP databases or the entire performance-sensitive working dataset could reside entirely in memory," she said. "Many of the largest financial, online retail and airline reservation systems fall between 500GB to 5TB with working sets that are significantly smaller."

"It’s entirely possible that within a few years you’ll be able to build distributed DRAM-based systems with capacities of 1-10 Petabytes at a cost less than $5/GB," Delaney continued. "It is also only a question of time before non-volatile RAM becomes viable."

Another new in-memory benefit is "new buffer pool extension support to non-volatile memory such as solid state drives (SSDs)," according to a SQL Server Blog post. This will "increase performance by extending SQL Server in-memory buffer pool to SSDs for faster paging."

Independent database expert Brent Ozar expounded on this subject, writing "SQL Server 2014 will automatically cache data [on SSDs] with zero risk of data loss."

"The best use case is for read-heavy OLTP workloads," Ozar continued. "This works with local SSDs in clusters, too--each node can have its own local SSDs (just like you would with TempDB) and preserve the SAN throughput for the data and log files. SSDs are cheap, and they’re only getting cheaper and faster."

Other in-memory features mentioned by the Microsoft SQL Server team include "enhanced in-memory ColumnStore for data warehousing," which supports real-time analytics and "new enhanced query processing" that speeds up database queries "regardless of workload."

Some readers expressed enthusiasm for the new features, but, of course, wanted more. "Ok the in-memory stuff (specifically OLTP and SSD support) is valuable but the rest is so so," read one comment from a reader named John on the Microsoft blog post. "Really I wish that we would see continued improvements in reporting and analysis services and in general less dependence on SharePoint which is a painful platform to manage. QlikView and Tableau are a real threat here."

Besides the in-memory capabilities, Microsoft also emphasized increased support for hybrid solutions where, for example, a company might have part of its system on-premises because of complex hardware configurations that don't lend themselves to hosting in the cloud. These companies can then use the cloud--Windows Azure--for backup, disaster recovery and many more applications. You can read more about that in this whitepaper (also a PDF download).

What do you think of the new in-memory capabilities of SQL Server 2014? Comment here or drop me a line.

Posted by David Ramel on 06/13/2013


comments powered by Disqus

Featured

  • Copilot Agent Mode Preview Highlights New Visual Studio 17.14 Release

    Agent mode, now in public preview for Visual Studio 17.14, marks a major step forward for AI-assisted development. Unlike previous Copilot features, agent mode can autonomously plan, edit, iterate, and invoke trusted tools-completing complex coding tasks from a single natural language prompt.

  • Microsoft Busts the 'Myth of AI/ML and Java'

    Microsoft, contradicting beliefs of Java developers responding to a survey, said they don't need to learn AI, master machine learning, or switch to Python to build intelligent, production-ready applications.

  • Predicting the Future Using Azure Machine Learning

    Eric D. Boyd of responsiveX previews his VSLive! 2025 session at Microsoft HQ in August where he explains how Azure ML empowers teams to build, deploy, and manage machine learning models with ease and confidence.

  • VS Code 1.10 Showcases New, Detailed Markdown Copilot Prompting

    The new way to get the most out of GitHub Copilot is from markdown prompting, or writing detailed, reusable natural-language instructions in files like README.md or copilot-instructions.md to guide different AI models in generating context-aware, accurate code.

  • Uno Platform Studio and 'Hot Design' Reach General Availability

    Uno Platform, a .NET-centric open source project for building single-codebase apps across multiple platforms, this week announced v6.0 of its flagship offering, which introduces a zero-install, web-based IDE for rapid cross-platform development, alongside a modernized app architecture that embraces MVU and .NET-style extensions.

Subscribe on YouTube