This presentation includes an overview of SQL Server 2012 & a deep dive into SQL Server 2014 columnstore indexes, including a review of the architecture, as well as the challenges, workarounds, travails, & big wins at customer sites
Data warehouse designers often ignore the specific needs of an OLAP database. In this session, John will outline the best ways to optimise your relational database to support your multidimensional OLAP cubes
Learn how to use the Power BI tools to help you analyse, specify and test the concepts and requirements of your next data Warehouse project.
Your relational Data Warehouse is suffering from performance problems. In this session we will take a look at some of the most common performance problems that tend to occur in a relational data warehouse.
This presentation discusses the process undertaken to achieve the certification of a 90TB SanDisk solution (HP DL580 Gen8 and Fusion ioMemory) whilst achieving an efficient out-of-box balance between SQL Server data processing capability and hardware
As data warehouse developers we deliver data to business users... but we don't deliver data as fast as we want. There are many things you can do to increase your own productivity and become a more efficient data warehouse developer.
This session goes beyond the classical star schema modeling, exploring new techniques to model data with Power Pivot and SSAS Tabular. You will see how brute-force power in DAX allows different data models than those used in SSAS Multidimensional
You have tuned your SQL server to it's limits, you have maxed the RAM on your server, and still your data warehouse is too slow. Is a Parallel Data Warehouse the next step?
Big Data has triggered the need for organisations to build real-time Operational Data Store (ODS) layers to feed the ‘Data Lake’ on Hadoop. Attend this lunchtime session to learn why a real-time ODS should be at the heart of your data architecture
T4 templating will be a first class citizen in SSDT for SQL Server 2014. This session will show why you should use this technology for SQL code generation and how you can automate the process. The session will be demo rich.
A deep dive in the internals of the database architecture, discovering how Vertipaq stores information, in order to gain better insights into the engine and understand the best way to model your data warehouse to leverage the features of VertiPaq.
“Just use partitioning” is the answer you hear, when you need to manage very large data sets in your Data Warehouse. But how do you design and implement it? We will walk through different ways to design partitioning, including layered partitioning.
This session will discuss what a modern strategy for data warehousing can be in this era, considering how the use of technologies like PowerPivot or Analysis Services Tabular affect the way you should model your data.
In this session, we are going to explain and test different DW features in SQL Server 2012, including star join optimization through bitmap filters, table partitioning, window functions, columnstore indices and more.
: We are looking at the dawn of a new workload: BigData. In this session I will talk about what BigData is, and which BigData technologies Microsoft are working on.
Users love flexible analytics but hate to wait for the data to be loaded into a traditional data warehouse. John will describe how to build an infrastructure to support real-time loading of your OLAP cubes so your user's get exactly what they want
SQL Server FastTrack References Architecture gives you the best practices and reference architectures from different hardware vendors to help you create a balanced environment for your data warehouse workload.
This session looks at some of the different methods available to load slowly changing dimension data into a data warehouse, and compares the relative performance given different data scenarios and traditional storage compared with FusionIO
This talk will describe how the new ColumnStore index technology in SQL Server 2012 makes queries go faster. Covering details of the storage and execution model, how this model interacts with modern CPUs to deliver significant performance benefits.
In this first of two sessions, we review the architecture of SQL Server and its BI components and deployment options for optimal performance. We'll also discuss how to optimize data warehouse load operations.
Snapshots without snapshots...is that possible? Take a "Classic" snapshot fact table, add some temporal data theory and you'll get a new fact table than can store snapshot data without doing snapshots. A life saver when you have a lot of data.
Do you have data warehouse queries that run too long? In this session we’ll address how columnstore indexes speed up queries, best practices for creating and using columnstore indexes, and how to diagnose and treat potential issues.
This session will drill into some of the details on the shipping HP BDA and BDW appliances and how they can work well together to deliver a comprehensive BI and DW solution for customers using the Microsoft Application Platform.
This session will provide expert advice & best practices for replicating Oracle (& other) data to SQL Server for data warehousing & BI initiatives. Also, become one of the first to see Attunity’s innovative new software in its historic unveiling!
Organizations risk being overwhelmed by data. How can you effectively provide a “single version of the truth”, while unlocking the key trends and insights that will allow your business to succeed? Come to this session to find out how.
Do you have complex dimensions in your data warehouse? Parent-child, late arriving, type 3 or type 6? In this session, we'll cover some SSIS patterns for handling each of these, along with tips for making them perform well.
Steps involved in implementing a near-real-time data warehousing solution.
Master Data Services has been given an overhaul in Denali, including a new Excel add-in and modified web front-end. Come to this session to see how MDS can be used to give greater control and process to your BI/DW dimension management.
When loading a Fast Track Data Warehouse it is important to ensure that your data is optimally laid out for Sequential I/O. Fragmentation is therefore the enemy. Know your enemy. Learn what it is, how it occurs and prevent it from happening to you!
In this session we will introduce the new modeling capabilities of Vertipaq, showing how the same scenarios can be modeled in both Multidimensional (MOLAP) and Tabular (Vertipaq), looking at how to enable your data warehouse to support both.
Many of us know what dimensions and fact tables are. But dimensional modelling is more than just dimensions and fact tables. This session is about advanced dimensional modelling topics, which are useful for anybody involved in the design of a DW.
In this session Noemi Moreno, BI Developer and Tom Austin, SQL Tools representative, will show you how to use SQL Developer tools to simplify and enhance your Business Intelligence application.
A new and fresh approach to datwarehousing that is agile, performant and easy to do. Data Vault is a new methodology that tries to overcome traditional problems that plagued traditional datawarehousing in the past.
This session provides an overview of the new Data Warehousing capabilities in SQL Server 2008 R2 Parallel Data Warehouse Edition
This session is taken from lessons learned in the field on being able to scale data models to handle multi-terabytes of data. Come to this session and understand the difficulties before you encounter them.
An open forum panel discussion with members of the SQL Customer Advisory team.
Fast Track is a new reference data warehousing architecture provided by Microsoft. More than this it represents a new way of thinking about data warehousing. A Fast Track system is measured by its raw compute power - not by a DBAs ability to tune an index. Fast Track is an appliance-like solution that delivers phenomenal performance from a pre-defined, balanced configuration of CPU, memory and storage using nothing but commodity hardware.
Of particular interest in a Fast Track system is the way in which the storage and SQL Server are configured. To achieve the fantastic throughput without using SSDs requires some careful configuration. This configuration is designed to make use of Sequential I/O to dramatically improve disk I/O performance.
Interested? If you have a large data warehouse that's seen better days or perhaps you are about to embark on a new warehousing project then you should be! Fast Track is a great solution with a fantastic value proposition.
In this one hour session we'll aim to get under the skin of Fast Track and get some answers as to how it delivers such great throughput on commodity hardware. In the process we'll aim to answer the following questions:
* When might I need Fast Track?
* What is Sequential I/O?
* How does Sequential I/O improve performance?
* What do I need to do to get Sequential I/O?
* How can I monitor for Sequential I/O ?
* What may I need to change in my ETL to get the benefit of sequential I/O?
Still reading? I'll save you a front row seat....
A typical day of DBA and new features of SQL Server 2008 can help - save a minute.
A step-by-step demo on BIDS of how to build an SSAS cube (DB) from an operational system (normalized database) such as Operational Data Store (ODS) or directly from the transactional business system, without building a star/snowflake schema Data Warehouse/Mart first.
Explore the Fast Track Data Warehouse appliances from HP & Microsoft with case studies, implementation thoughts, and roadmap.
Data warehousing features in SQL 2008
In this session we will take a deeper look at how SQL Server uses I/O and how you can design the I/O system to meet the requirements of your applications.
In this presentation, I will introduce the Madison architecture and provide a roadmap with major milestones for this product
Designing dimensional and fact tables using a case study to understand data modelling
SQL Server 2008 has a ton of great features to help you in your data warehousing endeavours. In this session find out what they are and how you can best use them.