This session goes beyond the classical star schema modeling, exploring new techniques to model data with Power Pivot and SSAS Tabular. You will see how brute-force power in DAX allows different data models than those used in SSAS Multidimensional
You have tuned your SQL server to it's limits, you have maxed the RAM on your server, and still your data warehouse is too slow. Is a Parallel Data Warehouse the next step?
Big Data has triggered the need for organisations to build real-time Operational Data Store (ODS) layers to feed the ‘Data Lake’ on Hadoop. Attend this lunchtime session to learn why a real-time ODS should be at the heart of your data architecture
T4 templating will be a first class citizen in SSDT for SQL Server 2014. This session will show why you should use this technology for SQL code generation and how you can automate the process. The session will be demo rich.
A deep dive in the internals of the database architecture, discovering how Vertipaq stores information, in order to gain better insights into the engine and understand the best way to model your data warehouse to leverage the features of VertiPaq.
“Just use partitioning” is the answer you hear, when you need to manage very large data sets in your Data Warehouse. But how do you design and implement it? We will walk through different ways to design partitioning, including layered partitioning.
This session will discuss what a modern strategy for data warehousing can be in this era, considering how the use of technologies like PowerPivot or Analysis Services Tabular affect the way you should model your data.
In this session, we are going to explain and test different DW features in SQL Server 2012, including star join optimization through bitmap filters, table partitioning, window functions, columnstore indices and more.
: We are looking at the dawn of a new workload: BigData. In this session I will talk about what BigData is, and which BigData technologies Microsoft are working on.
Users love flexible analytics but hate to wait for the data to be loaded into a traditional data warehouse. John will describe how to build an infrastructure to support real-time loading of your OLAP cubes so your user's get exactly what they want
SQL Server FastTrack References Architecture gives you the best practices and reference architectures from different hardware vendors to help you create a balanced environment for your data warehouse workload.
This session looks at some of the different methods available to load slowly changing dimension data into a data warehouse, and compares the relative performance given different data scenarios and traditional storage compared with FusionIO
This talk will describe how the new ColumnStore index technology in SQL Server 2012 makes queries go faster. Covering details of the storage and execution model, how this model interacts with modern CPUs to deliver significant performance benefits.
In this first of two sessions, we review the architecture of SQL Server and its BI components and deployment options for optimal performance. We'll also discuss how to optimize data warehouse load operations.
Snapshots without snapshots...is that possible? Take a "Classic" snapshot fact table, add some temporal data theory and you'll get a new fact table than can store snapshot data without doing snapshots. A life saver when you have a lot of data.
Do you have data warehouse queries that run too long? In this session we’ll address how columnstore indexes speed up queries, best practices for creating and using columnstore indexes, and how to diagnose and treat potential issues.
This session will drill into some of the details on the shipping HP BDA and BDW appliances and how they can work well together to deliver a comprehensive BI and DW solution for customers using the Microsoft Application Platform.
This session will provide expert advice & best practices for replicating Oracle (& other) data to SQL Server for data warehousing & BI initiatives. Also, become one of the first to see Attunity’s innovative new software in its historic unveiling!
Organizations risk being overwhelmed by data. How can you effectively provide a “single version of the truth”, while unlocking the key trends and insights that will allow your business to succeed? Come to this session to find out how.
Do you have complex dimensions in your data warehouse? Parent-child, late arriving, type 3 or type 6? In this session, we'll cover some SSIS patterns for handling each of these, along with tips for making them perform well.
Steps involved in implementing a near-real-time data warehousing solution.
Master Data Services has been given an overhaul in Denali, including a new Excel add-in and modified web front-end. Come to this session to see how MDS can be used to give greater control and process to your BI/DW dimension management.
When loading a Fast Track Data Warehouse it is important to ensure that your data is optimally laid out for Sequential I/O. Fragmentation is therefore the enemy. Know your enemy. Learn what it is, how it occurs and prevent it from happening to you!
In this session we will introduce the new modeling capabilities of Vertipaq, showing how the same scenarios can be modeled in both Multidimensional (MOLAP) and Tabular (Vertipaq), looking at how to enable your data warehouse to support both.
Many of us know what dimensions and fact tables are. But dimensional modelling is more than just dimensions and fact tables. This session is about advanced dimensional modelling topics, which are useful for anybody involved in the design of a DW.