SSIS in SQL Server 2012 has a lot of new features to help control larger projects. I'm doing a trial migration of our datawarehouse (100+ packages, 1TB+ db size) This session will be a will be a walkthrough of this process and a lessons learned
SQL Server Integration Services (SSIS) 2012 includes a new database referred to as the SSIS Catalog. Come to this session to learn what it does and how its going to change how you do SSIS development.
Learn how Varigence tools accelerate BI development, enhance reusability, and improve agility through the use of customizable patterns and automation scripts and transform your metadata into comprehensive BI solutions.
This session looks at some of the different methods available to load slowly changing dimension data into a data warehouse, and compares the relative performance given different data scenarios and traditional storage compared with FusionIO
This demo intensive session will show off the new features and enhancements designed to ease adoption for new users and increase productivity for experienced SSIS developers
The SQL Server 2012 release of SQL Server Integration Services (SSIS) includes a broad range of new capabilities designed to improve the deployment, management and monitoring of SSIS solutions. The core of these capabilities is the new SSIS Catalog.
Learn all about the integration of Business Intelligence Markup Language (BIML) into BIDS Helper and see what it can do for you. BIML can help you automate the creation of SSIS packages, reduce maintenance and help code re-use across projects.
Unit testing is incredibly beneficial, but is very difficult to do effectively in SSIS. If you have wanted to do this, but aren't sure how, this session will cover the options and demonstrate a free tool for unit testing packages.
In this session you'll get a good introduction of the concepts and advantages of Biml as we walk to the creation of Meta Data Driven SSIS Solution.
In this “demo-tastic” presentation, SSIS trainer, author, and consultant Andy Leonard explains the what, why, and how of an SSIS framework that delivers metadata-driven package execution, connections management, and centralizes logging.
Do you have complex dimensions in your data warehouse? Parent-child, late arriving, type 3 or type 6? In this session, we'll cover some SSIS patterns for handling each of these, along with tips for making them perform well.
This session shows off the new features of Microsoft SQL Server Integration Services in SQL Server code-named “Denali”.
The SQL Server provides much functionality that increases the scalability and flexibility of your solution by distributing data and jobs among low-priced commodity servers.
A look at the new SSIS features in SQL Server code-named Denali
Executing SSIS Packages and handling the errors
Writing your first SSIS custom components can seem like a very steep learning curve. In this session i shall walk you through a simple skeletal one to start you on your way.
Building performant data flows takes more than just dragging a few boxes onto a design surface. In this session I'll demonstrate that SSIS perf tuning is less about fine-grained tweaks and more about designing packages correctly in the first place.
Attunity will be presenting their award winning change data capture (CDC) products for SSIS. This session will discuss whatchange data capture is, why it's used, and will demonstrate how CDC processingcan be seamlessly integrated into SSIS.
I had to import and process 100+ files for my latest project and wondered if there was a way to generate SSIS packages based on a repository table containing the rules. As it turns out, this can be done but the programming model is a bit awkward.
The samples that you can find online are not allways functional and debugging is difficult. I did get it to work though and will show you how I did it and what problems I ran into.
After this session you will know how to generate packages that contain sources, destinations, cached lookups, derived columns, conditional splits, sorts, union all, etc.
This session is not about the script component inside SSIS but about generating SSIS packages and programmatically reading and changing properties of existing packages.This session should also be usefull if you're not a full blown developer but more of a DBA type. I'm a DBA and I got through this :-)
My Name is Thomas Keyser and I work for Microsoft in the SQLCAT team. I hold the land speed record for data loading in SQL Server so come here my talk on the options you have.
Need to eek out a bit more oomph from your dataflows. This session might be of some use.
After years of careful observation and recording of the Species SSIS, Genus ETL, in both natural and artificial environments, I’ve gathered a large collection of fieldnotes. Find out the little decisions that get made that can have big impacts on you project later on. What option to choose for better performance, or better management, and what are the trade-offs? This session won’t just focus on holy grail of performance, it will review the way you build and use SSIS taking into account the on going maintenance and management aspects as well.
Do you wonder about SSIS performance? Well I do, and I've compiled my research into this session. We'll cover various design patterns for solving common problems like inserts vs. updates, is it faster to use a lookup, or can you just catch the errors and process them afterwards? As well as the richer patterns we'll look at some straight comparisons between two components that can be used to do perform the same task and ask which one is quicker?
If you want to understand how to load data into Sql Server very quickly then this is the session for you.
Fast Track is a new reference data warehousing architecture provided by Microsoft. More than this it represents a new way of thinking about data warehousing. A Fast Track system is measured by its raw compute power - not by a DBAs ability to tune an index. Fast Track is an appliance-like solution that delivers phenomenal performance from a pre-defined, balanced configuration of CPU, memory and storage using nothing but commodity hardware.
Of particular interest in a Fast Track system is the way in which the storage and SQL Server are configured. To achieve the fantastic throughput without using SSDs requires some careful configuration. This configuration is designed to make use of Sequential I/O to dramatically improve disk I/O performance.
Interested? If you have a large data warehouse that's seen better days or perhaps you are about to embark on a new warehousing project then you should be! Fast Track is a great solution with a fantastic value proposition.
In this one hour session we'll aim to get under the skin of Fast Track and get some answers as to how it delivers such great throughput on commodity hardware. In the process we'll aim to answer the following questions:
* When might I need Fast Track?
* What is Sequential I/O?
* How does Sequential I/O improve performance?
* What do I need to do to get Sequential I/O?
* How can I monitor for Sequential I/O ?
* What may I need to change in my ETL to get the benefit of sequential I/O?
Still reading? I'll save you a front row seat....