Intermountain Data Event 2016 and SQL Saturday Sessions


Data Technologies for Developers

Data Distribution with Apache Kafka
  5  

Data isn’t stationary – it has to move and as applications grow it has to move faster and with more data stores than ever before. Managing how data moves between all of these different data stores in a timely and reliable manner can be challenging; unfortunately, most attempts result in immense webs of brittle point-to-point integrations that struggle to scale. This is where Apache Kafka comes in. In its nearly seven-year history, Apache Kafka has solved some of the biggest challenges in “big data” distribution: processing hundreds of billions of records and moving petabytes of data per day in some of the largest enterprises. In this session I will introduce you to Apache Kafka and show you how it can solve your challenges with data distribution.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Ryan Plant


Can You Beat My MNIST Deep Net?
  3  

This talk will be a brief intro to deep learning focusing on the MNIST hand written dataset. This talk will go through the details and tricks of what is needed to take your deep nets to the next level. Overfitting issues? Problems converging? Need higher accuracy? Questions on augmentation? This is also an open call to PLEASE try and beat me, the top 3 MNIST scores sent to me before the talk will be given a chance to explain their scores and approach to the audience. Also, if you can beat me using Python I'll give you a $100 (wish it could be more).

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Ben Taylor


Data Distribution with Apache Kafka
  2  

<b>Apache Kafka</b>

Volunteers: No Volunteers
Want to volunteer? I Can Present This!


Dynamic Dimension Data Loading
  1  

This presentation will explain how we automated the way that our dimensions are loaded from the client schema's into a central dimension schema that is shared between the 90+ client schema's. We also show how we can add new dimensions to the process without writing any code.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Scott Heffron


Query Store Primer
  1  

This session will take a look at the new Query Store feature in SQL Server 2016. Query Store tracks changes in execution plans, allowing you to easily view performance differences and revert to older plans with a few clicks of the mouse. We will take a hands on look at how to use Query Store to troubleshoot problems as well as some of the new DMVs you can use to find issues as they arise.

Level 100 - Introduction
Duration: Hour
Presenter: Tracy Boggiano



Machine Learning

Machine Learning with Spark, GPU and Possibly Sparkling GPUs
  16  

I will present an introduction to data pipelining and machine learning done in Spark such as classification and regression. In comparison, I will also present how the same things can be done on a GPU machine. However, single-GPU face the challenge of dealing with large scale datasets or non-trivial development effort. At the end I will present the current status of the merger of both methods in accomplishing big data predictive analytics.

Level 100 - Introduction
Duration: Hour
Presenter: Anton Wiranata


Deep Speech Demystified (+ live demo)
  11  

Have you ever talked to Siri and had it fail to understand your command? Frustrating? Coming to this talk will help demystify the entire speech to text pipeline including the latest breakthroughs. Superhuman speech is essentially here where computers will soon be able to understand what we say better than we do. Talk includes live demo and breakdown. Skills discussed and demoed in this talk include: PyAudio Bokeh Deep Learning (Keras) Deep Learning (Torch)

Level 100 - Introduction
Duration: Hour
Presenter: Ben Taylor


The Nuts and Bolts of Machine Learning
  9  

An introduction to what machine learning is and an overview of different algorithms you can use and which situations work best. We will go over the perceptron algorithm, nearest neighbour, and random forests.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Jackson Pontsler


Bayesian Machine Learning
  7  

An introduction to bayesian methods for machine learning. We will start with Bayes theorem and MCMC sampling methods and end with examples using pymc3. Use this link to prep your environment before the session https://github.com/tfolkman/intermountain-data-conf-2016

Level 100 - Introduction
Duration: Hour
Presenter: Tyler Folkman


A Brief History of Deep Learning
  6  

Ever wondered about the technology that allows computers to beat the best human Go players and answer Jeopardy questions? This hour of fast-paced, research-driven exploration of the past, present, and future of deep learning is for you. We will take a high-level view that is perfect for beginners. From the first paper by Ivakhnenko and Lapa 51 years ago to the algorithms and technologies fueling the industry today to the next milestones in the field, the presentation is sure to expand your understanding of deep learning.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Ilya Reznik


Robust Image Embeddings via Deep Triplet Networks
  4  

At Recursion Pharmaceuticals, we search for new treatments for rare genetic diseases by growing cells and taking pictures. Each week, we generate between 3 and 6 TB of image data that must be mathematically embedded into a vector space for further analysis. This talk will provide a glimpse at how to learn a deep convolutional neural network that is robust to experimental noise by using Triplet ResNets in a clever way. We will examine Keras, explain the ResNet and Triplet network architectures, discuss the importance of batch normalization and ways to work with it, and learn a little bit of biology as well.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Mason Victors


NLP Intro - Teaching a computer to read
  4  

The wide world of Natural Language Processing has moved forward leaps and bounds over the past couple of decades. Now, we're at the beginning of the chatbot revolution, where we'll go from touch and swipe to real conversations with computers like in Star Trek. Come learn about some of the technology at the heart of this new age and see what makes it tick. In this quick overview, we'll see how to parse sentences, identify topics within a conversation and summarize a document. Come join us for a tour of the tech of the future and discover what might be coming sooner than you think.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Robert Davis


Deep Q networks! Using ML to play computer games!
  2  

Can your computer play Halo/Call-of-duty/Battlefield for you? Come to this presentation to find out how deep Q networks work, why they work, and see demos with real computer games being played using deep learning! This approach predicts what actions the computer should do based on ONLY the sequence of frame data (what it is looking at). In the near future these same networks will be used to play more advanced games and may even power some of our future deep learning military applications. This session will be team taught by Ben Taylor and Derek Headley from Colorado.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Ben Taylor



Data Analytics

The Life-Changing Magic of Tidying Text
  15  

Unstructured, text-heavy data is increasingly important in many domains, and tidy data principles and tidy tools can make text mining and natural language processing easier and more effective. In this talk, I will explore what tidy data is, focusing particularly on text mining using tidy data principles. We will explore how we can manipulate, summarize, and visualize the characteristics of text, using R packages from the tidy tool ecosystem such as dplyr, ggplot2, and tidytext, and effectively perform tasks such as sentiment analysis and measuring tf-idf.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Julia Silge


0 to 50M Rows a Day: A Year of Data Science at Jane
  8  

Jane.com is a rapidly growing e-commerce platform that sells boutique items. This talk will be a reflection on how to create and grow a data science team in a fast-paced, high-growth company. I will talk about how we scaled our technology, our team, and our processes. This will be an overview with some of the successes and failures in the last year. It will also cover some of the projects we have completed and what we hope to be able to accomplish in the next several years as we work to become the most personalized social shopping platform in the world.

Level 100 - Introduction
Duration: Hour
Presenter: Jeff Potter


R for Absolute Beginners
  6  

The statistical programming language R is a foundational tool in data science. This session will introduce R, the RStudio IDE, important built-in functions, and the vast collection of packages that give R enormous capabilities. Sample scripts and data will be provided, as will one month’s free access to the R training resources at datalab.cc.

Level 100 - Introduction
Duration: Hour
Presenter: Barton Poulson


Introduction to Reinforcement Learning
  5  

The idea that we learn by interacting with our environment is probably the first to occur to us when we think about the nature of learning. Exercising this connection produces a wealth of information about cause and effect, about the consequences of actions, and about what to do in order to achieve goals. (from the book Outwitting the devil, the devil said: know the law of nature and adjust yourself to it and you will get what you want). Learning from interaction is a foundational ideal underlying nearly all theories of learning and intelligence. This was simply the idea of a learning system that wants something, that adapts its behavior in order to maximize a special signal from its environment.

Level 100 - Introduction
Duration: Half Hour
Presenter: Anton Wiranata


Introduction to Data Warehousing
  4  

An introduction to Data Warehousing for those unfamiliar with the needs and benefits of star schemas / dimensional models. Data Warehousing introduced fact and dimension tables to provide simple and faster report writing. Measures/Aggregates and Dimensions/Filters have become the basic building for several reporting structures such as pivot tables, cubes and Tableau visualizations.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: John Kerley-Weeks


What is the Hype about Hadoop
  3  

This session is intended to enable you to engage in intelligent conversations about the Hadoop Ecosystem of products and Big Data Analytics.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: John Kerley-Weeks


Introduction to Causal Inference
  3  

Correlation does not imply causation, but that doesn't mean that we can't infer causation. Causal inference techniques can help us learn about actual causal effects and relationships in our data. We'll cover the assumptions required for causal inference, outline the basics of a causal model technique, and work through an example using a bayesian framework.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Paul English


Big Data Analytics for BI and Data Science
  2  

Big Data Analytics HPE's Big Data Platform provides analytics for use by line of business analysts and data scientists alike. HPE provides insight and blazing fast analytics across all data, including unstructured, semi-structured and structured sources. This session will provide insight towards harnessing your data for innovative big data analytics leveraging machine learning and the HPE Big Data Platform solution.

Volunteers: Mike Eisterer
Want to volunteer? I Can Present This!



Professional Development

Data Science Done Wrong: Dissociation and Disempowerment
  6  

The development of extraordinary data science tools and the revelations of major data science projects tend to obscure a troubling disconnect: specifically, data science may be disempowering people and dissociating them from their own humanity. Billions of lives are affected by data science, of course, but only an infinitesimal set of people practice data science. Put another way, data science is typically not done BY people but done TO them, and the results are often neither pretty nor empowering. This presentation outlines some of the social, technical, and even philosophical issues that contribute to this suboptimal situation. More importantly, this session opens a discussion on how we, as data science practitioners and evangelists, can humanize our work and better serve both our clients and our communities.

Level 100 - Introduction
Duration: Hour
Presenter: Barton Poulson


Data Careers Deconstructed: Next Steps for All of Us
  6  

Data careers have exciting new possibilities and challenges. This talk addresses things employees and employers need to know. How are companies finding and using data professionals? What skills are companies looking for? Do we have reasonable expectations from employees and companies? Where do we fit? We'll pay special attention to career paths that emerge gracefully from common jobs you might already have.

Level 100 - Introduction
Duration: Hour
Presenter: David Richards


Docker for experimenting with new technologies
  3  

Is your team trying to test new technologies for a project? Are you worried that your local environment may get affected when playing with new tools? Moving from MySQL to MongoDB, and need all developers to upgrade their workstations? New web server? New build tool? Or do you work in multiple projects in a daily basis, and want a quick way to switch from project to project? Docker can help you and your team accomplish this with a lot of flexibility and minimum invasion to your workstation.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Andres Arias



Architecture

Using R for Data Profiling
  8  

The benefits of knowing your data before embarking on a BI project are endless. Sure you can buy a tool to help with this. Or you could use R, an open source tool. This session will dig into methods for using R to connect to your data source to see visual and tabular analyses of your dataset. You'll learn how to find missing data, outliers, and unexpected values. If you don’t know R or you are wanting to learn more functions within R, you’ll benefit from this session.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Michelle Kolbe


Columnar Databases: Adventures of Warehousing in Redshift
  6  

More companies are beginning to consider the performance and scalability of columnar databases for data warehouse applications over the traditional row-based architecture. Based on our experiences building out a new data warehouse at Jane.com, this presentation will cover some design considerations for implementing a warehouse in AWS Redshift as well as give pointers to SQL tuning in this kind of environment. I will attempt to contrast these ideas with those considerations used in a row-based database, for those coming from a traditional RDBMS background.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Nick Johnson


Alluxio: Speeding Up The Distributed File System
  3  

Distributed file systems (Hadoop) and distributed object stores (S3) are at the heart of any truly large data infrastructure. They provide an environment for fast writes and mostly quick reads on embarrassingly large datasets. At large enough scale, these solutions are slow due to the costs of disk and network I/O. Alluxio is an in-memory virtual distributed file system that works with existing solutions to provide up to 100X speed boosts over native distributed file and object stores. In this talk I'll walk through the motivation for using Alluxio, the ease of use, architecture and an example of Alluxio working on a Swift Object store with Apache Spark.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Jowanza Joseph


The Future Is Fast: Event-driven and Stream-based Architecture
  3  

For the last several years the conversation about data has mostly centered around how to process its increasing volume and variety. More recently, however, the focus is shifting more towards the need to process all of this data with more velocity. As systems become more distributed and artificially intelligent it is going to be more important to master fast data, not just big data. In this session I will expound on the philosophies behind “event-driven” or “stream-based” architecture and illustrate the concepts with reference designs and technologies that can serve as a guide for how to starting thinking about data architecture for the present and future.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Ryan Plant



Business of Data

Lean Analytics in the Enterprise
  5  

Data Sources. ETL. EDW. KPIs. Visualization & Monitoring. These are all aspects of supporting analytics within enterprise organizations, often with disparate teams handling each layer. In this talk, I'll walk through the methodology of lean analytics and how it addresses each layer within enterprise organizations to drive the business of data - currently separating companies who are surviving from those who are thriving.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: George Ortiz


Basic Reporting and the "Why" of Data
  5  

More and more SQL developers are asked to do more and more reporting. We will spend time going over the basic tools needed to build good reports.(SSRS, Power BI, Cubes) Also, as SQL and Data Nerds, we sometimes forget to ask the question...why? Why is this data important? Why do you need it? We're so focused on the "if" we can do it, we don't ask if we "should". I will show how answering this question will help your simple reports and BI tools because the coolest things ever.

Level 100 - Introduction
Duration: Hour
Presenter: Matt Horton


MicroStrategy Table Partitioning
  4  

We needed a way to transparently access client data in separate schema's, but make it appear to be from one area. This presentation will show how we accomplished this and how the internal and external users can access multiple clients data using the same report interface. I will show the gotcha's that we encountered.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Scott Heffron


Salary Surveys in Big Data
  2  

An overview of Salary Surveys in the Big Data Arena from Dice, O'Reilly and a variety of job boards

Level 100 - Introduction
Duration: Hour
Presenter: John Kerley-Weeks



Data Visualization / Story Telling

How to not kill your cool insight with a bad graph: ggplot and other visualization tips
  9  

Conveying results is a key part of data analysis and a bad graph can buzz kill any cool insights when presented to technical and not technical audiences alike. I be will discussing some key points of the science of data visualization and creation of visuals through the ggplot package available in R and Python. I will cover show some of the cool tricks to effectively and simply visually encode variables to create beautiful figures that excel at conveying the data insights you've work so hard to find.

Level 100 - Introduction
Duration: Hour
Presenter: Seth Paul


Think Broad. Mine Deep. Explain Simply. Effective Storytelling With Data
  6  

Storytelling with data is where art meets science, and thanks to the digital revolution it’s never been easier to collect information and translate it into actionable insights. In a world where 80% of data projects are failing, this presentation explains a simple three-step process to turn any data into a deliverable with transformational impact.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Luciano Pesci


Why storytelling matters and how to communicate actionable insights
  6  

The main difference between the top 10% of Data Scientists and the bottom 90% is a distinction most people overlook: Data Storytelling. No matter who you are, you’re going to have a message that you want other people to hear. Often they are product managers, other data scientists, marketing folks, user research or Execs. Great data analysts start by providing insights that have value to the business. However, insights provide no value if they stay in your head! Communicating data science effectively to your business partners is crucial. With access to open source tools, even novices can have an amazing tool kit at their disposal; An array of statistical techniques, algorithms, and more computational power than you know what to do with. However, all of your shiny cool skills will go to waste if you don’t provide information that is relevant to your stakeholders. Simply, a job isn’t paying you to construct a complex neural net, they are paying you to provide value towards a business question. In this workshop you will learn how to get your data science message to the next level and ultimately get buy-in from your audience.

Level 200 - (Beginner): Introductory / fast moving
Duration: Hour
Presenter: Alton Alexander



SQL Saturday Sessions

SQL Analytics - Beyond SQL 101
  7  

A rapid introduction to common SQL Analytic functions utilized by data analysts including SQL clauses such as GROUP BY, PARTITION BY, windowing, RANK, LAST, LAG, TOP and ROLLUP. Basic knowledge of SQL is required for this session.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: John Kerley-Weeks


SQL Server Administration Fly By
  6  

Overview of SQL Server Administration driven by the study guide for the 70-462 exam requirements.

Level 100 - Introduction
Duration: Hour
Presenter: Jacob Case


SQL Server's Crystal Ball - Predicting Your Future Performance
  3  

DBAs are constantly barraged with suggestions for improving their system. Maybe it's a new partitioning scheme, upgrading to SQL 2016, or implementing a high-availability solution. Each of these changes, whether large or small, will impact the performance of your system. But how can you be sure how large the impact will be? Google will tell you that upgrading to SQL 2016 results in a 34x performance increase and the impact of Always On is negligible, but those are based on a generic benchmarks or someone else's workload. How can you predict the impact on YOUR environment? In this session, we will show you how to use Distributed Replay to baseline performance characteristics and evaluate the impact of database changes. We'll show how to set up and configure it properly, and how to correctly analyze the results. At the end of the session you will be prepared to throw away generic internet benchmarks and confidently measure the impact of database changes in your environment.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Chad Crawford


When the Bough Breaks - Solutions for a Database Performance Catastrophe
  3  

Parameter sniffing and bad statistics are the two most common causes of abrupt and catastrophic database performance problems. They appear suddenly and sometimes disappear inexplicably. Understanding how to accurately diagnose and recover when they occur is a critical skill for maintaining performance SLAs and application up-time. This session will cover the optimizer internals as they relate to these two issues, and the data distributions that cause them to surface. We’ll use examples to show how to correctly diagnose the root cause, and help you differentiate between solutions that just hide the problem and those that actually resolve it.

Level 300 - (Intermediate): Basic knowledge of subject matter is suggested
Duration: Hour
Presenter: Chad Crawford


Sql Server Management Studio Tips and Tricks
  1  

Showing tips and tricks of using SSMS

Level 100 - Introduction
Duration: Hour
Presenter: Andrea Allred


Extended Events - Better than Profiler
  1  

Extended Events were introduced in SQL Server 2008. With SQL 2014, we have seen a significant upgrade to this feature. Join me for a little adventure into the realm of extended events. We will discuss how to use extended events to aid in performance tuning and in day to day administration. We will also explore some background and the architecture of extended events. This is a fast paced session and will be a lot of fun. Attendees will be able to take back several sample XE sessions to help troubleshoot different issues and get to more intimately know their data/environment.

Level 100 - Introduction
Duration: Hour
Presenter: Jason Brimhall


Alerts: Knowing Is Half the Battle!
  0  

Learn how to quickly configure alerts for several instances using tools that ship with SQL Server. Increasing visibility of the errors occurring in your servers can expose the cobras that have been hiding there all along. Scripts will be made available to attendees.

Level 100 - Introduction
Duration: Hour
Presenter: Lars Rasmussen


Creating an Always On Distributed Availability Group
  0  

Always On has developed and matured into an effective HA/DR solution with SQL 2016. Most enterprise DBAs have built a cluster and installed Always On Availability Groups (AOAG), this session will take the Always On technology one step further and discuss the new features in SQL 2016 that will allow a DBA to have live, read only database replicas in a geographically diverse location, in a completely different domain cluster and as a member of a completely separate Availability Group (AG). This session will go through the steps of identifying the pre-requisites, limitations and installation steps to setup a Distributed Availability Group from start to finish! Come and watch how to do it!

Level 100 - Introduction
Duration: Hour
Presenter: Dale Cunningham


Introduction to the SQL Server Query Store
  0  

Don't you wish there was a recording of everything running on your SQL Server instance tracking performance data? Well in SQL Server 2016, Microsoft has introduced a new feature called the Query Store. This session will go over how the query store works, how you configure it for your workloads and how to use the information to track performance data. When you leave this session, you will know how to get started using the Query Store.

Level 100 - Introduction
Duration: Hour
Presenter: Brandon Preece


Inside the SQLPS PowerShell Module
  0  

SQLPS has been evolving since 2012 and the SQL Provider/cmdlets have been around since 2008. Have you taken advantage of the PowerShell infrastructure that is on your servers? This session takes you behind the scenes of the SQLPS module that is installed with the tools. We will cover how to get it without the tools, how to use it once you have it and why you will want to use it for simple tasks or complex ones. When you learn how to navigate SQL Server and all its services using Paths, you won't go back. Join me for a fun filled hour and learn how to keep SQLPS from becoming frustrating and a black box. You will be glad you did. Let's continue the journey to becoming a PowerShell DBA.

Level 100 - Introduction
Duration: Hour
Presenter: Ben Miller


Power to the people
  0  

Over the years Microsoft has released many power products. We are all familiar with PowerPoint and PowerShell but how familiar are you with PowerQuery, Power Q&A, PowerView and Power BI? In this session we will take a closer look at the new tools Microsoft has released to enable true self service BI

Level 100 - Introduction
Duration: Hour
Presenter: Steven Seeley


Introduction to Azure SQL Data Warehouse
  0  

This session will introduce you to Microsoft Azure SQL Data Warehouse and show you how it compares to other technology options. You will learn about the benefits and idiosyncracies of working with SQL DW's massively parallel processing architecture. Then you will get to see a demo of provisioning, loading, and querying SQL DW.

Level 100 - Introduction
Duration: Hour
Presenter: Jimmy May


SQL Server 2016 Always On Availability Groups Perf & Internals
  0  

Always On Availability Groups (AGs) were introduced in SQL Server 2012. AGs significantly expanded our HA/DR options. However, in both SQL 2012 and SQL 2014 two limitations were extant in the code—both Network Transport and Log Redo significantly compromised performance. In SQL Server 2016, these limitations have been significantly improved. Our work shows enhancements of 7x – 10x, allowing over 1/2GB/s (4Gb/s) across the wire. These improvements expand solutions options for Tier 1 OLTP and Data Warehouse workloads, Tier 1 and Tier 2 consolidation, and more. This presentation explains the improvements, characterizes the new solutions, and shows commodity servers and modern non-shared in-server flash storage showing the profound contrast between SQL Server 2014 vs. 2016. Doing so moves the performance bottleneck from the SQL Server codebase to the CPU where it belongs.

Level 100 - Introduction
Duration: Hour
Presenter: Jimmy May


Advanced SSIS Techniques
  0  

As an ETL developer using SQL Server Integration Services have you ever wondered how to: 1. Import all .csv files from a folder 2. Replace dynamic SQL using variables in SSIS 3. How to configure the data flow buffer 4. Create custom logging solutions 5. Use a script component as a data flow transformation This session focuses on advanced techniques that are used in SSIS and is meant for the experienced SSIS developer. We will focus on tasks and techniques that are native to SSIS 2005-2014 as well as some of the more advanced and vague properties. Attendees must have a good working knowledge of SSIS as well as familiarity with VB.NET or C# as some time will be taken looking at extending SSIS with script tasks and components. Upon completion of this workshop attendees will be able to: 1. Import/export multiple sources/destinations 2. Use an execute SQL task with variables to replace dynamic SQL within a package 3. Dynamically handle errors within any part of a package 4.

Level 100 - Introduction
Duration: Hour
Presenter: David Dye


M&M&M: Making the Most of the M Language - Real World Power Query for Excel and Power BI
  0  

Learn how to leverage the diverse abilities of Power Query to acquire and transform data in this example driven session. Our journey will begin with CSV and we will work our way through HTML tables, JSON data, stop to examine REST and OData endpoints, and end with the spectacular view of connecting to SQL Server 2016. Each of these data types can present a unique opportunity to combine elements of the M language in Power Query to transform and massage the data before finally presenting to the visualization engine for display. Come learn to use the tricks of expanding records and nested tables to flatten out your data model and be better prepared to tackle any size or shape of data that comes your way.

Level 100 - Introduction
Duration: Hour
Presenter: Greg McMurray


Disaster Recovery Where to Begin
  0  

So you have a good backup strategy in place, but now you have to think about a disaster recovery. Where do you start and what are your options? SQL Server offers several options, but then how do you choose the right one to meet your business requirements? In this session, we will go over some of the common DR options including log shipping, replication, mirroring, clustering, and AlwaysOn. Then we will talk about how to choose the best one to meet your business needs.

Level 100 - Introduction
Duration: Hour
Presenter: Tracy Boggiano


PowerBI, from a DBA
  0  

Microsoft’s PowerBI is the new hotness. In this session, you will get your hands dirty and dive into PowerBI and create an actual report, with actual data. You will get to learn from examples from what a DBA has done to implement PowerBI in his shop and make his data accessible. As a DBA, we often need to collect and display data in a friendly and interactive fashion. We will play with some PowerShell to collect data from several systems, store this data in an Azure DB in the Cloud, then point PowerBI at this data to display and interact with said data. You will learn a little about PowerShell, Azure SQL DB, PowerBI, and how to make these work for you. You will be given permissions to an Azure DB, and get connected to this source. You will also need to have PowerBI desktop downloaded and installed. We will be using this to create the report during the session. So please come prepared to interact with the internet, with a cloud-based database you will be provided, and PowerBI Desk

Level 100 - Introduction
Duration: Hour
Presenter: tjay belt


Dealing With Errors in SSIS
  0  

Errors can occur anywhere within an SSIS package and the cause and error message can be somewhat cryptic. This session focuses on identifying error causes and handling them dynamically and proactively. The means of dealing with errors is defined by where the error occurs and we will cover error handling in control flow, data flow, and script tasks in great detail. You will learn: •To identify the different error handling methods in control flow, data flow, and script tasks •Be able to utilize event handlers to proactively handle errors •Be able to redirect data error and proactively handle data flow errors •Be able to identify error descriptions and log error details

Level 100 - Introduction
Duration: Hour
Presenter: David Dye


Database Restore Testing...Automated
  0  

How do we know the backup files of our databases are free from corruption? Why would we care? The hardware storing the backup files is reliable…right? What could possible go wrong? Right? The only way to know our database backup files are in good order is to validate them. How do we do that? One way is to restore the databases from those backup files and check the integrity of the restored databases. This session will show some benefits of database restore testing and walk through the code of a SQL Agent job that automates it.

Level 100 - Introduction
Duration: Hour
Presenter: Robert Every


Bare Naked Skinny on SQL Server
  0  

Ever feel like there are just too many layers with SQL Server? You peel off a layer and behold there is yet another layer. SQL Server is rather complex. The complexities can be rather frustrating from time to time. Join this NSFW (not safe for whining) session as we strip SQL down to the bare bones. I will help you undress the complexities of a few of the basic SQL Server concepts. These concepts will include locks, principals, query directives and deadlocks. In this session I will disrobe misunderstandings related to the aforementioned topics. I will show the different types of principals in SQL Server. I will also show various scenarios involving locks and deadlocks along with some of their effects with and without query directives. The goal is to better understand these topics and to hopefully stop using and perpetuating common mis-uses of terminology and common misconceptions about the technology.

Level 100 - Introduction
Duration: Hour
Presenter: Jason Brimhall


Introduction to SQL Server Internals
  0  

Getting started with the internals of SQL Server

Level 100 - Introduction
Duration: Hour
Presenter: Andrea Allred


SQL Server R Services in SQL 2016
  0  

One of the most exciting new features of SQL Server 2016 is the integration of the R statistical programming language into the platform with the brand-new SQL Server R Services. In this introductory session we'll briefly introduce the R language and then look at how to install and administer SQL Server R Services. We'll start to demonstrate the power of R by showing you how to create and run simple R scripts from within T-SQL. We'll then discuss some of the powerful open source packages that are available to aid your data analyses. Finally, we'll go through some examples of how to use R to integrate data analytical methods in real-world situations, and wow your customers with the power of statistics!

Level 100 - Introduction
Duration: Hour
Presenter: Chris hyde



Microsoft Technologies

There are no presentations in this track.


Suggest a Topic!

 

Track Name

Suggesting a topic does not mean you are presenting that suggestion.

Session Name

Track

Level

Duration

Session Abstract


Would you like a mentor to help you with this presentation?



The unconference track is designed to create sessions and ideas on the fly. We want to allow our attendees to use a room through the day for open conversations and concepts. Everyone can vote on the ideas you want to see and then we will publish them in the schedule. Feel free to submit any topic/idea for discussion.

Idea Name

Description