The Data Warrior

Changing the world, one data model at a time. How can I help you?

Archive for the tag “Snowflake”

Snowflake Summit 2021

It’s almost time! The annual Snowflake Summit is just a few weeks away.

Data is everywhere and it holds the keys to unlocking your organization’s success today and in the future. Attend Snowflake Summit and you’ll learn how to use the #DataCloud to unify, analyze, and share data previously out of your reach for more impact than you have ever imagined.

Data Together Now will inspire business and technical leaders, data scientists and engineers, data and analytics professionals, and application developers, to lead their organizations to a data-driven future, today. There is something for everyone in the organization.

You can be a DATA LEADER!

SAVE THE DATE:

AMERICAS: June 8-9  | 9  am PT  |  12 pm ET |  View Agenda

ASIA PACIFIC: June 9-10  | 10 am AEST  |  9 am JST |   View Agenda

EMEA AND INDIA: June 9-10  | 9 am BST  |  10 am CEST  |  1.30 pm IST |  View Agenda

So what are you waiting for? Register today.

Let it snow!

Kent

The Data Warrior

Building a Real-time Data Vault in Snowflake?

Yes you can! The #DataCloud loves #DataVault!

In this day and age, with the ever-increasing availability and volume of data from many types of sources such as IoT, mobile devices, and weblogs, there is a growing need, and yes, demand, to go from batch load processes to streaming or “real-time” (RT) loading of data. Businesses are changing at an alarming rate and are becoming more competitive all the time. Those that can harness the value of their data faster to drive better business outcomes will be the ones to prevail.

One of the benefits of using the Data Vault 2.0 architecture is that it was designed from inception not only to accept data loaded using traditional batch mode (which was the prevailing mode in the early 2000s when Dan introduced Data Vault) but also to easily accept data loading in real or near-realtime (NRT). In the early 2000s, that was a nice-to-have aspect of the approach and meant the methodology was effectively future-proofed from that perspective. Still, few database systems had the capacity to support that kind of requirement. Today, RT or at least NRT loading is almost becoming a mandatory requirement for modern data platforms. Granted, not all loads or use cases need to be NRT, but most forward-thinking organizations need to onboard data for analytics in an NRT manner.

See all the details (and some code) in the full post over on Data Vault Alliance.

Happy Vaulting!

Kent

The Data Warrior

The Data Cloud Tour: Mobilize a World of Data

6 INDUSTRY EVENTS

4 REGIONS!

Most years I have been at Snowflake, we have done Data for Breakfast events all over the world to kick off a new year. But this year is different, so instead we are launching a virtual #DataCloud Tour! Join us to learn how the Data Cloud can power your data strategies and help you deliver innovative and industry-leading products and services.

In this tour we will actually run seven different events in each region – one for the generalist and then separate sessions for these industries:

  1. Healthcare and Life Sciences
  2. Financial Services
  3. Public Sector
  4. Retail & CPG
  5. Media, Entertainment, & Advertising
  6. Manufacturing

Each event will bring you industry SMEs talking about how the Data Cloud has helped them be data leaders and succeed in getting the more value from their data in the most cost effective manner. In the talks you will learn how to break down data silos with The Data Cloud and how you can put the Snowflake platform to work for your company.

The Data Cloud is a global network where thousands of organizations mobilize data with near-unlimited scale, concurrency, and performance. Inside the Data Cloud, you can unite your siloed data, easily discover and securely share governed data, and execute diverse analytic workloads.

See the full agenda and reserve your spot (#Free) here: The Data Cloud Tour

#MobilizeYourData

Kent

The Data Warrior & Chief Technical Evangelist, Snowflake

Snowflake and #NetZero: The Case for Data Decarbonisation (Part One)

I love the fact that Snowflake is getting involved in the betterment of our planet!

This is the first in a series of three posts from one of my colleagues examining the concept of net zero data and how advances in technology can help the world’s largest world’s largest organisations—especially those which are particularly emissions-intensive like oil and gas—reduce the carbon emissions footprint of their data. 

Read the whole post here –  Snowflake and Net Zero: The Case for Data Decarbonisation (Part One)

Enjoy!

Kent

The Data Warrior & Chief Technical Evangelist at Snowflake

#LetItSnow

 

 

Data Vault 2.0 Automation with erwin and Snowflake

I am seeing a HUGE uptick in interest in Data Vault around the globe. Part of the interest is the need for agility in building a modern data platform. One of the benefits of the Data Vault 2.0 method is the repeatable patterns which lend themselves to automation.  I am please to pass on this great new post with details on how to automate building your Data Vault 2.0 architecture on Snowflake using erwin! Thanks to my buddy John Carter at erwin for taking this project on.

The Data Vault methodology can be applied to almost any data store and populated by almost any ETL or ELT data integration tool. As Snowflake Chief Technical Evangelist Kent Graziano mentions in one of his many blog posts, “DV (Data Vault) was developed specifically to address agility, flexibility, and scalability issues found in the other mainstream data modeling approaches used in the data warehousing space.” In other words, it enables you to build a scalable data warehouse that can incorporate disparate data sources over time. Traditional data warehousing typically requires refactoring to integrate new sources, but when implemented correctly, Data Vault 2.0 requires no refactoring.

Successfully implementing a Data Vault solution requires skilled resources and traditionally entails a lot of manual effort to define the Data Vault pipeline and create ETL (or ELT) code from scratch. The entire process can take months or even years, and it is often riddled with errors, slowing down the data pipeline. Automating design changes and the code to process data movement ensures organizations can accelerate development and deployment in a timely and cost-effective manner, speeding the time to value of the data.

Snowflake’s Data Cloud contains all the necessary components for building, populating, and managing Data Vault 2.0 solutions. erwin’s toolset models, maps, and automates the creation, population, and maintenance of Data Vault solutions on Snowflake. The combination of Snowflake and erwin provides an end-to-end solution for a governed Data Vault with powerful performance.

Get the rest of the details here: Data Vault Automation with erwin and Snowflake

Vault away my friends!

Kent

The Data Warrior

Post Navigation

%d bloggers like this: