The Informed Company. Dave Fowler
Чтение книги онлайн.
Читать онлайн книгу The Informed Company - Dave Fowler страница 10
Let's begin with an overview of each step.
Stage 1. Sources
Many teams only have a few data sources and commonly manipulate and visualize data in Excel or Google sheets. Two common sources that company products run on are Google Analytics and “transactional” database(s), such as PostgreSQL or MySQL. If only a few people at a company need to work with data sources, it's possible to set them up with direct or replica access; it's more agile and straightforward for them to work with the data directly.
We discuss how to configure proper access to data and how to lay the foundation for future organization into a single, unified collection of company data. We also discuss best practices when using tools like in‐app dashboards, spreadsheets, SQL, and BI solutions. Behind these assertions is the principle that teams should be vigorously deliberate when organizing their sources and picking their toolkit. Keeping this in mind not only makes analysts effective at dealing with data early on, but also prepares them to reorganize their input streams into a data lake later on.
This stage is right if:
Only a few people are going to be working with this dataset.
Data needs are minimal at the moment.
Only a few small data sources exist.
The only people who need to make new visuals are fairly technical.
It's time for the next stage if:
Data is accessed from multiple places/applications.
There are needs for unique or combined charts/dashboards for cloud application sources like Salesforce.
A growing number of people need access to data.
There are performance issues.
Data is getting too big for a transactional database to operate efficiently.
Nontechnical users need to create charts without help.
Stage 2. Lake
Once companies must run analyses on multiple sources of data, each of which need joining, filtering, and manipulation, a company must move to a data lake. Blended data sources enable several actors in an organization to query a large subset of the company's complete data. In turn, funneling various sources into a data lake supports database performance at a reasonably large (not necessarily “big data”) scale.
A central motivation for a data lake lies with the need for piping data to business intelligence tools. For example, when working with data from Salesforce, Hubspot, Jira, and Zendesk, each service has its own in‐app dashboards and unique data application programming interfaces (APIs). Configuring input data streams for each business tool is a confusing, time‐consuming, and unsustainable workflow. It cannot really be done, especially at scale. Likewise, performing in‐house analyses across various sources can wildly complicate otherwise simple queries. On the other hand, having a data lake, which holds all relevant data in one place, allows analysts to use straightforward SQL queries to obtain business insights.
The central challenge faced by companies in the lake stage is knowing what toolset and methodology will unify and (safely) store your data. Companies looking to combine their data also run into performance issues, which we offer solutions to. And perhaps most important of all, choosing an architecture during lake development informs how easy (or hard) it will be for a company to build their future data warehouse.
This stage is right if:
There's a need for unique or combined charts/dashboards for cloud application sources like Salesforce.
A core set of people can learn the ins and outs of the structure of the messy data.
You're intimidated by data modeling. (Don't be: that's why this book exists.)
There's no time for even light data modeling.
Large datasets need performant queries.
It's time for the next stage if:
More than a few people are going to be working with this dataset.
A clean source of truth would eliminate integrity issues.
There's a need to separate the structure of the data from the always‐changing transactional sources.
There's a need to adopt DRY (Don't Repeat Yourself) principles.
Data requires transformation so that it is more usable by different people or systems within a database. Modeling refers to this process of making these transformations to the data.
An acronym that represents a software design ethos that avoids repetitive patterns through modularity of design.
Stage 3. Warehouse (Single Source of Truth)
As more people begin to work with the data lake, questions begin to multiply: What data is where? Why? What particular criteria should queries use when looking for data insights? What do these schemata mean? Unavoidable complexities make it harder to obtain data, especially by less‐technical colleagues. Even among in‐house experts, more schemata and entities (i.e. tables and views) in turn cause more communication headaches. In time, the data lake serves all data but makes it harder to obtain the right data. It gets harder to write queries and share the knowledge within an organization.
All of these problems can be addressed with a clean and simplified version of the data, something we refer to as “a single source of truth.”
This stage—creating a data warehouse—has historically been quite a nightmare, and there are many books written on how best to model data for analytical processing. But these days, there are more straightforward paradigms that have been tried and tested: ones that not only streamline having to document the oddities found across an organization's schemata but also save time in having to repeat, edit, and maintain messy “boilerplate” query steps (e.g. “every time you query the order's table, make sure to adjust all orders from England to be in local time”).
In the data warehouse section of the book, we review how to clean data lakes and investigate standard practices for managing data complexity. In addition, we offer ways to establish an architecture with data integrity in mind. We provide modeling tool suggestions and an example SQL style guide. Finally, we give our recommendations for team structure, such as a lead to oversee this process and warehouse maintenance.