Data Engineering with dbt - a Pragmatic Approach
- Rhys Hanscombe

- Nov 27, 2024
- 3 min read
Updated: May 22, 2025
In a recent webinar hosted by the Data Vault User Group, Robert Zagni shared his expertise on building the lightweight solution “The Pragmatic Data Platform” (PDP) discussed in his book, Data Engineering with dbt.
His presentation covered essential aspects of data architecture, automation, and best practices for data management. Here are some of the key takeaways from his insightful session.
1. Pragmatic Data Platform Architecture
Robert began by introducing the concept of a pragmatic data platform, emphasizing the importance of a solid architecture. He highlighted the need for a robust ingestion and storage layer, which serves as the foundation for any data platform. This layer should be designed to handle various data sources, ensuring that data is ingested and stored efficiently.
2. Automation in Data Ingestion and Storage
Automation plays a crucial role in modern data platforms. Robert demonstrated how automation can streamline the ingestion and storage processes, reducing manual effort and minimizing errors. He showcased examples of how tools like dbt (data build tool) have revolutionized data engineering by providing capabilities similar to those in software engineering, such as automated testing and version control.
3. Organizing the Refine Layer
One of the critical points Robert stressed was the importance of organizing the refine layer. This layer is where raw data is transformed into meaningful information. He advocated for using best practices from software engineering, such as clean code principles and modular design, to ensure that the refine layer is maintainable and scalable.
4. Master Data, Data Domains, and Data Mesh
Robert also touched on the concepts of master data, data domains, and data mesh. He explained that a softer approach to data mesh, where data projects are divided according to organizational needs, can be more practical for many organizations. This approach allows for better management of data domains and ensures that master data is consistently used across the organization.
5. From Art to Industry
A recurring theme in Robert's presentation was the transition from artisanal data practices to industrialized processes. He emphasized the need to rely on best practices and repeatable patterns rather than reinventing the wheel for each new data project. By adopting standardized methods, organizations can achieve more reliable and scalable results.
6. Practical Examples and Pet Projects
To illustrate his points, Robert shared examples from his own projects. He demonstrated how he has applied these principles in real-world scenarios, making complex data processes more manageable and efficient. His pet projects served as practical case studies, showing the tangible benefits of a pragmatic data platform.
7. The Role of dbt in Data Engineering
Robert is a strong advocate for dbt, which he considers an ultimate tool for data engineering. He explained how dbt enables data engineers to apply software engineering practices to data projects, making the development process more efficient and reliable. He also discussed the importance of having a clear ingestion timestamp and metadata to ensure data integrity.
8. Handling Real-Time Data
Finally, Robert addressed the challenges of handling real-time data. He suggested that while tools like Snowflake and dbt may not be ideal for ultra-low latency requirements, they are well-suited for near real-time scenarios.
Conclusion
Robert Zagni's webinar provided valuable insights into building a pragmatic data platform. By focusing on automation, best practices, and practical examples, he demonstrated how organizations can improve their data management processes.
