Back to Glossary
Data modeling is the process of structuring data in a way that makes it easy to store, query, and analyze. It defines how data tables relate to each other and how metrics should be calculated. Good data modeling is the foundation of reliable BI and analytics.
The most common analytics models include:
Star schema (fact table + dimension tables)
Snowflake schema (normalized dimensions)
Wide tables (denormalized for simplicity)
Fact tables store measurable events like orders, sessions, or transactions. Dimension tables store descriptive attributes like customer, product, or date.
Data modeling helps:
Improve query performance
Ensure consistent metric definitions
Simplify BI dashboard creation
Enable self-service analytics
Reduce errors and confusion
From a technical perspective, data modeling involves:
Choosing the correct grain
Defining primary and foreign keys
Managing slowly changing dimensions (SCDs)
Handling nulls and edge cases
Documenting relationships
Poor data modeling leads to slow dashboards, duplicate logic, and conflicting reports. Strong modeling creates trust and scalability.
Modern analytics engineering uses tools like dbt to version, test, and document data models, bringing the software engineering discipline to analytics.




