Chord Analytics
Data Models
12 min
introduction data models, or transformations, are tables created by our analytics engineering team to present data in a shape that is easy to consume for analytics purposes orders facts, product dimensions, and user dimensions are all examples of potential entities (or tables) in a data model creating data models is essential for transforming raw data into structured, meaningful, and usable formats that drive analytics, reporting, and decision making well designed data models also serve as the foundation for powering more robust and useful ai systems by providing clean, consistent inputs that improve model accuracy, reduce noise, and support richer feature engineering in short, data models not only enable better business intelligence but also unlock the full potential of ai by ensuring data is trustworthy, well organized, and aligned with real world context why data models matter 🔧 structure and organization data models define how data is organized , related, and stored they bring consistency and clarity to how teams understand and query data 🤖 unlock full potential of ai with organized, accurate, and business ready context models provide the structured, high quality data that machine learning systems rely on for accuracy and relevance they enable richer feature engineering, reduce noise, and ensure ai outputs are grounded in business ready context 📊 enable reliable analytics models establish clean relationships between metrics and dimensions, reducing errors and ambiguity they ensure that calculations like revenue, retention, or conversion rate are accurate and consistent across tools and users, creating a single source of truth 🚀 power self service & scale with intuitive, documented models (e g , via dbt or semantic layers), non technical users can explore data confidently teams avoid reinventing logic and reduce dependency on data engineers or analysts for every question 🔒 governance and control data models provide defined logic, lineage, and ownership , supporting governance, auditing, and compliance they help enforce data quality checks, contracts, and access controls 📐 improve performance and efficiency well modeled data reduces duplication and streamlines queries optimized joins, indexing, and aggregations improve performance in warehouses and bi tools 🧠 support business understanding models reflect the business logic —how the company defines a customer, transaction, or lifecycle stage they bridge the gap between technical schemas and business needs how it works chord utilizes the kimball approach to data modeling this is a widely used method in analytics and data warehousing that enables teams to build models that are easy to understand and utilize it breaks data into two core types of tables 🧾 fact tables (fct ) – the “what happened” fact tables track events or transactions , such as purchases, logins, or ad clicks they include measurable data such as revenue, quantities, time of action, and user ids 👤 dimension tables (dim ) – the “who” and “what” dimension tables provide context for facts they describe things like customers, products, or regions these two types of tables are designed to join together easily, allowing users to answer questions like “how much revenue did we make from active customers last month?” “which products are most popular among new users?” 🛠️ how chord builds data models step 1 ingest raw data data often starts messy—coming from tools like shopify (orders, line items, products, etc ) chord cdp (website client side events and server side events) klaviyo (messaging) facebook ads (ad spend and metrics) this data is loaded into a cloud data warehouse however, at this point, it remains messy, inconsistent, and contains duplicates, inconsistent naming, or missing values additionally, each source typically contains tens or hundreds of separate tables that must be combined to make sense of the data step 2 clean and normalize the data chord utilizes dbt (short for data build tool ), a transformation tool that helps data teams clean, organize, and document data as it flows from raw sources to clean, usable models chord's analytics engineers build these data models in layers in dbt using programming languages like sql, python, jinja and yaml the various data modelling layers allow us to clean the raw data (e g , standardize names, remove junk values, handle nulls, trims, formatting, or simple derived fields deduplicate records convert data types (e g , string to timestamp) filte r out irrelevant rows (e g , deleted or test records) add flags like is active convert or standardize (e g , currency conversion, unit standardization) join related tables from different systems (e g , linking a cdp session to a shopify order) enrich records with helpful metadata or flags define metrics step 3 organize into facts and dimensions once cleaned, the data is structured into dimension models like dim customers, dim products, dim dates fact models like fct orders, fct pageviews, fct subscriptions step 4 explode! chord utilizes an explosion technique to join every applicable dimension and fact table into a single, consolidated table, such as orders, so that every field applicable to orders is now available in a single, unified table we call these x fct tables now, there's no need for one to many joins between dimension and fact tables, and instead, data is readily joined to immediately start analysis! these data models models are versioned, documented, and tested in dbt to ensure reliability they run daily (or even hourly), so dashboards and ai models always use fresh, trustworthy data getting started chord's data models power your ai, analytics, and activations in the chord platform you'll see data models such as orders, sessions, and more in your explores section of analytics for the full list of models, along with each field therein, see the documentation in chord data attribute definitions docid\ eehjlvljqmt3q98v e87n if you have any questions or need help, please reach out to us at help\@chord co mailto\ help\@chord co