Theta Glossary

Talk About CLV and CBCV

CLV and CBCV draw from the worlds of data science, analytics, finance and marketing, all of which have a language of their own. This glossary contains some of the common terms you may encounter in discussions about CLV and CBCV.

AJAX progress indicator
Search:
  • A mathematical formula or statistical process used to analyze data.
  • The maximum probability of making a Type I error. This probability is established by the experimenter and often set at 5%.
  • Statement of a change or difference; assumed to be true if the null hypothesis is rejected.
  • The systematic analysis of data. Analytics applications and toolkits contain mathematical algorithms and computational engines that can manipulate large datasets to uncover patterns, trends, relationships, and other intelligence that allow users to ask questions and gain useful insights about their business, operations, and markets.  
  • The process of identifying rare or unexpected items or events in a dataset that do not conform to other items in the dataset and do not match a projected pattern or expected behavior. Anomalies are also called outliers, exceptions, surprises or contaminants and they often provide critical and actionable information.
  • Making data anonymous; severing of links between people in a database and their records to prevent the discovery of the source of the records.
  • One-way ANOVA is a generalization of the 2-sample t-test, used to compare the means of more than two samples to each other.
  • The ANOVA table is the standard method of organizing the many calculations necessary for conducting an analysis of variance.
  • The apparent ability of a machine to apply information gained from previous experience accurately to new situations in a way that a human would.
  • The apparent ability of a machine to apply information gained from previous experience accurately to new situations in a way that a human would.
  • Batch data processing is an efficient way of processing high volumes of data where a group of transactions is collected over a period of time.
  • A theorem based on conditional probabilities. It uses relevant evidence, also known as conditional probability, to determine the probability of an event, based on prior knowledge of conditions that might be related to the event.
  • Using data about people’s behavior to understand intent and predict future actions.
  • Behavioral data provides information about a customer’s interaction with your business.
  • Creates more targeted and personalized offers to customers and prospective customers based on the knowledge of actions they previously performed.
  • A family of continuous probability distributions set on the interval [0, 1] having two positive parameters denoted by alpha (α) and beta (β). These parameters appear as exponents of the random variable and control the shape of the distribution.
  • Models the dropout process as a geometric distribution with a beta mixing distribution, and models the purchase frequency process as a negative binomial distribution.
  • The risk or probability of making a Type II error.
  • Extremely large datasets of structured, unstructured, and semi-structured data that are often characterized by the five Vs: volume of data collected, variety of data types, velocity at which the data is generated, veracity of the data, and value of it. 
  • Y variable takes on one of two outcomes (levels), e.g. pass/fail, agree/disagree.
  • The general term used for the identification, extraction and analysis of data.
  • The general term used for the identification, extraction and analysis of data.
  • The process of determining the economic value of a business; used to determine the fair value of a business for a variety of reasons, such as sale value and mergers and acquisition.
  • A model that helps explain the buying patterns of non-contractual customers by describing the rate at which customers make purchases and the rate at which they drop out or “die.” BTYD models all jointly model two processes: (1) a repeat purchase process, which explains how frequently customers make purchases while they are still "alive"; and (2) a dropout process, which models how likely a customer is to churn in any given time period. Common versions of the BTYD model include the Pareto/NBD model and the Beta-Geometric/NBD model.
  • A model that helps explain the buying patterns of non-contractual customers by describing the rate at which customers make purchases and the rate at which they drop out or “die.” BTYD models all jointly model two processes: (1) a repeat purchase process, which explains how frequently customers make purchases while they are still "alive"; and (2) a dropout process, which models how likely a customer is to churn in any given time period. Common versions of the BTYD model include the Pareto/NBD model and the Beta-Geometric/NBD model.
  • The series of actions and interactions that a consumer performs before, during, and after making a commercial transaction.
  • The rate at which customers stop purchasing or doing business with your company.  It is sometimes called customer attrition, customer turnover or customer defection. To calculate the rate of customer churn, divide the number of churned customers (over a given period) by the initial number of total customers: Churn rate = Churned Customers / Total Initial Customers
  • A systematic process for obtaining important and relevant information about data and assigning data to a particular group or class.
  • The analysis of web activity by users through the items they click on a page.
  • An aggregation of granular patient-centric health data usually collected from multiple-source IT systems and intended to support multiple uses. When a CDR holds data specifically organized for analytics it meets the definition of a clinical data warehouse.
  • An aggregation of granular patient-centric health data usually collected from multiple-source IT systems and intended to support multiple uses. When a CDR holds data specifically organized for analytics it meets the definition of a clinical data warehouse.
  • A Voice of the Customer (VOC) term that refers to the act of following up with customers who have provided either extremely positive or negative feedback.
  • The  process of acting on direct or indirect customer input collected from sources such as customer surveys, contact center interactions, and social media comments.
  • The process of identifying objects that are similar to each other and clustering them in order to understand the differences as well as the similarities within the data.
  • CLV to CAC is the ratio of Customer Lifetime Value to Customer Acquisition Cost. This metric helps you optimize your marketing campaigns to acquire profitable customers.
  • A standardized measure of dispersion of a probability distribution or frequency distribution. 
  • A group of customers that shares some relevant acquisition characteristic(s). In virtually every case, cohorts are defined primarily by the time period of acquisition. A cohort definition can also include other acquisition-related characteristics, such as channel, promotional campaign, product purchased, etc.
  • Standardized, precisely defined questions paired with a set of specific allowable responses, used systematically across different sites, studies, or clinical trials to ensure consistent data collection. CDEs may consist of a single data element, such as height, gender, or date of birth, or a collection of connected questions, such as a survey instrument used as a depression index or a quality-of-life scale.
  • Standardized, precisely defined questions paired with a set of specific allowable responses, used systematically across different sites, studies, or clinical trials to ensure consistent data collection. CDEs may consist of a single data element, such as height, gender, or date of birth, or a collection of connected questions, such as a survey instrument used as a depression index or a quality-of-life scale.
  • Data analysis that compares two or more data sets or processes to detect patterns within very large data sets.
  • A range of values which is likely to contain the population parameter of interest with a given level of confidence.
  • Data from a measurement scale that can be divided into finer and finer increments such as temperature, time or pressure.
  • A means to determine a statistical relationship between variables, often for the purpose of identifying predictive factors among the variables. A technique for quantifying the strength of the linear relationship between two variables.
  • A variable that may affect the result of what is being studied. 
  • The total cost of acquiring a new customer.  To calculate average CAC, total up all of the costs that go into your customer acquisition tactics for a given period. Then divide by the number of customers acquired during that period. CAC can also be calculated for individual customers.
  • The total cost of acquiring a new customer.  To calculate average CAC, total up all of the costs that go into your customer acquisition tactics for a given period. Then divide by the number of customers acquired during that period. CAC can also be calculated for individual customers.
  • A systematic review of the buying behavior of a firm’s customers using data captured by its transaction systems. It provides an understanding of how customers differ in their buying behavior and how their buying behavior evolves over time.
  • A process that employs various CLV models and customer behavior data to assess the lifetime value of a customer or cohort and the overall health of a company’s customer base. (See also Predictive Customer Base Analysis.)
  • A method that uses customer metrics to assess a firm’s underlying value; exploits basic accounting principles to make revenue projections from the bottom up instead of from the top down.
  • A method that uses customer metrics to assess a firm’s underlying value; exploits basic accounting principles to make revenue projections from the bottom up instead of from the top down.
  • Identifying your most valuable customers, then focusing most of your firm’s efforts on them and finding more customers like them.
  • Refers to the level of interaction between your brand and your customers across all channels over the course of your entire relationship. It can refer to what customers do to interact with your brand, as well as, how loyal they are towards your brand.
  • The customer’s perception of their interactions with your business, including all the actions, messages, and engagement you perform across all customer touchpoints.
  • The customer’s perception of their interactions with your business, including all the actions, messages, and engagement you perform across all customer touchpoints.
  • The full-cycle relationship between a business and its customers, from brand awareness to sales.
  • The various stages a consumer goes through before, during and after they complete a transaction. It can also refer to the phases a customer passes through during the course of an ongoing relationship with a brand.
  • Estimates the total amount of money a company receives from a customer during the entire engagement time before the customer churns; the net present value of ALL variable profits/costs (including customer acquisition costs).
  • Estimates the total amount of money a company receives from a customer during the entire engagement time before the customer churns; the net present value of ALL variable profits/costs (including customer acquisition costs).
  • The measure of a customer’s likelihood to do repeat business with a company or brand.
  • A measure of an organization’s ability to keep customers.
  • The process of segmenting individuals into groups based on common characteristics and attributes.
  • Information assets that organizations collect, process and store during regular business activities, but generally fail to use for other purposes such as analytics.
  • A graphical representation of analyses performed by algorithms.
  • The process of collecting data from multiple sources for the purpose of reporting or analysis.
  • A person responsible for the tasks of modeling, preparing and cleaning data for the purpose of deriving actionable information from it.
  • The overall design for the structure, policies, and rules that define an organization’s data and how it will be used and managed. 
  • The process of removing or correcting errors from a dataset, table, or database. Also called data scrubbing, the process finds duplicate data and other inconsistencies, like typos and numerical sets that don’t add up.  
  • The total amount of storage space across all storage systems required to accommodate an organization’s full collection of data. Also known as Storage Footprint.
  • A framework of people, policies, processes and technologies that define how you manage your organization’s data. 
  • The ongoing processes to ensure data is free of errors and in a format that enables it to be used for a variety of purposes.
  • The process of locating data – structured, unstructured or semi-structured ─ and specifying its key qualities that identify what it is.
  • A process for organizing data to maximize a query’s efficiency while searching.
  • The process of combining data from different sources into a single, unified view. Integration begins with the ingestion process, and includes steps such as cleansing, ETL (extract, transform, load) mapping, and transformation.
  • The overall accuracy, completeness, and consistency of data. Data integrity also refers to how well the data complies to regulatory requirements and security standards.  
  • A vast pool of raw data — unstructured, semi-structured, or structured — in an organization whose purpose is not yet defined, either at the company or department level. There are no limits on data sizes, files, sources, or structure; data is kept in its original format and machine-to-machine data logs flow in real time.
  • The act of extracting useful information from large datasets. Data mining is often done by business users employing analytics tools to uncover patterns, trends, anomalies, relationships, dependencies, and other useful intelligence. 
  • The process of taking siloed data from multiple data storage locations, combining and organizing it, and making it available for data analysis tools.
  • A set of tools and processes used to automate the movement and transformation of data between a source system and a target repository.
  • A task that prepares data for analytics. It involves combining data from various sources, then cleansing and transforming it. If done via a self-service interface, business users can access and manipulate the data they need with minimal training – and without asking IT for help.
  • The policies and practices for handling data in ways that protect it from unauthorized access or disclosure.  
  • A place to hold data, make data available for use, and organize data in a logical manner. Data repositories may have specific requirements concerning subject or research domain; data re-use and access; file format and data structure; and the types of metadata that can be used.
  • Storing of information for a specified period. Data retention is primarily relevant to businesses that store data to service their customers and comply with government or industry regulations.
  • A discipline that incorporates statistics, data visualization, computer programming, data mining, machine learning and database engineering to solve complex problems.
  • A person employed to analyze and interpret complex data, especially for assisting a business in its decision-making.
  • The tools, technologies, tactics and other methods used to keep data safe from harm, alteration, unauthorized access  or exposure, disaster, or system failure, and, at the same time, readily accessible to legitimate users and applications.
  • The ability and practice of distributing the same sets of data resources with multiple users or applications while maintaining data accuracy across all entities consuming the data.
  • The existence of data that is maintained and used by one group or department, but which is not easily or fully accessible by other groups in the same organization.
  • Covers the implementation and maintenance of the physical hardware or cloud-based infrastructure you use to collect, store, and manage your data, such as servers, data management platforms, data warehouses, and data lakes.
  • A central repository for structured, filtered data in an organization or department that has already been processed for a specific purpose (is currently in use). Data is stored in files and folders that help organize the data, provide a multidimensional view, and support strategic decision-making.
  • The process of taking raw data and transforming it into a format that is compatible with established databases and applications. The process may include structuring, cleaning, enriching, and validating data as necessary to make raw data useful.
  • A facility for organizing, storing, managing, safeguarding, and controlling access to data. Common types of databases include relational database management systems (RDBMS), in-memory databases, object-oriented databases (OODBMS), NoSQL databases, and SQL databases – each with its own advantages.
  • Data relating to the characteristics of a human population.
  • Condensing big numbers into smaller pieces of information. This is similar to summarizing the data story. Rather than listing every single number and detail, there is a general theme and narrative.
  • Reviewing past performance to determine what happened and why. Businesses use this type of analytics to complete root cause analysis.
  • A sales strategy where manufacturers and CPG (consumer packaged goods) brands sell their products directly to their customers instead of selling them through retailers and wholesalers.
  • A sales strategy where manufacturers and CPG (consumer packaged goods) brands sell their products directly to their customers instead of selling them through retailers and wholesalers.
  • Data which is not measured on a continuous scale. Examples are binomial (pass/fail), Counts per unit, Ordinal (small/medium/large), and Nominal (red/green/blue). Also known as attribute or categorical data.
  • A statistical analysis technique used to predict cluster membership from labeled data.
  • Reasonable steps taken to satisfy a legal requirement, especially in buying or selling something; the process a potential investor engages in to assess the desirability, value, financial viability and potential of an investment before committing capital.
  • An equation derived from the data that expresses a relationship between the inputs and an output (Y=f(x)).
  • A set of outcomes of an experiment to which a probability is assigned.
  • Data that is used as evidence to support a specific belief or proposition. It can be analyzed, presented, converted, etc. to validate, prove, or disprove a specific position.
  • A plan for a venture capitalist or private equity firm to get out of financial investment or sell off tangible business assets once certain predetermined conditions have been met. 
  • An approach to data analysis focused on identifying general patterns in data, including outliers and features of the data that are not anticipated by the experimenter’s current knowledge or preconceptions.
  • The processes a data pipeline uses to replicate data from a source system into a target system such as a data warehouse. It’s a modern variation of the older process of extract, transform, and load (ETL), in which transformations take place before the data is loaded
  • The processes a data pipeline uses to replicate data from a source system into a target system such as a data warehouse. It’s a modern variation of the older process of extract, transform, and load (ETL), in which transformations take place before the data is loaded
  • A three-phase process in which data is extracted, transformed and loaded into an output data container.  
  • A three-phase process in which data is extracted, transformed and loaded into an output data container.  
  • A hypothesis test for comparing variances.
  • A set period of time for estimating the value of a customer.
  • The average outcome predicted by a model. (See also overfitting and underfitting.)
  • A family of continuous probability distributions defined in terms of two positive parameters, a scale parameter denoted by k and a shape parameter denoted by theta (θ). The exponential distribution, erlang distribution, and the chi-square distribution are all special cases of the gamma distribution.
  • Distributions whose tails aren’t exponentially bounded. Unlike the bell curve with a normal distribution, heavy-tailed distributions approach zero at a slower rate and can have outliers with very high values.
  • Involving or serving as an aid to learning, discovery, or problem-solving by experimental and trial-and-error methods.
  • Representation of frequency of values by intervals.
  • The sum of all the transactions made by a customer in the past. It can be calculated using this formula: Historic CLV = (Transaction 1+ Transaction 2 + Transaction 3….+ Transaction N) x AGM Where N is the latest transaction and AGM is the Average Gross Margin (your profit margin after considering the cost of producing it).
  • The ideal customer that your business can sell to; a fictitious persona or company profile whose pain points precisely match the ones that your product seeks to solve. Also known as an Ideal Buyer Profile.
  • The ideal customer that your business can sell to; a fictitious persona or company profile whose pain points precisely match the ones that your product seeks to solve. Also known as an Ideal Buyer Profile.
  • A metric used to periodically track and evaluate performance toward the achievement of a specific objective or target.
  • A metric used to periodically track and evaluate performance toward the achievement of a specific objective or target.
  • A type of regression analysis used to analyze the direct association between a dependent variable that must be continuous and one or more independent variable(s) that can be any level of measurement, nominal, ordinal, interval, or ratio. A linear regression tests the changes in the mean of the dependent variable by the predictors included in our model, the independent variable(s).
  • Investigates the relationship between response (Ys) and one or more predictors (Xs) where Ys are categorical, not continuous and Xs can be either continuous or categorical. Types of logistic regression are:Binary logistic regression: The response variable can only belong to one of two categories.Multinomial logistic regression: The response variable can belong to one of three or more categories and there is no natural ordering among the categories.Ordinal logistic regression: The response variable can belong to one of three or more categories; there is a natural ordering among the categories.
  • A group of people on a social network that are similar to a group of your existing customers, based on factors like demographics, behavior, location, etc.
  • The lifetime spend of customers in aggregate. LTV is an aggregate metric, unlike CLV, which is calculated at the individual-customer level.
  • The lifetime spend of customers in aggregate. LTV is an aggregate metric, unlike CLV, which is calculated at the individual-customer level.
  • A type of artificial intelligence that provides computers the ability to learn from new data without being explicitly programmed. The computer can identify patterns of behavior that a human would likely not see.
  • A type of artificial intelligence that provides computers the ability to learn from new data without being explicitly programmed. The computer can identify patterns of behavior that a human would likely not see.
  • The process of associating specific marketing activities to customer sales and conversions.
  • The average of a data set.
  • The middle of the set of numbers.
  • A probability density function that is parameterized by two or more density functions and mixture parameters.
  • The most common number in a data set.
  • A classification technique based on Bayes Theorem with an assumption of independence among predictors. In simple terms, it assumes that the presence of a particular feature in a class is unrelated to the presence of any other feature.
  • A negative binomial distribution (also called the Pascal Distribution) is a discrete probability distribution for random variables in a negative binomial experiment.
  • The industry standard measure of customer experience (CX) and customer loyalty.
  • The industry standard measure of customer experience (CX) and customer loyalty.
  • Models the relationship between a set of predictors and a nominal response variable. A nominal response has at least three groups that don’t have a natural order.
  • A variable that has two or more categories, but there is no intrinsic ordering to the categories. Also known as Categorical Variable.
  • A probability distribution that is symmetric about the mean, showing that data near the mean are more frequent in occurrence than data far from the mean; the normal distribution appears as a bell curve.
  • Developed for handling unstructured data that SQL cannot support because of the lack of structure.
  • Statement of no change or difference; assumed to be true until sufficient evidence is presented to reject it.
  • Statement of no change or difference; assumed to be true until sufficient evidence is presented to reject it.
  • Data storage that manages data as objects, as opposed to other storage architectures like file systems that manage data as a file hierarchy, and block storage which manages data as blocks within sectors and tracks. Also known as S3 or S3-compatible storage. S3 is from Amazon Simple Storage Service, which is the most well-known object storage system. S3-compatible storage means that the storage employs the S3 API as its “language.”
  • The seamless integration of online and offline marketing channels a company uses to interact with customers.
  • Statistical test to compare the mean of one sample of data to a target. Uses the t-distribution.
  • The process of finding optimal problem parameters subject to constraints. Optimization algorithms heuristically test a large number of parameters. configurations in order to find an optimal result, determined by a characteristic function (also called a fitness function).
  • A statistical analysis method that can be used to model the relationship between an ordinal response variable and one or more explanatory variables. 
  • Similar to a nominal variable but has a clear ordering of the categories.
  • An object that deviates significantly from the general average within a dataset or a combination of data; numerically distant from the rest of the data and therefore indicates something unusual and generally requires additional analysis.
  • Producing an analysis that corresponds too closely or exactly to a particular set of data, so that it may fail to reliably fit to additional data or predict future observations.
  • A test used to compare the average difference between two samples of data that are linked in pairs. Special case of the 1-sample t-test. Uses the t-distribution.
  • Skewed distribution that has a ski slope shape — right skewed where mode and minimum are equal; has the longest tail of all probability distributions. It’s used to model any variable that has a minimum value and for which the probability density decreases geometrically towards zero.
  • A variation of the Pareto/NBD model that accounts for some level of regularity for inter-transaction times.
  • Models the dropout process as a Pareto Type II distribution and the purchase frequency process as a negative binomial distribution.
  • A heavy-tail probability distribution used in business, economics, actuarial science, queueing theory and Internet traffic modeling; essentially a Pareto distribution that has been shifted so that its support begins at zero. Also known as a Lomax Distribution.
  • Identifying patterns in data via algorithms to make predictions about new data coming from the same source.
  • A representation of a group of leads or customers with similar characteristics, features, and/or behaviors.
  • The  process of tailoring communications or customer experiences taking into account the unique aspects of each individual.
  • Information (data) that can be used to verify a person's identity.
  • Information (data) that can be used to verify a person's identity.
  • A discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space if these events occur with a known constant mean rate and independently of the time since the last event.
  • The net present value of variable profits, ignoring customer acquisition costs (CAC).
  • The net present value of variable profits, ignoring customer acquisition costs (CAC).
  • The ability of a statistical test to detect a real difference when there is one; the probability of correctly rejecting the null hypothesis. Determined by alpha and sample size.
  • A form of advanced analytics which examines data or content to answer the question “What is going to happen?” or more precisely, “What is likely to happen?,” and is characterized by techniques such as regression analysis, forecasting, multivariate statistics, pattern matching, predictive modeling, and forecasting.
  • A predictive analytics process that employs various customer lifetime value (CLV) models and customer behavior data to assess CLV and the overall health of a company’s customer base, as well as predict the future behaviors of a customer or cohort.  
  • The process of developing a model that will most likely predict a trend or outcome.
  • Builds on predictive analytics by including actions to make data-driven decisions by looking at the impacts of various actions.
  • A statistical approach that uses the effect of random occurrences or actions to forecast the possibility of future results.  
  • A statistical function that describes all the possible values and likelihoods that a random variable can take within a given range. Probability distributions may be discrete or continuous.
  • The practice of bringing new products to market and finding customers that want to buy them (as opposed to customer centricity.)
  • Psychographics can be described as the study of consumers based on their ideas, actions, and interests.
  • The series of actions and interactions that a consumer performs before, during, and after making a commercial transaction.
  • A type of data that cannot be counted, measured or easily expressed using numbers, mostly reflected textually.
  • Information that can be counted or measured, in other words numerically recognized and analyzed.
  • A request input by a user and executed by a database management system (DBMS). It can be a request for data results from a database, action on the data or both.
  • Difference between the largest and smallest measurement in a data set.
  • Data that is created, processed, stored, analyzed and visualized within milliseconds.
  • A simple method to determine customer value:Recency: How long ago did a customer purchase?Frequency: How often/consistently does a customer purchase?Monetary: How much does a customer spend on average?
  • A simple method to determine customer value:Recency: How long ago did a customer purchase?Frequency: How often/consistently does a customer purchase?Monetary: How much does a customer spend on average?
  • Statistical processes for estimating the relationships between a dependent variable  and one or more independent variables; can be used to explain the past and predict future events; the most common form is linear regression. 
  • The software used to store, manage, query, and retrieve data stored in a relational database, where the data points are related to one another. A RDBMS is an advanced version of a database management system where data is typically stored in the form of tables rather than files.
  • The software used to store, manage, query, and retrieve data stored in a relational database, where the data points are related to one another. A RDBMS is an advanced version of a database management system where data is typically stored in the form of tables rather than files.
  • The difference between reality (an actual measurement) and the fit (model output).
  • The amount of additional value we expect to collect from a customer over a specific and defined time period.
  • The amount of additional value we expect to collect from a customer over a specific and defined time period.
  • The ability of a system or process to maintain acceptable performance levels as workload or scope increases.
  • Information that doesn’t reside in a relational database but that has some organizational properties that make it easier to analyze. With some processes, it may be possible to store it in a relational database. For example, an email or a text file.
  • The application of statistical functions and probability theory to comments people make on the web or social networks to determine how they feel about a product, service or company.
  • A complete set of the data an organization has about a customer, which can be accessed in a single location.
  • A complete set of the data an organization has about a customer, which can be accessed in a single location.
  • Compares the variance of one sample of data to a target. Uses the Chi-square distribution.
  • A software licensing and delivery model in which a software provider delivers an application to users on the internet via a website or app. Unlike traditional software products, SaaS software is licensed on a subscription basis and is centrally hosted. 
  • A software licensing and delivery model in which a software provider delivers an application to users on the internet via a website or app. Unlike traditional software products, SaaS software is licensed on a subscription basis and is centrally hosted. 
  • A relational database that stores data in tables and rows. Data items (rows) are linked based on common data items to enable efficiency, avoid redundancy, and facilitate easy, flexible retrieval.
  • A model for estimating probability distributions of potential outcomes by allowing for random variation in one or more inputs over time.
  • Data that can be neatly formatted into rows and columns and mapped to predefined fields. It’s typically stored in Excel spreadsheets or relational databases.
  • A way of describing a set of observations where most observations fall close to the mean, and the rest of the observations make up the tails on either side; used for smaller sample sizes, where the variance in the data is unknown.
  • A statistical test that compares the means of two samples.
  • Compares the variance of two samples of data against each other. Uses the F distribution.
  • Analysis of well-defined data measured at repeated measures of time to identify time-based patterns.
  • A specific way of analyzing a sequence of data points collected over an interval of time; data points are recorded at consistent intervals over a set period of time rather than just recording the data points intermittently or randomly.
  • A record of all transaction activity.
  • Data that relates to the conducting of business, such as accounts payable and receivable data or product shipments data.
  • A statistical test to compare the means of two samples of data against each other. Uses the t-distribution.
  • The error that occurs when the null hypothesis is rejected when, in fact, it is true.
  • The error that occurs when the null hypothesis is not rejected when it is, in fact, false.
  • Describe a data model that can’t interpret the correlation between input and output variables; most often occurs when there is insufficient data or the wrong type of data for the task at hand.
  • Data that can’t be organized into rows and columns or processed and analyzed via conventional tools and methods. It is stored in its native format, and is often qualitative rather than quantitative, such as documents, web pages, email, social media content, mobile data, images, audio, video, and more.
  • The process of determining the worth of an asset or company.
  • The average squared deviation for all values from the mean. 
  • The process of collecting and analyzing customer input to determine their needs, preferences, and expectations about a brand’s products, services, and experiences.
  • The process of collecting and analyzing customer input to determine their needs, preferences, and expectations about a brand’s products, services, and experiences.

If you have questions about these terms or any other, Theta has the answers. Submit your questions or contact us to discuss how we can put a CBCV to work for you.

NEWSLETTER SIGN-UP

Sign up for the Centricity Newsletter

If you’re not already getting exclusive Centricity content delivered to your inbox, sign up to do so now. Check out our most recent Centricity email newsletter to preview the exclusive content you get when you sign up for Centricity.

Do you have any suggestions for content you’d like to see featured in the monthly Centricity newsletter? Let us know. Use the Ask Theta form.