Data Science Fundamentals

Expert-defined terms from the Professional Certificate in Data Science in E-commerce course at UK School of Management. Free to read, free to share, paired with a globally recognised certification pathway.

Data Science Fundamentals

Association Rule Mining #

Association rule mining is a technique used in data mining to discover interesti… #

It involves finding patterns where one set of items tends to appear together in the dataset. For example, in a retail setting, association rule mining can help identify which products are frequently bought together, allowing businesses to make targeted marketing or product placement decisions.

Big Data #

Big data refers to large, complex datasets that are difficult to manage and anal… #

Big data is characterized by its volume, velocity, and variety, as it often includes unstructured data from sources such as social media, sensors, and mobile devices. Data scientists use advanced analytics techniques to extract valuable insights from big data.

Clustering #

Clustering is a data mining technique used to group similar data points together… #

It is often used to uncover hidden patterns or structures in data and can help identify natural groupings within a dataset. For example, clustering can be used to segment customers based on their purchasing behavior or to group documents by topic.

Confusion Matrix #

A confusion matrix is a table that is used to evaluate the performance of a clas… #

It shows the number of true positives, true negatives, false positives, and false negatives predicted by the model. By analyzing the confusion matrix, data scientists can assess the accuracy, precision, recall, and other metrics of a classification algorithm.

Data Cleaning #

Data cleaning is the process of identifying and correcting errors or inconsisten… #

It involves removing duplicate records, handling missing values, correcting data formatting issues, and ensuring data quality. Data cleaning is an essential step in the data preprocessing pipeline to ensure accurate and reliable analysis results.

Data Exploration #

Data exploration is the initial phase of data analysis where data scientists exa… #

During data exploration, data scientists may use descriptive statistics, data visualization techniques, and exploratory data analysis to gain insights into the data before building predictive models.

Data Mining #

Data mining is the process of discovering patterns, relationships, or insights f… #

Data mining can help extract valuable knowledge from data and is used in various applications such as marketing, fraud detection, and healthcare. Common data mining techniques include clustering, classification, and association rule mining.

Data Preprocessing #

Data preprocessing is the initial step in the data analysis pipeline where raw d… #

It involves handling missing values, removing outliers, standardizing data formats, and encoding categorical variables. Data preprocessing is crucial for ensuring the quality and accuracy of data analysis results.

Data Science #

Data science is an interdisciplinary field that combines techniques from statist… #

Data scientists use advanced analytics, machine learning, and data visualization techniques to analyze complex datasets and make data-driven decisions. Data science is widely used in various industries such as e-commerce, healthcare, finance, and marketing.

Data Visualization #

Data visualization is the process of representing data graphically to help data… #

It involves creating charts, graphs, and interactive visualizations that highlight patterns, trends, and relationships in the data. Data visualization is essential for communicating insights and findings effectively to non-technical audiences.

Decision Tree #

A decision tree is a predictive modeling technique that uses a tree #

like structure to represent decisions and their possible consequences. Decision trees are used in classification and regression tasks to make predictions based on input features. Each internal node of the tree represents a decision based on a feature, while each leaf node represents the outcome or prediction.

Dimensionality Reduction #

Dimensionality reduction is a technique used to reduce the number of input featu… #

It helps simplify the data and improve the performance of machine learning algorithms by reducing noise and computational complexity. Common dimensionality reduction techniques include principal component analysis (PCA) and t-distributed stochastic neighbor embedding (t-SNE).

Ensemble Learning #

Ensemble learning is a machine learning technique that combines multiple models… #

It involves training multiple base models on different subsets of data and aggregating their predictions to make a final decision. Ensemble learning methods such as random forests, boosting, and bagging are commonly used to increase the accuracy and robustness of machine learning models.

Evaluation Metrics #

Evaluation metrics are measures used to assess the performance of a machine lear… #

They quantify how well the model predicts outcomes and help data scientists compare different models and algorithms. Common evaluation metrics include accuracy, precision, recall, F1 score, and area under the curve (AUC).

Feature Engineering #

Feature engineering is the process of creating new features or transforming exis… #

It involves selecting relevant features, encoding categorical variables, scaling numerical features, and creating interaction terms. Feature engineering is a critical step in building accurate and robust predictive models.

Hyperparameter Tuning #

Hyperparameter tuning is the process of selecting the optimal hyperparameters fo… #

Hyperparameters are parameters that control the learning process of the model, such as the learning rate, regularization strength, and tree depth. Data scientists use techniques like grid search, random search, and Bayesian optimization to find the best hyperparameters.

Machine Learning #

Machine learning is a subset of artificial intelligence that enables computers t… #

Machine learning algorithms learn patterns and relationships from data to make accurate predictions or classifications. Common machine learning techniques include supervised learning, unsupervised learning, and reinforcement learning.

Model Selection #

Model selection is the process of choosing the best machine learning algorithm f… #

Data scientists evaluate multiple models using cross-validation or holdout validation and select the one with the highest accuracy or other evaluation metrics. Model selection is crucial for building accurate predictive models.

Overfitting and Underfitting #

Overfitting and underfitting are common problems in machine learning where a mod… #

Overfitting occurs when a model is too complex and captures noise in the training data, leading to poor performance on unseen data. Underfitting, on the other hand, occurs when a model is too simple and fails to capture the underlying patterns in the data.

Principal Component Analysis (PCA) #

Principal component analysis (PCA) is a dimensionality reduction technique used… #

PCA identifies the principal components that explain the most variance in the data and projects the data onto these components. PCA is commonly used for visualization, noise reduction, and feature extraction.

Regression Analysis #

Regression analysis is a statistical technique used to model the relationship be… #

It is used to predict continuous outcomes and understand the impact of predictor variables on the target variable. Common regression models include linear regression, logistic regression, and polynomial regression.

Resampling Techniques #

Resampling techniques are methods used to estimate the performance of a machine… #

Resampling helps data scientists assess the stability and variability of model predictions and avoid overfitting. Common resampling techniques include cross-validation, bootstrapping, and leave-one-out validation.

Support Vector Machine (SVM) #

Support vector machine (SVM) is a supervised learning algorithm used for classif… #

SVM finds the optimal hyperplane that separates different classes in the feature space by maximizing the margin between classes. SVM is effective for high-dimensional data and can handle non-linear relationships using kernel functions.

Time Series Analysis #

Time series analysis is a statistical technique used to analyze and forecast tim… #

It involves identifying patterns, trends, and seasonality in time series data to make predictions about future values. Time series analysis is commonly used in finance, sales forecasting, and anomaly detection to understand and predict temporal patterns.

Unsupervised Learning #

Unsupervised learning is a machine learning technique used to find patterns and… #

It involves clustering similar data points together or reducing the dimensionality of the data to uncover hidden structures. Unsupervised learning is used in applications such as customer segmentation, anomaly detection, and recommendation systems.

Validation Set #

A validation set is a subset of data used to evaluate the performance of a machi… #

It is separate from the training set and is used to tune hyperparameters, optimize the model, and prevent overfitting. The validation set helps data scientists assess the generalization ability of the model on unseen data.

Variance and Bias #

Variance and bias are two sources of error in machine learning models that affec… #

Variance measures the model's sensitivity to fluctuations in the training data, while bias measures the model's tendency to make incorrect assumptions about the data. Balancing bias and variance is crucial for building accurate and robust predictive models.

Web Scraping #

Web scraping is the process of extracting data from websites using automated scr… #

It involves accessing the HTML of web pages, parsing the content, and extracting relevant information such as text, images, or links. Web scraping is commonly used in e-commerce to gather product information, prices, and reviews from online stores for analysis.

XGBoost #

XGBoost is an open #

source machine learning library that is widely used for gradient boosting algorithms. XGBoost stands for "eXtreme Gradient Boosting" and is known for its speed, performance, and scalability. XGBoost is used in various applications such as classification, regression, ranking, and recommendation systems to achieve high accuracy and efficiency.

May 2026 cohort · 29 days left
from £99 GBP
Enrol