Supervised Learning
Supervised learning is a core component of artificial intelligence and machine learning that empowers computers to make predictions by learning from labeled datasets. It plays a pivotal role in automating decisions across sectors, from healthcare diagnostics to financial forecasting. By training algorithms to recognize patterns in historical data, supervised learning complements areas like data science and analytics and supports real-world applications in computer vision and natural language processing (NLP).
In computer vision, for instance, supervised models help identify objects in images with precision. These capabilities extend to autonomous agents such as robotics and autonomous systems, where visual data must be interpreted in real time. With the support of cloud computing, supervised learning can be scaled to process vast image datasets and deliver insights across platforms that utilize different cloud deployment models.
As part of the broader STEM domain, supervised learning connects with numerous emerging technologies. It supports the control systems in smart manufacturing and Industry 4.0 and enables optimized data flow between interconnected devices in the Internet of Things (IoT) and smart technologies. Supervised models also drive advancements in space exploration technologies and the development of intelligent onboard systems in satellite technology.
Supervised learning operates alongside other paradigms, such as unsupervised learning and reinforcement learning, to form a diverse machine learning landscape. While supervised learning excels in structured prediction tasks, other approaches handle unlabeled data or sequential decision-making. Together, they power modern applications such as expert systems that mimic human reasoning.
At the algorithmic level, supervised learning benefits from innovations in deep learning, where multilayered neural networks learn complex representations from annotated inputs. These architectures may one day integrate with quantum algorithms emerging from fields like quantum computing, including concepts such as qubits, quantum superposition, and quantum gates and circuits. As these innovations converge, supervised learning could transform further, opening new pathways in optimization and inference.
The applications of supervised learning extend even into domains like internet and web technologies, where it powers search engines, spam filters, and recommendation engines. Within information technology, it underpins cybersecurity systems that identify anomalies and prevent intrusions. As future engineers and data scientists explore this field, supervised learning remains a foundational tool—essential for understanding intelligent systems and developing real-world solutions that learn from experience.
Table of Contents
Core Concepts of Supervised Learning
Training with Labeled Data:
The hallmark of supervised learning is the use of data that includes both features and corresponding target labels. For instance, in a dataset of emails, each email might be labeled as either “spam” or “not spam.” By exposing the model to numerous examples, it learns to distinguish subtle patterns that correlate specific features (such as certain keywords, sender addresses, or links) with these labels. Over time, the model refines its decision boundaries, improving its accuracy in predicting unseen cases.
Generalization to New Data:
One of the primary goals of supervised learning is to generalize beyond the training set. A well-trained model can handle data it has never encountered before, making it valuable for real-world applications. The model’s ability to generalize depends on factors like the quantity and diversity of training data, choice of algorithm, and the complexity of the model itself. Balancing these factors helps avoid overfitting (memorizing the training data too closely) or underfitting (failing to learn meaningful patterns).
Iterative Training Process:
Most supervised learning methods involve an iterative training cycle. Initially, the model starts with random or default parameters. As it processes the training data, it makes predictions and compares them to the known labels. By calculating an error measure (e.g., mean squared error for regression problems or accuracy for classification), it adjusts its parameters to reduce this error. Repeating this process many times allows the model to converge on an optimal set of parameters that produce accurate predictions.
Model Evaluation and Validation:
To assess a supervised learning model’s performance, the data is often split into training, validation, and test subsets. The training set teaches the model, the validation set helps tune hyperparameters and prevent overfitting, and the test set serves as a final, unbiased measure of how well the model generalizes. By carefully evaluating results on these sets, developers ensure the model is both accurate and reliable when deployed.
Common Techniques in Supervised Learning
Classification:
Classification tasks assign discrete labels to input data. Examples include:
Spam Detection:
Given an email’s content and metadata, a model learns to categorize emails as spam or not spam. After training on labeled examples, it can then accurately classify new incoming emails, alerting users or filtering out unwanted content.
Image Recognition:
In image classification, the model receives images along with labels indicating what each image depicts (e.g., “cat,” “dog,” or “car”). With enough labeled examples, the model can learn to recognize objects, animals, or scenes in new images.
Classification techniques often involve algorithms such as logistic regression, decision trees, random forests, support vector machines (SVMs), and neural networks. Evaluating a classifier’s performance involves metrics like accuracy, precision, recall, and F1-score, each offering unique insights into the model’s strengths and weaknesses.
Regression:
Regression tasks involve predicting continuous numerical values. Examples include:
Predicting Prices:
A model might learn to estimate housing prices based on features such as location, square footage, number of bedrooms, and recent sales data. Once trained, it can forecast the price of a home that the algorithm has never encountered before.
Forecasting Sales:
By examining historical sales data, seasonal trends, and economic indicators, a regression model can anticipate future sales volumes, aiding companies in inventory management and strategic planning.
Common regression algorithms include linear regression, ridge regression, lasso regression, and gradient boosting methods. Metrics like mean squared error (MSE), mean absolute error (MAE), and R² (coefficient of determination) help evaluate how closely the model’s predictions match the actual continuous values.
Practical Considerations:
Feature Selection and Engineering:
Choosing the right features is critical. Irrelevant or redundant features can confuse the model, while well-crafted features can significantly improve accuracy. Students and practitioners often experiment with feature scaling, dimensionality reduction, or domain-specific transformations to enhance model performance.
Handling Imbalanced Classes and Missing Data:
Real-world datasets are rarely perfect. Models must handle issues such as imbalanced classes (e.g., far fewer “spam” emails than “not spam”), missing values, or noisy data. Techniques like oversampling minority classes, imputing missing data, or cleaning outliers help create more robust models.
Ethical and Fairness Considerations:
Since supervised learning models rely on historical data, they can inadvertently perpetuate biases if the training data is not carefully vetted. Ensuring diverse and representative training sets, as well as applying fairness metrics and bias mitigation strategies, is essential for building ethical machine learning solutions.
Beyond the Basics
As students and developers gain confidence with supervised learning, they can explore more advanced topics such as ensemble methods (e.g., combining multiple models to improve performance), transfer learning (reusing models trained on related tasks), and active learning (efficiently querying labels for new data to reduce labeling costs).
Why Study Supervised Learning
Building the Foundation for Predictive Artificial Intelligence
Learning Key Algorithms and Their Applications
Developing Practical Skills in Data Preparation and Model Evaluation
Understanding Limitations and Ethical Considerations
Preparing for Further Study and Diverse Career Paths
Supervised Learning: Conclusion
In essence, supervised learning teaches models to replicate human-driven labeling decisions by finding patterns in historical examples. As one of the most widely used approaches in machine learning, it underpins countless applications—from fraud detection and disease diagnosis to personalized recommendations and financial forecasting—making it a foundational skill for those pursuing advanced data science and AI endeavors.
Supervised Learning: Review Questions and Answers:
1. What is supervised learning and how does it differ from other types of machine learning?
Answer: Supervised learning is a machine learning approach where models are trained using labeled datasets, meaning that each training example is paired with an output label. This method contrasts with unsupervised learning, where no explicit labels are provided, and reinforcement learning, where models learn through interactions with an environment. Supervised learning focuses on mapping input data to known outputs by minimizing the error between predictions and actual values. It is widely used for tasks such as regression and classification, making it a foundational concept in data-driven applications.
2. How does the training process work in supervised learning?
Answer: In supervised learning, the training process begins with a labeled dataset that is typically split into training and testing subsets. The model learns patterns from the training data by adjusting its internal parameters to minimize a predefined loss function. Once the model is trained, it is evaluated on the testing set to assess its performance and generalization ability. This iterative process of learning, validation, and fine-tuning is critical for developing accurate predictive models.
3. What are some common algorithms used in supervised learning?
Answer: Common algorithms in supervised learning include linear regression, logistic regression, decision trees, support vector machines, and neural networks. Each of these algorithms has unique strengths and is selected based on the problem type and data characteristics. For instance, linear regression is ideal for predicting continuous values, while logistic regression is suited for binary classification. These algorithms form the backbone of many modern data science applications and are continuously refined to enhance model performance.
4. What role does data labeling play in supervised learning?
Answer: Data labeling is a critical step in supervised learning because it provides the ground truth that models use to learn the relationship between input features and output responses. Accurate and comprehensive labels ensure that the model can correctly interpret patterns and make reliable predictions. The quality of the labels directly impacts the model’s performance, making data labeling a key factor in the overall success of the learning process. Furthermore, inconsistencies or errors in labeling can lead to biased models and diminished predictive accuracy.
5. How do regression techniques contribute to supervised learning tasks?
Answer: Regression techniques in supervised learning are used to predict continuous outcomes by modeling the relationship between dependent and independent variables. These techniques involve fitting a function to the data that best represents the underlying trend, allowing for accurate forecasting and analysis. Regression methods such as linear regression, polynomial regression, and ridge regression are fundamental tools for understanding and predicting quantitative phenomena. They provide a robust framework for evaluating how changes in input variables influence the target variable, thereby supporting data-driven decision-making.
6. What is classification in the context of supervised learning?
Answer: Classification is a supervised learning task where the goal is to assign input data into predefined categories or classes. It involves training a model on labeled examples so that it can accurately predict the class of new, unseen data. Techniques such as decision trees, k-nearest neighbors, and support vector machines are commonly used for classification problems. By learning from past examples, classification models help automate decision-making in applications like spam detection, image recognition, and medical diagnosis.
7. How are neural networks applied in supervised learning tasks?
Answer: Neural networks are a powerful class of models used in supervised learning for complex tasks such as image and speech recognition. They consist of layers of interconnected nodes that process and transform input data through weighted connections. During training, neural networks adjust these weights using backpropagation to minimize the error between predicted and actual outputs. Their ability to learn intricate, non-linear patterns makes them particularly effective in scenarios where traditional algorithms may struggle.
8. What challenges are commonly encountered in supervised learning?
Answer: Supervised learning faces challenges such as overfitting, where the model learns noise in the training data rather than general patterns, and underfitting, where the model is too simple to capture the underlying structure. Additionally, the quality and quantity of labeled data are crucial; insufficient or biased data can lead to poor model performance. Selecting the right model complexity and tuning hyperparameters are ongoing challenges that require careful consideration. These issues necessitate the use of validation techniques and regularization methods to ensure robust and accurate models.
9. How is model performance evaluated in supervised learning?
Answer: Model performance in supervised learning is typically evaluated using metrics such as accuracy, precision, recall, F1 score for classification tasks, and mean squared error or R-squared for regression tasks. These metrics provide quantitative insights into how well the model predicts unseen data and help identify areas for improvement. Evaluation is conducted on a separate testing dataset to ensure that the model’s performance generalizes beyond the training data. This rigorous assessment process is fundamental for comparing different models and selecting the best one for a given application.
10. What are some real-world applications of supervised learning?
Answer: Supervised learning is used in a variety of real-world applications including image and speech recognition, medical diagnosis, fraud detection, and financial forecasting. By leveraging labeled data, these applications can automate complex decision-making processes and improve operational efficiency. For example, in healthcare, supervised learning models help predict disease outcomes based on patient data. In finance, they assist in identifying fraudulent transactions and forecasting market trends, demonstrating the broad impact of supervised learning across diverse industries.
Supervised Learning: Thought-Provoking Questions and Answers
1. How might advancements in data labeling techniques influence the future of supervised learning?
Answer: Advancements in data labeling techniques, such as automated labeling, active learning, and crowdsourcing, have the potential to greatly enhance the scalability and accuracy of supervised learning models. These improvements could lead to the creation of larger and more diverse labeled datasets, which in turn would enable models to generalize better and reduce bias. Enhanced labeling methods may also lower the cost and time associated with data preparation, making high-quality data more accessible for research and development.
Improved data labeling techniques could foster innovation by enabling the development of more complex models that require vast amounts of data. They might also facilitate the integration of semi-supervised and transfer learning approaches, where limited labeled data is supplemented with unlabeled data. Ultimately, these advancements could transform industries by driving more reliable and efficient AI solutions.
2. What role could explainable AI play in enhancing trust in supervised learning models?
Answer: Explainable AI can demystify the decision-making process of supervised learning models, making it easier for users to understand how predictions are made. By providing insights into the internal workings of these models, explainable AI can help identify and mitigate biases or errors, thus increasing transparency and trust. This is particularly important in high-stakes applications such as healthcare and finance, where understanding the rationale behind decisions is critical.
Greater transparency achieved through explainability can also facilitate better collaboration between data scientists and domain experts. It allows stakeholders to validate and refine models based on clear, interpretable feedback. As a result, explainable AI not only builds trust but also enhances the overall robustness and reliability of supervised learning systems.
3. How can supervised learning be integrated with unsupervised or reinforcement learning to create hybrid models?
Answer: Integrating supervised learning with unsupervised or reinforcement learning can lead to the development of hybrid models that leverage the strengths of each approach. For example, unsupervised learning can be used to uncover hidden patterns in unlabeled data, which can then be used to inform supervised models. Reinforcement learning, on the other hand, can provide dynamic feedback in environments where the optimal action is not immediately clear, complementing the static nature of supervised learning.
Such hybrid models can enhance performance in complex scenarios where data is partially labeled or where environments are highly dynamic. The combination of these methods enables more robust feature extraction, improved generalization, and adaptive decision-making. This integration is expected to open new avenues for research and practical applications across various domains, from robotics to finance.
4. What potential societal impacts could arise from the widespread adoption of supervised learning technologies?
Answer: The widespread adoption of supervised learning technologies has the potential to revolutionize many aspects of society by automating complex tasks and improving decision-making processes. In sectors such as healthcare, transportation, and finance, these technologies can lead to significant improvements in efficiency, accuracy, and accessibility. However, they may also raise concerns related to job displacement, privacy, and algorithmic bias, which need to be carefully managed.
Addressing these societal impacts will require comprehensive strategies that include policy development, ethical guidelines, and continuous oversight. Engaging a broad range of stakeholders—from technologists and policymakers to the general public—will be essential to ensure that the benefits of supervised learning are equitably distributed. Balancing innovation with ethical responsibility will be key to fostering trust and acceptance in these transformative technologies.
5. How might the increasing complexity of supervised learning models affect their transparency and interpretability?
Answer: As supervised learning models become more complex, particularly with the rise of deep learning architectures, maintaining transparency and interpretability becomes a significant challenge. Complex models often operate as “black boxes,” making it difficult to understand how inputs are transformed into outputs. This opacity can hinder efforts to diagnose errors, identify biases, or comply with regulatory requirements.
To address these issues, researchers are exploring techniques such as model distillation, feature importance analysis, and visualization tools that can provide insights into complex model behaviors. These methods aim to strike a balance between leveraging the power of sophisticated models and ensuring they remain understandable to users and stakeholders. Enhancing transparency will be essential for the broader adoption of advanced supervised learning systems in sensitive applications.
6. Can the evolution of computing hardware reshape the efficiency and scalability of supervised learning algorithms?
Answer: The evolution of computing hardware, including the development of specialized processors like GPUs and TPUs, has already had a profound impact on the efficiency and scalability of supervised learning algorithms. Enhanced hardware capabilities enable the processing of massive datasets and the training of complex models in significantly reduced timeframes. This evolution allows researchers and practitioners to experiment with more sophisticated architectures and larger datasets than ever before.
In the future, continued improvements in hardware, such as quantum computing or neuromorphic chips, could further accelerate the pace of innovation in supervised learning. These advancements may lead to breakthroughs in real-time data processing, more energy-efficient training, and the ability to deploy models on a much larger scale. As a result, the potential applications of supervised learning could expand dramatically across various industries.
7. How might transfer learning techniques revolutionize the application of supervised learning in emerging fields?
Answer: Transfer learning techniques allow models to leverage knowledge gained from one task and apply it to another, thereby reducing the need for large labeled datasets in emerging fields. This approach is particularly valuable in areas where acquiring labeled data is challenging or expensive, as it can significantly shorten the training process while maintaining high performance. By reusing pre-trained models, practitioners can rapidly adapt to new tasks, enhancing the versatility and efficiency of supervised learning.
The adoption of transfer learning could revolutionize industries such as medical diagnostics, natural language processing, and environmental monitoring, where domain-specific data is often scarce. It enables the development of robust models with limited resources, fostering innovation in fields that were previously hindered by data constraints. Ultimately, transfer learning has the potential to democratize advanced AI applications and accelerate their real-world impact.
8. What challenges might arise from biases in labeled datasets, and how can they be mitigated in supervised learning?
Answer: Biases in labeled datasets can lead to skewed model predictions and unfair outcomes, as the learning process is heavily dependent on the quality and representativeness of the training data. Such biases may arise from historical inequities, sampling errors, or subjective labeling processes, and can negatively affect the performance of supervised learning models. Addressing these challenges requires meticulous data curation, robust validation techniques, and continuous monitoring of model outputs for bias.
Mitigation strategies include data augmentation, re-sampling methods, and the incorporation of fairness-aware algorithms that explicitly address potential imbalances. Additionally, transparency in the data collection and labeling process is crucial to identify and correct biases early on. These steps are essential to ensure that supervised learning models deliver equitable and reliable outcomes in diverse real-world scenarios.
9. How can interdisciplinary research contribute to solving complex problems in supervised learning?
Answer: Interdisciplinary research brings together diverse perspectives and expertise from fields such as computer science, statistics, domain-specific sciences, and ethics, which can lead to innovative solutions in supervised learning. By integrating knowledge from various disciplines, researchers can develop more robust models that address both technical challenges and societal concerns. This collaborative approach fosters the development of novel methodologies, such as advanced feature engineering and bias mitigation techniques, that enhance model performance.
Moreover, interdisciplinary efforts can help bridge the gap between theoretical research and practical applications, ensuring that supervised learning technologies are both effective and responsible. Such collaborations not only enrich the research process but also promote the adoption of best practices that align technological advancements with real-world needs.
10. What future trends in supervised learning could disrupt current AI applications in industry?
Answer: Future trends in supervised learning, such as the integration of deep learning with emerging data modalities and the rise of automated machine learning (AutoML), have the potential to disrupt existing AI applications in industry. These trends may lead to the development of more adaptive, accurate, and efficient models that can handle increasingly complex data environments. Innovations in model interpretability and transfer learning are also expected to play a significant role in reshaping how supervised learning is applied.
As these advancements mature, industries may witness a transformation in the way data-driven decisions are made, leading to enhanced productivity, reduced costs, and improved customer experiences. The convergence of cutting-edge research with practical applications is likely to drive a new era of AI innovation, setting new benchmarks for performance and reliability in supervised learning.
11. How might ethical considerations shape the development and deployment of supervised learning models in sensitive areas?
Answer: Ethical considerations are critical in guiding the development and deployment of supervised learning models, particularly in sensitive areas such as healthcare, criminal justice, and finance. These models must be designed to ensure fairness, transparency, and accountability, as errors or biases can have significant real-world consequences. Researchers and practitioners need to establish robust ethical frameworks and continuously assess the societal impacts of their models.
Incorporating ethical guidelines into the design process can help prevent discriminatory outcomes and build public trust. It also involves engaging stakeholders from diverse backgrounds to review and refine models, ensuring that they are aligned with broader social values and regulatory standards. This proactive approach is essential for fostering responsible AI practices.
12. What are the implications of combining supervised learning with real-time data analytics for dynamic decision-making systems?
Answer: Combining supervised learning with real-time data analytics can lead to highly dynamic decision-making systems capable of adapting to changing environments and rapidly evolving data patterns. This integration enables models to continuously learn from new data and make timely predictions that support critical business and operational decisions. The ability to process and analyze data in real time can significantly enhance responsiveness and accuracy, especially in fast-paced industries.
Such systems can revolutionize sectors like finance, emergency response, and supply chain management by providing actionable insights when they are needed most. The synergy between supervised learning and real-time analytics not only improves operational efficiency but also creates opportunities for proactive and informed decision-making in complex, dynamic settings.
Supervised Learning: Numerical Problems and Solutions
1. Calculating Mean Squared Error (MSE) for Regression Predictions
Solution:
Step 1: Given true values [3, -0.5, 2, 7] and predictions [2.5, 0.0, 2, 8], compute the error for each pair: errors are (0.5, -0.5, 0, -1).
Step 2: Square each error: (0.25, 0.25, 0, 1).
Step 3: Sum the squared errors and divide by the number of observations: MSE = (0.25 + 0.25 + 0 + 1) / 4 = 1.5 / 4 = 0.375.
2. Computing Root Mean Squared Error (RMSE) from Given Data
Solution:
Step 1: Use the MSE value calculated previously (0.375).
Step 2: Take the square root of the MSE: RMSE = √0.375.
Step 3: Calculate the value to obtain RMSE ≈ 0.612.
3. Determining the Coefficient of Determination (R²)
Solution:
Step 1: For true values [10, 20, 30, 40, 50] and predictions [12, 18, 33, 37, 52], compute the residuals and square them to obtain the Sum of Squared Residuals (SSR).
Step 2: Calculate the Total Sum of Squares (TSS) by subtracting the mean of true values from each true value, squaring the differences, and summing them.
Step 3: Use the formula R² = 1 – (SSR/TSS) to compute the coefficient of determination.
4. Calculating Accuracy from a Confusion Matrix
Solution:
Step 1: Given a confusion matrix with True Positives = 40, False Positives = 10, True Negatives = 35, and False Negatives = 15, compute the total correct predictions: 40 + 35 = 75.
Step 2: Sum all values in the confusion matrix to get the total predictions: 40 + 10 + 35 + 15 = 100.
Step 3: Calculate accuracy = 75 / 100 = 0.75 or 75%.
5. Determining Precision and Recall from Classification Results
Solution:
Step 1: Use the confusion matrix values: True Positives = 40, False Positives = 10, and False Negatives = 15.
Step 2: Compute precision = TP / (TP + FP) = 40 / (40 + 10) = 40 / 50 = 0.8.
Step 3: Compute recall = TP / (TP + FN) = 40 / (40 + 15) = 40 / 55 ≈ 0.727.
6. Computing the F1 Score from Precision and Recall
Solution:
Step 1: With precision = 0.8 and recall ≈ 0.727, compute the product: 0.8 × 0.727 ≈ 0.582.
Step 2: Calculate the sum: 0.8 + 0.727 ≈ 1.527.
Step 3: Compute F1 score = 2 × (0.582 / 1.527) ≈ 2 × 0.381 = 0.762.
7. Performing a Gradient Descent Update Step for Linear Regression
Solution:
Step 1: Assume an initial weight (w) = 0.5, a learning rate (α) = 0.1, and a computed gradient = 0.2 for a given iteration.
Step 2: Multiply the learning rate by the gradient: 0.1 × 0.2 = 0.02.
Step 3: Update the weight: new w = 0.5 – 0.02 = 0.48.
8. Estimating the Cost Function Value in Linear Regression
Solution:
Step 1: For a dataset with inputs X = [1, 2, 3] and true outputs Y = [2, 4, 6], and a hypothesis h(x) = θ₀ + θ₁·x with θ₀ = 0.5 and θ₁ = 1.5, compute the predictions: [2.0, 3.5, 5.0].
Step 2: Calculate the squared errors: (2 – 2.0)² = 0, (4 – 3.5)² = 0.25, (6 – 5.0)² = 1.
Step 3: Compute the mean squared error: (0 + 0.25 + 1) / 3 ≈ 0.4167.
9. Calculating the Effective Learning Rate with Decay
Solution:
Step 1: Start with an initial learning rate of 0.5 and a decay factor of 0.9 every 10 iterations.
Step 2: For 30 iterations, the learning rate is decayed three times: effective rate = 0.5 × (0.9)³.
Step 3: Compute effective learning rate = 0.5 × 0.729 = 0.3645.
10. Estimating the Number of Iterations for Error Reduction in Gradient Descent
Solution:
Step 1: Begin with an initial error of 100 that reduces by 5% each iteration, so error after n iterations is given by 100 × (0.95)ⁿ.
Step 2: Set the target error to 50 and solve 100 × (0.95)ⁿ ≤ 50, which simplifies to (0.95)ⁿ ≤ 0.5.
Step 3: Taking logarithms yields n ≥ log(0.5) / log(0.95) ≈ 13.5, so approximately 14 iterations are needed.
11. Evaluating the Cross-Validation Error from K-Fold Data
Solution:
Step 1: Given 5-fold cross-validation error values: [0.2, 0.25, 0.22, 0.24, 0.26], sum these errors: 0.2 + 0.25 + 0.22 + 0.24 + 0.26 = 1.17.
Step 2: Compute the average error: 1.17 / 5 = 0.234.
Step 3: To calculate the standard deviation, compute the squared differences from the mean for each fold, sum them, divide by (n – 1) = 4, and take the square root.
12. Computing Adjusted R-Squared for a Regression Model
Solution:
Step 1: Given R² = 0.85, number of predictors p = 3, and sample size n = 50, use the formula Adjusted R² = 1 – [(1 – R²) × (n – 1) / (n – p – 1)].
Step 2: Substitute the values: Adjusted R² = 1 – [(0.15) × (49) / (50 – 3 – 1)] = 1 – [(0.15 × 49) / 46].
Step 3: Compute the numerator (0.15 × 49 = 7.35) and divide by 46 to obtain ≈ 0.1598, then Adjusted R² = 1 – 0.1598 = 0.8402.