examhub .cc The most efficient path to the most valuable certifications.
In this note ≈ 31 min

AI and ML Core Concepts

6,100 words · ≈ 31 min read

The AI and ML Core Concepts topic is the conceptual backbone of the AWS Certified AI Practitioner (AIF-C01) exam. Task statement 1.1 — "Explain basic AI concepts and terminologies" — is the single most broadly cited task on the exam, because every other domain (Generative AI, Foundation Models, Responsible AI, Security) silently assumes you already know what a model, a feature, a label, a neural network, and an inference call are. If you misread the AI and ML Core Concepts vocabulary, you will misread the rest of the exam.

This study note walks through the entire AI and ML Core Concepts surface an AIF-C01 candidate is expected to recognize: the AI → ML → Deep Learning → Generative AI hierarchy, the narrow-vs-general-AI scope, the lexicon of model / training / inference / feature / label / parameter / hyperparameter, the physical layout of a neural network (neurons, layers, activation functions), the four canonical use-case patterns (classification, regression, clustering, generation), the evaluation-metric family (accuracy, precision, recall, F1), and the data-quality fundamentals that make or break every downstream system. A clear distinction note separates AIF-C01 recognition-level depth from AIP-C01 build-level depth so you do not over-study.

What Is Artificial Intelligence? Scope, History, and the AI/ML/DL Hierarchy

Artificial Intelligence (AI) is the umbrella discipline of building computer systems that perform tasks traditionally requiring human intelligence — perception, reasoning, language, decision-making, and creativity. The field was named at the 1956 Dartmouth Workshop and has moved through several waves: symbolic rule-based systems in the 1970s, statistical learning in the 1990s, deep learning breakthroughs starting in 2012, and the foundation-model / generative-AI wave that began accelerating in 2018 and exploded in 2022 with ChatGPT. The AIF-C01 exam does not test dates, but it does test the AI and ML Core Concepts scope: AI is the broadest circle, ML sits inside AI, Deep Learning sits inside ML, and Generative AI is a capability band that spans Deep Learning and Foundation Models.

Understanding the AI and ML Core Concepts hierarchy eliminates the most common vocabulary mistake candidates make on exam day — treating AI, ML, and Deep Learning as synonyms.

Artificial Intelligence (AI) is any system that simulates intelligent behavior. Machine Learning (ML) is a subset of AI in which systems learn patterns from data rather than follow hand-coded rules. Deep Learning (DL) is a subset of ML that uses multi-layer neural networks. Generative AI (GenAI) is a capability category, usually powered by deep-learning-based foundation models, that produces new content (text, images, audio, code). Source ↗

The Four Nested Circles

Think of AI, ML, DL, and GenAI as four concentric circles:

  1. AI (outermost) — any intelligent-seeming system, including rule-based expert systems and search algorithms that contain no learning at all.
  2. ML — systems that learn statistical patterns from data; requires training data and produces a model.
  3. DL — ML that uses neural networks with many layers; the algorithm of choice for images, speech, and language.
  4. GenAI (innermost capability) — DL-powered foundation models that create new artifacts rather than classify existing ones.

Every GenAI system is a DL system. Every DL system is an ML system. Every ML system is an AI system. The reverse direction does not hold: a chess engine using minimax search is AI but not ML; a linear-regression sales forecaster is ML but not DL; an image classifier is DL but not GenAI.

Why AIF-C01 Obsesses Over This Hierarchy

The AIF-C01 exam guide explicitly lists "Differentiate between AI, ML, and Deep Learning" as an in-scope knowledge statement under task 1.1. Community exam reports consistently flag two to three questions per attempt that test whether the candidate can place a scenario ("a retail company builds a customer-support chatbot on a large language model") into the right circle (GenAI inside DL inside ML inside AI). Miss the hierarchy and you will pick a plausible but wrong answer.

Machine Learning vs Deep Learning vs Traditional Programming

One of the cleanest ways to internalize AI and ML Core Concepts is to contrast the three paradigms of building software behavior.

Traditional Programming

In traditional programming, a human writes explicit rules and the computer applies those rules to input data to produce output. The logic lives in if / else statements, lookup tables, and mathematical formulas. This works when the rules are knowable and stable — payroll calculations, tax brackets, unit conversions.

Machine Learning

In ML, the human provides input data and desired outputs, and the machine learns the rules. The output of training is a model — a parameterized function that maps new inputs to predictions. ML is the right tool when the rules are too numerous, too nuanced, or too adaptive for a human to codify (spam detection, credit-risk scoring, medical-image triage).

Deep Learning

Deep Learning is ML that uses neural networks with many hidden layers. The "deep" refers to the number of layers, not to some philosophical profundity. Deep networks excel at extracting hierarchical features from unstructured data: edges → shapes → objects in images; phonemes → words → sentences in speech; tokens → grammar → semantics in text. Deep Learning needs more data and more compute than classical ML but achieves state-of-the-art accuracy on perception and language tasks.

The Three Paradigms Side-by-Side

Dimension Traditional Programming Machine Learning Deep Learning
Source of logic Human writes rules Algorithm learns from data Neural network learns from data
Input needed Rules + input data Input + labeled output examples Lots of input + lots of labeled output
Output Deterministic result Statistical prediction Statistical prediction
Best for Stable, well-understood domains Tabular data, moderate data size Images, audio, language, very large data
Compute cost Low Low to moderate High (often GPU/accelerator)

When an AIF-C01 scenario says "the team cannot manually write rules because there are too many edge cases," the answer almost certainly involves ML rather than traditional programming. When the scenario adds "working with unstructured data like images or natural language," add Deep Learning to the answer. Source ↗

Generative AI in the AI/ML/DL Taxonomy

Before 2020, most ML systems were discriminative: they classified, scored, or ranked existing inputs. Generative AI flips the direction — instead of asking "what label does this input belong to?", it asks "what new content best continues or answers this prompt?".

Discriminative vs Generative

  • Discriminative models learn the boundary between classes. Examples: email spam classifier, credit-default predictor, tumor-vs-benign image classifier.
  • Generative models learn the distribution of the data and can sample new examples from it. Examples: Large Language Models (LLMs) that produce new paragraphs, diffusion models that produce new images, code assistants that produce new source code.

Where GenAI Lives in the Hierarchy

Generative AI is implemented almost exclusively on top of deep neural networks — specifically the Transformer architecture for language and the diffusion architecture for images. GenAI is therefore a capability stripe across Deep Learning rather than a separate paradigm. AIF-C01 treats GenAI as a first-class domain (24% of exam weight), so a firm grasp of where it sits relative to classical ML is essential.

Foundation Models as the GenAI Engine

A Foundation Model (FM) is a very large deep-learning model, pre-trained on broad, diverse data, and adaptable to many downstream tasks without task-specific retraining. FMs are the current engine of GenAI on AWS, accessed through Amazon Bedrock. AIF-C01 expects recognition of FM characteristics (scale, generality, adaptability) but does not require you to build or fine-tune one at the code level — that depth belongs to AIP-C01.

Narrow AI vs General AI vs Artificial General Intelligence (AGI)

AI and ML Core Concepts questions sometimes reach into the philosophical classification of AI systems. Keep the distinctions short and exam-shaped.

Narrow AI (Artificial Narrow Intelligence, ANI)

Narrow AI performs a specific task extremely well but cannot generalize outside that task. Every commercially deployed AI system today — including Amazon Rekognition, Amazon Transcribe, GPT-class LLMs, AlphaGo, autonomous-driving stacks — is narrow AI. Narrow AI can be superhuman within its domain and completely helpless outside it.

General AI (Artificial General Intelligence, AGI)

AGI would match human cognitive flexibility across any intellectual task a human can perform: learning new skills from scratch, transferring knowledge across domains, reasoning about novel situations without retraining. AGI does not exist today. Research organizations pursue it; no product delivers it.

Superintelligence

A hypothetical future AI that surpasses human intelligence across all domains. Not relevant for AIF-C01 except as a distractor answer choice.

AIF-C01 loves to sneak "AGI" into answer choices as a distractor.

If a question describes a real AWS product (Rekognition, Bedrock Claude, SageMaker) and asks you to categorize it, the answer is always Narrow AI. No AWS service and no publicly available foundation model qualifies as AGI. Answer choices that say "the system demonstrates general intelligence" or "the model has achieved AGI" are wrong.

Similarly, when a scenario says "the model performs well on a narrow task," do not reach for the word "general" — it contradicts the setup. Source ↗

Plain-Language Explanation: AI and ML Core Concepts

Abstract taxonomy becomes intuitive when you anchor it to physical, everyday systems. Three very different analogies cover the full sweep of AI and ML Core Concepts.

Analogy 1: The Kitchen — AI, ML, Deep Learning, and Generative AI

Imagine a large restaurant kitchen. Artificial Intelligence is the whole kitchen operation — everything that produces a dish, whether by a cook following a laminated recipe card (rule-based) or by a chef improvising from experience (learning-based). Machine Learning is the chef who tastes thousands of sauces over a career and develops an intuitive sense of salt balance; the "rules" live in the chef's palate, not on paper. Deep Learning is a brigade of line cooks layered in stations — prep, sauté, sauce, garnish — where each station refines the work of the previous one and the final plate emerges from many sequential transformations. Generative AI is the pastry chef who, given the phrase "a warm dessert for a rainy Tuesday," invents a new dish that has never existed.

This kitchen analogy maps every core AI and ML Core Concepts term:

  • A model is the chef's trained palate.
  • Training is the years of tasting that calibrate the palate.
  • Inference is one service shift where the palate is applied to new sauces.
  • Features are the ingredients and techniques.
  • Labels are the customer feedback scores during training.
  • Neural network layers are the prep → sauté → sauce → garnish stations.

Analogy 2: The Library — Data Quality and the Training Process

A library's quality depends entirely on its cataloging process. Good librarians carefully label every book by genre, author, language, and subject. If the labels are wrong — a cookbook mis-shelved in the science section — every future reader who trusts the catalog ends up disappointed. AI and ML Core Concepts around data quality work identically. A model trained on mislabeled images (a cat tagged as a dog) will confidently output wrong predictions, because the model cannot distinguish between "the data is lying" and "the data is telling the truth." The training data is the catalog; the model is the librarian who internalized it. Garbage in, garbage out is not a slogan — it is the most deeply studied phenomenon in all of ML.

This analogy is especially useful for understanding training / validation / test splits. The training set is the collection the librarian learns from. The validation set is a small sample held back to quiz the librarian before certification. The test set is the surprise audit after certification: books the librarian has genuinely never seen.

Analogy 3: The Open-Book Exam — Parameters, Hyperparameters, and Inference

Picture a student preparing for an open-book exam. The student reads hundreds of practice problems and develops a study strategy: which chapters to review first, how many practice problems to attempt, how long to spend on each one. That strategy is the hyperparameter set — decisions made before studying begins.

As the student solves practice problems, they build up mental connections: "when I see a problem that mentions rate × time, I recall the distance formula." Those mental connections are the parameters — the learned weights inside the model. On exam day (inference), the student does not get to re-study; they apply the internalized parameters, guided by the strategy that shaped their study time, to produce answers to questions they have never seen.

This analogy clarifies why hyperparameter tuning is a distinct step from training, why a model with too few parameters underfits (the student did not build enough mental connections), and why a model with too many parameters relative to training data overfits (the student memorized practice problems word-for-word instead of learning concepts).

Which Analogy to Use on Exam Day

  • Questions about taxonomy (AI vs ML vs DL vs GenAI) → kitchen analogy.
  • Questions about data quality, labeling, training splits → library analogy.
  • Questions about parameters, hyperparameters, training vs inference → open-book-exam analogy.

Key ML Terminology: Model, Training, Inference, Feature, Label, Parameter, Hyperparameter

Every AIF-C01 study session must end with these seven words internalized. Mis-defining any one of them can cost two or three exam questions.

Model

A model is the learned function that maps inputs to predictions. A trained model is the artifact saved at the end of training — a file of numeric weights plus the architecture description that knows how to consume them. On AWS, a model might be a SageMaker-produced .tar.gz artifact, a Bedrock-hosted foundation model you never download, or a Rekognition Custom Labels project endpoint.

Training

Training is the process of showing labeled examples to an algorithm so it can adjust internal parameters until its predictions are good enough. Training is expensive, batch-oriented, and usually one-shot per model generation. It requires large compute (CPU, GPU, or AWS Trainium) and large data.

Inference

Inference is the act of using an already-trained model to make predictions on new inputs. Inference is cheap per call, runs in production continuously, and is usually latency-sensitive. AWS inference surfaces include SageMaker real-time endpoints, SageMaker Batch Transform, Bedrock InvokeModel, and AWS Inferentia chips. Confusing training with inference is a high-frequency AIF-C01 trap because cost and latency profiles differ by orders of magnitude.

Feature

A feature is a single measurable input variable fed to the model. For a credit-scoring model, features include income, credit-utilization ratio, and years of employment. Feature engineering — selecting, transforming, and combining features — is one of the highest-leverage activities in classical ML.

Label

A label is the correct answer attached to a training example. For a spam classifier, the label is spam or not spam; for a house-price model, the label is the sale price. Labels are expensive to produce (humans must attach them) and define what the model learns. Unlabeled data cannot be used for supervised learning.

Parameter

Parameters are the internal numeric values (weights and biases) the algorithm learns during training. A small linear-regression model might have a dozen parameters. A large foundation model might have hundreds of billions. Parameters are the output of training, not the input.

Hyperparameter

Hyperparameters are the knobs you set before training starts: learning rate, number of epochs, batch size, number of layers, dropout rate. Hyperparameters shape how the model learns. SageMaker Automatic Model Tuning searches hyperparameter space on your behalf.

Seven words, one sentence each — drill until reflexive:

  • Model = the trained function artifact
  • Training = one-time learning process that produces the model
  • Inference = production-time prediction using the trained model
  • Feature = one input variable fed to the model
  • Label = the ground-truth answer attached to a training example
  • Parameter = numeric weight learned during training (output of training)
  • Hyperparameter = knob set by humans before training starts (input to training)

Distractor cue: if an AIF-C01 answer choice swaps "parameter" and "hyperparameter," it is wrong. Parameters are learned; hyperparameters are chosen. Source ↗

Neural Networks 101: Layers, Weights, Activation Functions

AIF-C01 tests neural network basics at a conceptual, not mathematical, level. You do not need to compute a gradient; you need to recognize the parts and what they do.

The Neuron (Node)

A neuron (or node) is the atomic unit of a neural network. It accepts several numeric inputs, multiplies each by a learned weight, sums the weighted inputs plus a bias, and passes the result through an activation function to produce a single numeric output. A neuron by itself is weak; the power comes from connecting thousands of neurons in layers.

Layers

A neural network is organized into layers:

  • Input layer — one neuron per input feature.
  • Hidden layers — intermediate layers where the network builds up abstractions. A "deep" network has many hidden layers.
  • Output layer — produces the final prediction (one neuron per class for classification; one neuron for regression).

Each neuron in one layer typically connects to every neuron in the next layer (a "fully connected" or "dense" layer). Other architectures — convolutional layers for images, recurrent layers for sequences, transformer attention layers for language — are specialized variations on the same idea.

Weights and Biases

Weights are the numeric multipliers attached to every neuron-to-neuron connection. Biases are additive offsets per neuron. Together, weights and biases are the parameters the network learns during training. A modern foundation model like Claude 3 has on the order of hundreds of billions of these parameters.

Activation Functions

An activation function is applied to each neuron's weighted sum to introduce non-linearity. Without activation functions, a stack of neurons would collapse mathematically into a single linear model, and the network could not learn complex patterns. Common activation functions AIF-C01 may name:

  • ReLU (Rectified Linear Unit) — output = max(0, input); the workhorse of modern deep learning.
  • Sigmoid — squashes output to (0, 1); used for binary classification output layer.
  • Softmax — normalizes outputs to a probability distribution; used for multi-class output layer.
  • Tanh — squashes output to (-1, 1); common in older architectures and some recurrent networks.

How a Neural Network Learns: Forward Pass and Backpropagation

During training, every example goes through a forward pass: inputs flow through layers until the output layer produces a prediction. The prediction is compared to the correct label via a loss function, and an error score is computed. Backpropagation then runs the error backward through the network, calculating how each weight contributed to the error. Each weight is adjusted slightly in the direction that would reduce the error. Over millions of examples, the weights converge on values that produce accurate predictions.

AIF-C01 does not ask you to implement backpropagation. It does ask you to recognize that:

  • Training = many forward + backward passes adjusting weights.
  • Inference = forward passes only.
  • More layers + more parameters = more capacity but more compute and more risk of overfitting.

AIF-C01 tests neural network basics at the recognition level. You should be able to identify layers, weights, activation functions, and the distinction between forward pass (inference) and backward pass (training). You do not need to compute matrix multiplications or derive gradient formulas — that depth belongs to SageMaker model-development training, not to AIF-C01. Source ↗

Training Data, Validation Data, and Test Data

Every ML model is built from three disjoint datasets. Confusing their roles is one of the most heavily penalized AI and ML Core Concepts mistakes on the exam.

Training Set

The training set is the largest slice (typically 60–80% of available data). The model sees these examples and adjusts its parameters based on them. Model quality on the training set alone is meaningless — a memorizing model can achieve 100% training accuracy and fail disastrously in production.

Validation Set

The validation set (typically 10–20%) is held back from training and used to evaluate candidate model configurations during development. It guides hyperparameter choices, architecture comparisons, and early-stopping decisions. The model never directly trains on validation data, but it is indirectly shaped by it because you pick configurations based on validation scores.

Test Set

The test set (typically 10–20%) is the final, untouched evaluation set used once after all development is complete. It simulates production unseen data. If you peek at the test set during development, you contaminate it and lose the ability to honestly estimate real-world performance.

Why the Split Matters

Without separate splits, you cannot distinguish a model that learned the underlying patterns from a model that memorized the training data. The validation / test split is the primary defense against overfitting and the primary source of honest performance numbers.

Structured vs Unstructured Data

Data quality fundamentals begin with recognizing what kind of data you have.

Structured Data

Structured data lives in rows and columns with known schema: SQL tables, CSV files, Parquet, relational databases. Structured data is the natural input for classical ML algorithms (logistic regression, gradient boosting, random forests) and for AWS services like Amazon SageMaker tabular training and Amazon Personalize.

Unstructured Data

Unstructured data includes images, video, audio, and free-form text. There is no fixed schema. Unstructured data is the natural input for deep learning and for AWS services like Amazon Rekognition (images, video), Amazon Transcribe (audio), Amazon Comprehend (text), and Amazon Bedrock foundation models.

Semi-Structured Data

Semi-structured data (JSON, XML, log files) has some organizational markers but no fixed schema. AWS Glue, Amazon Athena, and SageMaker Data Wrangler excel at converting semi-structured inputs into training-ready forms.

Why This Matters for the Exam

AIF-C01 scenarios often hide the answer in data-type vocabulary. "The company has millions of scanned PDFs" → unstructured → reach for Textract, Rekognition, or foundation-model document understanding. "The company has a 10-million-row transactions table" → structured → reach for SageMaker Canvas, Amazon Personalize, or SageMaker training jobs.

Common Use-Case Patterns: Classification, Regression, Clustering, Generation

Four patterns cover the vast majority of AI and ML Core Concepts use cases tested on AIF-C01.

Classification

Classification predicts a discrete label from a finite set. Binary classification has two classes (spam / not spam, fraud / legitimate). Multi-class classification has many (dog / cat / rabbit / horse). Multi-label classification allows multiple labels per example (an image tagged both "beach" and "sunset"). Algorithms: logistic regression, decision trees, random forests, gradient boosting, neural networks.

Regression

Regression predicts a continuous numeric value. House-price prediction from square footage, energy-demand forecasting from weather, delivery-time estimation from distance. Algorithms: linear regression, polynomial regression, gradient boosting regressors, deep neural networks.

Clustering

Clustering groups similar examples together without labeled data. K-means is the canonical algorithm. Use cases: customer segmentation, anomaly detection, document topic discovery. Clustering is unsupervised — it needs no labels and produces no fixed labels, just cluster assignments.

Generation

Generation produces new content (text, images, audio, code) conditioned on a prompt or input. Implemented by foundation models and diffusion models. Use cases: content drafting, code assistance, synthetic-data creation, conversational agents. This is the generative AI pattern.

Memorize the keyword-to-pattern map:

  • "Predict a category / class / label" → classification
  • "Predict a numeric value / amount / price / duration" → regression
  • "Group similar items / segment customers / discover structure" → clustering
  • "Create / write / draw / compose new content" → generation

AIF-C01 scenario questions almost always telegraph the pattern through a single verb. Source ↗

Batch Inference vs Real-Time Inference

Both are inference; both run trained models on new data; their deployment shapes differ.

Real-Time (Online) Inference

Real-time inference serves a prediction within milliseconds, one request at a time, via an HTTPS endpoint. Use when a user or application is waiting for a response: fraud scoring during a checkout, chatbot replies, recommendation rendering. AWS surfaces: SageMaker real-time endpoints, Bedrock InvokeModel, SageMaker Serverless Inference (cold-start tolerant).

Batch Inference

Batch inference runs predictions on a whole dataset at once, asynchronously, writing results to storage (usually S3). Use when latency is not critical: nightly scoring of an entire customer base, weekly risk recomputation, monthly report generation. AWS surfaces: SageMaker Batch Transform, Bedrock Batch Inference, SageMaker Async Inference (middle ground).

Cost and Latency Profiles

  • Real-time endpoints incur cost while provisioned, even when idle.
  • Batch jobs incur cost only during the run.
  • Serverless and async inference attempt to bridge the two at the cost of some cold-start latency.

Knowing which profile suits which scenario is tested repeatedly on AIF-C01.

Evaluation Metrics: Accuracy, Precision, Recall, F1

Evaluation metrics are the numerical measures of model quality. For classification tasks — the most commonly tested family — four metrics dominate.

The Confusion Matrix

All four classification metrics derive from the confusion matrix:

  • TP (True Positive) — predicted positive, actually positive.
  • TN (True Negative) — predicted negative, actually negative.
  • FP (False Positive) — predicted positive, actually negative (type I error).
  • FN (False Negative) — predicted negative, actually positive (type II error).

Accuracy

Accuracy = (TP + TN) / (TP + TN + FP + FN). The fraction of predictions that are correct. Accuracy is intuitive but misleading when classes are imbalanced. A fraud detector that always predicts "not fraud" will score 99.5% accuracy on a dataset where fraud is 0.5% of transactions — and be completely useless.

Precision

Precision = TP / (TP + FP). Of everything the model predicted positive, how many actually were positive. High precision means few false alarms. Use precision when the cost of a false positive is high — marking a legitimate transaction as fraud and blocking a customer's card, or diagnosing a healthy patient with cancer.

Recall (Sensitivity)

Recall = TP / (TP + FN). Of everything that was actually positive, how many did the model catch. High recall means few missed cases. Use recall when the cost of a false negative is high — missing an actual cancer case, failing to detect an intrusion, overlooking a fraudulent transaction.

F1 Score

F1 = 2 × (Precision × Recall) / (Precision + Recall). The harmonic mean of precision and recall. F1 balances both, penalizing extreme imbalance between them. Use F1 when you care about both error types and the class distribution is imbalanced.

Regression Metrics (Briefer Coverage)

For regression, the common metrics are:

  • MAE (Mean Absolute Error) — average absolute difference between predicted and actual.
  • MSE (Mean Squared Error) — average squared difference; penalizes large errors more.
  • RMSE (Root Mean Squared Error) — square root of MSE, same units as target.
  • (coefficient of determination) — fraction of variance in the target explained by the model; 1.0 is perfect, 0 is no better than predicting the mean.

The four-metric classification cheat sheet:

  • Accuracy — overall correctness; unreliable on imbalanced data.
  • Precision — when you predict positive, how often are you right? Use when false positives are costly.
  • Recall — of all actual positives, how many did you catch? Use when false negatives are costly.
  • F1 — balanced average of precision and recall.

AIF-C01 scenario cues:

  • "Minimize false alarms / don't want to block innocent users" → precision
  • "Don't miss any actual cases / catch every instance" → recall
  • "Imbalanced data and both error types matter" → F1
  • "Balanced data, simple scorecard" → accuracy

Source ↗

Data Quality Fundamentals

Every AI and ML Core Concepts conversation ends here: without clean, representative, well-labeled data, no model can succeed. AIF-C01 includes data quality under task 1.3 (ML lifecycle) and task 4.1 (responsible AI) but the core vocabulary is introduced with core concepts.

The Dimensions of Data Quality

  • Accuracy — values correctly reflect the real-world state.
  • Completeness — no missing critical fields; missing-value strategies handle gaps.
  • Consistency — the same entity is represented identically across records ("USA" vs "United States" vs "U.S.A.").
  • Timeliness — data reflects the current state within the business's freshness requirements.
  • Validity — values conform to expected types, ranges, and formats.
  • Uniqueness — duplicate records are removed or merged.
  • Relevance — features relate to the problem being solved.

Common Data Quality Problems

  • Missing values — blank cells, null fields. Handled by imputation, deletion, or model-aware strategies.
  • Outliers — extreme values that distort learning. Handled by winsorizing, transforming, or investigation.
  • Class imbalance — one class overwhelms the others. Handled by resampling, class weights, or specialized metrics like F1.
  • Labeling errors — incorrect ground truth. Handled by re-labeling, consensus labeling, or tools like Amazon A2I.
  • Data leakage — information from the test set sneaks into training, inflating scores. Handled by disciplined splitting and feature engineering.
  • Sampling bias — training data does not represent the production population. Handled by stratified sampling and domain review.

Why Data Quality Matters for Generative AI

Foundation models and large language models are trained on vast datasets scraped from the public web. Their outputs inherit whatever biases, inaccuracies, and stale knowledge the training data contained. This is why AIF-C01 Domain 4 (Responsible AI) devotes so much attention to bias, fairness, and explainability — the mathematical properties of the model cannot overcome the properties of the data that shaped it.

Garbage in, garbage out is the single most durable principle in AI and ML Core Concepts. No amount of hyperparameter tuning, model-architecture upgrade, or fine-tuning budget can compensate for systematically flawed training data. When an AIF-C01 scenario mentions "model performance is poor after deployment," the first root-cause category to consider is always data quality — missing fields, label errors, distribution shift between training and production, or sampling bias. Source ↗

AI Service Categories on AWS: Pre-Built AI APIs vs Custom ML vs Generative AI

AWS organizes AI capabilities in three tiers; AIF-C01 expects recognition of which tier a service belongs to.

Tier 1: Pre-Built AI Services (AI APIs)

Fully managed, task-specific APIs. No ML expertise required. Examples: Amazon Rekognition (vision), Amazon Transcribe (speech-to-text), Amazon Polly (text-to-speech), Amazon Translate, Amazon Comprehend (NLP), Amazon Textract (document extraction), Amazon Personalize (recommendations), Amazon Forecast, Amazon Kendra (intelligent search), Amazon Lex (chatbots).

Tier 2: Custom ML Platform (Amazon SageMaker)

End-to-end ML platform for building, training, and deploying your own models. Includes SageMaker Studio, Training Jobs, Endpoints, Pipelines, JumpStart, Canvas, Data Wrangler. Use when the task requires custom logic beyond the pre-built APIs.

Tier 3: Generative AI Platform (Amazon Bedrock + Amazon Q)

Managed access to foundation models (Anthropic Claude, Amazon Titan, Meta Llama, Mistral, Cohere, Stability AI) with built-in features for Knowledge Bases, Agents, Guardrails, and Model Evaluation. Amazon Q adds purpose-built assistants (Q Business, Q Developer) layered on top of foundation-model capabilities.

Choosing the Right Tier

  • Start with Tier 1 if a pre-built API already solves the task.
  • Move to Tier 3 if the task is generative (content creation, summarization, Q&A, code).
  • Move to Tier 2 only if neither pre-built APIs nor foundation models fit — you need a bespoke custom model on your own data.

Key Numbers and Limits to Memorize

AIF-C01 rewards a handful of canonical numbers that contextualize the AI and ML Core Concepts vocabulary.

AIF-C01 cheat numbers for AI and ML Core Concepts:

  • 4 — tiers in the AI → ML → DL → GenAI hierarchy
  • 7 — core vocabulary terms (model, training, inference, feature, label, parameter, hyperparameter)
  • 3 — dataset splits required for honest evaluation (train / validation / test)
  • 60 / 20 / 20 — a common (not mandatory) train / validation / test ratio
  • 4 — main use-case patterns (classification, regression, clustering, generation)
  • 4 — classification evaluation metrics commonly tested (accuracy, precision, recall, F1)
  • 7 — dimensions of data quality (accuracy, completeness, consistency, timeliness, validity, uniqueness, relevance)
  • 90 minutes — AIF-C01 exam duration
  • 65 — total AIF-C01 questions (50 scored + 15 unscored)
  • 700 / 1000 — AIF-C01 passing score
  • USD 100 — AIF-C01 exam fee
  • 3 years — validity period before AIF-C01 recertification

Source ↗

Common Exam Traps: AI/ML/DL Conflation and AGI Misconceptions

AIF-C01 aggressively exploits five recurring trap patterns tied to AI and ML Core Concepts.

Trap 1: AI ≡ ML ≡ Deep Learning ≡ GenAI

The most common trap. Answer choices use the four terms interchangeably and expect you to notice. Only one placement is correct per scenario. Use the nested-circles mental model to eliminate wrong options.

Trap 2: Parameters vs Hyperparameters Swapped

Answer choices flip the definitions. Parameters are learned during training; hyperparameters are set by humans before training. If an option claims "learning rate is a parameter" it is wrong — learning rate is a hyperparameter.

Trap 3: Training vs Inference Cost and Latency Confusion

Scenarios that conflate the two are designed to punish candidates who skim. Training is expensive, one-shot, offline. Inference is cheap-per-call, continuous, latency-sensitive. "Reduce the cost of making predictions on live traffic" points at inference optimization (smaller model, Inferentia, serverless), not training infrastructure.

Trap 4: Accuracy Good Enough on Imbalanced Data

The 99.5%-accurate fraud detector that catches zero fraud. If a question highlights imbalanced classes, accuracy is the wrong metric; precision / recall / F1 are the right lens.

Trap 5: AGI in Answer Choices

As covered earlier — no deployed AWS service is AGI. Any answer claiming AGI capability is wrong.

The AI/ML/DL/GenAI conflation trap is the single highest-frequency AIF-C01 mistake.

When you see four answer choices like:

  • (A) The system uses AI.
  • (B) The system uses Machine Learning.
  • (C) The system uses Deep Learning.
  • (D) The system uses Generative AI.

All four can be simultaneously true if the scenario describes GenAI (which is deep learning, is ML, is AI). The exam wants the most specific correct answer. Apply the "innermost correct circle" rule: pick the tightest label that still holds. A scenario mentioning foundation models and content creation should resolve to (D) GenAI, not (A) AI, even though (A) is technically true. Source ↗

Distinction Note: AIF-C01 Recognition vs AIP-C01 Build Depth

AIF-C01 is positioned as a foundational AWS certification for business stakeholders, project managers, solution architects, and developers who interact with AI workloads. It tests recognition-level competence with the AI and ML Core Concepts vocabulary — can you identify, categorize, and match scenarios to the right concept?

AIP-C01 (the MLA / AI Engineer Associate track) tests build-level competence — can you implement, optimize, and operate ML and GenAI workloads on AWS? The same concepts appear, but with deeper implementation detail: gradient computation, specific hyperparameter search strategies, SageMaker Pipelines authoring, Bedrock Agents programming, fine-tuning job construction.

What AIF-C01 Expects of You

  • Recognize the AI/ML/DL/GenAI hierarchy.
  • Identify model, training, inference, feature, label, parameter, hyperparameter.
  • Name layers, weights, and activation functions at a conceptual level.
  • Match a scenario to classification / regression / clustering / generation.
  • Choose the right evaluation metric given business context.
  • Spot data-quality problems.
  • Pick the right AWS tier (AI API vs SageMaker vs Bedrock).

What AIF-C01 Does NOT Expect of You

  • Code a backpropagation pass or a gradient descent loop.
  • Configure SageMaker distributed training scripts.
  • Author a Bedrock Agents toolchain in Python.
  • Tune a foundation model with LoRA.
  • Derive the mathematical form of cross-entropy loss.

If you catch yourself studying any of the "does NOT expect" items, you have crossed into AIP-C01 territory. Redirect to AIF-C01 depth and move on.

Practice Anchors: Task 1.1 Concept Recognition Question Templates

AIF-C01 practice questions tied to AI and ML Core Concepts cluster into five shapes. Detailed questions with full explanations live in the ExamHub question bank.

Template A: Hierarchy Placement

A company builds a chatbot powered by Claude on Amazon Bedrock that generates marketing copy from prompts. Which of the following best describes the technology category? Correct answer: Generative AI. Distractors claim "traditional ML" or "narrow AI in general" (too vague).

Template B: Vocabulary Match

A data scientist reports that tuning the learning rate and the number of training epochs consumed most of a project's development time. Which ML concept describes these values? Correct answer: hyperparameters. Distractor claims "parameters" — wrong because parameters are learned, not chosen.

Template C: Metric Selection

A hospital builds a model to flag possible tumors for radiologist review. Missing a real tumor is far more damaging than flagging a false one. Which metric should optimization prioritize? Correct answer: recall. Distractors claim "precision" (would minimize false alarms at the cost of missing tumors) or "accuracy" (useless on imbalanced data).

Template D: Data Split Purpose

A team trains a model that scores 99% on its training data and 65% on data held back for evaluation. What is the most likely explanation? Correct answer: overfitting (the model memorized training data). This is the canonical training-vs-validation-gap signal.

Template E: Use-Case Pattern Match

An e-commerce company wants to group visitors into segments based on browsing behavior, without pre-defined categories. Which ML pattern applies? Correct answer: clustering (unsupervised). Distractors claim classification (requires labels) or regression (requires numeric target).

AI and ML Core Concepts Frequently Asked Questions (FAQ)

What is the difference between AI, ML, and Deep Learning for AIF-C01?

AI is the outer umbrella of any intelligent-seeming system. ML is the subset of AI where systems learn from data instead of following hand-coded rules. Deep Learning is the subset of ML that uses multi-layer neural networks and excels on images, speech, and language. Generative AI is a capability stripe powered by deep-learning-based foundation models that produces new content. AIF-C01 expects you to place any given scenario in the smallest correct circle and recognize when terms are being used interchangeably as a trap.

What is the difference between a parameter and a hyperparameter?

A parameter is a numeric weight learned by the algorithm during training — billions of them inside a foundation model. A hyperparameter is a value chosen by humans before training starts: learning rate, number of epochs, batch size, number of hidden layers, dropout rate. Parameters are the output of training; hyperparameters are the input to training. Swapping these terms is one of the most common AIF-C01 trap patterns, so drill the distinction until reflexive.

When should I use precision vs recall vs F1 vs accuracy on the AIF-C01 exam?

Use accuracy only when classes are balanced and both error types are equally costly. Use precision when false positives are expensive (blocking a legitimate transaction, diagnosing a healthy patient). Use recall when false negatives are expensive (missing an actual cancer, failing to flag real fraud). Use F1 when you need a single balanced metric on imbalanced data. Scenario wording almost always telegraphs the answer: "don't miss any" → recall; "don't falsely flag" → precision; "imbalanced and both matter" → F1.

Do I need to understand neural network math for AIF-C01?

No. AIF-C01 tests neural network basics at the recognition level only. You should know that a network has an input layer, hidden layers, and an output layer; that neurons have weights and biases learned during training; that activation functions (ReLU, Sigmoid, Softmax, Tanh) introduce non-linearity; and that training uses forward passes plus backpropagation while inference uses only forward passes. You will not be asked to derive gradients, compute matrix multiplications, or choose a specific activation function based on mathematical properties. That depth belongs to the AIP-C01 build-level exam.

What is the difference between batch inference and real-time inference on AWS?

Real-time inference serves predictions one request at a time over HTTPS with millisecond latency — use it when a user or application is waiting. AWS surfaces: SageMaker real-time endpoints, Bedrock InvokeModel. Batch inference runs predictions on a whole dataset asynchronously and writes results to storage — use it for nightly scoring, offline reports, or large jobs where latency is not critical. AWS surfaces: SageMaker Batch Transform, Bedrock Batch Inference. Cost profiles differ: real-time endpoints pay while provisioned even when idle; batch jobs pay only during execution.

Why is data quality emphasized so heavily in AIF-C01?

Because no model architecture, no hyperparameter search, and no fine-tuning budget can overcome systematically flawed training data. The seven dimensions of data quality — accuracy, completeness, consistency, timeliness, validity, uniqueness, relevance — directly determine the ceiling of model performance. AIF-C01 ties data quality to responsible AI (Domain 4) because data bias turns into model bias which turns into unfair outcomes at production scale. "Garbage in, garbage out" is not a slogan; it is the empirical backbone of every real-world ML outcome.

How does AIF-C01 scope differ from AIP-C01 scope for the same concepts?

AIF-C01 is a foundational, recognition-level certification: can you identify, categorize, and match scenarios to the right AI/ML concept or AWS service? AIP-C01 is an associate-level, build-level certification: can you implement, tune, and operate those systems? The same vocabulary appears in both, but AIF-C01 stops at "name the concept" while AIP-C01 pushes into "author the code, choose the hyperparameters, pick the deployment pattern." If you find yourself studying gradient derivations, LoRA fine-tuning config, or SageMaker Pipelines SDK, you have drifted into AIP-C01 territory.

Is Generative AI on AIF-C01 a separate topic from core ML concepts?

Generative AI has its own domain weight (24%, Domain 2) on AIF-C01, but it sits on top of core ML concepts and cannot be understood without them. Every generative model is a deep-learning model; every deep-learning model is an ML model; every ML model relies on features, parameters, and training processes. The AI and ML Core Concepts topic is the foundation that makes the generative AI domain intelligible. Expect cross-domain questions that reward candidates who internalized the core vocabulary first.

What data split ratio should I use and does AIF-C01 test specific percentages?

A common split is 60% training / 20% validation / 20% test, or 70% / 15% / 15%, or 80% / 10% / 10%. AIF-C01 does not test exact percentages — it tests the purpose of each split. Training data shapes the model's parameters. Validation data guides hyperparameter choices during development. Test data provides the final, honest evaluation after all development is complete. If an answer choice suggests training and testing on the same data, it is wrong.

Do foundation models eliminate the need to understand core ML concepts?

No — the opposite. Foundation models amplify the importance of core concepts because their scale means data quality, evaluation rigor, and metric selection translate into vastly larger downstream consequences. A foundation model trained on biased text will propagate that bias to millions of downstream users. A RAG pipeline evaluated only on accuracy (instead of faithfulness + context precision + context recall) will miss the most critical failure modes. AIF-C01 deliberately layers generative AI on top of core concepts precisely because the risks scale with model size, and a candidate who skipped fundamentals cannot reason about those risks responsibly.

Further Reading

Related ExamHub topics: Supervised, Unsupervised, and Reinforcement Learning, Overfitting, Bias, and Variance, ML Development Lifecycle, Generative AI Concepts, Foundation Models and LLMs.

Official sources