Module 02

Opening the Black Box:
Generative AI Mechanics

Generative AI represents a fundamental shift in computing. Unlike traditional software which follows explicit rules, Generative AI operates on probability. In this module, we peel back the layers of the LLM.

1

The Physics of GenAI

How does ChatGPT actually "write"? It's not magic; it's math. At its core, it is predicting the next most likely chunk of text (token) based on everything it has seen before. Explore the interactive diagram below to see how Transformers, Embeddings, and Diffusion work.

GenAI Architecture

The Physics of Generative AI

How models understand text, generate images, and capture meaning.

Thecatsatonthe___
Self-Attention Layer
Prediction: mat (98%)

Transformer Architecture (LLMs)

The engine behind GPT, Claude, and Llama. It uses 'Self-Attention' to understand the relationship between every word in a sequence simultaneously, not just sequentially.

Tokens

The basic units of text (words or sub-words). LLMs process tokens, not characters.

Context Window

The amount of text the model can 'see' at once (short-term memory).

Temperature

A parameter controlling randomness. High = Creative, Low = Deterministic.

2

Probabilistic vs. Deterministic

This is the most important concept for business leaders. Traditional software is rigid but accurate. Generative AI is flexible but can be wrong. Understanding this trade-off is key to choosing the right use cases.

Capabilities vs. Limitations

Navigating the shift from Deterministic Software to Probabilistic AI.

Dimension
Traditional Software
Generative AI
Core Logic
Deterministic
Input A always equals Output B. 100% predictable.
Probabilistic
Input A generates Output B... or B-Prime. Creative but variable.
Flexibility
Rigid
Built for specific pre-defined tasks.
Adaptable
Zero-shot learning. Can handle tasks it wasn't explicitly trained for.
Failure Mode
Crash / Error
System throws an exception and stops.
Hallucination
System confidently invents an incorrect answer.
  • Adaptability

    Unlike traditional rule-based systems, LLMs handle edge cases and "fuzzy" inputs naturally. You don't need to code every possible user intent.

  • Creation & Synthesis

    Can generate code, summaries, or marketing copy from scratch, massively boosting productivity for knowledge workers.

3

The Black Box Problem

Why did the AI say that? With traditional "Glass Box" models (like Excel formulas), we can trace the logic. With Deep Learning ("Black Box"), the logic is hidden inside billions of parameters. This has huge implications for compliance and auditing.

Transparency vs. Opacity

Not all AI is created equal. Understanding the "Black Box" problem.

Input: Image
Hidden Layers
Output: "Cat"
Deep Neural Networks, LLMs

Black Box Models

Opaque & Complex

The decision logic is hidden within millions (or billions) of parameters. You see inputs and outputs, but the 'Why' is mathematical noise.

Pros
  • High accuracy
  • Handles unstructured data
Cons
  • Hard to trust
  • Difficult to explain failure
4

Batch vs. Real-Time Inference

"Inference" is the act of using a model to generate a prediction. Depending on your business need, you might run this overnight for millions of records (Batch) or instantly when a user clicks a button (Real-Time).

Inference Strategies

How AI makes predictions: Batch vs. Real-Time.

10k Rows
Data Lake
Nightly Schedule
Batch Job

Batch Processing

Processing data in large groups (chunks) at scheduled intervals (e.g., every night). Ideal for non-urgent tasks where throughput matters more than speed.

Cost
Low
Latency
High (Minutes)
Use Case: Weekly Sales Forecasting, Sentiment Analysis of Archive.

Ready to apply this?

Now that you understand the mechanics, let's look at high-value business use cases.

Go to Module 3: Strategy & ROI