Monday, 2 June 2025

Mathematical Tools 4: How Are Complex Mathematical Expressions Like -log(P(x)) Invented

How Are Complex Mathematical Expressions Like -log(P(x)) Invented?

You may have come across expressions such as:

|log(1 / P(x))|

These don’t just appear out of nowhere — they are built purposefully. In this article, we break down how such mathematical expressions are constructed, not invented randomly. We’ll also show how you can engineer your own for specific data-driven goals.

🧭 1. They Are Designed to Serve a Purpose

These expressions are crafted to:

  • Compress large values or reduce skew
  • Normalize or transform scales
  • Penalize rare or extreme events
  • Fit into established frameworks like information theory or optimization

They are goal-driven, meaning the structure is chosen to produce certain behavior — not just mathematical decoration.

🧪 2. They Often Come from Known Theories

Let’s decode:

|log(1 / P(x))| = -log(P(x))

This is a classic concept from information theory, known as self-information or information content:

\[ I(x) = -\log P(x) \]

This expression quantifies how surprising an event is — rare events (low probability) have high information content.

Used in:

  • Shannon entropy
  • Cross-entropy loss in machine learning
  • KL divergence
  • Bayesian updating

🧠 3. Built from Design Heuristics or Objectives

Mathematicians don’t invent expressions out of thin air. They ask:

What properties do I want this expression to have?

For example:

  • Penalize low probabilities? → use -log(P(x))
  • Compress scale? → use log or sqrt
  • Ignore direction? → use |x|
  • Apply penalties exponentially? → use x^2, e^x

🔨 4. General Recipe to Construct Expressions

Step Action Example
1 Start with a base quantity P(x), x/y
2 Invert or complement 1/P(x), 1 - x
3 Apply log to compress/penalize log(1/P(x))
4 Apply absolute or square to remove sign |x|, x^2
5 Scale or combine with weight w * log(...)

🧬 5. Examples in Practice

Field Expression Purpose
Information Theory -log P(x) Information content
ML Loss -y log(p) Cross-entropy loss
Bioinformatics log(P1/P2) Likelihood ratio
NLP/IR TF * log(N/DF) TF-IDF weighting
Risk Analysis 1 / P(x) Inverse risk score

🧠 Final Thought

These expressions aren’t creative accidents. They’re engineered artifacts designed with clear intent: to model, penalize, compress, or rank something meaningfully.

Whenever you see a weird-looking formula, ask: “What is it trying to control, normalize, or emphasize?”

That question often reveals the logic behind the math.

No comments:

Post a Comment

🧠 You Only Laugh Once: Creativity and Humor in Deep Learning Community

It all started with a simple truth: Attention Is All You Need . Or at least, that’s what the transformers keep whispering at every AI confer...