DEV Community

Amritanshu Dash
Amritanshu Dash

Posted on

Linear Algebra for AI — Part 2: Deep Understanding of Vectors

Vectors are the language of machine learning. Humans think in words, but neural networks think only in vectors. If you understand vectors deeply, everything in AI becomes easier.

What is a Vector?

A vector is simply a list of numbers.
But these numbers represent two things:
• Magnitude
• Direction

The easiest example is a push force.
How strong you push = magnitude
Which way you push = direction

We will later show this diagram:


Column Vectors: Your Robot’s Secret Instructions in Math and AI

Imagine you’re in a magic treasure hunt. You have a map, and the map tells you exactly how far to go in each direction: forward, left, and up. Each instruction is crucial — missing or swapping one changes where you end up.

That’s exactly what a column vector does: it stores instructions for movement in multiple dimensions.

Column Vector Example
Mathematically, we write it like this:

3
v = −2
0.5

• 3 → move 3 units along the x-axis (forward)
• –2 → move 2 units along the y-axis (left/back)
• 0.5 → move 0.5 units along the z-axis (up)

💡 Memory Trick: “X forward, Y left, Z up — XYZ = exactly my path!”

A Vector in Machine Learning

[0.0123, -0.9431, 0.8567, -0.2210, …, 0.3312]

This is what embedding vectors look like.
In AI, vectors usually have 768 to 4096 numbers.

Physics Vectors vs Machine Learning Vectors

Here is the normal-block version of the table content:

Physics / Math:
• Represents force, velocity, movement
• Usually 2D or 3D
• Always has direction

Machine Learning:
• Represents meaning of a word, image, user, or sentence
• Hundreds or thousands of dimensions
• Meaning exists in a “semantic direction”

🛠️ Operations on Vectors — The Secret Tools of AI

Imagine vectors as little robots carrying instructions. Each robot has a strength (magnitude) and a direction (where it’s heading). Now, just like you can combine Lego blocks to build a bigger toy, vector operations let us combine and adjust these robots to do amazing things:

• Addition → putting robots together to work as a team.
• Scaling (Multiplication) → telling a robot to move faster, slower, or even backward.

These operations are the magic behind AI:
• Merging user preferences to suggest the perfect movie 🎬
• Combining forces to move a robot arm 🤖
• Adjusting image features for computer vision 🖼️

Think of it like this: each operation is a tool in your AI toolbox. The better you understand them, the more creative and powerful your AI creations can become.

1.Vector Addition

Two vectors are added element-wise:

Example:
(3, 1) + (-1, 4) = (2, 5)

Intuition:
Combining movements, combining signals, or merging features in ML.

2.Scalar Multiplication

Scaling a vector means multiplying every component:

Example:
5 × (2, 3) = (10, 15) → same direction, larger
-1 × (2, 3) = (-2, -3) → flipped direction

In AI, scaling adjusts magnitudes, gradients, and features.

Real AI Example: Netflix

• Your watch history → vector
• Movie metadata → vector
• Compare both vectors → similarity score
• Add bias vectors (time of day, mood, past behavior)
• Final recommendation vector → top results shown to you

Everything is vector math.

🧪 Vector Pooling — Combining the Power of Many Vectors

Imagine you’re a conductor of an orchestra. Each musician plays their own note — some loud, some soft. Now, if you want to capture the overall sound of the orchestra in one go, you need to combine all the individual notes into a single “summary sound”.

In AI, we face the same problem:
• A sentence has many word vectors.
• An image has many feature vectors.
• Audio has many signal vectors.

We can’t feed all of them individually into the next layer of a neural network. So, we use vector pooling: a technique that combines multiple vectors into a single vector while keeping important information intact.

Think of vector pooling as summing up all the little instructions into one “super instruction” that the AI can understand.


🔹 Types of Vector Pooling

Let’s see the three main types with realistic mini-examples:

1.Sum Pooling — Total Strength

We add all vectors component-wise, keeping the total magnitude.

Example:

v1 = [1, 2, 3]
v2 = [2, 0, 1]
v3 = [0, 1, 2]

Sum Pooling:

sum = [v1[0]+v2[0]+v3[0], v1[1]+v2[1]+v3[1], v1[2]+v2[2]+v3[2]]
sum = [1+2+0, 2+0+1, 3+1+2]
sum = [3, 3, 6]

Use case: Capturing total sentiment in a sentence, e.g., very excited + happy + surprised → overall excitement level.

2.Mean Pooling — Average Meaning

We average all vectors component-wise, focusing on the “typical” signal.

mean = [sum[0]/3, sum[1]/3, sum[2]/3]
mean = [3/3, 3/3, 6/3]
mean = [1, 1, 2]

Use case: Getting the overall meaning of a sentence in NLP or the average style of an image feature map.

3.Max Pooling — The Strongest Signal

We pick the maximum value component-wise, keeping the most dominant signal.

max_pool = [max(v1[0], v2[0], v3[0]),
max(v1[1], v2[1], v3[1]),
max(v1[2], v2[2], v3[2])]
max_pool = [2, 2, 3]

Use case: Detecting the strongest feature, e.g., the brightest pixel in a CNN, the loudest sound in audio, or the most important word in a sentence.

💡 Quick Analogy for Memory

• Sum Pooling: Like counting all your coins → total wealth.
• Mean Pooling: Like averaging your grades → overall performance.
• Max Pooling: Like picking your tallest friend in the class → the standout feature.

Conclusion

Vectors are more than numbers — they form a geometric world where neural networks “think.”
Once you understand vector addition, scaling, and pooling, you understand the foundation of embeddings, attention, and deep learning.

Top comments (0)