Feature scaling is important in machine learning when models rely on the magnitude of features. Without scaling, features with larger ranges can dominate the learning process.
For example:
age (0-100) and income (0–100,000), the model may give too much weight to income just because of its scale.Exactly. This is what happens when you don’t scale: features with larger values overshadow smaller-range features, even if they’re equally important.
There are several methods. You mentioned some:
x′=x−min(x)max(x)−min(x)x' = \frac{x - \min(x)}{\max(x) - \min(x)}
✅ Pros: Preserves relationships
❌ Cons: Sensitive to outliers