Derivative (Rate of Change) | fโฒ(x)=limhโ0โhf(x+h)โf(x)โ | For f(x)=x3, fโฒ(x)=3x2 | Measures sensitivity of outputs to inputs, marginal effects, rates of change |
Average vs Instantaneous Rate | Avg: x2โโx1โf(x2โ)โf(x1โ)โ; Inst: limit form above | Slope of secant vs slope of tangent | Distinguishes between overall vs. pointwise change |
Notation | fโฒ(x), dxdyโ, dxdโf(x) | yโฒ | Different notations useful in various contexts (Leibniz, Lagrange) |
Power Rule | dxdโ(xn)=nxnโ1 | (x3)โฒ=3x2, (xโ)โฒ=2xโ1โ | Core tool for polynomial/exponential changes |
Constant Multiple Rule | (cf(x))โฒ=cfโฒ(x) | (5x3)โฒ=15x2 | Simplifies scaling derivatives in models |
Sum/Difference Rule | (f(x)ยฑg(x))โฒ=fโฒ(x)ยฑgโฒ(x) | (4x2+7xโ2)โฒ=8x+7 | Enables decomposition of model functions |
Product Rule | (fg)โฒ=fโฒg+fgโฒ | (x2ex)โฒ=ex(2x+x2) | Needed for features interacting multiplicatively |
Quotient Rule | (gfโ)โฒ=g2fโฒgโfgโฒโ | (x+1x2โ)โฒ=(x+1)2x2+2xโ | Used when variables appear in ratios |
Chain Rule | (f(g(x)))โฒ=fโฒ(g(x))gโฒ(x) | ((x2+3x)5)โฒ=5(x2+3x)4(2x+3) | Central to backpropagation in neural networks |
Common Functions | (ex)โฒ=ex, (lnx)โฒ=x1โ, (sinx)โฒ=cosx, (cosx)โฒ=โsinx | Trigonometric and exponential forms | Key in interpreting exponential growth/periodicity |
Partial Derivatives | โxโfโ, hold others constant | For f(x,y)=x2y+3xy3, โxโfโ=2xy+3y3 | Measures effect of single feature |
Gradient | โf=(โx1โโfโ,โฆ,โxnโโfโ)T | For earlier f(x,y): (2xy+3y3,x2+9xy2) | Fundamental to gradient descent optimization |
Second-order Derivatives | Pure: โx2โ2fโ; Mixed: โxโyโ2fโ | Concavity/curvature along axes | Determines shape and inflection, used in convexity analysis |
Hessian Matrix | Matrix of all 2nd-order partials | For multivariable f(x1โ,โฆ,xnโ) โ Hessian H | Used in Newton's method, uncertainty estimation |
Applications | Optimization, sensitivity, feature importance, backpropagation, marginal analysis | Loss minimization, elasticity, customer growth rates | Core to machine learning training and interpretability |