Derivative Manipulation
We can think of derivative formulas as falling into two categories: (1) the specific, and (2) the general. A specific formula provides a way to differentiate a specific example. For example, has the derivative
In contrast, general formulas are those that apply to all functions. For example:
In the first formula above: The derivative of the sum of two functions is the sum of the derivatives of the two functions. And in the second: The derivative of a constant multiplied by a function is the constant multiplied by the derivative of the function. Both specific and general formulas are needed for polynomials.
The Sum Rule
What is the derivative of To compute this derivative, we must apply the sum rule for derivatives. To do so, let's derive the general rule. The function can be written as the sum of two functions: Suppose and Then That said, we apply the definition of a derivative:
Now we substitute and
Rearranging:
And applying the sum law for limits and the definition of a derivative:
The above derivation yields the following:
Sum Rule for Derivatives. Let and be differentiable functions. Then the derivative of the sum of and is the sum of the derivative of and the derivative of
Alternatively, given the differentiable functions and and then:
Thus, returning to our original question, what is the derivative of the derivative is simply the derivative of each term:
by the sum rule and the power rule.
The Difference Rule
Now let's consider the opposite of addition: What is the derivative of Here too we can think of as consisting of two separate functions: Given and Again we apply the definition of a derivative:
Now substitute: and
Then rearranging:
Applying the sum law for limits and the definition of a derivative:
From our analysis, we have the following rule:
Difference Rule for Derivatives. Let and be differentiable functions. The derivative of the difference of and is the difference of the derivative of and
Alternatively, given the differentiable functions and and that
Applying our newly derived rule, the derivative of is:
p'(x) = 10x^4 - 3x^2,}$ by the difference rule and power rule.
Constant Multiple Rule
What is the derivative of This may seem complex, but we know enough to handle this. To begin with, the term is just a constant. Thus, we can really write this function as: where and Let's come up with a rule for computing the derivative of
Because we can factor out constants from limits, applying the derivative definition is straightforward. First, the derivative of is the derivative Thus:
From our derivation above, we have the following rule:
Constant Multiple Rule. Given the function where is differentiable and is a constant:
Thus, the derivative of is a simple application of the constant multiple rule, difference rule, and power rule:
The Product Rule
It may be tempting to think that the derivative of a product is the product of the derivatives. This is wrong, and we've already seen why. Given the function the derivative of is However, the function can be written as If it were the case the derivative of a product is the product of the derivatives, then we would have Wrong. So what is the rule for products?
Well, let's apply the definition of a derivative. Suppose we have two functions, and both of which are differentiable. Now let's say that the function is the product of these two functions:
Now apply the definition of a derivative to the function
Let's apply the old trick of adding zero. We add and subtracting in the numerator:
Using the associative property, let's group the terms to break the quotient:
Applying the associative property again:
Now, we know that is differentiable (by assumption), so it follows that is continuous. And since is continuous, Hence:
We now have a general derivative formula, the product rule:
Product Rule. Given functions and the derivative of denoted is the sum of the derivative of the function denoted times and the derivative of the function denoted times I.e.,
For example, suppose we are given the functions and The product of these functions:
What is the derivative of We apply the product rule:
Quotient Rule
What is the derivative of For this, we need another rule. Never, ever, ever, think that This is absolutely wrong, and it is one of the most common mistakes made among calculus newcomers.
As always, let's generalize our problem, and construct a rule. Suppose and Thus, Accordingly, we want to compute Applying the quotient rule:
From the derivation above, we have the following rule:
Quotient Rule. Given functions and the derivative of denoted is provided by the following:
Accordingly, the derivative of is:
Chain Rule
In many of the previous examples, we abstracted the function's terms to arrive at a more general rule. For example, given the function we would write where and We would then arrive at a derivative from this more abstracted form. This technique results from two observations: First, every function can be expressed as the combination of smaller functions. Borrowing from computer science, we might call these smaller functions helper functions. For example, silly as it may be, the function can be written as where Or it can be written as , where Second, there are really only three ways to combine functions.
First, we can add functions. For example, the function can be written as where and The ability to add functions implies the ability to subtract functions, since addition is really just the addition of a negative. For example, the function can be written as where and
Second, we can multiply functions. The function can be written as where and The ability to multiply functions implies the ability to divide functions, since division is just multiplication with the reciprocal. The function can be written as where and
Third, we can compose, or nest, functions. For example, the function can be written as where Similarly, the function can be written as where
Borrowing again from computer science, these are the three means of combination. The remarkable aspect of functions is that every function can be written and rewritten, or expressed, with just these three means.
So far, we've used our abstraction technique for the first two means: addition and multiplication. What we haven't used it for, however, is with composition. This limitation leads to some problems. We haven't yet seen a way to apply the technique to a function like Fortunately, we can simply expand this function: then apply our familiar rules.
Expansion, however, begins falling apart once we have something like This is a polynomial that would be messy to expand. We'd sooner exhaust ourselves before seeing the polynomial's final form.
Worse, neither expansion nor our original method of generalization would help with functions like or With the rules we have thus far, it isn't exactly clear how we would compute the derivatives for these functions. All these functions are called composite functions β functions combined through composition, or nesting. The most efficient way to compute their derivatives is with a new theorem: the chain rule.
To understand the chain rule, we must first review how composite functions work. Composite functions are simply compositions of functions. Our original method of generalization is applicable to composing functions. With the composite function we generalize to Following this generalization, Now, let's compute
That was a straightforward application of the power rule. Now let's compute the derivative of
So now we have these two results:
But, what we ultimately want to find is Now, notice the way Leibniz's notation looks:
Those look an awful lot like fractions, and it's tempting to multiply them and cancel:
Let's put a pin on this observation for now, but we'll revisit it shortly. What we should do instead is think a little more carefully about what it means to compute the derivative of a nested function, say If suppose that then Looking at it this way, we can see that outputs a and that is an input for Now let's say that Viewing it this way, when we ask for the derivative of what we're really asking for is, how quickly (or slowly), does change as changes?
The answer to that question is the derivative. We know that a derivative is simply the application of a limit to the quotient rule. Thus, what we want to do is to consider the Newton quotient of the compositve function
Suppose has a derivative at and has a derivative at Then, by Newton's quotient:
This expression is a bit of an eye sore, so we'll replace some of the terms with simpler variables. Suppose:
Notice that with the above substitutions, depends on As approaches 0, tends to 0. Bearing this in mind, we can the corresponding terms and arrive at the following:
This looks very similar to the derivative but we have in the denominator and in the numerator. Recall that depends on And because depends on we have two cases to consider: and
The easier (and more common) case is when so let's deal with it it first. Suppose If then we can multiply and divide the our equation by
Examining the resulting equation, when approaches several things occur. First, the quotient:
tends to Second, (β tends towards 0β) as because and is continuous at (by our first assumptions). Thus, the quotient:
approaches as and the quotient:
approaches as Hence, we have the following rule:
The above rule, however, only applies when We must still consider the case where This is somewhat of an edge case, but because the possibility exists, we must address it. Otherwise, our preceding wouldn't apply.
First, say we have a number such that is defined. By the limit of Newton's quotient, we know that:
Encapsulating the above:
It follows then that, as approaches 0, In other words:
Multiplying both sides by we have:
We can rewrite this as:
Consider what happens when The equation is holds. But if the equation breaks down, because leaves a 0 in the denominator. We can avoid this by supposing that in which case the equation holds when since doing so would simply yield:
Of course, this is true. With this in mind, let's say and Then as tends towards 0, so too does Now, recall that the Newton quotient for the function is:
Based on our previous analysis, it follows that:
Substituting the value for we have:
Now we take the limit as as 0. Applying this limit, we see that the first term approaches Given that the limit of as or is 0, applying the limit to the second term results in:
Accordingly, we have the rule:
Formally stating this rule:
The Chain Rule. Suppose and are functions. For all in the domain of for which is differentiable at and is differentiable at the derivative of the composite function is given by:
Alternatively, if is a function of and is a function of then:
Once more we see that the chain rule applies even when This proves that the chain rule applies in general. Having proved the chain rule, we can now see that we can rest easy succumbing to our previous temptation:
This is because the chain rule can be expressed by the formula:
With the chain rule, derivatives of composite functions behave as if we could cancel For example, suppose we're asked to compute the following derivative:
We can compute this derivative quickly with the chain rule. First, we need the two terms, and The variable embodies the outer function. Thus:
Higher-order Derivatives
The mathematician Hugo Rossi famously quipped, βIn the fall of 1972, President Nixon announced that the rate of increase of inflation was decreasing. This was the first time a sitting president used the third derivative to advance his case for reelection.β What is this third derivative? It's simply a derivative of a derivative β a higher-order derivative.
Higher derivatives are denoted ordinally. I.e., the βsecond derivative,β the βthird derivative,β etc. For example, velocity is the first derivative of the position vector with respect to time, representing the change in an object's position over time. Acceleration is the second derivative of the position vector with respect to time, representing the moving object's change in velocity (e.g., how fast does this car go from 0mph to 100mph?). Jerk is the third derivative of the position vector with respect to time, representing the change in acceleration. Somewhat humorously and non-standardized: A snap is the fourth derivative representing the change in jerk. Crackle is the fifth derivative representing the change in snap. Pop is the sixth derivative representing the change in crackle.
The rules we've covered thus far equally apply to the computation of higher-order derivatives. For example, given the first derivative is The second derivative is
There are several forms of notation for higher derivatives:
The symbols and are operators we can apply to functions. Applying such operators, the derivative of the function is returned.