# Polynomials in Python#

• KEYWORDS: scipy.optimize.fsolve, numpy.roots, numpy.polyder, numpy.polyval, numpy.polyint, numpy.poly1d

## Special nonlinear systems - polynomials#

Polynomials are a special class of nonlinear algebraic equations that are especially easy to solve. A polynomial is linear in the coefficients in front of the variable. If we consider the following $$n^{th}$$ order polynomial:

$$p_0 x^n + p_1 x^{(n-1)} + ... + p_{n-1} x + p_n = 0$$

Letâ€™s be specific:

$$x^2 + 8x + 16 = 0$$

We express this as [1, 8, 16].

import numpy as np

p = [1, 8, 16]
r = np.roots(p)
r

array([-4., -4.])


Note we get all the roots. We can check that with the numpy.polyval command.

np.polyval(p, r)

array([0., 0.])

x = np.linspace(-10, 0)
y = np.polyval(p, x)

import matplotlib.pyplot as plt
plt.plot(x, y);


We can also use this to plot a polynomial.

import numpy as np

x = np.linspace(-5, -3)
y = np.polyval(p, x)

import matplotlib.pyplot as plt
plt.plot(x, y)
plt.xlabel('x')
plt.ylabel('y');


Why is this so convenient?

### Cubic equations of state#

There are applications of polynomials in thermodynamics. The van der waal equation is a cubic polynomial $$f(V) = V^3 - \frac{p n b + n R T}{p} V^2 + \frac{n^2 a}{p}V - \frac{n^3 a b}{p} = 0$$, where $$a$$ and $$b$$ are constants, $$p$$ is the pressure, $$R$$ is the gas constant, $$T$$ is an absolute temperature and $$n$$ is the number of moles. The roots of this equation tell you the volume of the gas at those conditions.

# numerical values of the constants
a = 3.49e4
b = 1.45
p = 679.7   # pressure in psi
T = 683     # T in Rankine
n = 1.136   # lb-moles
R = 10.73   # ft^3 * psi / R / lb-mol

ppar = [1.0,                           # V**3
-(p * n * b + n * R * T) / p,  # V**2
n**2 * a / p,                  # V
-n**3 * a * b / p]             # constant

print(np.roots(ppar))

R = np.roots(ppar)
R

[5.09432376+0.j         4.4006681 +1.43502848j 4.4006681 -1.43502848j]

array([5.09432376+0.j        , 4.4006681 +1.43502848j,
4.4006681 -1.43502848j])

print(f'V = {R[0]:1.2f}')

V = 5.09+0.00j

print(f'V = {R[0].real:1.2f}')

V = 5.09

R[0].real, R[0].imag

(np.float64(5.0943237645545985), np.float64(0.0))

float(R[0])

/tmp/ipykernel_2368/406724879.py:1: ComplexWarning: Casting complex values to real discards the imaginary part
float(R[0])

5.0943237645545985


Note that only one root is real (and even then, we have to interpret 0.j as not being imaginary. Also, in a cubic polynomial, there can only be two imaginary roots). In this case that means there is only one phase present.

### Other useful things to remember about polynomials#

You can easily get the parameters of the derivative of the polynomial with numpy.polyder.

p = [1, 8, 16]

pd = np.polyder(p)
pd

array([2, 8])


You can use these with numpy.polyval to compute the derivative at different points.

np.polyval(pd, [0, 1, 2])

array([ 8, 10, 12])


You can also get the coefficients of the integral of the polynomial. The integration constant is assumed to be 0 by default.

pint = np.polyint(p)
pint

array([ 0.33333333,  4.        , 16.        ,  0.        ])


You can use this to compute definite integrals, e.g. from x=1 to x=2:

np.polyval(pint, 2) - np.polyval(pint, 1)

np.float64(30.33333333333334)

X = np.linspace(1, 2, 100)
Y = np.polyval(p, X)
np.trapz(Y, X)

/tmp/ipykernel_2368/174692805.py:3: DeprecationWarning: trapz is deprecated. Use trapezoid instead, or one of the numerical integration functions in scipy.integrate.
np.trapz(Y, X)

np.float64(30.333350338400844)

def integrand(x):
return np.polyval(p, x)

from scipy.integrate import quad

(30.333333333333332, 3.367676508029641e-13)

%%timeit
def integrand(x):
return np.polyval(p, x)

from scipy.integrate import quad

130 Î¼s Â± 394 ns per loop (mean Â± std. dev. of 7 runs, 10,000 loops each)

%%timeit
def integrand(x):
return x**2 + 8 * x + 16

from scipy.integrate import quad

6.4 Î¼s Â± 17.4 ns per loop (mean Â± std. dev. of 7 runs, 100,000 loops each)


exercise Use another method to confirm the result above.

Finally, the syntax np.polyval(pint, 2) can be a little tedious. You can create a function with numpy.poly1d using the array of coefficients. Conveniently, you can use the function in the roots, polyder and polyint commands!

p = np.poly1d(pint)
print(p)
p(2) - p(1)

        3     2
0.3333 x + 4 x + 16 x

np.float64(30.33333333333334)

print(integrand)

<function integrand at 0x7fd8a21c7380>

p

poly1d([ 0.33333333,  4.        , 16.        ,  0.        ])

print(p)

        3     2
0.3333 x + 4 x + 16 x

np.roots(p)

array([-6.+3.46410162j, -6.-3.46410162j,  0.+0.j        ])

p([np.linspace(0, 1, 5)])

array([[ 0.        ,  4.25520833,  9.04166667, 14.390625  , 20.33333333]])

p

poly1d([ 0.33333333,  4.        , 16.        ,  0.        ])


## Systems of nonlinear equations#

Analogously to systems of ordinary differential equations, with systems of nonlinear equations we define functions that will return a zero for each equation in the system. Then we have to pass an initial guess for each variable to fsolve, and it will return an array of values, one for each variable.

It is considerably more difficult to visualize systems of nonlinear equations. With two equations and two unknowns it is sometimes easy to plot solutions, but not always.

\begin{align} y &=& x^2 \ y &=& 8 - x^2 \end{align}

One approach to visualizing this is to plot the two curves.

import numpy as np
import matplotlib.pyplot as plt

x = np.linspace(-4, 4)

y1 = x**2
y2 = 8 - x**2

plt.plot(x, y1, x, y2)
plt.xlabel('x')
plt.ylabel('y')
plt.legend(['y1', 'y2']);


You can see that on this domain, there is one place where the two curves intersect near the point (2, 5), which is a solution point. At this point there is one (x, y) pair that is a solution to both equations.

from scipy.optimize import fsolve

def objective(X):
x, y = X
z1 = y - x**2      # y = x**2
z2 = y - 8 + x**2  # y = 8 - x**2
return np.array([z1, z2])

guess = [-2, -5]
fsolve(objective, guess)

array([-2.,  4.])


It is not always easy to solve for one variable in terms of the other though. In that case, we can resort to an alternate graphical approach where we evaluate the objective function over a range of the variables, and look for regions where they overlap.

Consider the solution to these equations (adapted from https://www.mathworks.com/help/optim/ug/fsolve.html):

$$e^{-e^{-(x_1 + x_2)}} = x_2 (1 + x_1^2)$$

and

$$x_1 \cos(x_2) + x_2 \sin(x_1) = 1/2$$

It is not possible to solve either one for one variable in terms of the other. So instead, we will compute the objective function for a range of $$x_1, x_2$$ values, and then use a contour plot of each equation to see where there might be a solution.

The key to this visualization is where we draw the contours. A good choice is to highlight only the part of the solutions that bracket zero. Then we can see where they intersect, because there is probably a solution in that neighborhood.

def objective(X):
x1, x2 = X
z1 = np.exp(-np.exp(-(x1 + x2))) - x2 * (1 + x1**2)
z2 = x1 * np.cos(x2) + x2 * np.sin(x1) - 0.5
return np.array([z1, z2])

x1 = np.linspace(-1, 1)
x2 = np.linspace(-1, 1)

X1, X2 = np.meshgrid(x1, x2)

Z1, Z2 = objective([X1, X2])
plt.contourf(X1, X2, Z1, levels=10)
plt.colorbar();

plt.contour(X1, X2, Z1, levels=np.linspace(-0.01, 0.01, 100))
plt.contour(X1, X2, Z2, levels=np.linspace(-0.01, 0.01, 100))
plt.xlabel('$x_1$')
plt.ylabel('$x_2$')
plt.colorbar();


There is an intersection near $$x_1=0.4$$, and \$x_2 = 0.6. We can use that as an initial guess.

ans = fsolve(objective, [0.4, 0.6])  # note we do not need ans, because ans will have two values in it.
ans, objective(ans)

(array([0.35324662, 0.60608174]), array([-2.22044605e-16,  1.11022302e-16]))

for i in range(100):
guess = [(np.random.random() - 0.5) * 20, (np.random.random() - 0.5) * 20]
ans, info, ier, msg  = fsolve(objective, guess, full_output=True)
if ier != 1:
print(f'Failed at guess {guess}.\n{msg}')
continue
plt.plot([guess[0], ans[0]], [guess[1], ans[1]])

Failed at guess [-5.135320274707622, 5.801155693222169].
The iteration is not making good progress, as measured by the
improvement from the last ten iterations.

/tmp/ipykernel_2368/476845304.py:3: RuntimeWarning: overflow encountered in exp
z1 = np.exp(-np.exp(-(x1 + x2))) - x2 * (1 + x1**2)

all_answers = []
for i in range(100):
guess = [(np.random.random() - 0.5) * 20, (np.random.random() - 0.5) * 20]
ans, info, ier, msg  = fsolve(objective, guess, full_output=True)
if ier != 1:
print(f'Failed at guess {guess}.\n{msg}')
continue

array([6.16007917e-12, 1.27985616e-11])

AA = np.array(all_answers)
plt.plot(AA[:, 0], AA[:, 1], 'b.')
plt.xlim([0, 1])
plt.ylim([0, 1]);


This shows the solution, and that the objective is practically equal to zero at that point.

You can see that trying to do this in more than 2 dimensions can quickly get difficult to visualize!

## Summary#

• We learned about a special class of nonlinear functions that are polynomials, and a series of useful functions to manipulate them.

• We learned that you can use fsolve to find solutions to coupled non-linear algebraic equations.

• Next time, we will apply this to solving a nonlinear boundary value differential equation.