Yadu Chittampalli
\[ x_{n+1}^{*}=x_n+hf\left( t_n,x_n,y_n\right) \]
\[ y_{n+1}^{*}=y_n+hg\left( t_n,x_n,y_n\right) \]
\[ t_{n+1}=t_n+h \]
\[ x_{n+1}=x_n+0.5h\left[f\left( t_n,x_n,y_n\right) +f\left( t_{n+1},x_{n+1}^{*},y_{n+1}^{*}\right) \right] \]
\[ y_{n+1}=y_n+0.5h\left[g\left( t_n,x_n,y_n\right) +g\left(t_{n+1},x_{n+1}^{*},y_{n+1}^{*}\right) \right] \]
\[ \frac{dx}{dt} = f\left( t_n,x_n,y_n\right) = x - xy - 0.75y \] \[ \frac{dy}{dt} = g\left( t_n,x_n,y_n\right) = xy - y - 0.75x \]
All trajectories are diagonal straight lines.
Disregard the trajectory for step size 1.
x decreases as y increases.
The trajectory almost looks like that of the Euler's method. Only difference is that there is no steep drop to begin with.
The improved Euler's method displays divergence for step sizes less than 1.
The trajectories for the improved Euler's method are a straight diagonal line, whereas for the Euler's method, the trajectories are more curved.
\[ x_{k+1}=x_k - \frac{f^\prime \left( x_k \right)}{f^{\prime \prime} \left( x_k \right) } \]
\[ \epsilon = tolerance \]
\[ \mid x_{k+1} - x_k \mid < \epsilon \]
\[ \mid f^\prime \left( x_k \right) \mid < \epsilon \]
Keep on reiterating until one of the two conditions listed above is met.
\[ \left\{ \begin{array}{ll} 4x^{3} - 3x^{4} & x > 0 \\ 4x^{3} + 3x^{4} & x < 0 \\ \end{array} \right. \]
\[ \left\{ \begin{array}{ll} x^{2} + 2x & 3\leq x\leq 6 \\ \end{array} \right. \]
For x = 0.6, it does not matter what tolerance we pick because each of the tolerances yield the same trajectory of convergence.
For each tolerance convergence occurs at approximately 1.
The higher the tolerance, the lesser number of iterations required for convergence (shorter trajectory).
For the piecewise function where x1 = 0.6 and for the quadratic function it does not matter what tolerance we pick because convergence occurs at approximately the same number of iterations (7-9 for the piecewise function and 2-3 for the quadratic).
Generally, as the initial value of x increases, the number of iterations required for convergence increases.
However, when x is initially -1 only one iteration is required for convergence.
For all initial values of x greater than 1, convergence occurs at approximately 1.
As the initial value of x is increased, the length of the trajectory increases and the slope becomes initially steeper.
There are uniquely different trajectories for many of these initial values of x.
When x is initially less than or equal to 0.6 convergence occurs when x reaches 0.
When x is initially greater than 0.6, convergence occurs when x reaches 1.
No particular trend shown in the graph.
However, if the number of iterations required for the initial value of 0.5 was 4 or greater, the histogram would look like a bell-shaped curve or in other words a normal distribution.