**Batch gradient descent:**

One way to minimize that function and find the optimal θ is using the **gradient descent**

The gradient descent is an iterative algorithm that will get closer to the optimal parameters with time.

**First: **we start some random initial θ

**Second: **we repeatedly perform this iteration: