# Linear Regression with Elixir, Phoenix and LiveView. Part I

Phoenix LiveView has been around for a bit, but with the release of Phoenix 1.5 it became even easier to get started with it in a new Phoenix app! Simply pass the `--live`

flag when generating a new project and off you go! ðŸš€

In this two part series we’re getting our hands dirty with a basic linear regression algorithm that allows the user to click on a plane to add datapoints and the algorithm figure out the best fitting line through these points.

## Upgrading to Phoenix 1.5.0

To kick things off, we first need to get up-to-date with our Phoenix version, if you have not already done this. You can do so by running: `mix archive.install hex phx_new`

If you rather get LiveView to work on Phoenix 1.4.x, just follow the getting started guide.

## Setting up the project

Generating a new project that will help us hit the ground running is as simple as: `mix phx.new linreg --live --no-ecto`

.

Note the two flags there:

`--no-ecto`

skips setting up all Ecto related things since we won’t be using them in this project.`--live`

ensures everything is wired up to use LiveView!

Lastly; `cd`

into the `linreg`

folder and kick-off the Phoenix server by running `mix phx.server`

. You are now greeted by the getting started page over at `http://localhost:4000`

. ðŸŽ‰

## Basic model

Now that we have a new Elixir, Phoenix and LiveView project ready to go, we can get cracking on the real stuff!

First things first, let’s define a simple model to hold our data and make our predictions with. A simple struct to hold our weights will do! Fire up your favorite editor and create a new file under `lib/linreg/`

. Letâ€™s call it `model.ex`

by lack of a better name and drop in the following code:

```
1defmodule Linreg.Model do
2 defstruct m: 0.0, b: 0.0
3
4 alias __MODULE__
5
6 def new do
7 %Model{}
8 end
9
10 def predict(%Model{m: m, b: b}, x) do
11 b + m * x
12 end
13end
```

The module starts off pretty basic. It defines a struct to keep track of two values: M and B and define a function to make predictions based off of the model. For convenience I added a factory function that returns a new empty `%Model{}`

struct.

## What is training

Without training, our model would always return 0.0. This is because the weights aren’t adjusted yet. There’s a variety of ways to let the machine learn the correct values for M and B, but basically it all boils down to:

- Let the model make a prediction,
- Calculate how far it is off (often called error, loss or cost),
- Adjust the weights accordingly and repeat. Repeat until the error is acceptably low.

## Gather training data

In order to correctly train our model, we need training data. Training data in our case means a set of X and Y value pairs. The model can learn from these values by making a prediction based on X and see how far it is off by looking at the actual Y value.

In the `lib/linreg`

folder, go ahead and make a new file called `data.ex`

and write the following code:

```
1defmodule Linreg.Data do
2 defstruct points: []
3 alias __MODULE__
4
5 def new do
6 %Data{}
7 end
8
9 def add_point(%Data{points: points} = data, x, y) do
10 %Data{data | points: [{x, y}] ++ points}
11 end
12end
```

In this module we define a struct that holds our training data. A list of X and Y coordinates called *points* and some utility functions to initialize a new `%Data{}`

struct and to add a point to an existing dataset.

## Adjusting the weights

Now that we can record training data, we can implement a training function that adjusts the weights of our model. Open `lib/linreg/model.ex`

and write the following function:

```
1def train(%Model{m: m, b: b} = model, %Data{points: points}, opts \\ []) do
2 learning_rate = Keyword.get(opts, :learning_rate, 0.01)
3
4 m_error =
5 points
6 |> Enum.map(fn {x, y} -> x * (predict(model, x) - y) end)
7 |> Enum.sum()
8 |> Kernel./(length(points))
9
10 b_error =
11 points
12 |> Enum.map(fn {x, y} -> predict(model, x) - y end)
13 |> Enum.sum()
14 |> Kernel./(length(points))
15
16 %Model{model | m: m - m_error * learning_rate, b: b - b_error * learning_rate}
17end
```

The `train/3`

function takes the current model, training data and a list of options. Then for each point in the training set, it will make a prediction of the Y value given its X value and calculate how far it is off.

It will do this for both the M and the B values. Only difference is that, for M, weâ€™re constraining the result by multiplying it with its input X value. Once we have the errors for each datapoint, we take the average.

Once we have the average error of the model, we adjust each weight by subtracting the average corresponding error multiplied by the *learning rate*.

The learning rate is a value, often between 0 and 1, which determines

how quickthe machine learns. It ensures that the adjustments to the weights are done in very small steps as to not overshoot the optimum value. This process is called gradient descent and is a very common technique within machine learning algorithms.

## Taking it for a spin

Thatâ€™s pretty much all there is to it! Now letâ€™s take it for a spin and see how it works. Fire up an IEx shell again and letâ€™s make some predictions.

```
1# Initialize new trainings data and a new model
2iex> d = Linreg.Data.new
3%Linreg.Data{points: []}
4
5iex> m = Linreg.Model.new
6%Linreg.Model{b: 0.0, m: 0.0}
7
8# Add some known data to the training set
9iex> d = Linreg.Data.add_point(d, 2, 4)
10%Linreg.Data{points: [{2, 4}]}
11
12iex> d = Linreg.Data.add_point(d, 6, 12)
13%Linreg.Data{points: [{6, 12}, {2, 4}]}
14
15# Train the model
16iex> m = Model.train(m, d)
17%Linreg.Model{b: 0.08, m: 0.4}
18
19# Predict our first value
20iex> Model.predict(m, 5)
212.08
```

In the example above weâ€™ve created a model and fed it some pretty straight forward trainings data where `Y = 2 * X`

. We trained our model and let it make a prediction for 5. We expect it to predict 10, however it predicts 2 and some change.. We are getting closer, but the error is still pretty high.

What we see here is the learning rate in effect! We take smaller steps so we donâ€™t overshoot our goal, however that also means that simply iterating over the training set once, will not cut it.

In machine learning; iterating over the trainings set once is referred to as one epoch. In most machine learning algorithms, it requires iterating over your data set many times. Each time shaving some points off of the error, each time perfecting the modelâ€™s weights a bit more.

Running the training and prediction functions again yield in values a little closer to what we expect.

## Multiple epochs

In order to make training for multiple epochs a little easier, letâ€™s revisit the `train/3`

function one more time and make it iterate over the entire training set multiple times.

```
1def train(%Model{} = model, %Data{points: points}, opts \\ []) do
2 learning_rate = Keyword.get(opts, :learning_rate, 0.01)
3 epochs = Keyword.get(opts, :epochs, 100)
4
5 for _epoch <- 1..epochs, reduce: model do
6 %Model{m: m, b: b} = model ->
7 m_error =
8 points
9 |> Enum.map(fn {x, y} -> x * (predict(model, x) - y) end)
10 |> Enum.sum()
11 |> Kernel./(length(points))
12
13 b_error =
14 points
15 |> Enum.map(fn {x, y} -> predict(model, x) - y end)
16 |> Enum.sum()
17 |> Kernel./(length(points))
18
19 %Model{model | m: m - m_error * learning_rate, b: b - b_error * learning_rate}
20 end
21end
```

We made a few changes to the train function. First off: Weâ€™re defining an epoch value which will determine how many times we run through the entire training set. Secondly; we wrapped the whole function body in an for comprehension so that we can iterate through the entire training set multiple times, each time adjusting the weights a little bit more.

Note that weâ€™re using a comprehension with the reduce option. This is essentially the same as

`Enum.reduce/3`

but is a little easier on the eyes with larger bodies, but that might just be a matter of taste ðŸ˜—

## More training

Now that weâ€™ve advanced our training function, letâ€™s try it out and see if our predictions will get a little closer this time! Fire up the IEx again (`iex -S mix`

) and run:

```
1iex> d = Linreg.Data.add_point(d, 2, 4)
2%Linreg.Data{points: [{2, 4}]}
3
4iex> d = Linreg.Data.add_point(d, 8, 16)
5%Linreg.Data{points: [{8, 16}, {2, 4}]}
6
7iex> m = Linreg.Model.train(m, d)
8%Linreg.Model{b: 0.22374565348607184, m: 1.966843594782793}
9
10iex> Linreg.Model.predict(m, 5)
1110.057963627400037
```

Hooray! We trained our model the formula `Y = 2 * X + 0`

. More or less. As you can see the values for M and B are still a bit off, but given enough training, these values will approach the correct values more and more.

Iâ€™m saying approach here; they will never exactly be 2 and 0. This is mainly due to the learning rate, but floating point arithmetic is also to blame here.

## Try for yourself

The previous example is using a learning rate of 0.01 and trains for 100 epochs, these values are overridable in the options and will yield different results. Simply by calling our training function with the options list:

```
1iex> m = Model.train(m, d, learning_rate: 0.1, epochs: 500)
```

Iâ€™ll leave it as an exercise to the reader to play a bit with these values and see what happens.

## Closing words

Thatâ€™s it for this post! We built a machine learning model that can perform linear regressions and trained it using gradient descent in Elixir.

In the next part weâ€™ll look at how we can make this an interactive example using Phoenix LiveView where clicking on a SVG plane will generate training data and let the model predict the best fitting line through those points.