The algorithms we are going to create need to be able to handle lots of different data, the problems we can now solve with machine learning are usually relating image, text or audio inputs.
To handle these types of data, we try and convert them into formats that we can easily manipulate using mathematics and computers, a common strategy we'll see is to turn our input data into a vector.
For example given a black and white image that is pixels wide and
Thus we could write a very simple function to compute the brightness of an image by using the length of the input vector. Alternatively given two images (vectors), we could measure the length of their difference, if 0 the images are the same and the larger the the length of the difference the more the two images must differ
When we're dealing with supervised learning we are trying to predict properties about new data after it's seen lots of data with their true properties. The example we spoke of previous was doing object categorization
We know that our algorithm has access to a training set before hand, and now that we're equipped with knowledge of vectorization, we could first have a finite set of categories
Thus given this object categorizaiton problem we have a concrete way to represent what kind of data we'll be working with while creating our learning algorithms.