My name is Lovkush Agarwal. I recently decided to change careers and become a
data scientist. Following David Robinson’s’
advice, I decided to create this blog, to record my progress, learning and projects.
Summaries of the talks and my thoughts for Day 2
Jul 24, 2020
I am attending my first ever tech-related conference. Here I record my thoughts on Day I.
Jul 23, 2020
I end this series by describing what I learnt by reading other people's kernals on Kaggle.
Jul 18, 2020
I use the vanilla network from the first part on the MNIST data, achieving an accuracy of 94.9%.
Jul 14, 2020
I carry out several feature selection algorithms, with the hope of removing features that are reducing the performance of the models.
Jul 13, 2020
I create my first ever neural network. It is a vanilla network, written from scratch in Python.
Jul 9, 2020
I start a new project modelling another Kaggle dataset. To start things off, I create some default models, to establish a starting point for future models.
Jul 1, 2020
I end this project by summarising what I did and summarising what I learnt by having a look at other people's examples on Kaggle.
Jun 25, 2020
I finish this project by plotting various charts to summarise the data obtained in the previous two parts.
Jun 22, 2020
In this post, I describe the cleaning I did on the data.
Jun 17, 2020
During my previous job teaching mathematics at the University of Leicester, I did a project to investigate whether students did their homework last minute.
Jun 16, 2020
In light of the current prominence of BlackLivesMatter, I decided to investigate crime in relation to race. Here I describe how I collected the data I will be analysing.
Jun 15, 2020
I describe the various ways I made the algorithm from Part II more efficient. These resulted in big improvements in the efficiency.
Jun 9, 2020
I created an algorithm to search through the game-tree of pentago to a given maximum depth. The codes works, but it is highly inefficient.
Jun 4, 2020
I complete the hyper-parameter optimisations for the random forest and xgboost models. I then create a final model using these values to produce AUCs of 0.852 and 0.872.
May 30, 2020