Future of ml

10 May 2021

I’ve been thinking about the future of software and machine learning and reached some conclusions. Conclusions others have reached as well.

Machine Learning has started to converge on a few models and as time goes on this will narrow further. These models will live on specialized hardware and we will compile different, less resource intensive ones to ship on standard machines.

What this means is people will spend far more time on data. You will program the prompts you feed your GPT-N for text generation, you will carefully curate which samples you use to distill and prompt your GAN/VAE, you will bootstrap your own dataset using pre-existing models, you will write programs to generate a dataset, and on and on.

A lot of software is, or should be, very modular with several tiers of abstraction. For example LLVM takes your code and creates an intermediate graph which then gets optimized and compiled. As ML improves it will eat these layers of abstraction. Code will be written for humans to read but graph-based RL agents – or whatever else comes along – will optimize your program to run optimally on any arbitrary hardware.

So what should someone, who wants to make a career in software – while humans still have a hand in its creation – focus on now and in the future?