Presented by
Kiran Matty, Director of Product Management, Aerospike
About this talk
ML models run better with more data; and the more iterations and the more training, tuning and validation you can do, the better your results. So what’s the big deal?
It can often be about speeding up e.g. Spark jobs for operational efficiency. The faster a job runs, the more jobs can be run on the same cluster, maximizing returns on time and investments significantly. This in turn reduces the training and inference times significantly.
Data prep is painful. The problem is having an online system with streaming data and the need to make an inference in milliseconds. The problem is wanting to pull disparate signal data from sources from different countries and data centers in real-time.
AI/ML systems have insatiable appetites for data.
In this session, we’ll cover how to: