All content for Linear Digressions is the property of Ben Jaffe and Katie Malone and is served directly from their servers
with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
In each episode, your hosts explore machine learning and data science through interesting (and often very unusual) applications.
It’s pretty common to fit a function to a dataset when you’re a data scientist. But in many cases, it’s not clear what kind of function might be most appropriate—linear? quadratic? sinusoidal? some combination of these, and perhaps others? Gaussian processes introduce a nonparameteric option where you can fit over all the possible types of functions, using the data points in your datasets as constraints on the results that you get (the idea being that, no matter what the “true” underlying function is, it produced the data points you’re trying to fit). What this means is a very flexible, but depending on your parameters not-too-flexible, way to fit complex datasets.
The math underlying GPs gets complex, and the links below contain some excellent visualizations that help make the underlying concepts clearer. Check them out!
Relevant links:
http://katbailey.github.io/post/gaussian-processes-for-dummies/
https://thegradient.pub/gaussian-process-not-quite-for-dummies/
https://distill.pub/2019/visual-exploration-gaussian-processes/
Linear Digressions
In each episode, your hosts explore machine learning and data science through interesting (and often very unusual) applications.