- Few-Shot Image Classification with Meta-Learning - Mar 12, 2020.
Here is how you can teach your model to learn quickly from a few examples.
- Amazon Uses Self-Learning to Teach Alexa to Correct its Own Mistakes - Feb 10, 2020.
The digital assistant incorporates a reformulation engine that can learn to correct responses in real time based on customer interactions.
- The ravages of concept drift in stream learning applications and how to deal with it - Dec 18, 2019.
Stream data processing has gained progressive momentum with the arriving of new stream applications and big data scenarios. These streams of data evolve generally over time and may be occasionally affected by a change (concept drift). How to handle this change by using detection and adaptation mechanisms is crucial in many real-world systems.
- Probability Learning: Naive Bayes - Nov 26, 2019.
This post will describe various simplifications of Bayes' Theorem, that make it more practical and applicable to real world problems: these simplifications are known by the name of Naive Bayes. Also, to clarify everything we will see a very illustrative example of how Naive Bayes can be applied for classification.
- Live Webinar: Continual Learning with Human-in-the-loop - Nov 18, 2019.
Join this live webinar from cnvrg, Continual Learning with Human-in-the-loop, Nov 26 @ 12 PM EST, and learn the role of human-in-the-loop in your ML pipeline, how to close the loop in your pipeline, and much more.
- Probability Learning: Maximum Likelihood - Nov 5, 2019.
The maths behind Bayes will be better understood if we first cover the theory and maths underlying another fundamental method of probabilistic machine learning: Maximum Likelihood. This post will be dedicated to explaining it.
- A Concise Explanation of Learning Algorithms with the Mitchell Paradigm - Oct 5, 2018.
A single quote from Tom Mitchell can shed light on both the abstract concept and concrete implementations of machine learning algorithms.