# Tag: Model Performance (13)

**5 Machine Learning Projects You Should Not Overlook, June 2018**- Jun 12, 2018.

Here is a new installment of 5 more machine learning or machine learning-related projects you may not yet have heard of, but may want to consider checking out!**Train your Deep Learning Faster: FreezeOut**- Aug 3, 2017.

We explain another novel method for much faster training of Deep Learning models by freezing the intermediate layers, and show that it has little or no effect on accuracy.**Train your Deep Learning model faster and sharper: Snapshot Ensembling — M models for the cost of 1**- Aug 2, 2017.

We explain a novel Snapshot Ensembling method for increasing accuracy of Deep Learning models while also reducing training time.**The Top Predictive Analytics Pitfalls to Avoid**- Jan 23, 2017.

Predictive modelling and machine learning are significantly contributing to business, but they can be very sensitive to data and changes in it, which makes it very important to use proper techniques and avoid pitfalls in building data science models.**Sound Data Science: Avoiding the Most Pernicious Prediction Pitfall**- Jan 5, 2017.

Data science and predictive analytics can provide huge value, but they can mislead and backfire if not used with fail-safe measures. The author gives examples of such problems and provides guidelines to avoid them.**Continuous improvement for IoT through AI / Continuous learning**- Nov 25, 2016.

In reality, especially for IoT, it is not like once an analytics model is built, it will give the results with same accuracy till the end of time. Data pattern changes over the time which makes it absolutely important to learn from new data and improve/recalibrate the models to get correct result. Below article explain this phenomenon of continuous improvement in analytics for IoT.**Understanding the Bias-Variance Tradeoff: An Overview**- Aug 8, 2016.

A model's ability to minimize bias and minimize variance are often thought of as 2 opposing ends of a spectrum. Being able to understand these two types of errors are critical to diagnosing model results.**How to Compute the Statistical Significance of Two Classifiers Performance Difference**- Mar 30, 2016.

To determine whether a result is statistically significant, a researcher would have to calculate a p-value, which is the probability of observing an effect given that the null hypothesis is true. Here we are demonstrating how you can compute difference between two models using it.**Big Idea To Avoid Overfitting: Reusable Holdout to Preserve Validity in Adaptive Data Analysis**- Aug 17, 2015.

Big Data makes it all too easy find spurious "patterns" in data. A new approach helps avoid overfitting by using 2 key ideas: validation should not reveal any information about the holdout data, and adding of a small amount of noise to any validation result.**Overcoming Overfitting with the reusable holdout: Preserving validity in adaptive data analysis**- Aug 12, 2015.

Misapplication of statistical data analysis is a common cause of spurious discoveries in scientific research. We demonstrate a new approach for addressing the challenges of adaptivity based on insights from privacy-preserving data analysis.**How to Lead a Data Science Contest without Reading the Data**- May 17, 2015.

We examine a “wacky” boosting method that lets you climb the public leaderboard without even looking at the data . But there is a catch, so read on before trying to win Kaggle competitions with this approach.**Failing Optimally – Data Science’s Measurement Problem**- Mar 4, 2015.

Data science has a measurement problem. Simple metrics may not address complex situations. But complex metrics present myriad problems.**Interpreting Model Performance with Cost Functions**- Jan 13, 2014.

Cost functions are critical for the correct assessment of performance of data mining and predictive models. This series goes deep into the statistical properties and mathematical understanding of each cost function and explores their similarities and differences.