<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by Justin Tennenbaum on Medium]]></title>
        <description><![CDATA[Stories by Justin Tennenbaum on Medium]]></description>
        <link>https://medium.com/@jmtennenbaum?source=rss-6d54f7f47f17------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/2*EbrS5AGmts6cxij4G-oF3Q.jpeg</url>
            <title>Stories by Justin Tennenbaum on Medium</title>
            <link>https://medium.com/@jmtennenbaum?source=rss-6d54f7f47f17------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Sun, 17 May 2026 02:59:25 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@jmtennenbaum/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Predict Hospital Deaths with Biomedical Data and Machine Learning]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.datadriveninvestor.com/predict-hospital-deaths-with-biomedical-data-and-machine-learning-f45d010aebae?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/712/0*iv2H50vGimyp-Gg3.jpg" width="712"></a></p><p class="medium-feed-snippet">One of the best qualities of machine learning is that it applies to almost every type of work imaginable. Healthcare especially has been&#x2026;</p><p class="medium-feed-link"><a href="https://medium.datadriveninvestor.com/predict-hospital-deaths-with-biomedical-data-and-machine-learning-f45d010aebae?source=rss-6d54f7f47f17------2">Continue reading on DataDrivenInvestor »</a></p></div>]]></description>
            <link>https://medium.datadriveninvestor.com/predict-hospital-deaths-with-biomedical-data-and-machine-learning-f45d010aebae?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/f45d010aebae</guid>
            <category><![CDATA[health]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 09 Mar 2020 01:44:35 GMT</pubDate>
            <atom:updated>2020-03-10T13:23:26.693Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Using TensorFlow to Analyze Tweets]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/@jmtennenbaum/using-tensorflow-to-analyze-tweets-31ac95bddf9d?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/2028/1*CW5oAv3ERMMJTip4fwh5MA.png" width="2028"></a></p><p class="medium-feed-snippet">I recently decided to participate in a Kaggle competition where the goal is to predict where or not a tweet is referencing a disaster such&#x2026;</p><p class="medium-feed-link"><a href="https://medium.com/@jmtennenbaum/using-tensorflow-to-analyze-tweets-31ac95bddf9d?source=rss-6d54f7f47f17------2">Continue reading on Medium »</a></p></div>]]></description>
            <link>https://medium.com/@jmtennenbaum/using-tensorflow-to-analyze-tweets-31ac95bddf9d?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/31ac95bddf9d</guid>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[nlp]]></category>
            <category><![CDATA[tensorflow]]></category>
            <category><![CDATA[machine-learning]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 24 Feb 2020 01:59:49 GMT</pubDate>
            <atom:updated>2020-02-24T01:59:49.956Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Using Pandas to Impute Features for NLP]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/@jmtennenbaum/using-pandas-to-impute-features-for-nlp-942df2b77203?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/2012/1*ZDo8VfKTnuIXv-PkIzqtnA.png" width="2012"></a></p><p class="medium-feed-snippet">I was looking for a new project to work on and decided the best place to start would be to look on Kaggle. I checked out their&#x2026;</p><p class="medium-feed-link"><a href="https://medium.com/@jmtennenbaum/using-pandas-to-impute-features-for-nlp-942df2b77203?source=rss-6d54f7f47f17------2">Continue reading on Medium »</a></p></div>]]></description>
            <link>https://medium.com/@jmtennenbaum/using-pandas-to-impute-features-for-nlp-942df2b77203?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/942df2b77203</guid>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[python]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 17 Feb 2020 01:31:14 GMT</pubDate>
            <atom:updated>2020-02-17T01:31:14.583Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Reinforcement Learning: Dynamic Programming]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/data-science/reinforcement-learning-dynamic-programming-2b89da6ea1b?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/696/1*QlDoJCGO_PGXHpfU1Q3auA.png" width="696"></a></p><p class="medium-feed-snippet">Using Dynamic Programming to find the optimal policy in Grid World</p><p class="medium-feed-link"><a href="https://medium.com/data-science/reinforcement-learning-dynamic-programming-2b89da6ea1b?source=rss-6d54f7f47f17------2">Continue reading on TDS Archive »</a></p></div>]]></description>
            <link>https://medium.com/data-science/reinforcement-learning-dynamic-programming-2b89da6ea1b?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/2b89da6ea1b</guid>
            <category><![CDATA[reinforcement-learning]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[machine-learning]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 10 Feb 2020 02:49:57 GMT</pubDate>
            <atom:updated>2020-02-10T15:27:16.786Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Markov Decision Processes and Grid World]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/@jmtennenbaum/markov-decision-processes-and-grid-world-45dee8d85fdb?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/784/0*Vuswt6sNIMcX3H0F.png" width="784"></a></p><p class="medium-feed-snippet">In my previous article I discuss my first attempt at reinforcement learning by using it to tackle the Multi Armed Bandit problem. This&#x2026;</p><p class="medium-feed-link"><a href="https://medium.com/@jmtennenbaum/markov-decision-processes-and-grid-world-45dee8d85fdb?source=rss-6d54f7f47f17------2">Continue reading on Medium »</a></p></div>]]></description>
            <link>https://medium.com/@jmtennenbaum/markov-decision-processes-and-grid-world-45dee8d85fdb?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/45dee8d85fdb</guid>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[reinforcement-learning]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Sun, 02 Feb 2020 23:38:03 GMT</pubDate>
            <atom:updated>2020-02-02T23:38:03.173Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Starting with Reinforcement Learning: The Multi Armed Bandit Problem]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/@jmtennenbaum/starting-with-reinforcement-learning-the-multi-armed-bandit-problem-f124076775a4?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/1280/0*foDOUxlBXKBaluiT.jpg" width="1280"></a></p><p class="medium-feed-snippet">Going back to the basics of reinforcement learnings</p><p class="medium-feed-link"><a href="https://medium.com/@jmtennenbaum/starting-with-reinforcement-learning-the-multi-armed-bandit-problem-f124076775a4?source=rss-6d54f7f47f17------2">Continue reading on Medium »</a></p></div>]]></description>
            <link>https://medium.com/@jmtennenbaum/starting-with-reinforcement-learning-the-multi-armed-bandit-problem-f124076775a4?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/f124076775a4</guid>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[reinforcement-learning]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 27 Jan 2020 03:57:18 GMT</pubDate>
            <atom:updated>2020-01-27T17:48:46.456Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Coloring Photos with a Generative Adversarial Network]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/data-science/coloring-photos-with-a-generative-adversarial-network-a435c4403b5d?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/1458/1*N3nirBCA-K6NYDe3k9ZCGw.png" width="1458"></a></p><p class="medium-feed-snippet">Ever since I started learning about data science and machine learning, there has always been one algorithm that continually grabbed my&#x2026;</p><p class="medium-feed-link"><a href="https://medium.com/data-science/coloring-photos-with-a-generative-adversarial-network-a435c4403b5d?source=rss-6d54f7f47f17------2">Continue reading on TDS Archive »</a></p></div>]]></description>
            <link>https://medium.com/data-science/coloring-photos-with-a-generative-adversarial-network-a435c4403b5d?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/a435c4403b5d</guid>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[generative-adversarial]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Sun, 19 Jan 2020 18:42:41 GMT</pubDate>
            <atom:updated>2020-01-20T00:14:13.556Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Neural Networks and Image Recognition (CNN’s)]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/swlh/neural-networks-and-image-recognition-cnns-3bcd1a9f49b5?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/850/0*nUAhG6FgeRuSFwT-.png" width="850"></a></p><p class="medium-feed-snippet">Neural Networks have exploded in popularity the past couple of decades, and because of this we have adapted multiple variations to&#x2026;</p><p class="medium-feed-link"><a href="https://medium.com/swlh/neural-networks-and-image-recognition-cnns-3bcd1a9f49b5?source=rss-6d54f7f47f17------2">Continue reading on The Startup »</a></p></div>]]></description>
            <link>https://medium.com/swlh/neural-networks-and-image-recognition-cnns-3bcd1a9f49b5?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/3bcd1a9f49b5</guid>
            <category><![CDATA[neural-networks]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 04 Nov 2019 17:44:20 GMT</pubDate>
            <atom:updated>2019-11-14T19:52:34.918Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Automated Machine Learning]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/data-science/automated-machine-learning-d8568857bda1?source=rss-6d54f7f47f17------2"><img src="https://cdn-images-1.medium.com/max/1294/0*vcltcWFsDxDl-nzy.png" width="1294"></a></p><p class="medium-feed-snippet">Automated Machine Learning(AutoML) is currently one of the explosive subfields within Data Science. It sounds great for those who are not&#x2026;</p><p class="medium-feed-link"><a href="https://medium.com/data-science/automated-machine-learning-d8568857bda1?source=rss-6d54f7f47f17------2">Continue reading on TDS Archive »</a></p></div>]]></description>
            <link>https://medium.com/data-science/automated-machine-learning-d8568857bda1?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/d8568857bda1</guid>
            <category><![CDATA[machine-learning-ai]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 21 Oct 2019 13:30:46 GMT</pubDate>
            <atom:updated>2019-11-04T14:19:08.255Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Analysis of Variance and its Variations]]></title>
            <link>https://medium.com/data-science/analysis-of-variance-and-its-variations-6ef3f8fbeb05?source=rss-6d54f7f47f17------2</link>
            <guid isPermaLink="false">https://medium.com/p/6ef3f8fbeb05</guid>
            <category><![CDATA[anova]]></category>
            <category><![CDATA[statistics]]></category>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Justin Tennenbaum]]></dc:creator>
            <pubDate>Mon, 07 Oct 2019 16:40:03 GMT</pubDate>
            <atom:updated>2019-10-11T00:50:08.599Z</atom:updated>
            <content:encoded><![CDATA[<p>In statistics, when trying to compare samples, our first thought is to perform a student’s t-test. It compares the means of two samples (or a sample and population) relative to the standard error of the mean or pooled standard deviation. While the t-test is a robust and useful experiment, it limits itself to comparing only two groups at a time.</p><p>In order to compare multiple groups at once, we can look at the ANOVA, or Analysis of Variance. Unlike the t-test, it compares the variance within each sample relative to the variance between the samples. Ronald Fisher introduced the term <strong>Variance</strong> and its formal analysis in 1918, with Analysis of Variance becoming widely known in 1925 after Fisher’s <em>Statistical Methods for Research Workers</em>. The students t-test follows a t-distribution, follows normal distribution’s shape, however it has fatter tails to account for more values farther from the mean in samples.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*x9gXH3_V4nJXced4BoHALw.png" /><figcaption>Source: Wikipedia</figcaption></figure><p>Anova however follows the F-Distribution which is a right skewed distribution with a long tail. For only two groups, we can use the F-distribution directly to compare the variances.</p><p>If U1 and U2 are independent variables distributed by a Chi-Squared distribution with d1 and d2 degrees of freedom respectively, then the Random Variable X (where X = (U1/d1)/(U2/d2)) follows a F-Distribution with parameters X~F(d1,d2) and its PDF given by:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*TF-GNlpRUyhx6wnkfZdmTg.png" /><figcaption>Source: Wikipedia; (B=Beta Function)</figcaption></figure><p>Anova uses this same Distribution however the way it calculates its f-Value varies on the type of ANOVA test performed. The simplest form of Anova comes in the form of a 1-way Anova test which allows us to compare multiple groups by evaluating 1 independent variable and 1 dependent variable. In general Anova follows three main assumptions:</p><p><strong>·</strong> The distribution of the dependent variable should be continuous and approximately normal</p><p><strong>·</strong> Independence of Samples</p><p><strong>·</strong> Homogeneity of Variances</p><p>The Anova then evaluates the ratio of variance between the groups compared to variance within in order to calculate its f-value. The one way Anova is given by the following table:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*L1DZfj3wAoFJ3SRtYnQxlQ.png" /><figcaption>Source: Analytics Buddhu</figcaption></figure><p>Once we calculate our F-Ratio, we can compare it to our F-critical to determine if we can reject our null hypothesis. For an Anova test our alternative hypothesis is that <strong>at least one</strong> of the groups differ from each other so an ad hoc test such as Least Significant Squares or Tukey’s HSD (honestly significant test) . These filter through all combinations to determine which sample groups are different from each other. These tests are only performed if our Anova returns a statistically significant result.</p><p><strong>Variations of Anova</strong></p><p>As I said above the 1-way Anova can only account for a single independent and dependent variable. There are extensions of the 1-way Anova that allow us to circumvent these limitations. The first is the 2-way Anova. This test still requires us to have only 1 dependent variable, however we are able to include multiple independent variables to analyze the variance between groups. Since we have multiple variables, two calculations occur, the main effect and interaction effect. The main effect considers each independent variable separately, while the interaction effect looks all at the variables simultaneously.</p><p>A 2-way Anova is actually just a type of Factorial Anova, which means the test is going to contain multiple levels of independent variables (also called a factor). Simply put a two way Anova is a Factorial Anova with a level of 2. So a three way Anova has three independent variables, a 4-way has 4, and so on. The most common levels are 2 and 3, since above that can become difficult to interpret In a 1-way Anova the variability is compared between and within groups, while a factorial Anova compares the level of each factor with the other factors.</p><p>The other limitation of a 1-way Anova is fixed by performing a Manova, which allows us to compare the variances between groups on more than 1 dependent variable. The Manova will return a multivariate f-value as compared to the Anova, which returns a univariate f-value. The multivariate f-value will only indicate if the test is significant, it gives us no information on which particular variable is different between the groups. In order to learn which if 1 or more the dependent variables are significant, a 1-way Anova test needs to be performed to acquire the univariate f-value for each variable, followed by ad hoc tests.</p><p>A major assumption for all of the above tests is that our sample are independent of each other. This assumption means that we are unable to evaluate groups over time, or measure multiple results for the same subject. For only two groups this can easily be solved by using a dependent t-test. The Repeated Measure Anova is an extension of the dependent t-test and allows us to evaluate the same dependent subjects over multiple categories or times.</p><p>The repeated measure Anova is calculated very similarly to the 1-way Anova. Rather than being split between groups and within groups, it is split between times/conditions and within conditions/times has been split into 2 smaller categories, where the SSw equals SSsubjects + SSerror. Since we are using the same subjects in each group we can remove the variability of the subjects giving us a smaller error within-groups.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*vy25CfHPBiB8qWLlSQ3cUw.png" /><figcaption>Source: statistics.laerd.com</figcaption></figure><p>This partition ends up increasing our F-statistic which means a repeated measure Anova has an increased power in finding statistical differences. However, this only leads to a more powerful test if our decrease of within group variability outweighs the decrease in degrees of freedoms.</p><p>I recently did my own project comparing the average temperatures each year over the course of multiple decades. I originally made the mistake of running a 1-way Anova, forgetting that my different “groups” were dependent of each other. I couldn’t shake the feeling that I had performed the wrong test and I wanted to dig deeper, hence why I am writing this. I dove back into my data, re-ran the test in Python using a Repeated Measure Anova and got the following results.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*692TSBBVLsSsYzeIdUciwA.png" /></figure><p>Given my results and an alpha of 0.05, I found an F-critical value of 2.77, and therefore I was able to reject my null hypothesis that the temperatures across different decades remain unchanged. It is always important to understand the assumptions and requirements of tests in order to make sure we are performing accurate and correct tests. As a data scientist it is our job to make sure we are always performing the right test, not just the ones that give us the answers we want.</p><p>Soures:</p><ul><li><a href="https://en.wikipedia.org/wiki/Analysis_of_variance">Analysis of variance - Wikipedia</a></li><li><a href="https://www.statisticshowto.datasciencecentral.com/probability-and-statistics/hypothesis-testing/anova/#targetText=There%20are%20two%20main%20types,double%2Dtesting%20that%20same%20group.">ANOVA Test: Definition, Types, Examples</a></li><li><a href="https://en.wikipedia.org/wiki/F-distribution#targetText=In%20probability%20theory%20and%20statistics,%2C%20e.g.%2C%20F%2Dtest.">F-distribution</a></li><li><a href="https://statistics.laerd.com/statistical-guides/repeated-measures-anova-statistical-guide.php">Repeated Measures ANOVA</a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6ef3f8fbeb05" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/analysis-of-variance-and-its-variations-6ef3f8fbeb05">Analysis of Variance and its Variations</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>