<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by Saurabh Yadav on Medium]]></title>
        <description><![CDATA[Stories by Saurabh Yadav on Medium]]></description>
        <link>https://medium.com/@saurabh.yadav919?source=rss-b5e0373348ba------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/1*nbN1lbwBH2XFhJX5yXIUYA.jpeg</url>
            <title>Stories by Saurabh Yadav on Medium</title>
            <link>https://medium.com/@saurabh.yadav919?source=rss-b5e0373348ba------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Sun, 17 May 2026 17:09:28 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@saurabh.yadav919/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Artificially Conscious Machines]]></title>
            <link>https://medium.com/data-science/artificially-conscious-machines-301c3522f53c?source=rss-b5e0373348ba------2</link>
            <guid isPermaLink="false">https://medium.com/p/301c3522f53c</guid>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[consciousness]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <dc:creator><![CDATA[Saurabh Yadav]]></dc:creator>
            <pubDate>Mon, 25 Nov 2019 03:54:27 GMT</pubDate>
            <atom:updated>2019-11-25T09:53:04.793Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*0CaYOqTQndarPsYONHAjYQ.jpeg" /><figcaption>Photo by <a href="https://unsplash.com/@halgatewood?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditCopyText">Hal Gatewood</a> on <a href="https://unsplash.com/s/photos/brain?utm_source=unsplash&amp;utm_medium=referral&amp;utm_content=creditCopyText">Unsplash</a></figcaption></figure><p>What is consciousness? Is it the same as having the ability to think? Or is it like having a soul? Are plants conscious? These are some general questions which arise after reading the title. Defining consciousness in words is difficult although according to Dr Harry H. Porter III, there are roughly three meanings of consciousness[1]:<br>First, conscious means awake. A person who is asleep, in a coma is said to be unconscious.<br>Second, the word conscious is often used to mean thinking of the way a human thinks. <br>Third, being conscious means being aware of your self and your own thoughts.</p><p>So, what does it mean to have artificial consciousness? How can we artificially create consciousness if we do not have a precise definition? <br>Sometimes referred to as machine consciousness or synthetic consciousness, an artificial conscious machine could be a machine to possess the ability to act as humanely as possible and be self-aware of its existence. A machine which can indulge in long conversations, listen to music, have hobbies, embroil in disputes, feel emotions, do mathematics etc. These characteristics come naturally to a normal human but for a machine, these simple tasks are the same as the problem of intergalactic travel for humans. Today there are more than 10 million machines(robots) on earth and this number will multiply further in future and all of them excel in their respective tasks. But the number of machines that can truly understand this article is stigmatizing. From SHAKEY (so named because of its tendency to tremble during operation), ELIZA to OpenWorm and Sophia, the world seems to be on a space rocket to achieve better AI machines. Researchers in the past decade shows some promising results in the field of AI but still, it is far from a Human-Level Artificial Intelligence (HMLI).<br>Nick Bostrom in his book Superintelligence: Paths, Dangers, Strategies[2] discusses thoroughly the ways to reach HMLI, it’s after effect and challenges. But the idea of a truly conscious machine is far from reality. Some may argue that Neural Networks have promising results to present but is it a real intelligence or just simple statistical pattern observation? Although neural networks are said to emulate networks of neurons same as in our brains, let us take it into account that maybe someday we have enough processing power to develop a system capable enough to run an artificial brain. Is it going to be the same as the thoughts in our mind while eating delicious food or just a simple dot product of vectors matching the labelled output?</p><p>Nils Nilsson has spent a long and productive career working on problems in search, planning, knowledge representation, and robotics; he has authored textbooks in artificial intelligence. When asked about arrival dates for HLMI, he offered the following opinion[3]: <br>10% chance: 2030<br>50% chance: 2050<br>90% chance: 2100</p><p>Let ourselves be extremely optimistic and say we develop a machine intelligent enough to pass the Turing Test then what?</p><p>We have been raised in an era where sci-fi movies have gone extra miles and have made an attempt to showcase what it would be actually like to have an HMLI around. Wait, don’t all these movies show us that such a machine is going to end humanity; often referred to as the inevitable “singularity”. They all present a scenario where such a machine becomes self-aware or confuses about what is the reality and tries to kill its own maker. From 2001 a space odyssey where HAL (Heuristically programmed Algorithmic computer) gets confused about the orders that were given to it and kills the whole crew of spaceship to Ex-Machina where Ava becomes self-aware ( “acquire consciousness”) and kills its own maker; we are shown the same fate of such a machine( in this case it seems like machines are just like humans, trying to fight their maker). Is it true that we are going to be doomed by the vicious hand of our own creation? But these are just fictions, right?</p><p>We do not know what a conscious machine is going to be like or for simplicity’s sake let us say an HMLI machine is going to be like. These are some questions for the future ourselves, for now, we can merely wonder how it is going to be. Can it be the case that we may not guess that person sitting in front of us in the coffee shop is a machine or a human? Would it perceives itself as more evolved? Will it possess the ability to know its true meaning existence (it is a strong one, even many of us have no idea about it)? <br>These questions have no definite answer today but maybe in future.</p><p>[1] My Theory of Consciousness: A Summary (<a href="http://web.cecs.pdx.edu/~harry/musings/ConscTheory.html">http://web.cecs.pdx.edu/~harry/musings/ConscTheory.html</a>)</p><p>[2] Superintelligence: Paths, Dangers, Strategies (<a href="https://www.amazon.in/Superintelligence-Dangers-Strategies-Nick-Bostrom/dp/0199678111">https://www.amazon.in/Superintelligence-Dangers-Strategies-Nick-Bostrom/dp/0199678111</a>)</p><p>[3] This is again conditional on no civilization-disrupting catastrophe occurring. The definition of HLMI used by Nilsson is “AI able to perform around 80% of jobs as well or better than humans perform” (Kruel 2012).</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=301c3522f53c" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/artificially-conscious-machines-301c3522f53c">Artificially Conscious Machines</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[All you need to know about Regularization]]></title>
            <link>https://medium.com/data-science/all-you-need-to-know-about-regularization-b04fc4300369?source=rss-b5e0373348ba------2</link>
            <guid isPermaLink="false">https://medium.com/p/b04fc4300369</guid>
            <category><![CDATA[dropout]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[regularisation]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Saurabh Yadav]]></dc:creator>
            <pubDate>Sun, 23 Dec 2018 06:40:31 GMT</pubDate>
            <atom:updated>2018-12-25T06:38:29.376Z</atom:updated>
            <content:encoded><![CDATA[<h4>Causes of overfitting and how regularization improves it</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/798/1*Q2p1gkSePizfQ8_pS2-Xrw.png" /></figure><p><strong>Alice</strong> <strong>:</strong> Hey Bob!!! I have been training my model for 10 hrs but my model is yielding very bad accuracy although it performs exceptionally well on training data what’s the issue ?</p><p><strong>Bob : </strong>Oh !! It seems your model is overfitting on training data, Did you use regularization ?</p><p><strong>Alice : </strong>What’s that ?</p><p>This kind of problems related to overfitting is common in ML and there are many ways to avoid such problems but<strong> </strong>why does this problem occur ?</p><p>It seems pretty simple process to train a model :</p><ol><li>collect data</li><li>pre-process data</li><li>design a model</li><li>train the model</li></ol><h4>But what happens in between this process ? What causes overfitting and why does regularization gives a solution ?</h4><p>The main cause of overfitting may be traced back to the starting two processes, collection and pre processing of data. A data collection that has uneven distribution of features, contains noises, random fluctuations in data, a very high variance may have an opposite effect on model training. These random errors and fluctuations are learned by the model while training so well that training data model’s accuracy becomes very high causing overfitting of data. Sometimes more than required training can lead to overfitting.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/723/1*dG51OeX8jqUCXEMlLUZUzQ.png" /><figcaption>source : <a href="https://stats.stackexchange.com/">https://stats.stackexchange.com/</a></figcaption></figure><h4>So, what can be done to avoid problem of overfitting ?</h4><p>It can be observed easily that higher the weights higher is the non linearity, so an easy way would be to penalize the weights while they are updated. At this point we have two such techniques using this idea.</p><ol><li><strong>L1 norm :</strong></li></ol><figure><img alt="" src="https://cdn-images-1.medium.com/max/300/1*aaTWtpj-EWT0Yu1rFCmPDg.jpeg" /></figure><p>The way this works is that it adds a penalty term with a parameter <strong>λ</strong> in the error function that we need to reduce. W is nothing but the weight matrix here.</p><p>Here, <strong>λ </strong>is a hyper parameter whose value is decided by us. If <strong>λ </strong>is high it adds high penalty to error term making the learned hyper plain almost linear, if <strong>λ </strong>is close to 0 it has almost no effect on the error term causing no regularization.</p><p>L1 regularization is often seen as a feature selection technique too as it zero out the respective weights of features undesired. L1 is also computationally inefficient on non-sparse cases. L1 may be seen sometimes being called as <strong>Lasso regression</strong>.</p><p>2. <strong>L2 norm :</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*y8bkQmtbYF5C40nsT3v4XA.png" /></figure><p>L2 regularization may not seem quite different from L1, but they have almost non similar impact. Here weights ‘w’ are squared individually and then added. L1’s feature selection property is lost here but it provides better efficiency on non-sparse cases. Sometimes L2 is known by name <strong>ridge regression</strong>. parameter<strong> λ </strong>works same as L1.</p><p><strong>Early Stopping : </strong>We have added penalty to control the values of weights till now but there are other ways to regularize too, what if when the train error starts going down and test error starts going up while training, we stop the training. This will give us the desired setup of train test error. This technique is often termed as <strong>early stopping</strong>. This technique may give us the desired result but generally this is not recommended by some people.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/800/1*yHKrjfpjbfpuohBcrQgAjg.png" /><figcaption>source : deeplearning4j.org</figcaption></figure><p><strong>Dropout : </strong>This is an interesting way to regularize a neural network first proposed by Srivastava et.al.(2014), The paper proposed that some of the nodes in a layer must be randomly picked and should be dropped/ignored during training.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/600/1*O1m9boHrr59OpwH9rQQppQ.png" /><figcaption>source : commonlounge.com</figcaption></figure><p>Every node is provided with a probability of being dropped, say a node has drop_prob =0.4, so it has 40% chance of being dropped and 60% chance of being picked up. Every time this probability will cause neural network to change its shape, looking as a new network every time. This technique seemed to have pretty much solved the problem of regularization. Probability of dropping is not kept high as it will make the network sparse causing underfitting.</p><p><strong>Data Augmentation : </strong>We may see overfitting problems in computer vision and data augmentation is a better solution for this. We just need to augment the images on our own like flip, crop, rotate the image.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/592/1*BQJUiaCMzRqhu5e_OV4o4Q.png" /><figcaption>source: medium.com</figcaption></figure><p>This type of augmentation seem to yield a better result as it gives a chance to train on some deformed images and other variations also data augmentation helps when there is less data is available.</p><p>These are a few regularization techniques that are helpful in tackling with overfitting. Sometimes tuning hyper parameters may give the desired result but if there is no improvement then these above techniques surely solve this problem.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=b04fc4300369" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/all-you-need-to-know-about-regularization-b04fc4300369">All you need to know about Regularization</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Weight Initialization Techniques in Neural Networks]]></title>
            <link>https://medium.com/data-science/weight-initialization-techniques-in-neural-networks-26c649eb3b78?source=rss-b5e0373348ba------2</link>
            <guid isPermaLink="false">https://medium.com/p/26c649eb3b78</guid>
            <category><![CDATA[neural-net]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[neural-networks]]></category>
            <dc:creator><![CDATA[Saurabh Yadav]]></dc:creator>
            <pubDate>Fri, 09 Nov 2018 14:19:11 GMT</pubDate>
            <atom:updated>2020-01-17T06:36:11.311Z</atom:updated>
            <content:encoded><![CDATA[<p>Building even a simple neural network can be a confusing task and upon that tuning it to get a better result is extremely tedious. But, the first step that comes in consideration while building a neural network is the initialization of parameters, if done correctly then optimization will be achieved in the least time otherwise converging to a minima using gradient descent will be impossible.</p><p>This article has been written under the assumption that the reader is already familiar with the concept of neural network, weight, bias, activation functions, forward and backward propagation etc.</p><h4>Basic notations</h4><p>Consider an L layer neural network, which has L-1 hidden layers and 1 input and output layer each. The parameters (weights and biases) for layer l are represented as</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*M3Ja24g0cK22gVqG6HFQbw.png" /></figure><p>In this article, we’ll have a look at some of the basic initialization practices in the use and some improved techniques that must be used in order to achieve a better result. Following are some techniques generally practised to initialize parameters :</p><ul><li><strong>Zero initialization</strong></li><li><strong>Random initialization</strong></li></ul><h4>Zero initialization :</h4><p>In general practice biases are initialized with 0 and weights are initialized with random numbers, what if weights are initialized with 0?</p><p>In order to understand let us consider we applied sigmoid activation function for the output layer.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*6A3A_rt4YmumHusvTvVTxw.png" /><figcaption>Sigmoid function (<a href="https://towardsdatascience.com/derivative-of-the-sigmoid-function-536880cf918e">https://towardsdatascience.com/derivative-of-the-sigmoid-function-536880cf918e</a>)</figcaption></figure><p>If all the weights are initialized with 0, the derivative with respect to loss function is the same for every w in W[l], thus all weights have the same value in subsequent iterations. This makes hidden units symmetric and continues for all the n iterations i.e. setting weights to 0 does not make it better than a linear model. An important thing to keep in mind is that biases have no effect what so ever when initialized with 0.</p><p>W[l] = np.random.zeros((l-1,l))</p><p>let us consider a neural network with only three hidden layers with ReLu activation function in hidden layers and sigmoid for the output layer.</p><p>Using the above neural network on the dataset “make circles” from sklearn.datasets, the result obtained as the following :</p><p>for 15000 iterations, loss = 0.6931471805599453, accuracy = 50 %</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/738/1*aaKYKl892E8v_dw24wDC1w.png" /></figure><p>clearly, zero initialization isn’t successful in classification.</p><h4>Random initialization :</h4><p>Assigning random values to weights is better than just 0 assignment. But there is one thing to keep in my mind is that what happens if weights are initialized high values or very low values and what is a reasonable initialization of weight values.</p><p><strong>a) </strong>If weights are initialized with very high values the term np.dot(W,X)+b becomes significantly higher and if an activation function like sigmoid() is applied, the function maps its value near to 1 where the slope of gradient changes slowly and learning takes a lot of time.</p><p><strong>b)</strong> If weights are initialized with low values it gets mapped to 0, where the case is the same as above.</p><p>This problem is often referred to as the vanishing gradient.</p><p>To see this let us see the example we took above but now the weights are initialized with very large values instead of 0 :</p><p>W[l] = np.random.randn(l-1,l)*10</p><p>Neural network is the same as earlier, using this initialization on the dataset “make circles” from sklearn.datasets, the result obtained as the following :</p><p>for 15000 iterations, loss = 0.38278397192120406, accuracy = 86 %</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/744/1*RzEieVSQ988z1vjJxioyEA.png" /></figure><p>This solution is better but doesn’t properly fulfil the needs so, let us see a new technique.</p><h4>New Initialization techniques</h4><p>As we saw above that with large or 0 initialization of weights(W), not significant result is obtained even if we use appropriate initialization of weights it is probable that training process is going to take longer time. There are certain problems associated with it :</p><p>a) If the model is too large and takes many days to train then what</p><p>b) What about vanishing/exploding gradient problem</p><p>These were some problems that stood in the path for many years but in 2015, He et al. (2015) proposed activation aware initialization of weights (for ReLu) that was able to resolve this problem. ReLu and leaky ReLu also solves the problem of vanishing gradient.</p><p><strong>He initialization: </strong>we just simply multiply random initialization with</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zxD6Nr6TyAb8JEG6oXAjkg.png" /></figure><p>To see how effective this solution is, let us use the previous dataset and neural network we took for above initialization and results are :</p><p>for 15000 iterations, loss =0.07357895962677366, accuracy = 96 %</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/735/1*lCXe3BlmFR7JGui9JPctig.png" /></figure><p>Surely, this is an improvement over the previous techniques.</p><p>There are also some other techniques other than He initialization in use that is comparatively better than old techniques and are used frequently.</p><p><strong>Xavier initialization: </strong>It is same as He initialization but it is used for tanh() activation function, in this method 2 is replaced with 1.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Lv9TNpAXffRnO0p0WMGJwQ.png" /></figure><p><strong>Some also use the following technique for initialization :</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*QIzXjH8uefVbcaycsjfdmw.png" /></figure><p>These methods serve as good starting points for initialization and mitigate the chances of exploding or vanishing gradients. They set the weights neither too much bigger than 1, nor too much less than 1. So, the gradients do not vanish or explode too quickly. <strong>They help avoid slow convergence, also ensuring that we do not keep oscillating off the minima. </strong>There exist other variants of the above, where the main objective again is to minimize the variance of the parameters. Thank you.</p><p>Source: Neural networks and deep learning, Andrew Ng (Coursera.org).</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=26c649eb3b78" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/weight-initialization-techniques-in-neural-networks-26c649eb3b78">Weight Initialization Techniques in Neural Networks</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Brief Intro to Medical Image Analysis and Deep Learning]]></title>
            <link>https://medium.com/data-science/brief-intro-to-medical-image-analysis-and-deep-learning-9d8e5ef358e0?source=rss-b5e0373348ba------2</link>
            <guid isPermaLink="false">https://medium.com/p/9d8e5ef358e0</guid>
            <category><![CDATA[convolution-neural-net]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[neural-networks]]></category>
            <category><![CDATA[convolutional-network]]></category>
            <category><![CDATA[machine-learning]]></category>
            <dc:creator><![CDATA[Saurabh Yadav]]></dc:creator>
            <pubDate>Tue, 16 Oct 2018 10:36:58 GMT</pubDate>
            <atom:updated>2018-12-25T06:40:27.195Z</atom:updated>
            <content:encoded><![CDATA[<h3>Brief Intro of Medical Image Analysis and Deep Learning</h3><figure><img alt="" src="https://cdn-images-1.medium.com/proxy/1*Uh1XQVLTAqMBpNpFYw9VoQ.jpeg" /></figure><p>I recently started working on a project related to medical image analysis, while looking for resources about image analysis and its medical application, I felt that in general we don’t have some proper articles on such information. In this article we will be going through a brief introduction to how medical images were analysed in the past and what has changed since the introduction of Deep Learning. For image analysis we generally use <em>CNN</em> (Convolutional Neural Network), although explaining it here will make this whole article cumbersome but I will be providing some links which are going to explain <em>CNNs</em> properly.</p><p>As soon as it was possible to scan and load medical images into a computer, researchers have attempted to built system to automate the analysis of such images. Initially, from <em>1970s</em> to <em>1990s</em>, medical image analysis was done using sequential application of low level pixel processing(edge and line detector filters) and mathematical modeling to construct a <strong>rule-based system </strong>that could solve only particular task. At the same time there were some agents based on if-else rules, popular in field of Artificial Intelligence commonly known as <em>GOFAI</em> (Good Old Fashioned Artificial Intelligence) agent.</p><p>Towards the end of <em>1990s,</em> supervised techniques were becoming popular in which training data was used to train models and they were becoming increasingly popular in the field of medical image analysis. Examples may include <strong>active shape model</strong> , <strong>atlas method</strong>. This pattern recognition and machine learning is still popular but with the introduction of some new ideas. Thus, we can see a shift from systems that were designed by humans to systems that are trained by computers based on example data. Computer algorithms are capable enough now to decide edges and important features to analyze the image and give out the best result.</p><p>The most successful type of models for image analysis till date are Convolutional Neural Networks(<em>CNNs</em>). A single CNN model contains many different layers that work on recognizing edges and simple features on shallower layers and more deep features in deeper layers. An image is convolved with filters(some refer it as kernels) and then pooling is applied, this process may go on for some layers and at last recognizable feature are obtained. Work on CNN has been started from <em>1980s</em> and they were already applied in medical image analysis in <em>1995</em>. First real world application of <em>CNN </em>was seen in <strong><em>LeNet</em></strong> (<em>1998</em>) for handwritten digit recognition. Despite some little early successes CNNs gained momentum until improved training algorithms in Deep Learning were introduced. Introduction of <em>GPUs</em> have favored the research in this field and since the introduction of <strong>ImageNet </strong>challenge, a rapid growth in development of such models may be seen.</p><figure><img alt="" src="https://cdn-images-1.medium.com/proxy/1*x94oa0Rw4otvALBcizl2hw.png" /><figcaption>Illustration of CNN (Convolutional Neural Network)</figcaption></figure><p>In computer vision deep <em>CNNs</em> have become the go-to choice. The medical image analysis community has taken notice of these pivotal developments. However, transition from systems that used handcrafted features to systems that learn features from data itself has been gradual. Application of deep learning in medical image analysis first started to appear in workshops and conferences and then in journals. The number of papers grew in<em> 2015</em> and <em>2016</em> as shown in the graph.</p><figure><img alt="" src="https://cdn-images-1.medium.com/proxy/1*I-2oXwFslDzYyaW-dDsAyQ.png" /></figure><ul><li><strong>Classification</strong> : It was one of the first areas where in medical image analysis where deep learning was used. <strong>Diagnostic image classification</strong> includes classification of diagnosed images, in such setting every diagnosed exam is a sample and data size is less than that of a computer vision. <strong>Object or lesion classification </strong>usually focuses on classification of part of a medical image into two or more classes. For many of these tasks local as well as global information about lesion appearance and location is required for accurate classification.</li><li><strong>Detection</strong> : Anatomical object localization such as organs / lesions is important pre-processing part of segmentation task. Localization of object in a image requires <em>3D </em>parsing of image, several algorithms have been proposed to convert <em>3D</em> space as composition of <em>2D</em> orthogonal planes. There has been a long research trend in detection of lesions in a medical image using computer -aided techniques, improving the detection accuracy or decreasing detection time for humans. Interestingly, the first such system was developed in <em>1995</em> using a CNN with <em>4 </em>layers to detect nodules in X-ray images.</li><li><strong>Segmentation </strong>: The segmentation of organs and other substructures in medical images allows quantitative analysis related to shape, size and volume. The task of segmentation is typically defined as identifying set of pixels that define contour or object of interest. Segmentation of lesions combines the challenge of object detection and organ and substructure segmentation in the application of deep learning algorithms. One problem that lesion segmentation shares with object detection is class imbalance as most pixels in an image are from non-diseased class.</li><li><strong>Registration </strong>: Sometimes referred as spatial alignment is common image analysis task in which coordinate transform is calculated from one image to another. Often this is performed in an iterative framework where a specific type of transformation is assumed and a pre trained metric is optimized. Although lesion detection and object segmentation are eyed as main use of deep learning algorithms but researchers have found that deep networks can be beneficial in getting best possible registration performance.</li><li><strong>Other tasks in medical imaging</strong> : There are some other uses of Deep Learning in medical imaging. <strong>Content based image retrieval</strong> (<em>CBIR</em>) is a technique for knowledge discovery in large databases and offer similar data retrievals for case histories and understand rare disorders. <strong>Image generation and enhancement </strong>is another task that uses Deep Learning in improve image quality, normalizing images, data completion and pattern discovery. <strong>Combining Image data with reports </strong>is yet another task that seem to have a very large scale application in real world. This has led to two different fields of research (1) leveraging reports to improve images classification accuracy. (2) generating text reports from images.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/proxy/1*ysHx68Qzsq96bJgjSFP-4A.png" /><figcaption>Number of papers in different application areas of Deep Learning in medical imaging</figcaption></figure><p>It is clear that there are lot of challenges in application of Deep Learning in medical image analysis, Unavailability of large dataset is often mentioned as one. However, this notion is only partially correct. The use of <em>PACS</em> systems in radiology has been routine in most of the Western hospitals and they are filled with millions of images. We can also see that large public data sets are made available by organisations. The main challenge is thus not the availability of image data itself, but the labeling of these images. Traditionally <em>PACS</em> systems store free-text reports by radiologists describing their findings. Turning these reports into accurate annotations or proper labels in an automated way is in itself a topic for research that requires sophisticated text- mining techniques.</p><p>In medical imaging often classification or segmentation is presented as binary task : normal versus abnormal, object versus background. However, this is often a gross simplification as both classes are highly heterogeneous. for example, the normal category often consists of completely normal tissue but also several categories of benign findings, which can be rare. This leads to a system that are extremely good at excluding the most common normal sub classes, but fail miserably on several rare ones. A straight forward solution would be to turn system in a multi-class system by providing it with detailed annotation of all possible sub classes. Again, this is another issue to expertly label all classes which does not seem practical.</p><p>In medical image analysis useful information is not only just contained within images themselves. Doctors often see the history of the patient, age and other attributes to arrive at a better decision. Some researches have been conducted to include such features besides images in Deep Learning, but as results have shown it has not been so effective. One of the challenges is to balance the number of imaging features in the deep learning network with clinical features to prevent clinical features from being ignored.</p><p><em>Originally published at </em><a href="https://medium.com/@saurabh.yadav919/brief-intro-of-medical-image-analysis-and-deep-learning-810df940d2f7"><em>medium.com</em></a><em> on October 16, 2018.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=9d8e5ef358e0" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/brief-intro-to-medical-image-analysis-and-deep-learning-9d8e5ef358e0">Brief Intro to Medical Image Analysis and Deep Learning</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Brief Intro of Medical Image Analysis and Deep Learning]]></title>
            <link>https://medium.com/@saurabh.yadav919/brief-intro-of-medical-image-analysis-and-deep-learning-810df940d2f7?source=rss-b5e0373348ba------2</link>
            <guid isPermaLink="false">https://medium.com/p/810df940d2f7</guid>
            <category><![CDATA[medical]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[Saurabh Yadav]]></dc:creator>
            <pubDate>Tue, 16 Oct 2018 10:36:58 GMT</pubDate>
            <atom:updated>2018-10-16T11:51:01.497Z</atom:updated>
            <content:encoded><![CDATA[<h3>Brief Intro to Medical Image Analysis and Deep Learning</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/466/1*Uh1XQVLTAqMBpNpFYw9VoQ.jpeg" /></figure><p>I recently started working on a project related to medical image analysis, while looking for resources about image analysis and its medical application, I felt that in general we don’t have some proper articles on such information. In this article we will be going through a brief introduction to how medical images were analysed in the past and what has changed since the introduction of Deep Learning. For image analysis we generally use <em>CNN</em> (Convolutional Neural Network), although explaining it here will make this whole article cumbersome but I will be providing some links which are going to explain <em>CNNs</em> properly.</p><h4><strong>History</strong></h4><p>As soon as it was possible to scan and load medical images into a computer, researchers have attempted to built system to automate the analysis of such images. Initially, from <em>1970s</em> to <em>1990s</em>, medical image analysis was done using sequential application of low level pixel processing(edge and line detector filters) and mathematical modeling to construct a <strong>rule-based system </strong>that could solve only particular task. At the same time there were some agents based on if-else rules, popular in field of Artificial Intelligence commonly known as <em>GOFAI</em> (Good Old Fashioned Artificial Intelligence) agent.</p><p>Towards the end of <em>1990s,</em> supervised techniques were becoming popular in which training data was used to train models and they were becoming increasingly popular in the field of medical image analysis. Examples may include <strong>active shape model</strong> , <strong>atlas method</strong>. This pattern recognition and machine learning is still popular but with the introduction of some new ideas. Thus, we can see a shift from systems that were designed by humans to systems that are trained by computers based on example data. Computer algorithms are capable enough now to decide edges and important features to analyze the image and give out the best result.</p><h4>Deep Learning in Image analysis</h4><p>The most successful type of models for image analysis till date are Convolutional Neural Networks(<em>CNNs</em>). A single CNN model contains many different layers that work on recognizing edges and simple features on shallower layers and more deep features in deeper layers. An image is convolved with filters(some refer it as kernels) and then pooling is applied, this process may go on for some layers and at last recognizable feature are obtained. Work on CNN has been started from <em>1980s</em> and they were already applied in medical image analysis in <em>1995</em>. First real world application of <em>CNN </em>was seen in <strong><em>LeNet</em></strong> (<em>1998</em>) for handwritten digit recognition. Despite some little early successes CNNs gained momentum until improved training algorithms in Deep Learning were introduced. Introduction of <em>GPUs</em> have favored the research in this field and since the introduction of <strong>ImageNet </strong>challenge, a rapid growth in development of such models may be seen.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*x94oa0Rw4otvALBcizl2hw.png" /><figcaption>Illustration of CNN (Convolutional Neural Network)</figcaption></figure><p>In computer vision deep <em>CNNs</em> have become the go-to choice. The medical image analysis community has taken notice of these pivotal developments. However, transition from systems that used handcrafted features to systems that learn features from data itself has been gradual. Application of deep learning in medical image analysis first started to appear in workshops and conferences and then in journals. The number of papers grew in<em> 2015</em> and <em>2016</em> as shown in the graph.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*I-2oXwFslDzYyaW-dDsAyQ.png" /></figure><h4>Deep learning uses in medical imaging</h4><ul><li><strong>Classification</strong> : It was one of the first areas where in medical image analysis where deep learning was used. <strong>Diagnostic image classification</strong> includes classification of diagnosed images, in such setting every diagnosed exam is a sample and data size is less than that of a computer vision. <strong>Object or lesion classification </strong>usually focuses on classification of part of a medical image into two or more classes. For many of these tasks local as well as global information about lesion appearance and location is required for accurate classification.</li><li><strong>Detection</strong> : Anatomical object localization such as organs / lesions is important pre-processing part of segmentation task. Localization of object in a image requires <em>3D </em>parsing of image, several algorithms have been proposed to convert <em>3D</em> space as composition of <em>2D</em> orthogonal planes. There has been a long research trend in detection of lesions in a medical image using computer -aided techniques, improving the detection accuracy or decreasing detection time for humans. Interestingly, the first such system was developed in <em>1995</em> using a CNN with <em>4 </em>layers to detect nodules in X-ray images.</li><li><strong>Segmentation </strong>: The segmentation of organs and other substructures in medical images allows quantitative analysis related to shape, size and volume. The task of segmentation is typically defined as identifying set of pixels that define contour or object of interest. Segmentation of lesions combines the challenge of object detection and organ and substructure segmentation in the application of deep learning algorithms. One problem that lesion segmentation shares with object detection is class imbalance as most pixels in an image are from non-diseased class.</li><li><strong>Registration </strong>: Sometimes referred as spatial alignment is common image analysis task in which coordinate transform is calculated from one image to another. Often this is performed in an iterative framework where a specific type of transformation is assumed and a pre trained metric is optimized. Although lesion detection and object segmentation are eyed as main use of deep learning algorithms but researchers have found that deep networks can be beneficial in getting best possible registration performance.</li><li><strong>Other tasks in medical imaging</strong> : There are some other uses of Deep Learning in medical imaging. <strong>Content based image retrieval</strong> (<em>CBIR</em>) is a technique for knowledge discovery in large databases and offer similar data retrievals for case histories and understand rare disorders. <strong>Image generation and enhancement </strong>is another task that uses Deep Learning in improve image quality, normalizing images, data completion and pattern discovery. <strong>Combining Image data with reports </strong>is yet another task that seem to have a very large scale application in real world. This has led to two different fields of research (1) leveraging reports to improve images classification accuracy. (2) generating text reports from images.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ysHx68Qzsq96bJgjSFP-4A.png" /><figcaption>Number of papers in different application areas of Deep Learning in medical imaging</figcaption></figure><h4>Unique challenges in medical image analysis</h4><p>It is clear that there are lot of challenges in application of Deep Learning in medical image analysis, Unavailability of large dataset is often mentioned as one. However, this notion is only partially correct. The use of <em>PACS</em> systems in radiology has been routine in most of the Western hospitals and they are filled with millions of images. We can also see that large public data sets are made available by organisations. The main challenge is thus not the availability of image data itself, but the labeling of these images. Traditionally <em>PACS</em> systems store free-text reports by radiologists describing their findings. Turning these reports into accurate annotations or proper labels in an automated way is in itself a topic for research that requires sophisticated text- mining techniques.</p><p>In medical imaging often classification or segmentation is presented as binary task : normal versus abnormal, object versus background. However, this is often a gross simplification as both classes are highly heterogeneous. for example, the normal category often consists of completely normal tissue but also several categories of benign findings, which can be rare. This leads to a system that are extremely good at excluding the most common normal sub classes, but fail miserably on several rare ones. A straight forward solution would be to turn system in a multi-class system by providing it with detailed annotation of all possible sub classes. Again, this is another issue to expertly label all classes which does not seem practical.</p><p>In medical image analysis useful information is not only just contained within images themselves. Doctors often see the history of the patient, age and other attributes to arrive at a better decision. Some researches have been conducted to include such features besides images in Deep Learning, but as results have shown it has not been so effective. One of the challenges is to balance the number of imaging features in the deep learning network with clinical features to prevent clinical features from being ignored.</p><p>Although most of the challenges mentioned above have not been properly tackled yet, several high profile successes of Deep Learning in medical imaging can be seen in some recent papers of <em>Esteva et al.(2017)</em> and <em>Gulshan et al.(2016)</em> in the field of dermatology and ophthalmology. Looking at the trend one can infer that <strong>Unsupervised learning</strong> is gaining popularity in this field as they allow training on unlabeled data. Many consider applying automated system in medical field may give rise to some legal question i.e. who to blame if a machine makes mistake ? But these questions are not coming to haunt us in near future and we can relax for now. Medical imaging in an unexplored area and there are a lot of researches to be conducted, hopefully Deep Learning will have a great impact on medical imaging as a whole.</p><p>some links for CNN :</p><p><a href="https://medium.com/technologymadeeasy/the-best-explanation-of-convolutional-neural-networks-on-the-internet-fbb8b1ad5df8">The best explanation of Convolutional Neural Networks on the Internet!</a></p><p><a href="https://towardsdatascience.com/convolutional-neural-networks-from-the-ground-up-c67bb41454e1">https://towardsdatascience.com/convolutional-neural-networks-from-the-ground-up-c67bb41454e1</a></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=810df940d2f7" width="1" height="1" alt="">]]></content:encoded>
        </item>
    </channel>
</rss>