<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0">
  <channel>
    <title>Piccole Gioie </title>
    <link>https://livelovelet.tistory.com/</link>
    <description>Piccole Gioie: Finding Happiness in the Small Things!!
-
Goditi ogni momento del cammino, qualunque cosa ti capiti e ovunque ti porti. 
Un giorno, guarderai indietro e sarai contento di non averlo sprecato.
================================================</description>
    <language>ko</language>
    <pubDate>Fri, 10 Apr 2026 08:48:48 +0900</pubDate>
    <generator>TISTORY</generator>
    <ttl>100</ttl>
    <managingEditor>LiveLoveFlow</managingEditor>
    <item>
      <title>Markov chain Monte Carlo</title>
      <link>https://livelovelet.tistory.com/entry/Markov-chain-Monte-Carlo-1</link>
      <description>&lt;p data-ke-size=&quot;size16&quot;&gt;Markov chain Monte Carlo is a powerful technique that combines the concepts of Markov chain and Monte Carlo methods.&amp;nbsp;It provides a way to sample from otherwise intractable probability distributions. MCMC methods are statistical techniques that provide a way to sample from complex, high-dimensional probability distributions that are otherwise intractable or difficult to analyze mathematically. MCMC methods are particularly useful in Bayesian inference, a framework for updating beliefs or making predictions based on prior knowledge and observed data.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;MCMC methods arise from constructing a Markov chain - a sequence of random variables where the probability of each variable depends on the last variable in the sequence. The MCMC method aims to generate samples from a target probability distribution, often a posterior distribution in Bayesian inference. The key idea is constructing a Markov Chain whose stationary distribution equals the target distribution. We obtain a representative set of samples from the target distribution by iteratively generating samples from this Markov chain.&amp;nbsp;&lt;/p&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;548&quot; data-origin-height=&quot;395&quot;&gt;&lt;span data-url=&quot;https://blog.kakaocdn.net/dn/kFivh/btsgFvW4tX7/6LbP5hD4LkQ5ZPVAsSr0t0/img.png&quot; data-phocus=&quot;https://blog.kakaocdn.net/dn/kFivh/btsgFvW4tX7/6LbP5hD4LkQ5ZPVAsSr0t0/img.png&quot; data-alt=&quot;https://www.google.com/url?sa=i&amp;amp;amp;url=https%3A%2F%2Fwww.researchgate.net%2Ffigure%2FIllustration-of-Markov-Chain-Monte-Carlo-method_fig1_334001505&amp;amp;amp;psig=AOvVaw376MgcRDqSyf1XOBM9scPy&amp;amp;amp;ust=1684582332392000&amp;amp;amp;source=images&amp;amp;amp;cd=vfe&amp;amp;amp;ved=0CBMQjhxqFwoTCJiVlK6kgf8CFQAAAAAdAAAAABA5&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/kFivh/btsgFvW4tX7/6LbP5hD4LkQ5ZPVAsSr0t0/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2FkFivh%2FbtsgFvW4tX7%2F6LbP5hD4LkQ5ZPVAsSr0t0%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;548&quot; height=&quot;395&quot; data-origin-width=&quot;548&quot; data-origin-height=&quot;395&quot;/&gt;&lt;/span&gt;&lt;figcaption&gt;https://www.google.com/url?sa=i&amp;amp;url=https%3A%2F%2Fwww.researchgate.net%2Ffigure%2FIllustration-of-Markov-Chain-Monte-Carlo-method_fig1_334001505&amp;amp;psig=AOvVaw376MgcRDqSyf1XOBM9scPy&amp;amp;ust=1684582332392000&amp;amp;source=images&amp;amp;cd=vfe&amp;amp;ved=0CBMQjhxqFwoTCJiVlK6kgf8CFQAAAAAdAAAAABA5&lt;/figcaption&gt;
&lt;/figure&gt;
&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;The main advantage of MCMC methods is their ability to explore and sample from intractable probability distributions with no closed-form solution. This makes MCMC particularly valuable in cases where direct sampling or analytical approaches are not feasible.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Metropolis-Hastings algorithm&lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;1) Start with an initial sample from the target distribution.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;2) Propose a new sample by randomly perturbing the current sample&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;3) Evaluate the acceptance probability for the proposed sample based on the ratio of the target distribution values&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;4) Accept the proposed sample with a certain probability, or reject it and keep the current sample&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;5) Repeat steps 2 to 4 for a large number of iterations to obtain a sequence of samples&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;The&amp;nbsp;Metropolis-Hastings&amp;nbsp;algorithm&amp;nbsp;leverages&amp;nbsp;the&amp;nbsp;Markov&amp;nbsp;chain&amp;nbsp;property&amp;nbsp;to&amp;nbsp;explore&amp;nbsp;the&amp;nbsp;target&amp;nbsp;distribution&amp;nbsp;iteratively.&amp;nbsp;By&amp;nbsp;generating&amp;nbsp;proposals&amp;nbsp;and&amp;nbsp;accepting&amp;nbsp;or&amp;nbsp;rejecting&amp;nbsp;them&amp;nbsp;based&amp;nbsp;on&amp;nbsp;the&amp;nbsp;acceptance&amp;nbsp;probability,&amp;nbsp;the&amp;nbsp;algorithm&amp;nbsp;allows&amp;nbsp;for&amp;nbsp;efficient&amp;nbsp;sampling&amp;nbsp;from&amp;nbsp;high-dimensional&amp;nbsp;and&amp;nbsp;complicated&amp;nbsp;distributions.&amp;nbsp;Over&amp;nbsp;a&amp;nbsp;sufficient&amp;nbsp;number&amp;nbsp;of&amp;nbsp;iterations,&amp;nbsp;the&amp;nbsp;sequence&amp;nbsp;of&amp;nbsp;samples&amp;nbsp;produced&amp;nbsp;by&amp;nbsp;the&amp;nbsp;algorithm&amp;nbsp;converges&amp;nbsp;to&amp;nbsp;a&amp;nbsp;representative&amp;nbsp;sample&amp;nbsp;from&amp;nbsp;the&amp;nbsp;target&amp;nbsp;distribution.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;In Bayesian Inference, MCMC methods explore the posterior distribution of model parameters given observed data. By sampling from the posterior distribution, one can estimate the uncertainty in parameter values, assess model fit, and make predictions. It allows for the incorporation of prior knowledge and the quantification of uncertainty.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;In the Random-Walk Metropolis-Hastings (RW-MH) algorithms are accepted a proposed move from a point   to a point  ' with probability min ( 1,  ( | ') ( ')/ ( '| ) ( ) ).&amp;nbsp;&lt;/p&gt;
&lt;table style=&quot;color: #374151; text-align: left; border-collapse: collapse; width: 100%; height: 316px;&quot; border=&quot;1&quot; data-ke-align=&quot;alignLeft&quot;&gt;
&lt;tbody&gt;
&lt;tr style=&quot;height: 54px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 54px;&quot;&gt;Proposal &lt;br /&gt;Distribution&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 54px;&quot;&gt;The distribution is used to generate candidate points in the Markov chain. RW-MH often follows a random walk pattern, where the new point is obtained by perturbing the current point according to a specific distribution.&lt;/td&gt;
&lt;/tr&gt;
&lt;tr style=&quot;height: 36px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 36px;&quot;&gt; (&amp;theta;|&amp;theta;')&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 36px;&quot;&gt;The&lt;b&gt; density of proposing &amp;theta;' given &amp;theta;&lt;/b&gt;. It represents the &lt;b&gt;likelihood of transitioning from the current point &amp;theta; to the proposed point &amp;theta;'&lt;/b&gt;.&lt;/td&gt;
&lt;/tr&gt;
&lt;tr style=&quot;height: 36px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 36px;&quot;&gt; (&amp;theta;'|&amp;theta;)&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 36px;&quot;&gt;The density of proposing &amp;theta; given &amp;theta;'. It represents the &lt;b&gt;likelihood of transitioning from the proposed point &amp;theta;' back to the current point &amp;theta;.&lt;/b&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr style=&quot;height: 54px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 54px;&quot;&gt;Target &lt;br /&gt;Distribution&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 54px;&quot;&gt;The distribution from which we want to sample. Often, it is the posterior distribution in Bayesian inference. &amp;eta;(&amp;theta;) represents the density of the target distribution at point &amp;theta;, and &amp;eta;(&amp;theta;') represents the density at the proposed point &amp;theta;'.&lt;/td&gt;
&lt;/tr&gt;
&lt;tr style=&quot;height: 36px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 36px;&quot;&gt; (&amp;theta;)&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 36px;&quot;&gt;The density of the &lt;b&gt;target distribution at point &amp;theta; (The current state or parameter)&lt;/b&gt;. It indicates the relative probability of observing &amp;theta; according to the target distribution.&lt;/td&gt;
&lt;/tr&gt;
&lt;tr style=&quot;height: 36px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 36px;&quot;&gt; (&amp;theta;')&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 36px;&quot;&gt;The density of the&lt;b&gt; target distribution at the proposed point &amp;theta;' (New proposed state or parameter).&lt;/b&gt; It indicates the relative probability of observing &amp;theta;' according to the target distribution.&lt;/td&gt;
&lt;/tr&gt;
&lt;tr style=&quot;height: 54px;&quot;&gt;
&lt;td style=&quot;width: 11.744186%; height: 54px;&quot;&gt;Acceptance Probability&lt;/td&gt;
&lt;td style=&quot;width: 88.139535%; height: 54px;&quot;&gt;The probability of accepting a proposed move from point &amp;theta; to point &amp;theta;', determined by the formula min(1, (ϕ(&amp;theta;|&amp;theta;')&amp;eta;(&amp;theta;')) / (ϕ(&amp;theta;'|&amp;theta;)&amp;eta;(&amp;theta;))). It ensures that the Markov chain transitions between points with a probability proportional to the target distribution.&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;min( 1, x ) denotes taking the minimum between 1 and x. This means that this will ensure the proposed state  ' is accepted with a probability of at least one or proportional to the target and proposal densities ratio.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;This acceptance probability determines whether the proposed state is incorporated into Markov Chain or Rejected.&amp;nbsp;&lt;/p&gt;</description>
      <category>Data science/Machine Learning</category>
      <category>MCMC</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/95</guid>
      <comments>https://livelovelet.tistory.com/entry/Markov-chain-Monte-Carlo-1#entry95comment</comments>
      <pubDate>Sat, 20 May 2023 04:28:36 +0900</pubDate>
    </item>
    <item>
      <title>Unsupervised or Supervised Classification</title>
      <link>https://livelovelet.tistory.com/entry/Unsupervised-or-Supervised-Classification</link>
      <description>&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;b&gt;Unsupervised&amp;nbsp;or&amp;nbsp;Supervised&amp;nbsp;Classification&lt;/b&gt;&lt;/h2&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;1830&quot; data-origin-height=&quot;596&quot;&gt;&lt;span data-url=&quot;https://blog.kakaocdn.net/dn/lZXZ5/btsguiYLsVW/D8YiZ71MebRqkAnrgWCX8k/img.png&quot; data-phocus=&quot;https://blog.kakaocdn.net/dn/lZXZ5/btsguiYLsVW/D8YiZ71MebRqkAnrgWCX8k/img.png&quot; data-alt=&quot;Comparison Table&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/lZXZ5/btsguiYLsVW/D8YiZ71MebRqkAnrgWCX8k/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2FlZXZ5%2FbtsguiYLsVW%2FD8YiZ71MebRqkAnrgWCX8k%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;1830&quot; height=&quot;596&quot; data-origin-width=&quot;1830&quot; data-origin-height=&quot;596&quot;/&gt;&lt;/span&gt;&lt;figcaption&gt;Comparison Table&lt;/figcaption&gt;
&lt;/figure&gt;
&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;This post will focus on classification algorithms, such as K-means, model-based clustering, K nearest neighbours and probabilistic classifiers.&amp;nbsp;&lt;b&gt;&lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;table style=&quot;border-collapse: collapse; width: 100.232558%; height: 15px;&quot; border=&quot;1&quot; data-ke-align=&quot;alignLeft&quot; data-ke-style=&quot;style14&quot;&gt;
&lt;tbody&gt;
&lt;tr style=&quot;height: 18px;&quot;&gt;
&lt;td style=&quot;width: 100%; height: 18px;&quot;&gt;&amp;nbsp;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;h3 data-ke-size=&quot;size23&quot;&gt;&lt;b&gt;K-means&amp;nbsp;&lt;/b&gt;&lt;/h3&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;span style=&quot;color: #374151; text-align: start;&quot;&gt;K-means clustering is a popular unsupervised machine learning algorithm to partition a dataset into k distinct clusters. It aims to minimize the within-cluster sum of squares by assigning data points to clusters based on their proximity to cluster centres.&amp;nbsp;&lt;/span&gt;* distance can be measured by the standard Euclidean distance: d(x_j, x_n) = || x_j - x_k ||₂&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;-&amp;gt; which makes the total within-cluster sum of squares minimized.&lt;/p&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;601&quot; data-origin-height=&quot;844&quot;&gt;&lt;span data-url=&quot;https://blog.kakaocdn.net/dn/cSjKQW/btsgsQPDwVC/nKnJlwEHtuKfpM3YZjSBQ0/img.png&quot; data-phocus=&quot;https://blog.kakaocdn.net/dn/cSjKQW/btsgsQPDwVC/nKnJlwEHtuKfpM3YZjSBQ0/img.png&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/cSjKQW/btsgsQPDwVC/nKnJlwEHtuKfpM3YZjSBQ0/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2FcSjKQW%2FbtsgsQPDwVC%2FnKnJlwEHtuKfpM3YZjSBQ0%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;267&quot; height=&quot;375&quot; data-origin-width=&quot;601&quot; data-origin-height=&quot;844&quot;/&gt;&lt;/span&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;How to do the K-means?&lt;/b&gt;&lt;/p&gt;
&lt;ol style=&quot;list-style-type: decimal; color: #374151; text-align: start;&quot; data-ke-list-type=&quot;decimal&quot;&gt;
&lt;li&gt;Initialization: Start by selecting k initial cluster centres. Various methods can be used, such as choosing random points, points from the dataset, or a centrally located point.&lt;/li&gt;
&lt;li&gt;Assignment Step: For each data point, calculate the Euclidean distance to each cluster centre and assign the point to the cluster with the closest centre. The distance can be computed using the standard Euclidean distance formula.&lt;/li&gt;
&lt;li&gt;Update Step: After assigning all points to clusters, update the cluster centres by calculating the mean of the data points within each cluster. The new cluster centre represents the average position of the data points in that cluster.&lt;/li&gt;
&lt;li&gt;Convergence Check: Check if the cluster centres have converged. If the positions of the cluster centres remain unchanged or the change is below a predefined threshold, the algorithm stops. Otherwise, repeat steps 2 and 3.&lt;/li&gt;
&lt;/ol&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** Clustering depends on the initial centres. Thus, run the algorithms with different initial centres, then choose the result with the smallest sum of squares.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** To decide how many clusters to keep, using an elbow plot can be helpful. Choose the value of k where the break occurs in the slope.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;How to choose centres?&amp;nbsp;&lt;/b&gt;&lt;/p&gt;
&lt;ul style=&quot;list-style-type: disc; color: #374151; text-align: start;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Random Initialization: Choose k points uniformly and randomly from the data space. This approach is simple and suitable when the data distribution is uniform.&lt;/li&gt;
&lt;li&gt;Forgy's Method: Select k points from the dataset as the initial cluster centres. This method ensures that the initial centres are representative of the data points.&lt;/li&gt;
&lt;li&gt;Random Partition: Assign points to clusters uniformly at random, with each cluster initially having approximately n/k points. Use the cluster means as the initial centres.&lt;/li&gt;
&lt;li&gt;Centrally-Located Point: Choose the most centrally located point in the dataset as the first centre and subsequently select the points with the largest minimum distance to all existing centres.&lt;/li&gt;
&lt;li&gt;Maximin: Begin by selecting a centre arbitrarily and then choose subsequent centres as the data points with the largest minimum distance to all existing centres. This method ensures that the initial centres are well-separated.&lt;/li&gt;
&lt;/ul&gt;
&lt;div style=&quot;color: #000000; text-align: start;&quot;&gt;
&lt;div&gt;
&lt;div&gt;
&lt;div&gt;
&lt;div style=&quot;background-color: #000000; color: #000000;&quot;&gt;
&lt;div&gt;
&lt;div&gt;
&lt;div&gt;
&lt;div&gt;
&lt;div style=&quot;color: #000000;&quot;&gt;&amp;nbsp;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Determining&amp;nbsp;the&amp;nbsp;Number&amp;nbsp;of&amp;nbsp;Clusters:&lt;/b&gt;&lt;br /&gt;The&amp;nbsp;choice&amp;nbsp;of&amp;nbsp;the&amp;nbsp;number&amp;nbsp;of&amp;nbsp;clusters,&amp;nbsp;k,&amp;nbsp;is&amp;nbsp;crucial.&amp;nbsp;One&amp;nbsp;approach&amp;nbsp;uses&amp;nbsp;an&lt;span style=&quot;background-color: #9feec3;&quot;&gt;&amp;nbsp;elbow&amp;nbsp;plot&lt;/span&gt;,&amp;nbsp;which&amp;nbsp;displays&amp;nbsp;the&amp;nbsp;relationship&amp;nbsp;between&amp;nbsp;the&amp;nbsp;number&amp;nbsp;of&amp;nbsp;clusters&amp;nbsp;and&amp;nbsp;the&amp;nbsp;within-cluster&amp;nbsp;sum&amp;nbsp;of&amp;nbsp;squares.&amp;nbsp;Identify&amp;nbsp;the&amp;nbsp;value&amp;nbsp;of&amp;nbsp;k&amp;nbsp;where&amp;nbsp;the&amp;nbsp;plot&amp;nbsp;shows&amp;nbsp;a&amp;nbsp;significant&amp;nbsp;break&amp;nbsp;or&amp;nbsp;reduction&amp;nbsp;in&amp;nbsp;the&amp;nbsp;slope,&amp;nbsp;indicating&amp;nbsp;a&amp;nbsp;good&amp;nbsp;balance&amp;nbsp;between&amp;nbsp;compactness&amp;nbsp;and&amp;nbsp;separation&amp;nbsp;of&amp;nbsp;clusters.&lt;br /&gt;&lt;br /&gt;&lt;/p&gt;
&lt;table style=&quot;border-collapse: collapse; width: 100%;&quot; border=&quot;1&quot; data-ke-align=&quot;alignLeft&quot; data-ke-style=&quot;style14&quot;&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td style=&quot;width: 100%;&quot;&gt;&amp;nbsp;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;h3 data-ke-size=&quot;size23&quot;&gt;&lt;b&gt;Model-based clustering; EM algorithms&lt;/b&gt;&lt;/h3&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Model-based clustering is a powerful technique that allows us to discover underlying patterns in data by fitting a probabilistic model to the observed data. The Expectation-Maximization (EM) algorithms are commonly used to estimate the model's parameters in a maximum likelihood framework.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Model-based Clustering involves &lt;span style=&quot;background-color: #f6e199;&quot;&gt;specifying the number of clusters in advance&lt;/span&gt; and &lt;span style=&quot;background-color: #f6e199;&quot;&gt;then fitting a probabilistic model&lt;/span&gt; to the data. A probability distribution represents each cluster, and the goal is to estimate the parameters of these distributions that best explain the data.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;EM Algorithms:&lt;/b&gt; The EM algorithm is an iterative optimization technique to estimate the model parameters. It consists of two steps- The expectation and Maximization steps.&amp;nbsp;&lt;/p&gt;
&lt;ol style=&quot;list-style-type: decimal; color: #374151; text-align: start;&quot; data-ke-list-type=&quot;decimal&quot;&gt;
&lt;li&gt;Initialization: Start by assigning points to clusters uniformly at random and construct an initial class membership matrix, z, using one-hot encoding.&lt;/li&gt;
&lt;li&gt;Expectation step: Compute the mixture parameters, such as the mixing coefficients ( ), mean values ( ), and covariance matrices ( ), based on the current estimates.&lt;/li&gt;
&lt;li&gt;Maximization step: Update the cluster assignment probabilities using the newly estimated mixture parameters. This involves calculating the probabilities of points belonging to each cluster based on the model's parameters.&lt;/li&gt;
&lt;li&gt;Convergence check: If the cluster assignment probabilities, z_ij, are not changing significantly, stop the algorithm. Otherwise, repeat steps 2 and 3 until convergence is achieved.&lt;/li&gt;
&lt;/ol&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Benefits of Model-based Clustering&lt;/b&gt;&lt;/p&gt;
&lt;ul style=&quot;list-style-type: disc; color: #374151; text-align: start;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Flexibility: Model-based clustering allows for more flexible cluster shapes and can accommodate various types of data distributions.&lt;/li&gt;
&lt;li&gt;Uncertainty Estimation: The algorithm provides probabilities of class membership for each data point, allowing uncertainty quantification.&lt;/li&gt;
&lt;li&gt;Scalability: Model-based clustering can handle large datasets using efficient algorithms and approximation techniques.&lt;/li&gt;
&lt;/ul&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Choosing the number of clusters&lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Selecting the appropriate number of clusters is a crucial step in model-based clustering. BIC or cross-validation can be used to determine the optimal number of clusters that best balance model complexity and goodness of fit to the data.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;table style=&quot;border-collapse: collapse; width: 100%;&quot; border=&quot;1&quot; data-ke-align=&quot;alignLeft&quot; data-ke-style=&quot;style13&quot;&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td style=&quot;width: 100%;&quot;&gt;&amp;nbsp;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;K-nearest neighbours&lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;K-nearest neighbours (KNN) is a simple yet effective classification algorithm that makes predictions based on the proximity of training data points to a test feature vector.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;1) Given a test feature vector x, the KNN algorithm identifies the k nearest neighbours in the training set based on a specified distance metric. The distance metric can be Euclidean distance, Manhattan distance or any other suitable similarity measure.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;2) Take a majority vote from those K for the predicted class of the test datapoint ( K should be odd).&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;KNN offers several advantages, including its simplicity and flexibility. It can handle both classification and regression tasks and adapt to various data types.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** The choice of K: A small value of K may lead to overfitting, while a large value of K may result in underfitting (smooth boundary, but high bias and low variance, reduce noise impact).&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** KNN relies on distance calculations; it is crucial to normalize or scale the features to ensure that no single feature dominates the distance calculation. ( z-score normalization or min-max scaling)&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** Recommendation systems, image recognition and anomaly detection are useful when the decision boundaries are complex, or the training data is unbalanced.&amp;nbsp;&lt;/p&gt;
&lt;table style=&quot;border-collapse: collapse; width: 100%;&quot; border=&quot;1&quot; data-ke-align=&quot;alignLeft&quot; data-ke-style=&quot;style13&quot;&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td style=&quot;width: 100%;&quot;&gt;&amp;nbsp;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Probabilistic Classifiers&amp;nbsp;&lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Probabilistic Classifiers are a class of machine learning algorithms that assign probabilities to different classes or categories for a given input. Instead of providing a single predicted class, probabilistic classifiers estimate the &lt;span style=&quot;background-color: #f6e199;&quot;&gt;likelihood or probability of an input belonging to each class.&lt;/span&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;1) Probability estimation: Probabilistic classifiers provide a probability distribution over all possible classes for a given input.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;2) Decision boundary separates different classes based on the estimated probabilities.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;3) Training process: The algorithm learns the relationship between input features and class labels using a labelled dataset. The process involves estimating the parameters of the probability distribution for each class.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;4) Bayesian framework: Often based on the Bayesian framework, which incorporates prior knowledge and updates it with observed data using Bayes' theorem.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;5) Evaluation: Evaluated using various metrics such as prolonged loss, accuracy, precision, recall and F1-score.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** Logistic regression, Naive Bayes and support vector machines with probabilistic outputs.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;** Spam detection, sentiment analysis, medical diagnosis, and fraud detection&lt;/p&gt;</description>
      <category>Data science/Machine Learning</category>
      <category>emalgorithms</category>
      <category>kmeans</category>
      <category>KNN</category>
      <category>modelbasedclustering</category>
      <category>Probabilistic Classifiers</category>
      <category>supervisedclassification</category>
      <category>unsupervisedclassification</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/93</guid>
      <comments>https://livelovelet.tistory.com/entry/Unsupervised-or-Supervised-Classification#entry93comment</comments>
      <pubDate>Fri, 19 May 2023 08:50:54 +0900</pubDate>
    </item>
    <item>
      <title>PCA (Principal Components Analysis)</title>
      <link>https://livelovelet.tistory.com/entry/PCA-Principal-Components-Analysis</link>
      <description>&lt;p data-ke-size=&quot;size16&quot;&gt;Principal Components Analysis (PCA) is a popular technique for analysing high-dimensional data. PCA is a mathematical procedure that transforms a set of correlated variables into a new set of uncorrelated variables, called &quot;Principal Components&quot; while retaining most of the variability of the original data.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;To perform PCA on a dataset of N, p-dimensional data point &lt;span style=&quot;background-color: #f6e199;&quot;&gt;x_j &amp;isin; Rᴾ&lt;/span&gt;, we need to follow these steps.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;1. &lt;b&gt;Standardize&lt;/b&gt; the data: To ensure all variables are in a similar range, subtract the mean and divide by the standard deviation for each variable.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;2. &lt;b&gt;Compute the covariance matrix&lt;/b&gt;: The covariance matrix summarizes the relationships between the variables in the data. The covariance matrix can be computed by using the formula, &quot;&lt;span style=&quot;color: #000000; background-color: #f6e199;&quot;&gt;cov(x)= 1/(n-1)* XᵀX&lt;/span&gt;&quot; (X is the standardized data matrix).&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;3. &lt;b&gt;Compute the eigenvectors and eigenvalues (Eigensystem&lt;/b&gt;): These represent the directions and magnitude of the maximum data variability. (   and   )&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;4. &lt;b&gt;Sort the eigenvectors in descending&lt;/b&gt; order of their corresponding eigenvalues: The eigenvectors with the highest eigenvalues represent the directions of maximum variability in the data. These directions are called &quot;Principal components&quot;.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;5. &lt;b&gt;Choose the number of principal components&lt;/b&gt;: Use a scree plot to determine the number of components to keep. (scree plot: Plot of eigenvalues against the number of components)&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;6. &lt;b&gt;Compute the principal components:&lt;/b&gt; by multiplying the standardized data matrix by the eigenvectors of the covariance matrix&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;7. &lt;b&gt;Calculate the percentage of the total sample variance explained by the i-th principal components&lt;/b&gt; (  _i / &amp;sum;  _j ) * 100 %, j-th element of the i-th eigenvector v_i is called &quot;loading&quot; of the j-th variable onto the i-th principal component.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;(*Loading: Relationship between the original variables and principal components. ; PC_i = v_i1 * x_1 + ... + v_ip *x_p )&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;The result of PCA will produce principal components, which are linear combinations of original variables. These PCs will be uncorrelated.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;PCA can be used to reduce the dimensionality of data, making it easier to visualize and analyze. It can also be used to identify patterns and relationships in the data (correlation), and data compression.&amp;nbsp;&lt;/p&gt;</description>
      <category>Data science/Machine Learning</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/92</guid>
      <comments>https://livelovelet.tistory.com/entry/PCA-Principal-Components-Analysis#entry92comment</comments>
      <pubDate>Fri, 19 May 2023 02:41:48 +0900</pubDate>
    </item>
    <item>
      <title>Batch Normalization for Deep Learning</title>
      <link>https://livelovelet.tistory.com/entry/Batch-Normalization-for-Deep-Learning</link>
      <description>&lt;p data-ke-size=&quot;size16&quot;&gt;*** This post is a summary version of the below website. The website has more details and references.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;a href=&quot;https://machinelearningmastery.com/batch-normalization-for-training-of-deep-neural-networks/&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;https://machinelearningmastery.com/batch-normalization-for-training-of-deep-neural-networks/&lt;/a&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;One possible reason for the difficulties of having &lt;i&gt;changeable inputs&lt;/i&gt; to layers deep in the network(&quot;Internal Covariate Shift) is due to updating weights per mini-batch. Batch Normalisation can stabilise the learning process and reduce the number of training epochs required.&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;b&gt;Batch Normalisation&lt;/b&gt;&lt;/h2&gt;
&lt;p data-ke-size=&quot;size18&quot;&gt;: scales the layer output by standardising each input variable's activations per mini-batch. This is done by subtracting the mean of the batch from each value and dividing it by the standard deviation of the batch. This way, the rescaled data would have a mean of zero and a standard deviation of one. Therefore, it will allow all input variables to contribute to the model equally. &quot;&lt;b&gt;Whitening&lt;/b&gt;&quot; is another term for Batch Norm.&amp;nbsp;&lt;/p&gt;
&lt;h2 data-ke-size=&quot;size26&quot;&gt;&amp;nbsp;&lt;/h2&gt;
&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;b&gt;Use cases:&lt;/b&gt;&lt;/h2&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;- Dramatic speed improvement of an inception-based convolutional neural network for photo classification&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;- Used it after each convolution and before activation for a standard photo classification, significant outcome&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;- Used in the updated inception model (GoogleNet Inception-v3) for the ImageNet dataset, an incredible outcome&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;- Used in the recurrent neural networks in their end-to-end model for speech recognition, improve the final generalization error and accelerating training&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;b&gt;How to Use&amp;nbsp;&lt;/b&gt;&lt;/h2&gt;
&lt;ol style=&quot;list-style-type: decimal;&quot; data-ke-list-type=&quot;decimal&quot;&gt;
&lt;li&gt;Use before or after the activation function in the previous layer:
&lt;ol style=&quot;list-style-type: decimal;&quot; data-ke-list-type=&quot;decimal&quot;&gt;
&lt;li&gt;after it, if s-shaped functions ( logistic, hyperbolic tangent )&amp;nbsp;&lt;/li&gt;
&lt;li&gt;before it, if it results in non-Gaussian distributions like rectified linear activation function&amp;nbsp;&lt;/li&gt;
&lt;/ol&gt;
&lt;/li&gt;
&lt;li&gt;Use a Higher Learning Rate: Observed that the training speedup with higher learning rates.&lt;/li&gt;
&lt;li&gt;It can be used in data preprocessing purposes if the data has different scaled variables.&lt;/li&gt;
&lt;li&gt;DO NOT use Dropout: As it reduces the generalization error, dropout is not required.&amp;nbsp;&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;1422&quot; data-origin-height=&quot;668&quot;&gt;&lt;span data-url=&quot;https://blog.kakaocdn.net/dn/MGhdL/btr8axPwpLq/koWrUzcOHKmuVLrJjnKmB0/img.png&quot; data-phocus=&quot;https://blog.kakaocdn.net/dn/MGhdL/btr8axPwpLq/koWrUzcOHKmuVLrJjnKmB0/img.png&quot; data-alt=&quot;https://gaussian37.github.io/dl-concept-batchnorm/&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/MGhdL/btr8axPwpLq/koWrUzcOHKmuVLrJjnKmB0/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2FMGhdL%2Fbtr8axPwpLq%2FkoWrUzcOHKmuVLrJjnKmB0%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;1422&quot; height=&quot;668&quot; data-origin-width=&quot;1422&quot; data-origin-height=&quot;668&quot;/&gt;&lt;/span&gt;&lt;figcaption&gt;https://gaussian37.github.io/dl-concept-batchnorm/&lt;/figcaption&gt;
&lt;/figure&gt;
&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;</description>
      <category>Data science</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/91</guid>
      <comments>https://livelovelet.tistory.com/entry/Batch-Normalization-for-Deep-Learning#entry91comment</comments>
      <pubDate>Wed, 5 Apr 2023 06:21:37 +0900</pubDate>
    </item>
    <item>
      <title>Some quotes that I like related to Happiness</title>
      <link>https://livelovelet.tistory.com/entry/Some-quotes-that-I-like-related-to-Happiness</link>
      <description>&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;b&gt;Keywords&amp;nbsp;&lt;/b&gt;&lt;/h2&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Verbs&lt;/b&gt;&lt;span style=&quot;letter-spacing: 0px;&quot;&gt;:&lt;/span&gt;&lt;/p&gt;
&lt;p style=&quot;color: #333333; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;essere&amp;nbsp;(to&amp;nbsp;be)&lt;br /&gt;avere&amp;nbsp;(to&amp;nbsp;have)&lt;br /&gt;godere&amp;nbsp;(to&amp;nbsp;enjoy)&lt;br /&gt;trovare&amp;nbsp;(to&amp;nbsp;find)&lt;br /&gt;fare&amp;nbsp;(to&amp;nbsp;make/do)&lt;br /&gt;pensare&amp;nbsp;(to&amp;nbsp;think)&lt;br /&gt;aspettare&amp;nbsp;(to&amp;nbsp;wait)&lt;br /&gt;prendere&amp;nbsp;(to&amp;nbsp;take)&lt;br /&gt;durare&amp;nbsp;(to&amp;nbsp;last)&lt;br /&gt;vivere&amp;nbsp;(to&amp;nbsp;live)&lt;br /&gt;guardare&amp;nbsp;(to&amp;nbsp;look)&lt;br /&gt;rimpiangere&amp;nbsp;(to&amp;nbsp;regret)&lt;br /&gt;sprecare&amp;nbsp;(to&amp;nbsp;waste)&lt;/p&gt;
&lt;p style=&quot;color: #333333; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;&lt;br /&gt;&lt;b&gt;Subjects&lt;/b&gt;:&lt;br /&gt;la&amp;nbsp;felicit&amp;agrave;&amp;nbsp;(happiness)&lt;br /&gt;le&amp;nbsp;cose&amp;nbsp;belle&amp;nbsp;(beautiful&amp;nbsp;things)&lt;br /&gt;l'occasione&amp;nbsp;(opportunity)&lt;br /&gt;la&amp;nbsp;vita&amp;nbsp;(life)&lt;br /&gt;il&amp;nbsp;cammino&amp;nbsp;(journey)&lt;/p&gt;
&lt;p style=&quot;color: #333333; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;h2 style=&quot;color: #333333; text-align: start;&quot; data-ke-size=&quot;size26&quot;&gt;&lt;b&gt;Phrases&lt;/b&gt;&lt;/h2&gt;
&lt;ol style=&quot;list-style-type: decimal; color: #374151; text-align: start;&quot; data-ke-list-type=&quot;decimal&quot;&gt;
&lt;li&gt;&quot;La felicit&amp;agrave; &amp;egrave; un viaggio, non una meta.&quot; - Happiness is a journey, not a destination.&lt;/li&gt;
&lt;li&gt;&quot;La felicit&amp;agrave; &amp;egrave; fatta di piccole cose.&quot; - Happiness is made of small things.&lt;/li&gt;
&lt;li&gt;&quot;La felicit&amp;agrave; non &amp;egrave; avere tutto, ma saper godere di ci&amp;ograve; che si ha.&quot; - Happiness does not have everything, but being able to enjoy what you have.&lt;/li&gt;
&lt;li&gt;&quot;La felicit&amp;agrave; non &amp;egrave; qualcosa di gi&amp;agrave; fatto. Viene dalle tue azioni.&quot; - Happiness is not something already made. It comes from your actions.&lt;/li&gt;
&lt;li&gt;&quot;Essere felici non significa avere una vita perfetta, ma riuscire a trovare la perfezione nelle imperfezioni.&quot; - Being happy doesn't mean having a perfect life, but being able to find perfection in imperfections.&lt;/li&gt;
&lt;li&gt;&quot;Fallo e basta, non pensarci troppo. L'importante &amp;egrave; averci provato.&quot; (Do it, don't think too much about it. The important thing is to have tried.)&lt;/li&gt;
&lt;li&gt;&quot;Goditi ogni momento del cammino, qualunque cosa ti capiti e ovunque ti porti. Un giorno, guarderai indietro e sarai contento di non averlo sprecato.&quot; (Enjoy every moment of the journey, whatever happens and wherever it takes you. One day, you will look back and be glad you didn't waste it.)&lt;/li&gt;
&lt;li&gt;&quot;Non aspettare l'occasione perfetta, prendi l'occasione e rendila perfetta.&quot; (Don't wait for the perfect opportunity, take the opportunity and make it perfect.)&lt;/li&gt;
&lt;li&gt;&quot;Le cose belle non durano per sempre, ma i ricordi di esse restano per sempre nella tua mente.&quot; (Beautiful things don't last forever, but the memories of them stay with you forever in your mind.)&lt;/li&gt;
&lt;li&gt;&quot;Vivi la tua vita in modo che, quando la guardi indietro, il tuo unico rimpianto sia di non aver vissuto abbastanza.&quot; (Live your life in such a way that when you look back, your only regret is not having lived enough.)&lt;/li&gt;
&lt;/ol&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;div style=&quot;color: #000000; text-align: start;&quot;&gt;
&lt;div&gt;
&lt;div&gt;
&lt;div&gt;&amp;nbsp;&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;/div&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;</description>
      <category>Language/Italian</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/90</guid>
      <comments>https://livelovelet.tistory.com/entry/Some-quotes-that-I-like-related-to-Happiness#entry90comment</comments>
      <pubDate>Tue, 4 Apr 2023 23:04:07 +0900</pubDate>
    </item>
    <item>
      <title>Lecture Note - Parameter Tuning (Deep Learning) - TBU</title>
      <link>https://livelovelet.tistory.com/entry/Lecture-Note-Parameter-Tuning-Deep-Learning-TBU</link>
      <description>&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;a href=&quot;https://south-crown-f53.notion.site/Lecture-Note-Parameter-Tuning-Deep-Learning-109c66b3094c4bf08af1710947a5ea88&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot;&gt;https://south-crown-f53.notion.site/Lecture-Note-Parameter-Tuning-Deep-Learning-109c66b3094c4bf08af1710947a5ea88&lt;/a&gt;&lt;/p&gt;
&lt;figure id=&quot;og_1680127838357&quot; contenteditable=&quot;false&quot; data-ke-type=&quot;opengraph&quot; data-ke-align=&quot;alignCenter&quot; data-og-type=&quot;article&quot; data-og-title=&quot;Lecture Note - Parameter Tuning (Deep Learning)&quot; data-og-description=&quot;https://www.youtube.com/watch?v=1waHlpKiNyY&amp;amp;list=PLkDaE6sCZn6Hn0vK8co82zjQtt3T2Nkqc&quot; data-og-host=&quot;south-crown-f53.notion.site&quot; data-og-source-url=&quot;https://south-crown-f53.notion.site/Lecture-Note-Parameter-Tuning-Deep-Learning-109c66b3094c4bf08af1710947a5ea88&quot; data-og-url=&quot;https://south-crown-f53.notion.site/109c66b3094c4bf08af1710947a5ea88&quot; data-og-image=&quot;&quot;&gt;&lt;a href=&quot;https://south-crown-f53.notion.site/Lecture-Note-Parameter-Tuning-Deep-Learning-109c66b3094c4bf08af1710947a5ea88&quot; target=&quot;_blank&quot; rel=&quot;noopener&quot; data-source-url=&quot;https://south-crown-f53.notion.site/Lecture-Note-Parameter-Tuning-Deep-Learning-109c66b3094c4bf08af1710947a5ea88&quot;&gt;
&lt;div class=&quot;og-image&quot; style=&quot;background-image: url();&quot;&gt;&amp;nbsp;&lt;/div&gt;
&lt;div class=&quot;og-text&quot;&gt;
&lt;p class=&quot;og-title&quot; data-ke-size=&quot;size16&quot;&gt;Lecture Note - Parameter Tuning (Deep Learning)&lt;/p&gt;
&lt;p class=&quot;og-desc&quot; data-ke-size=&quot;size16&quot;&gt;https://www.youtube.com/watch?v=1waHlpKiNyY&amp;amp;list=PLkDaE6sCZn6Hn0vK8co82zjQtt3T2Nkqc&lt;/p&gt;
&lt;p class=&quot;og-host&quot; data-ke-size=&quot;size16&quot;&gt;south-crown-f53.notion.site&lt;/p&gt;
&lt;/div&gt;
&lt;/a&gt;&lt;/figure&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;</description>
      <category>Data science</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/89</guid>
      <comments>https://livelovelet.tistory.com/entry/Lecture-Note-Parameter-Tuning-Deep-Learning-TBU#entry89comment</comments>
      <pubDate>Thu, 30 Mar 2023 07:10:50 +0900</pubDate>
    </item>
    <item>
      <title>[Book Review] Methods for interpreting the deep neural networks - TBU</title>
      <link>https://livelovelet.tistory.com/entry/Book-Review-Methods-for-interpreting-the-deep-neural-networks-TBU</link>
      <description>&lt;p data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;Interpretation of the deep neural networks&lt;/span&gt;&lt;/span&gt;&lt;/h2&gt;
&lt;p data-ke-size=&quot;size18&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;Deep neural networks can be challenging to interpret and understand with their multiple layers of nonlinear transformations. While they may produce highly accurate predictions, it can be challenging to determine how the model arrived at those predictions and which features of the input data were most important in making the prediction.&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;This summarises a few methods that can be applied to interpreting deep neural networks.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;851&quot; data-origin-height=&quot;317&quot;&gt;&lt;span data-url=&quot;https://blog.kakaocdn.net/dn/DyOcL/btr6ofdfMwX/5Gz7mu4WqPxAqaCYm0fFVk/img.png&quot; data-phocus=&quot;https://blog.kakaocdn.net/dn/DyOcL/btr6ofdfMwX/5Gz7mu4WqPxAqaCYm0fFVk/img.png&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/DyOcL/btr6ofdfMwX/5Gz7mu4WqPxAqaCYm0fFVk/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2FDyOcL%2Fbtr6ofdfMwX%2F5Gz7mu4WqPxAqaCYm0fFVk%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;886&quot; height=&quot;330&quot; data-origin-width=&quot;851&quot; data-origin-height=&quot;317&quot;/&gt;&lt;/span&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;h2 data-ke-size=&quot;size26&quot;&gt;The following is the book &quot;Interpretable Machine Learning Chapter 10 - Neural Network&quot; summary.&amp;nbsp;&lt;/h2&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;[Introduction]&amp;nbsp;&lt;/b&gt;&lt;/p&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Interpretation methods are used to visualize features and concepts learned by neural networks, explain individual predictions, and simplify neural networks&lt;/li&gt;
&lt;li&gt;Deep learning has been successful in image and text tasks, but the mapping from input to prediction is too complex for humans to understand without interpretation methods&lt;/li&gt;
&lt;li&gt;Specific interpretation methods are needed for neural networks because they learn hidden features and concepts and because the gradient can be utilized for more efficient methods&lt;/li&gt;
&lt;li&gt;The following techniques are covered in the following chapters:
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Learned Features&lt;/li&gt;
&lt;li&gt;Pixel Attribution (Saliency Maps)&lt;/li&gt;
&lt;li&gt;Concepts&lt;/li&gt;
&lt;li&gt;Adversarial Examples&lt;/li&gt;
&lt;li&gt;Influential Instances&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;[10.1 Learned Features]&lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;We can use &lt;i&gt;activation maximisation&lt;/i&gt; to visualize the learned features in a convolutional neural network. This involves starting with a random input image and iteratively adjusting its pixel values to &lt;span style=&quot;background-color: #f6e199;&quot;&gt;&lt;i&gt;maximize the activation of a specific feature or neuron in the network&lt;/i&gt;&lt;/span&gt;. By doing this for multiple features or neurons, we can &lt;i&gt;generate images highlighting the learned features in each network layer&lt;/i&gt;. These images can give us insights into what types of patterns and concepts the network is learning.&lt;/span&gt;&lt;/p&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;639&quot; data-origin-height=&quot;536&quot;&gt;&lt;a href=&quot;https://christophm.github.io/interpretable-ml-book/cnn-features.html&quot; target=&quot;_blank&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/l3BGX/btr53E6rKgH/ZnRyXlNwkdVYU1kiiwUpK0/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2Fl3BGX%2Fbtr53E6rKgH%2FZnRyXlNwkdVYU1kiiwUpK0%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;639&quot; height=&quot;536&quot; data-origin-width=&quot;639&quot; data-origin-height=&quot;536&quot;/&gt;&lt;/a&gt;&lt;figcaption&gt;https://christophm.github.io/interpretable-ml-book/cnn-features.html&lt;/figcaption&gt;
&lt;/figure&gt;
&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;[10.1.1. Feature Visualization] &lt;/b&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Feature Visualization is making the learned features explicit in a neural network. It is done by f&lt;span style=&quot;background-color: #f6e199;&quot;&gt;&lt;i&gt;inding the input that maximizes the activation of a unit&lt;/i&gt;&lt;/span&gt;, such as a single neuron, channel, entire layer, or the final class probability in classification. (Feature visualization can be carried out for different units, neurons, channels, layers, hidden layers, and pre-softmax neurons)&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&lt;span style=&quot;color: #ee2323;&quot;&gt;TBD...&amp;nbsp;&lt;/span&gt;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Mathematically, feature visualization is an optimization problem, where the goal is to find a new image that maximizes the (mean) activation of a unit, assuming that the neural network weights are fixed. The optimization can be done by generating new images starting from random noise while applying constraints such as minor changes, jittering, rotation, scaling, frequency penalization, or generating images with learned priors.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;Different units can be used for feature visualization, including individual neurons, channels, entire layers, or the final class probability in classification. Individual neurons are atomic units of the network, but visualizing each neuron&amp;rsquo;s feature would be time-consuming due to the millions of neurons in a network. Channels as units are a good choice for feature visualization, and we can visualize an entire convolutional layer. Layers as a unit are used for Google&amp;rsquo;s DeepDream, which repeatedly adds the visualized features of a layer to the original image, resulting in a dream-like version of the input.&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;TBD...&amp;nbsp;&lt;/p&gt;</description>
      <category>Data science</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/88</guid>
      <comments>https://livelovelet.tistory.com/entry/Book-Review-Methods-for-interpreting-the-deep-neural-networks-TBU#entry88comment</comments>
      <pubDate>Thu, 30 Mar 2023 07:06:44 +0900</pubDate>
    </item>
    <item>
      <title>[Literature Review] Estimating Networks of Sustainable Development Goals</title>
      <link>https://livelovelet.tistory.com/entry/Literature-Review-Estimating-Networks-of-Sustainable-Development-Goals</link>
      <description>&lt;h2 data-ke-size=&quot;size26&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;&quot;Estimating Networks of Sustainable Development Goals&quot; by Luis Ospina-Forero&lt;/span&gt;&lt;/h2&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;&lt;span style=&quot;color: #000000;&quot;&gt;The paper &quot;Estimating Networks of Sustainable Development Goals&quot; by Luis Ospina-Forero focuses on the interlinkages and interactions between the 17 Sustainable Development Goals (SDGs) outlined by the United Nations. The paper presents a &lt;span style=&quot;background-color: #f6e199;&quot;&gt;&lt;i&gt;method for estimating networks of SDGs&lt;/i&gt;&lt;/span&gt;, which can be used to identify the &lt;i&gt;most influential SDGs and how they interact.&lt;/i&gt;&lt;/span&gt;&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;The paper begins with an introduction to the SDGs and their importance in global development efforts. It then discusses the challenges in measuring the progress towards achieving the SDGs and the need to understand the interlinkages between them better.&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;The proposed method for estimating SDG networks involves using Bayesian networks, which can model the causal relationships between the SDGs. The paper describes the data used in the analysis, which includes data from the World Bank and the United Nations and the statistical methods used to estimate the network.&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;The analysis results show that the SDGs are highly interconnected, with some SDGs being more influential than others. The paper discusses the implications of the findings for policymakers and suggests that the approach can be used to prioritize interventions and measure progress towards achieving the SDGs.&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;Overall, the paper provides a novel method for estimating networks of SDGs and contributes to understanding the interlinkages and interactions between the SDGs. The findings have important implications for global development efforts and can inform policymakers' decision-making processes.&lt;/p&gt;
&lt;p style=&quot;text-align: justify;&quot; data-ke-size=&quot;size18&quot;&gt;&amp;nbsp;&lt;/p&gt;
&lt;ol style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Introduction&lt;/li&gt;
&lt;/ol&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Background: Provides an overview of the Sustainable Development Goals (SDGs) and their importance in global development efforts - &lt;span style=&quot;color: #006dd7;&quot;&gt;introduction to the SDGs and their importance in global development efforts&lt;/span&gt;&lt;/li&gt;
&lt;li&gt;Research question: States the main research question addressed in the study - &lt;span&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;span style=&quot;color: #006dd7;&quot;&gt;challenges in measuring the progress towards achieving the SDGs and the need to understand the interlinkages between them better.&lt;/span&gt;&lt;/span&gt;&lt;/li&gt;
&lt;li&gt;Objectives: Outline the objectives of the study&lt;/li&gt;
&lt;li&gt;Contribution: Summarizes the contributions of the study&lt;/li&gt;
&lt;li&gt;Outline: Provides a brief overview of the structure of the paper&lt;/li&gt;
&lt;/ul&gt;
&lt;ol style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Literature review&lt;/li&gt;
&lt;/ol&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Describes the concept of the SDGs and their importance in sustainable development&lt;/li&gt;
&lt;li&gt;Discusses previous research on the SDGs, including studies that have examined their relationships and interdependencies&lt;/li&gt;
&lt;li&gt;Identifies gaps in the existing knowledge and explains how the current study addresses those gaps&lt;/li&gt;
&lt;/ul&gt;
&lt;ol style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Methods&lt;/li&gt;
&lt;/ol&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Methodologies mentioned in the paper
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Bayesian networks are a probabilistic graphical model used to represent and reason uncertain relationships between variables. This paper uses Bayesian networks to model the causal relationships between the SDGs.&lt;/li&gt;
&lt;li&gt;Partial correlation analysis: a method for identifying correlations between variables while controlling for the effects of other variables.&lt;/li&gt;
&lt;li&gt;Ridge regression: a linear regression method used to analyze multiple regression data that suffer from multicollinearity.&lt;/li&gt;
&lt;li&gt;Network inference using time-delayed mutual information: a method for inferring causal relationships between variables from time-series data.&lt;/li&gt;
&lt;li&gt;Granger causality: a statistical method for inferring causal relationships between variables in time-series data.&lt;/li&gt;
&lt;li&gt;Structural equation modelling: a statistical technique for modelling complex relationships between variables using a combination of observed variables and latent variables.&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;Data sources: Describes the data sources used in the study, including the SDG indicators from the UN Sustainable Development Goals Knowledge Platform&lt;/li&gt;
&lt;li&gt;Analytical techniques: Describes the analytical techniques used to estimate the networks of SDGs, including correlation analysis and network analysis&lt;/li&gt;
&lt;li&gt;Statistical models: Describes the statistical models used to estimate the networks of SDGs, including partial correlation and graphical LASSO&lt;/li&gt;
&lt;/ul&gt;
&lt;ol style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Results&lt;/li&gt;
&lt;/ol&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Describes the networks of SDGs estimated using the analytical techniques and statistical models described in the methods section&lt;/li&gt;
&lt;li&gt;Presents data visualizations of the networks, including network graphs and heatmaps&lt;/li&gt;
&lt;li&gt;Discusses the findings of the study, including the most important SDGs and their relationships to other SDGs&lt;/li&gt;
&lt;/ul&gt;
&lt;ol style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Discussion&lt;/li&gt;
&lt;/ol&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Interprets the results and discusses their implications, including the policy implications of the findings&lt;/li&gt;
&lt;li&gt;Highlights the strengths and limitations of the study&lt;/li&gt;
&lt;li&gt;Suggests avenues for future research&lt;/li&gt;
&lt;/ul&gt;
&lt;ol style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Conclusion&lt;/li&gt;
&lt;/ol&gt;
&lt;ul style=&quot;list-style-type: disc;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Summarizes the main findings of the study&lt;/li&gt;
&lt;li&gt;Restates the objectives of the study&lt;/li&gt;
&lt;li&gt;Highlights the contributions of the study&lt;/li&gt;
&lt;/ul&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;</description>
      <category>Data science</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/87</guid>
      <comments>https://livelovelet.tistory.com/entry/Literature-Review-Estimating-Networks-of-Sustainable-Development-Goals#entry87comment</comments>
      <pubDate>Tue, 28 Mar 2023 09:18:28 +0900</pubDate>
    </item>
    <item>
      <title>[Literature Summary] Causal Discovery and Inference: concepts and recent methodological advances</title>
      <link>https://livelovelet.tistory.com/entry/Literature-Summary-Causal-discovery-and-inference-concepts-and-recent-methodological-advances</link>
      <description>&lt;h1 style=&quot;color: #000000; text-align: start;&quot;&gt;Ref1. Causal Discovery and Inference: concepts and&amp;nbsp;recent methodological advances&lt;/h1&gt;
&lt;h2 style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size26&quot;&gt;Notes&lt;/h2&gt;
&lt;h3 style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size23&quot;&gt;[What to expect from this paper]&lt;/h3&gt;
&lt;ul style=&quot;list-style-type: disc; color: #000000; text-align: start;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;Provides an overview of&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;automated causal inference&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;and emerging approaches to causal discovery from i.i.d data and time series&lt;/li&gt;
&lt;li&gt;Reviews fundamental concepts such as manipulations, causal models, sample predictive modelling, causal predictive modelling, and structural equation models&lt;/li&gt;
&lt;li&gt;Discusses the&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;constraint-based approach to causal discovery,&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;which relies on conditional independence relationships in the data, and the assumptions underlying its validity&lt;/li&gt;
&lt;li&gt;Focuses on&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;causal discovery based on structural equation models&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;and discusses the identifiability of the causal structure implied by appropriately defined structural equation models&lt;/li&gt;
&lt;li&gt;Shows that&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;independence between the error term and causes,&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;together with appropriate structural constraints on the structural equation, makes it possible to identify the causal direction between two variables in the two-variable case&lt;/li&gt;
&lt;li&gt;Discusses&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;recent advances in causal discovery from time series&lt;/b&gt;, mentioning traditionally challenging problems such as causal discovery from subsampled data and in the presence of confounding time series&lt;/li&gt;
&lt;li&gt;Lists several open questions in the field of causal discovery and inference&lt;/li&gt;
&lt;li&gt;Provides a valuable resource for researchers interested in automated causal inference and emerging approaches to causal discovery from i.i.d data and time series.&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size26&quot;&gt;Key takeaways&lt;/h2&gt;
&lt;ul style=&quot;list-style-type: disc; color: #000000; text-align: start;&quot; data-ke-list-type=&quot;disc&quot;&gt;
&lt;li&gt;&lt;b&gt;Keywords&lt;/b&gt;: Causal inference, Causal discovery, Structural equation model, Conditional independence, Statistical independence, Identifiability&lt;/li&gt;
&lt;/ul&gt;
&lt;ol style=&quot;list-style-type: decimal; color: #000000; text-align: start;&quot; data-ke-list-type=&quot;decimal&quot;&gt;
&lt;li&gt;Manipulation is changing one or more variables in a system and observing the effect on other variables to establish causality.&lt;/li&gt;
&lt;li&gt;Causal models: mathematical models that describe the causal relationships between variables in a system.&lt;/li&gt;
&lt;li&gt;Sample predictive modelling: a method of causal inference that uses data from observational studies to predict the outcome of interventions on new data.&lt;/li&gt;
&lt;li&gt;Causal predictive modelling: a causal inference method that uses data from observational studies to estimate the causal effects of interventions.&lt;/li&gt;
&lt;li&gt;Structural equation models: a statistical model that describes the relationships between variables in terms of structural equations.&lt;/li&gt;
&lt;li&gt;Constraint-based approach: a method of causal discovery that relies on identifying conditional independence relationships in the data.&lt;/li&gt;
&lt;li&gt;Causal discovery based on structural equations: a method of causal discovery that involves fitting structural equation models to the data and using constraints to identify causal relationships.&lt;/li&gt;
&lt;li&gt;Linear causal models with non-Gaussian noise: a class of models that assumes linear causal relationships between variables with non-Gaussian noise.&lt;/li&gt;
&lt;li&gt;Causal discovery from time series: a causal discovery method involving analyzing time series data to identify causal relationships.&lt;/li&gt;
&lt;li&gt;Causal discovery from subsampled data: a causal discovery method that involves inferring causal relationships from subsets of time series data.&lt;/li&gt;
&lt;li&gt;Confounding time series: a problem in causal discovery from time series data where the observed data is influenced by unobserved variables that affect both the cause and effect.&lt;/li&gt;
&lt;li&gt;Counterfactual analysis: a method of causal inference that involves comparing the outcomes of a treatment group to a control group to estimate the causal effect of the treatment.&lt;/li&gt;
&lt;li&gt;Bayesian networks: a probabilistic graphical model representing the causal relationships between variables as a directed acyclic graph.&lt;/li&gt;
&lt;/ol&gt;
&lt;h2 style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size26&quot;&gt;Summary&lt;/h2&gt;
&lt;p style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;Causal discovery and inference are related but distinct concepts in statistics and machine learning.&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;Causal discovery involves&lt;/b&gt;&amp;nbsp;identifying the causal relationships between variables in a system, while causal inference involves using these relationships to make predictions or draw conclusions about the system. Recent methodological advances have improved our ability to perform causal discovery and inference in complex systems, using techniques such as&lt;span&gt;&amp;nbsp;&lt;/span&gt;&lt;b&gt;Bayesian networks, structural equation modelling, and counterfactual analysis.&lt;/b&gt;&lt;/p&gt;
&lt;p style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Bayesian networks&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;are a popular method for performing causal discovery, allowing researchers to model the relationships between variables in a probabilistic framework. Bayesian networks can be used to infer causal relationships from observational data, and can also be used to make predictions about the behaviour of the system under different conditions. Recent advances in Bayesian network modelling have included the development of efficient algorithms for learning the network's structure from data and integrating causal discovery with other statistical techniques such as clustering and regression.&lt;/p&gt;
&lt;p style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Structural equation modelling (SEM)&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;is another method for performing causal inference, which involves modelling the relationships between variables in terms of their underlying causal mechanisms. SEM allows researchers to test hypotheses about the causal relationships between variables, and to estimate the strength and direction of these relationships. Recent advances in SEM have included the development of new methods for dealing with missing data and the integration of SEM with machine learning techniques such as deep neural networks.&lt;/p&gt;
&lt;p style=&quot;color: #000000; text-align: start;&quot; data-ke-size=&quot;size16&quot;&gt;&lt;b&gt;Counterfactual analysis&lt;/b&gt;&lt;span&gt;&amp;nbsp;&lt;/span&gt;is a third approach to causal inference, which involves modelling the system's behaviour under hypothetical scenarios. The counterfactual analysis allows researchers to estimate the causal effects of interventions, by comparing the system's behaviour with and without the intervention. Recent advances in the counterfactual analysis have included the development of new methods for estimating causal effects from observational data and integrating counterfactual analysis with machine learning techniques such as causal forests and propensity score matching.&lt;/p&gt;</description>
      <category>Data science</category>
      <category>causal inference</category>
      <category>인과추론</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/86</guid>
      <comments>https://livelovelet.tistory.com/entry/Literature-Summary-Causal-discovery-and-inference-concepts-and-recent-methodological-advances#entry86comment</comments>
      <pubDate>Tue, 28 Mar 2023 09:10:19 +0900</pubDate>
    </item>
    <item>
      <title>[Korean Teaching Material] I can speak Korean, I can do!</title>
      <link>https://livelovelet.tistory.com/entry/Korean-Teaching-Material-I-can-speak-Korean-I-can-do</link>
      <description>&lt;p style=&quot;text-align: center;&quot; data-ke-size=&quot;size16&quot;&gt;This is the lecture material I made for the teaching demonstration: &lt;span style=&quot;background-color: #f3c000;&quot;&gt;&lt;b&gt;&quot;I can do ~ / I cannot do~.&quot;&amp;nbsp;&lt;/b&gt;&lt;/span&gt;&lt;/p&gt;
&lt;p style=&quot;text-align: center;&quot; data-ke-size=&quot;size16&quot;&gt;&lt;i&gt;&lt;b&gt;-을 수 있어요/없어요!&lt;/b&gt;&lt;/i&gt;&lt;/p&gt;
&lt;p&gt;&lt;figure class=&quot;imageblock alignCenter&quot; data-ke-mobileStyle=&quot;widthOrigin&quot; data-origin-width=&quot;720&quot; data-origin-height=&quot;540&quot;&gt;&lt;span data-url=&quot;https://blog.kakaocdn.net/dn/zH8AS/btr53F3E5wF/Voz3GDAKqDL2hKx2kmG701/img.png&quot; data-phocus=&quot;https://blog.kakaocdn.net/dn/zH8AS/btr53F3E5wF/Voz3GDAKqDL2hKx2kmG701/img.png&quot;&gt;&lt;img src=&quot;https://blog.kakaocdn.net/dn/zH8AS/btr53F3E5wF/Voz3GDAKqDL2hKx2kmG701/img.png&quot; srcset=&quot;https://img1.daumcdn.net/thumb/R1280x0/?scode=mtistory2&amp;fname=https%3A%2F%2Fblog.kakaocdn.net%2Fdn%2FzH8AS%2Fbtr53F3E5wF%2FVoz3GDAKqDL2hKx2kmG701%2Fimg.png&quot; onerror=&quot;this.onerror=null; this.src='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png'; this.srcset='//t1.daumcdn.net/tistory_admin/static/images/no-image-v1.png';&quot; loading=&quot;lazy&quot; width=&quot;720&quot; height=&quot;540&quot; data-origin-width=&quot;720&quot; data-origin-height=&quot;540&quot;/&gt;&lt;/span&gt;&lt;/figure&gt;
&lt;figure class=&quot;fileblock&quot; data-ke-align=&quot;alignCenter&quot;&gt;&lt;a href=&quot;https://blog.kakaocdn.net/dn/bTj4Wq/btr6eWX59Wa/Pmk2wIdtzY11yB93FACbIK/%ED%95%9C%EA%B5%AD%EB%A7%90%EC%9D%84%20%ED%95%A0%20%EC%88%98%20%EC%9E%88%EC%96%B4%EC%9A%94%202.pdf?attach=1&amp;amp;knm=tfile.pdf&quot; class=&quot;&quot;&gt;
    &lt;div class=&quot;image&quot;&gt;&lt;/div&gt;
    &lt;div class=&quot;desc&quot;&gt;&lt;div class=&quot;filename&quot;&gt;&lt;span class=&quot;name&quot;&gt;한국말을 할 수 있어요 2.pdf&lt;/span&gt;&lt;/div&gt;
&lt;div class=&quot;size&quot;&gt;1.70MB&lt;/div&gt;
&lt;/div&gt;
  &lt;/a&gt;&lt;/figure&gt;
&lt;/p&gt;
&lt;p data-ke-size=&quot;size16&quot;&gt;&amp;nbsp;&lt;/p&gt;</description>
      <category>한국말을할수있어요</category>
      <category>한국어수업</category>
      <author>LiveLoveFlow</author>
      <guid isPermaLink="true">https://livelovelet.tistory.com/84</guid>
      <comments>https://livelovelet.tistory.com/entry/Korean-Teaching-Material-I-can-speak-Korean-I-can-do#entry84comment</comments>
      <pubDate>Mon, 27 Mar 2023 03:33:01 +0900</pubDate>
    </item>
  </channel>
</rss>