{"id":13,"date":"2019-06-30T20:47:04","date_gmt":"2019-06-30T20:47:04","guid":{"rendered":"http:\/\/randomfields.org\/?page_id=13"},"modified":"2019-07-02T22:26:54","modified_gmt":"2019-07-02T22:26:54","slug":"deep-random-fields","status":"publish","type":"page","link":"https:\/\/randomfields.org\/?page_id=13","title":{"rendered":"Deep Boltzmann Trees"},"content":{"rendered":"\n\n\n<p><\/p>\n\n\n<div id=\"themify_builder_content-13\" data-postid=\"13\" class=\"themify_builder_content themify_builder_content-13 themify_builder\">\n    \t<!-- module_row -->\n\t<div  class=\"themify_builder_row module_row clearfix module_row_0 themify_builder_13_row module_row_13-0 tb_g07f215\">\n\t    \t    <div class=\"row_inner col_align_top\" >\n\t\t\t<div  class=\"module_column tb-column col-full first tb_13_column module_column_0 module_column_13-0-0 tb_7kvq215\">\n\t    \t    \t        <div class=\"tb-column-inner\">\n\t\t    <!-- module text -->\n<div  class=\"module module-text tb_zy6t726  repeat  \">\n            <div  class=\"tb_text_wrap\">\n    <p>Deep neural networks achieve state-of-the-art results in various classification and synthetic data generation tasks. However, only little is known about why depth improves a model. We investigate the structure of stochastic deep neural works, also known as Deep Boltzmann Machines, to shed some light on this issue. While the best known results postulate an exponential dependence between the number of visible units and the depth of the model, we show that the required depth is upper bounded by the longest path in the underlying junction tree, which is at most <em>linear<\/em> in the number of visible units. Moreover, we show that the conditional independence structure of any categorical Deep Boltzmann Machine contains a sub-tree that allows the consistent estimation of the full joint probability mass function of all visible units. We connect our results to L1-regularized maximum-likelihood estimation and Chow-Liu trees. Based on our theoretical findings, we present a new tractable version of Deep Boltzmann Machines, namely the <em>Deep Boltzmann Tree<\/em> (DBT). We provide a hyper-parameter-free algorithm for learning the DBT from data, and propose a new initialization method to enforce convergence to good solutions. Our findings provide some theoretical evidence for why a deep model might be beneficial. Experimental results on benchmark data show, that the DBT is a theoretical sound alternative to likelihood-free generative models.<\/p>    <\/div>\n<\/div>\n<!-- \/module text -->\n\t        <\/div>\n\t    \t<\/div>\n\t\t    <\/div>\n\t    <!-- \/row_inner -->\n\t<\/div>\n\t<!-- \/module_row -->\n\t<\/div>\n\n","protected":false},"excerpt":{"rendered":"<p>Deep neural networks achieve state-of-the-art results in various classification and synthetic data generation tasks. However, only little is known about why depth improves a model. We investigate the structure of stochastic deep neural works, also known as Deep Boltzmann Machines, to shed some light on this issue. While the best known results postulate an exponential [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"builder_content":"<p>Deep neural networks achieve state-of-the-art results in various classification and synthetic data generation tasks. However, only little is known about why depth improves a model. We investigate the structure of stochastic deep neural works, also known as Deep Boltzmann Machines, to shed some light on this issue. While the best known results postulate an exponential dependence between the number of visible units and the depth of the model, we show that the required depth is upper bounded by the longest path in the underlying junction tree, which is at most <em>linear<\/em> in the number of visible units. Moreover, we show that the conditional independence structure of any categorical Deep Boltzmann Machine contains a sub-tree that allows the consistent estimation of the full joint probability mass function of all visible units. We connect our results to L1-regularized maximum-likelihood estimation and Chow-Liu trees. Based on our theoretical findings, we present a new tractable version of Deep Boltzmann Machines, namely the <em>Deep Boltzmann Tree<\/em> (DBT). We provide a hyper-parameter-free algorithm for learning the DBT from data, and propose a new initialization method to enforce convergence to good solutions. Our findings provide some theoretical evidence for why a deep model might be beneficial. Experimental results on benchmark data show, that the DBT is a theoretical sound alternative to likelihood-free generative models.<\/p>","_links":{"self":[{"href":"https:\/\/randomfields.org\/index.php?rest_route=\/wp\/v2\/pages\/13"}],"collection":[{"href":"https:\/\/randomfields.org\/index.php?rest_route=\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/randomfields.org\/index.php?rest_route=\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/randomfields.org\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/randomfields.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=13"}],"version-history":[{"count":24,"href":"https:\/\/randomfields.org\/index.php?rest_route=\/wp\/v2\/pages\/13\/revisions"}],"predecessor-version":[{"id":112,"href":"https:\/\/randomfields.org\/index.php?rest_route=\/wp\/v2\/pages\/13\/revisions\/112"}],"wp:attachment":[{"href":"https:\/\/randomfields.org\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=13"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}