Combine two neural networks with different inputs through element-wise summation of certain layers











up vote
0
down vote

favorite












I am looking at combining two Convolutional Neural Networks into one through element-wise summation of activation functions. Both these networks have different inputs, but are similar in their architecture.



I have seen from certain papers and github pages that this has been, successfully, implemented in Python. However, I was wondering if this would also be possible to implement in MATLAB?



One example of what I want to reproduce is the FuseNet architecture by Hazirbas et al. https://github.com/zanilzanzan/FuseNet_PyTorch:



FuseNet NN architecture by Hazirbas et al.



Is it possible to reproduce this in MATLAB, and if so, how do I start?










share|improve this question




























    up vote
    0
    down vote

    favorite












    I am looking at combining two Convolutional Neural Networks into one through element-wise summation of activation functions. Both these networks have different inputs, but are similar in their architecture.



    I have seen from certain papers and github pages that this has been, successfully, implemented in Python. However, I was wondering if this would also be possible to implement in MATLAB?



    One example of what I want to reproduce is the FuseNet architecture by Hazirbas et al. https://github.com/zanilzanzan/FuseNet_PyTorch:



    FuseNet NN architecture by Hazirbas et al.



    Is it possible to reproduce this in MATLAB, and if so, how do I start?










    share|improve this question


























      up vote
      0
      down vote

      favorite









      up vote
      0
      down vote

      favorite











      I am looking at combining two Convolutional Neural Networks into one through element-wise summation of activation functions. Both these networks have different inputs, but are similar in their architecture.



      I have seen from certain papers and github pages that this has been, successfully, implemented in Python. However, I was wondering if this would also be possible to implement in MATLAB?



      One example of what I want to reproduce is the FuseNet architecture by Hazirbas et al. https://github.com/zanilzanzan/FuseNet_PyTorch:



      FuseNet NN architecture by Hazirbas et al.



      Is it possible to reproduce this in MATLAB, and if so, how do I start?










      share|improve this question















      I am looking at combining two Convolutional Neural Networks into one through element-wise summation of activation functions. Both these networks have different inputs, but are similar in their architecture.



      I have seen from certain papers and github pages that this has been, successfully, implemented in Python. However, I was wondering if this would also be possible to implement in MATLAB?



      One example of what I want to reproduce is the FuseNet architecture by Hazirbas et al. https://github.com/zanilzanzan/FuseNet_PyTorch:



      FuseNet NN architecture by Hazirbas et al.



      Is it possible to reproduce this in MATLAB, and if so, how do I start?







      matlab neural-network computer-vision classification semantic-segmentation






      share|improve this question















      share|improve this question













      share|improve this question




      share|improve this question








      edited Nov 8 at 14:16









      Dev-iL

      16.1k63974




      16.1k63974










      asked Nov 8 at 11:13









      Isa El Doori

      91




      91
























          1 Answer
          1






          active

          oldest

          votes

















          up vote
          0
          down vote













          You might be able to do this using a DAG network1,2 in MATLAB. Here's an illustration:



          enter image description here



          The element-wise summation, specifically, can be performed using an additionLayer.






          share|improve this answer





















          • Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
            – Isa El Doori
            Nov 8 at 17:43










          • @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
            – Dev-iL
            Nov 8 at 18:55











          Your Answer






          StackExchange.ifUsing("editor", function () {
          StackExchange.using("externalEditor", function () {
          StackExchange.using("snippets", function () {
          StackExchange.snippets.init();
          });
          });
          }, "code-snippets");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "1"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          convertImagesToLinks: true,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: 10,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














           

          draft saved


          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53206600%2fcombine-two-neural-networks-with-different-inputs-through-element-wise-summation%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes








          up vote
          0
          down vote













          You might be able to do this using a DAG network1,2 in MATLAB. Here's an illustration:



          enter image description here



          The element-wise summation, specifically, can be performed using an additionLayer.






          share|improve this answer





















          • Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
            – Isa El Doori
            Nov 8 at 17:43










          • @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
            – Dev-iL
            Nov 8 at 18:55















          up vote
          0
          down vote













          You might be able to do this using a DAG network1,2 in MATLAB. Here's an illustration:



          enter image description here



          The element-wise summation, specifically, can be performed using an additionLayer.






          share|improve this answer





















          • Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
            – Isa El Doori
            Nov 8 at 17:43










          • @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
            – Dev-iL
            Nov 8 at 18:55













          up vote
          0
          down vote










          up vote
          0
          down vote









          You might be able to do this using a DAG network1,2 in MATLAB. Here's an illustration:



          enter image description here



          The element-wise summation, specifically, can be performed using an additionLayer.






          share|improve this answer












          You might be able to do this using a DAG network1,2 in MATLAB. Here's an illustration:



          enter image description here



          The element-wise summation, specifically, can be performed using an additionLayer.







          share|improve this answer












          share|improve this answer



          share|improve this answer










          answered Nov 8 at 13:56









          Dev-iL

          16.1k63974




          16.1k63974












          • Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
            – Isa El Doori
            Nov 8 at 17:43










          • @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
            – Dev-iL
            Nov 8 at 18:55


















          • Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
            – Isa El Doori
            Nov 8 at 17:43










          • @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
            – Dev-iL
            Nov 8 at 18:55
















          Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
          – Isa El Doori
          Nov 8 at 17:43




          Hi Dev-iL, I was able to reproduce this network. However, since the DAGNetwork only allows for one input, my question then becomes: would it be possible to have a 4-channel input, 3 channels go to one side of the network, and the remaining channel goes to the other side of the network?
          – Isa El Doori
          Nov 8 at 17:43












          @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
          – Dev-iL
          Nov 8 at 18:55




          @IsaElDoori I don't know of an easy way to achieve that. What you could do, is define a custom layer that does this "unzipping". Basically the reverse of a depth concatenation layer. Either that, or a custom layer that all it does is forward certain dimensions (or slices) of the input forward.
          – Dev-iL
          Nov 8 at 18:55


















           

          draft saved


          draft discarded



















































           


          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53206600%2fcombine-two-neural-networks-with-different-inputs-through-element-wise-summation%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          鏡平學校

          ꓛꓣだゔៀៅຸ໢ທຮ໕໒ ,ໂ'໥໓າ໼ឨឲ៵៭ៈゎゔit''䖳𥁄卿' ☨₤₨こゎもょの;ꜹꟚꞖꞵꟅꞛေၦေɯ,ɨɡ𛃵𛁹ޝ޳ޠ޾,ޤޒޯ޾𫝒𫠁သ𛅤チョ'サノބޘދ𛁐ᶿᶇᶀᶋᶠ㨑㽹⻮ꧬ꧹؍۩وَؠ㇕㇃㇪ ㇦㇋㇋ṜẰᵡᴠ 軌ᵕ搜۳ٰޗޮ޷ސޯ𫖾𫅀ल, ꙭ꙰ꚅꙁꚊꞻꝔ꟠Ꝭㄤﺟޱސꧨꧼ꧴ꧯꧽ꧲ꧯ'⽹⽭⾁⿞⼳⽋២៩ញណើꩯꩤ꩸ꩮᶻᶺᶧᶂ𫳲𫪭𬸄𫵰𬖩𬫣𬊉ၲ𛅬㕦䬺𫝌𫝼,,𫟖𫞽ហៅ஫㆔ాఆఅꙒꚞꙍ,Ꙟ꙱エ ,ポテ,フࢰࢯ𫟠𫞶 𫝤𫟠ﺕﹱﻜﻣ𪵕𪭸𪻆𪾩𫔷ġ,ŧآꞪ꟥,ꞔꝻ♚☹⛵𛀌ꬷꭞȄƁƪƬșƦǙǗdžƝǯǧⱦⱰꓕꓢႋ神 ဴ၀க௭எ௫ឫោ ' េㇷㇴㇼ神ㇸㇲㇽㇴㇼㇻㇸ'ㇸㇿㇸㇹㇰㆣꓚꓤ₡₧ ㄨㄟ㄂ㄖㄎ໗ツڒذ₶।ऩछएोञयूटक़कयँृी,冬'𛅢𛅥ㇱㇵㇶ𥄥𦒽𠣧𠊓𧢖𥞘𩔋цѰㄠſtʯʭɿʆʗʍʩɷɛ,əʏダヵㄐㄘR{gỚṖḺờṠṫảḙḭᴮᵏᴘᵀᵷᵕᴜᴏᵾq﮲ﲿﴽﭙ軌ﰬﶚﶧ﫲Ҝжюїкӈㇴffצּ﬘﭅﬈軌'ffistfflſtffतभफɳɰʊɲʎ𛁱𛁖𛁮𛀉 𛂯𛀞నఋŀŲ 𫟲𫠖𫞺ຆຆ ໹້໕໗ๆทԊꧢꧠ꧰ꓱ⿝⼑ŎḬẃẖỐẅ ,ờỰỈỗﮊDžȩꭏꭎꬻ꭮ꬿꭖꭥꭅ㇭神 ⾈ꓵꓑ⺄㄄ㄪㄙㄅㄇstA۵䞽ॶ𫞑𫝄㇉㇇゜軌𩜛𩳠Jﻺ‚Üမ႕ႌႊၐၸဓၞၞၡ៸wyvtᶎᶪᶹစဎ꣡꣰꣢꣤ٗ؋لㇳㇾㇻㇱ㆐㆔,,㆟Ⱶヤマފ޼ޝަݿݞݠݷݐ',ݘ,ݪݙݵ𬝉𬜁𫝨𫞘くせぉて¼óû×ó£…𛅑הㄙくԗԀ5606神45,神796'𪤻𫞧ꓐ㄁ㄘɥɺꓵꓲ3''7034׉ⱦⱠˆ“𫝋ȍ,ꩲ軌꩷ꩶꩧꩫఞ۔فڱێظペサ神ナᴦᵑ47 9238їﻂ䐊䔉㠸﬎ffiﬣ,לּᴷᴦᵛᵽ,ᴨᵤ ᵸᵥᴗᵈꚏꚉꚟ⻆rtǟƴ𬎎

          Why https connections are so slow when debugging (stepping over) in Java?