How to safely limit the no process running without using multiprocessing.pool












0















I have a list containing process objects, and i want only 100 of them to be active and running at any time, and after they are done they should exit from memory, and the next 100 process should start, and so on.., I've writen a demo code in python3, and i want to know if there are any problems or limitation with it.



process = [List of process]
while len(process) != 0:
i=0
for i in range (100):
process[0].start()
copy = process[0]
del process[0]
print(process[0])

copy.join()
print("joining")









share|improve this question























  • Have you looked into process pools? Or process pool executors.

    – jdehesa
    Nov 19 '18 at 16:36













  • have you looked at the question?

    – user1779646
    Nov 19 '18 at 16:44
















0















I have a list containing process objects, and i want only 100 of them to be active and running at any time, and after they are done they should exit from memory, and the next 100 process should start, and so on.., I've writen a demo code in python3, and i want to know if there are any problems or limitation with it.



process = [List of process]
while len(process) != 0:
i=0
for i in range (100):
process[0].start()
copy = process[0]
del process[0]
print(process[0])

copy.join()
print("joining")









share|improve this question























  • Have you looked into process pools? Or process pool executors.

    – jdehesa
    Nov 19 '18 at 16:36













  • have you looked at the question?

    – user1779646
    Nov 19 '18 at 16:44














0












0








0








I have a list containing process objects, and i want only 100 of them to be active and running at any time, and after they are done they should exit from memory, and the next 100 process should start, and so on.., I've writen a demo code in python3, and i want to know if there are any problems or limitation with it.



process = [List of process]
while len(process) != 0:
i=0
for i in range (100):
process[0].start()
copy = process[0]
del process[0]
print(process[0])

copy.join()
print("joining")









share|improve this question














I have a list containing process objects, and i want only 100 of them to be active and running at any time, and after they are done they should exit from memory, and the next 100 process should start, and so on.., I've writen a demo code in python3, and i want to know if there are any problems or limitation with it.



process = [List of process]
while len(process) != 0:
i=0
for i in range (100):
process[0].start()
copy = process[0]
del process[0]
print(process[0])

copy.join()
print("joining")






python multiprocessing






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Nov 19 '18 at 16:31









user1779646user1779646

3861414




3861414













  • Have you looked into process pools? Or process pool executors.

    – jdehesa
    Nov 19 '18 at 16:36













  • have you looked at the question?

    – user1779646
    Nov 19 '18 at 16:44



















  • Have you looked into process pools? Or process pool executors.

    – jdehesa
    Nov 19 '18 at 16:36













  • have you looked at the question?

    – user1779646
    Nov 19 '18 at 16:44

















Have you looked into process pools? Or process pool executors.

– jdehesa
Nov 19 '18 at 16:36







Have you looked into process pools? Or process pool executors.

– jdehesa
Nov 19 '18 at 16:36















have you looked at the question?

– user1779646
Nov 19 '18 at 16:44





have you looked at the question?

– user1779646
Nov 19 '18 at 16:44












1 Answer
1






active

oldest

votes


















0














It might be most sensible to use multiprocessing.Pool which produces a pool of worker processes based on the max number of cores available on your system, and then basically feeds tasks in as the cores become available.



Hardcoding number of process' might actually slow your execution and more importantly, there is a threat of process' entering deadlock state.



In python, multiple process' are spawned according to POSIX standard(using fork). During this fork, everything from the parent except threads are copied into the child process. Be careful of shared memory space and inheriting config from parent to child. More on this if you are interested - How can I inherit parent logger when using Python's multiprocessing? Especially for paramiko



import multiprocessing

def f(name):
print 'hello', name

if __name__ == '__main__':
pool = multiprocessing.Pool() #use all available cores, otherwise specify the number you want as an argument
for i in xrange(0, 512):
pool.apply_async(f, args=(i,)) #process function f asynchronously.
pool.close() #safely close the pool and all associated process.
pool.join() #execute process' in pool.


Hardcoding something like p = multiprocessing.pool(999999) is likely to suffer a catastrophic death on any machine by grinding disk and grokking RAM.



Number of process's should always be determined by Python and it depends on:




  1. Hardware capability to run process' simultaneously.


  2. OS deciding to give resources to process'



If you still want to hardcode number of process, using semaphore restricted number of process is safe:



pool = multiprocessing.Semaphore(4) # no of cpus of your system.


Hope this helps.






share|improve this answer























    Your Answer






    StackExchange.ifUsing("editor", function () {
    StackExchange.using("externalEditor", function () {
    StackExchange.using("snippets", function () {
    StackExchange.snippets.init();
    });
    });
    }, "code-snippets");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "1"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53378981%2fhow-to-safely-limit-the-no-process-running-without-using-multiprocessing-pool%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    0














    It might be most sensible to use multiprocessing.Pool which produces a pool of worker processes based on the max number of cores available on your system, and then basically feeds tasks in as the cores become available.



    Hardcoding number of process' might actually slow your execution and more importantly, there is a threat of process' entering deadlock state.



    In python, multiple process' are spawned according to POSIX standard(using fork). During this fork, everything from the parent except threads are copied into the child process. Be careful of shared memory space and inheriting config from parent to child. More on this if you are interested - How can I inherit parent logger when using Python's multiprocessing? Especially for paramiko



    import multiprocessing

    def f(name):
    print 'hello', name

    if __name__ == '__main__':
    pool = multiprocessing.Pool() #use all available cores, otherwise specify the number you want as an argument
    for i in xrange(0, 512):
    pool.apply_async(f, args=(i,)) #process function f asynchronously.
    pool.close() #safely close the pool and all associated process.
    pool.join() #execute process' in pool.


    Hardcoding something like p = multiprocessing.pool(999999) is likely to suffer a catastrophic death on any machine by grinding disk and grokking RAM.



    Number of process's should always be determined by Python and it depends on:




    1. Hardware capability to run process' simultaneously.


    2. OS deciding to give resources to process'



    If you still want to hardcode number of process, using semaphore restricted number of process is safe:



    pool = multiprocessing.Semaphore(4) # no of cpus of your system.


    Hope this helps.






    share|improve this answer




























      0














      It might be most sensible to use multiprocessing.Pool which produces a pool of worker processes based on the max number of cores available on your system, and then basically feeds tasks in as the cores become available.



      Hardcoding number of process' might actually slow your execution and more importantly, there is a threat of process' entering deadlock state.



      In python, multiple process' are spawned according to POSIX standard(using fork). During this fork, everything from the parent except threads are copied into the child process. Be careful of shared memory space and inheriting config from parent to child. More on this if you are interested - How can I inherit parent logger when using Python's multiprocessing? Especially for paramiko



      import multiprocessing

      def f(name):
      print 'hello', name

      if __name__ == '__main__':
      pool = multiprocessing.Pool() #use all available cores, otherwise specify the number you want as an argument
      for i in xrange(0, 512):
      pool.apply_async(f, args=(i,)) #process function f asynchronously.
      pool.close() #safely close the pool and all associated process.
      pool.join() #execute process' in pool.


      Hardcoding something like p = multiprocessing.pool(999999) is likely to suffer a catastrophic death on any machine by grinding disk and grokking RAM.



      Number of process's should always be determined by Python and it depends on:




      1. Hardware capability to run process' simultaneously.


      2. OS deciding to give resources to process'



      If you still want to hardcode number of process, using semaphore restricted number of process is safe:



      pool = multiprocessing.Semaphore(4) # no of cpus of your system.


      Hope this helps.






      share|improve this answer


























        0












        0








        0







        It might be most sensible to use multiprocessing.Pool which produces a pool of worker processes based on the max number of cores available on your system, and then basically feeds tasks in as the cores become available.



        Hardcoding number of process' might actually slow your execution and more importantly, there is a threat of process' entering deadlock state.



        In python, multiple process' are spawned according to POSIX standard(using fork). During this fork, everything from the parent except threads are copied into the child process. Be careful of shared memory space and inheriting config from parent to child. More on this if you are interested - How can I inherit parent logger when using Python's multiprocessing? Especially for paramiko



        import multiprocessing

        def f(name):
        print 'hello', name

        if __name__ == '__main__':
        pool = multiprocessing.Pool() #use all available cores, otherwise specify the number you want as an argument
        for i in xrange(0, 512):
        pool.apply_async(f, args=(i,)) #process function f asynchronously.
        pool.close() #safely close the pool and all associated process.
        pool.join() #execute process' in pool.


        Hardcoding something like p = multiprocessing.pool(999999) is likely to suffer a catastrophic death on any machine by grinding disk and grokking RAM.



        Number of process's should always be determined by Python and it depends on:




        1. Hardware capability to run process' simultaneously.


        2. OS deciding to give resources to process'



        If you still want to hardcode number of process, using semaphore restricted number of process is safe:



        pool = multiprocessing.Semaphore(4) # no of cpus of your system.


        Hope this helps.






        share|improve this answer













        It might be most sensible to use multiprocessing.Pool which produces a pool of worker processes based on the max number of cores available on your system, and then basically feeds tasks in as the cores become available.



        Hardcoding number of process' might actually slow your execution and more importantly, there is a threat of process' entering deadlock state.



        In python, multiple process' are spawned according to POSIX standard(using fork). During this fork, everything from the parent except threads are copied into the child process. Be careful of shared memory space and inheriting config from parent to child. More on this if you are interested - How can I inherit parent logger when using Python's multiprocessing? Especially for paramiko



        import multiprocessing

        def f(name):
        print 'hello', name

        if __name__ == '__main__':
        pool = multiprocessing.Pool() #use all available cores, otherwise specify the number you want as an argument
        for i in xrange(0, 512):
        pool.apply_async(f, args=(i,)) #process function f asynchronously.
        pool.close() #safely close the pool and all associated process.
        pool.join() #execute process' in pool.


        Hardcoding something like p = multiprocessing.pool(999999) is likely to suffer a catastrophic death on any machine by grinding disk and grokking RAM.



        Number of process's should always be determined by Python and it depends on:




        1. Hardware capability to run process' simultaneously.


        2. OS deciding to give resources to process'



        If you still want to hardcode number of process, using semaphore restricted number of process is safe:



        pool = multiprocessing.Semaphore(4) # no of cpus of your system.


        Hope this helps.







        share|improve this answer












        share|improve this answer



        share|improve this answer










        answered Nov 20 '18 at 4:02









        Pruthvi KumarPruthvi Kumar

        59729




        59729






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Stack Overflow!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53378981%2fhow-to-safely-limit-the-no-process-running-without-using-multiprocessing-pool%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            鏡平學校

            ꓛꓣだゔៀៅຸ໢ທຮ໕໒ ,ໂ'໥໓າ໼ឨឲ៵៭ៈゎゔit''䖳𥁄卿' ☨₤₨こゎもょの;ꜹꟚꞖꞵꟅꞛေၦေɯ,ɨɡ𛃵𛁹ޝ޳ޠ޾,ޤޒޯ޾𫝒𫠁သ𛅤チョ'サノބޘދ𛁐ᶿᶇᶀᶋᶠ㨑㽹⻮ꧬ꧹؍۩وَؠ㇕㇃㇪ ㇦㇋㇋ṜẰᵡᴠ 軌ᵕ搜۳ٰޗޮ޷ސޯ𫖾𫅀ल, ꙭ꙰ꚅꙁꚊꞻꝔ꟠Ꝭㄤﺟޱސꧨꧼ꧴ꧯꧽ꧲ꧯ'⽹⽭⾁⿞⼳⽋២៩ញណើꩯꩤ꩸ꩮᶻᶺᶧᶂ𫳲𫪭𬸄𫵰𬖩𬫣𬊉ၲ𛅬㕦䬺𫝌𫝼,,𫟖𫞽ហៅ஫㆔ాఆఅꙒꚞꙍ,Ꙟ꙱エ ,ポテ,フࢰࢯ𫟠𫞶 𫝤𫟠ﺕﹱﻜﻣ𪵕𪭸𪻆𪾩𫔷ġ,ŧآꞪ꟥,ꞔꝻ♚☹⛵𛀌ꬷꭞȄƁƪƬșƦǙǗdžƝǯǧⱦⱰꓕꓢႋ神 ဴ၀க௭எ௫ឫោ ' េㇷㇴㇼ神ㇸㇲㇽㇴㇼㇻㇸ'ㇸㇿㇸㇹㇰㆣꓚꓤ₡₧ ㄨㄟ㄂ㄖㄎ໗ツڒذ₶।ऩछएोञयूटक़कयँृी,冬'𛅢𛅥ㇱㇵㇶ𥄥𦒽𠣧𠊓𧢖𥞘𩔋цѰㄠſtʯʭɿʆʗʍʩɷɛ,əʏダヵㄐㄘR{gỚṖḺờṠṫảḙḭᴮᵏᴘᵀᵷᵕᴜᴏᵾq﮲ﲿﴽﭙ軌ﰬﶚﶧ﫲Ҝжюїкӈㇴffצּ﬘﭅﬈軌'ffistfflſtffतभफɳɰʊɲʎ𛁱𛁖𛁮𛀉 𛂯𛀞నఋŀŲ 𫟲𫠖𫞺ຆຆ ໹້໕໗ๆทԊꧢꧠ꧰ꓱ⿝⼑ŎḬẃẖỐẅ ,ờỰỈỗﮊDžȩꭏꭎꬻ꭮ꬿꭖꭥꭅ㇭神 ⾈ꓵꓑ⺄㄄ㄪㄙㄅㄇstA۵䞽ॶ𫞑𫝄㇉㇇゜軌𩜛𩳠Jﻺ‚Üမ႕ႌႊၐၸဓၞၞၡ៸wyvtᶎᶪᶹစဎ꣡꣰꣢꣤ٗ؋لㇳㇾㇻㇱ㆐㆔,,㆟Ⱶヤマފ޼ޝަݿݞݠݷݐ',ݘ,ݪݙݵ𬝉𬜁𫝨𫞘くせぉて¼óû×ó£…𛅑הㄙくԗԀ5606神45,神796'𪤻𫞧ꓐ㄁ㄘɥɺꓵꓲ3''7034׉ⱦⱠˆ“𫝋ȍ,ꩲ軌꩷ꩶꩧꩫఞ۔فڱێظペサ神ナᴦᵑ47 9238їﻂ䐊䔉㠸﬎ffiﬣ,לּᴷᴦᵛᵽ,ᴨᵤ ᵸᵥᴗᵈꚏꚉꚟ⻆rtǟƴ𬎎

            Why https connections are so slow when debugging (stepping over) in Java?