Python pandas read_csv for specfic records in columns












1















I am trying to import data from a large csv file 15GB+. I have to select few columns with specific values (there are over 50 columns) but as an example. I have used



df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', State'])


Is there a way where I can specify something like that:



df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', 'State'='abc'])


Can't find any option to do that










share|improve this question



























    1















    I am trying to import data from a large csv file 15GB+. I have to select few columns with specific values (there are over 50 columns) but as an example. I have used



    df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', State'])


    Is there a way where I can specify something like that:



    df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', 'State'='abc'])


    Can't find any option to do that










    share|improve this question

























      1












      1








      1








      I am trying to import data from a large csv file 15GB+. I have to select few columns with specific values (there are over 50 columns) but as an example. I have used



      df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', State'])


      Is there a way where I can specify something like that:



      df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', 'State'='abc'])


      Can't find any option to do that










      share|improve this question














      I am trying to import data from a large csv file 15GB+. I have to select few columns with specific values (there are over 50 columns) but as an example. I have used



      df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', State'])


      Is there a way where I can specify something like that:



      df=pd.read_csv('filename.csv', nrows=10000, usecols=['ID', 'State'='abc'])


      Can't find any option to do that







      python-3.x






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Nov 21 '18 at 1:37









      aus_fasaus_fas

      207




      207
























          2 Answers
          2






          active

          oldest

          votes


















          1














          There's no option to filter rows like that while reading csv files.
          What you can do is create an iterator then apply your filter to each chunk then concat the chunks. It would look something like:



          iterable = pd.read_csv('filename.csv', usecols=['ID', 'State'], iterator=True, chunksize=10000)
          df = pd.concat([chunk[chunk['State'] == 'abc'] for chunk in iterable])





          share|improve this answer
























          • Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

            – aus_fas
            Nov 21 '18 at 6:10





















          0














          Assuming that the resulting DataFrame for a selection where 'State' == 'abc' is small enough to be accommodated in RAM, you could extract those from the csv as follows. df is the resultant DataFrame.



          import pandas as pd 

          inPath = 'filename.csv'
          chunkSize = 10000 #size of chunks relies on your available memory

          tmpDf = pd.read_csv(inPath,chunksize=chunkSize,
          usecols=['ID', 'State'])
          for chunk in tmpDf:
          try:
          df
          except NameError:
          df = tmpDf[tmpDf['State'] == 'abc']
          else:
          df = pd.concat([df, tmpDf[tmpDf['State'] == 'abc']])





          share|improve this answer

























            Your Answer






            StackExchange.ifUsing("editor", function () {
            StackExchange.using("externalEditor", function () {
            StackExchange.using("snippets", function () {
            StackExchange.snippets.init();
            });
            });
            }, "code-snippets");

            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "1"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














            draft saved

            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53404127%2fpython-pandas-read-csv-for-specfic-records-in-columns%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown

























            2 Answers
            2






            active

            oldest

            votes








            2 Answers
            2






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            1














            There's no option to filter rows like that while reading csv files.
            What you can do is create an iterator then apply your filter to each chunk then concat the chunks. It would look something like:



            iterable = pd.read_csv('filename.csv', usecols=['ID', 'State'], iterator=True, chunksize=10000)
            df = pd.concat([chunk[chunk['State'] == 'abc'] for chunk in iterable])





            share|improve this answer
























            • Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

              – aus_fas
              Nov 21 '18 at 6:10


















            1














            There's no option to filter rows like that while reading csv files.
            What you can do is create an iterator then apply your filter to each chunk then concat the chunks. It would look something like:



            iterable = pd.read_csv('filename.csv', usecols=['ID', 'State'], iterator=True, chunksize=10000)
            df = pd.concat([chunk[chunk['State'] == 'abc'] for chunk in iterable])





            share|improve this answer
























            • Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

              – aus_fas
              Nov 21 '18 at 6:10
















            1












            1








            1







            There's no option to filter rows like that while reading csv files.
            What you can do is create an iterator then apply your filter to each chunk then concat the chunks. It would look something like:



            iterable = pd.read_csv('filename.csv', usecols=['ID', 'State'], iterator=True, chunksize=10000)
            df = pd.concat([chunk[chunk['State'] == 'abc'] for chunk in iterable])





            share|improve this answer













            There's no option to filter rows like that while reading csv files.
            What you can do is create an iterator then apply your filter to each chunk then concat the chunks. It would look something like:



            iterable = pd.read_csv('filename.csv', usecols=['ID', 'State'], iterator=True, chunksize=10000)
            df = pd.concat([chunk[chunk['State'] == 'abc'] for chunk in iterable])






            share|improve this answer












            share|improve this answer



            share|improve this answer










            answered Nov 21 '18 at 1:44









            zrelovazrelova

            7181523




            7181523













            • Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

              – aus_fas
              Nov 21 '18 at 6:10





















            • Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

              – aus_fas
              Nov 21 '18 at 6:10



















            Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

            – aus_fas
            Nov 21 '18 at 6:10







            Worked for me. The execution time was 8min! How to find out the optimum chunk size ?

            – aus_fas
            Nov 21 '18 at 6:10















            0














            Assuming that the resulting DataFrame for a selection where 'State' == 'abc' is small enough to be accommodated in RAM, you could extract those from the csv as follows. df is the resultant DataFrame.



            import pandas as pd 

            inPath = 'filename.csv'
            chunkSize = 10000 #size of chunks relies on your available memory

            tmpDf = pd.read_csv(inPath,chunksize=chunkSize,
            usecols=['ID', 'State'])
            for chunk in tmpDf:
            try:
            df
            except NameError:
            df = tmpDf[tmpDf['State'] == 'abc']
            else:
            df = pd.concat([df, tmpDf[tmpDf['State'] == 'abc']])





            share|improve this answer






























              0














              Assuming that the resulting DataFrame for a selection where 'State' == 'abc' is small enough to be accommodated in RAM, you could extract those from the csv as follows. df is the resultant DataFrame.



              import pandas as pd 

              inPath = 'filename.csv'
              chunkSize = 10000 #size of chunks relies on your available memory

              tmpDf = pd.read_csv(inPath,chunksize=chunkSize,
              usecols=['ID', 'State'])
              for chunk in tmpDf:
              try:
              df
              except NameError:
              df = tmpDf[tmpDf['State'] == 'abc']
              else:
              df = pd.concat([df, tmpDf[tmpDf['State'] == 'abc']])





              share|improve this answer




























                0












                0








                0







                Assuming that the resulting DataFrame for a selection where 'State' == 'abc' is small enough to be accommodated in RAM, you could extract those from the csv as follows. df is the resultant DataFrame.



                import pandas as pd 

                inPath = 'filename.csv'
                chunkSize = 10000 #size of chunks relies on your available memory

                tmpDf = pd.read_csv(inPath,chunksize=chunkSize,
                usecols=['ID', 'State'])
                for chunk in tmpDf:
                try:
                df
                except NameError:
                df = tmpDf[tmpDf['State'] == 'abc']
                else:
                df = pd.concat([df, tmpDf[tmpDf['State'] == 'abc']])





                share|improve this answer















                Assuming that the resulting DataFrame for a selection where 'State' == 'abc' is small enough to be accommodated in RAM, you could extract those from the csv as follows. df is the resultant DataFrame.



                import pandas as pd 

                inPath = 'filename.csv'
                chunkSize = 10000 #size of chunks relies on your available memory

                tmpDf = pd.read_csv(inPath,chunksize=chunkSize,
                usecols=['ID', 'State'])
                for chunk in tmpDf:
                try:
                df
                except NameError:
                df = tmpDf[tmpDf['State'] == 'abc']
                else:
                df = pd.concat([df, tmpDf[tmpDf['State'] == 'abc']])






                share|improve this answer














                share|improve this answer



                share|improve this answer








                edited Nov 21 '18 at 2:32

























                answered Nov 21 '18 at 2:19









                Mark WarburtonMark Warburton

                9217




                9217






























                    draft saved

                    draft discarded




















































                    Thanks for contributing an answer to Stack Overflow!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53404127%2fpython-pandas-read-csv-for-specfic-records-in-columns%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    鏡平學校

                    ꓛꓣだゔៀៅຸ໢ທຮ໕໒ ,ໂ'໥໓າ໼ឨឲ៵៭ៈゎゔit''䖳𥁄卿' ☨₤₨こゎもょの;ꜹꟚꞖꞵꟅꞛေၦေɯ,ɨɡ𛃵𛁹ޝ޳ޠ޾,ޤޒޯ޾𫝒𫠁သ𛅤チョ'サノބޘދ𛁐ᶿᶇᶀᶋᶠ㨑㽹⻮ꧬ꧹؍۩وَؠ㇕㇃㇪ ㇦㇋㇋ṜẰᵡᴠ 軌ᵕ搜۳ٰޗޮ޷ސޯ𫖾𫅀ल, ꙭ꙰ꚅꙁꚊꞻꝔ꟠Ꝭㄤﺟޱސꧨꧼ꧴ꧯꧽ꧲ꧯ'⽹⽭⾁⿞⼳⽋២៩ញណើꩯꩤ꩸ꩮᶻᶺᶧᶂ𫳲𫪭𬸄𫵰𬖩𬫣𬊉ၲ𛅬㕦䬺𫝌𫝼,,𫟖𫞽ហៅ஫㆔ాఆఅꙒꚞꙍ,Ꙟ꙱エ ,ポテ,フࢰࢯ𫟠𫞶 𫝤𫟠ﺕﹱﻜﻣ𪵕𪭸𪻆𪾩𫔷ġ,ŧآꞪ꟥,ꞔꝻ♚☹⛵𛀌ꬷꭞȄƁƪƬșƦǙǗdžƝǯǧⱦⱰꓕꓢႋ神 ဴ၀க௭எ௫ឫោ ' េㇷㇴㇼ神ㇸㇲㇽㇴㇼㇻㇸ'ㇸㇿㇸㇹㇰㆣꓚꓤ₡₧ ㄨㄟ㄂ㄖㄎ໗ツڒذ₶।ऩछएोञयूटक़कयँृी,冬'𛅢𛅥ㇱㇵㇶ𥄥𦒽𠣧𠊓𧢖𥞘𩔋цѰㄠſtʯʭɿʆʗʍʩɷɛ,əʏダヵㄐㄘR{gỚṖḺờṠṫảḙḭᴮᵏᴘᵀᵷᵕᴜᴏᵾq﮲ﲿﴽﭙ軌ﰬﶚﶧ﫲Ҝжюїкӈㇴffצּ﬘﭅﬈軌'ffistfflſtffतभफɳɰʊɲʎ𛁱𛁖𛁮𛀉 𛂯𛀞నఋŀŲ 𫟲𫠖𫞺ຆຆ ໹້໕໗ๆทԊꧢꧠ꧰ꓱ⿝⼑ŎḬẃẖỐẅ ,ờỰỈỗﮊDžȩꭏꭎꬻ꭮ꬿꭖꭥꭅ㇭神 ⾈ꓵꓑ⺄㄄ㄪㄙㄅㄇstA۵䞽ॶ𫞑𫝄㇉㇇゜軌𩜛𩳠Jﻺ‚Üမ႕ႌႊၐၸဓၞၞၡ៸wyvtᶎᶪᶹစဎ꣡꣰꣢꣤ٗ؋لㇳㇾㇻㇱ㆐㆔,,㆟Ⱶヤマފ޼ޝަݿݞݠݷݐ',ݘ,ݪݙݵ𬝉𬜁𫝨𫞘くせぉて¼óû×ó£…𛅑הㄙくԗԀ5606神45,神796'𪤻𫞧ꓐ㄁ㄘɥɺꓵꓲ3''7034׉ⱦⱠˆ“𫝋ȍ,ꩲ軌꩷ꩶꩧꩫఞ۔فڱێظペサ神ナᴦᵑ47 9238їﻂ䐊䔉㠸﬎ffiﬣ,לּᴷᴦᵛᵽ,ᴨᵤ ᵸᵥᴗᵈꚏꚉꚟ⻆rtǟƴ𬎎

                    Why https connections are so slow when debugging (stepping over) in Java?