spark override the dataframe variable without using var










0















I have one API which perform delete operation on dataframe like below



def deleteColmns(df:DataFrame,clmList :List[org.apache.spark.sql.Column]):DataFrame
var ddf:DataFrame = null
for(clm<-clmList)
ddf.drop(clm)

return ddf



Since it is not good practice to use the var in functional programming , how to avoid this situation










share|improve this question


























    0















    I have one API which perform delete operation on dataframe like below



    def deleteColmns(df:DataFrame,clmList :List[org.apache.spark.sql.Column]):DataFrame
    var ddf:DataFrame = null
    for(clm<-clmList)
    ddf.drop(clm)

    return ddf



    Since it is not good practice to use the var in functional programming , how to avoid this situation










    share|improve this question
























      0












      0








      0








      I have one API which perform delete operation on dataframe like below



      def deleteColmns(df:DataFrame,clmList :List[org.apache.spark.sql.Column]):DataFrame
      var ddf:DataFrame = null
      for(clm<-clmList)
      ddf.drop(clm)

      return ddf



      Since it is not good practice to use the var in functional programming , how to avoid this situation










      share|improve this question














      I have one API which perform delete operation on dataframe like below



      def deleteColmns(df:DataFrame,clmList :List[org.apache.spark.sql.Column]):DataFrame
      var ddf:DataFrame = null
      for(clm<-clmList)
      ddf.drop(clm)

      return ddf



      Since it is not good practice to use the var in functional programming , how to avoid this situation







      apache-spark apache-spark-sql






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Nov 15 '18 at 13:25









      user3607698user3607698

      378117




      378117






















          1 Answer
          1






          active

          oldest

          votes


















          3














          With Spark >2.0, you can drop multiple columns using a sequence of column name :



          val clmList: Seq[Column] = _
          val strList: Seq[String] = clmList.map(c => s"$c")
          df.drop(strList: _*)


          Otherwise, you can always use foldLeft to fold left on the DataFrame and drop your columns :



          clmList.foldLeft(df)((acc, c) => acc.drop(c))


          I hope this helps.






          share|improve this answer
























            Your Answer






            StackExchange.ifUsing("editor", function ()
            StackExchange.using("externalEditor", function ()
            StackExchange.using("snippets", function ()
            StackExchange.snippets.init();
            );
            );
            , "code-snippets");

            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "1"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53320504%2fspark-override-the-dataframe-variable-without-using-var%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            3














            With Spark >2.0, you can drop multiple columns using a sequence of column name :



            val clmList: Seq[Column] = _
            val strList: Seq[String] = clmList.map(c => s"$c")
            df.drop(strList: _*)


            Otherwise, you can always use foldLeft to fold left on the DataFrame and drop your columns :



            clmList.foldLeft(df)((acc, c) => acc.drop(c))


            I hope this helps.






            share|improve this answer





























              3














              With Spark >2.0, you can drop multiple columns using a sequence of column name :



              val clmList: Seq[Column] = _
              val strList: Seq[String] = clmList.map(c => s"$c")
              df.drop(strList: _*)


              Otherwise, you can always use foldLeft to fold left on the DataFrame and drop your columns :



              clmList.foldLeft(df)((acc, c) => acc.drop(c))


              I hope this helps.






              share|improve this answer



























                3












                3








                3







                With Spark >2.0, you can drop multiple columns using a sequence of column name :



                val clmList: Seq[Column] = _
                val strList: Seq[String] = clmList.map(c => s"$c")
                df.drop(strList: _*)


                Otherwise, you can always use foldLeft to fold left on the DataFrame and drop your columns :



                clmList.foldLeft(df)((acc, c) => acc.drop(c))


                I hope this helps.






                share|improve this answer















                With Spark >2.0, you can drop multiple columns using a sequence of column name :



                val clmList: Seq[Column] = _
                val strList: Seq[String] = clmList.map(c => s"$c")
                df.drop(strList: _*)


                Otherwise, you can always use foldLeft to fold left on the DataFrame and drop your columns :



                clmList.foldLeft(df)((acc, c) => acc.drop(c))


                I hope this helps.







                share|improve this answer














                share|improve this answer



                share|improve this answer








                edited Nov 15 '18 at 21:46

























                answered Nov 15 '18 at 14:03









                eliasaheliasah

                27.5k774117




                27.5k774117





























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Stack Overflow!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53320504%2fspark-override-the-dataframe-variable-without-using-var%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    這個網誌中的熱門文章

                    How to read a connectionString WITH PROVIDER in .NET Core?

                    Node.js Script on GitHub Pages or Amazon S3

                    Museum of Modern and Contemporary Art of Trento and Rovereto