Is it possible to build spark code on fly and execute?












-1















I am trying to create a generic function to read a csv file using databriks CSV READER.But the option's are not mandatory it can differ based on the my input json configuration file.



Example1 :



 "ReaderOption":{

"delimiter":";",
"header":"true",
"inferSchema":"true",
"schema":"""some custome schema.."""
},


Example2:



   "ReaderOption":{

"delimiter":";",
"schema":"""some custome schema.."""
},


Is it possible to construct options or the entire read statement in runtime and run in spark ?



like below,



  def readCsvWithOptions(): DataFrame=
{

val options:Map[String,String]= Map("inferSchema"->"true")
val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
.option(options)
.load(inputPath)
readDF
}









share|improve this question























  • .options(options) helps to achieve the same.

    – William R
    Nov 19 '18 at 15:31
















-1















I am trying to create a generic function to read a csv file using databriks CSV READER.But the option's are not mandatory it can differ based on the my input json configuration file.



Example1 :



 "ReaderOption":{

"delimiter":";",
"header":"true",
"inferSchema":"true",
"schema":"""some custome schema.."""
},


Example2:



   "ReaderOption":{

"delimiter":";",
"schema":"""some custome schema.."""
},


Is it possible to construct options or the entire read statement in runtime and run in spark ?



like below,



  def readCsvWithOptions(): DataFrame=
{

val options:Map[String,String]= Map("inferSchema"->"true")
val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
.option(options)
.load(inputPath)
readDF
}









share|improve this question























  • .options(options) helps to achieve the same.

    – William R
    Nov 19 '18 at 15:31














-1












-1








-1








I am trying to create a generic function to read a csv file using databriks CSV READER.But the option's are not mandatory it can differ based on the my input json configuration file.



Example1 :



 "ReaderOption":{

"delimiter":";",
"header":"true",
"inferSchema":"true",
"schema":"""some custome schema.."""
},


Example2:



   "ReaderOption":{

"delimiter":";",
"schema":"""some custome schema.."""
},


Is it possible to construct options or the entire read statement in runtime and run in spark ?



like below,



  def readCsvWithOptions(): DataFrame=
{

val options:Map[String,String]= Map("inferSchema"->"true")
val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
.option(options)
.load(inputPath)
readDF
}









share|improve this question














I am trying to create a generic function to read a csv file using databriks CSV READER.But the option's are not mandatory it can differ based on the my input json configuration file.



Example1 :



 "ReaderOption":{

"delimiter":";",
"header":"true",
"inferSchema":"true",
"schema":"""some custome schema.."""
},


Example2:



   "ReaderOption":{

"delimiter":";",
"schema":"""some custome schema.."""
},


Is it possible to construct options or the entire read statement in runtime and run in spark ?



like below,



  def readCsvWithOptions(): DataFrame=
{

val options:Map[String,String]= Map("inferSchema"->"true")
val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
.option(options)
.load(inputPath)
readDF
}






apache-spark apache-spark-sql databricks






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Nov 19 '18 at 10:37









William RWilliam R

303419




303419













  • .options(options) helps to achieve the same.

    – William R
    Nov 19 '18 at 15:31



















  • .options(options) helps to achieve the same.

    – William R
    Nov 19 '18 at 15:31

















.options(options) helps to achieve the same.

– William R
Nov 19 '18 at 15:31





.options(options) helps to achieve the same.

– William R
Nov 19 '18 at 15:31












1 Answer
1






active

oldest

votes


















0














def readCsvWithOptions(): DataFrame=
{



val options:Map[String,String]= Map("inferSchema"->"true")
val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
.options(options)
.load(inputPath)
readDF
}



There is an options which takes key, value pair.






share|improve this answer























    Your Answer






    StackExchange.ifUsing("editor", function () {
    StackExchange.using("externalEditor", function () {
    StackExchange.using("snippets", function () {
    StackExchange.snippets.init();
    });
    });
    }, "code-snippets");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "1"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53372781%2fis-it-possible-to-build-spark-code-on-fly-and-execute%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    0














    def readCsvWithOptions(): DataFrame=
    {



    val options:Map[String,String]= Map("inferSchema"->"true")
    val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
    .options(options)
    .load(inputPath)
    readDF
    }



    There is an options which takes key, value pair.






    share|improve this answer




























      0














      def readCsvWithOptions(): DataFrame=
      {



      val options:Map[String,String]= Map("inferSchema"->"true")
      val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
      .options(options)
      .load(inputPath)
      readDF
      }



      There is an options which takes key, value pair.






      share|improve this answer


























        0












        0








        0







        def readCsvWithOptions(): DataFrame=
        {



        val options:Map[String,String]= Map("inferSchema"->"true")
        val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
        .options(options)
        .load(inputPath)
        readDF
        }



        There is an options which takes key, value pair.






        share|improve this answer













        def readCsvWithOptions(): DataFrame=
        {



        val options:Map[String,String]= Map("inferSchema"->"true")
        val readDF = jobContext.spark.read.format("com.databricks.spark.csv")
        .options(options)
        .load(inputPath)
        readDF
        }



        There is an options which takes key, value pair.







        share|improve this answer












        share|improve this answer



        share|improve this answer










        answered Nov 24 '18 at 20:47









        William RWilliam R

        303419




        303419






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Stack Overflow!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53372781%2fis-it-possible-to-build-spark-code-on-fly-and-execute%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            How to change which sound is reproduced for terminal bell?

            Title Spacing in Bjornstrup Chapter, Removing Chapter Number From Contents

            Can I use Tabulator js library in my java Spring + Thymeleaf project?