Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-30542

Two Spark structured streaming jobs cannot write to same base path

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Invalid
    • 2.3.0
    • None
    • Structured Streaming
    • None

    Description

      Hi All,

      Spark Structured Streaming doesn't allow two structured streaming jobs to write data to the same base directory which is possible with using dstreams.

      As _spark_metadata directory will be created by default for one job, second job cannot use the same directory as base path as already _spark_metadata directory is created by other job, It is throwing exception.

      Is there any workaround for this, other than creating separate base path's for both the jobs.

      Is it possible to create the _spark_metadata directory else where or disable without any data loss.

      If I had to change the base path for both the jobs, then my whole framework will get impacted, So i don't want to do that.

       

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              SparkSiva Sivakumar
              Votes:
              0 Vote for this issue
              Watchers:
              5 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: