Uploaded image for project: 'Beam'
  1. Beam
  2. BEAM-12792

Multiple jobs running on Flink session cluster reuse the persistent Python environment.

Details

    • Bug
    • Status: Open
    • P1
    • Resolution: Unresolved
    • 2.27.0, 2.28.0, 2.29.0, 2.30.0, 2.31.0
    • None
    • sdk-py-harness
    • Kubernetes 1.20 on Ubuntu 18.04.
    • Important

    Description

      I'm running TFX pipelines on a Flink cluster using Beam in k8s. However, extra python packages passed to the Flink runner (or rather beam worker side-car) are only installed once per deployment cycle. Example:

      1. Flink is deployed and is up and running
      2. A TFX pipeline starts, submits a job to Flink along with a python whl of custom code and beam ops.
      3. The beam worker installs the package and the pipeline finishes succesfully.
      4. A new TFX pipeline is build where a new beam fn is introduced, the pipline is started and the new whl is submitted as in step 2).
      5. This time, the new package is not being installed in the beam worker causing the job to fail due to a reference which does not exist in the beam worker, since it didn't install the new package.

       

      I started using Flink from beam version 2.27 and it has been an issue all the time.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              ConverJens Jens Wiren
              Votes:
              2 Vote for this issue
              Watchers:
              5 Start watching this issue

              Dates

                Created:
                Updated:

                Time Tracking

                  Estimated:
                  Original Estimate - Not Specified
                  Not Specified
                  Remaining:
                  Remaining Estimate - 0h
                  0h
                  Logged:
                  Time Spent - 19.5h
                  19.5h