Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-14531 [Umbrella] Improve S3A error handling & reporting
  3. HADOOP-14714

handle InternalError in bulk object delete through retries

    XMLWordPrintableJSON

Details

    • Sub-task
    • Status: Resolved
    • Major
    • Resolution: Duplicate
    • 2.8.0
    • None
    • fs/s3
    • None

    Description

      There's some more detail appearing on HADOOP-11572 about the errors seen here; sounds like its large fileset related (or just probability working against you). Most importantly: retries may make it go away.

      Proposed: implement a retry policy.

      Issue: delete is not idempotent, not if someone else adds things.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              stevel@apache.org Steve Loughran
              Votes:
              0 Vote for this issue
              Watchers:
              2 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: