Class org.broadinstitute.hellbender.tools.spark.pipelines.MarkDuplicatesSparkIntegrationTest
|
100%
successful |
Tests
Test | Duration | Result |
---|---|---|
assertThatExpectedOutputUpdateToggleIsDisabled | 0s | passed |
testAssertAssumeUnsortedFilesAreQueryGroupedFiles | 0.543s | passed |
testAssertCorrectSortOrderMultipleBams | 0.148s | passed |
testAssertCorrectSortOrderMultipleBamsOverriding | 0.499s | passed |
testBulkFragmentsNoDuplicates | 1.074s | passed |
testBulkFragmentsWithDuplicates | 2.141s | passed |
testDifferentChromosomesInOppositeOrder | 0.321s | passed |
testDuplicateDetectionDataProviderWithMetrics[0](src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/example.chr1.1-1K.unmarkedDups.bam, src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/expected.picard-2.15.0.sorted.chr1.1-1K.unmarkedDups.markDuplicate.metrics) | 0.271s | passed |
testDuplicateDetectionDataProviderWithMetrics[1](src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/optical_dupes.bam, src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/expected.picard-2.15.0.sorted.optical_dupes.markDuplicate.metrics) | 0.250s | passed |
testDuplicateDetectionDataProviderWithMetrics[2](src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/inputSingleLibrarySolexa16404.bam, src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/expected.picard-2.15.0.sorted.inputSingleLibrarySolexa16404.metrics) | 0.251s | passed |
testFlowModeFlag | 1.673s | passed |
testHashCollisionHandling | 1.030s | passed |
testMDOrder[0](src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/mdOrderBug.bam, src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/expected.mdOrderBug.bam) | 0.278s | passed |
testMDOrder[1](src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/mdOrderBug2.bam, src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/expected.mdOrderBug2.bam) | 0.279s | passed |
testManyOpticalDuplicateClusterOneEndSamePositionOneCluster[0](true, true, true) | 0.335s | passed |
testManyOpticalDuplicateClusterOneEndSamePositionOneCluster[1](true, false, true) | 0.318s | passed |
testManyOpticalDuplicateClusterOneEndSamePositionOneCluster[2](false, true, true) | 0.317s | passed |
testManyOpticalDuplicateClusterOneEndSamePositionOneCluster[3](true, true, false) | 0.303s | passed |
testManyOpticalDuplicateClusterOneEndSamePositionOneCluster[4](true, false, false) | 0.323s | passed |
testManyOpticalDuplicateClusterOneEndSamePositionOneCluster[5](false, true, false) | 0.314s | passed |
testMappedFragmentAndMappedPairFirstOfPairNonPrimary | 0.317s | passed |
testMappedFragmentAndMatePairFirstUnmapped | 0.317s | passed |
testMappedFragmentAndMatePairSecondUnmapped | 0.300s | passed |
testMappedPairAndMappedFragmentAndMatePairFirstUnmapped | 0.307s | passed |
testMappedPairAndMappedFragmentAndMatePairSecondUnmapped | 0.318s | passed |
testMappedPairAndMatePairFirstOppositeStrandSecondUnmapped | 0.307s | passed |
testMappedPairAndMatePairFirstUnmapped | 0.309s | passed |
testMappedPairAndMatePairSecondUnmapped | 0.302s | passed |
testMappedPairWithFirstEndSamePositionAndOther | 0.309s | passed |
testMappedPairWithSamePosition | 0.306s | passed |
testMappedPairWithSamePositionSameCigar | 0.309s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[0]([Ljava.io.File;@59002daa, 20, 0, {Solexa-16419=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16416=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16404=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16406=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16412=[0, 1, 0, 0, 0, 0, 0.0, 0]}) | 0.351s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[1]([Ljava.io.File;@4aba9704, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.360s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[2]([Ljava.io.File;@71f2b12d, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.374s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[3]([Ljava.io.File;@6769a486, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.335s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[4]([Ljava.io.File;@76b89abc, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.339s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[5]([Ljava.io.File;@29a2c491, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.190s | passed |
testMarkDuplicatesSparkDeletingDuplicateReads[6]([Ljava.io.File;@37545110, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.221s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[0]([Ljava.io.File;@722c91ca, 20, 0, {Solexa-16419=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16416=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16404=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16406=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16412=[0, 1, 0, 0, 0, 0, 0.0, 0]}) | 0.368s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[1]([Ljava.io.File;@30999cf, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.350s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[2]([Ljava.io.File;@28334f90, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.358s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[3]([Ljava.io.File;@17809ed7, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.335s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[4]([Ljava.io.File;@37dc5cb4, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.343s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[5]([Ljava.io.File;@435d2c70, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.193s | passed |
testMarkDuplicatesSparkDeletingOpticalDuplicateReads[6]([Ljava.io.File;@50826f9f, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.157s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[0]([Ljava.io.File;@6b3c06fc, 20, 0, {Solexa-16419=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16416=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16404=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16406=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16412=[0, 1, 0, 0, 0, 0, 0.0, 0]}) | 0.329s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[1]([Ljava.io.File;@4678b390, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.351s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[2]([Ljava.io.File;@dbdac72, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.362s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[3]([Ljava.io.File;@71bb5c2, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.308s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[4]([Ljava.io.File;@3daed2a5, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.329s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[5]([Ljava.io.File;@63d8b56e, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.170s | passed |
testMarkDuplicatesSparkIntegrationTestLocal[6]([Ljava.io.File;@1cc1b097, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.129s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[0]([Ljava.io.File;@598287f0, 20, 0, {Solexa-16419=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16416=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16404=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16406=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16412=[0, 1, 0, 0, 0, 0, 0.0, 0]}) | 0.345s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[1]([Ljava.io.File;@251c503f, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.343s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[2]([Ljava.io.File;@2a88bb75, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.349s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[3]([Ljava.io.File;@df1f0c5, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.318s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[4]([Ljava.io.File;@791f270d, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.306s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[5]([Ljava.io.File;@7a67bac2, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.162s | passed |
testMarkDuplicatesSparkMarkingAllDuplicatesWithTagging[6]([Ljava.io.File;@38266b3a, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.197s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[0]([Ljava.io.File;@6142abde, 20, 0, {Solexa-16419=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16416=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16404=[0, 3, 0, 0, 0, 0, 0.0, 0], Solexa-16406=[0, 1, 0, 0, 0, 0, 0.0, 0], Solexa-16412=[0, 1, 0, 0, 0, 0, 0.0, 0]}) | 0.341s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[1]([Ljava.io.File;@122713d6, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.343s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[2]([Ljava.io.File;@5e792f38, 90, 6, {Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 0.349s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[3]([Ljava.io.File;@600de63a, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.341s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[4]([Ljava.io.File;@515e4df5, 4, 2, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0]}) | 0.322s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[5]([Ljava.io.File;@10851d4b, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.173s | passed |
testMarkDuplicatesSparkMarkingOpticalDuplicatesWithTagging[6]([Ljava.io.File;@31f758a2, 94, 8, {mylib=[0, 2, 0, 0, 1, 1, 0.5, 0], Solexa-16419=[4, 4, 4, 0, 0, 0, 0.0, 0], Solexa-16416=[2, 2, 2, 0, 0, 0, 0.0, 0], Solexa-16404=[3, 9, 3, 0, 2, 0, 0.190476, 17], Solexa-16406=[1, 10, 1, 0, 0, 0, 0.0, 0], Solexa-16412=[3, 6, 3, 0, 1, 0, 0.133333, 15]}) | 1.187s | passed |
testMatePairFirstUnmapped | 0.313s | passed |
testMatePairSecondUnmapped | 0.307s | passed |
testNoReadGroupInRead | 0.255s | passed |
testNonExistantReadGroupInRead | 0.241s | passed |
testNullOpticalDuplicates | 0.311s | passed |
testOpticalDuplicateClusterOneEndSamePositionOneCluster | 0.310s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicatesWithinPixelDistance | 0.309s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[0](RUNID:7:1203:2886:82292, RUNID:7:1205:3886:16834) | 0.329s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[10](RUNID:7:1203:6886:16756, RUNID:7:1203:4886:26756) | 0.298s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[11](RUNID:7:1203:2886:34756, RUNID:7:1203:2886:36756) | 0.311s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[12](RUNID:7:1203:2886:35756, RUNID:7:1203:2886:36756) | 0.311s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[13](RUNID:7:1203:2886:37756, RUNID:7:1203:2886:36756) | 0.305s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[14](RUNID:7:1203:2886:38756, RUNID:7:1203:2886:36756) | 0.304s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[15](HK3T5CCXX160204:3:1112:11586:37067, HK3T5CCXX160204:3:1112:11586:32144) | 0.308s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[1](RUNID:7:1203:2886:16756, RUNID:7:1205:3886:16756) | 0.315s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[2](RUNID:7:1204:2886:16756, RUNID:7:1205:3886:16756) | 0.310s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[3](RUNID:7:1205:2886:16756, RUNID:7:1205:3886:16756) | 0.314s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[4](RUNID:7:1206:2886:16756, RUNID:7:1205:3886:16756) | 0.298s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[5](RUNID:7:1207:2886:16756, RUNID:7:1205:3886:16756) | 0.313s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[6](RUNID:7:1203:2886:16756, RUNID:7:1203:4886:26756) | 0.301s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[7](RUNID:7:1203:3886:16756, RUNID:7:1203:4886:26756) | 0.308s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[8](RUNID:7:1203:4886:16756, RUNID:7:1203:4886:26756) | 0.304s | passed |
testOpticalDuplicateClusterSamePositionNoOpticalDuplicates[9](RUNID:7:1203:5886:16756, RUNID:7:1203:4886:26756) | 0.312s | passed |
testOpticalDuplicateClusterSamePositionOneOpticalDuplicatesWithinPixelDistance | 0.300s | passed |
testOpticalDuplicateClustersAddingSecondEndFirstSameCoordinate | 0.310s | passed |
testOpticalDuplicateFinding | 0.300s | passed |
testOpticalDuplicateFindingPxelDistance | 0.310s | passed |
testOpticalDuplicatesAndPCRDuplicatesOrientationDifference | 0.298s | passed |
testOpticalDuplicatesDifferentReadGroups | 0.306s | passed |
testOpticalDuplicatesTheSameReadGroup | 0.301s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[0](RUNID:7:1203:2886:82292, RUNID:7:1205:3886:16834) | 0.299s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[10](RUNID:7:1203:6886:16756, RUNID:7:1203:4886:26756) | 0.313s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[11](RUNID:7:1203:2886:34756, RUNID:7:1203:2886:36756) | 0.310s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[12](RUNID:7:1203:2886:35756, RUNID:7:1203:2886:36756) | 0.305s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[13](RUNID:7:1203:2886:37756, RUNID:7:1203:2886:36756) | 0.304s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[14](RUNID:7:1203:2886:38756, RUNID:7:1203:2886:36756) | 0.306s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[15](HK3T5CCXX160204:3:1112:11586:37067, HK3T5CCXX160204:3:1112:11586:32144) | 0.295s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[1](RUNID:7:1203:2886:16756, RUNID:7:1205:3886:16756) | 0.305s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[2](RUNID:7:1204:2886:16756, RUNID:7:1205:3886:16756) | 0.304s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[3](RUNID:7:1205:2886:16756, RUNID:7:1205:3886:16756) | 0.305s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[4](RUNID:7:1206:2886:16756, RUNID:7:1205:3886:16756) | 0.309s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[5](RUNID:7:1207:2886:16756, RUNID:7:1205:3886:16756) | 0.309s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[6](RUNID:7:1203:2886:16756, RUNID:7:1203:4886:26756) | 0.297s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[7](RUNID:7:1203:3886:16756, RUNID:7:1203:4886:26756) | 0.310s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[8](RUNID:7:1203:4886:16756, RUNID:7:1203:4886:26756) | 0.306s | passed |
testOpticalDuplicatesTiebrokenByPhysicalLocationNotStartPosition[9](RUNID:7:1203:5886:16756, RUNID:7:1203:4886:26756) | 0.307s | passed |
testPathologicalOrderingAtTheSamePosition | 0.306s | passed |
testReadSameStartPositionOrientationOverride | 0.300s | passed |
testReadsHaveSameStartPositionButDifferentChromosomeNonEquivalence | 0.316s | passed |
testSecondEndIsBeforeFirstInCoordinate | 0.306s | passed |
testSingleMappedFragment | 0.307s | passed |
testSingleMappedFragmentAndSingleMappedPair | 0.313s | passed |
testSingleMappedFragmentAndTwoMappedPairs | 0.302s | passed |
testSingleMappedPair | 0.304s | passed |
testSingleUnmappedFragment | 0.290s | passed |
testSingleUnmappedPair | 0.296s | passed |
testStackOverFlowPairSetSwap | 0.300s | passed |
testSupplementaryReadUnmappedMate | 0.314s | passed |
testThreeGroupsOnDifferentChromosomesOfThreeMappedPairs | 0.302s | passed |
testThreeMappedPairs | 0.302s | passed |
testThreeMappedPairsWithMatchingSecondMate | 0.300s | passed |
testTwoGroupsOnDifferentChromosomesOfThreeMappedPairs | 0.302s | passed |
testTwoGroupsOnDifferentChromosomesOfTwoFragments | 0.311s | passed |
testTwoGroupsOnDifferentChromosomesOfTwoMappedPairs | 0.330s | passed |
testTwoMappedFragments | 0.314s | passed |
testTwoMappedPairWithSamePosition | 0.292s | passed |
testTwoMappedPairWithSamePositionDifferentStrands | 0.297s | passed |
testTwoMappedPairWithSamePositionDifferentStrands2 | 0.291s | passed |
testTwoMappedPairs | 0.296s | passed |
testTwoMappedPairsAndMappedSecondaryFragment | 0.288s | passed |
testTwoMappedPairsAndTerminalUnmappedFragment | 0.299s | passed |
testTwoMappedPairsAndTerminalUnmappedPair | 0.294s | passed |
testTwoMappedPairsMatesSoftClipped | 0.302s | passed |
testTwoMappedPairsWithOppositeOrientations | 0.303s | passed |
testTwoMappedPairsWithOppositeOrientationsNumberTwo | 0.291s | passed |
testTwoMappedPairsWithSoftClipping | 0.305s | passed |
testTwoMappedPairsWithSoftClippingBoth | 0.298s | passed |
testTwoMappedPairsWithSoftClippingFirstOfPairOnly | 0.300s | passed |
testTwoMappedPairsWithSoftClippingFirstOfPairOnlyNoMateCigar | 0.300s | passed |
testTwoMappedPairsWithSupplementaryReads[0](true, true, true) | 0.302s | passed |
testTwoMappedPairsWithSupplementaryReads[1](true, false, true) | 0.291s | passed |
testTwoMappedPairsWithSupplementaryReads[2](false, true, true) | 0.302s | passed |
testTwoMappedPairsWithSupplementaryReads[3](true, true, false) | 0.300s | passed |
testTwoMappedPairsWithSupplementaryReads[4](true, false, false) | 0.301s | passed |
testTwoMappedPairsWithSupplementaryReads[5](false, true, false) | 0.301s | passed |
testTwoUnmappedFragments | 0.293s | passed |
Standard error
[May 19, 2025 at 4:01:38 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.01 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:38 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:39 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.01 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:40 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:42 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 16:01:42.052 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:01:42.052 INFO MarkDuplicatesSpark - The Genome Analysis Toolkit (GATK) va6bd333-SNAPSHOT 16:01:42.052 INFO MarkDuplicatesSpark - For support and documentation go to https://software.broadinstitute.org/gatk/ 16:01:42.052 INFO MarkDuplicatesSpark - Executing as root@67ac59543751 on Linux v6.11.0-1014-azure amd64 16:01:42.052 INFO MarkDuplicatesSpark - Java runtime: OpenJDK 64-Bit Server VM v17.0.12+7-Ubuntu-1ubuntu222.04 16:01:42.052 INFO MarkDuplicatesSpark - Start Date/Time: May 19, 2025 at 4:01:42 PM GMT 16:01:42.052 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:01:42.052 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:01:42.053 INFO MarkDuplicatesSpark - HTSJDK Version: test_cram_wiring-SNAPSHOT 16:01:42.053 INFO MarkDuplicatesSpark - Picard Version: 3.4.0 16:01:42.053 INFO MarkDuplicatesSpark - Built for Spark Version: 3.5.0 16:01:42.053 INFO MarkDuplicatesSpark - HTSJDK Defaults.COMPRESSION_LEVEL : 2 16:01:42.053 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_READ_FOR_SAMTOOLS : false 16:01:42.053 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_SAMTOOLS : true 16:01:42.053 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_TRIBBLE : false 16:01:42.053 INFO MarkDuplicatesSpark - Deflater: IntelDeflater 16:01:42.053 INFO MarkDuplicatesSpark - Inflater: IntelInflater 16:01:42.053 INFO MarkDuplicatesSpark - GCS max retries/reopens: 20 16:01:42.053 INFO MarkDuplicatesSpark - Requester pays: disabled 16:01:42.053 INFO MarkDuplicatesSpark - Initializing engine 16:01:42.053 INFO MarkDuplicatesSpark - Done initializing engine 16:01:42.053 INFO MarkDuplicatesSpark - Spark verbosity set to INFO (see --spark-verbosity argument) 16:01:42.056 INFO MemoryStore - Block broadcast_1411 stored as values in memory (estimated size 305.5 KiB, free 1903.3 MiB) 16:01:42.062 INFO MemoryStore - Block broadcast_1411_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1903.2 MiB) 16:01:42.063 INFO BlockManagerInfo - Added broadcast_1411_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1918.9 MiB) 16:01:42.063 INFO SparkContext - Created broadcast 1411 from newAPIHadoopFile at PathSplitSource.java:96 16:01:42.083 INFO MemoryStore - Block broadcast_1412 stored as values in memory (estimated size 305.5 KiB, free 1902.9 MiB) 16:01:42.089 INFO MemoryStore - Block broadcast_1412_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1902.9 MiB) 16:01:42.089 INFO BlockManagerInfo - Added broadcast_1412_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1918.9 MiB) 16:01:42.089 INFO SparkContext - Created broadcast 1412 from newAPIHadoopFile at PathSplitSource.java:96 16:01:42.121 INFO SparkContext - Starting job: collect at SparkUtils.java:205 16:01:42.122 INFO FileInputFormat - Total input files to process : 1 16:01:42.122 INFO DAGScheduler - Registering RDD 3614 (mapToPair at SparkUtils.java:161) as input to shuffle 97 16:01:42.123 INFO DAGScheduler - Got job 433 (collect at SparkUtils.java:205) with 1 output partitions 16:01:42.123 INFO DAGScheduler - Final stage: ResultStage 741 (collect at SparkUtils.java:205) 16:01:42.123 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 740) 16:01:42.123 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 740) 16:01:42.123 INFO DAGScheduler - Submitting ShuffleMapStage 740 (MapPartitionsRDD[3614] at mapToPair at SparkUtils.java:161), which has no missing parents 16:01:42.140 INFO MemoryStore - Block broadcast_1413 stored as values in memory (estimated size 459.2 KiB, free 1902.4 MiB) 16:01:42.142 INFO MemoryStore - Block broadcast_1413_piece0 stored as bytes in memory (estimated size 201.4 KiB, free 1902.2 MiB) 16:01:42.143 INFO BlockManagerInfo - Added broadcast_1413_piece0 in memory on localhost:42397 (size: 201.4 KiB, free: 1918.7 MiB) 16:01:42.143 INFO SparkContext - Created broadcast 1413 from broadcast at DAGScheduler.scala:1580 16:01:42.143 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 740 (MapPartitionsRDD[3614] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:01:42.143 INFO TaskSchedulerImpl - Adding task set 740.0 with 1 tasks resource profile 0 16:01:42.144 INFO TaskSetManager - Starting task 0.0 in stage 740.0 (TID 670) (localhost, executor driver, partition 0, PROCESS_LOCAL, 9650 bytes) 16:01:42.145 INFO Executor - Running task 0.0 in stage 740.0 (TID 670) 16:01:42.173 INFO NewHadoopRDD - Input split: file:/gatkCloneMountPoint/src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/mdOrderBug.bam:0+1507 16:01:42.177 INFO Executor - Finished task 0.0 in stage 740.0 (TID 670). 1148 bytes result sent to driver 16:01:42.178 INFO TaskSetManager - Finished task 0.0 in stage 740.0 (TID 670) in 34 ms on localhost (executor driver) (1/1) 16:01:42.178 INFO TaskSchedulerImpl - Removed TaskSet 740.0, whose tasks have all completed, from pool 16:01:42.178 INFO DAGScheduler - ShuffleMapStage 740 (mapToPair at SparkUtils.java:161) finished in 0.054 s 16:01:42.178 INFO DAGScheduler - looking for newly runnable stages 16:01:42.178 INFO DAGScheduler - running: HashSet() 16:01:42.178 INFO DAGScheduler - waiting: HashSet(ResultStage 741) 16:01:42.178 INFO DAGScheduler - failed: HashSet() 16:01:42.178 INFO DAGScheduler - Submitting ResultStage 741 (MapPartitionsRDD[3617] at mapPartitions at SparkUtils.java:188), which has no missing parents 16:01:42.179 INFO MemoryStore - Block broadcast_1414 stored as values in memory (estimated size 7.4 KiB, free 1902.2 MiB) 16:01:42.179 INFO MemoryStore - Block broadcast_1414_piece0 stored as bytes in memory (estimated size 4.1 KiB, free 1902.2 MiB) 16:01:42.180 INFO BlockManagerInfo - Added broadcast_1414_piece0 in memory on localhost:42397 (size: 4.1 KiB, free: 1918.7 MiB) 16:01:42.180 INFO SparkContext - Created broadcast 1414 from broadcast at DAGScheduler.scala:1580 16:01:42.180 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 741 (MapPartitionsRDD[3617] at mapPartitions at SparkUtils.java:188) (first 15 tasks are for partitions Vector(0)) 16:01:42.180 INFO TaskSchedulerImpl - Adding task set 741.0 with 1 tasks resource profile 0 16:01:42.181 INFO TaskSetManager - Starting task 0.0 in stage 741.0 (TID 671) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:01:42.181 INFO Executor - Running task 0.0 in stage 741.0 (TID 671) 16:01:42.182 INFO ShuffleBlockFetcherIterator - Getting 1 (405.0 B) non-empty blocks including 1 (405.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.182 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.184 INFO Executor - Finished task 0.0 in stage 741.0 (TID 671). 2238 bytes result sent to driver 16:01:42.184 INFO TaskSetManager - Finished task 0.0 in stage 741.0 (TID 671) in 4 ms on localhost (executor driver) (1/1) 16:01:42.184 INFO TaskSchedulerImpl - Removed TaskSet 741.0, whose tasks have all completed, from pool 16:01:42.185 INFO DAGScheduler - ResultStage 741 (collect at SparkUtils.java:205) finished in 0.005 s 16:01:42.185 INFO DAGScheduler - Job 433 is finished. Cancelling potential speculative or zombie tasks for this job 16:01:42.185 INFO TaskSchedulerImpl - Killing all running tasks in stage 741: Stage finished 16:01:42.185 INFO DAGScheduler - Job 433 finished: collect at SparkUtils.java:205, took 0.063488 s 16:01:42.189 INFO MemoryStore - Block broadcast_1415 stored as values in memory (estimated size 1632.0 B, free 1902.2 MiB) 16:01:42.189 INFO MemoryStore - Block broadcast_1415_piece0 stored as bytes in memory (estimated size 145.0 B, free 1902.2 MiB) 16:01:42.190 INFO BlockManagerInfo - Added broadcast_1415_piece0 in memory on localhost:42397 (size: 145.0 B, free: 1918.7 MiB) 16:01:42.190 INFO SparkContext - Created broadcast 1415 from broadcast at MarkDuplicatesSparkUtils.java:126 16:01:42.190 INFO MemoryStore - Block broadcast_1416 stored as values in memory (estimated size 304.0 B, free 1902.2 MiB) 16:01:42.190 INFO MemoryStore - Block broadcast_1416_piece0 stored as bytes in memory (estimated size 37.0 B, free 1902.2 MiB) 16:01:42.191 INFO BlockManagerInfo - Added broadcast_1416_piece0 in memory on localhost:42397 (size: 37.0 B, free: 1918.7 MiB) 16:01:42.191 INFO SparkContext - Created broadcast 1416 from broadcast at MarkDuplicatesSparkUtils.java:127 16:01:42.202 INFO SparkContext - Starting job: collectAsMap at MarkDuplicatesSparkUtils.java:542 16:01:42.203 INFO DAGScheduler - Registering RDD 3626 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) as input to shuffle 100 16:01:42.203 INFO DAGScheduler - Registering RDD 3630 (mapToPair at MarkDuplicatesSpark.java:217) as input to shuffle 99 16:01:42.203 INFO DAGScheduler - Registering RDD 3634 (mapToPair at MarkDuplicatesSparkUtils.java:497) as input to shuffle 98 16:01:42.203 INFO DAGScheduler - Got job 434 (collectAsMap at MarkDuplicatesSparkUtils.java:542) with 1 output partitions 16:01:42.203 INFO DAGScheduler - Final stage: ResultStage 746 (collectAsMap at MarkDuplicatesSparkUtils.java:542) 16:01:42.203 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 745) 16:01:42.204 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 745) 16:01:42.204 INFO DAGScheduler - Submitting ShuffleMapStage 743 (MapPartitionsRDD[3626] at flatMapToPair at MarkDuplicatesSparkUtils.java:130), which has no missing parents 16:01:42.205 INFO MemoryStore - Block broadcast_1417 stored as values in memory (estimated size 24.2 KiB, free 1902.2 MiB) 16:01:42.205 INFO MemoryStore - Block broadcast_1417_piece0 stored as bytes in memory (estimated size 12.6 KiB, free 1902.2 MiB) 16:01:42.205 INFO BlockManagerInfo - Added broadcast_1417_piece0 in memory on localhost:42397 (size: 12.6 KiB, free: 1918.6 MiB) 16:01:42.206 INFO SparkContext - Created broadcast 1417 from broadcast at DAGScheduler.scala:1580 16:01:42.206 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 743 (MapPartitionsRDD[3626] at flatMapToPair at MarkDuplicatesSparkUtils.java:130) (first 15 tasks are for partitions Vector(0)) 16:01:42.206 INFO TaskSchedulerImpl - Adding task set 743.0 with 1 tasks resource profile 0 16:01:42.206 INFO TaskSetManager - Starting task 0.0 in stage 743.0 (TID 672) (localhost, executor driver, partition 0, NODE_LOCAL, 9813 bytes) 16:01:42.207 INFO Executor - Running task 0.0 in stage 743.0 (TID 672) 16:01:42.208 INFO ShuffleBlockFetcherIterator - Getting 1 (405.0 B) non-empty blocks including 1 (405.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.208 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.211 INFO Executor - Finished task 0.0 in stage 743.0 (TID 672). 1922 bytes result sent to driver 16:01:42.211 INFO TaskSetManager - Finished task 0.0 in stage 743.0 (TID 672) in 5 ms on localhost (executor driver) (1/1) 16:01:42.212 INFO TaskSchedulerImpl - Removed TaskSet 743.0, whose tasks have all completed, from pool 16:01:42.212 INFO DAGScheduler - ShuffleMapStage 743 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) finished in 0.008 s 16:01:42.212 INFO DAGScheduler - looking for newly runnable stages 16:01:42.212 INFO DAGScheduler - running: HashSet() 16:01:42.212 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 744, ShuffleMapStage 745, ResultStage 746) 16:01:42.212 INFO DAGScheduler - failed: HashSet() 16:01:42.212 INFO DAGScheduler - Submitting ShuffleMapStage 744 (MapPartitionsRDD[3630] at mapToPair at MarkDuplicatesSpark.java:217), which has no missing parents 16:01:42.213 INFO MemoryStore - Block broadcast_1418 stored as values in memory (estimated size 28.3 KiB, free 1902.1 MiB) 16:01:42.213 INFO MemoryStore - Block broadcast_1418_piece0 stored as bytes in memory (estimated size 14.8 KiB, free 1902.1 MiB) 16:01:42.214 INFO BlockManagerInfo - Added broadcast_1418_piece0 in memory on localhost:42397 (size: 14.8 KiB, free: 1918.6 MiB) 16:01:42.214 INFO SparkContext - Created broadcast 1418 from broadcast at DAGScheduler.scala:1580 16:01:42.214 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 744 (MapPartitionsRDD[3630] at mapToPair at MarkDuplicatesSpark.java:217) (first 15 tasks are for partitions Vector(0)) 16:01:42.214 INFO TaskSchedulerImpl - Adding task set 744.0 with 1 tasks resource profile 0 16:01:42.214 INFO TaskSetManager - Starting task 0.0 in stage 744.0 (TID 673) (localhost, executor driver, partition 0, NODE_LOCAL, 9288 bytes) 16:01:42.215 INFO Executor - Running task 0.0 in stage 744.0 (TID 673) 16:01:42.217 INFO ShuffleBlockFetcherIterator - Getting 1 (129.0 B) non-empty blocks including 1 (129.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.217 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.220 INFO Executor - Finished task 0.0 in stage 744.0 (TID 673). 1922 bytes result sent to driver 16:01:42.220 INFO TaskSetManager - Finished task 0.0 in stage 744.0 (TID 673) in 6 ms on localhost (executor driver) (1/1) 16:01:42.220 INFO TaskSchedulerImpl - Removed TaskSet 744.0, whose tasks have all completed, from pool 16:01:42.220 INFO DAGScheduler - ShuffleMapStage 744 (mapToPair at MarkDuplicatesSpark.java:217) finished in 0.008 s 16:01:42.221 INFO DAGScheduler - looking for newly runnable stages 16:01:42.221 INFO DAGScheduler - running: HashSet() 16:01:42.221 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 745, ResultStage 746) 16:01:42.221 INFO DAGScheduler - failed: HashSet() 16:01:42.221 INFO DAGScheduler - Submitting ShuffleMapStage 745 (MapPartitionsRDD[3634] at mapToPair at MarkDuplicatesSparkUtils.java:497), which has no missing parents 16:01:42.222 INFO MemoryStore - Block broadcast_1419 stored as values in memory (estimated size 19.0 KiB, free 1902.1 MiB) 16:01:42.222 INFO MemoryStore - Block broadcast_1419_piece0 stored as bytes in memory (estimated size 9.7 KiB, free 1902.1 MiB) 16:01:42.222 INFO BlockManagerInfo - Added broadcast_1419_piece0 in memory on localhost:42397 (size: 9.7 KiB, free: 1918.6 MiB) 16:01:42.222 INFO SparkContext - Created broadcast 1419 from broadcast at DAGScheduler.scala:1580 16:01:42.222 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 745 (MapPartitionsRDD[3634] at mapToPair at MarkDuplicatesSparkUtils.java:497) (first 15 tasks are for partitions Vector(0)) 16:01:42.223 INFO TaskSchedulerImpl - Adding task set 745.0 with 1 tasks resource profile 0 16:01:42.223 INFO TaskSetManager - Starting task 0.0 in stage 745.0 (TID 674) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:01:42.223 INFO Executor - Running task 0.0 in stage 745.0 (TID 674) 16:01:42.225 INFO ShuffleBlockFetcherIterator - Getting 1 (405.0 B) non-empty blocks including 1 (405.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.225 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.226 INFO ShuffleBlockFetcherIterator - Getting 1 (117.0 B) non-empty blocks including 1 (117.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.226 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.228 INFO Executor - Finished task 0.0 in stage 745.0 (TID 674). 1922 bytes result sent to driver 16:01:42.228 INFO TaskSetManager - Finished task 0.0 in stage 745.0 (TID 674) in 5 ms on localhost (executor driver) (1/1) 16:01:42.228 INFO TaskSchedulerImpl - Removed TaskSet 745.0, whose tasks have all completed, from pool 16:01:42.228 INFO DAGScheduler - ShuffleMapStage 745 (mapToPair at MarkDuplicatesSparkUtils.java:497) finished in 0.007 s 16:01:42.228 INFO DAGScheduler - looking for newly runnable stages 16:01:42.228 INFO DAGScheduler - running: HashSet() 16:01:42.229 INFO DAGScheduler - waiting: HashSet(ResultStage 746) 16:01:42.229 INFO DAGScheduler - failed: HashSet() 16:01:42.229 INFO DAGScheduler - Submitting ResultStage 746 (MapPartitionsRDD[3636] at mapValues at MarkDuplicatesSparkUtils.java:519), which has no missing parents 16:01:42.229 INFO MemoryStore - Block broadcast_1420 stored as values in memory (estimated size 20.2 KiB, free 1902.1 MiB) 16:01:42.230 INFO MemoryStore - Block broadcast_1420_piece0 stored as bytes in memory (estimated size 10.6 KiB, free 1902.1 MiB) 16:01:42.230 INFO BlockManagerInfo - Added broadcast_1420_piece0 in memory on localhost:42397 (size: 10.6 KiB, free: 1918.6 MiB) 16:01:42.230 INFO SparkContext - Created broadcast 1420 from broadcast at DAGScheduler.scala:1580 16:01:42.230 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 746 (MapPartitionsRDD[3636] at mapValues at MarkDuplicatesSparkUtils.java:519) (first 15 tasks are for partitions Vector(0)) 16:01:42.230 INFO TaskSchedulerImpl - Adding task set 746.0 with 1 tasks resource profile 0 16:01:42.231 INFO TaskSetManager - Starting task 0.0 in stage 746.0 (TID 675) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:01:42.231 INFO Executor - Running task 0.0 in stage 746.0 (TID 675) 16:01:42.232 INFO ShuffleBlockFetcherIterator - Getting 1 (117.0 B) non-empty blocks including 1 (117.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.232 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.234 INFO Executor - Finished task 0.0 in stage 746.0 (TID 675). 1901 bytes result sent to driver 16:01:42.234 INFO TaskSetManager - Finished task 0.0 in stage 746.0 (TID 675) in 3 ms on localhost (executor driver) (1/1) 16:01:42.234 INFO TaskSchedulerImpl - Removed TaskSet 746.0, whose tasks have all completed, from pool 16:01:42.234 INFO DAGScheduler - ResultStage 746 (collectAsMap at MarkDuplicatesSparkUtils.java:542) finished in 0.005 s 16:01:42.234 INFO DAGScheduler - Job 434 is finished. Cancelling potential speculative or zombie tasks for this job 16:01:42.234 INFO TaskSchedulerImpl - Killing all running tasks in stage 746: Stage finished 16:01:42.234 INFO DAGScheduler - Job 434 finished: collectAsMap at MarkDuplicatesSparkUtils.java:542, took 0.031984 s 16:01:42.244 INFO MemoryStore - Block broadcast_1421 stored as values in memory (estimated size 20.3 KiB, free 1902.0 MiB) 16:01:42.245 INFO MemoryStore - Block broadcast_1421_piece0 stored as bytes in memory (estimated size 1850.0 B, free 1902.0 MiB) 16:01:42.245 INFO BlockManagerInfo - Added broadcast_1421_piece0 in memory on localhost:42397 (size: 1850.0 B, free: 1918.6 MiB) 16:01:42.245 INFO SparkContext - Created broadcast 1421 from broadcast at ReadsSparkSink.java:133 16:01:42.246 INFO MemoryStore - Block broadcast_1422 stored as values in memory (estimated size 20.4 KiB, free 1902.0 MiB) 16:01:42.246 INFO MemoryStore - Block broadcast_1422_piece0 stored as bytes in memory (estimated size 1850.0 B, free 1902.0 MiB) 16:01:42.246 INFO BlockManagerInfo - Added broadcast_1422_piece0 in memory on localhost:42397 (size: 1850.0 B, free: 1918.6 MiB) 16:01:42.246 INFO SparkContext - Created broadcast 1422 from broadcast at BamSink.java:76 16:01:42.248 INFO PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 16:01:42.248 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:01:42.248 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:01:42.259 INFO SparkContext - Starting job: runJob at SparkHadoopWriter.scala:83 16:01:42.259 INFO DAGScheduler - Registering RDD 3638 (mapToPair at SparkUtils.java:161) as input to shuffle 101 16:01:42.260 INFO DAGScheduler - Got job 435 (runJob at SparkHadoopWriter.scala:83) with 1 output partitions 16:01:42.260 INFO DAGScheduler - Final stage: ResultStage 751 (runJob at SparkHadoopWriter.scala:83) 16:01:42.260 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 750) 16:01:42.260 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 750) 16:01:42.260 INFO DAGScheduler - Submitting ShuffleMapStage 750 (MapPartitionsRDD[3638] at mapToPair at SparkUtils.java:161), which has no missing parents 16:01:42.261 INFO MemoryStore - Block broadcast_1423 stored as values in memory (estimated size 18.1 KiB, free 1902.0 MiB) 16:01:42.261 INFO MemoryStore - Block broadcast_1423_piece0 stored as bytes in memory (estimated size 9.5 KiB, free 1902.0 MiB) 16:01:42.261 INFO BlockManagerInfo - Added broadcast_1423_piece0 in memory on localhost:42397 (size: 9.5 KiB, free: 1918.6 MiB) 16:01:42.261 INFO SparkContext - Created broadcast 1423 from broadcast at DAGScheduler.scala:1580 16:01:42.262 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 750 (MapPartitionsRDD[3638] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:01:42.262 INFO TaskSchedulerImpl - Adding task set 750.0 with 1 tasks resource profile 0 16:01:42.262 INFO TaskSetManager - Starting task 0.0 in stage 750.0 (TID 676) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:01:42.262 INFO Executor - Running task 0.0 in stage 750.0 (TID 676) 16:01:42.264 INFO ShuffleBlockFetcherIterator - Getting 1 (405.0 B) non-empty blocks including 1 (405.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.264 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.266 INFO ShuffleBlockFetcherIterator - Getting 1 (117.0 B) non-empty blocks including 1 (117.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.266 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.267 INFO Executor - Finished task 0.0 in stage 750.0 (TID 676). 1879 bytes result sent to driver 16:01:42.267 INFO TaskSetManager - Finished task 0.0 in stage 750.0 (TID 676) in 5 ms on localhost (executor driver) (1/1) 16:01:42.268 INFO TaskSchedulerImpl - Removed TaskSet 750.0, whose tasks have all completed, from pool 16:01:42.268 INFO DAGScheduler - ShuffleMapStage 750 (mapToPair at SparkUtils.java:161) finished in 0.008 s 16:01:42.268 INFO DAGScheduler - looking for newly runnable stages 16:01:42.268 INFO DAGScheduler - running: HashSet() 16:01:42.268 INFO DAGScheduler - waiting: HashSet(ResultStage 751) 16:01:42.268 INFO DAGScheduler - failed: HashSet() 16:01:42.268 INFO DAGScheduler - Submitting ResultStage 751 (MapPartitionsRDD[3643] at mapToPair at BamSink.java:91), which has no missing parents 16:01:42.274 INFO MemoryStore - Block broadcast_1424 stored as values in memory (estimated size 163.3 KiB, free 1901.8 MiB) 16:01:42.275 INFO MemoryStore - Block broadcast_1424_piece0 stored as bytes in memory (estimated size 73.4 KiB, free 1901.8 MiB) 16:01:42.275 INFO BlockManagerInfo - Added broadcast_1424_piece0 in memory on localhost:42397 (size: 73.4 KiB, free: 1918.5 MiB) 16:01:42.276 INFO SparkContext - Created broadcast 1424 from broadcast at DAGScheduler.scala:1580 16:01:42.276 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 751 (MapPartitionsRDD[3643] at mapToPair at BamSink.java:91) (first 15 tasks are for partitions Vector(0)) 16:01:42.276 INFO TaskSchedulerImpl - Adding task set 751.0 with 1 tasks resource profile 0 16:01:42.276 INFO TaskSetManager - Starting task 0.0 in stage 751.0 (TID 677) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:01:42.276 INFO Executor - Running task 0.0 in stage 751.0 (TID 677) 16:01:42.280 INFO ShuffleBlockFetcherIterator - Getting 1 (405.0 B) non-empty blocks including 1 (405.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.280 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.282 INFO PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 16:01:42.282 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:01:42.282 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:01:42.282 INFO PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 16:01:42.282 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:01:42.282 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:01:42.292 INFO FileOutputCommitter - Saved output of task 'attempt_20250519160142968355963334720284_3643_r_000000_0' to file:/tmp/local13774000465916765003/markdups2178781706614791083.bam.parts/_temporary/0/task_20250519160142968355963334720284_3643_r_000000 16:01:42.292 INFO SparkHadoopMapRedUtil - attempt_20250519160142968355963334720284_3643_r_000000_0: Committed. Elapsed time: 0 ms. 16:01:42.292 INFO Executor - Finished task 0.0 in stage 751.0 (TID 677). 1858 bytes result sent to driver 16:01:42.292 INFO TaskSetManager - Finished task 0.0 in stage 751.0 (TID 677) in 16 ms on localhost (executor driver) (1/1) 16:01:42.292 INFO TaskSchedulerImpl - Removed TaskSet 751.0, whose tasks have all completed, from pool 16:01:42.293 INFO DAGScheduler - ResultStage 751 (runJob at SparkHadoopWriter.scala:83) finished in 0.025 s 16:01:42.293 INFO DAGScheduler - Job 435 is finished. Cancelling potential speculative or zombie tasks for this job 16:01:42.293 INFO TaskSchedulerImpl - Killing all running tasks in stage 751: Stage finished 16:01:42.293 INFO DAGScheduler - Job 435 finished: runJob at SparkHadoopWriter.scala:83, took 0.033837 s 16:01:42.293 INFO SparkHadoopWriter - Start to commit write Job job_20250519160142968355963334720284_3643. 16:01:42.296 INFO SparkHadoopWriter - Write Job job_20250519160142968355963334720284_3643 committed. Elapsed time: 3 ms. 16:01:42.304 INFO HadoopFileSystemWrapper - Concatenating 3 parts to /tmp/local13774000465916765003/markdups2178781706614791083.bam 16:01:42.306 INFO HadoopFileSystemWrapper - Concatenating to /tmp/local13774000465916765003/markdups2178781706614791083.bam done 16:01:42.306 INFO IndexFileMerger - Merging .sbi files in temp directory /tmp/local13774000465916765003/markdups2178781706614791083.bam.parts/ to /tmp/local13774000465916765003/markdups2178781706614791083.bam.sbi 16:01:42.309 INFO IndexFileMerger - Done merging .sbi files 16:01:42.309 INFO IndexFileMerger - Merging .bai files in temp directory /tmp/local13774000465916765003/markdups2178781706614791083.bam.parts/ to /tmp/local13774000465916765003/markdups2178781706614791083.bam.bai 16:01:42.312 INFO IndexFileMerger - Done merging .bai files 16:01:42.312 INFO MarkDuplicatesSpark - Shutting down engine [May 19, 2025 at 4:01:42 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 16:01:42.329 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:01:42.329 INFO MarkDuplicatesSpark - The Genome Analysis Toolkit (GATK) va6bd333-SNAPSHOT 16:01:42.329 INFO MarkDuplicatesSpark - For support and documentation go to https://software.broadinstitute.org/gatk/ 16:01:42.329 INFO MarkDuplicatesSpark - Executing as root@67ac59543751 on Linux v6.11.0-1014-azure amd64 16:01:42.329 INFO MarkDuplicatesSpark - Java runtime: OpenJDK 64-Bit Server VM v17.0.12+7-Ubuntu-1ubuntu222.04 16:01:42.330 INFO MarkDuplicatesSpark - Start Date/Time: May 19, 2025 at 4:01:42 PM GMT 16:01:42.330 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:01:42.330 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:01:42.330 INFO MarkDuplicatesSpark - HTSJDK Version: test_cram_wiring-SNAPSHOT 16:01:42.330 INFO MarkDuplicatesSpark - Picard Version: 3.4.0 16:01:42.330 INFO MarkDuplicatesSpark - Built for Spark Version: 3.5.0 16:01:42.330 INFO MarkDuplicatesSpark - HTSJDK Defaults.COMPRESSION_LEVEL : 2 16:01:42.330 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_READ_FOR_SAMTOOLS : false 16:01:42.330 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_SAMTOOLS : true 16:01:42.330 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_TRIBBLE : false 16:01:42.330 INFO MarkDuplicatesSpark - Deflater: IntelDeflater 16:01:42.330 INFO MarkDuplicatesSpark - Inflater: IntelInflater 16:01:42.330 INFO MarkDuplicatesSpark - GCS max retries/reopens: 20 16:01:42.330 INFO MarkDuplicatesSpark - Requester pays: disabled 16:01:42.330 INFO MarkDuplicatesSpark - Initializing engine 16:01:42.330 INFO MarkDuplicatesSpark - Done initializing engine 16:01:42.330 INFO MarkDuplicatesSpark - Spark verbosity set to INFO (see --spark-verbosity argument) 16:01:42.332 INFO MemoryStore - Block broadcast_1425 stored as values in memory (estimated size 305.5 KiB, free 1901.5 MiB) 16:01:42.339 INFO MemoryStore - Block broadcast_1425_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1901.4 MiB) 16:01:42.339 INFO BlockManagerInfo - Added broadcast_1425_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1918.5 MiB) 16:01:42.339 INFO SparkContext - Created broadcast 1425 from newAPIHadoopFile at PathSplitSource.java:96 16:01:42.359 INFO MemoryStore - Block broadcast_1426 stored as values in memory (estimated size 305.5 KiB, free 1901.1 MiB) 16:01:42.365 INFO MemoryStore - Block broadcast_1426_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1901.0 MiB) 16:01:42.366 INFO BlockManagerInfo - Added broadcast_1426_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1918.4 MiB) 16:01:42.366 INFO SparkContext - Created broadcast 1426 from newAPIHadoopFile at PathSplitSource.java:96 16:01:42.397 INFO SparkContext - Starting job: collect at SparkUtils.java:205 16:01:42.397 INFO FileInputFormat - Total input files to process : 1 16:01:42.398 INFO DAGScheduler - Registering RDD 3657 (mapToPair at SparkUtils.java:161) as input to shuffle 102 16:01:42.398 INFO DAGScheduler - Got job 436 (collect at SparkUtils.java:205) with 1 output partitions 16:01:42.398 INFO DAGScheduler - Final stage: ResultStage 753 (collect at SparkUtils.java:205) 16:01:42.398 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 752) 16:01:42.398 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 752) 16:01:42.398 INFO DAGScheduler - Submitting ShuffleMapStage 752 (MapPartitionsRDD[3657] at mapToPair at SparkUtils.java:161), which has no missing parents 16:01:42.415 INFO MemoryStore - Block broadcast_1427 stored as values in memory (estimated size 459.2 KiB, free 1900.6 MiB) 16:01:42.417 INFO MemoryStore - Block broadcast_1427_piece0 stored as bytes in memory (estimated size 201.4 KiB, free 1900.4 MiB) 16:01:42.417 INFO BlockManagerInfo - Added broadcast_1427_piece0 in memory on localhost:42397 (size: 201.4 KiB, free: 1918.2 MiB) 16:01:42.418 INFO SparkContext - Created broadcast 1427 from broadcast at DAGScheduler.scala:1580 16:01:42.418 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 752 (MapPartitionsRDD[3657] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:01:42.418 INFO TaskSchedulerImpl - Adding task set 752.0 with 1 tasks resource profile 0 16:01:42.418 INFO TaskSetManager - Starting task 0.0 in stage 752.0 (TID 678) (localhost, executor driver, partition 0, PROCESS_LOCAL, 9651 bytes) 16:01:42.418 INFO Executor - Running task 0.0 in stage 752.0 (TID 678) 16:01:42.447 INFO NewHadoopRDD - Input split: file:/gatkCloneMountPoint/src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/mdOrderBug2.bam:0+1856 16:01:42.451 INFO Executor - Finished task 0.0 in stage 752.0 (TID 678). 1148 bytes result sent to driver 16:01:42.451 INFO TaskSetManager - Finished task 0.0 in stage 752.0 (TID 678) in 33 ms on localhost (executor driver) (1/1) 16:01:42.451 INFO TaskSchedulerImpl - Removed TaskSet 752.0, whose tasks have all completed, from pool 16:01:42.451 INFO DAGScheduler - ShuffleMapStage 752 (mapToPair at SparkUtils.java:161) finished in 0.053 s 16:01:42.451 INFO DAGScheduler - looking for newly runnable stages 16:01:42.451 INFO DAGScheduler - running: HashSet() 16:01:42.451 INFO DAGScheduler - waiting: HashSet(ResultStage 753) 16:01:42.451 INFO DAGScheduler - failed: HashSet() 16:01:42.452 INFO DAGScheduler - Submitting ResultStage 753 (MapPartitionsRDD[3660] at mapPartitions at SparkUtils.java:188), which has no missing parents 16:01:42.452 INFO MemoryStore - Block broadcast_1428 stored as values in memory (estimated size 7.4 KiB, free 1900.4 MiB) 16:01:42.453 INFO MemoryStore - Block broadcast_1428_piece0 stored as bytes in memory (estimated size 4.1 KiB, free 1900.4 MiB) 16:01:42.453 INFO BlockManagerInfo - Added broadcast_1428_piece0 in memory on localhost:42397 (size: 4.1 KiB, free: 1918.2 MiB) 16:01:42.453 INFO SparkContext - Created broadcast 1428 from broadcast at DAGScheduler.scala:1580 16:01:42.453 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 753 (MapPartitionsRDD[3660] at mapPartitions at SparkUtils.java:188) (first 15 tasks are for partitions Vector(0)) 16:01:42.453 INFO TaskSchedulerImpl - Adding task set 753.0 with 1 tasks resource profile 0 16:01:42.454 INFO TaskSetManager - Starting task 0.0 in stage 753.0 (TID 679) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:01:42.454 INFO Executor - Running task 0.0 in stage 753.0 (TID 679) 16:01:42.455 INFO ShuffleBlockFetcherIterator - Getting 1 (868.0 B) non-empty blocks including 1 (868.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.455 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.457 INFO Executor - Finished task 0.0 in stage 753.0 (TID 679). 2495 bytes result sent to driver 16:01:42.457 INFO TaskSetManager - Finished task 0.0 in stage 753.0 (TID 679) in 4 ms on localhost (executor driver) (1/1) 16:01:42.457 INFO TaskSchedulerImpl - Removed TaskSet 753.0, whose tasks have all completed, from pool 16:01:42.457 INFO DAGScheduler - ResultStage 753 (collect at SparkUtils.java:205) finished in 0.005 s 16:01:42.457 INFO DAGScheduler - Job 436 is finished. Cancelling potential speculative or zombie tasks for this job 16:01:42.457 INFO TaskSchedulerImpl - Killing all running tasks in stage 753: Stage finished 16:01:42.457 INFO DAGScheduler - Job 436 finished: collect at SparkUtils.java:205, took 0.060395 s 16:01:42.463 INFO MemoryStore - Block broadcast_1429 stored as values in memory (estimated size 1632.0 B, free 1900.4 MiB) 16:01:42.463 INFO MemoryStore - Block broadcast_1429_piece0 stored as bytes in memory (estimated size 145.0 B, free 1900.4 MiB) 16:01:42.463 INFO BlockManagerInfo - Added broadcast_1429_piece0 in memory on localhost:42397 (size: 145.0 B, free: 1918.2 MiB) 16:01:42.463 INFO SparkContext - Created broadcast 1429 from broadcast at MarkDuplicatesSparkUtils.java:126 16:01:42.464 INFO MemoryStore - Block broadcast_1430 stored as values in memory (estimated size 304.0 B, free 1900.4 MiB) 16:01:42.464 INFO MemoryStore - Block broadcast_1430_piece0 stored as bytes in memory (estimated size 37.0 B, free 1900.4 MiB) 16:01:42.464 INFO BlockManagerInfo - Added broadcast_1430_piece0 in memory on localhost:42397 (size: 37.0 B, free: 1918.2 MiB) 16:01:42.464 INFO SparkContext - Created broadcast 1430 from broadcast at MarkDuplicatesSparkUtils.java:127 16:01:42.476 INFO SparkContext - Starting job: collectAsMap at MarkDuplicatesSparkUtils.java:542 16:01:42.476 INFO DAGScheduler - Registering RDD 3669 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) as input to shuffle 105 16:01:42.476 INFO DAGScheduler - Registering RDD 3673 (mapToPair at MarkDuplicatesSpark.java:217) as input to shuffle 104 16:01:42.476 INFO DAGScheduler - Registering RDD 3677 (mapToPair at MarkDuplicatesSparkUtils.java:497) as input to shuffle 103 16:01:42.477 INFO DAGScheduler - Got job 437 (collectAsMap at MarkDuplicatesSparkUtils.java:542) with 1 output partitions 16:01:42.477 INFO DAGScheduler - Final stage: ResultStage 758 (collectAsMap at MarkDuplicatesSparkUtils.java:542) 16:01:42.477 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 757) 16:01:42.477 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 757) 16:01:42.477 INFO DAGScheduler - Submitting ShuffleMapStage 755 (MapPartitionsRDD[3669] at flatMapToPair at MarkDuplicatesSparkUtils.java:130), which has no missing parents 16:01:42.478 INFO MemoryStore - Block broadcast_1431 stored as values in memory (estimated size 24.2 KiB, free 1900.4 MiB) 16:01:42.478 INFO MemoryStore - Block broadcast_1431_piece0 stored as bytes in memory (estimated size 12.6 KiB, free 1900.3 MiB) 16:01:42.478 INFO BlockManagerInfo - Added broadcast_1431_piece0 in memory on localhost:42397 (size: 12.6 KiB, free: 1918.2 MiB) 16:01:42.479 INFO SparkContext - Created broadcast 1431 from broadcast at DAGScheduler.scala:1580 16:01:42.479 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 755 (MapPartitionsRDD[3669] at flatMapToPair at MarkDuplicatesSparkUtils.java:130) (first 15 tasks are for partitions Vector(0)) 16:01:42.479 INFO TaskSchedulerImpl - Adding task set 755.0 with 1 tasks resource profile 0 16:01:42.479 INFO TaskSetManager - Starting task 0.0 in stage 755.0 (TID 680) (localhost, executor driver, partition 0, NODE_LOCAL, 9813 bytes) 16:01:42.480 INFO Executor - Running task 0.0 in stage 755.0 (TID 680) 16:01:42.481 INFO ShuffleBlockFetcherIterator - Getting 1 (868.0 B) non-empty blocks including 1 (868.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.481 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.484 INFO Executor - Finished task 0.0 in stage 755.0 (TID 680). 1922 bytes result sent to driver 16:01:42.484 INFO TaskSetManager - Finished task 0.0 in stage 755.0 (TID 680) in 5 ms on localhost (executor driver) (1/1) 16:01:42.484 INFO TaskSchedulerImpl - Removed TaskSet 755.0, whose tasks have all completed, from pool 16:01:42.484 INFO DAGScheduler - ShuffleMapStage 755 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) finished in 0.007 s 16:01:42.484 INFO DAGScheduler - looking for newly runnable stages 16:01:42.484 INFO DAGScheduler - running: HashSet() 16:01:42.484 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 756, ShuffleMapStage 757, ResultStage 758) 16:01:42.485 INFO DAGScheduler - failed: HashSet() 16:01:42.485 INFO DAGScheduler - Submitting ShuffleMapStage 756 (MapPartitionsRDD[3673] at mapToPair at MarkDuplicatesSpark.java:217), which has no missing parents 16:01:42.485 INFO MemoryStore - Block broadcast_1432 stored as values in memory (estimated size 28.3 KiB, free 1900.3 MiB) 16:01:42.486 INFO MemoryStore - Block broadcast_1432_piece0 stored as bytes in memory (estimated size 14.8 KiB, free 1900.3 MiB) 16:01:42.486 INFO BlockManagerInfo - Added broadcast_1432_piece0 in memory on localhost:42397 (size: 14.8 KiB, free: 1918.2 MiB) 16:01:42.486 INFO SparkContext - Created broadcast 1432 from broadcast at DAGScheduler.scala:1580 16:01:42.486 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 756 (MapPartitionsRDD[3673] at mapToPair at MarkDuplicatesSpark.java:217) (first 15 tasks are for partitions Vector(0)) 16:01:42.486 INFO TaskSchedulerImpl - Adding task set 756.0 with 1 tasks resource profile 0 16:01:42.487 INFO TaskSetManager - Starting task 0.0 in stage 756.0 (TID 681) (localhost, executor driver, partition 0, NODE_LOCAL, 9288 bytes) 16:01:42.487 INFO Executor - Running task 0.0 in stage 756.0 (TID 681) 16:01:42.488 INFO ShuffleBlockFetcherIterator - Getting 1 (142.0 B) non-empty blocks including 1 (142.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.488 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.491 INFO Executor - Finished task 0.0 in stage 756.0 (TID 681). 1879 bytes result sent to driver 16:01:42.491 INFO TaskSetManager - Finished task 0.0 in stage 756.0 (TID 681) in 5 ms on localhost (executor driver) (1/1) 16:01:42.491 INFO TaskSchedulerImpl - Removed TaskSet 756.0, whose tasks have all completed, from pool 16:01:42.492 INFO DAGScheduler - ShuffleMapStage 756 (mapToPair at MarkDuplicatesSpark.java:217) finished in 0.007 s 16:01:42.492 INFO DAGScheduler - looking for newly runnable stages 16:01:42.492 INFO DAGScheduler - running: HashSet() 16:01:42.492 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 757, ResultStage 758) 16:01:42.492 INFO DAGScheduler - failed: HashSet() 16:01:42.492 INFO DAGScheduler - Submitting ShuffleMapStage 757 (MapPartitionsRDD[3677] at mapToPair at MarkDuplicatesSparkUtils.java:497), which has no missing parents 16:01:42.493 INFO MemoryStore - Block broadcast_1433 stored as values in memory (estimated size 19.0 KiB, free 1900.3 MiB) 16:01:42.493 INFO MemoryStore - Block broadcast_1433_piece0 stored as bytes in memory (estimated size 9.7 KiB, free 1900.3 MiB) 16:01:42.493 INFO BlockManagerInfo - Added broadcast_1433_piece0 in memory on localhost:42397 (size: 9.7 KiB, free: 1918.2 MiB) 16:01:42.493 INFO SparkContext - Created broadcast 1433 from broadcast at DAGScheduler.scala:1580 16:01:42.493 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 757 (MapPartitionsRDD[3677] at mapToPair at MarkDuplicatesSparkUtils.java:497) (first 15 tasks are for partitions Vector(0)) 16:01:42.494 INFO TaskSchedulerImpl - Adding task set 757.0 with 1 tasks resource profile 0 16:01:42.494 INFO TaskSetManager - Starting task 0.0 in stage 757.0 (TID 682) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:01:42.494 INFO Executor - Running task 0.0 in stage 757.0 (TID 682) 16:01:42.496 INFO ShuffleBlockFetcherIterator - Getting 1 (868.0 B) non-empty blocks including 1 (868.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.496 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.497 INFO ShuffleBlockFetcherIterator - Getting 1 (117.0 B) non-empty blocks including 1 (117.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.497 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.498 INFO Executor - Finished task 0.0 in stage 757.0 (TID 682). 1879 bytes result sent to driver 16:01:42.499 INFO TaskSetManager - Finished task 0.0 in stage 757.0 (TID 682) in 5 ms on localhost (executor driver) (1/1) 16:01:42.499 INFO TaskSchedulerImpl - Removed TaskSet 757.0, whose tasks have all completed, from pool 16:01:42.499 INFO DAGScheduler - ShuffleMapStage 757 (mapToPair at MarkDuplicatesSparkUtils.java:497) finished in 0.007 s 16:01:42.499 INFO DAGScheduler - looking for newly runnable stages 16:01:42.499 INFO DAGScheduler - running: HashSet() 16:01:42.499 INFO DAGScheduler - waiting: HashSet(ResultStage 758) 16:01:42.499 INFO DAGScheduler - failed: HashSet() 16:01:42.499 INFO DAGScheduler - Submitting ResultStage 758 (MapPartitionsRDD[3679] at mapValues at MarkDuplicatesSparkUtils.java:519), which has no missing parents 16:01:42.500 INFO MemoryStore - Block broadcast_1434 stored as values in memory (estimated size 20.2 KiB, free 1900.3 MiB) 16:01:42.500 INFO MemoryStore - Block broadcast_1434_piece0 stored as bytes in memory (estimated size 10.6 KiB, free 1900.2 MiB) 16:01:42.500 INFO BlockManagerInfo - Added broadcast_1434_piece0 in memory on localhost:42397 (size: 10.6 KiB, free: 1918.1 MiB) 16:01:42.500 INFO SparkContext - Created broadcast 1434 from broadcast at DAGScheduler.scala:1580 16:01:42.501 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 758 (MapPartitionsRDD[3679] at mapValues at MarkDuplicatesSparkUtils.java:519) (first 15 tasks are for partitions Vector(0)) 16:01:42.501 INFO TaskSchedulerImpl - Adding task set 758.0 with 1 tasks resource profile 0 16:01:42.501 INFO TaskSetManager - Starting task 0.0 in stage 758.0 (TID 683) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:01:42.501 INFO Executor - Running task 0.0 in stage 758.0 (TID 683) 16:01:42.503 INFO ShuffleBlockFetcherIterator - Getting 1 (142.0 B) non-empty blocks including 1 (142.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.503 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.505 INFO Executor - Finished task 0.0 in stage 758.0 (TID 683). 2053 bytes result sent to driver 16:01:42.505 INFO TaskSetManager - Finished task 0.0 in stage 758.0 (TID 683) in 4 ms on localhost (executor driver) (1/1) 16:01:42.505 INFO TaskSchedulerImpl - Removed TaskSet 758.0, whose tasks have all completed, from pool 16:01:42.505 INFO DAGScheduler - ResultStage 758 (collectAsMap at MarkDuplicatesSparkUtils.java:542) finished in 0.006 s 16:01:42.506 INFO DAGScheduler - Job 437 is finished. Cancelling potential speculative or zombie tasks for this job 16:01:42.506 INFO TaskSchedulerImpl - Killing all running tasks in stage 758: Stage finished 16:01:42.506 INFO DAGScheduler - Job 437 finished: collectAsMap at MarkDuplicatesSparkUtils.java:542, took 0.029992 s 16:01:42.516 INFO MemoryStore - Block broadcast_1435 stored as values in memory (estimated size 20.3 KiB, free 1900.2 MiB) 16:01:42.516 INFO MemoryStore - Block broadcast_1435_piece0 stored as bytes in memory (estimated size 1850.0 B, free 1900.2 MiB) 16:01:42.516 INFO BlockManagerInfo - Added broadcast_1435_piece0 in memory on localhost:42397 (size: 1850.0 B, free: 1918.1 MiB) 16:01:42.517 INFO SparkContext - Created broadcast 1435 from broadcast at ReadsSparkSink.java:133 16:01:42.517 INFO MemoryStore - Block broadcast_1436 stored as values in memory (estimated size 20.4 KiB, free 1900.2 MiB) 16:01:42.518 INFO MemoryStore - Block broadcast_1436_piece0 stored as bytes in memory (estimated size 1850.0 B, free 1900.2 MiB) 16:01:42.518 INFO BlockManagerInfo - Added broadcast_1436_piece0 in memory on localhost:42397 (size: 1850.0 B, free: 1918.1 MiB) 16:01:42.518 INFO SparkContext - Created broadcast 1436 from broadcast at BamSink.java:76 16:01:42.519 INFO PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 16:01:42.519 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:01:42.519 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:01:42.531 INFO SparkContext - Starting job: runJob at SparkHadoopWriter.scala:83 16:01:42.531 INFO DAGScheduler - Registering RDD 3681 (mapToPair at SparkUtils.java:161) as input to shuffle 106 16:01:42.531 INFO DAGScheduler - Got job 438 (runJob at SparkHadoopWriter.scala:83) with 1 output partitions 16:01:42.531 INFO DAGScheduler - Final stage: ResultStage 763 (runJob at SparkHadoopWriter.scala:83) 16:01:42.532 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 762) 16:01:42.532 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 762) 16:01:42.532 INFO DAGScheduler - Submitting ShuffleMapStage 762 (MapPartitionsRDD[3681] at mapToPair at SparkUtils.java:161), which has no missing parents 16:01:42.533 INFO MemoryStore - Block broadcast_1437 stored as values in memory (estimated size 18.1 KiB, free 1900.2 MiB) 16:01:42.533 INFO MemoryStore - Block broadcast_1437_piece0 stored as bytes in memory (estimated size 9.5 KiB, free 1900.2 MiB) 16:01:42.533 INFO BlockManagerInfo - Added broadcast_1437_piece0 in memory on localhost:42397 (size: 9.5 KiB, free: 1918.1 MiB) 16:01:42.533 INFO SparkContext - Created broadcast 1437 from broadcast at DAGScheduler.scala:1580 16:01:42.533 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 762 (MapPartitionsRDD[3681] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:01:42.534 INFO TaskSchedulerImpl - Adding task set 762.0 with 1 tasks resource profile 0 16:01:42.534 INFO TaskSetManager - Starting task 0.0 in stage 762.0 (TID 684) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:01:42.534 INFO Executor - Running task 0.0 in stage 762.0 (TID 684) 16:01:42.536 INFO ShuffleBlockFetcherIterator - Getting 1 (868.0 B) non-empty blocks including 1 (868.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.536 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.537 INFO ShuffleBlockFetcherIterator - Getting 1 (117.0 B) non-empty blocks including 1 (117.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.538 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.539 INFO Executor - Finished task 0.0 in stage 762.0 (TID 684). 1879 bytes result sent to driver 16:01:42.539 INFO TaskSetManager - Finished task 0.0 in stage 762.0 (TID 684) in 5 ms on localhost (executor driver) (1/1) 16:01:42.539 INFO TaskSchedulerImpl - Removed TaskSet 762.0, whose tasks have all completed, from pool 16:01:42.539 INFO DAGScheduler - ShuffleMapStage 762 (mapToPair at SparkUtils.java:161) finished in 0.007 s 16:01:42.539 INFO DAGScheduler - looking for newly runnable stages 16:01:42.539 INFO DAGScheduler - running: HashSet() 16:01:42.540 INFO DAGScheduler - waiting: HashSet(ResultStage 763) 16:01:42.540 INFO DAGScheduler - failed: HashSet() 16:01:42.540 INFO DAGScheduler - Submitting ResultStage 763 (MapPartitionsRDD[3686] at mapToPair at BamSink.java:91), which has no missing parents 16:01:42.546 INFO MemoryStore - Block broadcast_1438 stored as values in memory (estimated size 163.3 KiB, free 1900.0 MiB) 16:01:42.547 INFO MemoryStore - Block broadcast_1438_piece0 stored as bytes in memory (estimated size 73.4 KiB, free 1899.9 MiB) 16:01:42.547 INFO BlockManagerInfo - Added broadcast_1438_piece0 in memory on localhost:42397 (size: 73.4 KiB, free: 1918.1 MiB) 16:01:42.547 INFO SparkContext - Created broadcast 1438 from broadcast at DAGScheduler.scala:1580 16:01:42.548 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 763 (MapPartitionsRDD[3686] at mapToPair at BamSink.java:91) (first 15 tasks are for partitions Vector(0)) 16:01:42.548 INFO TaskSchedulerImpl - Adding task set 763.0 with 1 tasks resource profile 0 16:01:42.548 INFO TaskSetManager - Starting task 0.0 in stage 763.0 (TID 685) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:01:42.548 INFO Executor - Running task 0.0 in stage 763.0 (TID 685) 16:01:42.552 INFO ShuffleBlockFetcherIterator - Getting 1 (868.0 B) non-empty blocks including 1 (868.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:01:42.552 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:01:42.554 INFO PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 16:01:42.554 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:01:42.554 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:01:42.554 INFO PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 16:01:42.554 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:01:42.554 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:01:42.565 INFO FileOutputCommitter - Saved output of task 'attempt_202505191601427956918440701693653_3686_r_000000_0' to file:/tmp/local13774000465916765003/markdups4450785990897483788.bam.parts/_temporary/0/task_202505191601427956918440701693653_3686_r_000000 16:01:42.565 INFO SparkHadoopMapRedUtil - attempt_202505191601427956918440701693653_3686_r_000000_0: Committed. Elapsed time: 0 ms. 16:01:42.571 INFO Executor - Finished task 0.0 in stage 763.0 (TID 685). 1944 bytes result sent to driver 16:01:42.571 INFO BlockManagerInfo - Removed broadcast_1431_piece0 on localhost:42397 in memory (size: 12.6 KiB, free: 1918.1 MiB) 16:01:42.571 INFO TaskSetManager - Finished task 0.0 in stage 763.0 (TID 685) in 23 ms on localhost (executor driver) (1/1) 16:01:42.571 INFO TaskSchedulerImpl - Removed TaskSet 763.0, whose tasks have all completed, from pool 16:01:42.572 INFO BlockManagerInfo - Removed broadcast_1401_piece0 on localhost:42397 in memory (size: 64.1 KiB, free: 1918.1 MiB) 16:01:42.572 INFO DAGScheduler - ResultStage 763 (runJob at SparkHadoopWriter.scala:83) finished in 0.032 s 16:01:42.572 INFO DAGScheduler - Job 438 is finished. Cancelling potential speculative or zombie tasks for this job 16:01:42.572 INFO TaskSchedulerImpl - Killing all running tasks in stage 763: Stage finished 16:01:42.572 INFO DAGScheduler - Job 438 finished: runJob at SparkHadoopWriter.scala:83, took 0.040922 s 16:01:42.572 INFO BlockManagerInfo - Removed broadcast_1437_piece0 on localhost:42397 in memory (size: 9.5 KiB, free: 1918.1 MiB) 16:01:42.572 INFO SparkHadoopWriter - Start to commit write Job job_202505191601427956918440701693653_3686. 16:01:42.572 INFO BlockManagerInfo - Removed broadcast_1410_piece0 on localhost:42397 in memory (size: 323.5 KiB, free: 1918.5 MiB) 16:01:42.573 INFO BlockManagerInfo - Removed broadcast_1420_piece0 on localhost:42397 in memory (size: 10.6 KiB, free: 1918.5 MiB) 16:01:42.573 INFO BlockManagerInfo - Removed broadcast_1414_piece0 on localhost:42397 in memory (size: 4.1 KiB, free: 1918.5 MiB) 16:01:42.574 INFO BlockManagerInfo - Removed broadcast_1413_piece0 on localhost:42397 in memory (size: 201.4 KiB, free: 1918.7 MiB) 16:01:42.574 INFO BlockManagerInfo - Removed broadcast_1411_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1918.7 MiB) 16:01:42.575 INFO BlockManagerInfo - Removed broadcast_1424_piece0 on localhost:42397 in memory (size: 73.4 KiB, free: 1918.8 MiB) 16:01:42.575 INFO BlockManagerInfo - Removed broadcast_1404_piece0 on localhost:42397 in memory (size: 20.0 B, free: 1918.8 MiB) 16:01:42.575 INFO BlockManagerInfo - Removed broadcast_1405_piece0 on localhost:42397 in memory (size: 228.2 KiB, free: 1919.0 MiB) 16:01:42.576 INFO BlockManagerInfo - Removed broadcast_1406_piece0 on localhost:42397 in memory (size: 228.2 KiB, free: 1919.3 MiB) 16:01:42.576 INFO BlockManagerInfo - Removed broadcast_1422_piece0 on localhost:42397 in memory (size: 1850.0 B, free: 1919.3 MiB) 16:01:42.576 INFO SparkHadoopWriter - Write Job job_202505191601427956918440701693653_3686 committed. Elapsed time: 4 ms. 16:01:42.576 INFO BlockManagerInfo - Removed broadcast_1425_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1919.3 MiB) 16:01:42.577 INFO BlockManagerInfo - Removed broadcast_1432_piece0 on localhost:42397 in memory (size: 14.8 KiB, free: 1919.3 MiB) 16:01:42.577 INFO BlockManagerInfo - Removed broadcast_1433_piece0 on localhost:42397 in memory (size: 9.7 KiB, free: 1919.3 MiB) 16:01:42.578 INFO BlockManagerInfo - Removed broadcast_1434_piece0 on localhost:42397 in memory (size: 10.6 KiB, free: 1919.4 MiB) 16:01:42.579 INFO BlockManagerInfo - Removed broadcast_1416_piece0 on localhost:42397 in memory (size: 37.0 B, free: 1919.4 MiB) 16:01:42.579 INFO BlockManagerInfo - Removed broadcast_1417_piece0 on localhost:42397 in memory (size: 12.6 KiB, free: 1919.4 MiB) 16:01:42.579 INFO BlockManagerInfo - Removed broadcast_1423_piece0 on localhost:42397 in memory (size: 9.5 KiB, free: 1919.4 MiB) 16:01:42.580 INFO BlockManagerInfo - Removed broadcast_1412_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1919.4 MiB) 16:01:42.580 INFO BlockManagerInfo - Removed broadcast_1427_piece0 on localhost:42397 in memory (size: 201.4 KiB, free: 1919.6 MiB) 16:01:42.581 INFO BlockManagerInfo - Removed broadcast_1403_piece0 on localhost:42397 in memory (size: 19.0 B, free: 1919.6 MiB) 16:01:42.581 INFO BlockManagerInfo - Removed broadcast_1421_piece0 on localhost:42397 in memory (size: 1850.0 B, free: 1919.6 MiB) 16:01:42.582 INFO BlockManagerInfo - Removed broadcast_1415_piece0 on localhost:42397 in memory (size: 145.0 B, free: 1919.6 MiB) 16:01:42.582 INFO BlockManagerInfo - Removed broadcast_1428_piece0 on localhost:42397 in memory (size: 4.1 KiB, free: 1919.6 MiB) 16:01:42.583 INFO BlockManagerInfo - Removed broadcast_1419_piece0 on localhost:42397 in memory (size: 9.7 KiB, free: 1919.6 MiB) 16:01:42.584 INFO BlockManagerInfo - Removed broadcast_1418_piece0 on localhost:42397 in memory (size: 14.8 KiB, free: 1919.7 MiB) 16:01:42.584 INFO HadoopFileSystemWrapper - Concatenating 3 parts to /tmp/local13774000465916765003/markdups4450785990897483788.bam 16:01:42.586 INFO HadoopFileSystemWrapper - Concatenating to /tmp/local13774000465916765003/markdups4450785990897483788.bam done 16:01:42.586 INFO IndexFileMerger - Merging .sbi files in temp directory /tmp/local13774000465916765003/markdups4450785990897483788.bam.parts/ to /tmp/local13774000465916765003/markdups4450785990897483788.bam.sbi 16:01:42.589 INFO IndexFileMerger - Done merging .sbi files 16:01:42.589 INFO IndexFileMerger - Merging .bai files in temp directory /tmp/local13774000465916765003/markdups4450785990897483788.bam.parts/ to /tmp/local13774000465916765003/markdups4450785990897483788.bam.bai 16:01:42.592 INFO IndexFileMerger - Done merging .bai files 16:01:42.592 INFO MarkDuplicatesSpark - Shutting down engine [May 19, 2025 at 4:01:42 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:42 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:43 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:43 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:43 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:44 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:44 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:45 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:47 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:47 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:47 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:48 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:48 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:48 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:50 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:51 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:51 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:51 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:52 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:52 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:52 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:54 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:55 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:55 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:55 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:56 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:56 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:56 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:57 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:59 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:59 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:01:59 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:00 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:00 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:00 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:02 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:03 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.02 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:03 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:04 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.01 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:06 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.03 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:07 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 16:02:07.112 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.112 INFO MarkDuplicatesSpark - The Genome Analysis Toolkit (GATK) va6bd333-SNAPSHOT 16:02:07.112 INFO MarkDuplicatesSpark - For support and documentation go to https://software.broadinstitute.org/gatk/ 16:02:07.112 INFO MarkDuplicatesSpark - Executing as root@67ac59543751 on Linux v6.11.0-1014-azure amd64 16:02:07.112 INFO MarkDuplicatesSpark - Java runtime: OpenJDK 64-Bit Server VM v17.0.12+7-Ubuntu-1ubuntu222.04 16:02:07.112 INFO MarkDuplicatesSpark - Start Date/Time: May 19, 2025 at 4:02:07 PM GMT 16:02:07.112 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.112 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.112 INFO MarkDuplicatesSpark - HTSJDK Version: test_cram_wiring-SNAPSHOT 16:02:07.113 INFO MarkDuplicatesSpark - Picard Version: 3.4.0 16:02:07.113 INFO MarkDuplicatesSpark - Built for Spark Version: 3.5.0 16:02:07.113 INFO MarkDuplicatesSpark - HTSJDK Defaults.COMPRESSION_LEVEL : 2 16:02:07.113 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_READ_FOR_SAMTOOLS : false 16:02:07.113 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_SAMTOOLS : true 16:02:07.113 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_TRIBBLE : false 16:02:07.113 INFO MarkDuplicatesSpark - Deflater: IntelDeflater 16:02:07.113 INFO MarkDuplicatesSpark - Inflater: IntelInflater 16:02:07.113 INFO MarkDuplicatesSpark - GCS max retries/reopens: 20 16:02:07.113 INFO MarkDuplicatesSpark - Requester pays: disabled 16:02:07.113 INFO MarkDuplicatesSpark - Initializing engine 16:02:07.113 INFO MarkDuplicatesSpark - Done initializing engine 16:02:07.113 INFO MarkDuplicatesSpark - Spark verbosity set to INFO (see --spark-verbosity argument) 16:02:07.115 INFO MemoryStore - Block broadcast_2015 stored as values in memory (estimated size 305.5 KiB, free 1917.1 MiB) 16:02:07.122 INFO MemoryStore - Block broadcast_2015_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1917.1 MiB) 16:02:07.122 INFO BlockManagerInfo - Added broadcast_2015_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1919.3 MiB) 16:02:07.123 INFO SparkContext - Created broadcast 2015 from newAPIHadoopFile at PathSplitSource.java:96 16:02:07.143 INFO MemoryStore - Block broadcast_2016 stored as values in memory (estimated size 305.5 KiB, free 1916.8 MiB) 16:02:07.149 INFO MemoryStore - Block broadcast_2016_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1916.7 MiB) 16:02:07.150 INFO BlockManagerInfo - Added broadcast_2016_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1919.2 MiB) 16:02:07.150 INFO SparkContext - Created broadcast 2016 from newAPIHadoopFile at PathSplitSource.java:96 16:02:07.182 INFO SparkContext - Starting job: collect at SparkUtils.java:205 16:02:07.182 INFO FileInputFormat - Total input files to process : 1 16:02:07.183 INFO DAGScheduler - Registering RDD 5578 (mapToPair at SparkUtils.java:161) as input to shuffle 297 16:02:07.183 INFO DAGScheduler - Got job 569 (collect at SparkUtils.java:205) with 1 output partitions 16:02:07.183 INFO DAGScheduler - Final stage: ResultStage 1225 (collect at SparkUtils.java:205) 16:02:07.183 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1224) 16:02:07.183 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1224) 16:02:07.183 INFO DAGScheduler - Submitting ShuffleMapStage 1224 (MapPartitionsRDD[5578] at mapToPair at SparkUtils.java:161), which has no missing parents 16:02:07.200 INFO MemoryStore - Block broadcast_2017 stored as values in memory (estimated size 492.6 KiB, free 1916.2 MiB) 16:02:07.203 INFO MemoryStore - Block broadcast_2017_piece0 stored as bytes in memory (estimated size 208.9 KiB, free 1916.0 MiB) 16:02:07.203 INFO BlockManagerInfo - Added broadcast_2017_piece0 in memory on localhost:42397 (size: 208.9 KiB, free: 1919.0 MiB) 16:02:07.203 INFO SparkContext - Created broadcast 2017 from broadcast at DAGScheduler.scala:1580 16:02:07.203 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1224 (MapPartitionsRDD[5578] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:02:07.204 INFO TaskSchedulerImpl - Adding task set 1224.0 with 1 tasks resource profile 0 16:02:07.204 INFO TaskSetManager - Starting task 0.0 in stage 1224.0 (TID 1056) (localhost, executor driver, partition 0, PROCESS_LOCAL, 9674 bytes) 16:02:07.204 INFO Executor - Running task 0.0 in stage 1224.0 (TID 1056) 16:02:07.233 INFO NewHadoopRDD - Input split: file:/gatkCloneMountPoint/src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/example.chr1.1-1K.unmarkedDups.bam:0+13412 16:02:07.239 INFO Executor - Finished task 0.0 in stage 1224.0 (TID 1056). 1148 bytes result sent to driver 16:02:07.239 INFO TaskSetManager - Finished task 0.0 in stage 1224.0 (TID 1056) in 35 ms on localhost (executor driver) (1/1) 16:02:07.239 INFO TaskSchedulerImpl - Removed TaskSet 1224.0, whose tasks have all completed, from pool 16:02:07.239 INFO DAGScheduler - ShuffleMapStage 1224 (mapToPair at SparkUtils.java:161) finished in 0.056 s 16:02:07.239 INFO DAGScheduler - looking for newly runnable stages 16:02:07.239 INFO DAGScheduler - running: HashSet() 16:02:07.239 INFO DAGScheduler - waiting: HashSet(ResultStage 1225) 16:02:07.239 INFO DAGScheduler - failed: HashSet() 16:02:07.240 INFO DAGScheduler - Submitting ResultStage 1225 (MapPartitionsRDD[5581] at mapPartitions at SparkUtils.java:188), which has no missing parents 16:02:07.240 INFO MemoryStore - Block broadcast_2018 stored as values in memory (estimated size 7.4 KiB, free 1916.0 MiB) 16:02:07.240 INFO MemoryStore - Block broadcast_2018_piece0 stored as bytes in memory (estimated size 4.1 KiB, free 1916.0 MiB) 16:02:07.241 INFO BlockManagerInfo - Added broadcast_2018_piece0 in memory on localhost:42397 (size: 4.1 KiB, free: 1919.0 MiB) 16:02:07.241 INFO SparkContext - Created broadcast 2018 from broadcast at DAGScheduler.scala:1580 16:02:07.241 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1225 (MapPartitionsRDD[5581] at mapPartitions at SparkUtils.java:188) (first 15 tasks are for partitions Vector(0)) 16:02:07.241 INFO TaskSchedulerImpl - Adding task set 1225.0 with 1 tasks resource profile 0 16:02:07.241 INFO TaskSetManager - Starting task 0.0 in stage 1225.0 (TID 1057) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.241 INFO Executor - Running task 0.0 in stage 1225.0 (TID 1057) 16:02:07.242 INFO ShuffleBlockFetcherIterator - Getting 1 (16.3 KiB) non-empty blocks including 1 (16.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.242 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.245 INFO Executor - Finished task 0.0 in stage 1225.0 (TID 1057). 2329 bytes result sent to driver 16:02:07.245 INFO TaskSetManager - Finished task 0.0 in stage 1225.0 (TID 1057) in 4 ms on localhost (executor driver) (1/1) 16:02:07.245 INFO TaskSchedulerImpl - Removed TaskSet 1225.0, whose tasks have all completed, from pool 16:02:07.245 INFO DAGScheduler - ResultStage 1225 (collect at SparkUtils.java:205) finished in 0.005 s 16:02:07.245 INFO DAGScheduler - Job 569 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.245 INFO TaskSchedulerImpl - Killing all running tasks in stage 1225: Stage finished 16:02:07.245 INFO DAGScheduler - Job 569 finished: collect at SparkUtils.java:205, took 0.063396 s 16:02:07.249 INFO MemoryStore - Block broadcast_2019 stored as values in memory (estimated size 7.8 KiB, free 1916.0 MiB) 16:02:07.250 INFO MemoryStore - Block broadcast_2019_piece0 stored as bytes in memory (estimated size 540.0 B, free 1916.0 MiB) 16:02:07.250 INFO BlockManagerInfo - Added broadcast_2019_piece0 in memory on localhost:42397 (size: 540.0 B, free: 1919.0 MiB) 16:02:07.250 INFO SparkContext - Created broadcast 2019 from broadcast at MarkDuplicatesSparkUtils.java:126 16:02:07.250 INFO MemoryStore - Block broadcast_2020 stored as values in memory (estimated size 2.6 KiB, free 1916.0 MiB) 16:02:07.251 INFO MemoryStore - Block broadcast_2020_piece0 stored as bytes in memory (estimated size 209.0 B, free 1916.0 MiB) 16:02:07.251 INFO BlockManagerInfo - Added broadcast_2020_piece0 in memory on localhost:42397 (size: 209.0 B, free: 1919.0 MiB) 16:02:07.251 INFO SparkContext - Created broadcast 2020 from broadcast at MarkDuplicatesSparkUtils.java:127 16:02:07.265 INFO SparkContext - Starting job: collectAsMap at MarkDuplicatesSparkUtils.java:542 16:02:07.265 INFO DAGScheduler - Registering RDD 5590 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) as input to shuffle 300 16:02:07.265 INFO DAGScheduler - Registering RDD 5594 (mapToPair at MarkDuplicatesSpark.java:217) as input to shuffle 299 16:02:07.266 INFO DAGScheduler - Registering RDD 5598 (mapToPair at MarkDuplicatesSparkUtils.java:497) as input to shuffle 298 16:02:07.266 INFO DAGScheduler - Got job 570 (collectAsMap at MarkDuplicatesSparkUtils.java:542) with 1 output partitions 16:02:07.266 INFO DAGScheduler - Final stage: ResultStage 1230 (collectAsMap at MarkDuplicatesSparkUtils.java:542) 16:02:07.266 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1229) 16:02:07.266 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1229) 16:02:07.266 INFO DAGScheduler - Submitting ShuffleMapStage 1227 (MapPartitionsRDD[5590] at flatMapToPair at MarkDuplicatesSparkUtils.java:130), which has no missing parents 16:02:07.267 INFO MemoryStore - Block broadcast_2021 stored as values in memory (estimated size 57.6 KiB, free 1916.0 MiB) 16:02:07.267 INFO MemoryStore - Block broadcast_2021_piece0 stored as bytes in memory (estimated size 19.8 KiB, free 1915.9 MiB) 16:02:07.268 INFO BlockManagerInfo - Added broadcast_2021_piece0 in memory on localhost:42397 (size: 19.8 KiB, free: 1919.0 MiB) 16:02:07.268 INFO SparkContext - Created broadcast 2021 from broadcast at DAGScheduler.scala:1580 16:02:07.268 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1227 (MapPartitionsRDD[5590] at flatMapToPair at MarkDuplicatesSparkUtils.java:130) (first 15 tasks are for partitions Vector(0)) 16:02:07.268 INFO TaskSchedulerImpl - Adding task set 1227.0 with 1 tasks resource profile 0 16:02:07.268 INFO TaskSetManager - Starting task 0.0 in stage 1227.0 (TID 1058) (localhost, executor driver, partition 0, NODE_LOCAL, 9813 bytes) 16:02:07.268 INFO Executor - Running task 0.0 in stage 1227.0 (TID 1058) 16:02:07.270 INFO ShuffleBlockFetcherIterator - Getting 1 (16.3 KiB) non-empty blocks including 1 (16.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.270 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.274 INFO Executor - Finished task 0.0 in stage 1227.0 (TID 1058). 1922 bytes result sent to driver 16:02:07.274 INFO TaskSetManager - Finished task 0.0 in stage 1227.0 (TID 1058) in 6 ms on localhost (executor driver) (1/1) 16:02:07.274 INFO TaskSchedulerImpl - Removed TaskSet 1227.0, whose tasks have all completed, from pool 16:02:07.274 INFO DAGScheduler - ShuffleMapStage 1227 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) finished in 0.008 s 16:02:07.274 INFO DAGScheduler - looking for newly runnable stages 16:02:07.274 INFO DAGScheduler - running: HashSet() 16:02:07.274 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 1228, ShuffleMapStage 1229, ResultStage 1230) 16:02:07.274 INFO DAGScheduler - failed: HashSet() 16:02:07.275 INFO DAGScheduler - Submitting ShuffleMapStage 1228 (MapPartitionsRDD[5594] at mapToPair at MarkDuplicatesSpark.java:217), which has no missing parents 16:02:07.276 INFO MemoryStore - Block broadcast_2022 stored as values in memory (estimated size 61.7 KiB, free 1915.9 MiB) 16:02:07.276 INFO MemoryStore - Block broadcast_2022_piece0 stored as bytes in memory (estimated size 22.0 KiB, free 1915.9 MiB) 16:02:07.276 INFO BlockManagerInfo - Added broadcast_2022_piece0 in memory on localhost:42397 (size: 22.0 KiB, free: 1919.0 MiB) 16:02:07.276 INFO SparkContext - Created broadcast 2022 from broadcast at DAGScheduler.scala:1580 16:02:07.277 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1228 (MapPartitionsRDD[5594] at mapToPair at MarkDuplicatesSpark.java:217) (first 15 tasks are for partitions Vector(0)) 16:02:07.277 INFO TaskSchedulerImpl - Adding task set 1228.0 with 1 tasks resource profile 0 16:02:07.277 INFO TaskSetManager - Starting task 0.0 in stage 1228.0 (TID 1059) (localhost, executor driver, partition 0, NODE_LOCAL, 9288 bytes) 16:02:07.277 INFO Executor - Running task 0.0 in stage 1228.0 (TID 1059) 16:02:07.279 INFO ShuffleBlockFetcherIterator - Getting 1 (2.0 KiB) non-empty blocks including 1 (2.0 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.279 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.283 INFO Executor - Finished task 0.0 in stage 1228.0 (TID 1059). 1922 bytes result sent to driver 16:02:07.284 INFO TaskSetManager - Finished task 0.0 in stage 1228.0 (TID 1059) in 7 ms on localhost (executor driver) (1/1) 16:02:07.284 INFO TaskSchedulerImpl - Removed TaskSet 1228.0, whose tasks have all completed, from pool 16:02:07.284 INFO DAGScheduler - ShuffleMapStage 1228 (mapToPair at MarkDuplicatesSpark.java:217) finished in 0.009 s 16:02:07.284 INFO DAGScheduler - looking for newly runnable stages 16:02:07.284 INFO DAGScheduler - running: HashSet() 16:02:07.284 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 1229, ResultStage 1230) 16:02:07.284 INFO DAGScheduler - failed: HashSet() 16:02:07.284 INFO DAGScheduler - Submitting ShuffleMapStage 1229 (MapPartitionsRDD[5598] at mapToPair at MarkDuplicatesSparkUtils.java:497), which has no missing parents 16:02:07.285 INFO MemoryStore - Block broadcast_2023 stored as values in memory (estimated size 52.4 KiB, free 1915.8 MiB) 16:02:07.285 INFO MemoryStore - Block broadcast_2023_piece0 stored as bytes in memory (estimated size 17.3 KiB, free 1915.8 MiB) 16:02:07.286 INFO BlockManagerInfo - Added broadcast_2023_piece0 in memory on localhost:42397 (size: 17.3 KiB, free: 1919.0 MiB) 16:02:07.286 INFO SparkContext - Created broadcast 2023 from broadcast at DAGScheduler.scala:1580 16:02:07.286 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1229 (MapPartitionsRDD[5598] at mapToPair at MarkDuplicatesSparkUtils.java:497) (first 15 tasks are for partitions Vector(0)) 16:02:07.286 INFO TaskSchedulerImpl - Adding task set 1229.0 with 1 tasks resource profile 0 16:02:07.286 INFO TaskSetManager - Starting task 0.0 in stage 1229.0 (TID 1060) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:02:07.286 INFO Executor - Running task 0.0 in stage 1229.0 (TID 1060) 16:02:07.288 INFO ShuffleBlockFetcherIterator - Getting 1 (16.3 KiB) non-empty blocks including 1 (16.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.288 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.290 INFO ShuffleBlockFetcherIterator - Getting 1 (2.7 KiB) non-empty blocks including 1 (2.7 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.290 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.292 INFO Executor - Finished task 0.0 in stage 1229.0 (TID 1060). 1922 bytes result sent to driver 16:02:07.292 INFO TaskSetManager - Finished task 0.0 in stage 1229.0 (TID 1060) in 6 ms on localhost (executor driver) (1/1) 16:02:07.292 INFO TaskSchedulerImpl - Removed TaskSet 1229.0, whose tasks have all completed, from pool 16:02:07.292 INFO DAGScheduler - ShuffleMapStage 1229 (mapToPair at MarkDuplicatesSparkUtils.java:497) finished in 0.008 s 16:02:07.292 INFO DAGScheduler - looking for newly runnable stages 16:02:07.292 INFO DAGScheduler - running: HashSet() 16:02:07.292 INFO DAGScheduler - waiting: HashSet(ResultStage 1230) 16:02:07.292 INFO DAGScheduler - failed: HashSet() 16:02:07.292 INFO DAGScheduler - Submitting ResultStage 1230 (MapPartitionsRDD[5600] at mapValues at MarkDuplicatesSparkUtils.java:519), which has no missing parents 16:02:07.293 INFO MemoryStore - Block broadcast_2024 stored as values in memory (estimated size 53.6 KiB, free 1915.7 MiB) 16:02:07.294 INFO MemoryStore - Block broadcast_2024_piece0 stored as bytes in memory (estimated size 18.0 KiB, free 1915.7 MiB) 16:02:07.294 INFO BlockManagerInfo - Added broadcast_2024_piece0 in memory on localhost:42397 (size: 18.0 KiB, free: 1918.9 MiB) 16:02:07.294 INFO SparkContext - Created broadcast 2024 from broadcast at DAGScheduler.scala:1580 16:02:07.294 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1230 (MapPartitionsRDD[5600] at mapValues at MarkDuplicatesSparkUtils.java:519) (first 15 tasks are for partitions Vector(0)) 16:02:07.294 INFO TaskSchedulerImpl - Adding task set 1230.0 with 1 tasks resource profile 0 16:02:07.295 INFO TaskSetManager - Starting task 0.0 in stage 1230.0 (TID 1061) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.295 INFO Executor - Running task 0.0 in stage 1230.0 (TID 1061) 16:02:07.296 INFO ShuffleBlockFetcherIterator - Getting 1 (251.0 B) non-empty blocks including 1 (251.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.296 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.297 INFO Executor - Finished task 0.0 in stage 1230.0 (TID 1061). 2371 bytes result sent to driver 16:02:07.298 INFO TaskSetManager - Finished task 0.0 in stage 1230.0 (TID 1061) in 4 ms on localhost (executor driver) (1/1) 16:02:07.298 INFO TaskSchedulerImpl - Removed TaskSet 1230.0, whose tasks have all completed, from pool 16:02:07.298 INFO DAGScheduler - ResultStage 1230 (collectAsMap at MarkDuplicatesSparkUtils.java:542) finished in 0.005 s 16:02:07.298 INFO DAGScheduler - Job 570 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.298 INFO TaskSchedulerImpl - Killing all running tasks in stage 1230: Stage finished 16:02:07.298 INFO DAGScheduler - Job 570 finished: collectAsMap at MarkDuplicatesSparkUtils.java:542, took 0.032986 s 16:02:07.311 INFO MemoryStore - Block broadcast_2025 stored as values in memory (estimated size 179.4 KiB, free 1915.5 MiB) 16:02:07.311 INFO MemoryStore - Block broadcast_2025_piece0 stored as bytes in memory (estimated size 7.9 KiB, free 1915.5 MiB) 16:02:07.311 INFO BlockManagerInfo - Added broadcast_2025_piece0 in memory on localhost:42397 (size: 7.9 KiB, free: 1918.9 MiB) 16:02:07.312 INFO SparkContext - Created broadcast 2025 from broadcast at ReadsSparkSink.java:133 16:02:07.315 INFO HadoopMapRedCommitProtocol - Using output committer class org.apache.hadoop.mapred.FileOutputCommitter 16:02:07.315 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:02:07.315 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:02:07.326 INFO SparkContext - Starting job: runJob at SparkHadoopWriter.scala:83 16:02:07.326 INFO DAGScheduler - Registering RDD 5602 (mapToPair at SparkUtils.java:161) as input to shuffle 301 16:02:07.326 INFO DAGScheduler - Got job 571 (runJob at SparkHadoopWriter.scala:83) with 1 output partitions 16:02:07.327 INFO DAGScheduler - Final stage: ResultStage 1235 (runJob at SparkHadoopWriter.scala:83) 16:02:07.327 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1234) 16:02:07.327 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1234) 16:02:07.327 INFO DAGScheduler - Submitting ShuffleMapStage 1234 (MapPartitionsRDD[5602] at mapToPair at SparkUtils.java:161), which has no missing parents 16:02:07.328 INFO MemoryStore - Block broadcast_2026 stored as values in memory (estimated size 51.5 KiB, free 1915.5 MiB) 16:02:07.328 INFO MemoryStore - Block broadcast_2026_piece0 stored as bytes in memory (estimated size 16.8 KiB, free 1915.5 MiB) 16:02:07.328 INFO BlockManagerInfo - Added broadcast_2026_piece0 in memory on localhost:42397 (size: 16.8 KiB, free: 1918.9 MiB) 16:02:07.328 INFO SparkContext - Created broadcast 2026 from broadcast at DAGScheduler.scala:1580 16:02:07.329 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1234 (MapPartitionsRDD[5602] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:02:07.329 INFO TaskSchedulerImpl - Adding task set 1234.0 with 1 tasks resource profile 0 16:02:07.329 INFO TaskSetManager - Starting task 0.0 in stage 1234.0 (TID 1062) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:02:07.329 INFO Executor - Running task 0.0 in stage 1234.0 (TID 1062) 16:02:07.331 INFO ShuffleBlockFetcherIterator - Getting 1 (16.3 KiB) non-empty blocks including 1 (16.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.331 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.334 INFO ShuffleBlockFetcherIterator - Getting 1 (2.7 KiB) non-empty blocks including 1 (2.7 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.334 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.336 INFO Executor - Finished task 0.0 in stage 1234.0 (TID 1062). 1922 bytes result sent to driver 16:02:07.336 INFO TaskSetManager - Finished task 0.0 in stage 1234.0 (TID 1062) in 7 ms on localhost (executor driver) (1/1) 16:02:07.336 INFO TaskSchedulerImpl - Removed TaskSet 1234.0, whose tasks have all completed, from pool 16:02:07.336 INFO DAGScheduler - ShuffleMapStage 1234 (mapToPair at SparkUtils.java:161) finished in 0.009 s 16:02:07.336 INFO DAGScheduler - looking for newly runnable stages 16:02:07.336 INFO DAGScheduler - running: HashSet() 16:02:07.337 INFO DAGScheduler - waiting: HashSet(ResultStage 1235) 16:02:07.337 INFO DAGScheduler - failed: HashSet() 16:02:07.337 INFO DAGScheduler - Submitting ResultStage 1235 (MapPartitionsRDD[5608] at saveAsTextFile at SamSink.java:65), which has no missing parents 16:02:07.344 INFO MemoryStore - Block broadcast_2027 stored as values in memory (estimated size 196.3 KiB, free 1915.3 MiB) 16:02:07.345 INFO MemoryStore - Block broadcast_2027_piece0 stored as bytes in memory (estimated size 80.5 KiB, free 1915.2 MiB) 16:02:07.345 INFO BlockManagerInfo - Added broadcast_2027_piece0 in memory on localhost:42397 (size: 80.5 KiB, free: 1918.8 MiB) 16:02:07.345 INFO SparkContext - Created broadcast 2027 from broadcast at DAGScheduler.scala:1580 16:02:07.345 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1235 (MapPartitionsRDD[5608] at saveAsTextFile at SamSink.java:65) (first 15 tasks are for partitions Vector(0)) 16:02:07.345 INFO TaskSchedulerImpl - Adding task set 1235.0 with 1 tasks resource profile 0 16:02:07.346 INFO TaskSetManager - Starting task 0.0 in stage 1235.0 (TID 1063) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.346 INFO Executor - Running task 0.0 in stage 1235.0 (TID 1063) 16:02:07.349 INFO ShuffleBlockFetcherIterator - Getting 1 (14.8 KiB) non-empty blocks including 1 (14.8 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.349 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.352 INFO HadoopMapRedCommitProtocol - Using output committer class org.apache.hadoop.mapred.FileOutputCommitter 16:02:07.352 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:02:07.352 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:02:07.358 INFO FileOutputCommitter - Saved output of task 'attempt_202505191602073571397478377297836_5608_m_000000_0' to file:/tmp/MarkDups2704444477632446738/MarkDups.sam.parts/_temporary/0/task_202505191602073571397478377297836_5608_m_000000 16:02:07.358 INFO SparkHadoopMapRedUtil - attempt_202505191602073571397478377297836_5608_m_000000_0: Committed. Elapsed time: 0 ms. 16:02:07.358 INFO Executor - Finished task 0.0 in stage 1235.0 (TID 1063). 1858 bytes result sent to driver 16:02:07.358 INFO TaskSetManager - Finished task 0.0 in stage 1235.0 (TID 1063) in 13 ms on localhost (executor driver) (1/1) 16:02:07.358 INFO TaskSchedulerImpl - Removed TaskSet 1235.0, whose tasks have all completed, from pool 16:02:07.358 INFO DAGScheduler - ResultStage 1235 (runJob at SparkHadoopWriter.scala:83) finished in 0.021 s 16:02:07.359 INFO DAGScheduler - Job 571 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.359 INFO TaskSchedulerImpl - Killing all running tasks in stage 1235: Stage finished 16:02:07.359 INFO DAGScheduler - Job 571 finished: runJob at SparkHadoopWriter.scala:83, took 0.032563 s 16:02:07.359 INFO SparkHadoopWriter - Start to commit write Job job_202505191602073571397478377297836_5608. 16:02:07.361 INFO SparkHadoopWriter - Write Job job_202505191602073571397478377297836_5608 committed. Elapsed time: 2 ms. 16:02:07.365 INFO HadoopFileSystemWrapper - Concatenating 2 parts to /tmp/MarkDups2704444477632446738/MarkDups.sam 16:02:07.367 INFO HadoopFileSystemWrapper - Concatenating to /tmp/MarkDups2704444477632446738/MarkDups.sam done 16:02:07.368 INFO MarkDuplicatesSpark - Shutting down engine [May 19, 2025 at 4:02:07 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 16:02:07.383 INFO BlockManagerInfo - Removed broadcast_2013_piece0 on localhost:42397 in memory (size: 8.0 KiB, free: 1918.9 MiB) 16:02:07.383 INFO BlockManagerInfo - Removed broadcast_2021_piece0 on localhost:42397 in memory (size: 19.8 KiB, free: 1918.9 MiB) 16:02:07.383 INFO BlockManagerInfo - Removed broadcast_2025_piece0 on localhost:42397 in memory (size: 7.9 KiB, free: 1918.9 MiB) 16:02:07.384 INFO BlockManagerInfo - Removed broadcast_2016_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1918.9 MiB) 16:02:07.384 INFO BlockManagerInfo - Removed broadcast_2002_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1919.0 MiB) 16:02:07.384 INFO BlockManagerInfo - Removed broadcast_2024_piece0 on localhost:42397 in memory (size: 18.0 KiB, free: 1919.0 MiB) 16:02:07.385 INFO BlockManagerInfo - Removed broadcast_2014_piece0 on localhost:42397 in memory (size: 71.6 KiB, free: 1919.1 MiB) 16:02:07.385 INFO BlockManagerInfo - Removed broadcast_2022_piece0 on localhost:42397 in memory (size: 22.0 KiB, free: 1919.1 MiB) 16:02:07.386 INFO BlockManagerInfo - Removed broadcast_2019_piece0 on localhost:42397 in memory (size: 540.0 B, free: 1919.1 MiB) 16:02:07.386 INFO BlockManagerInfo - Removed broadcast_2027_piece0 on localhost:42397 in memory (size: 80.5 KiB, free: 1919.2 MiB) 16:02:07.387 INFO BlockManagerInfo - Removed broadcast_2006_piece0 on localhost:42397 in memory (size: 19.0 B, free: 1919.2 MiB) 16:02:07.388 INFO BlockManagerInfo - Removed broadcast_2018_piece0 on localhost:42397 in memory (size: 4.1 KiB, free: 1919.2 MiB) 16:02:07.388 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.388 INFO BlockManagerInfo - Removed broadcast_2023_piece0 on localhost:42397 in memory (size: 17.3 KiB, free: 1919.2 MiB) 16:02:07.388 INFO MarkDuplicatesSpark - The Genome Analysis Toolkit (GATK) va6bd333-SNAPSHOT 16:02:07.388 INFO MarkDuplicatesSpark - For support and documentation go to https://software.broadinstitute.org/gatk/ 16:02:07.388 INFO MarkDuplicatesSpark - Executing as root@67ac59543751 on Linux v6.11.0-1014-azure amd64 16:02:07.388 INFO MarkDuplicatesSpark - Java runtime: OpenJDK 64-Bit Server VM v17.0.12+7-Ubuntu-1ubuntu222.04 16:02:07.388 INFO MarkDuplicatesSpark - Start Date/Time: May 19, 2025 at 4:02:07 PM GMT 16:02:07.388 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.388 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.388 INFO MarkDuplicatesSpark - HTSJDK Version: test_cram_wiring-SNAPSHOT 16:02:07.388 INFO MarkDuplicatesSpark - Picard Version: 3.4.0 16:02:07.388 INFO MarkDuplicatesSpark - Built for Spark Version: 3.5.0 16:02:07.388 INFO MarkDuplicatesSpark - HTSJDK Defaults.COMPRESSION_LEVEL : 2 16:02:07.388 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_READ_FOR_SAMTOOLS : false 16:02:07.388 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_SAMTOOLS : true 16:02:07.388 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_TRIBBLE : false 16:02:07.388 INFO BlockManagerInfo - Removed broadcast_2015_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1919.3 MiB) 16:02:07.388 INFO MarkDuplicatesSpark - Deflater: IntelDeflater 16:02:07.388 INFO MarkDuplicatesSpark - Inflater: IntelInflater 16:02:07.388 INFO MarkDuplicatesSpark - GCS max retries/reopens: 20 16:02:07.389 INFO MarkDuplicatesSpark - Requester pays: disabled 16:02:07.389 INFO MarkDuplicatesSpark - Initializing engine 16:02:07.389 INFO MarkDuplicatesSpark - Done initializing engine 16:02:07.389 INFO MarkDuplicatesSpark - Spark verbosity set to INFO (see --spark-verbosity argument) 16:02:07.389 INFO BlockManagerInfo - Removed broadcast_2012_piece0 on localhost:42397 in memory (size: 663.0 B, free: 1919.3 MiB) 16:02:07.389 INFO BlockManagerInfo - Removed broadcast_2011_piece0 on localhost:42397 in memory (size: 9.1 KiB, free: 1919.3 MiB) 16:02:07.390 INFO BlockManagerInfo - Removed broadcast_2009_piece0 on localhost:42397 in memory (size: 13.2 KiB, free: 1919.3 MiB) 16:02:07.390 INFO BlockManagerInfo - Removed broadcast_2003_piece0 on localhost:42397 in memory (size: 64.0 KiB, free: 1919.4 MiB) 16:02:07.390 INFO BlockManagerInfo - Removed broadcast_2026_piece0 on localhost:42397 in memory (size: 16.8 KiB, free: 1919.4 MiB) 16:02:07.391 INFO BlockManagerInfo - Removed broadcast_2007_piece0 on localhost:42397 in memory (size: 31.0 B, free: 1919.4 MiB) 16:02:07.391 INFO MemoryStore - Block broadcast_2028 stored as values in memory (estimated size 305.5 KiB, free 1917.5 MiB) 16:02:07.391 INFO BlockManagerInfo - Removed broadcast_2008_piece0 on localhost:42397 in memory (size: 11.2 KiB, free: 1919.4 MiB) 16:02:07.391 INFO BlockManagerInfo - Removed broadcast_2010_piece0 on localhost:42397 in memory (size: 8.5 KiB, free: 1919.4 MiB) 16:02:07.392 INFO BlockManagerInfo - Removed broadcast_2017_piece0 on localhost:42397 in memory (size: 208.9 KiB, free: 1919.6 MiB) 16:02:07.393 INFO BlockManagerInfo - Removed broadcast_2004_piece0 on localhost:42397 in memory (size: 200.3 KiB, free: 1919.8 MiB) 16:02:07.393 INFO BlockManagerInfo - Removed broadcast_2020_piece0 on localhost:42397 in memory (size: 209.0 B, free: 1919.8 MiB) 16:02:07.394 INFO BlockManagerInfo - Removed broadcast_2005_piece0 on localhost:42397 in memory (size: 4.1 KiB, free: 1919.8 MiB) 16:02:07.399 INFO MemoryStore - Block broadcast_2028_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1918.8 MiB) 16:02:07.399 INFO BlockManagerInfo - Added broadcast_2028_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1919.7 MiB) 16:02:07.399 INFO SparkContext - Created broadcast 2028 from newAPIHadoopFile at PathSplitSource.java:96 16:02:07.419 INFO MemoryStore - Block broadcast_2029 stored as values in memory (estimated size 305.5 KiB, free 1918.5 MiB) 16:02:07.426 INFO MemoryStore - Block broadcast_2029_piece0 stored as bytes in memory (estimated size 64.0 KiB, free 1918.5 MiB) 16:02:07.426 INFO BlockManagerInfo - Added broadcast_2029_piece0 in memory on localhost:42397 (size: 64.0 KiB, free: 1919.7 MiB) 16:02:07.426 INFO SparkContext - Created broadcast 2029 from newAPIHadoopFile at PathSplitSource.java:96 16:02:07.458 INFO SparkContext - Starting job: collect at SparkUtils.java:205 16:02:07.459 INFO FileInputFormat - Total input files to process : 1 16:02:07.459 INFO DAGScheduler - Registering RDD 5622 (mapToPair at SparkUtils.java:161) as input to shuffle 302 16:02:07.459 INFO DAGScheduler - Got job 572 (collect at SparkUtils.java:205) with 1 output partitions 16:02:07.459 INFO DAGScheduler - Final stage: ResultStage 1237 (collect at SparkUtils.java:205) 16:02:07.459 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1236) 16:02:07.459 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1236) 16:02:07.459 INFO DAGScheduler - Submitting ShuffleMapStage 1236 (MapPartitionsRDD[5622] at mapToPair at SparkUtils.java:161), which has no missing parents 16:02:07.476 INFO MemoryStore - Block broadcast_2030 stored as values in memory (estimated size 456.1 KiB, free 1918.0 MiB) 16:02:07.478 INFO MemoryStore - Block broadcast_2030_piece0 stored as bytes in memory (estimated size 200.0 KiB, free 1917.8 MiB) 16:02:07.479 INFO BlockManagerInfo - Added broadcast_2030_piece0 in memory on localhost:42397 (size: 200.0 KiB, free: 1919.5 MiB) 16:02:07.479 INFO SparkContext - Created broadcast 2030 from broadcast at DAGScheduler.scala:1580 16:02:07.479 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1236 (MapPartitionsRDD[5622] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:02:07.479 INFO TaskSchedulerImpl - Adding task set 1236.0 with 1 tasks resource profile 0 16:02:07.479 INFO TaskSetManager - Starting task 0.0 in stage 1236.0 (TID 1064) (localhost, executor driver, partition 0, PROCESS_LOCAL, 9653 bytes) 16:02:07.480 INFO Executor - Running task 0.0 in stage 1236.0 (TID 1064) 16:02:07.508 INFO NewHadoopRDD - Input split: file:/gatkCloneMountPoint/src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/optical_dupes.bam:0+678 16:02:07.512 INFO Executor - Finished task 0.0 in stage 1236.0 (TID 1064). 1148 bytes result sent to driver 16:02:07.512 INFO TaskSetManager - Finished task 0.0 in stage 1236.0 (TID 1064) in 33 ms on localhost (executor driver) (1/1) 16:02:07.512 INFO TaskSchedulerImpl - Removed TaskSet 1236.0, whose tasks have all completed, from pool 16:02:07.512 INFO DAGScheduler - ShuffleMapStage 1236 (mapToPair at SparkUtils.java:161) finished in 0.053 s 16:02:07.512 INFO DAGScheduler - looking for newly runnable stages 16:02:07.512 INFO DAGScheduler - running: HashSet() 16:02:07.512 INFO DAGScheduler - waiting: HashSet(ResultStage 1237) 16:02:07.512 INFO DAGScheduler - failed: HashSet() 16:02:07.513 INFO DAGScheduler - Submitting ResultStage 1237 (MapPartitionsRDD[5625] at mapPartitions at SparkUtils.java:188), which has no missing parents 16:02:07.513 INFO MemoryStore - Block broadcast_2031 stored as values in memory (estimated size 7.4 KiB, free 1917.8 MiB) 16:02:07.513 INFO MemoryStore - Block broadcast_2031_piece0 stored as bytes in memory (estimated size 4.1 KiB, free 1917.8 MiB) 16:02:07.513 INFO BlockManagerInfo - Added broadcast_2031_piece0 in memory on localhost:42397 (size: 4.1 KiB, free: 1919.5 MiB) 16:02:07.514 INFO SparkContext - Created broadcast 2031 from broadcast at DAGScheduler.scala:1580 16:02:07.514 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1237 (MapPartitionsRDD[5625] at mapPartitions at SparkUtils.java:188) (first 15 tasks are for partitions Vector(0)) 16:02:07.514 INFO TaskSchedulerImpl - Adding task set 1237.0 with 1 tasks resource profile 0 16:02:07.514 INFO TaskSetManager - Starting task 0.0 in stage 1237.0 (TID 1065) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.514 INFO Executor - Running task 0.0 in stage 1237.0 (TID 1065) 16:02:07.515 INFO ShuffleBlockFetcherIterator - Getting 1 (490.0 B) non-empty blocks including 1 (490.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.515 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.516 INFO Executor - Finished task 0.0 in stage 1237.0 (TID 1065). 2269 bytes result sent to driver 16:02:07.517 INFO TaskSetManager - Finished task 0.0 in stage 1237.0 (TID 1065) in 3 ms on localhost (executor driver) (1/1) 16:02:07.517 INFO TaskSchedulerImpl - Removed TaskSet 1237.0, whose tasks have all completed, from pool 16:02:07.517 INFO DAGScheduler - ResultStage 1237 (collect at SparkUtils.java:205) finished in 0.004 s 16:02:07.517 INFO DAGScheduler - Job 572 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.517 INFO TaskSchedulerImpl - Killing all running tasks in stage 1237: Stage finished 16:02:07.517 INFO DAGScheduler - Job 572 finished: collect at SparkUtils.java:205, took 0.058656 s 16:02:07.521 INFO MemoryStore - Block broadcast_2032 stored as values in memory (estimated size 136.0 B, free 1917.8 MiB) 16:02:07.521 INFO MemoryStore - Block broadcast_2032_piece0 stored as bytes in memory (estimated size 24.0 B, free 1917.8 MiB) 16:02:07.521 INFO BlockManagerInfo - Added broadcast_2032_piece0 in memory on localhost:42397 (size: 24.0 B, free: 1919.5 MiB) 16:02:07.522 INFO SparkContext - Created broadcast 2032 from broadcast at MarkDuplicatesSparkUtils.java:126 16:02:07.522 INFO MemoryStore - Block broadcast_2033 stored as values in memory (estimated size 136.0 B, free 1917.8 MiB) 16:02:07.522 INFO MemoryStore - Block broadcast_2033_piece0 stored as bytes in memory (estimated size 21.0 B, free 1917.8 MiB) 16:02:07.522 INFO BlockManagerInfo - Added broadcast_2033_piece0 in memory on localhost:42397 (size: 21.0 B, free: 1919.5 MiB) 16:02:07.523 INFO SparkContext - Created broadcast 2033 from broadcast at MarkDuplicatesSparkUtils.java:127 16:02:07.536 INFO SparkContext - Starting job: collectAsMap at MarkDuplicatesSparkUtils.java:542 16:02:07.536 INFO DAGScheduler - Registering RDD 5634 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) as input to shuffle 305 16:02:07.536 INFO DAGScheduler - Registering RDD 5638 (mapToPair at MarkDuplicatesSpark.java:217) as input to shuffle 304 16:02:07.537 INFO DAGScheduler - Registering RDD 5642 (mapToPair at MarkDuplicatesSparkUtils.java:497) as input to shuffle 303 16:02:07.537 INFO DAGScheduler - Got job 573 (collectAsMap at MarkDuplicatesSparkUtils.java:542) with 1 output partitions 16:02:07.537 INFO DAGScheduler - Final stage: ResultStage 1242 (collectAsMap at MarkDuplicatesSparkUtils.java:542) 16:02:07.537 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1241) 16:02:07.537 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1241) 16:02:07.537 INFO DAGScheduler - Submitting ShuffleMapStage 1239 (MapPartitionsRDD[5634] at flatMapToPair at MarkDuplicatesSparkUtils.java:130), which has no missing parents 16:02:07.537 INFO MemoryStore - Block broadcast_2034 stored as values in memory (estimated size 21.1 KiB, free 1917.8 MiB) 16:02:07.538 INFO MemoryStore - Block broadcast_2034_piece0 stored as bytes in memory (estimated size 10.9 KiB, free 1917.8 MiB) 16:02:07.538 INFO BlockManagerInfo - Added broadcast_2034_piece0 in memory on localhost:42397 (size: 10.9 KiB, free: 1919.5 MiB) 16:02:07.538 INFO SparkContext - Created broadcast 2034 from broadcast at DAGScheduler.scala:1580 16:02:07.538 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1239 (MapPartitionsRDD[5634] at flatMapToPair at MarkDuplicatesSparkUtils.java:130) (first 15 tasks are for partitions Vector(0)) 16:02:07.538 INFO TaskSchedulerImpl - Adding task set 1239.0 with 1 tasks resource profile 0 16:02:07.539 INFO TaskSetManager - Starting task 0.0 in stage 1239.0 (TID 1066) (localhost, executor driver, partition 0, NODE_LOCAL, 9813 bytes) 16:02:07.539 INFO Executor - Running task 0.0 in stage 1239.0 (TID 1066) 16:02:07.540 INFO ShuffleBlockFetcherIterator - Getting 1 (490.0 B) non-empty blocks including 1 (490.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.540 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.542 INFO Executor - Finished task 0.0 in stage 1239.0 (TID 1066). 1879 bytes result sent to driver 16:02:07.543 INFO TaskSetManager - Finished task 0.0 in stage 1239.0 (TID 1066) in 4 ms on localhost (executor driver) (1/1) 16:02:07.543 INFO TaskSchedulerImpl - Removed TaskSet 1239.0, whose tasks have all completed, from pool 16:02:07.543 INFO DAGScheduler - ShuffleMapStage 1239 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) finished in 0.006 s 16:02:07.543 INFO DAGScheduler - looking for newly runnable stages 16:02:07.543 INFO DAGScheduler - running: HashSet() 16:02:07.543 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 1240, ShuffleMapStage 1241, ResultStage 1242) 16:02:07.543 INFO DAGScheduler - failed: HashSet() 16:02:07.543 INFO DAGScheduler - Submitting ShuffleMapStage 1240 (MapPartitionsRDD[5638] at mapToPair at MarkDuplicatesSpark.java:217), which has no missing parents 16:02:07.543 INFO MemoryStore - Block broadcast_2035 stored as values in memory (estimated size 25.2 KiB, free 1917.8 MiB) 16:02:07.544 INFO MemoryStore - Block broadcast_2035_piece0 stored as bytes in memory (estimated size 13.1 KiB, free 1917.8 MiB) 16:02:07.544 INFO BlockManagerInfo - Added broadcast_2035_piece0 in memory on localhost:42397 (size: 13.1 KiB, free: 1919.5 MiB) 16:02:07.544 INFO SparkContext - Created broadcast 2035 from broadcast at DAGScheduler.scala:1580 16:02:07.544 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1240 (MapPartitionsRDD[5638] at mapToPair at MarkDuplicatesSpark.java:217) (first 15 tasks are for partitions Vector(0)) 16:02:07.544 INFO TaskSchedulerImpl - Adding task set 1240.0 with 1 tasks resource profile 0 16:02:07.545 INFO TaskSetManager - Starting task 0.0 in stage 1240.0 (TID 1067) (localhost, executor driver, partition 0, NODE_LOCAL, 9288 bytes) 16:02:07.545 INFO Executor - Running task 0.0 in stage 1240.0 (TID 1067) 16:02:07.546 INFO ShuffleBlockFetcherIterator - Getting 1 (106.0 B) non-empty blocks including 1 (106.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.546 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.549 INFO Executor - Finished task 0.0 in stage 1240.0 (TID 1067). 1879 bytes result sent to driver 16:02:07.549 INFO TaskSetManager - Finished task 0.0 in stage 1240.0 (TID 1067) in 4 ms on localhost (executor driver) (1/1) 16:02:07.549 INFO TaskSchedulerImpl - Removed TaskSet 1240.0, whose tasks have all completed, from pool 16:02:07.549 INFO DAGScheduler - ShuffleMapStage 1240 (mapToPair at MarkDuplicatesSpark.java:217) finished in 0.006 s 16:02:07.549 INFO DAGScheduler - looking for newly runnable stages 16:02:07.549 INFO DAGScheduler - running: HashSet() 16:02:07.549 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 1241, ResultStage 1242) 16:02:07.549 INFO DAGScheduler - failed: HashSet() 16:02:07.549 INFO DAGScheduler - Submitting ShuffleMapStage 1241 (MapPartitionsRDD[5642] at mapToPair at MarkDuplicatesSparkUtils.java:497), which has no missing parents 16:02:07.550 INFO MemoryStore - Block broadcast_2036 stored as values in memory (estimated size 15.9 KiB, free 1917.7 MiB) 16:02:07.550 INFO MemoryStore - Block broadcast_2036_piece0 stored as bytes in memory (estimated size 8.3 KiB, free 1917.7 MiB) 16:02:07.550 INFO BlockManagerInfo - Added broadcast_2036_piece0 in memory on localhost:42397 (size: 8.3 KiB, free: 1919.4 MiB) 16:02:07.550 INFO SparkContext - Created broadcast 2036 from broadcast at DAGScheduler.scala:1580 16:02:07.550 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1241 (MapPartitionsRDD[5642] at mapToPair at MarkDuplicatesSparkUtils.java:497) (first 15 tasks are for partitions Vector(0)) 16:02:07.550 INFO TaskSchedulerImpl - Adding task set 1241.0 with 1 tasks resource profile 0 16:02:07.551 INFO TaskSetManager - Starting task 0.0 in stage 1241.0 (TID 1068) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:02:07.551 INFO Executor - Running task 0.0 in stage 1241.0 (TID 1068) 16:02:07.552 INFO ShuffleBlockFetcherIterator - Getting 1 (490.0 B) non-empty blocks including 1 (490.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.552 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.553 INFO ShuffleBlockFetcherIterator - Getting 1 (60.0 B) non-empty blocks including 1 (60.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.553 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.554 INFO Executor - Finished task 0.0 in stage 1241.0 (TID 1068). 1879 bytes result sent to driver 16:02:07.554 INFO TaskSetManager - Finished task 0.0 in stage 1241.0 (TID 1068) in 3 ms on localhost (executor driver) (1/1) 16:02:07.554 INFO TaskSchedulerImpl - Removed TaskSet 1241.0, whose tasks have all completed, from pool 16:02:07.554 INFO DAGScheduler - ShuffleMapStage 1241 (mapToPair at MarkDuplicatesSparkUtils.java:497) finished in 0.005 s 16:02:07.554 INFO DAGScheduler - looking for newly runnable stages 16:02:07.554 INFO DAGScheduler - running: HashSet() 16:02:07.554 INFO DAGScheduler - waiting: HashSet(ResultStage 1242) 16:02:07.554 INFO DAGScheduler - failed: HashSet() 16:02:07.554 INFO DAGScheduler - Submitting ResultStage 1242 (MapPartitionsRDD[5644] at mapValues at MarkDuplicatesSparkUtils.java:519), which has no missing parents 16:02:07.555 INFO MemoryStore - Block broadcast_2037 stored as values in memory (estimated size 17.1 KiB, free 1917.7 MiB) 16:02:07.555 INFO MemoryStore - Block broadcast_2037_piece0 stored as bytes in memory (estimated size 9.0 KiB, free 1917.7 MiB) 16:02:07.555 INFO BlockManagerInfo - Added broadcast_2037_piece0 in memory on localhost:42397 (size: 9.0 KiB, free: 1919.4 MiB) 16:02:07.555 INFO SparkContext - Created broadcast 2037 from broadcast at DAGScheduler.scala:1580 16:02:07.556 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1242 (MapPartitionsRDD[5644] at mapValues at MarkDuplicatesSparkUtils.java:519) (first 15 tasks are for partitions Vector(0)) 16:02:07.556 INFO TaskSchedulerImpl - Adding task set 1242.0 with 1 tasks resource profile 0 16:02:07.556 INFO TaskSetManager - Starting task 0.0 in stage 1242.0 (TID 1069) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.556 INFO Executor - Running task 0.0 in stage 1242.0 (TID 1069) 16:02:07.557 INFO ShuffleBlockFetcherIterator - Getting 1 (129.0 B) non-empty blocks including 1 (129.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.557 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.558 INFO Executor - Finished task 0.0 in stage 1242.0 (TID 1069). 1887 bytes result sent to driver 16:02:07.558 INFO TaskSetManager - Finished task 0.0 in stage 1242.0 (TID 1069) in 2 ms on localhost (executor driver) (1/1) 16:02:07.558 INFO TaskSchedulerImpl - Removed TaskSet 1242.0, whose tasks have all completed, from pool 16:02:07.558 INFO DAGScheduler - ResultStage 1242 (collectAsMap at MarkDuplicatesSparkUtils.java:542) finished in 0.003 s 16:02:07.558 INFO DAGScheduler - Job 573 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.559 INFO TaskSchedulerImpl - Killing all running tasks in stage 1242: Stage finished 16:02:07.559 INFO DAGScheduler - Job 573 finished: collectAsMap at MarkDuplicatesSparkUtils.java:542, took 0.022600 s 16:02:07.568 INFO MemoryStore - Block broadcast_2038 stored as values in memory (estimated size 5.6 KiB, free 1917.7 MiB) 16:02:07.569 INFO MemoryStore - Block broadcast_2038_piece0 stored as bytes in memory (estimated size 500.0 B, free 1917.7 MiB) 16:02:07.569 INFO BlockManagerInfo - Added broadcast_2038_piece0 in memory on localhost:42397 (size: 500.0 B, free: 1919.4 MiB) 16:02:07.569 INFO SparkContext - Created broadcast 2038 from broadcast at ReadsSparkSink.java:133 16:02:07.572 INFO HadoopMapRedCommitProtocol - Using output committer class org.apache.hadoop.mapred.FileOutputCommitter 16:02:07.572 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:02:07.572 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:02:07.584 INFO SparkContext - Starting job: runJob at SparkHadoopWriter.scala:83 16:02:07.584 INFO DAGScheduler - Registering RDD 5646 (mapToPair at SparkUtils.java:161) as input to shuffle 306 16:02:07.584 INFO DAGScheduler - Got job 574 (runJob at SparkHadoopWriter.scala:83) with 1 output partitions 16:02:07.584 INFO DAGScheduler - Final stage: ResultStage 1247 (runJob at SparkHadoopWriter.scala:83) 16:02:07.584 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1246) 16:02:07.584 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1246) 16:02:07.584 INFO DAGScheduler - Submitting ShuffleMapStage 1246 (MapPartitionsRDD[5646] at mapToPair at SparkUtils.java:161), which has no missing parents 16:02:07.585 INFO MemoryStore - Block broadcast_2039 stored as values in memory (estimated size 15.0 KiB, free 1917.7 MiB) 16:02:07.585 INFO MemoryStore - Block broadcast_2039_piece0 stored as bytes in memory (estimated size 7.9 KiB, free 1917.7 MiB) 16:02:07.585 INFO BlockManagerInfo - Added broadcast_2039_piece0 in memory on localhost:42397 (size: 7.9 KiB, free: 1919.4 MiB) 16:02:07.585 INFO SparkContext - Created broadcast 2039 from broadcast at DAGScheduler.scala:1580 16:02:07.586 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1246 (MapPartitionsRDD[5646] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:02:07.586 INFO TaskSchedulerImpl - Adding task set 1246.0 with 1 tasks resource profile 0 16:02:07.586 INFO TaskSetManager - Starting task 0.0 in stage 1246.0 (TID 1070) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:02:07.586 INFO Executor - Running task 0.0 in stage 1246.0 (TID 1070) 16:02:07.587 INFO ShuffleBlockFetcherIterator - Getting 1 (490.0 B) non-empty blocks including 1 (490.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.587 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.589 INFO ShuffleBlockFetcherIterator - Getting 1 (60.0 B) non-empty blocks including 1 (60.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.589 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.590 INFO Executor - Finished task 0.0 in stage 1246.0 (TID 1070). 1879 bytes result sent to driver 16:02:07.590 INFO TaskSetManager - Finished task 0.0 in stage 1246.0 (TID 1070) in 4 ms on localhost (executor driver) (1/1) 16:02:07.590 INFO TaskSchedulerImpl - Removed TaskSet 1246.0, whose tasks have all completed, from pool 16:02:07.590 INFO DAGScheduler - ShuffleMapStage 1246 (mapToPair at SparkUtils.java:161) finished in 0.006 s 16:02:07.590 INFO DAGScheduler - looking for newly runnable stages 16:02:07.590 INFO DAGScheduler - running: HashSet() 16:02:07.590 INFO DAGScheduler - waiting: HashSet(ResultStage 1247) 16:02:07.590 INFO DAGScheduler - failed: HashSet() 16:02:07.590 INFO DAGScheduler - Submitting ResultStage 1247 (MapPartitionsRDD[5652] at saveAsTextFile at SamSink.java:65), which has no missing parents 16:02:07.597 INFO MemoryStore - Block broadcast_2040 stored as values in memory (estimated size 159.8 KiB, free 1917.5 MiB) 16:02:07.598 INFO MemoryStore - Block broadcast_2040_piece0 stored as bytes in memory (estimated size 71.5 KiB, free 1917.5 MiB) 16:02:07.598 INFO BlockManagerInfo - Added broadcast_2040_piece0 in memory on localhost:42397 (size: 71.5 KiB, free: 1919.4 MiB) 16:02:07.598 INFO SparkContext - Created broadcast 2040 from broadcast at DAGScheduler.scala:1580 16:02:07.599 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1247 (MapPartitionsRDD[5652] at saveAsTextFile at SamSink.java:65) (first 15 tasks are for partitions Vector(0)) 16:02:07.599 INFO TaskSchedulerImpl - Adding task set 1247.0 with 1 tasks resource profile 0 16:02:07.599 INFO TaskSetManager - Starting task 0.0 in stage 1247.0 (TID 1071) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.599 INFO Executor - Running task 0.0 in stage 1247.0 (TID 1071) 16:02:07.602 INFO ShuffleBlockFetcherIterator - Getting 1 (490.0 B) non-empty blocks including 1 (490.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.602 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.604 INFO HadoopMapRedCommitProtocol - Using output committer class org.apache.hadoop.mapred.FileOutputCommitter 16:02:07.604 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:02:07.604 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:02:07.610 INFO FileOutputCommitter - Saved output of task 'attempt_20250519160207289338380496033331_5652_m_000000_0' to file:/tmp/MarkDups10436947302969018299/MarkDups.sam.parts/_temporary/0/task_20250519160207289338380496033331_5652_m_000000 16:02:07.610 INFO SparkHadoopMapRedUtil - attempt_20250519160207289338380496033331_5652_m_000000_0: Committed. Elapsed time: 0 ms. 16:02:07.610 INFO Executor - Finished task 0.0 in stage 1247.0 (TID 1071). 1858 bytes result sent to driver 16:02:07.610 INFO TaskSetManager - Finished task 0.0 in stage 1247.0 (TID 1071) in 11 ms on localhost (executor driver) (1/1) 16:02:07.610 INFO TaskSchedulerImpl - Removed TaskSet 1247.0, whose tasks have all completed, from pool 16:02:07.610 INFO DAGScheduler - ResultStage 1247 (runJob at SparkHadoopWriter.scala:83) finished in 0.020 s 16:02:07.610 INFO DAGScheduler - Job 574 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.610 INFO TaskSchedulerImpl - Killing all running tasks in stage 1247: Stage finished 16:02:07.610 INFO DAGScheduler - Job 574 finished: runJob at SparkHadoopWriter.scala:83, took 0.026606 s 16:02:07.610 INFO SparkHadoopWriter - Start to commit write Job job_20250519160207289338380496033331_5652. 16:02:07.613 INFO SparkHadoopWriter - Write Job job_20250519160207289338380496033331_5652 committed. Elapsed time: 2 ms. 16:02:07.616 INFO HadoopFileSystemWrapper - Concatenating 2 parts to /tmp/MarkDups10436947302969018299/MarkDups.sam 16:02:07.618 INFO HadoopFileSystemWrapper - Concatenating to /tmp/MarkDups10436947302969018299/MarkDups.sam done 16:02:07.619 INFO MarkDuplicatesSpark - Shutting down engine [May 19, 2025 at 4:02:07 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 16:02:07.634 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.634 INFO MarkDuplicatesSpark - The Genome Analysis Toolkit (GATK) va6bd333-SNAPSHOT 16:02:07.634 INFO MarkDuplicatesSpark - For support and documentation go to https://software.broadinstitute.org/gatk/ 16:02:07.634 INFO MarkDuplicatesSpark - Executing as root@67ac59543751 on Linux v6.11.0-1014-azure amd64 16:02:07.634 INFO MarkDuplicatesSpark - Java runtime: OpenJDK 64-Bit Server VM v17.0.12+7-Ubuntu-1ubuntu222.04 16:02:07.634 INFO MarkDuplicatesSpark - Start Date/Time: May 19, 2025 at 4:02:07 PM GMT 16:02:07.634 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.634 INFO MarkDuplicatesSpark - ------------------------------------------------------------ 16:02:07.634 INFO MarkDuplicatesSpark - HTSJDK Version: test_cram_wiring-SNAPSHOT 16:02:07.634 INFO MarkDuplicatesSpark - Picard Version: 3.4.0 16:02:07.634 INFO MarkDuplicatesSpark - Built for Spark Version: 3.5.0 16:02:07.634 INFO MarkDuplicatesSpark - HTSJDK Defaults.COMPRESSION_LEVEL : 2 16:02:07.634 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_READ_FOR_SAMTOOLS : false 16:02:07.634 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_SAMTOOLS : true 16:02:07.634 INFO MarkDuplicatesSpark - HTSJDK Defaults.USE_ASYNC_IO_WRITE_FOR_TRIBBLE : false 16:02:07.634 INFO MarkDuplicatesSpark - Deflater: IntelDeflater 16:02:07.634 INFO MarkDuplicatesSpark - Inflater: IntelInflater 16:02:07.634 INFO MarkDuplicatesSpark - GCS max retries/reopens: 20 16:02:07.634 INFO MarkDuplicatesSpark - Requester pays: disabled 16:02:07.634 INFO MarkDuplicatesSpark - Initializing engine 16:02:07.634 INFO MarkDuplicatesSpark - Done initializing engine 16:02:07.634 INFO MarkDuplicatesSpark - Spark verbosity set to INFO (see --spark-verbosity argument) 16:02:07.636 INFO MemoryStore - Block broadcast_2041 stored as values in memory (estimated size 305.5 KiB, free 1917.2 MiB) 16:02:07.643 INFO MemoryStore - Block broadcast_2041_piece0 stored as bytes in memory (estimated size 64.1 KiB, free 1917.1 MiB) 16:02:07.643 INFO BlockManagerInfo - Added broadcast_2041_piece0 in memory on localhost:42397 (size: 64.1 KiB, free: 1919.3 MiB) 16:02:07.643 INFO SparkContext - Created broadcast 2041 from newAPIHadoopFile at PathSplitSource.java:96 16:02:07.663 INFO MemoryStore - Block broadcast_2042 stored as values in memory (estimated size 305.5 KiB, free 1916.8 MiB) 16:02:07.669 INFO MemoryStore - Block broadcast_2042_piece0 stored as bytes in memory (estimated size 64.1 KiB, free 1916.7 MiB) 16:02:07.669 INFO BlockManagerInfo - Added broadcast_2042_piece0 in memory on localhost:42397 (size: 64.1 KiB, free: 1919.2 MiB) 16:02:07.669 INFO SparkContext - Created broadcast 2042 from newAPIHadoopFile at PathSplitSource.java:96 16:02:07.701 INFO SparkContext - Starting job: collect at SparkUtils.java:205 16:02:07.701 INFO FileInputFormat - Total input files to process : 1 16:02:07.702 INFO DAGScheduler - Registering RDD 5666 (mapToPair at SparkUtils.java:161) as input to shuffle 307 16:02:07.702 INFO DAGScheduler - Got job 575 (collect at SparkUtils.java:205) with 1 output partitions 16:02:07.702 INFO DAGScheduler - Final stage: ResultStage 1249 (collect at SparkUtils.java:205) 16:02:07.702 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1248) 16:02:07.702 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1248) 16:02:07.702 INFO DAGScheduler - Submitting ShuffleMapStage 1248 (MapPartitionsRDD[5666] at mapToPair at SparkUtils.java:161), which has no missing parents 16:02:07.720 INFO MemoryStore - Block broadcast_2043 stored as values in memory (estimated size 480.2 KiB, free 1916.3 MiB) 16:02:07.722 INFO MemoryStore - Block broadcast_2043_piece0 stored as bytes in memory (estimated size 206.2 KiB, free 1916.1 MiB) 16:02:07.722 INFO BlockManagerInfo - Added broadcast_2043_piece0 in memory on localhost:42397 (size: 206.2 KiB, free: 1919.0 MiB) 16:02:07.722 INFO SparkContext - Created broadcast 2043 from broadcast at DAGScheduler.scala:1580 16:02:07.722 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1248 (MapPartitionsRDD[5666] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:02:07.722 INFO TaskSchedulerImpl - Adding task set 1248.0 with 1 tasks resource profile 0 16:02:07.723 INFO TaskSetManager - Starting task 0.0 in stage 1248.0 (TID 1072) (localhost, executor driver, partition 0, PROCESS_LOCAL, 9673 bytes) 16:02:07.723 INFO Executor - Running task 0.0 in stage 1248.0 (TID 1072) 16:02:07.751 INFO NewHadoopRDD - Input split: file:/gatkCloneMountPoint/src/test/resources/org/broadinstitute/hellbender/tools/walkers/MarkDuplicatesGATK/inputSingleLibrarySolexa16404.bam:0+5416 16:02:07.756 INFO Executor - Finished task 0.0 in stage 1248.0 (TID 1072). 1148 bytes result sent to driver 16:02:07.757 INFO TaskSetManager - Finished task 0.0 in stage 1248.0 (TID 1072) in 34 ms on localhost (executor driver) (1/1) 16:02:07.757 INFO TaskSchedulerImpl - Removed TaskSet 1248.0, whose tasks have all completed, from pool 16:02:07.757 INFO DAGScheduler - ShuffleMapStage 1248 (mapToPair at SparkUtils.java:161) finished in 0.055 s 16:02:07.757 INFO DAGScheduler - looking for newly runnable stages 16:02:07.757 INFO DAGScheduler - running: HashSet() 16:02:07.757 INFO DAGScheduler - waiting: HashSet(ResultStage 1249) 16:02:07.757 INFO DAGScheduler - failed: HashSet() 16:02:07.757 INFO DAGScheduler - Submitting ResultStage 1249 (MapPartitionsRDD[5669] at mapPartitions at SparkUtils.java:188), which has no missing parents 16:02:07.757 INFO MemoryStore - Block broadcast_2044 stored as values in memory (estimated size 7.4 KiB, free 1916.1 MiB) 16:02:07.758 INFO MemoryStore - Block broadcast_2044_piece0 stored as bytes in memory (estimated size 4.1 KiB, free 1916.1 MiB) 16:02:07.758 INFO BlockManagerInfo - Added broadcast_2044_piece0 in memory on localhost:42397 (size: 4.1 KiB, free: 1919.0 MiB) 16:02:07.758 INFO SparkContext - Created broadcast 2044 from broadcast at DAGScheduler.scala:1580 16:02:07.758 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1249 (MapPartitionsRDD[5669] at mapPartitions at SparkUtils.java:188) (first 15 tasks are for partitions Vector(0)) 16:02:07.758 INFO TaskSchedulerImpl - Adding task set 1249.0 with 1 tasks resource profile 0 16:02:07.759 INFO TaskSetManager - Starting task 0.0 in stage 1249.0 (TID 1073) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.759 INFO Executor - Running task 0.0 in stage 1249.0 (TID 1073) 16:02:07.760 INFO ShuffleBlockFetcherIterator - Getting 1 (4.3 KiB) non-empty blocks including 1 (4.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.760 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.761 INFO Executor - Finished task 0.0 in stage 1249.0 (TID 1073). 2039 bytes result sent to driver 16:02:07.761 INFO TaskSetManager - Finished task 0.0 in stage 1249.0 (TID 1073) in 2 ms on localhost (executor driver) (1/1) 16:02:07.761 INFO TaskSchedulerImpl - Removed TaskSet 1249.0, whose tasks have all completed, from pool 16:02:07.761 INFO DAGScheduler - ResultStage 1249 (collect at SparkUtils.java:205) finished in 0.004 s 16:02:07.761 INFO DAGScheduler - Job 575 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.761 INFO TaskSchedulerImpl - Killing all running tasks in stage 1249: Stage finished 16:02:07.762 INFO DAGScheduler - Job 575 finished: collect at SparkUtils.java:205, took 0.060667 s 16:02:07.765 INFO MemoryStore - Block broadcast_2045 stored as values in memory (estimated size 392.0 B, free 1916.1 MiB) 16:02:07.766 INFO MemoryStore - Block broadcast_2045_piece0 stored as bytes in memory (estimated size 33.0 B, free 1916.1 MiB) 16:02:07.766 INFO BlockManagerInfo - Added broadcast_2045_piece0 in memory on localhost:42397 (size: 33.0 B, free: 1919.0 MiB) 16:02:07.766 INFO SparkContext - Created broadcast 2045 from broadcast at MarkDuplicatesSparkUtils.java:126 16:02:07.766 INFO MemoryStore - Block broadcast_2046 stored as values in memory (estimated size 144.0 B, free 1916.1 MiB) 16:02:07.766 INFO MemoryStore - Block broadcast_2046_piece0 stored as bytes in memory (estimated size 28.0 B, free 1916.1 MiB) 16:02:07.767 INFO BlockManagerInfo - Added broadcast_2046_piece0 in memory on localhost:42397 (size: 28.0 B, free: 1919.0 MiB) 16:02:07.767 INFO SparkContext - Created broadcast 2046 from broadcast at MarkDuplicatesSparkUtils.java:127 16:02:07.779 INFO SparkContext - Starting job: collectAsMap at MarkDuplicatesSparkUtils.java:542 16:02:07.780 INFO DAGScheduler - Registering RDD 5678 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) as input to shuffle 310 16:02:07.780 INFO DAGScheduler - Registering RDD 5682 (mapToPair at MarkDuplicatesSpark.java:217) as input to shuffle 309 16:02:07.780 INFO DAGScheduler - Registering RDD 5686 (mapToPair at MarkDuplicatesSparkUtils.java:497) as input to shuffle 308 16:02:07.780 INFO DAGScheduler - Got job 576 (collectAsMap at MarkDuplicatesSparkUtils.java:542) with 1 output partitions 16:02:07.780 INFO DAGScheduler - Final stage: ResultStage 1254 (collectAsMap at MarkDuplicatesSparkUtils.java:542) 16:02:07.780 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1253) 16:02:07.780 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1253) 16:02:07.780 INFO DAGScheduler - Submitting ShuffleMapStage 1251 (MapPartitionsRDD[5678] at flatMapToPair at MarkDuplicatesSparkUtils.java:130), which has no missing parents 16:02:07.781 INFO MemoryStore - Block broadcast_2047 stored as values in memory (estimated size 45.2 KiB, free 1916.0 MiB) 16:02:07.782 INFO MemoryStore - Block broadcast_2047_piece0 stored as bytes in memory (estimated size 17.1 KiB, free 1916.0 MiB) 16:02:07.782 INFO BlockManagerInfo - Added broadcast_2047_piece0 in memory on localhost:42397 (size: 17.1 KiB, free: 1919.0 MiB) 16:02:07.782 INFO SparkContext - Created broadcast 2047 from broadcast at DAGScheduler.scala:1580 16:02:07.782 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1251 (MapPartitionsRDD[5678] at flatMapToPair at MarkDuplicatesSparkUtils.java:130) (first 15 tasks are for partitions Vector(0)) 16:02:07.782 INFO TaskSchedulerImpl - Adding task set 1251.0 with 1 tasks resource profile 0 16:02:07.783 INFO TaskSetManager - Starting task 0.0 in stage 1251.0 (TID 1074) (localhost, executor driver, partition 0, NODE_LOCAL, 9813 bytes) 16:02:07.783 INFO Executor - Running task 0.0 in stage 1251.0 (TID 1074) 16:02:07.784 INFO ShuffleBlockFetcherIterator - Getting 1 (4.3 KiB) non-empty blocks including 1 (4.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.784 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.787 INFO Executor - Finished task 0.0 in stage 1251.0 (TID 1074). 1922 bytes result sent to driver 16:02:07.787 INFO TaskSetManager - Finished task 0.0 in stage 1251.0 (TID 1074) in 5 ms on localhost (executor driver) (1/1) 16:02:07.787 INFO TaskSchedulerImpl - Removed TaskSet 1251.0, whose tasks have all completed, from pool 16:02:07.787 INFO DAGScheduler - ShuffleMapStage 1251 (flatMapToPair at MarkDuplicatesSparkUtils.java:130) finished in 0.007 s 16:02:07.787 INFO DAGScheduler - looking for newly runnable stages 16:02:07.787 INFO DAGScheduler - running: HashSet() 16:02:07.787 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 1252, ShuffleMapStage 1253, ResultStage 1254) 16:02:07.787 INFO DAGScheduler - failed: HashSet() 16:02:07.787 INFO DAGScheduler - Submitting ShuffleMapStage 1252 (MapPartitionsRDD[5682] at mapToPair at MarkDuplicatesSpark.java:217), which has no missing parents 16:02:07.788 INFO MemoryStore - Block broadcast_2048 stored as values in memory (estimated size 49.3 KiB, free 1915.9 MiB) 16:02:07.789 INFO MemoryStore - Block broadcast_2048_piece0 stored as bytes in memory (estimated size 19.3 KiB, free 1915.9 MiB) 16:02:07.789 INFO BlockManagerInfo - Added broadcast_2048_piece0 in memory on localhost:42397 (size: 19.3 KiB, free: 1919.0 MiB) 16:02:07.789 INFO SparkContext - Created broadcast 2048 from broadcast at DAGScheduler.scala:1580 16:02:07.789 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1252 (MapPartitionsRDD[5682] at mapToPair at MarkDuplicatesSpark.java:217) (first 15 tasks are for partitions Vector(0)) 16:02:07.789 INFO TaskSchedulerImpl - Adding task set 1252.0 with 1 tasks resource profile 0 16:02:07.789 INFO TaskSetManager - Starting task 0.0 in stage 1252.0 (TID 1075) (localhost, executor driver, partition 0, NODE_LOCAL, 9288 bytes) 16:02:07.790 INFO Executor - Running task 0.0 in stage 1252.0 (TID 1075) 16:02:07.791 INFO ShuffleBlockFetcherIterator - Getting 1 (593.0 B) non-empty blocks including 1 (593.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.791 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.794 INFO Executor - Finished task 0.0 in stage 1252.0 (TID 1075). 1922 bytes result sent to driver 16:02:07.795 INFO TaskSetManager - Finished task 0.0 in stage 1252.0 (TID 1075) in 6 ms on localhost (executor driver) (1/1) 16:02:07.795 INFO TaskSchedulerImpl - Removed TaskSet 1252.0, whose tasks have all completed, from pool 16:02:07.795 INFO DAGScheduler - ShuffleMapStage 1252 (mapToPair at MarkDuplicatesSpark.java:217) finished in 0.007 s 16:02:07.795 INFO DAGScheduler - looking for newly runnable stages 16:02:07.795 INFO DAGScheduler - running: HashSet() 16:02:07.795 INFO DAGScheduler - waiting: HashSet(ShuffleMapStage 1253, ResultStage 1254) 16:02:07.795 INFO DAGScheduler - failed: HashSet() 16:02:07.795 INFO DAGScheduler - Submitting ShuffleMapStage 1253 (MapPartitionsRDD[5686] at mapToPair at MarkDuplicatesSparkUtils.java:497), which has no missing parents 16:02:07.796 INFO MemoryStore - Block broadcast_2049 stored as values in memory (estimated size 40.0 KiB, free 1915.9 MiB) 16:02:07.796 INFO MemoryStore - Block broadcast_2049_piece0 stored as bytes in memory (estimated size 14.5 KiB, free 1915.9 MiB) 16:02:07.796 INFO BlockManagerInfo - Added broadcast_2049_piece0 in memory on localhost:42397 (size: 14.5 KiB, free: 1919.0 MiB) 16:02:07.796 INFO SparkContext - Created broadcast 2049 from broadcast at DAGScheduler.scala:1580 16:02:07.797 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1253 (MapPartitionsRDD[5686] at mapToPair at MarkDuplicatesSparkUtils.java:497) (first 15 tasks are for partitions Vector(0)) 16:02:07.797 INFO TaskSchedulerImpl - Adding task set 1253.0 with 1 tasks resource profile 0 16:02:07.797 INFO TaskSetManager - Starting task 0.0 in stage 1253.0 (TID 1076) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:02:07.797 INFO Executor - Running task 0.0 in stage 1253.0 (TID 1076) 16:02:07.798 INFO ShuffleBlockFetcherIterator - Getting 1 (4.3 KiB) non-empty blocks including 1 (4.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.798 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.799 INFO ShuffleBlockFetcherIterator - Getting 1 (652.0 B) non-empty blocks including 1 (652.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.800 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.801 INFO Executor - Finished task 0.0 in stage 1253.0 (TID 1076). 1879 bytes result sent to driver 16:02:07.801 INFO TaskSetManager - Finished task 0.0 in stage 1253.0 (TID 1076) in 4 ms on localhost (executor driver) (1/1) 16:02:07.801 INFO TaskSchedulerImpl - Removed TaskSet 1253.0, whose tasks have all completed, from pool 16:02:07.801 INFO DAGScheduler - ShuffleMapStage 1253 (mapToPair at MarkDuplicatesSparkUtils.java:497) finished in 0.006 s 16:02:07.801 INFO DAGScheduler - looking for newly runnable stages 16:02:07.801 INFO DAGScheduler - running: HashSet() 16:02:07.801 INFO DAGScheduler - waiting: HashSet(ResultStage 1254) 16:02:07.801 INFO DAGScheduler - failed: HashSet() 16:02:07.801 INFO DAGScheduler - Submitting ResultStage 1254 (MapPartitionsRDD[5688] at mapValues at MarkDuplicatesSparkUtils.java:519), which has no missing parents 16:02:07.802 INFO MemoryStore - Block broadcast_2050 stored as values in memory (estimated size 41.2 KiB, free 1915.8 MiB) 16:02:07.802 INFO MemoryStore - Block broadcast_2050_piece0 stored as bytes in memory (estimated size 15.3 KiB, free 1915.8 MiB) 16:02:07.802 INFO BlockManagerInfo - Added broadcast_2050_piece0 in memory on localhost:42397 (size: 15.3 KiB, free: 1919.0 MiB) 16:02:07.803 INFO SparkContext - Created broadcast 2050 from broadcast at DAGScheduler.scala:1580 16:02:07.803 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1254 (MapPartitionsRDD[5688] at mapValues at MarkDuplicatesSparkUtils.java:519) (first 15 tasks are for partitions Vector(0)) 16:02:07.803 INFO TaskSchedulerImpl - Adding task set 1254.0 with 1 tasks resource profile 0 16:02:07.803 INFO TaskSetManager - Starting task 0.0 in stage 1254.0 (TID 1077) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.803 INFO Executor - Running task 0.0 in stage 1254.0 (TID 1077) 16:02:07.805 INFO ShuffleBlockFetcherIterator - Getting 1 (142.0 B) non-empty blocks including 1 (142.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.805 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.806 INFO Executor - Finished task 0.0 in stage 1254.0 (TID 1077). 1944 bytes result sent to driver 16:02:07.806 INFO TaskSetManager - Finished task 0.0 in stage 1254.0 (TID 1077) in 3 ms on localhost (executor driver) (1/1) 16:02:07.807 INFO TaskSchedulerImpl - Removed TaskSet 1254.0, whose tasks have all completed, from pool 16:02:07.807 INFO DAGScheduler - ResultStage 1254 (collectAsMap at MarkDuplicatesSparkUtils.java:542) finished in 0.006 s 16:02:07.807 INFO DAGScheduler - Job 576 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.807 INFO TaskSchedulerImpl - Killing all running tasks in stage 1254: Stage finished 16:02:07.807 INFO DAGScheduler - Job 576 finished: collectAsMap at MarkDuplicatesSparkUtils.java:542, took 0.027441 s 16:02:07.818 INFO MemoryStore - Block broadcast_2051 stored as values in memory (estimated size 93.5 KiB, free 1915.7 MiB) 16:02:07.818 INFO MemoryStore - Block broadcast_2051_piece0 stored as bytes in memory (estimated size 5.6 KiB, free 1915.7 MiB) 16:02:07.818 INFO BlockManagerInfo - Added broadcast_2051_piece0 in memory on localhost:42397 (size: 5.6 KiB, free: 1919.0 MiB) 16:02:07.818 INFO SparkContext - Created broadcast 2051 from broadcast at ReadsSparkSink.java:133 16:02:07.821 INFO HadoopMapRedCommitProtocol - Using output committer class org.apache.hadoop.mapred.FileOutputCommitter 16:02:07.822 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:02:07.822 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:02:07.833 INFO SparkContext - Starting job: runJob at SparkHadoopWriter.scala:83 16:02:07.833 INFO DAGScheduler - Registering RDD 5690 (mapToPair at SparkUtils.java:161) as input to shuffle 311 16:02:07.833 INFO DAGScheduler - Got job 577 (runJob at SparkHadoopWriter.scala:83) with 1 output partitions 16:02:07.833 INFO DAGScheduler - Final stage: ResultStage 1259 (runJob at SparkHadoopWriter.scala:83) 16:02:07.833 INFO DAGScheduler - Parents of final stage: List(ShuffleMapStage 1258) 16:02:07.833 INFO DAGScheduler - Missing parents: List(ShuffleMapStage 1258) 16:02:07.833 INFO DAGScheduler - Submitting ShuffleMapStage 1258 (MapPartitionsRDD[5690] at mapToPair at SparkUtils.java:161), which has no missing parents 16:02:07.834 INFO MemoryStore - Block broadcast_2052 stored as values in memory (estimated size 39.1 KiB, free 1915.7 MiB) 16:02:07.834 INFO MemoryStore - Block broadcast_2052_piece0 stored as bytes in memory (estimated size 14.1 KiB, free 1915.7 MiB) 16:02:07.834 INFO BlockManagerInfo - Added broadcast_2052_piece0 in memory on localhost:42397 (size: 14.1 KiB, free: 1918.9 MiB) 16:02:07.835 INFO SparkContext - Created broadcast 2052 from broadcast at DAGScheduler.scala:1580 16:02:07.835 INFO DAGScheduler - Submitting 1 missing tasks from ShuffleMapStage 1258 (MapPartitionsRDD[5690] at mapToPair at SparkUtils.java:161) (first 15 tasks are for partitions Vector(0)) 16:02:07.835 INFO TaskSchedulerImpl - Adding task set 1258.0 with 1 tasks resource profile 0 16:02:07.835 INFO TaskSetManager - Starting task 0.0 in stage 1258.0 (TID 1078) (localhost, executor driver, partition 0, NODE_LOCAL, 9871 bytes) 16:02:07.835 INFO Executor - Running task 0.0 in stage 1258.0 (TID 1078) 16:02:07.837 INFO ShuffleBlockFetcherIterator - Getting 1 (4.3 KiB) non-empty blocks including 1 (4.3 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.837 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.838 INFO ShuffleBlockFetcherIterator - Getting 1 (652.0 B) non-empty blocks including 1 (652.0 B) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.838 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.839 INFO Executor - Finished task 0.0 in stage 1258.0 (TID 1078). 1922 bytes result sent to driver 16:02:07.840 INFO TaskSetManager - Finished task 0.0 in stage 1258.0 (TID 1078) in 5 ms on localhost (executor driver) (1/1) 16:02:07.840 INFO TaskSchedulerImpl - Removed TaskSet 1258.0, whose tasks have all completed, from pool 16:02:07.840 INFO DAGScheduler - ShuffleMapStage 1258 (mapToPair at SparkUtils.java:161) finished in 0.007 s 16:02:07.840 INFO DAGScheduler - looking for newly runnable stages 16:02:07.840 INFO DAGScheduler - running: HashSet() 16:02:07.840 INFO DAGScheduler - waiting: HashSet(ResultStage 1259) 16:02:07.840 INFO DAGScheduler - failed: HashSet() 16:02:07.840 INFO DAGScheduler - Submitting ResultStage 1259 (MapPartitionsRDD[5696] at saveAsTextFile at SamSink.java:65), which has no missing parents 16:02:07.846 INFO MemoryStore - Block broadcast_2053 stored as values in memory (estimated size 183.9 KiB, free 1915.5 MiB) 16:02:07.847 INFO MemoryStore - Block broadcast_2053_piece0 stored as bytes in memory (estimated size 77.7 KiB, free 1915.4 MiB) 16:02:07.847 INFO BlockManagerInfo - Added broadcast_2053_piece0 in memory on localhost:42397 (size: 77.7 KiB, free: 1918.9 MiB) 16:02:07.848 INFO SparkContext - Created broadcast 2053 from broadcast at DAGScheduler.scala:1580 16:02:07.848 INFO DAGScheduler - Submitting 1 missing tasks from ResultStage 1259 (MapPartitionsRDD[5696] at saveAsTextFile at SamSink.java:65) (first 15 tasks are for partitions Vector(0)) 16:02:07.848 INFO TaskSchedulerImpl - Adding task set 1259.0 with 1 tasks resource profile 0 16:02:07.848 INFO TaskSetManager - Starting task 0.0 in stage 1259.0 (TID 1079) (localhost, executor driver, partition 0, NODE_LOCAL, 9299 bytes) 16:02:07.848 INFO Executor - Running task 0.0 in stage 1259.0 (TID 1079) 16:02:07.852 INFO ShuffleBlockFetcherIterator - Getting 1 (3.9 KiB) non-empty blocks including 1 (3.9 KiB) local and 0 (0.0 B) host-local and 0 (0.0 B) push-merged-local and 0 (0.0 B) remote blocks 16:02:07.852 INFO ShuffleBlockFetcherIterator - Started 0 remote fetches in 0 ms 16:02:07.853 INFO HadoopMapRedCommitProtocol - Using output committer class org.apache.hadoop.mapred.FileOutputCommitter 16:02:07.853 INFO FileOutputCommitter - File Output Committer Algorithm version is 1 16:02:07.853 INFO FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 16:02:07.859 INFO FileOutputCommitter - Saved output of task 'attempt_202505191602075483652289307443919_5696_m_000000_0' to file:/tmp/MarkDups12998395014682154056/MarkDups.sam.parts/_temporary/0/task_202505191602075483652289307443919_5696_m_000000 16:02:07.859 INFO SparkHadoopMapRedUtil - attempt_202505191602075483652289307443919_5696_m_000000_0: Committed. Elapsed time: 0 ms. 16:02:07.859 INFO Executor - Finished task 0.0 in stage 1259.0 (TID 1079). 1858 bytes result sent to driver 16:02:07.860 INFO TaskSetManager - Finished task 0.0 in stage 1259.0 (TID 1079) in 12 ms on localhost (executor driver) (1/1) 16:02:07.860 INFO TaskSchedulerImpl - Removed TaskSet 1259.0, whose tasks have all completed, from pool 16:02:07.860 INFO DAGScheduler - ResultStage 1259 (runJob at SparkHadoopWriter.scala:83) finished in 0.020 s 16:02:07.860 INFO DAGScheduler - Job 577 is finished. Cancelling potential speculative or zombie tasks for this job 16:02:07.860 INFO TaskSchedulerImpl - Killing all running tasks in stage 1259: Stage finished 16:02:07.860 INFO DAGScheduler - Job 577 finished: runJob at SparkHadoopWriter.scala:83, took 0.027415 s 16:02:07.860 INFO SparkHadoopWriter - Start to commit write Job job_202505191602075483652289307443919_5696. 16:02:07.863 INFO SparkHadoopWriter - Write Job job_202505191602075483652289307443919_5696 committed. Elapsed time: 2 ms. 16:02:07.867 INFO HadoopFileSystemWrapper - Concatenating 2 parts to /tmp/MarkDups12998395014682154056/MarkDups.sam 16:02:07.869 INFO HadoopFileSystemWrapper - Concatenating to /tmp/MarkDups12998395014682154056/MarkDups.sam done 16:02:07.869 INFO MarkDuplicatesSpark - Shutting down engine [May 19, 2025 at 4:02:07 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:08 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:08 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:08 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:09 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:09 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:09 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:10 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:10 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:10 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:11 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:11 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:11 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:11 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:12 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:12 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:12 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:13 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:13 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 16:02:13.729 ERROR Executor - Exception in task 0.0 in stage 1479.0 (TID 1226) org.broadinstitute.hellbender.exceptions.UserException$ReadMissingReadGroup: SAM/BAM/CRAM file (unknown) is malformed: Read RUNID:7:1203:2886:82292 is missing the read group (RG) tag, which is required by the GATK. Please use http://gatkforums.broadinstitute.org/discussion/59/companion-utilities-replacereadgroups to fix this problem at org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSparkUtils.getLibraryForRead(MarkDuplicatesSparkUtils.java:64) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.broadinstitute.hellbender.utils.read.markduplicates.sparkrecords.EmptyFragment.<init>(EmptyFragment.java:36) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.broadinstitute.hellbender.utils.read.markduplicates.sparkrecords.MarkDuplicatesSparkRecord.newEmptyFragment(MarkDuplicatesSparkRecord.java:44) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSparkUtils.lambda$transformToDuplicateNames$0(MarkDuplicatesSparkUtils.java:141) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at java.base/java.util.stream.ReferencePipeline$15$1.accept(ReferencePipeline.java:540) ~[?:?] at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1625) ~[?:?] at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:509) ~[?:?] at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:499) ~[?:?] at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:921) ~[?:?] at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) ~[?:?] at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:682) ~[?:?] at org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSparkUtils.lambda$transformToDuplicateNames$5ac2632f$1(MarkDuplicatesSparkUtils.java:150) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.api.java.JavaRDDLike.$anonfun$flatMapToPair$1(JavaRDDLike.scala:143) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:594) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:608) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.scheduler.Task.run(Task.scala:141) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623) [gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) [?:?] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) [?:?] at java.base/java.lang.Thread.run(Thread.java:840) [?:?] 16:02:13.744 ERROR TaskSetManager - Task 0 in stage 1479.0 failed 1 times; aborting job [May 19, 2025 at 4:02:13 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 16:02:13.985 ERROR Executor - Exception in task 0.0 in stage 1486.0 (TID 1229) org.broadinstitute.hellbender.exceptions.UserException$HeaderMissingReadGroup: SAM/BAM/CRAM file (unknown) is malformed: Read RUNID:7:1203:2886:82292 contains an (RG) tag with the group NotADuplicateGroup which is not found in the file header. at org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSparkUtils.getLibraryForRead(MarkDuplicatesSparkUtils.java:66) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.broadinstitute.hellbender.utils.read.markduplicates.sparkrecords.EmptyFragment.<init>(EmptyFragment.java:36) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.broadinstitute.hellbender.utils.read.markduplicates.sparkrecords.MarkDuplicatesSparkRecord.newEmptyFragment(MarkDuplicatesSparkRecord.java:44) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSparkUtils.lambda$transformToDuplicateNames$0(MarkDuplicatesSparkUtils.java:141) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at java.base/java.util.stream.ReferencePipeline$15$1.accept(ReferencePipeline.java:540) ~[?:?] at java.base/java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1625) ~[?:?] at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:509) ~[?:?] at java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:499) ~[?:?] at java.base/java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:921) ~[?:?] at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) ~[?:?] at java.base/java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:682) ~[?:?] at org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSparkUtils.lambda$transformToDuplicateNames$5ac2632f$1(MarkDuplicatesSparkUtils.java:150) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.api.java.JavaRDDLike.$anonfun$flatMapToPair$1(JavaRDDLike.scala:143) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at scala.collection.Iterator$$anon$10.nextCur(Iterator.scala:594) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:608) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:140) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.scheduler.Task.run(Task.scala:141) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94) ~[gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623) [gatk-package-a6bd333-SNAPSHOT-local.jar:a6bd333-SNAPSHOT] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) [?:?] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) [?:?] at java.base/java.lang.Thread.run(Thread.java:840) [?:?] 16:02:13.987 ERROR TaskSetManager - Task 0 in stage 1486.0 failed 1 times; aborting job [May 19, 2025 at 4:02:13 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:14 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:14 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:14 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:15 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:15 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:15 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:16 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:16 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:16 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:17 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:17 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:17 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:18 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:18 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:18 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:18 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:19 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:19 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:19 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:20 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:20 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:20 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:21 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:21 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:21 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:21 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:22 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:22 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:22 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:23 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:23 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:23 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:24 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:24 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:24 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:25 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:25 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:25 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:25 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:26 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:26 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:26 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:27 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:27 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:27 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:28 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:28 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:28 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:29 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:29 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:29 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:29 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:30 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:30 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:30 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:31 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:31 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:31 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:32 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:32 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:32 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:32 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:33 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:33 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:33 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:34 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:34 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:34 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:35 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:35 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:35 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:35 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:36 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:36 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:36 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:37 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:37 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:37 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:38 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:38 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744 [May 19, 2025 at 4:02:38 PM GMT] org.broadinstitute.hellbender.tools.spark.transforms.markduplicates.MarkDuplicatesSpark done. Elapsed time: 0.00 minutes. Runtime.totalMemory()=989855744