spark-instrumented-optimizer/project
Josh Rosen 38700ea40c [SPARK-10381] Fix mixup of taskAttemptNumber & attemptId in OutputCommitCoordinator
When speculative execution is enabled, consider a scenario where the authorized committer of a particular output partition fails during the OutputCommitter.commitTask() call. In this case, the OutputCommitCoordinator is supposed to release that committer's exclusive lock on committing once that task fails. However, due to a unit mismatch (we used task attempt number in one place and task attempt id in another) the lock will not be released, causing Spark to go into an infinite retry loop.

This bug was masked by the fact that the OutputCommitCoordinator does not have enough end-to-end tests (the current tests use many mocks). Other factors contributing to this bug are the fact that we have many similarly-named identifiers that have different semantics but the same data types (e.g. attemptNumber and taskAttemptId, with inconsistent variable naming which makes them difficult to distinguish).

This patch adds a regression test and fixes this bug by always using task attempt numbers throughout this code.

Author: Josh Rosen <joshrosen@databricks.com>

Closes #8544 from JoshRosen/SPARK-10381.
2015-09-15 17:11:21 -07:00
..
project [SPARK-6750] Upgrade ScalaStyle to 0.7. 2015-04-07 12:37:33 -07:00
build.properties [SPARK-5415] bump sbt to version to 0.13.7 2015-01-28 02:13:06 -08:00
MimaBuild.scala Update version to 1.6.0-SNAPSHOT. 2015-09-15 00:54:20 -07:00
MimaExcludes.scala [SPARK-10381] Fix mixup of taskAttemptNumber & attemptId in OutputCommitCoordinator 2015-09-15 17:11:21 -07:00
plugins.sbt [SPARK-10556] Remove explicit Scala version for sbt project build files 2015-09-11 13:06:14 +01:00
SparkBuild.scala Revert "[SPARK-10300] [BUILD] [TESTS] Add support for test tags in run-tests.py." 2015-09-15 13:03:38 -07:00