Apache Spark community uses various resources to maintain the community test coverage.
GitHub Actions provides the following on Ubuntu 22.04.
Spark’s default build strategy is to assemble a jar including all of its dependencies. This can be cumbersome when doing iterative development. When developing locally, it is possible to create an assembly jar including all of Spark’s dependencies and then re-package only Spark itself when making changes.
$ build/sbt clean package$ ./bin/spark-shell$ export SPARK_PREPEND_CLASSES=true$ ./bin/spark-shell # Now it's using compiled classes# ... do some local development ... #$ build/sbt compile$ unset SPARK_PREPEND_CLASSES$ ./bin/spark-shell # You can also use ~ to let sbt do incremental builds on file changes without running a new sbt session every time$ build/sbt ~compile
For instance, you can build the Spark Core module using:
$ # sbt$ build/sbt> project core> package$ # or you can build the spark-core module with sbt directly using:$ build/sbt core/package$ # Maven$ build/mvn package -DskipTests -pl core
When developing locally, it’s often convenient to run a single test or a few tests, rather than running the entire test suite.
The fastest way to run individual tests is to use thesbt
console. It’s fastest to keep asbt
console open, and use it to re-run tests as necessary. For example, to run all of the tests in a particular project, e.g.,core
:
$ build/sbt> project core> test
You can run a single test suite using thetestOnly
command. For example, to run the DAGSchedulerSuite:
> testOnly org.apache.spark.scheduler.DAGSchedulerSuite
ThetestOnly
command accepts wildcards; e.g., you can also run theDAGSchedulerSuite
with:
> testOnly *DAGSchedulerSuite
Or you could run all of the tests in the scheduler package:
> testOnly org.apache.spark.scheduler.*
If you’d like to run just a single test in theDAGSchedulerSuite
, e.g., a test that includes “SPARK-12345” in the name, you run the following command in the sbt console:
> testOnly *DAGSchedulerSuite -- -z "SPARK-12345"
If you’d prefer, you can run all of these commands on the command line (but this will be slower than running tests using an open console). To do this, you need to surroundtestOnly
and the following arguments in quotes:
$ build/sbt "core/testOnly *DAGSchedulerSuite -- -z SPARK-12345"
For more about how to run individual tests with sbt, see thesbt documentation.
With Maven, you can use the-DwildcardSuites
flag to run individual Scala tests:
build/mvn -Dtest=none -DwildcardSuites=org.apache.spark.scheduler.DAGSchedulerSuite test
You need-Dtest=none
to avoid running the Java tests. For more information about the ScalaTest Maven Plugin, refer to theScalaTest documentation.
To run individual Java tests, you can use the-Dtest
flag:
build/mvn test -DwildcardSuites=none -Dtest=org.apache.spark.streaming.JavaAPISuite test
To run individual PySpark tests, you can userun-tests
script underpython
directory. Test cases are located attests
package under each PySpark packages.Note that, if you add some changes into Scala or Python side in Apache Spark, you need to manually build Apache Spark again before running PySpark tests in order to apply the changes.Running PySpark testing script does not automatically build it.
Also, note that there is an ongoing issue to use PySpark on macOS High Serria+.OBJC_DISABLE_INITIALIZE_FORK_SAFETY
should be set toYES
in order to run some of tests.SeePySpark issue andPython issue for more details.
To run test cases in a specific module:
$ python/run-tests --testnames pyspark.sql.tests.test_arrow
To run test cases in a specific class:
$ python/run-tests --testnames 'pyspark.sql.tests.test_arrow ArrowTests'
To run single test case in a specific class:
$ python/run-tests --testnames 'pyspark.sql.tests.test_arrow ArrowTests.test_null_conversion'
You can also run doctests in a specific module:
$ python/run-tests --testnames pyspark.sql.dataframe
Lastly, there is another script calledrun-tests-with-coverage
in the same location, which generates coverage report for PySpark tests. It accepts same arguments withrun-tests
.
$ python/run-tests-with-coverage --testnames pyspark.sql.tests.test_arrow --python-executables=python...Name Stmts Miss Branch BrPart Cover-------------------------------------------------------------------pyspark/__init__.py 42 4 8 2 84%pyspark/_globals.py 16 3 4 2 75%...Generating HTML files for PySpark coverage under /.../spark/python/test_coverage/htmlcov
You can check the coverage report visually by HTMLs under/.../spark/python/test_coverage/htmlcov
.
Please check other available options viapython/run-tests[-with-coverage] --help
.
Although GitHub Actions provide both K8s unit test and integration test coverage, you can run it locally. For example, Volcano batch scheduler integration test should be done manually. Please refer the integration test documentation for the detail.
https://github.com/apache/spark/blob/master/resource-managers/kubernetes/integration-tests/README.md
Docker integration tests are covered by GitHub Actions. However, you can run it locally to speedup deveplopment and testing.Please refer theDocker integration test documentation for the detail.
Apache Spark leverages GitHub Actions that enables continuous integration and a wide range of automation. Apache Spark repository provides several GitHub Actions workflows for developers to run before creating a pull request.
Apache Spark repository provides an easy way to run benchmarks in GitHub Actions. When you update the benchmark results in a pull request, it is recommended to use GitHub Actions to run and generate the benchmark results in order to run them on the environment as same as possible.
org.apache.spark.sql.*
.11
.true
, it fails right away. Whenfalse
, it runs all whether it fails or not.If the following error occurs when running ScalaTest
An internal error occurred during: "Launching XYZSuite.scala".java.lang.NullPointerException
It is due to an incorrect Scala library in the classpath. To fix it:
Build Path | Configure Build Path
Add Library | Scala Library
scala-library-2.10.4.jar - lib_managed\jars
In the event of “Could not find resource path for Web UI: org/apache/spark/ui/static”, it’s due to a classpath issue (some classes were probably not compiled). To fix this, it sufficient to run a test from the command line:
build/sbt "testOnly org.apache.spark.rdd.SortingSuite"
To ensure binary compatibility, Spark usesMiMa.
When working on an issue, it’s always a good idea to check that your changes donot introduce binary incompatibilities before opening a pull request.
You can do so by running the following command:
$ dev/mima
A binary incompatibility reported by MiMa might look like the following:
[error] method this(org.apache.spark.sql.Dataset)Unit in class org.apache.spark.SomeClass does not have a correspondent in current version[error] filter with: ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.SomeClass.this")
If you believe that your binary incompatibilies are justified or that MiMareported false positives (e.g. the reported binary incompatibilities are about anon-user facing API), you can filter them out by adding an exclusion inproject/MimaExcludes.scalacontaining what was suggested by the MiMa report and a comment containing theJIRA number of the issue you’re working on as well as its title.
For the problem described above, we might add the following:
// [SPARK-zz][CORE] Fix an issueProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.SomeClass.this")
Otherwise, you will have to resolve those incompatibilies before opening orupdating your pull request. Usually, the problems reported by MiMa areself-explanatory and revolve around missing members (methods or fields) thatyou will have to add back in order to maintain binary compatibility.
Git provides a mechanism for fetching remote pull requests into your own local repository. This is useful when reviewing code or testing patches locally. If you haven’t yet cloned the Spark Git repository, use the following command:
$ git clone https://github.com/apache/spark.git$ cd spark
To enable this feature you’ll need to configure the git remote repository to fetch pull request data. Do this by modifying the.git/config
file inside of your Spark directory. The remote may not be named “origin” if you’ve named it something else:
[remote "origin"] url = git@github.com:apache/spark.git fetch = +refs/heads/*:refs/remotes/origin/* fetch = +refs/pull/*/head:refs/remotes/origin/pr/* # Add this line
Once you’ve done this you can fetch remote pull requests
# Fetch remote pull requests$ git fetch origin# Checkout a remote pull request$ git checkout origin/pr/112# Create a local branch from a remote pull request$ git checkout origin/pr/112 -b new-branch
$ # sbt$ build/sbt dependencyTree $ # Maven$ build/mvn -DskipTests install$ build/mvn dependency:tree
You can use aIntelliJ Imports Organizer from Aaron Davidson to help you organize the imports in your code. It can be configured to match the import ordering from the style guide.
To format Scala code, run the following command prior to submitting a PR:
$ ./dev/scalafmt
By default, this script will format files that differ from git master. For more information, seescalafmt documentation, but use the existing script not a locally installed version of scalafmt.
While many of the Spark developers use SBT or Maven on the command line, the most common IDE we use is IntelliJ IDEA. You can get the community edition for free (Apache committers can get free IntelliJ Ultimate Edition licenses) and install the JetBrains Scala plugin fromPreferences > Plugins
.
To create a Spark project for IntelliJ:
File -> Import Project
, locate the spark source directory, and select “Maven Project”.-P[profile name]
above may be enabled on the Profiles screen in the Import wizard. For example, if developing for Hadoop 2.7 with YARN support, enable profilesyarn
andhadoop-2.7
. These selections can be changed later by accessing the “Maven Projects” tool window from the View menu, and expanding the Profiles section.Other tips:
Preference -> Build, Execution, Deployment -> Maven -> Maven home directory
) of your project to point to a newer installation of Maven. You may also build Spark with the scriptbuild/mvn
first.If the script cannot locate a new enough Maven installation, it will download and install a recent version of Maven to folderbuild/apache-maven-<version>/
./Users/irashid/github/spark/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodeGenerator.scalaError:(147, 9) value q is not a member of StringContext Note: implicit class Evaluate2 is not applicable here because it comes after the application point and it lacks an explicit result type q""" ^
This part will show you how to debug Spark remotely with IntelliJ.
FollowRun > Edit Configurations > + > Remote to open a default Remote Configuration template:
Normally, the default values should be good enough to use. Make sure that you chooseListen to remote JVMasDebugger mode and select the right JDK version to generate properCommand line arguments for remote JVM.
Once you finish configuration and save it. You can followRun > Run > Your_Remote_Debug_Name > Debug to start remote debugprocess and wait for SBT console to connect:
In general, there are 2 steps:
The following is an example of how to trigger the remote debugging using SBT unit tests.
Enter in SBT console
./build/sbt
Switch to project where the target test locates, e.g.:
sbt > project core
Copy pasting theCommand line arguments for remote JVM
sbt > set javaOptions in Test += "-agentlib:jdwp=transport=dt_socket,server=n,suspend=n,address=localhost:5005"
Set breakpoints with IntelliJ and run the test with SBT, e.g.:
sbt > testOnly *SparkContextSuite -- -t "Only one SparkContext may be active at a time"
It should be successfully connected to IntelliJ when you see “Connected to the target VM, address: ‘localhost:5005’, transport: ‘socket’” in IntelliJ console. And then, you can startdebug in IntelliJ as usual.
To exit remote debug mode (so that you don’t have to keep starting the remote debugger),type “session clear” in SBT console while you’re in a project.
Eclipse can be used to develop and test Spark. The following configuration is known to work:
The easiest way is to download the Scala IDE bundle from the Scala IDE download page. It comes pre-installed with ScalaTest. Alternatively, use the Scala IDE update site or Eclipse Marketplace.
SBT can create Eclipse.project
and.classpath
files. To create these files for each Spark sub project, use this command:
sbt/sbt eclipse
To import a specific project, e.g. spark-core, selectFile | Import | Existing Projects
into Workspace. Do not select “Copy projects into workspace”.
If you want to develop on Scala 2.10 you need to configure a Scala installation for the exact Scala version that’s used to compile Spark. Since Scala IDE bundles the latest versions (2.10.5 and 2.11.8 at this point), you need to add one inEclipse Preferences -> Scala -> Installations
by pointing to thelib/
directory of your Scala 2.10.5 distribution. Once this is done, select all Spark projects and right-click, chooseScala -> Set Scala Installation
and point to the 2.10.5 installation. This should clear all errors about invalid cross-compiled libraries. A clean build should succeed now.
ScalaTest can execute unit tests by right clicking a source file and selectingRun As | Scala Test
.
If Java memory errors occur, it might be necessary to increase the settings ineclipse.ini
in the Eclipse install directory. Increase the following setting as needed:
--launcher.XXMaxPermSize256M
Spark publishes SNAPSHOT releases of its Maven artifacts for both master and maintenance branches on a nightly basis. To link to a SNAPSHOT you need to add the ASF snapshot repository to your build. Note that SNAPSHOT artifacts are ephemeral and may change orbe removed. To use these you must add the ASF snapshot repository athttps://repository.apache.org/snapshots/.
groupId: org.apache.sparkartifactId: spark-core_2.12version: 3.0.0-SNAPSHOT
Here are instructions on profiling Spark applications using YourKit Java Profiler.
/root
in our case):unzip YourKit-JavaProfiler-2017.02-b66.zip
~/spark-ec2/copy-dir /root/YourKit-JavaProfiler-2017.02
~/spark/conf/spark-env.sh
and adding the linesSPARK_DAEMON_JAVA_OPTS+=" -agentpath:/root/YourKit-JavaProfiler-2017.02/bin/linux-x86-64/libyjpagent.so=sampling"export SPARK_DAEMON_JAVA_OPTSSPARK_EXECUTOR_OPTS+=" -agentpath:/root/YourKit-JavaProfiler-2017.02/bin/linux-x86-64/libyjpagent.so=sampling"export SPARK_EXECUTOR_OPTS
~/spark-ec2/copy-dir ~/spark/conf/spark-env.sh
~/spark/bin/stop-all.sh
and~/spark/bin/start-all.sh
10001-10010
. To connect the YourKit desktopapplication to the remote profiler agents, you’ll have to open these ports in the cluster’s EC2 security groups. To do this, sign into the AWS Management Console. Go to the EC2 section and selectSecurity Groups
from theNetwork & Security
section on the left side of the page. Find the security groups corresponding to your cluster; if you launched a cluster namedtest_cluster
, then you will want to modify the settings for thetest_cluster-slaves
andtest_cluster-master
security groups. For each group, select it from the list, click theInbound
tab, and create a newCustom TCP Rule
opening the port range10001-10010
. Finally, clickApply Rule Changes
. Make sure to do this for both security groups.Note: by default,spark-ec2
re-uses security groups: if you stop this cluster and launch another cluster with the same name, your security group settings will be re-used.ec2--.compute-1.amazonaws.com
Please see the full YourKit documentation for the full list of profiler agentstartup options.
When running Spark tests through SBT, addjavaOptions in Test += "-agentpath:/path/to/yjp"
toSparkBuild.scala
to launch the tests with the YourKit profiler agent enabled.
The platform-specific paths to the profiler agents are listed in theYourKit documentation.
In general, the ASF allows contributions co-authored using generative AI tools. However, there are several considerations when you submit a patch containing generated content.
Foremost, you are required to disclose usage of such tool. Furthermore, you are responsible for ensuring that the terms and conditions of the tool in question arecompatible with usage in an Open Source project and inclusion of the generated content doesn’t pose a risk of copyright violation.
Please refer toThe ASF Generative Tooling Guidance for details and developments.