Flintrock's exit codes are carefully chosen it offers options to disable interactive prompts and when appropriate it prints output in YAML, which is both human- and machine-friendly. Most people will use Flintrock interactively from the command line, but Flintrock is also designed to be used as part of an automated pipeline. EBS-optimized volumes on EC2) which makes it easy to create a cluster with predictable performance for Spark performance testing. Performance testingįlintrock exposes many options of its underlying providers (e.g. If you want to play around with Spark, develop a prototype application, run a one-off job, or otherwise just experiment, Flintrock is the fastest way to get you a working Spark cluster.
#Flint rock how to#
If you want to contribute, follow the instructions in our contributing guide on how to install Flintrock. To get the latest release of Flintrock, simply run pip: Only on OS X, but it should run on all POSIX systems.Ī motivated contributor should be able to add
Installationīefore using Flintrock, take a quick look at theĪnd make sure you're OK with their terms.įlintrock requires Python 3.7 or newer, unless you are using one As long as the assigned IAM roleĪllows it, Spark will be able to read and write data to S3 simply by With this approach you don't need to copy around your AWS credentials Typically Hadoop 3.2 or 2.7), even if the version of Hadoop that you're deploying to Of hadoop-aws to the version of Hadoop that Spark was built against (which is As a rule of thumb, you should match the version If you have issues using the package, consult the hadoop-aws troubleshootingĪnd try adjusting the version. Spark-submit -packages :hadoop-aws:3.3.2 my-app.py Call Spark with the hadoop-aws package to enable s3a://.Here if you're using a vanilla configuration. Versions of Hadoop do not have solid implementations of s3a://.įlintrock's default is Hadoop 3.3.2, so you don't need to do anything Make sure Flintrock is configured to use Hadoop/HDFS 2.7+.
Since it is actively developed, supports larger files, and offers The Hadoop project recommends using s3a:// s3a:// isīackwards compatible with s3n:// and replaces both s3n:// and s3://.