Skip to content

Commit

Permalink
Update README to include DataFrames and zinc.
Browse files Browse the repository at this point in the history
Also cut trailing whitespaces.

Author: Reynold Xin <[email protected]>

Closes apache#6548 from rxin/readme and squashes the following commits:

630efc3 [Reynold Xin] Update README to include DataFrames and zinc.
  • Loading branch information
rxin committed Jun 1, 2015
1 parent a0e46a0 commit 3c01568
Showing 1 changed file with 8 additions and 8 deletions.
16 changes: 8 additions & 8 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@
Spark is a fast and general cluster computing system for Big Data. It provides
high-level APIs in Scala, Java, and Python, and an optimized engine that
supports general computation graphs for data analysis. It also supports a
rich set of higher-level tools including Spark SQL for SQL and structured
data processing, MLlib for machine learning, GraphX for graph processing,
rich set of higher-level tools including Spark SQL for SQL and DataFrames,
MLlib for machine learning, GraphX for graph processing,
and Spark Streaming for stream processing.

<http://spark.apache.org/>
Expand All @@ -22,7 +22,7 @@ This README file only contains basic setup instructions.
Spark is built using [Apache Maven](http://maven.apache.org/).
To build Spark and its example programs, run:

mvn -DskipTests clean package
build/mvn -DskipTests clean package

(You do not need to do this if you downloaded a pre-built package.)
More detailed documentation is available from the project site, at
Expand All @@ -43,7 +43,7 @@ Try the following command, which should return 1000:
Alternatively, if you prefer Python, you can use the Python shell:

./bin/pyspark

And run the following command, which should also return 1000:

>>> sc.parallelize(range(1000)).count()
Expand All @@ -58,9 +58,9 @@ To run one of them, use `./bin/run-example <class> [params]`. For example:
will run the Pi example locally.

You can set the MASTER environment variable when running examples to submit
examples to a cluster. This can be a mesos:// or spark:// URL,
"yarn-cluster" or "yarn-client" to run on YARN, and "local" to run
locally with one thread, or "local[N]" to run locally with N threads. You
examples to a cluster. This can be a mesos:// or spark:// URL,
"yarn-cluster" or "yarn-client" to run on YARN, and "local" to run
locally with one thread, or "local[N]" to run locally with N threads. You
can also use an abbreviated class name if the class is in the `examples`
package. For instance:

Expand All @@ -75,7 +75,7 @@ can be run using:

./dev/run-tests

Please see the guidance on how to
Please see the guidance on how to
[run tests for a module, or individual tests](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools).

## A Note About Hadoop Versions
Expand Down

0 comments on commit 3c01568

Please sign in to comment.