Why Spark with MongoDB?
- High-performance, official known as 100x faster, because you can run the full memory, performance is certainly obvious
- Easy to use, support for Java, Python, Scala, SQL and other languages, so that is very simple to build analytic applications
- Construction of unity, supports a variety of data sources, by masking the underlying data Spark RDD differences, the same application may run on a different analysis of data sources;
- A wide range of application scenarios, can support both batch and streaming
MongoDB Spark Connector as the official launch for adapting Spark operating MongoDB data; paper Python for example, describes the use of MongoDB Spark Connector to help you build your first MongoDB-based analytic applications.
MongoDB ready environment
Installation MongoDB Reference Install MongoDB Community Ed