On this page
Spark Structured Streaming is a data stream processing engine you can use through the Dataset or DataFrame API. The MongoDB Spark Connector enables you to stream to and from MongoDB using Spark Structured Streaming.
To learn more about Structured Streaming, see the Spark Programming Guide.
When reading a stream from a MongoDB database, the MongoDB Spark Connector supports both micro-batch processing and continuous processing. Micro-batch processing is the default processing engine, while continuous processing is an experimental feature introduced in Spark version 2.3. To learn more about continuous processing, see the Spark documentation.
The connector reads from your MongoDB deployment's change stream. To generate change events on the change stream, perform update operations on your database.
To learn more about change streams, see Change Streams in the MongoDB manual.
The following examples show Spark Structured Streaming configurations for streaming to and from MongoDB.
To stream data from a CSV file to MongoDB:
To stream data from MongoDB to your console: