MongoDB Developer Center
Developer Topics

Integrating MongoDB with Amazon Managed Streaming for Apache Kafka (MSK)

Igor Alekseev, Robert WaltersPublished May 06, 2022 • Updated May 26, 2022
facebook icontwitter iconlinkedin icon
random alt
Rate this tutorial
Amazon Managed Streaming for Apache Kafka
(MSK) is a fully managed, highly available Apache Kafka service. MSK makes it easy to ingest and process streaming data in real time and leverage that data easily within the AWS ecosystem. By being able to quickly stand up a Kafka solution, you spend less time managing infrastructure and more time solving your business problems, dramatically increasing productivity. MSK also supports integration of data sources such as MongoDB via the
AWS MSK Connect
(Connect) service. This Connect service works with the
MongoDB Connector for Apache Kafka
, enabling you to easily integrate MongoDB data.
In this blog post, we will walk through how to set up MSK, configure the MongoDB Connector for Apache Kafka, and create a secured VPC Peered connection with MSK and a MongoDB Atlas cluster. The high-level process is as follows:
  • Configure Amazon Managed Streaming for Apache Kafka
  • Configure EC2 client
  • Configure a MongoDB Atlas Cluster
  • Configure Atlas Private Link including VPC and subnet of the MSK
  • Configure plugin in MSK for MongoDB Connector
  • Create topic on MSK Cluster
  • Install MongoSH command line tool on client
  • Configure MongoDB Connector as a source or sink
In this example, we will have two collections in the same MongoDB cluster—the “source” and the “sink.” We will insert sample data into the source collection from the client, and this data will be consumed by MSK via the MongoDB Connector for Apache Kafka running as an MSK connector. As data arrives in the MSK topic, another instance of the MongoDB Connector for Apache Kafka will write the data to the MongoDB Atlas cluster “sink” collection. To align with best practices for secure configuration, we will set up an AWS Network Peered connection between the MongoDB Atlas cluster and the VPC containing MSK and the client EC2 instance.
Figure 1. Data flow between AWS MSK and MongoDB Atlas

Configure AWS Managed Service for Kafka

To create an Amazon MSK cluster using the AWS Management Console, sign in to the AWS Management Console, and open the
Amazon MSK console
  • Choose Create cluster and select Quick create.
For Cluster name, enter MongoDBMSKCluster.
For Apache Kafka version, select one that is 2.6.2 or above.
For broker type, select kafla.t3.small.
From the table under All cluster settings, copy the values of the following settings and save them because you will need them later in this blog:
  • VPC
  • Subnets
  • Security groups associated with VPC
Figure 2. Cluster settings showing VPC and Subnets
  • Choose “Create cluster.”

Configure an EC2 client

Next, let's configure an EC2 instance to create a topic. This is where the MongoDB Atlas source collection will write to. This client can also be used to query the MSK topic and monitor the flow of messages from the source to the sink.
To create a client machine, open the Amazon EC2 console at
  • Choose Launch instances.
  • Choose Select to create an instance of Amazon Linux 2 AMI (HVM) - Kernel 5.10, SSD Volume Type.
  • Choose the t2.micro instance type by selecting the check box.
  • Choose Next: Configure Instance Details.
  • Navigate to the Network list and choose the VPC whose ID you saved in the previous step.
  • Go to Auto-assign Public IP list and choose Enable.
  • In the menu near the top, select Add Tags.
  • Enter Name for the Key and MongoDBMSKCluster for the Value.
  • Choose Review and Launch, and then click Launch.
  • Choose Create a new key pair, enter MongoDBMSKKeyPair for Key pair name, and then choose Download Key Pair. Alternatively, you can use an existing key pair if you prefer.
  • Start the new instance by pressing Launch Instances.
Next, we will need to configure the networking to allow connectivity between the client instance and the MSK cluster.
  • Select View Instances. Then, in the Security Groups column, choose the security group that is associated with the MSKTutorialClient instance.
  • Copy the name of the security group, and save it for later.
  • Open the Amazon VPC console at
  • In the navigation pane, click on Security Groups. Find the security group whose ID you saved in Step 1 (Create an Amazon MSK Cluster). Choose this row by selecting the check box in the first column.
  • In the Inbound Rules tab, choose Edit inbound rules.
  • Choose Add rule.
  • In the new rule, choose All traffic in the Type column. In the second field in the Source column, select the security group of the client machine. This is the group whose name you saved earlier in this step.
  • Click Save rules.
The cluster's security group can now accept traffic that comes from the client machine's security group.

Create MongoDB Atlas Cluster

To create a MongoDB Atlas Cluster, follow the Getting Started with Atlas tutorial. Note that in this blog, you will need to create an M30 Atlas cluster or above—as VPC peering is not available for M0, M2, and M5 clusters.
Figure 3: Create new cluster dialog
Once the cluster is created, configure an AWS private endpoint in the Atlas Network Access UI supplying the same subnets and VPC.
  • Click on Network Access.
  • Click on Private Endpoint, and then the Add Private Endpoint button.
Figure 4. Add private endpoint
  • Fill out the VPC and subnet IDs from the previous section.
Figure 5. Create private endpoint
  • SSH into the client machine created earlier and issue the following command in the Atlas portal: **aws ec2 create-vpc-endpoint **
    • Note that you may have to first configure the AWS CLI command using aws configure before you can create the VPC through this tool. See
      Configuration Basics
      for more information.

Configure MSK plugin

Next, we need to create a custom plugin for MSK. This custom plugin will be the MongoDB Connector for Apache Kafka. For reference, note that the connector will need to be uploaded to an S3 repository before you can create the plugin. You can download the MongoDB Connector for Apache Kafka from
Confluent Hub
  • Select “Create custom plugin” from the Custom Plugins menu within MSK.
  • Fill out the custom plugin form, including the S3 location of the downloaded connector, and click “Create custom plugin.”
Figure 6. Create custom plugin

Create topic on MSK cluster

When we start reading the data from MongoDB, we also need to create a topic in MSK to accept the data. On the client EC2 instance, let’s install Apache Kafka, which includes some basic tools.
To begin, run the following command to install Java:
sudo yum install java-1.8.0
Next, run the command below to download Apache Kafka.
Building off the previous step, run this command in the directory where you downloaded the TAR file:
tar -xzf kafka_2.12-2.6.2.tgz
The distribution of Kafka includes a bin folder with tools that can be used to manage topics. Go to the kafka_2.12-2.6.2 directory.
To create the topic that will be used to write MongoDB events, issue this command:
bin/ --create --zookeeper (INSERT YOUR ZOOKEEPER INFO HERE)--replication-factor 1 --partitions 1 --topic MongoDBMSKDemo.Source
Also, remember that you can copy the Zookeeper server endpoint from the “View Client Information” page on your MSK Cluster. In this example, we are using plaintext.
Figure 7. Apache ZooKeeper connection

Configure source connector

Once the plugin is created, we can create an instance of the MongoDB connector by selecting “Create connector” from the Connectors menu.
Figure 8. Custom plugin dialog
  • Select the MongoDB plug in and click “Next.”
  • Fill out the form as follows:
Conector name: MongoDB Source Connector
Cluster Type: MSK Connector
Select the MSK cluster that was created previously, and select “None” under the authentication drop down menu.
Enter your connector configuration (shown below) in the configuration settings text area.
connector.class=com.mongodb.kafka.connect.MongoSourceConnector database=MongoDBMSKDemo collection=Source tasks.max=1 connection.uri=(MONGODB CONNECTION STRING HERE)
Note: You can find your Atlas connection string by clicking on the Connect button on your Atlas cluster. Select “Private Endpoint” if you have already configured the Private Endpoint above, then press “Choose a connection method.” Next, select “Connect your application” and copy the mongodb+srv connection string.
Figure 9. Access permissions
In the “Access Permissions” section, you will need to create an IAM role with the
required trust policy
Figure 10. Log delivery dialog
Once this is done, click “Next.” The last section will offer you the ability to use logs—which we highly recommend, as it will simplify the troubleshooting process.

Configure sink connector

Now that we have the source connector up and running, let’s configure a sink connector to complete the round trip. Create another instance of the MongoDB connector by selecting “Create connector” from the Connectors menu.
Figure 8. Custom plugin dialog
Select the same plugin that was created previously, and fill out the form as follows:
Connector name: MongoDB Sink Connector Cluster type: MSK Connector Select the MSK cluster that was created previously and select “None” under the authentication drop down menu. Enter your connector configuration (shown below) in the Configuration Settings text area.
connector.class=com.mongodb.kafka.connect.MongoSinkConnector database=MongoDBMSKDemo collection=Sink tasks.max=1 topics=MongoDBMSKDemo.Source connection.uri=(MongoDB Atlas Connection String Gos Here)
In the Access Permissions section, select the IAM role created earlier that has the
required trust policy
. As with the previous connector, be sure to leverage a log service like CloudWatch.
Once the connector is successfully configured, we can test the round trip by writing to the Source collection and seeing the same data in the Sink collection.
We can insert data in one of two ways: either through the intuitive Atlas UI, or with the new MongoSH (mongoshell) command line tool. Using
, you can interact directly with a MongoDB cluster to test queries, perform ad hoc database operations, and more.
For your reference, we’ve added a section on how to use the mongoshell on your client EC2 instance below.

Install MongoDB shell on client

On the client EC2 instance, create a /etc/yum.repos.d/mongodb-org-5.0.repo file by typing:
sudo nano /etc/yum.repos.d/mongodb-org-5.0.repo
Paste in the following:
[mongodb-org-5.0] name=MongoDB Repository baseurl= gpgcheck=1 enabled=1 gpgkey=[](
Next, install the MongoSH shell with this command:
sudo yum install -y mongodb-mongosh
Use the template below to connect to your MongoDB cluster via mongoshell:
mongosh “ (paste in your Atlas connection string here) “
Once connected, type: Use MongoDBMSKDemo db.Source.insertOne({“Testing”:123})
To check the data on the sink collection, use this command: db.Sink.find({})
If you run into any issues, be sure to check the log files. In this example, we used CloudWatch to read the events that were generated from MSK and the MongoDB Connector for Apache Kafka.
Figure 12. Log events


Amazon Managed Streaming for Apache Kafka (
) is a fully managed, secure, and highly available Apache Kafka service that makes it easy to ingest and process streaming data in real time. MSK allows you to import Kafka connectors such as the MongoDB Connector for Apache Kafka. These connectors make working with data sources seamless within MSK. In this article, you learned how to set up MSK, MSK Connect, and the MongoDB Connector for Apache Kafka. You also learned how to set up a MongoDB Atlas cluster and configure it to use AWS network peering. To continue your learning, check out the following resources:

Copy Link
facebook icontwitter iconlinkedin icon
Rate this tutorial
News & Announcements
The 2022 MongoDB Java Developer Survey

Jul 02, 2022
Java - Mapping POJOs

May 12, 2022
Java - Change Streams

May 26, 2022
Implementing Bulk Writes using Spring Data for MongoDB

Aug 11, 2022
Table of Contents