Docs Menu
Docs Home
MongoDB Ops Manager
/ /

Deploy a Cluster through the API

On this page

  • Prerequisites
  • Examples
  • Variables for Cluster Creation API Resources
  • Prerequisites
  • Procedures
  • Next Steps

This tutorial manipulates the Ops Manager Administration API's automation configuration to deploy a sharded cluster that is owned by another user. The tutorial first creates a new project, then a new user as owner of the project, and then a sharded cluster owned by the new user. You can create a script to automate these procedures for use in routine operations.

To perform these steps, you must have sufficient access to Ops Manager. A user with the Global Owner or Project Owner role has sufficient access.

The procedures install a cluster with two shards. Each shard comprises a three-member replica set. The tutorial installs one mongos and three config servers. Each component of the cluster resides on its own server, requiring a total of 10 hosts.

The tutorial installs the MongoDB Agent on each host.

Ops Manager must have an existing user. If you are deploying the sharded cluster on a fresh install of Ops Manager, you must register the first user.

You must have the URL of the Ops Manager host, as set in the mmsbaseurl setting of the MongoDB Agent configuration file.

Provision ten hosts to serve the components of the sharded cluster. For host requirements, see the Production Notes in the MongoDB manual.

Each host must provide its MongoDB Agent with full networking access to the hostnames and ports of the MongoDB Agents on all the other hosts. Each agent runs the command hostname -f to self-identify its hostname and port and report them to Ops Manager.


To ensure agents can reach each other, provision the hosts using Automation. This installs the MongoDB Agents with correct network access. Use this tutorial to reinstall the Automations on those machines.

As you work with the API, you can view examples on the GitHub example page.

The API resources use one or more of these variables. Replace these variables with your desired values before calling these API resources.

Your public API Key for your API credentials.
Your private API Key for your API credentials.
URL of your Ops Manager instance.
Unique identifier of your project from your project settings.

Use the Ops Manager Administration API to send a projects document to create the new project.

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--header "Content-Type: application/json" \
--request POST "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups?pretty=true" \
--data '
"name": "{GROUP-NAME}"

The API returns a document that includes the project's agentApiKey and id.


Record these values for use in this procedure and in other procedures in this tutorial.


Use the /users endpoint to add a user to the new project.

The body of the request should contain a users JSON document with the user's information.

Set the user's roles.roleName to GROUP_OWNER and the user's roles.groupId set to the new group's' id.

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--header "Content-Type: application/json" \
--request POST "https://<OpsManagerHost>:<Port>/api/public/v1.0/users?pretty=true" \
--data '
"username": "<>",
"emailAddress": "<>",
"firstName": "<First>",
"lastName": "<Last>",
"password": "<password>",
"roles": [{
"groupId": "{PROJECT-ID}",
"roleName": "GROUP_OWNER"

The user you use to create the project is automatically added to the project. If you used a user with the Global Owner role, you can remove the user from the project without losing the ability to make changes to the project in the future. As long as you have the project's agentApiKey and id, you have full access to the project when logged in as the global owner.

GET the global owner's ID. Issue the following command to request the project's users:

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--request GET "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups/{PROJECT-ID}/users?pretty=true"

The API returns a JSON document that lists all the project's users. Locate the user with roles.roleName set to GLOBAL_OWNER. Copy the user's id value, and issue the following to remove the user from the project, replacing {USER-ID} with the user's id value:

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--request GET "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups/{PROJECT-ID}/users/{USER-ID}?pretty=true"

If Ops Manager removes the user successfully, the API returns the HTTP 200 OK status code.


To learn how to install the MongoDB Agent, follow the procedure for the appropriate platform.


When the MongoDB Agent first runs, it downloads the mms-cluster-config-backup.json file, which describes the desired state of the automation configuration.

On one of the hosts, navigate to /var/lib/mongodb-mms-automation/ and open mms-cluster-config-backup.json. Confirm that the file's version field is set to 1. Ops Manager automatically increments this field as changes occur.

To add or update a deployment, retrieve the configuration, make changes as needed, and send the updated configuration though the API to Ops Manager.

The following procedure deploys an updated automation configuration through the API:

  1. Use the automationConfig resource to retrieve the configuration. Issue the following command, replacing the placeholders with the Variables for Cluster Creation API Resources.

    curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
    --request GET "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups/{PROJECT-ID}/automationConfig?pretty=true" \
    --output currentAutomationConfig.json
  2. Validate the downloaded Automation Configuration file.

    Compare the version field of the currentAutomationConfig.json with that of the Automation Configuration backup file, mms-cluster-config-backup.json. The version value is the last element in both JSON documents. You can find this file on any host running the MongoDB Agent at:

    • Linux and macOS: /var/lib/mongodb-mms-automation/mms-cluster-config-backup.json

    • Windows: %SystemDrive%\MMSAutomation\versions\mms-cluster-config-backup.json

    If the version values match, you are working with the current version of the Automation Configuration file.


Create a document with the following fields. As you build the configuration document, refer the description of an automation configuration for detailed explanations of the settings. For examples, see the MongoDB Labs page.

2 "options": {
3 "downloadBase": "/var/lib/mongodb-mms-automation",
4 },
5 "mongoDbVersions": [],
6 "monitoringVersions": [],
7 "backupVersions": [],
8 "processes": [],
9 "replicaSets": [],
10 "sharding": []

In the monitoringVersions.hostname field, enter the hostname of the server where Ops Manager should install the Monitoring. Use the fully qualified domain name that running hostname -f on the server returns, as in the following:

1"monitoringVersions": [
2 {
3 "hostname": "<>",
4 "logPath": "/var/log/mongodb-mms-automation/monitoring-agent.log",
5 "logRotate": {
6 "sizeThresholdMB": 1000,
7 "timeThresholdHrs": 24
8 }
9 }

This configuration example also includes the logPath field, which specifies the log location, and logRotate, which specifies the log thresholds.


This sharded cluster has 10 MongoDB instances, as described in the Deploy a Cluster through the API, each running on its own server. Thus, the automation configuration's processes array will have 10 documents, one for each MongoDB instance.

The following example adds the first document to the processes array. Replace <process_name_1> with any name you choose, and replace <> with the FQDN of the host.

Add 9 documents: one for each MongoDB instance in your sharded cluster.

Specify the args2_6 syntax for the processes.<args> field. The processes.args2_6 object accepts most MongoDB settings and parameters for MongoDB versions 2.6 and later. To learn more, see MongoDB Settings and Automation Support.

1"processes": [
2 {
3 "version": "4.0.6",
4 "name": "<process_name_1>",
5 "hostname": "<>",
6 "logRotate": {
7 "sizeThresholdMB": 1000,
8 "timeThresholdHrs": 24
9 },
10 "authSchemaVersion": 5,
11 "featureCompatibilityVersion": "4.0",
12 "processType": "mongod",
13 "args2_6": {
14 "net": {
15 "port": 27017
16 },
17 "storage": {
18 "dbPath": "/data/"
19 },
20 "systemLog": {
21 "path": "/data/mongodb.log",
22 "destination": "file"
23 },
24 "replication": {
25 "replSetName": "rs1"
26 }
27 }
28 },

Add two replica set documents to the replicaSets array. Add three members to each document.


This section adds one replica set member to the first replica set document:


You must include "protocolVersion": 1 in the root document for each replica set.

1"replicaSets": [
2 {
3 "_id": "rs1",
4 "members": [
5 {
6 "_id": 0,
7 "host": "<process_name_1>",
8 "priority": 1,
9 "votes": 1,
10 "secondaryDelaySecs": 0,
11 "hidden": false,
12 "arbiterOnly": false
13 }
14 ],
15 "protocolVersion": 1
16 }

In the sharding array, add the replica sets to the shards, and add the config server replica set name, as in the following:

1"sharding": [
2 {
3 "shards": [
4 {
5 "tags": [],
6 "_id": "shard1",
7 "rs": "rs1"
8 },
9 {
10 "tags": [],
11 "_id": "shard2",
12 "rs": "rs2"
13 }
14 ],
15 "name": "sharded_cluster_via_api",
16 "configServerReplica": "rs-config"
17 }

Use the automationConfig resource to send the updated automation configuration.

Issue the following command with path to the updated configuration document and replace the placeholders with the Variables for Cluster Creation API Resources.

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--header "Content-Type: application/json"
--request PUT "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups/{PROJECT-ID}/automationConfig?pretty=true" \
--data @currentAutomationConfig.json

Upon successful update of the configuration, the API returns the HTTP 200 OK status code to indicate the request has succeeded.


Retrieve the automation configuration from Ops Manager and confirm it contains the changes. To retrieve the configuration, issue the following command, replacing the placeholders with the Variables for Cluster Creation API Resources.

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--request GET "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups/{PROJECT-ID}/automationConfig?pretty=true"

Use the automationStatus resource to verify the configuration update is fully deployed. Issue the following command:

curl --user "{PUBLIC-KEY}:{PRIVATE-KEY}" --digest \
--request GET "https://<OpsManagerHost>:<Port>/api/public/v1.0/groups/{PROJECT-ID}/automationStatus?pretty=true"

The curl command returns a JSON object containing the processes array and the goalVersion key and value. The processes array contains a document for each server that hosts a MongoDB instance. The new configuration is successfully deployed when all lastGoalVersionAchieved fields in the processes array equal the value specified for goalVersion.


In this response, processes[2].lastGoalVersionAchieved is behind goalVersion. This indicates that the MongoDB instance at is running one version behind the goalVersion. Wait several seconds and issue the curl command again.

2 "goalVersion": 2,
3 "processes": [{
4 "hostname": "",
5 "lastGoalVersionAchieved": 2,
6 "name": "ReplSet_0",
7 "plan": []
8 }, {
9 "hostname": "",
10 "lastGoalVersionAchieved": 2,
11 "name": "ReplSet_1",
12 "plan": []
13 }, {
14 "hostname": "",
15 "lastGoalVersionAchieved": 1,
16 "name": "ReplSet_2",
17 "plan":[]
18 }]

To view the new configuration in the Ops Manager console, click Deployment.

To make an additional version of MongoDB available in the cluster, see Update the MongoDB Version of a Deployment.

← Ops Manager Administration API Tutorials