- Deploy and Configure Ops Manager Resources >
- Configure an Ops Manager Resource to use Local Mode
Configure an Ops Manager Resource to use Local Mode¶
On this page
Important
Configuring Ops Manager to use Local Mode in Kubernetes is not recommended. Consider configuring Ops Manager to use Remote Mode instead.
In a default configuration, the MongoDB Agents and Backup Daemons access MongoDB installation archives over the Internet from MongoDB, Inc.
You can configure Ops Manager to run in Local Mode with the Kubernetes Operator if the nodes in your Kubernetes cluster don’t have access to the Internet. The Backup Daemons and managed MongoDB resources download installation archives only from a Persistent Volume that you create for the Ops Manager StatefulSet.
This procedure covers uploading installation archives to Ops Manager.
If Ops Manager has no internet access, see Configure Deployment to Have Limited Internet Access.
For compatbility, see MongoDB Enterprise Kubernetes Operator Compatibility. To view all available versions for each image, see Container Images.
Prerequisites¶
Deploy an Ops Manager Resource. The following procedure shows you how to update your Ops Manager Kubernetes object to enable Local Mode.
To avoid downtime when you enable Local Mode, ensure that you set
spec.replicas
to a value greater than1
in your Ops Manager resource definition.If you updated your Ops Manager resource definition to make Ops Manager highly available, apply your changes before you begin this tutorial:
Procedure¶
Configure kubectl
to default to your namespace.¶
If you have not already, run the following command to execute all
kubectl
commands in the namespace you created.
Note
If you are deploying an Ops Manager resource in a multi-Kubernetes-cluster deployment:
- Set the
context
to the name of the central cluster, such as:kubectl config set context "$MDB_CENTRAL_CLUSTER_FULL_NAME"
. - Set the
--namespace
to the same scope that you used for your multi-Kubernetes-cluster deployment, such as:kubectl config --namespace "mongodb"
.
Delete the StatefulSet that manages your Ops Manager Pods.¶
In this tutorial, you update the StatefulSet that manages the Ops Manager Pods in your Kubernetes cluster.
You must first delete the Ops Manager StatefulSet so that Kubernetes can apply the updates that Local Mode requires.
Find the name of your Ops Manager StatefulSet:
The entry in the response that matches the
metadata.name
of yourYour Ops Manager StatefulSet is the entry in the response that matches the
metadata.name
in your Ops Manager resource definition.Delete the Ops Manager StatefulSet:
Warning
Ensure that you include the
--cascade=false
flag when you delete your Ops Manager StatefulSet. If you don’t include this flag, Kubernetes also deletes your Ops Manager Pods.
Copy the fields of this Ops Manager resource.¶
Copy Lines 9-31 of this example to:
- Use the Ops Manager configuration setting
automation.versions.source: local
inspec.configuration
to enable Local Mode. - Define a Persistent Volume for the Ops Manager StatefulSet to store the MongoDB installation archive. MongoDB Agents running in MongoDB database resource containers that you create with the Kubernetes Operator download the installation archives from Ops Manager instead of from the Internet.
Save your Ops Manager config file.¶
Apply changes to your Ops Manager deployment.¶
Invoke the following
kubectl
command on the filename of the Ops Manager resource definition:Kubernetes creates a new Ops Manager StatefulSet when you apply the changes to your Ops Manager resource definition. Before proceeding to the next step, run the following command to ensure that the Ops Manager StatefulSet exists:
The new Ops Manager StatefulSet should show 0 members ready:
In a rolling fashion, delete your old Ops Manager Pods.¶
List the Ops Manager Pods in your Kubernetes cluster:
Delete one Ops Manager Pod:
Kubernetes recreates the Ops Manager Pod you deleted. Continue to get the status of the new Pod until it is ready:
When the new Pod is initializing, the output is similar to the following example:
When the new Pod is ready, the output is similar to the following example:
- Repeat Steps b and c until you’ve deleted all of your Ops Manager Pods and confirmed that all of the new Pods are ready.
Track the status of your Ops Manager instance.¶
To check the status of your Ops Manager resource, invoke the following command:
See Troubleshoot the Kubernetes Operator for information about the resource deployment statuses.
After the Ops Manager resource completes the Reconciling
phase, the
command returns output similar to the following:
Copy the value of the status.opsManager.url
field, which states
the resource’s connection URL. You use this value when you create a
ConfigMap later in the procedure.
Download the MongoDB installation archive to your local machine.¶
The installers that you download depend on the environment to which you deployed the operator:
Note
The following example includes a link that allows you to download the specified version of MongoDB Community Edition. To download any other version of MongoDB Community Edition, visit the MongoDB Community Edition Download Center. To download MongoDB Enterprise Edition, visit the MongoDB Enterprise Download Center.
Download the RHEL installation tarball for the MongoDB Server version
you want the Kubernetes Operator to deploy. For example, to download the
6.0.1
release:
Copy the MongoDB archive to the Ops Manager Persistent Volume.¶
Copy the MongoDB archive for each MongoDB version you intend to deploy to the Ops Manager Persistent Volume.
The commands that you use depend on the environment to which you deployed the Kubernetes Operator:
Note
If you deployed more than one Ops Manager
replica
, copy only the MongoDB
installation tarball
packages to Replica 1
and
beyond.
- Vanilla Kubernetes
- OpenShift
To copy the MongoDB installation archive to the Ops Manager PersistentVolume:
Copy the MongoDB Server installation tarball to the
Ops Manager PersistentVolume. For example, to copy the 6.0.1
release:
To copy the MongoDB installation archive to the
Ops Manager PersistentVolume, copy the MongoDB Server installation tarball
to the
Ops Manager PersistentVolume. For example, to copy the 6.0.1
release:
Deploy a MongoDB Database Resource.¶
- If you have not done so already, complete the following prerequisites:
- Deploy a MongoDB database resource
in the same namespace to which you deployed Ops Manager.
Ensure that you:
- Match the
spec.opsManager.configMapRef.name
of the resource to themetadata.name
of your ConfigMap. - Match the
spec.credentials
of the resource to the name of the secret you created that contains an Ops Manager programmatic API key pair.
- Match the
MongoDB Agents run in MongoDB database resource containers that you create with the Kubernetes Operator. Download the installation archives from Ops Manager instead of downloading them from the Internet.