MongoDB with drivers
This page documents a mongosh method. To see the equivalent
method in a MongoDB driver, see the corresponding page for your
programming language:
Definition
db.collection.aggregate(pipeline, options)Calculates aggregate values for the data in a collection or a view.
Returns: - A cursor for the documents produced by the final stage of the aggregation pipeline.
- If the pipeline includes the
explainoption, the query returns a document that provides details on the processing of the aggregation operation. - If the pipeline includes the
$outor$mergeoperators, the query returns an empty cursor.
Compatibility
You can use db.collection.aggregate() for deployments hosted in the following
environments:
MongoDB Atlas: The fully managed service for MongoDB deployments in the cloud
MongoDB Enterprise: The subscription-based, self-managed version of MongoDB
MongoDB Community: The source-available, free-to-use, and self-managed version of MongoDB
Syntax
The aggregate() method has the following
form:
db.collection.aggregate( <pipeline>, <options> )
The aggregate() method takes the following
parameters:
Parameter | Type | Description |
|---|---|---|
| array | A sequence of data aggregation operations or stages. See the aggregation pipeline operators for details. The method can still accept the pipeline stages as separate
arguments instead of as elements in an array; however, if you do
not specify the |
| document | Optional. Additional options that |
Behavior
Error Handling
If an error occurs, the aggregate() helper
throws an exception.
Cursor Behavior
In mongosh, if the cursor returned from the
db.collection.aggregate() is not assigned to a variable using
the var keyword, then mongosh automatically
iterates the cursor up to 20 times. See
Iterate a Cursor in mongosh for handling cursors in
mongosh.
Cursors returned from aggregation only supports cursor methods that operate on evaluated cursors (i.e. cursors whose first batch has been retrieved), such as the following methods:
For more information, see:
Sessions
For cursors created inside a session, you cannot call
getMore outside the session.
Similarly, for cursors created outside of a session, you cannot call
getMore inside a session.
Session Idle Timeout
MongoDB drivers and mongosh
associate all operations with a server session, with the exception of unacknowledged
write operations. For operations not explicitly associated with a
session (i.e. using Mongo.startSession()), MongoDB drivers
and mongosh create an implicit session and associate it
with the operation.
If a session is idle for longer than 30 minutes, the MongoDB server
marks that session as expired and may close it at any time. When the
MongoDB server closes the session, it also kills any in-progress
operations and open cursors associated with the session. This
includes cursors configured with noCursorTimeout() or
a maxTimeMS() greater than 30 minutes.
For operations that return a cursor, if the cursor may be idle for
longer than 30 minutes, issue the operation within an explicit session
using Mongo.startSession() and periodically refresh the
session using the refreshSessions command. See
Session Idle Timeout for more information.
Transactions
db.collection.aggregate() can be used inside distributed transactions.
However, the following stages are not allowed within transactions:
You also cannot specify the explain option.
For cursors created outside of a transaction, you cannot call
getMoreinside the transaction.For cursors created in a transaction, you cannot call
getMoreoutside the transaction.
Important
In most cases, a distributed transaction incurs a greater performance cost over single document writes, and the availability of distributed transactions should not be a replacement for effective schema design. For many scenarios, the denormalized data model (embedded documents and arrays) will continue to be optimal for your data and use cases. That is, for many scenarios, modeling your data appropriately will minimize the need for distributed transactions.
For additional transactions usage considerations (such as runtime limit and oplog size limit), see also Production Considerations.
Client Disconnection
For db.collection.aggregate() operation that do not include
the $out or $merge stages:
If the client that issued db.collection.aggregate() disconnects before the operation
completes, MongoDB marks db.collection.aggregate() for termination using
killOp.
Query Settings
New in version 8.0.
You can use query settings to set index hints, set operation rejection filters, and other fields. The settings apply to the query shape on the entire cluster. The cluster retains the settings after shutdown.
The query optimizer uses the query settings as an additional input during query planning, which affects the plan selected to run the query. You can also use query settings to block a query shape.
To add query settings and explore examples, see
setQuerySettings.
You can add query settings for find, distinct,
and aggregate commands.
Query settings have more functionality and are preferred over deprecated index filters.
To remove query settings, use removeQuerySettings. To
obtain the query settings, use a $querySettings stage in an
aggregation pipeline.
Examples
The following examples use the collection orders that contains the
following documents:
db.orders.insertMany( [ { _id: 1, cust_id: "abc1", ord_date: ISODate("2012-11-02T17:04:11.102Z"), status: "A", amount: 50 }, { _id: 2, cust_id: "xyz1", ord_date: ISODate("2013-10-01T17:04:11.102Z"), status: "A", amount: 100 }, { _id: 3, cust_id: "xyz1", ord_date: ISODate("2013-10-12T17:04:11.102Z"), status: "D", amount: 25 }, { _id: 4, cust_id: "xyz1", ord_date: ISODate("2013-10-11T17:04:11.102Z"), status: "D", amount: 125 }, { _id: 5, cust_id: "abc1", ord_date: ISODate("2013-11-12T17:04:11.102Z"), status: "A", amount: 25 } ] )
Group by and Calculate a Sum
The following aggregation operation selects documents with status equal
to "A", groups the matching documents by the cust_id field and
calculates the total for each cust_id field from the sum of the
amount field, and sorts the results by the total field in
descending order:
db.orders.aggregate( [ { $match: { status: "A" } }, { $group: { _id: "$cust_id", total: { $sum: "$amount" } } }, { $sort: { total: -1 } } ] )
The operation returns a cursor with the following documents:
[ { _id: "xyz1", total: 100 }, { _id: "abc1", total: 75 } ]
mongosh iterates the returned cursor automatically
to print the results. See Iterate a Cursor in mongosh for
handling cursors manually in mongosh.
Return Information on Aggregation Pipeline Operation
The following example uses db.collection.explain() to view
detailed information regarding the execution plan of the aggregation
pipeline.
db.orders.explain().aggregate( [ { $match: { status: "A" } }, { $group: { _id: "$cust_id", total: { $sum: "$amount" } } }, { $sort: { total: -1 } } ] )
The operation returns a document that details the processing of the
aggregation pipeline. For example, the document may show, among other
details, which index, if any, the operation used. [1]
If the orders collection is a sharded collection, the document
also shows the division of labor between the shards and the merge
operation, and for targeted queries, the targeted shards.
Note
The intended readers of the explain output document are humans, and
not machines, and the output format is subject to change between
releases.
You can view more verbose explain output by passing the
executionStats or allPlansExecution explain modes to the
db.collection.explain() method.
| [1] | Index Filters can affect the choice of index used. See Index Filters for details. |
Interaction with allowDiskUseByDefault
Starting in MongoDB 6.0, pipeline stages that require more than 100
megabytes of memory to execute write temporary files to disk by
default. These temporary files last for the duration of the pipeline
execution and can influence storage space on your instance. In earlier
versions of MongoDB, you must pass { allowDiskUse: true } to
individual find and aggregate commands to enable this
behavior.
Individual find and aggregate commands can override the
allowDiskUseByDefault parameter by either:
Using
{ allowDiskUse: true }to allow writing temporary files out to disk whenallowDiskUseByDefaultis set tofalseUsing
{ allowDiskUse: false }to prohibit writing temporary files out to disk whenallowDiskUseByDefaultis set totrue
The profiler log messages and diagnostic log
messages includes a usedDisk
indicator if any aggregation stage wrote data to temporary files due
to memory restrictions.
For more information, see Aggregation Pipeline Limits.
Specify an Initial Batch Size
To specify an initial batch size for the cursor, use the following
syntax for the cursor option:
cursor: { batchSize: <int> }
For example, the following aggregation operation specifies the
initial batch size of 0 for the cursor:
db.orders.aggregate( [ { $match: { status: "A" } }, { $group: { _id: "$cust_id", total: { $sum: "$amount" } } }, { $sort: { total: -1 } }, { $limit: 2 } ], { cursor: { batchSize: 0 } } )
The { cursor: { batchSize: 0 } } document, which specifies the size of the
initial batch size, indicates an empty first batch. This batch size is useful
for quickly returning a cursor or failure message without doing significant
server-side work.
To specify batch size for subsequent getMore operations
(after the initial batch), use the batchSize field when running the
getMore command.
mongosh iterates the returned cursor automatically
to print the results. See Iterate a Cursor in mongosh for
handling cursors manually in mongosh.
Specify a Collation
Collation allows users to specify language-specific rules for string comparison, such as rules for lettercase and accent marks.
A restaurants collection has the following documents:
db.restaurants.insertMany( [ { _id: 1, category: "café", status: "Open" }, { _id: 2, category: "cafe", status: "open" }, { _id: 3, category: "cafE", status: "open" } ] )
The following aggregation operation includes the collation option:
db.restaurants.aggregate( [ { $match: { status: "Open" } }, { $group: { _id: "$category", count: { $sum: 1 } } } ], { collation: { locale: "fr", strength: 1 } } )
[ { _id: 'café', count: 3 } ]
The aggregation specifies a collation with strength: 1, which
means the collation ignores differences between case and letter
variants. As a result:
The
$matchfilter matches all documents in the collection and passes those documents to the$groupstage.The
$groupstage ignores differences in letter variants in thecategoryfield and puts all documents in a single group.
Note
If performing an aggregation that involves multiple views, such as
with $lookup or $graphLookup, the views must
have the same collation.
For descriptions on the collation fields, see Collation Document.
Hint an Index
Create a collection food with the following documents:
db.food.insertMany( [ { _id: 1, category: "cake", type: "chocolate", qty: 10 }, { _id: 2, category: "cake", type: "ice cream", qty: 25 }, { _id: 3, category: "pie", type: "boston cream", qty: 20 }, { _id: 4, category: "pie", type: "blueberry", qty: 15 } ] )
Create the following indexes:
db.food.createIndex( { qty: 1, type: 1 } ); db.food.createIndex( { qty: 1, category: 1 } );
The following aggregation operation includes the hint option to
force the usage of the specified index:
db.food.aggregate( [ { $sort: { qty: 1 }}, { $match: { category: "cake", qty: 10 } }, { $sort: { type: -1 } } ], { hint: { qty: 1, category: 1 } } )
Override readConcern
Use the readConcern option to specify the read concern for
the operation.
You cannot use the $out or the $merge stage
in conjunction with read concern "linearizable". That
is, if you specify "linearizable" read concern for
db.collection.aggregate(), you cannot include either
stages in the pipeline.
The following operation on a replica set specifies a
Read Concern of "majority" to read the
most recent copy of the data confirmed as having been written to a
majority of the nodes.
Note
To ensure that a single thread can read its own writes, use
"majority"read concern and"majority"write concern against the primary of the replica set.You can specify read concern level
"majority"for an aggregation that includes an$outstage.Regardless of the read concern level, the most recent data on a node may not reflect the most recent version of the data in the system.
db.restaurants.aggregate( [ { $match: { rating: { $lt: 5 } } } ], { readConcern: { level: "majority" } } )
Specify a Comment
A collection named movies contains documents formatted as such:
db.movies.insertOne( { _id: ObjectId("599b3b54b8ffff5d1cd323d8"), title: "Jaws", year: 1975, imdb: "tt0073195" } )
The following aggregation operation finds movies created in 1995 and includes
the comment option to provide tracking information in the logs,
the db.system.profile collection, and db.currentOp.
db.movies.aggregate( [ { $match: { year : 1995 } } ], { comment : "match_all_movies_from_1995" } ).pretty()
On a system with profiling enabled, you can then query the system.profile
collection to see all recent similar aggregations, as shown below:
db.system.profile.find( { "command.aggregate": "movies", "command.comment" : "match_all_movies_from_1995" } ).sort( { ts : -1 } ).pretty()
This will return a set of profiler results in the following format:
{ "op" : "command", "ns" : "video.movies", "command" : { "aggregate" : "movies", "pipeline" : [ { "$match" : { "year" : 1995 } } ], "comment" : "match_all_movies_from_1995", "cursor" : { }, "$db" : "video" }, ... }
An application can encode any arbitrary information in the comment in order to more easily trace or identify specific operations through the system. For instance, an application might attach a string comment incorporating its process ID, thread ID, client hostname, and the user who issued the command.
Use Variables in let
New in version 5.0.
To define variables that you can access elsewhere in the command, use
the let option.
Note
Create a collection cakeSales containing sales for cake flavors:
db.cakeSales.insertMany( [ { _id: 1, flavor: "chocolate", salesTotal: 1580 }, { _id: 2, flavor: "strawberry", salesTotal: 4350 }, { _id: 3, flavor: "cherry", salesTotal: 2150 } ] )
The following example:
retrieves the cake that has a
salesTotalgreater than 3000, which is the cake with an_idof 2defines a
targetTotalvariable inlet, which is referenced in$gtas$$targetTotal
db.cakeSales.aggregate( [ { $match: { $expr: { $gt: [ "$salesTotal", "$$targetTotal" ] } } } ], { let: { targetTotal: 3000 } } )