multi-cloud

2760 results

Transforming Industries with MongoDB and AI: Healthcare

This is the sixth in a six-part series focusing on critical AI use cases across several industries . The series covers the manufacturing and motion, financial services, retail, telecommunications and media, insurance, and healthcare industries. In healthcare, transforming data into actionable insights is vital for enhancing clinical outcomes and advancing patient care. From medical professionals improving care delivery to administrators optimizing workflows and researchers advancing knowledge, data is the lifeblood of the healthcare ecosystem. Today, AI emerges as a pivotal technology, with the potential to enhance decision-making, improve patient experiences, and streamline operations — and to do so more efficiently than traditional systems. Patient experience and engagement While they may not expect it based on past experiences, patients crave a seamless experience with healthcare providers. Ideally, patient data from healthcare services, including telehealth platforms, patient portals, wearable devices, and EHR, can be shared – securely – across interoperable channels. Unfortunately, disparate data sources, burdensome and time-consuming administrative work for providers, and overly complex and bloated solution stacks at the health system level all stand in the way of that friction-free experience. AI can synthesize vast amounts of data and provide actionable insights, leading to personalized and proactive patient care, automated administrative processes, and real-time health insights. AI technologies, such as machine learning algorithms, natural language processing, and chatbots, are being used to enhance and quantify interactions. Additionally, AI-powered systems can automatically schedule appointments, send notifications, and optimize clinic schedules, all reducing wait times for patients. AI-enabled chatbots and virtual health assistants provide 24/7 support, offering instant responses, medication reminders, and personalized health education. AI can even identify trends and predict health events, allowing for early intervention and reduction in adverse outcomes. MongoDB’s flexible data model can unify disparate data sources, providing a single view of the patient that integrates EHRs, wearable data, and patient-generated health data for personalized care and better patient outcomes. For wearables and medical devices, MongoDB is the ideal underlying data platform to house time series data, significantly cutting down on storage costs while enhancing performance. With Atlas for the Edge, synchronization with edge applications, including hospital-at-home setups, becomes seamless. On the patient care front, MongoDB can support AI-driven recommendations for personalized patient education and engagement based on the analysis of individual health records and engagement patterns, and Vector Search can power search functionalities within patient portals, allowing patients to easily find relevant information and resources, thereby improving the self-service experience. Enhanced clinical decision making Healthcare decision-making is critically dependent on the ability to aggregate, analyze, and act on an exponentially growing volume of data. From EHRs and imaging studies to genomic data and wearable device data, the challenge is not just the sheer volume but the diversity and complexity of data. Healthcare professionals need to synthesize information across various dimensions to make informed, real-time, accurate decisions. Interoperability issues, data silos, lack of data quality, and the manual effort required to integrate and interpret this data all stand in the way of better decision-making processes. The advent of AI technologies, particularly NLP and LLMs, offers transformative potential for healthcare decision-making by automating the extraction and analysis of data from disparate sources, including structured data in EHRs and unstructured text in medical literature or patient notes. By enabling the querying of databases using natural language, clinicians can access and integrate patient information more rapidly and accurately, enhancing diagnostic precision and personalizing treatment approaches. Moreover, AI can support real-time decision-making by analyzing streaming data from wearable devices, alerting healthcare providers to changes in patient conditions that require immediate attention. MongoDB, with its flexible data model and powerful data development platform, is uniquely positioned to support the complex data needs of healthcare decision-making applications. It can seamlessly integrate diverse data types, from FHIR-formatted clinical data to unstructured text and real-time sensor data, in a single platform. By integrating MongoDB with Large Language Models (LLMs), healthcare organizations can create intuitive, AI-enhanced interfaces for data retrieval and analysis. This integration not only reduces the cognitive load on clinicians but also enables them to access and interpret patient data more efficiently, focusing their efforts on patient care rather than navigating complex data systems. MongoDB's scalability ensures that healthcare organizations can manage growing data volumes efficiently, supporting the implementation of AI-driven decision support systems. These systems analyze patient data in real-time against extensive medical knowledge bases, providing clinicians with actionable insights and recommendations, thereby enhancing the quality and timeliness of care provided. MongoDB's Vector Search further enriches decision-making processes by enabling semantic search across vast datasets directly within the database. This integrated approach enables the application of pre-filters based on extensive metadata, enhancing the efficiency and relevance of search results without the need to synchronize with dedicated search engines or vector stores, meaning healthcare professionals can utilize previously undiscoverable insights, streamlining the identification of relevant information and patterns. Clinical trials and precision medicine The need for innovation and transformation isn’t just limited to the patient-provider-healthcare system experience. The challenges of conducting clinical trials and advancing precision medicine are significant, from identifying and enrolling suitable participants to data management practices are fraught with the potential for errors, compromising the accuracy and reliability of trial outcomes. Moreover, the traditional one-size-fits-all approach to treatment development fails to address the unique genetic makeup of individual patients, limiting the effectiveness of therapeutic interventions. AI can make clinical trials faster and treatments more personalized. It's like having a super-smart assistant that can quickly find the right people for studies, keep track of all the data without making mistakes, and even predict which medicines will work best for different people. This means doctors can create safe, efficient treatments that fit you perfectly, just like a tailor-made suit. Plus, with AI's help, these custom treatments can be developed quicker and be more affordable, bringing us closer to a future where everyone gets the care they need, designed just for them. It's a big step towards making medicine not just about treating sickness but about creating health plans that are as unique as patients are. MongoDB plays a pivotal role in modernizing clinical trials and advancing precision medicine by addressing complex data challenges. Its flexible data model excels in integrating diverse data types, from EHRs and genomic data to real-time patient monitoring streams. This capability is crucial for clinical trials and precision medicine, where combining various data sources is necessary, sometimes through a project purpose ODL, to develop a comprehensive understanding of patient health and treatment responses. For clinical trials, MongoDB can streamline participant selection by efficiently managing and querying vast datasets to identify candidates who meet specific criteria, significantly reducing the recruitment time. Its ability to handle large-scale, complex datasets in real-time also facilitates the dynamic monitoring of trial participants, enhancing the safety and accuracy of trials. Other notable use cases Patient Flow Optimization and Emergency Department Efficiency: AI algorithms can process historical and real-time data to forecast patient volumes, predict bed availability, and identify optimal staffing levels, enabling proactive resource allocation and patient routing. Virtual Health Assistants for Chronic Disease Management: Utilizing AI-powered virtual assistants to monitor patients' health status, provide personalized advice, and support medication adherence for chronic conditions such as diabetes and hypertension. AI-Enhanced Digital Pathology and Medical Imaging: Build modern VNA (Vendor Neutral Archive and Digital pathology solutions with innovative approaches, dealing with interoperable data, and manage extensive metadata associated with all your resources enabling fast findings and automated annotations. Operational Efficiency in Hospital Resource Management: Implementing AI to optimize hospital operations, from staff scheduling to inventory management, ensuring resources are used efficiently and patient care is prioritized. Learn more about AI use cases for top industries in our new ebook, How Leading Industries are Transforming with AI and MongoDB Atlas .

April 22, 2024

Retrieval Augmented Generation for Claim Processing: Combining MongoDB Atlas Vector Search and Large Language Models

Following up on our previous blog, AI, Vectors, and the Future of Claims Processing: Why Insurance Needs to Understand The Power of Vector Databases , we’ll pick up the conversation right where we left it. We discussed extensively how Atlas Vector Search can benefit the claim process in insurance and briefly covered Retrieval Augmented Generation (RAG) and Large Language Models (LLMs). MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More One of the biggest challenges for claim adjusters is pulling and aggregating information from disparate systems and diverse data formats. PDFs of policy guidelines might be stored in a content-sharing platform, customer information locked in a legacy CRM, and claim-related pictures and voice reports in yet another tool. All of this data is not just fragmented across siloed sources and hard to find but also in formats that have been historically nearly impossible to index with traditional methods. Over the years, insurance companies have accumulated terabytes of unstructured data in their data stores but have failed to capitalize on the possibility of accessing and leveraging it to uncover business insights, deliver better customer experiences, and streamline operations. Some of our customers even admit they’re not fully aware of all the data in their archives. There’s a tremendous opportunity to leverage this unstructured data to benefit the insurer and its customers. Our image search post covered part of the solution to these challenges, opening the door to working more easily with unstructured data. RAG takes it a step further, integrating Atlas Vector Search and LLMs, thus allowing insurers to go beyond the limitations of baseline foundational models, making them context-aware by feeding them proprietary data. Figure 1 shows how the interaction works in practice: through a chat prompt, we can ask questions to the system, and the LLM returns answers to the user and shows what references it used to retrieve the information contained in the response. Great! We’ve got a nice UI, but how can we build an RAG application? Let’s open the hood and see what’s in it! Figure 1: UI of the claim adjuster RAG-powered chatbot Architecture and flow Before we start building our application, we need to ensure that our data is easily accessible and in one secure place. Operational Data Layers (ODLs) are the recommended pattern for wrangling data to create single views. This post walks the reader through the process of modernizing insurance data models with Relational Migrator, helping insurers migrate off legacy systems to create ODLs. Once the data is organized in our MongoDB collections and ready to be consumed, we can start architecting our solution. Building upon the schema developed in the image search post , we augment our documents by adding a few fields that will allow adjusters to ask more complex questions about the data and solve harder business challenges, such as resolving a claim in a fraction of the time with increased accuracy. Figure 2 shows the resulting document with two highlighted fields, “claimDescription” and its vector representation, “claimDescriptionEmbedding” . We can now create a Vector Search index on this array, a key step to facilitate retrieving the information fed to the LLM. Figure 2: document schema of the claim collection, the highlighted fields are used to retrieve the data that will be passed as context to the LLM Having prepared our data, building the RAG interaction is straightforward; refer to this GitHub repository for the implementation details. Here, we’ll just discuss the high-level architecture and the data flow, as shown in Figure 3 below: The user enters the prompt, a question in natural language. The prompt is vectorized and sent to Atlas Vector Search; similar documents are retrieved. The prompt and the retrieved documents are passed to the LLM as context. The LLM produces an answer to the user (in natural language), considering the context and the prompt. Figure 3: RAG architecture and interaction flow It is important to note how the semantics of the question are preserved throughout the different steps. The reference to “adverse weather” related accidents in the prompt is captured and passed to Atlas Vector Search, which surfaces claim documents whose claim description relates to similar concepts (e.g., rain) without needing to mention them explicitly. Finally, the LLM consumes the relevant documents to produce a context-aware question referencing rain, hail, and fire, as we’d expect based on the user's initial question. So what? To sum it all up, what’s the benefit of combining Atlas Vector Search and LLMs in a Claim Processing RAG application? Speed and accuracy: Having the data centrally organized and ready to be consumed by LLMs, adjusters can find all the necessary information in a fraction of the time. Flexibility: LLMs can answer a wide spectrum of questions, meaning applications require less upfront system design. There is no need to build custom APIs for each piece of information you’re trying to retrieve; just ask the LLM to do it for you. Natural interaction: Applications can be interrogated in plain English without programming skills or system training. Data accessibility: Insurers can finally leverage and explore unstructured data that was previously hard to access. Not just claim processing The same data model and architecture can serve additional personas and use cases within the organization: Customer Service: Operators can quickly pull customer data and answer complex questions without navigating different systems. For example, “Summarize this customer's past interactions,” “What coverages does this customer have?” or “What coverages can I recommend to this customer?” Customer self-service: Simplify your members’ experience by enabling them to ask questions themselves. For example, “My apartment is flooded. Am I covered?” or “How long do windshield repairs take on average?” Underwriting: Underwriters can quickly aggregate and summarize information, providing quotes in a fraction of the time. For example, “Summarize this customer claim history.” “I Am renewing a customer policy. What are the customer's current coverages? Pull everything related to the policy entity/customer. I need to get baseline info. Find relevant underwriting guidelines.” If you would like to discover more about Converged AI and Application Data Stores with MongoDB, take a look at the following resources: RAG for claim processing GitHub repository From Relational Databases to AI: An Insurance Data Modernization Journey Modernize your insurance data models with MongoDB and Relational Migrator

April 18, 2024

VertexAI and MongoDB for Intelligent Retail Pricing

In today’s competitive retail environment, the ability to quickly adjust pricing in response to market trends, consumer demand, and competitors’ moves is not just an advantage — it's essential for survival. This is where dynamic pricing comes into play, serving as a strategic tool for businesses to pull in their quest for market dominance. Dynamic pricing goes beyond changing numbers; it’s a strategic approach that reflects the dynamic nature of the market, powered by data-driven insights that enable prices to be adjusted in real-time for maximum effectiveness. This shift towards a more agile, data-driven pricing strategy underscores a broader trend in the business world: the recognition of data as a foundational element in decision-making processes. By leveraging real-time data, businesses can ensure their pricing strategies are not only responsive to market fluctuations but also strategically aligned with their overall business objectives, thus driving retail competitiveness to new heights. Let’s uncover how integrating both platforms empowers developers when it comes to delivering best-in-class, data-driven applications. MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More Google Cloud: A platform for real-time analytics and AI Google Cloud stands out as a powerhouse in real-time analytics and artificial intelligence (AI), offering the infrastructure necessary for dynamic pricing strategies and other data-driven business approaches. It's designed to facilitate big data analysis, machine learning, and operational agility. Built-in tools form the backbone of an effective dynamic pricing strategy. These include Vertex AI for advanced machine learning models following best- in- class MLOps practices, and Pub/Sub for real-time messaging to solve real- time data ingestion. By harnessing the power of Google Cloud, retailers can analyze vast quantities of data in real-time, from current market trends to customer behavior and competitor pricing. This enables businesses to make informed decisions swiftly, adjusting their pricing strategies to reflect the ever-changing market conditions. MongoDB: Flexible data modeling and rapid application development MongoDB complements Google Cloud by offering a high performance document- based database with a flexible data model that allows rapid application development. For pricing data in particular, where there may be different variants for different sizes of store or country, the flexibility allows for the ease of storage of complex or hierarchical data. In addition, polymorphic capabilities allow you to use a single interface to represent different types, making your system more flexible. It also supports scalability as new types can be easily integrated. Lastly, it enhances efficiency by allowing the same operation to behave differently based on the object, reducing code redundancy. This flexible schema also enables seamless integration with AI models. MongoDB Atlas supports workload isolation , ensuring dedicated resources for AI tasks and smooth operation alongside core application workloads. Additionally, change streams and triggers can be utilized to capture real-time updates in the pricing data, allowing the AI model to be called upon for immediate analysis and adaptation and enabling in-app analytics for retailers to gain a competitive edge. Figure 1: MongoDB replica set: Workload Isolation In the dynamic pricing reference architecture, Atlas collections function as an ML feature store. By leveraging the capabilities of MongoDB Atlas as a developer data platform, we are able to embed real-time automated decision-making into our e-commerce applications and reduce operational overhead for both business operations and MLOps model fine-tuning. This is achieved through implementing a streamlined approach to data management, incorporating real-time, automated decision making, workload isolation, change streams, triggers for immediate updates, and seamless integration with AI models. Dynamic prising microservice overview Building an event-driven AI architecture leveraging MongoDB Atlas in Google Cloud is straightforward. We can summarize our dynamic pricing microservice by first describing the different components of its architecture, what they are used for, and how they interact with each other: Figure 2: Description of the different technology components of a dynamic pricing microservice and what they are used for. Handling data sources The proposed solution uses Google Cloud Pub/Sub to ingest data sources like customer behavior events in JSON format. Using a technology like Pub/Sub allows for scaling to handle a large number of messages and efficiently distribute them to many subscribers. This is partly because it allows for parallel processing of messages and can be distributed across multiple servers or instances. It is often a fundamental pattern in event-driven architectures, where the flow of the program is determined by events or messages, supporting reactive programming and making the system more responsive and efficient. Data federation We’ll use Vertex AI Notebooks to clean the data and train a TensorFlow model. This model will learn the non-linear relation between customer events, products names, and prices, enabling it to calculate the optimal predicted price. Orchestrating Using Cloud Functions, we orchestrate the customer events coming from the Pub/Sub topic to be converted into tensors, which are then stored in a MongoDB Atlas collection. This collection acts as a feature store serving as a centralized repository designed to store, manage, and serve features for machine learning (ML) models. Features represent individual measurable properties or characteristics used by ML models to make predictions or decisions. MongoDB’s document model flexibility paired with the document versioning pattern will allow us to design time-sensitive chunks of events and granularly manage the training datasets for our models. Serving The Cloud Function will use the event tensor to invoke our trained model that is served in a Vertex AI endpoint. The model will provide a predicted price score that can then be inserted into our product catalog stored in MongoDB so our e-commerce application can read the price change in real time. Dynamic pricing architecture: Putting it all together In the following architecture diagram, the blue data flow illustrates how customer event data is ingested into a Pub/Sub topic. This allows us to make a push subscription to a Cloud Function from the topic. This function orchestrates the data transformation from raw event into a tensor and calls an endpoint to then update the predicted price into our MongoDB product catalog collection. By using this architectural approach, we can isolate raw events threads and build different services around them, reacting in real time for dynamic pricing or asynchronously for model training. With every component loosely coupled, we prevent the system from crashing completely. Moreover, publishers and subscribers can continue to process their logic without the need for the other components to receive or publish messages. Figure 3: Dynamic pricing architecture integrating different Google Cloud components and MongoDB Atlas as a Feature Store For businesses, this translates into more precise and responsive pricing strategies. In the model building and optimization phase, by utilizing TensorFlow within Google Cloud Vertex AI notebooks, retailers can harness the power of deep learning capabilities. The neural network model is capable of analyzing intricate patterns and relationships within large datasets. This is how businesses may capture nuanced market dynamics, customer behavior, and pricing elasticity with greater accuracy, leading to more optimized pricing decisions. But even the best of the models should be consistently optimized. Maintaining model effectiveness requires continuous adaptation. Regularly evaluating accuracy and performing feature engineering ensures your models stay sensitive to market changes. This underscores the importance of retraining as a core principle in a continuous improvement data science approach. Using MongoDB Atlas as your operational data layer means that your feature store is always accessible, reducing downtime and improving the efficiency of machine learning operations. On the other hand, cross-region deployments can bring features closer to where machine learning models are being trained or served, reducing latency and improving model performance. Get started The integration of Google Cloud and MongoDB presents an easy approach to modernizing dynamic pricing strategies. Leveraging real-time analytics, flexible data modeling, and reactive microservices architecture, it empowers businesses to achieve operational efficiencies and gain a competitive advantage in their pricing strategies. For retailers looking to elevate their pricing strategies, considering a strategic partnership with both technologies is essential. For a deeper dive into integrating the different components of this architecture, make sure to check our GitHub repository.

April 17, 2024

The Journey of MongoDB with COVESA in the Connected Vehicle Landscape

There’s a popular saying: “If you want to go fast, go alone; if you want to go far, go together.” I would argue The Connected Vehicle Systems Alliance (COVESA) in partnership with their extensive member network, turns this saying on its head. They have found a way to go fast, together and also go far, together. COVESA is an industry alliance focused on enabling the widespread adoption of connected vehicle systems. This group aims to accelerate the development of these technologies through collaboration and standardization. It's made up of various stakeholders in the automotive and technology sectors, including car manufacturers, suppliers, and tech companies. COVESA’s collaborative approach allows members to accelerate progress. Shared solutions eliminate the need for individual members to reinvent the wheel. This frees up their resources to tackle new challenges, as the community collectively builds, tests, and refines foundational components. As vehicles become more connected, the data they generate explodes in volume, variety, and velocity. Cars are no longer just a mode of transportation, but a platform for advanced technology and data-driven services. This is where MongoDB steps in. MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More MongoDB and COVESA As the database trusted for mission-critical systems by enterprises such as Cathay Pacific , Volvo Connect , or Cox Automotive ; MongoDB has gained expertise in automotive, along with many other industries, building cross-industry knowledge in handling large-scale, diverse data sets. This in turn enables us to contribute significantly to vehicle applications and provide a unique view, especially in the data architecture discussions within COVESA. MongoDB solutions support these kinds of innovations, enabling automotive companies to leverage data for advanced features. One of the main features we provide is Atlas Device SDKs : a low-footprint, embedded database directly living on ECUs. It can synchronize data automatically with the cloud using Atlas Device Sync , our data transfer protocol that compresses the data handles conflict resolution, and only syncs delta changes, making it extremely efficient in terms of operations and maintenance. VSS: The backbone of connected vehicle data An important area of COVESA’s work is the Vehicle Signal Specification (VSS). VSS is a standardized framework used to describe data of a vehicle, such as speed, location, and diagnostic information. This standardization is essential for interoperability between different systems and components within a vehicle, as well as for external communication with other vehicles and infrastructure. VSS has been gaining more and more adoption, and it’s backed by ongoing contributions from BMW, Volvo Cars, Jaguar LR, Robert Bosch and Geotab, among others. MongoDB’s BSON and our Object-oriented Device SDKs uniquely position us to contribute to VSS implementation. The VSS data structured maps 1 to 1 to documents in MongoDB and objects in Atlas Device SDKs , which simplifies development, and speeds up applications by completely skipping any Relational Mapper layer. For every read or write, there is no need to transform the data between relational and VSS. Our insights into data structuring, querying, and management can help optimize the way data is stored and accessed in connected vehicles, making it more efficient and robust. Where MongoDB contributes MongoDB, within COVESA, finds its most meaningful contributions in areas where data complexities and community collaboration intersect. First, we can share insights into managing vast and varied data emerging from connected vehicles generating data on everything from engine performance to driver behavior. Second, we have an important role in supporting the standardization efforts, crucial for ensuring different systems within vehicles can communicate seamlessly. Our inputs can help ensure these standards are robust and practical, considering the real-world scenarios of data usage in vehicles. Some of our contributions include an Over the Air update architectural review presented at Troy COVESA’s AMM in October 2023; sharing insights about the Data Middleware PoC with BMW; and weekly contributions at the Data Expert Group. You can find some of our contributions on COVESA’s Wiki page . In essence, MongoDB's role in COVESA is about providing a unique perspective from the database management point of view, offering our understanding from different industries and use cases to support the developments towards more connected and intelligent vehicles. MongoDB, COVESA, and AWS together at CES2024 MongoDB’s most recent collaboration with COVESA was at the Consumer Electronics Show CES 2024 during which MongoDB’s Connected Vehicle solution was showcased. This solution leverages Atlas Device SDKs, such as the SDK for C++ , which enables local data storage, in-vehicle data synchronization, and also uni and bi-directional data transfer with the cloud. Below is a schematic illustrating the integration of MongoDB within the software-defined vehicle: Schema 1: End to end integration for the connected vehicle At CES 2024, MongoDB also teamed up with AWS for a compelling presentation, " AI-powered Connected Vehicles with MongoDB and AWS " led by Dr. Humza Akhtar and Mohan Yellapantula, Head of Automotive Solutions & Go To Market at AWS. The session delved into the intricacies of building connected vehicle user experiences using MongoDB Atlas. It showcased the combined strengths of MongoDB's expertise and AWS's generative AI tools, emphasizing how Atlas Vector Search unlocks the full lifecycle value of connected vehicle data. During the event, MongoDB also engaged in a conversation with The Six Five, exploring various aspects of mobility, self-driving vehicles (SDVs), and the MongoDB and AWS partnership. This discussion extended to merging IT and OT, GenAI, Atlas Edger Server, and Atlas Device SDK. Going forward At the end of the road, it’s all about enhancing the end-user experience and providing unique value propositions. Defect diagnosis based on the acoustics of the engine, improved crash assistance with mobile and vehicle telemetry data, just-in-time food ordering while on the road, in-vehicle payments, and much, much more. What all these experiences have in common is the combination of interconnected data from different systems. At MongoDB, we are laser-focused on empowering OEMs to create, transform, and disrupt the automotive industry by unleashing the power of software and data. We enable this by: Partnering with alliances such as COVESA to build a strong ecosystem or collaboration. Having one single API for In-vehicle Data Storage, Edge to Cloud Synchronization, Time Series storage, and more, improves the developer experience. Focusing on having a robust, scalable, and secure suite of services trusted by tens of thousands of customers in more than 100 countries. Together with COVESA’s vision for connected vehicles, we’re driving a future where this industry is safer, more efficient, and seamlessly integrated into the digital world. The journey is just beginning. To learn more about MongoDB-connected mobility solutions, visit the MongoDB for Manufacturing & Motion webpage . Achieving fast, reliable and compressed data exchange is one of the pillars of Software Defined Vehicles, learn how MongoDB Atlas and Edge Server can help in this short demo .

April 15, 2024

What’s your Developer Persona?

Cool stuff is getting built with MongoDB every day. Behind the scenes are developers, data scientists, and UX designers who use our products in truly innovative ways. We believe developers are doing way more than just writing code and we’ve outlined a few personas we commonly see. Pick your Archetype for custom backgrounds and other shareable content ! MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More The architect Ideal team member: The Storyteller. Together you’d be unstoppable. Description People are usually intimidated by you at first. You’ll start all quiet and casually drop how you created the very board game everyone is playing at the dinner party you’re at. You have a knack for strategy and probably have a pretty good poker face. Under your calm and collected demeanor is a perpetual motion machine. Scientists said it couldn’t be done but it’s happening inside your brain right now. It’s so effortless you probably don’t even notice it. You’re deeply intuitive and you have a unique combination of both big ideas and informed action. The scientist Ideal team member: The Architect Description You love a healthy debate. Not in the devil’s advocate way but you genuinely like hearing both sides (though you're likely always right). When starting a project, you’re already thinking about things to research so you can hit the ground running. You’re always up to date with the newest technology and constantly thinking about how you can implement it in your workflow. You’re a lifelong learner and you’re super sharp, which makes you an asset to any team. The storyteller Ideal team member: The Scientist Description You’re deeply interested in the human experience. When people enter your house they are wowed by how beautiful and well thought out everything is. You also probably greet them with a seasonally spiced espresso. You have very niche-specific hobbies like crafting modern candlesticks for your friends. People come to you when they want someone to use their product or idea - you know how to craft an experience everyone will keep coming back to! Beauty is in the eye of the beholder, but you’re the beholder and you always know what’s best. The visionary Ideal team member: The Sprinter, they can ground you. Description Leaders are constantly asking you for opinions if you aren’t one yet. You’re interested in the big stuff that will change how people think and interact with the world around them. Your dinner party conversation is strong and you pride yourself on being able to predict the next big thing. Some people may write you off as arrogant until that thing you swore to them would happen, finally did, and they text you saying, “Did you see the news?!?!? You called it!”. You’re uber confident and for good reason. You know exactly how to, as they say, strike while the iron is hot! The sprinter Ideal team member: The Grasshopper Description Your desk is full of half-drunk red-bulls. Or maybe not, but that’s what everyone assumes. You outpace all your colleagues because you simply love getting things done. We imagine you’re into low-effort activities in your spare time like bouldering or Ironman triathlons. You are often the one people call when they want something done right but they are too lazy to do it themselves. You love a good checklist and you're probably the one who does all the work on that new startup venture you and your college friends have. You are deeply passionate about your work and think this meeting could have been an email. The grasshopper Ideal team member: The Architect, you’ll learn a lot! Description No is not in your vocabulary. You’re somewhat of a Renaissance person, people are usually in awe of how you can already be so natural at something you just started. You have a unique blend of independent and collaborative energy that touches everything you set your mind to. You get things done in style. Your unique brand of optimism makes everything you do look effortless. Some people may write you off as a newbie. Don’t worry though, they’re probably just grumpy. The world is truly your oyster, the sky’s the limit, and whatever other saying there is out there you’ve probably thought of 10 already. Meet other developers at .local NYC! We are coming to NYC! Check out our event page to sign up for the conference, including a schedule of events, speaker profiles, and attendee resources, like how to justify your trip and travel details. Register today and meet your ideal collaborator!

April 15, 2024

Transforming Industries with MongoDB and AI: Insurance

This is the fifth in a six-part series focusing on critical AI use cases across several industries . The series covers the manufacturing and motion, financial services, retail, telecommunications and media, insurance, and healthcare industries. With its ability to streamline processes, enhance decision-making, and improve customer experiences in far less time, resources, and staff than traditional IT systems, artificial intelligence offers insurers great promise. In an inherently information-driven industry, insurance companies ingest, analyze, and process massive amounts of data. Whether it’s agents and brokers selling more policies, underwriters adequately pricing, renewing and steering product portfolios, claim handlers adjudicating claims, or service representatives providing assurance and support, data is at the heart of it all. Given the volumes of data, and the amount of decision-making that needs to occur based on it, insurance companies have a myriad of technologies and IT support staff within their technology investment portfolios. It’s no surprise that AI is at the top of the list when it comes to current or prospective IT investments. With its ability to streamline processes, enhance decision-making, and improve customer experiences with far less time, resources, and staff than traditional IT systems, AI offers insurers great promise. Underwriting & risk management Few roles within insurance are as important as that of the underwriters who strike the right balance between profit and risk, bring real-world variables to the actuarial models at the heart of the insurer, and help steer product portfolios, markets, pricing, and coverages. Achieving equilibrium between exposures and premiums means constantly gathering and analyzing information from a myriad of sources to build a risk profile sufficient and detailed enough to make effective policy decisions. While many well-established insurers have access to a wealth of their own underwriting and claims experience data, integrating newer and real-time sources of information, keeping up with regulatory changes, and modeling out what-if risk scenarios still involve significant manual effort. Perhaps the single greatest advantage of AI will be its ability to quickly analyze more information with fewer people and resources. The long-term impact will likely be profound, and there is tremendous promise within underwriting. Advanced analytics: Traditional IT systems are slow to respond to changing formats and requirements surrounding data retrieval. The burden falls on the underwriter to summarize data and turn that into information and insight. Large Language Models are now being leveraged to help speed up the process of wrangling data sources and summarizing the results, helping underwriting teams make quicker decisions from that data. Workload and triage assistance: AI models are mitigating seasonal demands, market shifts, and even staff availability that impact the workload and productivity of underwriting teams, saving underwriting time for high-value accounts and customers where their expertise is truly needed. Amid high volumes for new and renewal underwriting, traditional AI models can help classify and triage risk, sending very low-risk policies to ‘touchless’ automated workflows, low to moderate risk to trained service center staff, and high-risk and high-value accounts to dedicated underwriters. Decision-making support: Determining if a quoted rate needs adjustment before binding and issuing can take significant time and manual effort. So can preparing and issuing renewals of existing policies, another large portion of the underwriters’ day-to-day responsibilities. Automated underwriting workflows leveraging AI are being employed to analyze and classify risk with far less manual effort. This frees up significant time and intellectual capital for the underwriter. Check out our machine learning solutions page to learn more about automated digital underwriting. Vast amounts of data analyzed by underwriters are kept on the underwriter's desktop rather than IT-managed databases. MongoDB offers an unparalleled ability to store data from a vast amount of sources and formats and deliver the ability to respond quickly to requests to ingest new data. As data and requirements change, the Document Model allows insurers to simply add more data and fields without the costly change cycle associated with databases that rely on single, fixed structures. For every major business entity found within the underwriting process, such as a broker, policy, account, and claim, there is a wealth of unstructured data sources, waiting to be leveraged by generative AI. MongoDB offers insurers a platform that consolidates complex data from legacy systems, builds new applications, and extends those same data assets to AI-augmented workflows. By eliminating the need for niche databases for these AI-specific workloads, MongoDB reduces technology evaluation and onboarding time, development time, and developer friction. Claim processing Efficient claim processing is critical for an insurer. Timely resolution of a claim and good communication and information transparency throughout the process is key to maintaining positive relationships and customer satisfaction. In addition, insurers are on the hook to pay and process claims according to jurisdictional regulations and requirements, which may include penalties for failing to comply with specific timelines and stipulations. To process a claim accurately, a wealth of information is needed. A typical automobile accident may include not only verbal and written descriptions from claimants and damage appraisers but also unstructured content from police reports, traffic and vehicle dashboard cameras, photos, and even vehicle telemetry data. Aligning the right technology and the right amount of your workforce in either single or multi-claimant scenarios is crucial to meeting the high demands of claim processing. Taming the flood of data: AI is helping insurers accelerate the process of making sense of a trove of data and allowing insurers to do so in real-time. From Natural Language Processing to image classification and vector embedding, all the pieces of the puzzle are now on the board for insurers to make a generation leap forward when it comes to transforming their IT systems and business workflows for faster information processing. Claims experience: Generating accurate impact assessments for catastrophic events in a timely fashion to inform the market of your exposure can now be done with far less time, and with far more accuracy, by cross-referencing real-time and historical claims experience data, thanks to the power of Generative AI and vector-embedding of unstructured data. Claim expediter: Using vector embeddings from photo, text, and voice sources, insurers are now able to decorate inbound claims with richer and more insightful metadata so that they can more quickly classify, triage, and route work. In addition, real-time insight into workload and staff skills and availability is allowing insurers to be even more prescriptive when it comes to work assignments, driving towards higher output and higher customer satisfaction. Litigation assistance: Claims details are not always black and white, parties do not always act in good faith, and insurers expend significant resources in the pursuit of resolving matters. AI is helping insurers drive to resolution faster and even avoid litigation and subrogation altogether, thanks to its ability to help us analyze more data more effectively and more quickly. Risk prevention: Many insurers provide risk-assessment services to customers using drones, sensors, or cameras, to capture and analyze data. This data offers the promise of preventing losses altogether for customers and lowering exposures, liability, and expenses for the insurer. This is possible thanks to a combination of vector-embedding, traditional, and generative AI models. Learn more about AI-enhanced claim adjustment for automotive insurance on our solutions page. Customer experience Accessing information consistently during a customer service interaction, and expecting the representative to quickly interpret it, are perennial challenges with any customer service desk. Add in the volume, variety, and complexity of information within insurance, and it’s easy to understand why many insurers are investing heavily in the transformation of their customer experience call center systems and processes. 24/7 virtual assistance: As with many AI-based chat agents, the advantage is that it can free up your call center staff to work on more complex and high-touch cases. Handling routine inquiries can now include far more complex scenarios than before, thanks to the power of vector-embedded content and Large Language Models. Claims assistance: Generative AI can deliver specific claim-handling guidelines to claim-handling staff in real time, while traditional ML models can interrogate real-time streams of collected information to alert either the customer or the claim-handler to issues with quality, content, or compliance. AI capabilities allow insurers to process more claims more quickly and significantly reduce errors or incomplete information. Customer profiles: Every interaction is an opportunity to learn more about your customers. Technologies such as voice-to-text streaming, vector embedding, and generative AI help insurers build out a more robust ‘social profile’ of their customers in near real-time. Real-time fraud detection: According to estimates from the Coalition Against Insurance Fraud , the U.S. insurance industry lost over $308 billion to fraud in 2022. With vector-embedding of unstructured data sources, semantic and similarity searches across both vector and structured metadata, and traditional machine learning models, insurers can detect and prevent fraud in ways that were simply not ever before possible. Other notable use cases Predictive Analytics: AI-powered predictive analytics can anticipate customer needs, preferences, and behaviors based on historical data and trends. By leveraging predictive models, insurers can identify at-risk customers, anticipate churn, and proactively engage with customers to prevent issues and enhance satisfaction. Crop Insurance and Precision Farming: AI is being used in agricultural insurance to assess crop health, predict yields, and mitigate risks associated with weather events and crop diseases, which helps insurers offer more accurate and tailored crop insurance products to farmers. Predictive Maintenance for Property Insurance: AI-powered predictive maintenance solutions, leveraging IoT sensors installed in buildings and infrastructure, are used in property insurance to prevent losses and minimize damage to insured properties. Usage-Based Insurance (UBI) for Commercial Fleets: AI-enabled telematics devices installed in commercial vehicles collect data on driving behavior, including speed, acceleration, braking, and location. Machine learning algorithms analyze this data to assess risk and determine insurance premiums for commercial fleets to help promote safer driving practices, reduce accidents, and lower insurance costs for businesses. Learn more about AI use cases for top industries in our new ebook, How Leading Industries are Transforming with AI and MongoDB Atlas. Read the full ebook here .

April 11, 2024

Embracing Neurodiversity During Autism Awareness Month

April is Autism Awareness Month. According to a 2023 study by The Tavistock Institute of Human Relations, nearly half of neurodivergent employees feel impacted by their conditions in the workplace. On top of this, many neurodivergent employees choose to not disclose their conditions due to fear of stigma or reduced career opportunities. It’s safe to say that organizations can do more to create supportive, inclusive, and empowering environments for neurodivergent employees. Config.MDB is an employee resource group aiming to raise awareness for and build support around neurodiversity in the workplace at MongoDB. Sarah Lin, Senior Information & Content Architect, and Config member, speaks about her experience as a parent to a child with autism and shares her perspective on creating inclusive spaces for neurodivergent employees, customers, and members of our communities. According to the CDC, one in every 36 children in the U.S . is on the autism spectrum. My child is one of those. Parenting a child with autism is, in many ways, just parenting. There’s often awe, wonder, frustration, exhaustion, and fountains of love all mixed up in the same day. Parenting is hard, period. As a parent of one neurodiverse and one neurotypical child, I'm able to see a little bit of both worlds and while there are highs and lows for both, they’re usually different. People with autism often have a particular area of interest, and my child has opened up a world to me I was only vaguely familiar with. Learning, enjoying, and being part of their world is a gift. I’m so grateful to experience it with them every day. Parenting children with different abilities can be lonely and isolating, though. I get support where I can, but many folks we encounter are actively unsupportive, assuring me “they’ll grow out of it” or expecting I can make them behave in a certain way. My child needs lower levels of support to navigate the world, which means that people often just dismiss their diagnosis entirely as misbehavior or even say that their challenges don’t exist. Understanding autism spectrum disorder I learned pretty early in my journey that if you’ve met one person with autism, you’ve met one person with autism. Autism Spectrum Disorder is just that, a spectrum, with individuals needing more or less support, so interacting with one person on this spectrum isn’t generalizable to everyone else. I’ve found that each person presents a unique constellation of abilities and challenges, and the best way forward is to ask and get to know each person I meet. Learning about autism and how my child experiences the world has made me a better person, which I hope has carried over into work as well. The most significant impact is raising awareness of sensory needs in the workplace. We all have sensory needs to varying degrees, and understanding that they are genuine makes me more empathetic, accommodating, and patient. Considering the needs of the neurodivergent community If my child grows up wanting to work in the tech industry like I do, I’d advise them to remember their needs and seek a role at a company that seems like the best fit. For example, does working from home help them avoid auditory overstimulation? Awareness of their legal rights is paramount, though; they should ask for the accommodations they need to succeed. I’m sure it’s the parent in me, but the most essential advice is that they take care of themselves. Having time off, structuring their workday and environment to meet their needs, and having a healthy sensory diet are all foundational for doing their best work. To workplaces and colleagues, I encourage you to consider that creating inclusive spaces can be both physical and metaphorical. I’m reminded of a previous employer that set out a specific neurodivergent space at its customer conferences for attendees. I can only speak for what has helped my family, but having alternative options and authority figures willing to compromise when accommodation is outside the norm is impactful. As I teach my child to advocate for their own needs, I rely on the rest of us to meet them with understanding and flexibility. The impact of increasing awareness When talking about impact, there’s no denying that millions of individuals and families benefit from increased awareness, understanding, and inclusivity in our society. Speaking from my own experience as a parent, even if we don’t take advantage of specifically autism-friendly events or spaces, just knowing they are available takes some stress off of daily “how am I going to make this work?” questions. Taking advantage of sensory-friendly activities, for example, allows my child to have the same experiences as everyone else, making them feel included and accepted. For kids, shared experiences are essential for social belonging, and when you’re already experiencing the world differently, that can be a very challenging area to begin with. As I reflect on Autism Awareness Month, I encourage my friends and colleagues to work towards learning about and supporting neurodiversity in their communities. Some of the best ways to do this are simply through asking and researching. Information abounds, and so does the work of learning rather than expecting to be taught. You can also participate in (or start) a community group or employee resource group — like Config at MongoDB , a global employee resource group focused on disability and neurodiversity at MongoDB. The more we strive to educate ourselves and create inclusive environments for everyone, the better our workplaces and communities will become. Learn more about Diversity & Inclusion at MongoDB.

April 10, 2024

Enabling Commerce Innovation with the Power of MongoDB and Google Cloud

Across all industries, business leaders are grappling with economic uncertainty, cost concerns, disruption to supply chains, and pressure to embrace new technologies like generative AI. In this dynamic landscape, having a performant and future-proofed technology foundation is critical to your business’s success. Kin + Carta, a Premier Google Cloud Partner and MongoDB Systems Integrator Partner, recently launched the Integrated Commerce Network . The Integrated Commerce Network is an Accelerator that enables clients to modernize to a composable commerce platform and create value with their commerce data on Google Cloud with a pre-integrated solution in as little as six weeks. This article explains the concept of composable commerce and explores how MongoDB and Google Cloud form a powerful combination that enables innovation in commerce. Finally, it explains how Kin + Carta can help you navigate the complexity facing businesses today with their approach to digital decoupling. MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More Unraveling the complexity: What is composable commerce? Why microservices and APIs? The evolution of commerce architecture Traditional monolithic architectures, once the cornerstone of commerce platforms, are facing challenges in meeting the demands of today's fast-paced digital environment. Microservices, a paradigm that breaks down applications into small, independent services, offer a solution to the limitations of monoliths. This architectural shift allows for improved agility, scalability, and maintainability. Defining composable commerce Composable commerce is a component-based, API-driven design approach that gives businesses the flexibility to build and run outstanding buying experiences free of constraints found in legacy platforms. To be truly composable, the platform must support key tenets: Support continuous delivery without downtime at the component level Have API as the contract of implementation between services, with open, industry-standard protocols providing the glue between components Be SaaS based, or portable to run on any modern public cloud environment Allow the open egress and ingress of data — no black-boxes of vendor data ownership Defining APIs and microservices APIs play a pivotal role in connecting microservices, enabling seamless communication and data exchange. This modular approach empowers businesses to adapt quickly to market changes, launch new features efficiently, and scale resources as needed. Enhanced scalability, resilience, and agility Taking a microservices approach provides businesses with options and now represents a mature and battle-tested approach with commoditized architectures, infrastructure-as-code, and open-source design patterns to enable robust, resilient, and scalable commerce workloads at lower cost and risk. Additionally, the decoupled nature of microservices facilitates faster development cycles. Development teams can work on isolated components, allowing for parallel development and quicker releases. This agility is a game-changer in the competitive e-commerce landscape, where rapid innovation is essential for staying ahead. Microservices and API-based commerce solutions (like commercetools, which is powered by MongoDB) have begun to dominate the market with their composable approach, and for good reason. These solutions remove the dead-end of legacy commerce suite software and enable a brand to pick and choose to enhance its environment on its own terms and schedule. MongoDB Atlas: The backbone of intelligent, generative AI-driven experiences As e-commerce has developed, customers are expecting more from their interactions — flat, unsophisticated experiences just don’t cut it anymore and brands need to deliver on the expectation of immediacy and contextual relevance. Taking a microservices approach enables richer and more granular data to be surfaced, analyzed, and fed back into the loop, perhaps leveraging generative AI to synthesize information that previously would have been difficult or impossible without huge computing capabilities. However, to do this well you need core data infrastructure that underpins the platform and provides the performance, resilience, and advanced features required. MongoDB Atlas on Google Cloud can play a pivotal role in this enablement. Flexible data models: Microservices often require diverse data models. MongoDB Atlas, a fully managed database service, accommodates these varying needs with its flexible schema design, which allows businesses to adapt their data structures without compromising performance. Horizontal scalability: Modern commerce moves a lot of data. MongoDB Atlas excels in distributing data across multiple nodes, ensuring that the database can handle increased loads effortlessly. Real-time data access: Delivering on expectations relies on real-time data access. MongoDB Atlas supports real-time, event-driven data updates, ensuring you are using the most up-to-date information about your customers. Serverless deployment: Rather than spend time and money managing complex database infrastructure, MongoDB Atlas can leverage serverless deployment, allowing developers to focus on building features that delight customers and impact the bottom line. Unleashing generative AI with MongoDB and Google Cloud Generative AI applications thrive on massive datasets and require robust data management. MongoDB effortlessly handles the complex and ever-evolving nature of gen AI data. This includes text, code, images, and more, allowing you to train your models on a richer data tapestry. MongoDB Atlas: Streamlined gen AI development on Google Cloud MongoDB Atlas, the cloud-based deployment option for MongoDB, integrates seamlessly with Google Cloud. Atlas offers scalability and manageability, letting you focus on building groundbreaking gen AI applications. Here's how this powerful duo functions together: Data ingestion and storage: Effortlessly ingest your training data, regardless of format, into MongoDB Atlas on Google Cloud. This data can include text for natural language processing, code for programming tasks, or images for creative generation. AI model training: Leverage Google Cloud's AI services like Vertex AI to train your gen AI models using the data stored in MongoDB Atlas. Vertex AI provides pre-built algorithms and tools to streamline model development. Operationalization and serving: Once trained, deploy your gen AI model seamlessly within your application. MongoDB Atlas ensures the smooth data flow to and from your model, enabling real-time generation. Vector search with MongoDB Atlas: MongoDB Atlas Vector Search allows for efficient retrieval of similar data points within your gen AI dataset. This is crucial for tasks like image generation or recommendation systems. Advantages of this open approach By leveraging a microservices architecture, APIs, and the scalability and flexibility of Atlas, businesses can build agile and adaptable composable platforms. Atlas seamlessly integrates with Google Cloud, providing a streamlined environment for developing and deploying generative AI models. This integrated approach offers several benefits: Simplified development: The combined power of MongoDB Atlas and Google Cloud streamlines the development process, allowing you to focus on core gen AI functionalities. Scalability and flexibility: Both MongoDB Atlas and Google Cloud offer on-demand scalability, ensuring your infrastructure adapts to your gen AI application's growing needs. Faster time to market: The ease of integration and development offered by this combination helps you get your gen AI applications to market quickly. Cost-effectiveness: Both MongoDB Atlas and Google Cloud offer flexible pricing models, allowing you to optimize costs based on your specific gen AI project requirements. Digital decoupling, a legacy modernization approach With so much digital disruption, technology leaders are constantly being challenged. Existing legacy architectures and infrastructure can be extremely rigid and hard to unravel. Over 94% of senior leaders reported experiencing tech anxiety . So how do you manage this noise, meet the needs of the business, stay relevant, and evolve your technology so that you can deliver the kinds of experiences audiences expect? Digital decoupling is a legacy modernization approach that enables large, often well-established organizations to present a unified online experience to their users, take full advantage of their data, innovate safely, and compete effectively with digital natives. Technology evolves rapidly, and an effective microservices solution should be designed with future scalability and adaptability in mind. Kin + Carta helps to ensure that your solution is not only robust for current requirements but also capable of evolving with emerging technologies and business needs. It all starts with a clear modernization strategy that allows you to iteratively untangle from legacy systems, while also meeting the needs of business stakeholders seeking innovation. Navigating commerce complexity with Kin + Carta on Google Cloud Commerce is undergoing a significant transformation, and businesses need a future-proof technology foundation to handle the demands of complex models and massive datasets. That’s why Kin + Carta launched their Integrated Commerce Network , the first commerce-related solution that’s part of Google’s Industry Value Network . With the right tools and partners, your business can be at the forefront of innovation with generative AI, through automating tasks in revolutionary new ways, creating entirely new content formats, and delivering more personalized customer experiences. The complexities of commerce transformation can be daunting. But you can master the art of digital decoupling and leverage the strengths of the Integrated Commerce Network to unlock limitless possibilities and gain an edge over your competition. Check out Kin + Carta’s guide: Flipping the script — A new vision of legacy modernization enabled by digital decoupling . Get started with MongoDB Atlas on Google Cloud today.

April 9, 2024

A Smarter Factory Floor with MongoDB Atlas and Google Cloud's Manufacturing Data Engine

The manufacturing industry is undergoing a transformative shift from traditional to digital, propelled by data-driven insights, intelligent automation, and artificial intelligence. Traditional methods of data collection and analysis are no longer sufficient to keep pace with the demands of today's competitive landscape. This is precisely where Google Cloud’s Manufacturing Data Engine (MDE) and MongoDB Atlas come into play, offering a powerful combination for optimizing your factory floor. Unlock the power of your factory data MDE is positioned to transform the factory floor with the power of cloud-driven insights. MDE simplifies communication with your factory floor, regardless of the diverse protocols your machines might use. It effortlessly connects legacy equipment with modern systems, ensuring a comprehensive data stream. MDE doesn't just collect data, it transforms it. By intelligently processing and contextualizing the information, you gain a clearer picture of your production processes in real-time with a historical pretext. It offers pre-built analytics and AI tools directly addressing common manufacturing pain points. This means you can start finding solutions faster, whether it's identifying bottlenecks, reducing downtime, or optimizing resource utilization. Conveniently, it also offers great support for integrations that can further enhance the potential of the data (e.g. additional data sinks). The MongoDB Atlas developer data platform enhances MDE by providing scalability and flexibility through automated scaling to adapt to evolving data requirements. This makes it particularly suitable for dynamic manufacturing environments. MongoDB’s document model can handle diverse data types and structures effortlessly while being incredibly flexible because of its native JSON format. This allows for enriching MDE data with other relevant data, such as supply chain logistics, for a deeper understanding of the factory business. You can gain immediate insights into your operations through real-time analytics, enabling informed decision-making based on up-to-date data. While MDE offers a robust solution for collecting, contextualizing, and managing industrial data, leveraging it alongside MongoDB Atlas offers tremendous advantages Inside the MDE integration Google Cloud’s Manufacturing Data Engine (MDE) acts as a central hub for your factory data. It not only processes and enriches the data with context, but also offers flexible storage options like BigQuery and Cloud Storage. Now, customers already using MongoDB Atlas can skip the hassle of application re-integration and make this data readily accessible for applications. Through this joint solution developed by Google Cloud and MongoDB, you can seamlessly move the processed streaming data from MDE to MongoDB Atlas using Dataflow jobs. MDE publishes the data via a Pub/Sub subscription, which is then picked up by a custom Dataflow job built by MongoDB. This job transforms the data into the desired format and writes it to your MongoDB Atlas cluster. Google Cloud’s MDE and MongoDB Atlas utilize compatible data structures, simplifying data integration through a shared semantic configuration. Once the data resides in MongoDB Atlas, your existing applications can access it seamlessly without any code modifications, saving you time and effort. The flexibility of MDE, combined with the scalability and ease of use of MongoDB Atlas, makes this a powerful and versatile solution for various data-driven use cases such as predictive maintenance and quality control, while still providing factory ownership of the data. Instructions on how to set up the dataflow job are available in the MongoDB github repository. Conclusion If you want to level up your manufacturing data analytics, pairing MDE with MongoDB Atlas provides a proven, easy-to-implement solution. It's easy to get started with MDE and MongoDB Atlas . MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More

April 9, 2024

Unleashing Developer Potential–and Managing Costs–with MongoDB Atlas

In today's business landscape, where unpredictability has become the norm, engineering leaders have to balance the dual challenges of managing uncertainty while optimizing IT costs. Indeed, the 2024 MarginPLUS Deloitte survey—which draws on insights from over 300 business leaders—emphasizes a collective pivot towards growth initiatives and cost transformations amidst the fluctuating global economic climate. MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More MongoDB Atlas: A developer's ally for cost-effective productivity Executives across industries want to cut costs without impacting innovation; based on the Deloitte survey , 83% of companies are looking to change how they run their business margin improvement efforts. This is where MongoDB Atlas , the most advanced cloud database service on the market, comes in. An integrated suite of data services that simplify how developers build with data, MongoDB Atlas helps teams enhance their productivity without compromising on cost efficiency by offering visibility and control over spending—balancing developer freedom with data governance and cost management. This helps organizations escape the modernization hamster wheel—or the vicious cycle of continuously updating technology without making real progress, and draining resources while failing to deliver meaningful improvements. Put another way, MongoDB gives teams more time to innovate, instead of just maintaining the status quo. Outlined below are the built-in features of MongoDB Atlas, which enable customers to get the most out of their data while also focusing on budget optimization. Strategic features for cost optimization with MongoDB Atlas Right-sizing your cluster Use MongoDB Atlas’s Cluster Sizing Guide or auto-scalability to match your cluster with your workload, optimizing resource use with options for every requirement, including Low CPU options for lighter workloads. Pausing clusters and global distribution Save costs by pausing your cluster , and securely storing data for up to 30 days with auto-resume. Furthermore, Global Clusters improve performance across regions while maintaining cost efficiency and compliance. Index and storage management Enhance performance and reduce costs with MongoDB Atlas’s Performance Advisor , which provides tailored index and schema optimizations for better query execution and potential reductions in cluster size. Strategic data management Reduce storage expenses using Online Archive for infrequently accessed data and TTL indexes for efficient Time Series data management, ensuring only essential data is stored. Securely backup data before deletion with mongodump. Enhanced spend management Use spend analysis, billing alerts , and usage insights via the Billing Cost Explorer for detailed financial management and optimization. Resource tagging and customizable dashboards provide in-depth financial reporting and visual expense tracking, supporting effective budgeting and cost optimization. Additionally, Opt for serverless instances to adjust to your workload's scale, offering a pay-for-what-you-use model that eliminates overprovisioning concerns. Transforming uncertainty into advancement MongoDB Atlas equips IT decision-makers and developers with the features and tools to balance developer productivity with strategic cost management, transforming economic uncertainty into a platform for strategic advancement. MongoDB Atlas is more than a database management solution; it’s a strategic partner in optimizing your IT spending, ensuring that your organization remains agile, efficient, and cost-effective in the face of change. Need expert assistance in taking control of your MongoDB Atlas costs? MongoDB’s Professional Services team can provide a deep-dive assessment of your environment to build a tailored optimization plan—and to help you execute. Reach out to learn how we can support your cost optimization goals! If you haven't yet set up your free cluster on MongoDB Atlas , now is a great time to do so. You have all the instructions in this DevCenter article .

April 8, 2024

Transforming Industries with MongoDB and AI: Financial Services

This is the fourth in a six-part series focusing on critical AI use cases across several industries . The series covers the manufacturing and motion, financial services, retail, telecommunications and media, insurance, and healthcare industries. In the dynamic world of financial services, the partnership between artificial intelligence (AI) and banking services is reshaping traditional practices, offering innovative solutions across critical functions. MongoDB.local NYC Join us in person on May 2, 2024 for our keynote address, announcements, and technical sessions to help you build and deploy mission-critical applications at scale. Use Code Web50 for 50% off your ticket! Learn More Relationship management support with chatbots One key service that relationship managers provide to their private banking customers is aggregating and condensing information. Because banks typically operate on fragmented infrastructure with information spread across different departments, solutions, and applications, this can require a lot of detailed knowledge about this infrastructure and how to source information such as: When are the next coupon dates for bonds in the portfolio? What has been the cost of transactions for a given portfolio? What would be a summary of our latest research? Please generate a summary of my conversation with the client. Until now, these activities would be highly manual and exploratory. For example, a relationship manager (RM) looking for the next coupon dates would likely have to go into each of the clients' individual positions and manually look up the coupon dates. If this is a frequent enough activity, the RM could raise a request for change with the product manager of the portfolio management software to add this as a standardized report. But even if such a standardized report existed, the RM might struggle to find the report quickly. Overall, the process is time-consuming. Generative AI systems can facilitate such tasks. Even without specifically trained models, RAG can be used to have the AI generate the correct answers, provide the inquirer with a detailed explanation of how to get to the data, and, in the same cases directly execute the query against the system and report back the results. Similar to a human, it is critical that the algorithm has access to not only the primary business data, e.g. the portfolio data of the customer, but also user manuals and static data. Detailed customer data, in machine-readable format and as text documents, is used to personalize the output for the individual customer. In an interactive process, the RM can instruct the AI to add more information about specific topics, tweak the text, or make any other necessary changes. Ultimately, the RM will be the quality control for the AI’s output to mitigate hallucinations or information gaps. As outlined above, not only will the AI need highly heterogeneous data from highly structured portfolio information to text documents and system manuals to provide a flexible natural language interface for the RMs, it will also have to have timely processing information about a customer's transactions, positions, and investment objectives. Providing transactional database capabilities as well as vector search makes it easy to build RAG-based applications using MongoDB’s developer data platform. Risk management and regulatory compliance Risk and fraud prevention Banks are tasked with safeguarding customer assets and detecting fraud , verifying customer identities, supporting sanctions regimes (Sanctions), and preventing various illegal activities (AML). The challenge is magnified by the sheer volume and complexity of regulations, making the integration of new rules into bank infrastructure costly, time-consuming, and often inadequate. For instance, when the EU's Fifth Anti-Money Laundering Directive was implemented, it broadened regulations to cover virtual currencies and prepaid cards . Banks had to update their onboarding processes swiftly, and software, train staff, and possibly update their customer interfaces to comply with these new requirements. AI offers a transformative approach to fraud detection and risk management by automating the interpretation of regulations, supporting data cleansing, and enhancing the efficacy of surveillance systems. Unlike static, rules-based frameworks that may miss or misidentify fraud due to narrow scope or limited data, AI can adaptively learn and analyze vast datasets to identify suspicious activities more accurately. Machine learning, in particular, has shown promise in trade surveillance, offering a more dynamic and comprehensive approach to fraud prevention. Regulatory compliance and code change assistance The regulatory landscape for banks has grown increasingly complex, demanding significant resources for the implementation of numerous regulations. Traditionally, adapting to new regulations has required the manual translation of legal text into code, provisioning of data, and thorough quality control—a process that is both costly and time-consuming, often leading to incomplete or insufficient compliance. For instance, to comply with the Basel III international banking regulations , developers must undertake extensive coding changes to accommodate the requirements laid out in thousands of pages of documentation. AI has the capacity to revolutionize compliance by automating the translation of regulatory texts into actionable data requirements and validating compliance through intelligent analysis. This approach is not without its challenges, as AI-based systems may produce non-deterministic outcomes and unexpected errors. However, the ability to rapidly adapt to new regulations and provide detailed records of compliance processes can significantly enhance regulatory adherence. Financial document search and summarization Financial institutions, encompassing both retail banks and capital market firms, handle a broad spectrum of documents critical to their operations. Retail banks focus on contracts, policies, credit memos, underwriting documents, and regulatory filings, which are pivotal for daily banking services. On the other hand, capital market firms delve into company filings, transcripts, reports, and intricate data sets to grasp global market dynamics and risk assessments. These documents often arrive in unstructured formats, presenting challenges in efficiently locating and synthesizing the necessary information. While retail banks aim to streamline customer and internal operations, capital market firms prioritize the rapid and effective analysis of diverse data to inform their investment strategies. Both retail banks and capital market firms allocate considerable time to searching for and condensing information from documents internally, resulting in reduced direct engagement with their clients. Generative AI can streamline the process of finding and integrating information from documents by using NLP and machine learning to understand and summarize content. This reduces the need for manual searches, allowing bank staff to access relevant information more quickly. MongoDB can store vast amounts of both live and historical data, regardless of its format which is typically needed for AI applications. It offers Vector Search capabilities essential for retrieval-augmented generation (RAG). MongoDB supports transactions, ensuring data accuracy and consistency for AI model retraining with live data. It facilitates data access for both deterministic algorithms and AI-driven rules through a single interface. MongoDB boasts a strong partnership ecosystem , including companies like Radiant AI and Mistral AI, to speed solution development. ESG analysis Environmental, social, and governance (ESG) considerations can have a profound impact on organizations. For example, regulatory changes—especially in Europe—have compelled financial institutions to integrate ESG into investment and lending decisions. Regulations such as the EU Sustainable Finance Disclosure Regulation (SFDR) and the EU Taxonomy Regulation are examples of such directives that require financial institutions to consider environmental sustainability in their operations and investment products. Investors' demand for sustainable options has surged, leading to increased ESG-focused funds. The regulatory and commercial requirements, in turn, drive banks to also improve their green lending practices . This shift is strategic for financial institutions, attracting clients, managing risks, and creating long-term value. However, financial institutions face many challenges in managing different aspects of improving their ESG analysis. The key challenges include defining and aligning standards, and processes and managing the flood of rapidly changing and varied data to be included for ESG analysis purposes. AI can help to address these key challenges in not only an automatic but also adaptive manner via techniques like machine learning. Financial institutions and ESG solution providers have already leveraged AI to extract insights from corporate reports, social media, and environmental data, improving the accuracy and depth of ESG analysis. As the market demands a more sustainable and equitable society, predictive AI combined with generative AI can also help to reduce bias in lending to create fairer and more inclusive financing while improving the predictive powers. The power of AI can help facilitate the development of sophisticated sustainability models and strategies, marking a leap forward in integrating ESG into broader financial and corporate practices. Credit scoring The convergence of alternative data, artificial intelligence, and generative AI is reshaping the foundations of credit scoring, marking a pivotal moment in the financial industry. The challenges of traditional models are being overcome by adopting alternative credit scoring methods, offering a more inclusive and nuanced assessment. Generative AI, while introducing the potential challenge of hallucination, represents the forefront of innovation, not only revolutionizing technological capabilities but fundamentally redefining how credit is evaluated, fostering a new era of financial inclusivity, efficiency, and fairness. The use of artificial intelligence, in particular generative artificial intelligence, as an alternative method to credit scoring has emerged as a transformative force to address the challenges of traditional credit scoring methods for several reasons: Alternative data analysis: AI models can process a myriad of information, including alternative data such as utility payments and rental history, to create a more comprehensive assessment of an individual's creditworthiness. AI offers unparalleled adaptability : As economic conditions change and consumer behaviors evolve, AI-powered models can quickly adjust. Fraud detection: AI algorithms can detect fraudulent behavior by identifying anomalies and suspicious patterns in credit applications and transaction data. Predictive analysts: AI algorithms, particularly ML techniques, can be used to build predictive models that identify patterns and correlations in historical credit data. Behavioral analysis: AI algorithms can analyze behavioral data sets to understand financial habits and risk propensity. By harnessing the power of artificial intelligence, lenders can make more informed lending decisions, expand access to credit, and better serve consumers (especially those with limited credit history). However, to mitigate potential biases and ensure consumer trust, it's crucial to ensure transparency, fairness, and regulatory compliance when deploying artificial intelligence in credit scoring. AI in payments A lack of developer capacity is one of the biggest challenges for banks when delivering payment product innovation. Banks believe the product enhancements they could not deliver in the past two years due to resource constraints would have supported a 5.3% growth in payments revenues . With this in mind and the revolutionary transformation with the integration of AI, it is imperative to consider how to free up developer resources to make the most of these opportunities. There are several areas in which banks can apply AI to unlock new revenue streams and efficiency gains. The image below provides a high-level view of eight of the principal themes and areas. This is not an exhaustive view but does demonstrate the depth and breadth of current opportunities. In each example, there are already banks that have begun to bring services or enhancements to the market using AI technologies or are otherwise experimenting with the technology. Learn more about AI use cases for top industries in our new ebook, How Leading Industries are Transforming with AI and MongoDB Atlas .

April 4, 2024

MongoDB Announces Support for Punch Cards, Arrakis

MongoDB is excited to announce a groundbreaking expansion of our capabilities. Beginning today, April 1, MongoDB is expanding its support of programming languages like INTERCAL, FORTRAN, and even UNITYPER punch cards, as well as support for ultra-remote regions of the known universe with new subspace, faster-than-light (FTL) data transfer and networking protocols. Bringing vintage programming languages into the modern age Support for legacy programming languages presents a unique opportunity for our developers. INTERCAL, known for its quirkiness and obscurity, challenges developers with its unconventional syntax. Meanwhile, the experience many pioneers in computing gained by encoding data onto UNITYPER punch cards paved the way for the modern computing we know today, and supporting punch card languages underscores MongoDB’s long-standing commitment to honoring the rich heritage of computing history. We are thrilled to resurrect these relics of computing and, by doing so, to highlight the importance of developers in the past, present, and future. Expanding availability of MongoDB to the earth’s crust, the planet Arrakis (Dune), and beyond Following MongoDB’s long-standing tradition of pushing boundaries, we are also excited to announce the expanded availability of MongoDB Atlas to regions that were once thought unreachable. Beginning today, MongoDB Atlas will support even the most remote regions of the known universe. Whether you find yourself trapped in the Earth’s crust, or you’re riding on a Guild heighliner to harvest spice on the planet Arrakis, MongoDB will be there to support your data needs every step of the way. As part of this planned expansion, MongoDB will provide network connections to the center of the Earth using new types of heat-resistant polymers that can withstand the extreme temperatures of Earth’s molten core. Additionally, new subspace data transfer and networking protocols derived from advancements in quantum computing will allow low-latency applications to function across solar systems, galaxies, and parallel dimensions. ...APRIL FOOLS! Alright, alright, that’s enough, we’ll let you in on the joke—we won’t be supporting any vintage coding languages any time soon. While bringing back punch cards and riding sandworms might not be on MongoDB’s roadmap, we remain as committed as ever to supporting our developer community . No matter the future, developers will build it. Happy April Fools’ Day! For more developer updates, head to our Developer Center .

April 1, 2024