Innovate With AI Future Enterprise
Innovate With AI Future Enterprise
3
Sept 2024
AI and Application
Modernization
Modernizing apps faster for
a better user experience
1
Table of Contents
Why read it, you ask? Because within these pages lie invaluable insights
into the critical role of AI. Understanding its significance and harnessing
its power is paramount for businesses striving for success. You can also
delve into our partner section highlighting organizations that have built
AI solutions using MongoDB and our MongoDB AI Application Program
(MAAP). Whether a SaaS end-to-end solution you can implement, or
component-based solution you can plug in, there is something here for
you.
Data Model
query API. Together, they empower developers
to accelerate innovation, gain a competitive
edge, and seize new market opportunities
presented by gen AI. Documents, which align
seamlessly with code objects, offer an intuitive
and adaptable way to manage data of any
structure. Unlike traditional tabular data
models, documents afford the flexibility to
accommodate diverse data types and
application features, facilitating data
rationalization and utilization in ways
previously
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. unattainable.
All rights reserved.
Rapid
Querying
Paired with the document model, the search, enriching the user experience and
MongoDB Query API provides developers with expanding application capabilities. MongoDB
a unified and consistent approach to data Atlas stands apart by offering a
manipulation across various data services. comprehensive suite of query functionality
From basic CRUD operations to complex within a single, unified experience. This
analytics and stream processing, the MongoDB eliminates the need for developers to manually
Query API offers developers the flexibility to integrate query results from multiple
query and process data according to the databases, reducing complexity, errors, costs,
application's requirements. In the realm of gen and latency. Moreover, it maintains a compact
AI, this flexibility enables developers to define and agile technology footprint, enabling
additional filters on vector-based queries, such developers to focus on building end-user
as combining metadata, aggregations, and functionality with greater ease and efficiency.
geospatial
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Vectors, Unstructured Data, and
MongoDB Atlas Vector Search
To feed AI models with proprietary data, there to with structured business data. Considering
is a need to create vector embeddings. Data in that more than 80% of data is unstructured,
any digital format and of any structure – i.e., it’s easy to appreciate how transformational
text, video, audio, images, code, tables – can vector search combined with gen AI really is.
be transformed into a vector by processing it
with a suitable vector embedding model. This Once data has been transformed into vector
incredible transformation turns data that was embeddings, it is persisted and indexed in a
previously unstructured and, therefore, vector store such as MongoDB Atlas Vector
completely opaque to a computer into data Search. To retrieve similar vectors, the store is
that contains meaning and structure inferred queried with an Approximate Nearest
and represented via these embeddings. Now Neighbor (ANN) algorithm to perform a K
users can search and compute unstructured Nearest Neighbor (KNN) search using an
data in the same way they’ve always been able algorithm such as 'Hierarchical Navigable
Small Worlds' (HNSW).
Figure 1: Data is transformed from unstructured internal sources to actionable, impactful insights
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Atlas for Industries
Manufacturing
and Motion
The integration of AI
within the manufacturing
and automotive industry
has transformed the
conventional value chain,
presenting a spectrum
of opportunities.
AI investments in the manufacturing industry
Inventory Management
and Optimization
Inventory management is essential for right inventory in the right place at the right
achieving the goals of efficient supply chains, time to minimize system costs and meet
controlling costs, and delivering to customers customer needs.
with minimal delays. Inventory management
is primarily concerned with the planning and Usually, companies do supply chain planning
controlling an industry’s inventory. It includes at several levels, namely strategic, tactical, and
business processes such as estimating operational. Each level differs in its objectives,
material requirements at various points in the planning horizon, and level of detail. Strategic
supply chain, determining necessary material’s and tactical planning are crucial to successful
amount, ordering frequency, and safety stock supply chain management. The so-called
levels. It also includes inventory visibility, Pareto's law could be applicable here as 20%
inventory forecasting, lead time management, of efforts in strategic and tactical planning
inventory shipping costs, inventory valuation, brings 80% of the total effect. At the strategic
forecasting future inventory prices, available level, the leadership team makes high-level
physical space, quality management, returns decisions that affect the entire organization.
and defective goods, and demand forecasting. Scenario planning is done at this level. The
It plays a very important role in reducing analysts go through scores of internal and
overall costs and rapid response objectives. external data including global news, political
Effective inventory management requires the developments, think tank studies and scientific
literature to pinpoint the most strategic integration challenge. The spread of data
concerns and trends that the organization across multiple systems and locations
needs to focus on. The team can then use (on-prem systems, cloud regions etc.) can lead
these outputs to develop a set of draft to data silos and hinder visibility into overall
scenarios for consideration. inventory levels and movements. Finally, when
there is sparse historical inventory data
This tedious process comes with its own set of available, then traditional ML models may
challenges. Predicting long-term demand, suffer in accuracy.
market trends and economic conditions is
challenging because of the long-term horizon. Generative AI and IoT technologies hold
This long planning horizon increases the potential to address some of these challenges.
uncertainty in predicting demand, as market Generative AI in particular can assist in
conditions, consumer preferences, and scenario planning by generating various
technological advancements can change potential outcomes based on a wide range of
significantly over time. data, allowing the organization to prepare
better for an uncertain future.
At the tactical and operational level, for
manufacturers, to manage and optimize How AI and MongoDB Help
inventory levels, the first step is to maintain an We will start with scenario planning to
accurate and real-time view of inventory levels generate hypothetical situations which could
across multiple plants, warehouses, and affect inventory requirements, supply chain
suppliers. This is absolutely essential as performance and overall operations. Effective
without having visibility on the current scenario planning helps companies plan for
inventory levels, it is impossible to optimize. optimal inventory levels.
The second step is to reduce inventory carrying
costs while still ensuring that the required parts After scenario planning, we will look into
and materials are available to ship out when inventory classification use cases where AI can
needed. Finally, the data from multiple be used to categorize inventory based on
customers needs to be aggregated and factors such as demand variability, lead times
analyzed despite being in different formats, and criticality.
each with its own unique lead times and order
quantities. Scenario planning
A scenario planning process has two stages
However, efficient inventory management for
manufacturers presents complex data ● Scenario generation
challenges too, primarily in forecasting ● Strategy generation
demand accurately and optimizing stock
In scenario generation, a generative AI
levels. One issue routinely faced is the
application can look at a vast amount of data
variability and unpredictability of customer
including internal and external business data,
demand patterns, making it difficult to
competition data, political news and events
precisely anticipate inventory needs. Managing
and social media news, find the correlation
diverse data streams from sales records,
between each piece of unstructured
production schedules, supplier information,
information and then rank these areas of
and market trends poses a considerable data
concern in terms of their estimated
significance. It is important to utilize general
knowledge that a Large Language Model
processes as well as internal company data in a
Retrieval Augmented Generation (RAG) model
to avoid hallucinations.
The figure above shows a scenario generation factors due to certain reasons, including but not
application data flow. This retrieval-augmented limited to the political landscape and global
generation (RAG) system consists of four parts. supply chain disruptions etc.
First, an AI data store aggregates and
operationalizes structured and unstructured MongoDB Atlas streamlines RAG
data. In our case, the majority of the data is in implementation as it handles everything under
unstructured format such as news sources, the hood. App data, metadata, context data,
social media and competitive information and vectors are all stored in the same place. As
typically stored as unstructured PDF files. The the app evolves, the document model is
PDF files are vectorized and stored in MongoDB inherently flexible and ideal for storing
Atlas. Atlas Vector Search is then utilized to structured and unstructured data. You can add
perform semantic searches and to find data to the same collections inside the
meaningful context from the PDF embeddings. database as needed, vectorize it, and store the
vectors alongside it.
Atlas Vector Search can be triggered using an
AI application, connecting to MongoDB Atlas to Once the data is stored, vector search
retrieve the right context, which is then fed into capabilities are provided right out of the box,
the large language model to answer questions and search operations can be optimized using
like "Can you generate hypothetical trends that dedicated search nodes.
can affect our firm?" The response might
suggest customer behavior shifts or economic
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
With MongoDB Atlas, it's just one query in one Once the demand is known, AI can play a
technology, one language, and one pivotal role in stock optimization. By analyzing
infrastructure to manage and scale, with no historical sales data, market trends,
data duplication, ultimately leading to a lower manufacturers can determine the most
total cost of ownership and a unified developer efficient stock levels. AI systems can also place
experience. orders automatically based on predicted
demand and targeted stock levels. This
The same application can then categorize the automation not just saves time but also
trends in terms of their probability of reduces human error. Finally, AI utility can be
occurrence and impact. Next, a strategy can extended to supplier selection and relationship
be created automatically, which provides management. By analyzing supplier
guidance on what steps to take in response to performance data, AI based software can
the trends and scenarios generated. The assist in choosing the right suppliers who will
strategy can also contain information about meet the company’s quality, delivery and cost
the financial implications and risks associated requirements.
with the response, for example any significant
initial investment in hiring more AI scientists to MongoDB Atlas provides a flexible, scalable,
develop RAG applications for inventory and highly available developer data platform
management. for managing inventory data. The document
data model can handle complex inventory
For scenario planning, generative AI can be structures and hierarchies, making it easy to
used to generate and evaluate strategies. manage inventory across multiple plants and
However, it is important to provide enough suppliers.
context to the LLM so that it does not
hallucinate. MongoDB Atlas Vector Search is At the warehouse, the inventory can be
key to creating a RAG application. Additionally, scanned using a mobile device. This data can
these AI-generated strategies should be be persisted in a MongoDB collection. Once
viewed as initial concepts for further data is in Atlas, it can serve as the central
exploration rather than as final solutions to be repository for all inventory-related data which
adopted without additional analysis. includes stock, supplier, and customer
Generative AI should complement human information, bill of materials and production
efforts by supporting the identification, line data. This repository becomes the source
evaluation and timely execution of appropriate of data for the inventory management AI
strategies. applications. This approach removes data silos
and improves visibility into overall inventory
Inventory Optimization levels and movements.
One of the most significant applications of AI
in inventory management is in demand However, the challenge of poor or sparse data
forecasting. AI algorithms can be used to at the source systems may still remain. To solve
analyze complex datasets to predict future this, manufacturers can take advantage of
demand of products or parts. Improvement in generative AI and Atlas Vector Search to
demand forecasting accuracy is crucial for implement a Retrieval Augmented Generation
maintaining optimal inventory levels. AI-based (RAG) architecture to generate synthetic data
time series forecasting can assist in adapting whenever needed. They can take multimodal
to rapid changes in customer demand. content such as product descriptions and
specifications, customer feedback and reviews RAG. They can easily categorize products
and inventory notes, vectorize them and store based on their seasonal attributes, cluster
vector embeddings alongside the operational products with similar seasonal demand
data in MongoDB Atlas. This allows them to patterns and provide context to the foundation
supercharge their inventory optimization using model to improve the accuracy of synthetic
inventory data generation.
The figure below shows a reference customer behaviors. This approach not only
architecture of generative AI+AI enabled fills the gap of missing historical data for new
demand forecasting with MongoDB Atlas. The products but also provides a robust foundation
accurate demand forecasting results will help for demand forecasting enabling
in stocking up on right inventory levels. For manufacturers to optimize their inventory
new products, the historical sales data is not levels.
available. Generative AI models can create
realistic and diverse synthetic data by learning
patterns from existing datasets of similar
products. This synthetic data can mimic the
sales trends and seasonality that new products
Solution demo
Discover how to building an
might experience. Atlas Vector search can find
event-driven inventory management
similar products attributes and feed that
system.
context into the generative AI model. By
finding semantics in similar products, Atlas
vector search can help refine the synthetic
data generation, ensuring that it closely
reflects potential market conditions and
Figure 5: Gen AI enabled demand forecasting with MongoDB Atlas
A well-defined maintenance strategy can be a Third, based on the prediction, the system can
game-changer for any organization, driving generate work orders, schedule maintenance
significant revenue and cost savings. Here's activities, and even provide guidance to
how it works: maintenance personnel. This ensures that
maintenance is performed only when
First, identify the equipment that is most necessary, optimizing resource allocation.
crucial for your operations. Downtime for this
equipment can lead to bottlenecks, halting This series of activities delivers tangible
production. benefits. Costs are reduced through saved
labor hours and extended machine lifespan.
Second, equip these critical assets with Additionally, revenue increases as your
sensors to enable condition monitoring. This machines operate at optimal performance
allows you to monitor the health of the levels.
equipment in real time, identifying potential
issues before they escalate into catastrophic
failures.
Figure 7: Different challenges seen at each stage of the predictive maintenance deployment
Now, let's envision a factory containing both maintenance managers and leaders and
automated and manual machines. Let's call it conducting quantitative analyses, a rather
Gear Transmission Systems Ltd, whose primary manual process.
output is gearboxes.
Once you’ve identified the machines, the next
Within this factory, we have an array of step is to install sensors and train the machine
equipment: cutting machines, milling learning model. However, two major challenges
machines, measurement devices, and more. As arise. First, you lack sufficient "run to failure"
its general manager, you are tasked with data to effectively train the model. Secondly,
managing the budget allocated for machine health deteriorates over time, leading
maintenance processes and improving to evolving failure modes with the age of the
strategies. One pressing question you must machine.
address is: which machines should take priority
for the predictive maintenance projects, and Assuming you manage to overcome these
why? This involves consulting with the hurdles, the next phase involves maintenance
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
scheduling and execution. You’re faced with a ● Prioritization of
myriad of data silos, including inventory data machines/criticality analysis: At this
and resource availability data, which need to stage, we require both structured and
be integrated to formulate a comprehensive unstructured data. We need previous
repair plan. Furthermore, it's essential to machine failure data as well as expert
ensure that operators are adept at addressing analysis/opinion on which machines to
minor machine issues to reduce reliance on prioritize for predictive maintenance
external experts. While complex issues may and why.
still require OEM or SI support, internal ● Failure prediction: This stage involves
troubleshooting capabilities are invaluable. structured data such as sensor data
Therefore, developing easy-to-follow and maintenance logs to identify the
documentation tailored to the skill levels of our root cause of imminent failure.
staff is important. ● Maintenance plan and guidance
generation: In both of these stages,
As the Figure below shows, different data is we deal with both structured and
required for solving above listed challenges: unstructured data. The objective is to
combine this data to generate an
optimal repair plan and operator
guidance.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
How AI and MongoDB Help and to provide the right context to the large
language model, ensuring it generates
MongoDB Atlas is the only multi-cloud responses based on factory data without
developer data platform designed to hallucinating. Such a system can result in
accelerate and simplify how developers work positive business outcomes. Here’s how the
with data. Using MongoDB Atlas, developers system can look with MongoDB Atlas as the AI
can power end-to-end value chain data store:
optimization with AI/ML, advanced analytics,
and real-time data processing for innovative
mobile, edge, and IoT applications.
Figure 10: Smart milling machine uses real-time data to predict failures
As seen in the Figure above, we have our topic. This allows data transformation and
prioritized machine, which is a milling machine enrichment before it even lands in the
with attached sensors that collect data such as database.
air temperature, rotational speed, torque, and
tool wear. This data will be processed through Once the data is in MongoDB, another
Atlas Stream Processing, enabling the application can react to sensor values and run
processing of streams of complex data using a trained model designed to predict failures.
the same data model and Query API used in The model results can be stored back into
Atlas databases. Atlas Stream Processing Atlas (between steps 4 and 5/Inference in the
enables developers to build aggregation Figure above). These results can then be
pipelines to continuously operate on streaming visualized using Atlas Charts. Finally, Atlas
data without the delays inherent to batch Triggers and Functions can be used to push
processing. Results can be continuously notifications to on-site users. This establishes
published to MongoDB Atlas or to a Kafka an end-to-end system for failure prediction.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Stage 3: Repair plan generators schedule, and resource availability information.
In this case, both structured and unstructured
Having identified the nature of the equipment data are involved. The repair instructions will
failures, the implementation of a come from the machine manual. For this
comprehensive repair strategy becomes process, MongoDB Atlas acts as the
paramount. First, we have to generate a operational data layer, seamlessly integrating
maintenance work order. This order should structured and unstructured data.
include repair instructions, spare parts needed,
Figure 11: MongoDB Atlas as the operational data layer for structured and unstructured data
The Figure above shows the process of work old work orders. What sets this approach apart
order generation using generative AI. First, we is the ability to use the same MongoDB
must extract chunks of information from a database to store structured data such as
milling machine's operating and repair manual, failure types, spare parts inventory, and
as well as from old work orders stored as PDF resource information. By employing the
files, and convert them into vectors. These aggregation framework to extract relevant
embeddings are then stored in MongoDB Atlas. information from structured data and vector
MongoDB's versatility allows for the storage of search to glean insights from vectors, the LLM
both structured and unstructured data within within the work order generator application
the same database. Leveraging Atlas Vector gains contextual understanding.
Search and Aggregation pipelines, we can
integrate this data to feed into a large This application seamlessly utilizes the LLM to
language model (LLM) powering a work order generate work orders with the correct
generator application. The LLM analyzes the template, filling in inventory and resource
data to generate the appropriate work order details through aggregations, and ultimately
and template, drawing from past examples. It creating repair plans based on machine
populates inventory and resource details using manuals. This application can run inside a
aggregation
US 866-237-8815 techniques
| INTL +1-650-440-4474 and structured
| [email protected] 2023 MongoDB, Inc.central
| © data. maintenance
All rights reserved. management system.
Finally, it generates a repair plan similar to the
Stage 4: Maintenance guidance generation
Figure 12: Using the RAG Approach for Operator Work Instructions
Let's walk through an example scenario here. then we can publish notifications down to the
The repair plan was generated in the last step. technician's application via Atlas Triggers and
Now, the computerized maintenance Functions.
management system (CMMS) has found some
service notes uploaded to the platform by In summary, we are essentially integrating AI
another technician, but they're written in and gen AI apps to implement an end-to-end
another language, let's say Spanish. We can predictive maintenance strategy (shown in the
use the RAG architecture again to intelligently Figure on the next page).
merge these service notes with the repair
instructions generated in the previous step.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Figure 13: Model chaining with unified data store
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Knowledge
Management
Figure| 14:
US 866-237-8815 INTLPreserving knowledge
+1-650-440-4474 in MongoDB
| [email protected] Atlas Inc. All rights reserved.
| © 2023 MongoDB,
Knowledge collection from shop With a knowledge management application on
the shopfloor all the input from the production
floor
equipment as well as from the workers can be
The time being wasted for consolidating data
collected. The application takes all the
from different systems to take decisions on a
structured and unstructured text input and
daily basis can be heavily reduced by using
categorizes it into one of many categories
gen AI. A shift leader, as an example, spends a
(defects, breakdowns, alarms, etc.).
lot of time to collects data from different
sources like MES, SCADA, or from handwritten
Having all that data collected, contextualized,
notes of the night shift workers to get an
and indexed allows a chatbot application to
overview about the condition of the equipment
get an immediate overview on the status of
after the last shift. Traditionally it takes a lot of
the shopfloor by prompting: Provide me a list
time to collect all that data from various
of machines with problems in the last shift,
sources and locations for getting a holistic
followed by prompting on how to solve that
overview to understand current safety,
problem. The RAG application can therefore
maintenance, inventory and quality needs.
use all the preserved information from the
experienced workers, stored as vectors in
A shifter leader may struggle to make timely
MongoDB Atlas.
decisions due to the fragmented nature of the
data, leading to delays in addressing
production issues and optimizing workflows.
Improving staff productivity at Enel The solution uses Amazon Titan, a family of
using Amazon Bedrock models exclusive to Amazon Bedrock.
Specifically, it uses the Amazon Titan Text
Enel is a leading integrated electric utility
Embeddings model to generate
with a presence across 32 countries and an
embeddings (vectors capturing semantics
82-GW generation capacity
of text) from Enel’s knowledge base, which
Enel identified the opportunity to use consists of a series of runbooks containing
generative AI to boost IT service desk incidents classes, preconditions, root
efficiency by extending automation to causes, resolutions steps, and operations
nontrivial tasks through basic information related to the applications.
troubleshooting, providing resolution steps Embeddings are computed and persisted in
and ticket routing without human a vector database instance using MongoDB
involvement. Atlas Vector Search, which supports
similarity search.
The solution is designed around a Learn more
retrieval-augmented generation
architecture using Amazon Bedrock.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Other Notable Use Cases
Quality Control
Personalized Product
and Defect Detection
Recommendations
Computer or machine vision can be used to
AI can be used to analyze user behavior
identify irregularities in the products as they
and preferences to deliver personalized
are manufactured. This ensures that product
product recommendations via a mobile or
standards are met with precision.
a web app, enhancing customer satisfaction
and driving sales.
Production Optimization
By analyzing time series data from sensors
installed on production lines, waste can be
identified and reduced, thereby improving
throughput and efficiency.
Contact
Information
Dr. Humza Akhtar Raphael Schor
Manufacturing & Motion Manufacturing & Motion
Industry Solutions Principal Industry Solutions Principal
[email protected] [email protected]
Atlas for Industries
Telecommunications
and Media
Faced with high
operational costs and
low margins, the
telecommunications and
media industries are
exploring new ways to
create value and enhance
revenue streams with AI.
AI investments in the telco & media industry
Generative AI is set to revolutionize the media These are some of the key findings from an
and telecommunications sectors, significantly industry survey of over 300 digital leaders
impacting content creation, distribution, across 50 countries:
service assurance, anti-fraud measures, and
overall industry dynamics. ● Referral Traffic: 63% are worried
about declining traffic from social
Content Creation and Distribution media, with significant drops reported
AI will dramatically enhance both content from platforms like Facebook and
creation and distribution. Search generative X/Twitter. Traffic to news sites from
experiences (SGE) and AI-driven chatbots are Facebook fell 48% in 2023, with traffic
expected to provide faster, more intuitive from X/Twitter declining by 27%.
access to information, shifting audience flows ● Platform Strategies: Publishers are
and financial dynamics within the industry. As prioritizing platforms like WhatsApp
referral traffic from traditional social media and Instagram while maintaining
platforms like Facebook and X (formerly interest in video networks such as
Twitter) declines, media companies face the TikTok and YouTube.
challenge of maintaining their audience and ● Content Focus: There is a strong
revenue streams. emphasis on increasing video,
newsletters, and podcasts to engage
In response to these challenges, media audiences, while concerns about news
organizations are focusing on building direct fatigue and selective avoidance persist.
relationships with their audiences and
investing in intellectual property protection. Revenue and AI Integration
This strategy aims to reduce dependence on Subscription and membership models are seen
major tech platforms but risks alienating as vital revenue streams, with publishers also
younger, tech-savvy audiences who are exploring lucrative AI platform licensing deals.
already comfortable with algorithmically AI is primarily viewed as a tool for backend
generated news. automation, improved recommendations, and
commercial uses.
Telecom providers today are utilizing an The models can adapt over time, learning from
advanced array of techniques for detecting new data and emerging fraud tactics, thus
and preventing fraud, constantly adjusting to enabling real-time detection and the
the dynamic nature of telecom fraud. Routine automation of fraud prevention measures,
activities for detecting fraud consist of reducing manual checks, and speeding up
tracking unusual call trends and data usage, response times.
along with safeguarding against SIM swap
incidents, a method frequently used for To deal with fraud successfully, a multitude of
identity theft. To prevent fraud, strategies are data dimensions need to be considered, with
applied at various levels, starting with reaction time being a critical factor in
stringent verification for new customers, preventing the worst things from happening,
during SIM swaps, or for transactions with so the solution must also support fast,
elevated risk, taking into account the unique sub-second decisions. By vectorizing the data
risk profile of each customer. with an appropriate ML model, normal
(healthy) business can be defined, and in turn,
Machine learning offers telecommunications deviations from the norm identified, for
companies a powerful tool to enhance their instance, suspicious user activities. In addition
fraud detection and prevention capabilities by to Atlas Vector Search, the MongoDB Query
training ML models on historical data like Call API supports stream processing, simplifying
Detail Records (CDR). Moreover, these data ingestion from various sources.
algorithms can assess the individual risk
profile of each customer, tailoring detection One of the customers utilizing MongoDB for its
and prevention strategies to their specific anti-fraud strategy is AT&T. The company
patterns of use. selected MongoDB for its ability to ingest and
store rapidly changing data. Moreover,
MongoDB Atlas meets AT&T's key requirements
for performance, availability, and security. The
fraud prevention platform now operates over
50 different AI models, utilizes more than 1,200
features, stores 30TB of data, and processes
up to 20 million events daily. Events are
processed in under 200 milliseconds, enabling
real-time fraud detection and prevention. As a
result, fraud—particularly iPhone-related fraud,
which was previously a $1 billion problem—has
been reduced by over 80%.
Content Discovery
Content Discovery
Solution demo
Discover how to leverage MongoDB
and Vector Search effectively to
create more engaging, personalized
user experiences.
Figure 16: Reference architecture for to creating more engaging, personalized user experiences with
MongoDB and Vector Search
Search Generative
Content Discovery
Experiences (SGE)
Search generative experiences (SGE) represent the most relevant articles for a topic of
a transformative shift in how media and interest, but also combining them into a
content industries interact with information concise summary, elevating the value for the
retrieval and user engagement. SGE utilizes
customer.
artificial intelligence to generate search results
that are not just links to existing content, but
Smart Conversational Interfaces
comprehensive, synthesized responses tailored
to the user's query. By providing immediate, Content retrieval products that are accessible
concise answers, SGE helps media platforms from smartwatches, AR and VR glasses, lapel
keep users engaged longer, reducing bounce pins, and other wearable devices, or from the
rates and increasing the likelihood of deeper car, allow for the consumption of data in more
interaction with the content. day-to-day situations.
Contact
Information
Benjamin Lorenz
Telco & Media Industry
Solutions Principal
[email protected]
Atlas for Industries
Retail
AI is transforming retailers’
ability to maximize their
competitive advantage
through better understanding
of their customers and
improving their operating
margins through intelligent
decision making.
Artificial Intelligence (AI) is revolutionizing the
retail industry across the globe, driving
innovation and enhancing efficiency. AI itself is
evolving from traditional AI to generative AI.
Retail leaders who invest more in To ensure that the right products in the right
technology to improve customer quantities are in the right place at the right
experience are 17% more likely to time, back office teams leveraged machine
outperform other retailers in learning arithmetic algorithms for years.
organic sales growth and get
about 37% more sales from As technology has advanced and the barrier
digital channels (Bain and for entry is lowered for adopting AI, retailers
Retail is a customer centric business. analytics from AI algorithms. Also, it's now
Customers have more choice than ever in possible to incorporate factors such as the
where they purchase a product. To retain and current shopping cart or basket and customer
grow their customer base, retailers need to clickstream and trending purchases across
keep innovating in order to offer each shoppers.
customer a differentiated buying experience.
To do this, it is necessary to use a large The first step in truly understanding the
amount of data from the customers such as customer is to build a customer operational
buying patterns, interests, and interactions data store that combines data from disparate
and to be able to quickly make complex systems and silos in the organization: support,
decisions on that data. e-commerce transactions, in-store interactions,
wish lists, reviews, and more. MongoDB’s
One of the key customer interactions in an flexible document model enables bringing
e-commerce experience is search. Through the data of different types and formats in one
implementation of full-text search engines, document to get a clear view of the customer
customers can more easily find items that in one place. As the retailer captures more
match their search, and retailers are given the data points about the customer, they can
opportunity to rank those results in a way that easily add fields without the need for
will give the customer the best option. downtime due to schema changes.
Traditionally, decisions on how to rank search
results in a personalized way were made by Then comes the ability to run analytics in
segmentation of customers through data real time rather than retroactively in another
acquisition from various operational systems, separate system. MongoDB’s architecture
moving it all into a data warehouse, and allows for workload isolation, meaning
subsequently running classical AI with various operational workloads (the customer's actions
machine learning algorithms on such data. on the ecommerce site) and the analytical or
Typically, this would run in a batch mode AI workload (calculating what the next best
(every 24, 48, or even 72 hours or a few days), offer should be) can be run simultaneously
and the next time a customer logs in, they will without interrupting the other. Retailers can
have a personalized experience. It does not, build dynamic ranking by using the MongoDB
however, capture the customer's true desire in aggregation framework for advanced
real time. analytical queries or triggering an AI model in
real time to give an answer that can be
Modern retailers augment search ranking with embedded into the search ranking.
data from real-time responses and/or
41
The benefit of an all-in-one platform is huge
here, as instead of having to update your
Vector search technology in retail
search indexing to incorporate your AI
provides notable economic
augmentation, MongoDB has Lucene Search
benefits. As highlighted by
built in. This whole flow can be completed in
Deloitte: Sales Uplift and
one data platform automatically—as your data
Customer Engagement, Deloitte
is being augmented with AI results, the search
indexing will sync to match.
reported that retailers
implementing personalized
MongoDB Atlas Vector Search brings the next search have seen a sales uplift of
generation of search capability. By using LLMs about 40% (Deloitte).
to create vector embeddings for each product
and then turning on a vector index, retailers
are able to offer semantic search to their
customers. AI will calculate the complex
similarities between items in vector space and
give the customer a unique set of results READ MORE
matched to their true desire. AI-Enhanced Search in
E-commerce With MongoDB
Figure 18: Architecture of an AI-enhanced search engine explaining the different MongoDB Atlas
components and Data Intelligent Platforms and workflows used for data cleaning and preparation,
product scoring, dynamic pricing, and vector search
Delivery Hero Helps Customers The solution was MongoDB Atlas, a
Navigate more than 100 Million scalable, high-performing developer data
Products with MongoDB Atlas Search platform that integrates easily with the AI
ecosystem. With it the new Item
Replacement Tool is being piloted first in the
Delivery Hero, a food delivery service
Middle East. By providing personalized
based out of Germany, has built a new Item
recommendations against live inventory,
Replacement Tool providing
Delivery Hero expects to see an increase in
hyper-personalized product
its monthly gross merchandise value.
recommendations in real time using
state-of-the-art AI models and MongoDB
Atlas Vector Search. “We can implement things that
would take six months on other
The challenge was that around 10% of the
platforms in just a few days.”
inventory is perishable produce that can
quickly go out of stock. Without being able Learn more
to recommend a suitable alternative to the Andrii Hrachov
customer, the company risks revenue loss Principal Software Engineer at Delivery
and customer churn. Hero
Personalized
Advertising, and marketing material are vital
to capturing a customer’s interest and driving
The above shows a reference architecture By doing so, media organizations can suggest
highlighting where MongoDB can be leveraged content that aligns more closely with individual
to achieve AI-powered personalization. By preferences and past interactions. This not only
leveraging user data and the enhances user engagement but also increases
multi-dimensional vectorization of media the likelihood of converting free users into
content, MongoDB Atlas can be applied to paying subscribers.
multiple AI use cases. This allows for the
utilization of media channels to more
effectively improve end-user experiences.
Figure 20: Example of the data flow for an AI-generated personalized newsletter. The prompt is
entered by a user on the left hand side and context is added via the vectorized data in
MongoDB—wishlist, loyalty data, purchase history, and customer behavior. Using RAG, the LLM
can produce a personalized newsletter per customer in seconds, allowing the retailer to create
vast amounts of customized content
& Predictive
products. It also enables data-driven
decisions, leading to improved sales forecasts
● [1] This produces events about the cost increase and places them" with "The price
increases are placed in the message stream where the event queue makes them available.
All microservices are listening for such messages.
● [2-3-4] The pricing microservice consumes the event, analyzes it against existing data,
and further conveys the new pricing into the message stream.
● [5 - 6] The database pushes those messages to the event queue, which makes them
available to all consumers listening for messages. Microservices directly impacted by
pricing changes—such as those that manage inventory, marketing, promotions, coupons,
point of sale (POS), and the e-commerce provider’s order management system
(OMS)—consume the price change events and update their individual databases
accordingly.
● [7] The centralized database aggregates and persists events, enriches event streams with
data from other sources, including historical data, and provides a central repository for
multiple event streams.
Conversational Conversational chatbots powered by
generative AI are revolutionizing the retail
Chatbots industry by enhancing customer service.
These chatbots can handle a wide range of
customer inquiries, from product
recommendations to order tracking,
providing instant and accurate responses.
This reduces wait times and improves the
overall customer experience, leading to
higher satisfaction and increased loyalty.
Additionally, chatbots can operate on real-
time data 24/7, ensuring customers receive
support at any time, which is especially
beneficial for global retailers.
Beyond customer service, AI chatbots are also Operational efficiency is another area where
transforming marketing and sales strategies in AI chatbots are making a significant impact.
retail. They can analyze customer data to By automating routine tasks such as
personalize shopping experiences, offering answering FAQs, managing inventory inquiries,
tailored recommendations and promotions and processing returns, chatbots free up
based on individual preferences and behavior. employees to focus on more complex and
This personalization helps retailers boost value-added activities. This not only reduces
conversion rates and increase sales. Moreover, operational costs but also improves accuracy
chatbots can engage customers through and consistency in service delivery.
various digital channels, including social Furthermore, the data collected by chatbots
media, websites, and messaging apps, can provide valuable insights into customer
broadening the reach and effectiveness of preferences and behavior, helping retailers
marketing campaigns. refine their strategies and improve their
offerings.
Figure 22: Example of the data flow for a chatbot RAG architecture
L’Oréal Improves App Performance The result was reducing latency from
and Velocity with MongoDB Atlas seconds to just 10 milliseconds.
The challenge for L’Oréal was to complete “MongoDB Atlas doesnʼt just solve
complex calculations on vast volumes of our performance issues It makes
data—without causing latency. The solution life easier. We have a hyper agile
was simplifying management and
maintenance while boosting performance
DevOps model.”
with MongoDB Atlas. Learn more
Moutia Khatiri
CTO, Tech Accelerator, L’Oréal
Artificial intelligence is revolutionizing the way benefits of enhanced and semantic search
retailers enhance their competitive edge by capabilities, create targeted marketing content
providing deeper insights into customer based on current market trends, effectively
behavior and optimizing profit margins utilize predictive analytics for demand
through smart decision-making processes. By forecasting, employ conversational chatbots,
incorporating both traditional and generative and significantly elevate the overall customer
AI, retailers can harness the experience.
Contact
Information
Genevieve Broadhead Prashant Juttukonda
MongoDB Global Lead, Retail Industry
Retail Solutions Solutions Principal
[email protected] [email protected]
Rodrigo Leal
Retail Industry
Solutions Principal
[email protected]
Atlas for Industries
Financial
Services
Analyst firms, Celent and
McKinsey, both identify
financial services as one
of the industries most
likely to be transformed
AI. While the initial
benefits are likely to be
process-related as firms
slowly ramp up adoption,
it won't be long before
customer interactions are
also completely
re-invented by AI-driven
experiences.
AI investments in the financial services industry
Interest in leveraging AI
for banking use cases is
expected to increase in
the next few years.
While the industry has always had to deal with A strategic and comprehensive approach
persistent issues like risk management and encompassing aspects of technology, data,
governance, adopting generative AI and ethics, and organizational readiness is critical
machine learning introduces new challenges to overcoming these challenges:
that AI specialists have always dealt with, like
inherent biases and ethical concerns. One ● Hallucination mitigation: One promising
challenge that stands out for generative AI is strategy is using retrieval augmented
hallucination—the generation of content that is generation (RAG) to mitigate hallucination
not accurate, factual, or reflective of the real in gen AI models–incorporating
world. information retrieval mechanisms to
enhance the generation process to ensure
Generative AI models, especially in natural content is grounded in real-world
language processing, might generate text that knowledge. Atlas Vector Search is a
is coherent and contextually appropriate but popular mechanism to support RAG,
lacks factual accuracy. This poses challenges, which uses vector embeddings to retrieve
for example: relevant documents based on the input
query.
● Misleading financial planning advice: ● Data quality and availability: Take a
In financial advisory services, hallucinated step back before adopting AI to ensure
information may result in misleading the quality, relevance, and accuracy of
advice, leading to unexpected risks or data being used for AI training and
missed opportunities. decision-making can be accessed in real
● Incorrect risk assessments for lending: time.
Inaccurate risk profiles may lead to poor ● AI education: The key is to invest in
risk assessments for loan applicants, training programs to address skill gaps,
which can cause financial institutions to create a culture of learning and
approve a loan with a higher risk of development, and promote awareness
default than the firm would normally about vulnerabilities.
accept. ● Develop new governance, frameworks,
● Sensitive information in generated and controls: Before going live, create
text: When generating text, models may safe and secure environments for testing.
inadvertently include sensitive information ● Implement monitoring systems:
from the training data. Adversaries can Measure and understand financial
craft input prompts to coax the model impacts, change impacts, scale, and
into generating outputs that expose complexity associated with AI.
confidential details present in the training ● Security and privacy: Implement secure
corpus. data access privileges and authentication
measures to safeguard AI models and the
data they rely on.
Relationship Management
Support with Chat Bots
One key service that relationship managers ● When are the next coupon dates for
provide to their private banking customers bonds in the portfolio?
is aggregating and condensing information. ● What has been the cost of
transactions for a given portfolio?
Because banks typically operate on ● What would be a summary of our
fragmented infrastructure, this can latest research?
require a lot of detailed knowledge ● Please generate a summary of my
about this infrastructure and how conversation with the client.
to source information such as:
Data loading from raw data sources mentioned text chunks into vector
Begin by extracting data from databases or representations. The chunks can then
files. Examples of data can include customer be indexed using MongoDB vector
profiles, transaction histories, and product indexing that will then enable an
information. The source of the data may be efficient retrieval of vectorized data
located in different databases or files and can using Atlas Vector Search.
be consolidated into MongoDB and serve as
an operational data store to provide a Pre-query Retrieval
customer 360-degree view. Advanced RAG methods explore various
transformations of user queries to enhance
Data indexing: chunking and vectorization retrieval accuracy before it is passed to the
● Data chunking involves breaking LLM for response generation. As an example, If
down the documents (such as the RM asks for all recent fund transfers for a
customer service records, emails, and given customer, the gen AI application can
transaction logs) into smaller, more pre-process the query by first checking if the
manageable chunks (eg. individual RM has access privileges to the customer data.
interactions or sentences) and then It can then refine the query to add a specific
creating an index that allows for quick time period, and add the relevant documents
searching and retrieval of relevant retrieved from the vector search. This will
information. augment the search, giving it context to
● Data vectorization is implemented by optimize relevancy of the results.
using the embedding model (eg.
Cohere, Anthropic, etc) to convert the
above
Post retrieval also be trained to specialize in a specific
With the transformed query executed, the domain for accuracy and efficiency. For
relevant documents will be retrieved (such as example in the context of financial advisory
with regular, text, vector or even hybrid support, an agent could specialized in
searches). The retrieved documents (such as Natural Language Understanding of
the customer details and transactions) can be financial market information, another AI
further filtered and reranked to provide an Agent can specialist in Risk Assessment for
even more relevant set of documents to be evaluating client and product risk profiles
used as a data augmentation when prompting while yet another can specialized in
the LLM later. Compliance for regulatory adherence to
ensure the investment decisions especially
LLM for corporate clients are also compliant.
The LLM engine can be a single or a group of Together they can be “composed” using
LLM engines deployed on an AI platform (eg. orchestration tools (eg. Langchain and
Fireworks.AI) to work together leveraging LLamaIndex) to support the tasks required
different specialization to provide a relevant, for the generative AI interactions.
accurate and combined response to the users.
Let us generalize and refer to these engines as
AI agents so to not limit our imagination to
just Language Models (be it large or small).
These AI Agents not only can perform
orchestration tasks but can agent can
Risk Management
and Regulatory Compliance
Banks are tasked not only with particular, has shown promise in trade
safeguarding customer assets but also with surveillance, offering a more dynamic and
detecting fraud, verifying customer comprehensive approach to fraud
identities (KYC), supporting sanctions prevention.
regimes, and preventing various illegal
activities (AML). The challenge is magnified Modernizing fraud systems improves
by the sheer volume and complexity of detection accuracy, increases flexibility and
regulations, making the integration of new scalability, enhances security, and reduces
rules into bank infrastructure costly, the cost of operation. MongoDB's developer
time-consuming, and often inadequate. For data platform is a great fit for modern
instance, when the EU's Fifth Anti-Money fraud prevention systems due to:
Laundering Directive was implemented,
Data visibility: MongoDB’s document
it broadened regulations to cover virtual
model incorporates any kind of data no
currencies and prepaid cards. Banks had
matter how often it changes, allowing you
to swiftly update their onboarding
to create a holistic picture of customers to
processes and software, train staff, and
better predict transaction anomalies in real
possibly update their customer interfaces to
time. As well as process large amounts of
comply with these new requirements.
data and analyze in real-time to identify
AI offers a transformative approach to suspicious activity. Change Streams in
fraud detection and risk management by MongoDB enable the monitoring of
automating the interpretation of database changes in real time meaning
regulations, supporting data cleansing, and immediate detection of suspicious activity
enhancing the efficacy of surveillance or transactions. By receiving instant
systems. Unlike static, rules-based notifications, fraud prevention teams can
frameworks that may miss or misidentify quickly respond to potential threats.
fraud due to narrow scope or limited data,
Real-time data processing: By leveraging
AI can adaptively learn and analyze vast
MongoDB as the operational data store
datasets to identify suspicious activities
(ODS) you can handle large amounts of
more accurately. Machine learning, in
data to effortlessly capture, store, and
process high-volume transactional data in Besides the use of AI/ML techniques, there
real-time. MongoDB’s aggregation the option of leveraging vector data store
framework allows for complex data capabilities such as MongoDB Atlas Vector
processing and analytics directly within the Search to perform fraud and money
database, crucial for implementing laundering detection. This can be also be
sophisticated, rules-based fraud detection used in conjunction with the AI/ML
algorithms. approach to create a hybrid
implementation to further enhance the
Integrating with modern AI/ML accuracy of detection and minimize false
technologies: MongoDB’s flexible data positives. Learn More
model makes it easier to integrate with
AI/ML platforms to develop models for Below is an example architecture of an
real-time, predictive fraud prevention that ML-based fraud solution leveraging
are trained on the most accurate and MongoDB and Databricks. Key features
relevant data available. include data completeness through
integration with external sources, real-time
Regulatory compliance: Compliance with processing, AI/ML modeling to identify
regulations like PCI, GDPR, CCPA, and PSD2 potential fraud patterns, real-time
is crucial in the finance sector. MongoDB's monitoring for instant analysis, and model
security features ensure data handling observability for full data visibility into
meets these regulatory standards. fraud behaviors. Learn more
Enhanced authentication and role-based
authorization controls further fortify the
system.
Financial institutions, encompassing both retail MongoDB can store vast amounts of both live
banks and capital market firms, handle a and historical data, regardless of its format,
broad spectrum of documents critical to their which is typically needed for AI applications.
operations. Retail banks focus on contracts, It offers vector search capabilities essential for
policies, credit memos, underwriting retrieval-augmented generation (RAG).
documents, and regulatory filings, which are MongoDB supports transactions, ensuring
pivotal for daily banking services. On the other data accuracy and consistency for AI model
hand, capital market firms delve into company retraining with live data. It facilitates data
filings, transcripts, reports, and intricate data access for both deterministic algorithms and
sets to grasp global market dynamics and risk AI-driven rules through a single interface.
assessments. MongoDB boasts a strong partnership
ecosystem, including companies like Radiant
These documents often arrive in unstructured AI and Mistral LLM, to speed up solution
formats, presenting challenges in efficiently development.
locating and synthesizing the necessary
information. While retail banks aim to
streamline customer and internal operations,
Discover more about MongoDB’s
capital market firms prioritize the rapid and
full full-text search and vector
effective analysis of diverse data to inform
search capabilities.
their investment strategies. Both retail banks
and capital market firms allocate considerable
time to searching for and condensing
information from documents internally, To achieve search and summarization using
resulting in reduced direct engagement with RAG, a conceptual architecture leveraging
their clients. MongoDB and generative AI can be used.
Generative AI can streamline the process This architecture aims to centralize document
of finding and integrating information from storage, enhance search capabilities, and
documents by using NLP and machine automate the extraction and synthesis of
learning to understand and summarize information. The following diagram illustrates
content. This reduces the need for manual this conceptual architecture.
searches, allowing bank staff to access
relevant information more quickly.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Figure 27: Atlas hybrid search workflow: query embedding, document retrieval, summarization, user
presentation
ESG Analysis
The profound impact of environmental, banks to also improve their green lending
social, and governance (ESG) principles is practices. This shift is strategic for financial
evident, driven by regulatory changes, institutions, attracting clients, managing
especially in Europe, compelling financial risks, and creating long-term value.
institutions to integrate ESG into investment
and lending decisions. Regulations such as However, financial institutions face many
the EU Sustainable Finance Disclosure challenges in managing different aspects
Regulation (SFDR) and the EU Taxonomy of improving their ESG analysis. The key
Regulation are examples of such directives challenges include defining and aligning
that require financial institutions to consider standards and processes, and managing
environmental sustainability in their the flood of rapidly changing and varied
operations and investment products. data to be included for ESG analysis
Investors' demand for sustainable options purposes.
has surged, leading to increased
ESG-focused funds. The regulatory and
commercial requirements in turn, drive
AI can help to address these key challenges
in not only an automatic but also adaptive
manner via techniques like machine learning.
Financial institutions and ESG solution
providers have already leveraged AI to
extract insights from corporate reports,
social media, and environmental data,
improving the accuracy and depth of ESG
analysis. As the market demands a more
sustainable and equitable society, predictive
AI combined with generative AI can also help
to reduce bias in lending to create a fairer and
more inclusive financing while improving the
predictive powers. The power of AI can help
facilitate the development of sophisticated
sustainability models and strategies, marking
a leap forward in integrating ESG into broader
Incorporating LLMs enhances MongoDB's
financial and corporate practices.
capacity to process ESG textual content,
automating sentiment extraction,
MongoDB's dynamic architecture
summarization, and trend identification.
revolutionizes ESG data management,
Combining LLMs with vector data
handling semi-structured and unstructured
management capabilities, generative AI
data. Its flexible schema nature allows the
applications can be created to interpret the
adaptation of data models as ESG strategies
complex and evolving sustainability taxonomy
evolve. Advanced text search capabilities
and guide the investment and financing
efficiently analyze vast semi-structured data
processes in a compliant manner. This
for informed ESG reporting. Support for vector
AI-driven approach, supported by MongoDB's
search enriches ESG analysis with multimedia
robust data management, offers a
content insights.
sophisticated means of analyzing extensive
narrative data in ESG reporting.
Figure 28: Blueprint for Enterprise ESG Solution Architecture Using MongoDB
Transforming Credit
Scoring with AI
capture process and also the performance declination: when it comes to credit
representation proves highly conducive for reasons behind it is crucial. Let’s explore
organizing the multifaceted information how MongoDB and large language models
within credit applications, encompassing (LLMs) can shed light on XGBoost model
personal, financial, and employment details. predictions (the model used in this tutorial).
The risk profiling ML pipeline employed
Enhance credit scoring with AI: provides a probability score that defines the
Leveraging MongoDB’s developer data risk associated with the profile for product
platform — an integrated suite of data recommendation. This message is
services centered around a cloud database communicated back to the user in a
— we can create a comprehensive templatized manner where only the final
customer/user banking profile by status of the application is communicated
combining relevant data points. to the end user.
Figure 29: Architecture of the data processing pipeline for the predicting probability of
delinquency and credit scoring
Below you will see the architecture
explaining credit scoring using an LLM.
Learn how with this credit card Credit card application with generative AI tutorial
Assessing Business Loan
Risks
Business loans are a cornerstone of banking For instance, ChatGPT-4.0 can be prompted to
operations, providing significant benefits to perform a flood risk assessment. However, due
both financial institutions and the broader to insufficient flood data, it first indicated the
economy. In 2023, the value of commercial location was in a low flood risk zone but later
and industrial loans at commercial banks in claimed it had significant flooding. This
the United States reached nearly 2.8 trillion inconsistency can confuse users and lead to
U.S. dollars. However, these loans involve incorrect risk assessments.
significant risks, including credit and business
risks.
Figure 33: Choosing a location on the map and writing a brief plan description
Upon submission, the AI performs a RAG
analysis. Using prompt engineering, it delivers
a simplified business analysis considering the
location and flood risk data from external
sources. McKinsey's survey found that 65% of
respondents report their organizations are
regularly using generative AI, including RAG, in
at least one business function. This
demonstrates the increasing adoption and
integration of RAG within business processes.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Emerging use cases
for AI in payments
A lack of developer capacity is one of the are several areas in which banks can apply AI
biggest challenges for banks when delivering to unlock new revenue streams and efficiency
payment product innovation. Banks believe the gains. The image below provides a high-level
product enhancements they could not deliver view on eight of the principal themes and
in the past two years due to resource areas. This is not an exhaustive view but does
constraints would have supported a 5.3% demonstrate the depth and breadth of current
growth in payments revenues. With this in opportunities. In each example, there are
mind and the revolutionary transformation already banks that have begun to bring
with the integration of AI, it is imperative to services or enhancements to market using AI
consider how to free up developer resources to technologies or are otherwise experimenting
make the most of these opportunities. There with the technology.
Below are some areas according to tech value-added services making the most of
analyst Celent in which AI can be leveraged payments data. Analysis of payables,
across financial institutions in front, middle receivables, and balance information in
and back office operations. aggregate can be used to deliver a range of
different insights and service improvements
Front office for new and richer insights: in real-time. Below are three areas where
AI has huge potential to support
financial institutions can use AI to enhance ● Product and service
their customer-facing services: enhancements: Personalizing
services is a huge area AI can help.
● Cash flow analysis and
forecasting: Using the information Cross-functional capabilities to
in payment messages to provide increasing agility and efficiency
granular and real-time insights into AI technologies can also be used to support
the cash position for a client. several cross-functional use cases within a
● Working capital optimization: bank:
Using AI in conjunction with
payments data can support a range ● Using gen AI to support code
of use cases around liquidity and generation: This is a huge
management with insights and opportunity for institutions to
recommendations. increase developer efficiency.
● Payment insights: Gen AI can ● Risk, compliance, and reporting:
support the delivery of payment and Technologies such as machine
account insights to customers. learning, NLP, computer vision, and
gen AI can all be deployed in
Middle and back office for workflow and various combinations to support
process improvements: processes such as risk management
AI and advanced analytics are already and regulatory reporting.
commonplace across large areas of middle
and back-office operations for payment
processing, but many potential
Dive into Celent’s report to learn
opportunities for further enhancement still
more about the use cases above
exist:
and understand how you can
● Payment processing harnessing the benefits of AI in
optimization: One use case for payments.
machine learning is to further
automate the payments repair
process. AI technologies to improve
transaction routing is another
important opportunity. A third would
be in message translation or
conversion between older standards
and ISO formats, e.g. to translate
unstructured name and address in
ISO 8583 messages to structured
ISO 20022 messages.
● Risk and fraud: As with the use
case for payments repair, applying
AI technologies in transaction
screening can reduce the need for
manual intervention.
Regulatory Compliance and
Code Change Assistance
The regulatory landscape for banks has grown AI has the capacity to revolutionize
increasingly complex, demanding significant compliance by automating the translation of
resources for the implementation of numerous regulatory texts into actionable data
regulations. Traditionally, adapting to new requirements and validating compliance
regulations has required the manual through intelligent analysis. This approach is
translation of legal text into code, provisioning not without its challenges, as AI-based
of data, and thorough quality control—a systems may produce non-deterministic
process that is both costly and outcomes and unexpected errors. However, the
time-consuming, often leading to incomplete ability to rapidly adapt to new regulations and
or insufficient compliance. For instance, to provide detailed records of compliance
comply with the Basel III, Dodd-Frank, PSD2, processes can significantly enhance regulatory
or various other regulations, developers must adherence.
undertake extensive coding changes to
accommodate the requirements laid out in
thousands of pages of documentation.
Recent research highlights that generative in identifying gaps and interpreting new
AI not only identifies regulatory obligations regulatory requirements improves. Eventually,
and gaps but also provides accurate code generative AI may even write new controls,
recommendations. For example, generative driving greater efficiency.
AI can help a bank ingest all applicable
regulatory requirements and internal This seamless process is facilitated by using a
controls, quickly alerting it to any multimodal database like MongoDB, which
discrepancies. By harnessing its addresses the common issue of documents
comprehensive and sophisticated being scattered and lacking proper metadata
capabilities, generative AI can significantly and documentation. MongoDB offers several
reduce manual effort and minimize errors, key differentiating features:
all while strictly adhering to factual data.
● Converged operational and AI data
It then determines an organization’s storage: MongoDB allows
obligations and evaluates its compliance. If organizations to keep all their
operational, unstructured, and vector
obligations are unmet, AI maps them to
data in one place, eliminating data
specific controls or generates new ones. As
silos and making it easier to access
AI models continue to train, their accuracy and manage.
● Efficient data handling: MongoDB 2. Translate to controls: Convert these
simplifies working with data at scale. summaries into actionable controls. In this
Centralizing data reduces manual context, control is a specific measure often
effort, leading to quicker and more implemented through code that ensures
accurate compliance processes. compliance with regulatory requirements by
● Enhanced search and generative AI automating necessary actions, such as
capabilities: MongoDB facilitates easy enforcing multi-factor authentication (MFA) for
searching through documents and certain transactions.
metadata, crucial for quickly
identifying regulatory requirements 3. Convert into executable code:
and checking them against existing
● Check the controls against the
controls. Utilizing RAG capabilities with
organization's existing internal
MongoDB provides better context to
controls, identify gaps, and create new
LLMs for generating code suggestions.
controls where needed.
Below is an example of a conceptual workflow ● Convert controls into executable code.
using MongoDB for code generation:
Figure 39: ChatGPT 4.0 Example of PSD2 regulation converted into code
Figure 40: MongoDB setup for multi-factor Figure 41: Transaction handling and strong
authentication customer authentication (SCA) enforcement
Other Notable Use Cases
Below are a few more examples of where AI These examples highlight several avenues for
can be leveraged in financial operations: integrating AI within financial institutions.
Embracing AI in financial applications
Risk modeling: AI can create synthetic promises enhanced risk management,
scenarios and data that can be used to operational efficiency, and superior customer
stress test financial systems and models experiences.
gen AI and MongoDB to modernize migrated 32 banking applications to AWS in just 30 days,
leveraging MongoDB to accelerate development and deliver
applications
enhanced services to its customer base.
Contact Information
Paul Rogers
Principal, Financial
Services Industry
[email protected]
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
Atlas for Industries
Insurance
Generative AI and large language models, The data that supports those applications
however, give us core NLP (natural language resides in operational data stores. If we want
processing) capabilities that are particularly both our applications and our AI to work with
well-suited to augmenting data processing real-time data, it should be accessible within
abilities. These capabilities, as applied to the the same underlying data store. What serves
claim-handling workflow, include: our apps should also serve our AI.
The above diagram shows practical use of core for example, these capabilities can reduce
NLP capabilities that include text and image data hotspots, resulting in lower processing
classification, interactive chat, entity times and costs, and improved customer
extraction, and text summarization. When experiences.
applied to a claim-handling workflow,
Before AI can transform our organizations,
we first have to bring it into our applications
The above diagram describes the different AI-powered apps and experiences: The
stages of AI adoption within enterprises, from interface and experiences that end-users
early interest to a pervasive and structural interact with, and app frameworks that
integration into processes and decisions. Many simplify the process of building AI experiences.
organizations struggle with moving out of
Level 2, analytics experimentation, to Level 3, Foundational models are extremely powerful,
deploying AI features within business but being trained on public datasets, they lack
applications, in order to deliver meaningful the domain knowledge and data context
business value and outcomes. needed to adequately support enterprise
applications.
The three layers of generative AI
Gen AI applications can be separated into This is where the second layer comes into play:
three major layers: the data and tooling that will glue everything
together and enable generative AI-enhanced
Compute and AI models: Underlying applications to be fully operational, moving
processing power plus foundational and your organization from the experimentation
embedding models. level of Level 2 into full-fledged production
apps of Level 3 maturity.
Tooling to fine tune and build applications:
Tools that provide foundational models with
context by feeding them proprietary data.
MongoDB’s operational database stores
AI-powered
proprietary structured and vector data, and
apps and experiences
makes it available to LLMs when an
application makes a request, effectively Data and tooling to use the
providing the context necessary for a model
foundational model to answer questions
Compute and AI models
beyond its initial knowledge boundary.
The above diagram illustrates in greater detail processed. Part of the processing includes the
the three layers of gen AI, showing the vectorization and subsequent persistence of
structured data flowing from the application those vectors in the operational data store,
into the operational database, raw where it can be readily accessed by the
unstructured data managed in object storage, application.
accessible by applications where it can be
In an inherently information-driven industry,
insurance companies ingest, analyze, and
process massive amounts of data. Whether
it’s agents and brokers selling more policies,
underwriters adequately pricing, renewing
and steering product portfolios, claim
handlers adjudicating claims, or service
representatives providing assurance and
support, data is at the heart of it all.
Given the volumes of data, and the amount IT investments. With its ability to streamline
of decision-making that needs to occur processes, enhance decision-making, and
based on it, insurance companies have a improve customer experiences with far less
myriad of technologies and IT support staff time, resources, and staff than traditional IT
within their technology investment portfolios. systems, AI offers insurers great promise.
It’s no surprise that AI is at the top of the list
when it comes to current or prospective
Decision-making support
Determining if a quoted rate needs adjustment
can take significant time and manual effort. So
can preparing and issuing renewals of existing
policies, another large portion of the
underwriters’ day-to-day responsibilities.
Automated underwriting workflows leveraging
AI are being employed to analyze and classify
risk with far less manual effort. This frees up
significant time and intellectual capital for the
underwriter.
Vast amounts of data analyzed by For every major business entity found within
underwriters are kept on the underwriters the underwriting process, such as broker,
desktop rather than IT-managed databases. policy, account, and claim, there is a wealth
MongoDB offers an unparalleled ability to of unstructured data sources, waiting to be
store data from a vast amount of sources leveraged by generative AI. MongoDB offers
and formats, and respond quickly to insurers a platform that consolidates
requests to ingest new data. As data and complex data from legacy systems, supports
requirements change, the document model modern applications, and extends those
allows insurers to simply add more data and same data assets to AI-augmented
fields without the costly change-cycle workflows. By eliminating the need for niche
associated with databases that rely on databases for these AI-specific workloads,
single, fixed structures. MongoDB reduces technology evaluation
and on-boarding time, development time,
and developer friction.
Litigation assistance
Claim details are not always black and white,
parties do not always act in good faith, and
Taming the flood of data insurers expend significant resources in the
AI is helping insurers make sense of a trove pursuit of resolving matters. AI is helping
of data faster and in real time. From natural insurers drive to resolution faster and even
language processing to image classification avoid litigation and subrogation altogether,
and vector embedding, all the pieces of the thanks to its ability to help analyze more data
puzzle are now on the board for insurers to more effectively and in less time.
make a generational leap forward when it
comes to transforming their IT systems and
business workflows for faster information
processing.
Risk prevention
Many insurers provide risk-assessment services
to customers using drones, sensors, or
cameras to capture and analyze data. This
data offers the promise of preventing losses
Claims experience
altogether for customers and lowering
Generating accurate impact assessments for
exposures, liability, and expenses for the
catastrophic events in a timely fashion in order
insurer. This is possible thanks to a
to inform the market of your exposure can now
combination of vector embedding and
be done with far less time, and with far more
traditional and generative AI models.
accuracy, by cross-referencing real-time and
historical claims experience data, thanks to the
power of generative AI and vector-
embedding of unstructured data.
Customer Accessing information consistently during a
customer service interaction, and expecting
Experience the representative to quickly interpret it, are
perennial challenges with any customer
service desk. Add in the volume, variety,
and complexity of information within
insurance and it’s easy to understand why
many insurers are investing heavily in
transformation of their customer experience
call center systems and processes.
Figure 45: Atlas Vector Search provides answers to user questions about insurance claims, including
calculations and detailed claim examples
AI-enhanced claim adjustment for auto insurance
By leveraging AI and vector image search, Centralized data management in MongoDB
this solution automates auto insurance Atlas enhances efficiency and consistency
claim adjustments. Accident photos are throughout the claims process.
compared to a database of past claims,
significantly accelerating damage estimates. Learn more
Figure 46: An image similarity query is performed, and the 5 top similar images are returned
Figure 47: PDF chunking, embedding creation, and storage, orchestrated with Superduper
Other Notable Use Cases
Contact
Information
Jeff Needham
Insurance Industry
Solutions Principal
[email protected]
Healthcare and
Life Sciences
AI is transforming
healthcare by enhancing
decision-making, improving
patient experiences, and
boosting operational
efficiency.
In life sciences, AI
accelerates research, drives
drug discovery, and
personalizes treatments,
leading to breakthroughs in
innovation and tailored
medicine.
Transformative Potential of Generative AI in
Healthcare and Life Sciences
The healthcare and life sciences sectors have Additionally, 57% of healthcare providers
witnessed a significant surge in the application report that excessive documentation
of artificial intelligence (AI) and machine contributes to clinician burnout. This
learning (ML) over recent years. These administrative overload not only affects the
technologies have been increasingly integrated well-being of healthcare providers but also has
into various aspects of healthcare, from a negative impact on patient care, with two
diagnostics and treatment planning to out of three patients reporting a lack of
operational efficiencies and patient empathy from their healthcare providers due
engagement. The acceleration of AI and ML to these constraints.
adoption is evident in the exponential growth
of FDA-approved algorithms, reflecting a The solution: Leveraging gen AI to
broader acceptance and reliance on these alleviate administrative burden
technologies to enhance healthcare delivery.
Gen AI has the potential to dramatically
reduce the administrative workload on
However, the advent of generative AI marks a
healthcare professionals, allowing them to
new frontier in this evolution, bringing the
focus more on patient care. Technologies such
potential to disrupt and revolutionize even
as ambient listening, combined with advanced
more areas of the industry. Unlike traditional
patient records, could automate much of the
AI, which primarily focuses on analyzing and
documentation process, reducing the time
predicting, gen AI can autonomously read and
clinicians spend on EHRs. By harnessing these
interpret multimodal data, and generate new
capabilities, healthcare providers can reclaim
content, offering unprecedented possibilities in
valuable time, improve the quality of patient
automation, personalized medicine, drug
interactions, and ultimately enhance overall
discovery, and beyond.
patient outcomes.
Generative AI offers a transformative solution including reports, and metadata from DICOM
by automating the extraction, analysis, and files. Its support for vector search capabilities
summarization of information from medical is essential for implementing retrieval-
images. By leveraging advanced NLP, augmented generation (RAG) approaches,
embeddings, and ML techniques, generative AI which enhance the accuracy and relevance of
can rapidly analyze large datasets, identify key AI-generated summaries.
insights, and generate accurate,
comprehensive summaries. This significantly Medical visual question answering (MVQA)
reduces the time radiologists spend on manual Building on the broader capabilities of
data review, allowing them to focus more on generative AI in radiology, MVQA specifically
patient care and decision-making. targets the intersection of medical imaging,
clinical reporting, and NLP. In this sub-use
For instance, AI-powered tools can be case, generative AI is utilized to not only
integrated into radiology workflows to generate summaries but also to directly
automatically generate impressions and answer clinician-specific queries regarding
diagnostic summaries based on imaging data. medical images and reports.
These tools not only streamline the reporting
process but also enhance the accuracy of Aa an example, a clinician might ask, “What
diagnostics by cross-referencing findings with are the signs of pulmonary embolism in this CT
historical data and known medical conditions. scan?” The MVQA system would analyze the CT
Moreover, chat-based AI assistants can image, cross-reference it with the clinical
provide real-time support to clinicians by report, and provide a precise answer such as,
answering queries and offering contextually “Presence of filling defects in the pulmonary
relevant information derived from patient arteries consistent with pulmonary embolism.”
records and imaging results.
This application significantly enhances the
MongoDB’s role in supporting AI comprehension of imaging reports, providing
applications concise and clinically relevant answers, which
MongoDB plays a crucial role in enabling these improves diagnostic accuracy and speeds up
AI-driven applications. By providing a flexible decision-making processes.
and scalable database architecture, MongoDB
MVQA combined with vector search
allows for the efficient storage and retrieval of
streamlines diagnostics and supports a more
vast amounts of unstructured medical data,
targeted and effective approach to analyzing
complex medical images.
Vector embeddings for biopsy image rapidly identify and highlight suspicious tiles
images. Similar to text analysis, where large ecosystem, including collaborations with AI
texts are divided into smaller chunks, large pioneers, accelerates the development and
medical images can be segmented into tile deployment of these cutting-edge solutions
characteristics. By leveraging Atlas Vector this use case using MongoDB Atlas.
Figure 49: Gen AI in medical imaging with MongoDB Atlas for the detection of suspicious biopsies
combined with a knowledge base chatbot
By integrating generative AI into the medical
imaging workflow, healthcare providers can
achieve significant improvements in diagnostic
accuracy, reporting efficiency, and overall
patient care. The synergy between MongoDB’s
robust data management capabilities and
advanced AI technologies paves the way for a
future where healthcare professionals can
deliver faster, more reliable diagnoses,
ultimately improving patient outcomes and
reducing operational burdens.
The pharmaceutical industry faces immense required to generate CSRs from weeks to mere
pressure to expedite the regulatory approval minutes, allowing pharmaceutical companies
process for new drugs and therapies. A critical to accelerate their time-to-market for new
component of this process is the creation of drugs.
Clinical Study Reports (CSRs), which are
comprehensive documents detailing the With MongoDB Atlas, companies can leverage
methodology, execution, and results of clinical a flexible, scalable database environment that
trials. Traditionally, compiling a CSR is a supports the dynamic and varied data
labor-intensive task, often requiring several structures inherent in clinical trials. This
weeks to complete and involving flexibility is crucial for managing the diverse
multidisciplinary teams. This prolonged data types involved in CSR generation,
timeline not only delays the introduction of including text, tables, and complex statistical
potentially life-saving treatments but also data. By using generative AI models,
incurs significant costs associated with companies can automate the drafting of CSRs,
prolonged R&D cycles. producing high-quality, compliant documents
that require minimal human intervention.
Challenge
The process of generating CSRs is complex, For example, these AI models can automate
involving the integration of vast amounts of the importation and transformation of data
clinical data, including statistical outputs and tables, generate accurate narratives, and
detailed narratives. Manual methods are ensure that the final documents meet the
time-consuming and error prone, which can compliance standards required by regulatory
further delay regulatory approvals. Plus, the bodies. MongoDB’s Vector Search capabilities
need for compliance with stringent regulatory further enhance this process by enabling the
standards, such as those set by the FDA and retrieval of relevant data with high precision,
EMA, adds another layer of complexity to the which the AI uses to generate consistent and
document creation process. accurate content.
Figure 50: Screenshot from Exafluence AI-powered platform showcasing the automated
synthesis planning of DEET using integrated web search and reaction planning tools
Other Notable Use Cases
Contact
Information
Francesc Mateu Amengual
Healthcare Industry
Solutions Principal
[email protected]
AI Leaders
in Action
Hear from AI pioneers sharing their expert
insights on the future of AI
Ben Ellencweig,
McKinsey senior partner
and global alliance leader
for QuantumBlack, AI by
McKinsey
arm, supports transformation recently, I’ve been focused primarily on AI, and
am currently the global leader of alliances and
initiatives using the power of
acquisitions for QuantumBlack, AI by McKinsey.
technology, technical expertise,
I also led our acquisition of Iguazio, a leader in
and industry experts.
AI and machine learning technology, in 2023.
QuantumBlack Labs is the R&D
and software development arm Before McKinsey, I spent nearly a decade in roles
of QuantumBlack. leading the deployment of new technologies in
government and military settings. The constant
theme throughout my career has been
implementing cutting-edge technology in
innovative ways that transform large
organizations and create significant impact at
scale. This is why it makes sense that I've
become focused on AI—there’s simply no other
technology in our lifetime with the same kind of
potential to improve society and our lives.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
QuantumBlack has been instrumental in
leveraging AI across various sectors. Can
you highlight some of the most impactful
AI projects or initiatives led by
QuantumBlack?
I’m especially proud of our ongoing work with It’s all too easy to overestimate AI's capabilities
One Ocean, a foundation that leads research in the short-term, expecting human-level
efforts to support the development of an reasoning across all domains, even if it’s working
eco-friendly “blue economy.” Identifying which with low-quality or poorly structured data. In
companies are truly having a positive terms of gaining long-term value from AI, I’d say
environmental impact has always been the biggest mistake is executives focusing solely
difficult, so we helped One Ocean develop an on productivity through automation and
AI-powered analytics platform that cost-cutting rather than AI's potential for
streamlines the process dramatically. business model innovation and value creation.
Given the varying rates of generative AI Data is a critical component of successful
adoption across industries, what specific AI implementations. What are the unique
industries do you see as leading the charge data challenges and privacy
in terms of deriving maximum value from considerations for different industries, and
the technology? What are the key factors how can organizations effectively address
driving this disparity? them?
Right now, technology, financial services, and One that comes immediately to mind is
healthcare are at the forefront when it comes healthcare: There’s tremendous value to be
to harnessing the power of gen AI. Tech gained, but you’re also dealing with incredibly
companies are leveraging it for software sensitive patient data that requires robust
development, content creation, and customer anonymization and consent management
service. Financial services firms are using it for practices. Both healthcare and financial
personalized financial advice, risk assessment, services face strict regulations on data usage,
and fraud detection. Healthcare is seeing and finance in particular has to worry about
significant applications in drug discovery, cross-border data flows. Retail and
medical imaging analysis, and clinical decision e-commerce face similar but less extreme
support. issues, as the more they use data to
personalize experiences, the more privacy
Although these three sectors are very different, concerns they bump up against. Regulations
they have a few things in common that have like GDPR have already made these concerns
contributed to their early adoption. All three broadly understood, but there’s still plenty
are very data-driven with well-developed being worked out. Any sector that deals with
technological infrastructure, meaning they classified or proprietary information, including
have access to the well-structured, high-quality manufacturing, government, and defense,
data needed for meaningful AI action, and must put serious effort into ensuring that data
may not need to make big investments in is protected.
hardware. Software and finance are also
sectors where AI-savvy talent is already Every sector faces issues of data quality and
abundant. They have clear use cases integration to some degree, and very large
(mentioned above) that are closely aligned industries with a wide range of customers (like
with core business processes. Last but not healthcare,) often struggle trying to manage a
least, these are competitive sectors that are variety of datasets with different formats and
used to making investments in order to quality levels. Strong data governance
maintain an edge, and stay ahead of the pack. frameworks can go a long way to addressing
these issues, and data quality and integration
tools are getting better all the time. There’s
also been rapid improvement in privacy-
preserving technologies like federated learning
and differential privacy. Ultimately, good data
management is a strategic and even cultural
issue, and one that’s essential for AI success.
How do you envision generative AI QuantumBlack emphasizes the integration
reshaping specific industries in the next of hybrid intelligence, combining human
five to 10 years? What are the key trends expertise with AI capabilities. Can you
and disruptive innovations to watch for? elaborate on how this approach has been
implemented in specific projects or
Predicting the future is always a challenging sectors? What benefits have organizations
task, but if I had to guess I think some of the experienced from leveraging this blend of
most visible changes will be in healthcare, with human and artificial intelligence, and how
AI-assisted drug discovery, personalized does it enhance decision-making and
treatment, and AI-powered diagnostic tools. operational efficiency?
You can expect more effective, more targeted
medical treatments, and potentially a That’s true, hybrid intelligence is something
reduction or at least a leveling of costs, which QuantumBlack has championed since its
all have exciting implications for our quality of beginning and successfully implemented
lives. Several other industries will also start across a wide range of projects and sectors. To
offering hyper-personalized products and give one example, we integrated AI-powered
services combining text, image and video, predictive maintenance models with human
from financial services to education to expert knowledge in a large oil and gas project
shopping, in ways that we have never to optimize equipment maintenance schedules,
experienced before. which reduced downtime by 25% and brought
significant cost savings. For a financial
I'm quite excited to see how the intersection of services client, we developed an AI system
AI and blockchain impacts legal services, and that assists human analysts in detecting
how IoT sensors can be leveraged through AI complex fraud patterns, improving detection
analysis to revolutionize manufacturing, rates by 35% while reducing false positives. In
realizing unprecedented levels of efficiency healthcare, we helped one provider with a
and predictive maintenance. hybrid system for patient triage: AI provides
initial assessments that are then refined by
human clinicians, leading to patient routing
that’s 20% faster, and more accurate as well.
Asaf Somekh is the Founder & CEO at Iguazio Iguazio offers an AI platform
(acquired by McKinsey) that enables enterprises to
develop, deploy, and manage
Can you share your personal journey in the ML and gen AI applications at
world of AI that led you to found Iguazio?
scale. It serves as a unified
What were the specific challenges you
place for data ingestion,
encountered in the industry that inspired
storage, processing, model
you to create a platform to address them?
training, and deployment.
In late 2014, after the successful acquisition of
Voltaire (which went public on NASDAQ in 2007
and was later on acquired by Mellanox, now a
part of NVIDIA) I climbed Mount Kilimanjaro In early 2023, after eight years of building and
with one of my co-founders. Together, we scaling the company, Iguazio was acquired by
decided to create Iguazio to address the McKinsey. Today, Iguazio is a part of
challenge of operationalizing and de-risking QuantumBlack, McKinsey’s AI arm. Together
AI. We saw firsthand the challenges that with McKinsey, we help global organizations,
enterprises were struggling with when trying to including Fortune 500 companies, with special
implement machine learning in the interest from with the financial industry,
organization. Many models were either unable implement and scale gen AI responsibly.
to reach production or, worse, introduced
errors and risks that had a negative impact on Could you explain how Iguazio’s platform
the business. These weren’t just technical simplifies the development, deployment,
setbacks but rather financial losses, and management of gen AI applications?
inefficiencies, and missed opportunities for
innovation. At Iguazio, our mission is to bring AI to life,
making AI implementation not just possible,
ChatGPT galvanized the adoption of AI in but scalable, effective and responsible in live
2022, making the challenges of bringing AI business environments across industries.
from development to production and impact
even bigger.
US 866-237-8815 | INTL +1-650-440-4474 | [email protected] | © 2023 MongoDB, Inc. All rights reserved.
We focus on two critical aspects that set our
platform apart: operationalization and
de-risking.
122
MongoDB Atlas for
Industries: Innovation
Workshops
Accelerate your
modernization
journey Industry-specific solutions The art of the possible
Gain insights into how MongoDB and Explore the art of the possible with
adjacent technologies can provide MongoDB-powered solutions. Learn
end-to-end solutions for from real-world use cases that
Learn more
industry-specific challenges. demonstrate how MongoDB can
Understand how similar issues have transform data management and
been addressed successfully by other modern application development
leading companies in your field. within your industry.
Accelerate
your AI
journey MAAP Ecosystem Composable Architectures
Technology and expertise from Pre-designed architectures serve
the world’s leading AI and tech as accelerated frameworks for
organizations, across the entire fully customizable and secure
gen AI stack, for a cohesive, applications to accommodate
fully supported application ever-evolving gen AI use cases.
development experience.
The AI Stack
Learn more
While the MAAP ecosystem is just getting
MAAP started, it already includes industry leaders
from every part of the AI stack, including
Ecosystem Anthropic, Anyscale, Arcee AI, AWS, Cohere,
Credal, Fireworks AI, Google Cloud,
LangChain, LlamaIndex, Microsoft Azure,
Nomic, and Together AI.
The AI Stack
The cornerstone of MAAP is MongoDB technology. MongoDB sits at the heart of the AI
application stack, the data layer, removing the friction of integrating, securing, and maintaining
the essential data components required to build AI applications. MongoDB underpins these
applications, securely unifying real-time, operational, unstructured, and AI-related data without
the need for bolt-on solutions. MongoDB’s open and integrated architecture provides easy access
to the MAAP ecosystem and enables the extension and customization of applications to tackle
any use case.
The MongoDB AI Applications
Program and its ecosystem of
companies offers customers the
right expertise and solutions for their
use cases. It removes integration
risk, making it easy for businesses to
use the industry-leading
technologies of their choice to
unlock the value of AI with their
data. Learn more about each
member of the MAAP ecosystem in
the MongoDB Partner Ecosystem
Catalog.
The MAAP framework provides access to a
Composable set of downloadable, pre-designed,
pre-integrated RAG architectures with
Architectures pre-configured UIs that serve as
accelerated frameworks for building AI
applications. Architectures are fully
customizable and extendable to
accommodate ever-evolving generative AI
use cases, like retrieval-augmented
generation (RAG) or advanced AI
capabilities like Agentic AI and advanced
RAG technique integrations. With
MongoDB’s open and integrated platform
at its core, innovation with MAAP’s
composable architectures is unlimited.
Figure 51: The MAAP framework is a set of libraries that you can use to build your RAG application
using MongoDB and Atlas Vector Search. Above is the reference architecture diagram of the
framework with various components
MAAP expert services, combining the strengths
of MongoDB Professional Services and
industry-leading gen AI consultancies, will enable
Expert Services customers to rapidly innovate with AI. MAAP
offers strategic guidance on roadmaps and
skillsets, assists with data integration into
advanced AI technologies, and can even develop
production-ready applications.
MAAP goes beyond development, empowering
teams with best practices for securely
integrating your data into scalable gen AI
solutions, ensuring businesses are equipped to
tackle future AI initiatives.
MAAP Framework
The MAAP Framework is a set of libraries that you
can use to build your RAG application using
MongoDB, Atlas Vector Search, and associated
MAAP ecosystem partners.
MongoDB University
Learn to get the most from MongoDB products
with our labs, courses and live training. Create a
foundation with the Introduction to AI and Vector
Search course or advance your learning with
instructor-led training in AI and Vector Search
Basics.
Solutions Library
Drawing from experience with over 49,000
customers, the Solutions Library is curated with
tailored solutions to help developers kick-start
their projects. Get inspired by real gen AI solutions
spanning diverse industries.
AI Resources Hub
View our library of articles, analyst reports, and
case studies designed to help you build
AI-powered applications.
AI Partnerships
How partners are leveraging MongoDB to
build AI solutions
AI with MongoDB and
hyperscaler cloud services
134
Amazon Web Services
(AWS)
Unlock the full potential of generative AI-powered applications with MongoDB Atlas and AWS.
MongoDB Atlas on AWS allows you to build intelligent applications that are highly available,
performant at global scale, and compliant with the most demanding security and privacy standards.
MongoDB Atlas is a Knowledge Base for “Using Amazon Bedrock and MongoDB
Amazon Bedrock, making it even easier to Atlas… we are the first in the industry to
build generative AI applications backed by generate complete Clinical Study Reports
enterprise data.
in minutes rather than weeks.”
MongoDB Atlas on Microsoft Azure empowers enterprises to build intelligent applications that drive
real-world results derived from your data. Integrations with key Azure services, like Microsoft Fabric,
offer a seamless, scalable, and secure platform to unify your data and launch experiences that
harness the best of AI/ML and gen AI.
A seamless integration with Microsoft Fabric Enterprises need to take advantage of gen AI,
enables you to run large scale AI/ML, analytics, AI/ML, and analytics to get the most from their
and BI reports across your unified data estate data.
on MongoDB Atlas.
MongoDB Atlas, Microsoft Fabric, and Azure AI
Reimagine how teams work with data by Studio operate as an ecosystem, driving
bringing everyone together on a single, actionable intelligence on historical data and
AI-powered platform designed to simplify and real-time intelligence to power AI/ML and gen
accelerate intelligent application development. AI use cases.
Google Cloud
MongoDB Atlas and Google Cloud bring together the two halves of the gen AI equation—data and
AI—to help organizations pursue the full potential of gen AI for their applications and software.
The integration between Atlas and Vertex AI, Google Cloud’s unified ML and AI platform, empowers
organizations to unlock the value of their data for AI, gen AI, and ML.
MongoDB Atlas seamlessly integrates with Your teams can query MongoDB Atlas in
Google cloud infrastructure, simplifying the natural language within Vertex AI with Google
connection between your data source and gen Gemini. Automatically turn human language
AI models. into MongoDB-specific query syntax.
Thomas Edison's 10,000 failures led to 10,000 “Accenture sees growing demand from
successful ways that didn't work. Organizations
companies for solutions that can easily
need this spirit of experimentation to innovate.
migrate applications and data migrations
Mark Porter, MongoDB CTO, and Michael Ljung,
Global Software Engineering Lead and Chief to the cloud. Our Smart Data Mover
Software Engineer for Accenture, discuss how solution accelerates this process for
to encourage a culture of experimentation and moving data into MongoDB and other
normalize failed experiments.
target platforms. We continue to invest in
the Smart Data Mover application and
look forward to expanding our
capabilities using MongoDB as a target
platform.”
Learn More
Shail Jain
Global Managing Director, Data & AI at
Accenture
MAAP Partner
Pureinsights
Pureinsights transforms the way organizations interact with information, leveraging cutting-edge AI
technologies like Generative AI, Vector Search, and NLP to build intuitive, human-centered
applications that go beyond traditional search. From information retrieval to innovative use cases
like audio diagnostics, we help businesses uncover new insights and drive innovation.
Pureinsights, MongoDB's premier services partner for search and AI, has built one of the best
Retrieval Augmented Generation (RAG) architectures for a large European car manufacturer using
MongoDB's tech stack and by leveraging our comprehensive ecosystem for AI. Pureinsights’
expertise in search and AI, along with the Pureinsights Discovery platform, have demonstrated the
capability to help customers take their ideas for search and AI from business idea, to working
prototype, to phased production rollout – all while delivering business value.
“Pureinsights exemplifies the innovation and dedication to customer needs that define
our leading partners, earning them this year's Services AI Partner of the Year award.”
Alan Chhabra, Executive
Executive Vice President at MongoDB
As a MongoDB BSI partner and MAAP launch ● GenAI Build Essentials is an initial
partner, Pureinsights is a preferred service project assessment engagement to
provider for MongoDB’s GenAI Build initiatives. scope out a plan to build a Generative
The purpose of these initiatives is to accelerate AI (GenAI) solution with
the planning and prototyping of AI-powered Retrieval-Augmented Generation (RAG)
applications on MongoDB platforms which will capabilities on MongoDB Atlas.
result in faster time-to-value for customers. ● GenAI Build Implementations are a
Pureinsights staff have over 15 years of follow-up to Essentials. This service
experience with commercial or open-source leverages Pureinsights’ and MongoDB’s
data, search and AI platform. technical and development expertise to
build a working prototype of the GenAI
solution implemented on MongoDB
Atlas with retrieval-augmented
generation (RAG) using synthetic data
based on your existing datasets in a
MongoDB environment.
MAAP Partner
gravity9
gravity9 is a leading technology partner, driving businesses beyond outdated legacy systems into the
future with AI-led, cloud-based platforms designed for continuous innovation. Our expertise in
application modernization harnesses advanced cloud technologies, artificial intelligence, and
microservices architecture to streamline operations, elevate user experiences, and foster ongoing
growth. Through our sleek micro-UI platform, we empower organizations to thrive in a rapidly
evolving market, ensuring agility, scalability, and a lasting competitive edge.
gravity9 is proud to be one of only six global strategic MongoDB implementation partners, offering
comprehensive solutions that combine MongoDB's powerful data platform with gravity9’s expertise
in Generative AI and data analytics. gravity9 harnesses the power of MongoDB Atlas to deliver
advanced AI solutions and develop applications based on Large Language Models (LLMs), Retrieval
Augmented Generation (RAG), graph knowledge base and GraphRAG. Our expertise also extends to
Agentic AI solutions for workflow automation, driving efficiency and innovation. By integrating these
cutting-edge techniques and leveraging exclusive access to MongoDB resources and insights, we
help businesses embrace the future of AI and data-driven growth with confidence.
gravity9’s deep partnership with MongoDB gravity9’s highly skilled professionals are
enables them to deliver a flexible, best-practice dedicated to transforming digital applications,
approach to technology modernization. Backed rapidly revitalizing outdated systems with
by rigorous training and certification, gravity9’s tailored, innovative solutions. Their approach
team is fully equipped to design, implement, unlocks new possibilities for businesses across
and optimize MongoDB solutions tailored to various industries by delivering efficiency,
each client's needs. Whether migrating, flexibility, and enhanced functionality.
optimizing, or building new applications, they
ensure seamless integration and rapid
deployment, maximizing the value of each
investment.
MAAP Partner
PeerIslands
Since 2018, PeerIslands has been leading the AI revolution with an exceptional
team and state-of-the-art technology, propelling AI-driven software solutions into the Enterprise and
SMB marketplace. Central to our mission is a handpicked team of top-tier, Top 1% developers,
leveraging advanced AI platforms and our distinctive ‘Human in the Loop’ approach. We go beyond
conventional software development to turbocharge your enterprise applications, reimagine outdated
systems, and harness the power of your data.
We work very closely with MongoDB and complement a great database product with end to end
solutions that customers want. With a talent pool of over 150 MongoDB certified Polyglot engineers
and our active collaboration with Enterprise clients, PeerIslands has been named MongoDB's
Boutique SI of the Year for 3 of the last 4 years. Our software delivery expertise, combined with
MongoDB technology and Solution Architects, has been a recipe for success for clients seeking to
modernize their applications and migrate their data to a modern, scalable, and flexible data
architecture.
MAAP Partner
Unlocking the Power
of AI With SaaS
From building your first Gen AI app, to a full blown Gen AI Factory
By automating and streamlining AI, Iguazio accelerates time-to-market, lowers operating
costs, de-risks, provides guardrails and enhances business impact and profitability. This
enables Iguazio to support enterprise needs, either in a self-serve or managed services
model, with an open and flexible architecture.
Iguazio provides you with the latest capabilities for:
1. Gen AI Ops: Operationalizing AI / Gen AI apps efficiently at scale to create real
business impact.
2. Gen AI Guardrails: De-risking Gen AI to meet compliance, regulations and controls
relevant to your industry while ensuring peak performance.
Iguazio supports data management, training and fine-tuning LLMs, application deployment
and LiveOps that enables monitoring models and data for feedback.
MongoDB and Iguazio can be used for Guardrails for Protecting Against LLM
creating a smart customer care agent that Risks
documents call details, provides live
Iguazio eliminates LLM risks with guardrails
contextual recommendations as a co-pilot,
that ensure:
provides live agent support, customizes
offers and recommendations and more. ● Fair and unbiased outputs
● Intellectual property protection
First, the joint architecture processes and
● PII elimination to safeguard user
analyzes raw data (e.g., web pages, PDFs,
privacy
images) inputted by the customer or the
● Improved LLM accuracy and
enterprise.
performance for minimizing AI
Then, the data is processed in a batch hallucinations
pipeline for analyzing customer logs and a ● Filtering of offensive or harmful
stream pipeline for live interactions. content
Finally, results are stored in MongoDB, ● Alignment with legal and regulatory
leveraging its capabilities for managing standards
unstructured data like user age, preferences ● Ethical use of LLMs
and historical transactions, together with
structured data like account balance and
product lists.
Reliability Trustworthiness
Claude powers business-critical use cases, Claude combines best-in-class jailbreak
exhibiting lower hallucination rates and higher resistance and misuse prevention to mitigate
accuracy. brand risk for enterprises.
Safety Flexibility
Claude offers robust security and compliance The Claude model family offers a range of AI
features, meets SOC 2 Type II and HIPAA solutions across the price-performance
standards, and ensures enterprise-grade spectrum, providing enterprises with options for
protection. any use case.
Bring enterprise data to gen AI and use a family of flexible foundation models
Anthropic and MongoDB offer collaborative ● Tailored AI for Your Needs: Deploy and
support to help you navigate the complexities scale generative AI applications tailored to
of building safe, reliable, enterprise-grade your use case and leveraging your data.
generative AI solutions. Together, we provide
● Compliance Built-In: Ensure your
comprehensive solutions for enterprises—so you
generative AI apps are secure and compliant
can launch and scale generative AI applications
out of the box.
safely and reliably, customized with your data.
● RAG Made Easy: Abstract the challenges of
building a Retrieval-Augmented Generation
(RAG) workflow with your data.
MAAP Partner
© 2024 MongoDB, Inc. All rights reserved.
Figure 53: Cost vs intelligence of Claude models.
The Claude model family offers state-of-the-art performance across a wide range of tasks,
allowing users to select the optimal balance of intelligence, speed, and cost for their specific
applications. Claude’s capabilities include:
● Advanced reasoning: Claude can perform complex cognitive tasks that go beyond
simple pattern recognition or text generation
● Vision analysis: Transcribe and analyze almost any static image, from handwritten notes
and graphs to photographs
● Code generation: Start creating websites in HTML and CSS, turning images into
structured JSON data, or debugging complex code bases
Developers and researchers can accelerate the Developers simply add a few lines of Python
entire AI development and deployment process code to access GPU and HPC infrastructure,
- from training, fine-tuning, benchmarking, data without the burden of learning Kubernetes,
synthesis, to inference - in a cost-effective way. Slurm, or other DevOps tools.
Take AI agents to the next level by expressing Covalent automatically pools, assigns, and
complex interactions as a seamless Python scales resources to workloads dynamically, in
workflow and even deploying each agent in its order to meet changing demands for compute
own separate compute environment. and optimizing infrastructure efficiency.
Agnostiq chose MongoDB as their default MongoDB is used in the front-end, allowing a
NoSQL database for the free, open source high volume of metadata and other assets
version of Covalent. Without any DBAs as a to be published and cached in accordance
small agile team, MongoDB gave Agnostiq with an event-driven architecture. This near
the freedom to build and manage data real-time experience is key to a product
workflows without the need for a specialist. aimed at delivering a unified view over
distributed resources. MongoDB Atlas further
As their customer base grew along with the provides the autoscaling required to grow
demand for cloud computing access, with the user base and the number of
Agnostiq moved to MongoDB Atlas, gaining workloads while keeping costs in check.
the freedom to move data seamlessly
between AWS, Google Cloud, and Microsoft “MongoDB Atlas helps us provide an ideal
Azure. This gave Covalent the flexibility to foundation for modern HPC and AI
reach multi-cloud compatibility at a faster applications which require serverless
rate than with standard tooling. compute, autoscaling resources, distributed
workloads, and rapidly reconfigurable
Covalent provides a workflow management infrastructure.”
service by registering jobs, dispatching IDs, Santosh Kumar Radha, Head of Product at
and collecting other metadata that allows Agnostiq
fellow researchers and developers to
reproduce the original work.
Give every employee the ability to create the perfect AI Assistant for their work.
MAAP Partner
© 2024 MongoDB, Inc. All rights reserved.
Figure 55: Bringing your data to LLMs and bring LLMs to your operations.
MAAP Partner
© 2024 MongoDB, Inc. All rights reserved.
Figure 56: Multi-Modal Search Application - application is composed of multiple components -
including legacy search and the AI-enabled search - that work together to provide a hybrid search
experience
Multi-Modal Search is a cutting-edge use case that leverages Anyscale's distributed AI compute
engine and MongoDB Atlas's flexible NoSQL database to create a seamless search experience
across diverse data types, including text, images, and structured data. Anyscale provides the
infrastructure to easily scale and optimize compute resources, enabling rapid processing of
multimodal data, while MongoDB Atlas efficiently stores and indexes these datasets.
The combine solution makes it possible to perform fast and accurate similarity searches across
large volumes of complex, unstructured data. With this architecture, users can quickly retrieve
relevant content from a variety of formats, such as images, documents, and even video, without
being burdened by infrastructure management. This makes it ideal for building scalable, AI-driven
search applications across multiple domains while optimizing for resources utilization and
scalability.
● Enable vector search across multiple fields and dimensions via MongoDB Atlas Vector
Search
● Supporting the storage of multi-modal data like images, text, and structured data in
MongoDB Atlas
● Run performant LLM Batch Inference compute jobs with Anyscale’s Platform
● Optimally scale and maximize utilization of costly compute resources with Anyscale’s
platform and intelligent infrastructure management
As the financing landscape evolves, there's a enables financial institutions to remain agile
growing need for configurable, cloud-based and responsive in today's competitive
software built on microservices. This type of landscape. This integrated approach not only
software should be able to manage the streamlines processes but also enhances
entire life cycle of any financing product. transparency and accountability throughout
Basikon, as illustrated in the Figure above, the financing journey, ultimately driving
fulfills this need by orchestrating the digital greater customer satisfaction and loyalty.
journey across all stages, from initial
customer interaction to loan approval and “MongoDB Atlas is very stable – in 4 years,
management. It also empowers financial we did not experience a single interruption
institutions to manage their distribution of service or find a single bug. Upgrades are
networks and partner relationships directly done in seconds with just the press of a
through the platform. button, increasing our agility 10x. At Basikon,
MongoDB has played a crucial role in our
Moreover, Basikon's cloud-based software success and we wouldn't be where we are
architecture built on microservices ensures today without it.”
adaptability to changing market dynamics Thomas Nokin, Founder and CEO at Basikon
and regulatory requirements. By offering a
comprehensive solution for managing the
entire financing product life cycle, Basikon
Cognigy is a pioneering force in AI-driven customer service solutions on a global scale. They are at
the forefront of revolutionizing the customer service industry by providing the most cutting-edge AI
workforce on the market. Trusted by giants like Toyota, Bosch, and Lufthansa, their award-winning
solution empowers businesses to deliver exceptional customer service: instant, personalized, in any
language, and on any channel.
● MongoDB's JSON document storage aligns diverse data types, ensuring peak
perfectly with Cognigy's application performance under high loads.
language, facilitating seamless integration
● MongoDB empowered Cognigy.AI to handle
with Typescript and intuitive querying
expanding user interactions while
processes
maintaining peak performance, ensuring
● MongoDB's scalability via sharding aligns scalability and responsiveness in scaling
with Cognigy's growth vision, enabling conversational agents.
expansion across cloud providers and
● MongoDB's document model flexibility
on-premises setups.
enables easy data model modifications,
● MongoDB's developer data platform reducing concerns about data and schema
empowered Cognigy to efficiently manage migrations.
Figure 59: Cognigy’s replica-sets in production
Have you ever built a chatbot that struggled This tech allows Cognigy.AI to effortlessly
to keep up with user demands? Imagine a manage a growing number of user
platform that can handle hundreds of interactions, processing all sorts of data
queries per second, even during peak hours, easily.
all while storing massive amounts of data.
That's the power of MongoDB at work for Imagine your chatbot being able to learn
Cognigy.AI! and improve over time, This is what
MongoDB's flexible data model enables for
Cognigy constructed the platform by Cognigy.AI. As new data and user
employing a composable architecture model interactions flow in, Cognigy.AI can
with over 30 specialized microservices, which continuously update and refine its
they adeptly orchestrated through understanding of how to best serve your
Kubernetes. These microservices were customers. This collaboration is a prime
strategically fortified with MongoDB's replica example of how powerful technology can be
sets, spanning across three availability the driving force behind groundbreaking
zones, a move aimed at bolstering reliability products like Cognigy.AI. Imagine the
and fault tolerance. possibilities: chatbots that can provide
personalized recommendations, troubleshoot
As you can see in the Figure above, complex issues, and even have engaging
MongoDB's magic isn't just marketing hype. conversations.
Devnagri is India's leading AI-powered translation engine, enabling brands to localize content five
times faster and more accurately. As a SaaS platform, it focuses on translating Indian languages,
utilizing a hybrid approach of 80% machine and 20% human effort to achieve 99% accuracy in
translating millions of words daily.
Devnagri’s real-time translation engine helps over 100 Indian brands connect with their
customers over digital channels for the first time
The real-time translation engine has helped The platform's focus spans diverse industries
over 100 Indian brands connect with their such as e-learning, banking, e-commerce,
customers over digital channels for the first and media publishing, offering a tailored
time. This achievement signifies a solution beyond a general consumer tool.
breakthrough in overcoming the language Powered by custom transformer models and
barrier in India, where 90% of the population advancements like OpenAI GPT-4,
are not fluent in English, and more than 22 Devnagri's technology strives to democratize
Indian languages are in use. internet access for India's non-English
speakers.
WINN.AI is more than just a tool; it’s a productivity powerhouse designed to transform the way sales
teams operate. By reducing administrative busywork, WINN.AI is helping organizations save time,
money, and resources, enabling sales teams to better invest their working hours in serving
customers.
● Flexibility and Agility: Ada can easily scale ● Less Dependency on One Central Cloud
their database as their business grows and Vendor: By using a cloud-agnostic solution,
adapt to new channels and modalities Ada avoids being locked into a single cloud
without being restricted by their database provider. This gives them more freedom and
infrastructure. flexibility.
● Performance and Support: Ada has found ● Distributed Event Processing System: Ada
that the performance of MongoDB Atlas is using MongoDB Change Streams to build a
meets their needs, and they appreciate the distributed event processing system that
great support from the MongoDB team. powers bots and analytics.
Figure 61: Ada’s AI customer service
At the Heart of XOLTAR lies a sophisticated data, such as video transcripts. This data
array of state-of-the-art machine learning provides both long-term memory for each
models working across multiple modalities — patient as well as input for ongoing model
voice and text, as well as vision for visual training and tuning.
perception of micro-expressions and
non-verbal communication. These custom MongoDB also powers XOLTAR’S
multilingual models are trained and event-driven data pipelines. Follow-on
deployed to create a truthful, grounded, and actions generated from patient interactions
aligned free-guided conversation, along with are persisted in MongoDB, with Atlas
various transformers for real-time automatic Triggers notifying downstream consuming
speech recognition. applications so they can react in real-time to
new treatment recommendations and
XOLTAR’s models personalize each patient’s regimes.
experience by retrieving data stored in
MongoDB Atlas. Taking advantage of the Through its participation in the MongoDB AI
flexible document model, XOLTAR developers Innovators program, XOLTAR’s development
store both structured data, such as patient team receives access to free Atlas credits
details and sensor measurements from and expert technical support, helping them
wearables, alongside unstructured de-risk new feature development.
Observe.AI, a California-based company funded by over $200 million, is the leading provider of live
conversation intelligence for contact centers. Trusted by industry leaders like Accolade and Pearson,
Observe.AI empowers businesses to transform the way they interact with customers. The company is
focused on being the fastest way to boost contact center performance with live conversation
intelligence.
The MongoDB developer data platform gives because it enables us to quickly innovate,
the company’s developers and data scientists a scale to handle large and unpredictable
unified solution to build smarter AI workloads, and meet the security
applications.
requirements of our largest enterprise
customers.”
“OBSERVE.AI processes and runs models on Jithendra Vepa
millions of support touchpoints daily to Ph.D, Chief Scientist & India General Manager
generate insights for our customers. Most of at Observe.AI
this rich, unstructured data is stored in
MongoDB. We chose to build on MongoDB
Figure 63: Observe.AI’s conversation intelligence
Flagler Health is dedicated to supporting patients with chronic diseases by matching them with the
right physician for the right care. Typically, patients grappling with severe pain conditions face
limited options, often relying on prolonged opioid use or exploring costly and invasive surgical
interventions. Unfortunately, the latter approach is not only expensive but also has a long recovery
period. Flagler finds these patients and triages them to the appropriate specialist for an advanced
and comprehensive evaluation.
Flagler Health employs sophisticated AI techniques to rapidly process, synthesize, and analyze
patient health records to aid physicians in treating patients with advanced pain conditions. This
enables medical teams to make well-informed decisions, resulting in improved patient outcomes with
an accuracy rate exceeding 90% in identifying and diagnosing patients.
As the company built out its offerings, it identified the need to perform similarity searches across
patient records to match conditions. Flagler’s engineers identified the need for a vector database
but found standalone systems to be inefficient. They decided to use MongoDB Atlas Vector Search.
Creating an integrated platform to store all data in a single location with a unified
interface, facilitating quick access and efficient data querying.
Flagler Health collaborates with many clinics, This comprehensive application architecture,
first processing millions of electronic health consolidated on MongoDB's developer data
record (EHR) files in Databricks and platform, simplifies Flagler Health's
transforming PDFs into raw text. Using the operations, enabling efficient development
MongoDB Spark Connector and Atlas Data and increased productivity. By preventing
Federation, the company seamlessly streams administrative loops, the platform ensures
data from AWS S3 to MongoDB. Combined timely access to potentially life-saving care
with the transformed data from Databricks, for patients.
Flagler’s real-time application data in
MongoDB is used to generate accurate and Looking ahead, Flagler Health aims to
personalized treatment plans for its users. enhance patient experiences by developing
MongoDB Atlas Search facilitates efficient new features, such as a digital portal
data search across Flagler Health's extensive offering virtual therapy and mental health
patient records. Beyond AI applications, services, treatment and recovery tracking,
MongoDB serves critical functions in Flagler and a repository of physical therapy videos.
Health's business, including its web Leveraging MongoDB’s AI Innovators
application and patient engagement suite, program for technical support and free Atlas
fostering seamless communication between credits, Flagler Health is rapidly integrating
patients and clinics. new AI-backed functionalities on the
MongoDB Atlas developer data platform to
further aid patients in need.
The Dataworkz gen AI applications platform provides an all-in-one RAG as a Service to rapidly build,
deploy, operationalize and scale gen AI applications, and eliminates the complexity involved in
building reliable and scalable RAG applications. It includes advanced search and retrieval to provide
relevant context to LLMs, and monitoring with traceability to observe and optimize application
performance.
Build Optimize
Visually Create Gen AI Applications: Comprehensive Visibility &
Develop gen AI applications using a visual Customization: Achieve complete AI stack
RAG builder. This eliminates the need to transparency and easily customize data
worry about the complexity of the underlying processing steps in a user-friendly, no-code
infrastructure. interface.
Observe Scale
End-to-End Traceability: Get full visibility Rapid Application Development: Build
of your gen AI apps with end-to-end diverse gen AI applications efficiently by
traceability for better performance utilizing pre-defined templates for various
optimization. use cases.
Centralized Monitoring: Track all system Embeddable RAG: Integrate gen AI apps
activity, like LLM calls, SLM calls, indexing, with Slack, Azure Studio, and HTML widgets
and retrieval, with one unified tool. via RAG APIs for enhanced accessibility in
workflows.
VISO TRUST is an AI-powered platform that helps companies quickly assess the cybersecurity risk of
their vendors. It provides actionable security information in minutes, allowing businesses to make
informed decisions with ease. VISO TRUST boasts a 90% reduction in workload and an 80% faster
risk assessment process, with near-universal vendor adoption by their clients.
VISO TRUST uses AI to streamline third-party Curated AI extracts insights from source
risk assessments, enabling instant evaluation artifacts, automatically determining vendor
without extra analysts. It eliminates lengthy security posture. This frictionless due diligence
questionnaires and manual document analysis process simplifies assessing any number of
for a more efficient approach. third parties.
On the platform, users can gain a Continuously exceeding ISO, NIST, AICPA, and
comprehensive overview of their organization’s other standards without impeding business
cyber risk posture, enabling them to make operations is made possible by VISO TRUST. It
data-driven decisions to reduce risk across all empowers organizations to take control of their
third-party relationships. third-party security posture.
● VISO TRUST deploys discriminator models ● The outcomes of RAG serve as the
that produce high-confidence predictions foundation for seeding LLM prompts and
about features of the artifact. linking their outputs in a chain, resulting in
the generation of highly precise factual
● The artifacts undergo a process where their
details regarding the artifact in the pipeline.
text content is extracted and integrated into
This data facilitates the swift delivery of
MongoDB Atlas, thus becoming integrated
intelligence to customers, a task that
into the dense retrieval system. This system
previously required weeks to accomplish.
executes Retrieval-Augmented Generation
(RAG) by leveraging MongoDB functionalities
such as Atlas Vector Search. Its aim is to
furnish ranked context to prompts for large
language models (LLMs).
Figure 66: Insights dashboard
The multi-cloud architecture of Atlas provides flexibility and choice that proprietary offerings from
the hyperscalers can’t match. While DevRev today runs on AWS, in the early days of the company,
they evaluated multiple cloud vendors. Knowing that MongoDB Atlas could run anywhere gave them
the confidence to make a choice on the platform, knowing they would not be locked into that choice
in the future.
DevRev manages critical customer data, and so relies on MongoDB Atlas’ native encryption and
backup for data protection and regulatory compliance. The ability to provide multi-region databases
in Atlas means global customers get further control over data residency, latency, and high
availability requirements.
MongoDB is the primary database backing and release faster. Developers can
OneCRM, managing users, customer and experiment locally, then move to integration
product data, tickets, and more. DevRev testing, and then production — all running in
selected MongoDB Atlas from the very different environments — without changing
outset of the company. The flexibility of its a single line of code. This is core to DevRev’s
data model, freedom to run anywhere,
velocity in handling over 4,000 pull requests
reliability and compliance, and operational
per month:
efficiency of the Atlas managed service all
impact how quickly DevRev can build and
● Developers can experiment and test with
ship high-quality features to its customers.
MongoDB on local instances — for
The flexibility of the document data model example adding indexes or evaluating
enables DevRev’s engineers to handle the new query operators, enabling them to
massive variety of data structures their catch issues earlier in the development
microservices need to work with. Documents cycle.
are large, and each can have many custom ● Once unit tests are complete, developers
fields. To efficiently store, index, and query can move to temporary instances in
this data, developers use MongoDB’s Docker containers for end-to-end
Attribute pattern and have the flexibility to
integration testing.
add, modify, and remove fields at any time.
● When ready, teams can deploy to
The freedom to run MongoDB anywhere production in MongoDB Atlas.
helps the engineering team develop, test,
● Cloneable integrates seamlessly with ● Utilizing Cloneable and Atlas Vector Search
MongoDB, enabling the persistence of data to generate vector embeddings from images
locally on devices and its synchronized and device data enables users to efficiently
transfer to the cloud-based Atlas database. search and analyze field-collected events,
This ensures that enterprises can track, thereby enhancing decision-making and
measure, and respond to events across insights.
their operations in real-time.
Figure 68: Cloneable components
Patronus AI is a company that develops tools to help businesses safely use large language models
(LLMs). Their main product is an automated evaluation platform that can identify errors and
unreliable outputs from LLMs. This is especially important for regulated industries where mistakes
can have serious consequences.
Founded by machine learning experts from Meta AI and Meta Reality Labs, Patronus AI is on a
mission to boost enterprise confidence in gen AI-powered apps, leading the way in shaping a
trustworthy AI landscape.
In recently published and widely cited research While retrieval augmented generation (RAG) is
based on the FinanceBench question answering a common way of feeding models with
(QA) evaluation suite, Patronus made a up-to-date, domain-specific context, a key
startling discovery. Researchers found that a question faced by app owners is how to test the
range of widely used state-of-the-art LLMs reliability of model outputs in a scalable way.
frequently hallucinated, incorrectly answering or This is where Patronus comes in. The company
refusing to answer up to 81% of financial has partnered with the leading technologies in
analysts’ questions! This error rate occurred the gen AI ecosystem — from model providers
despite the models’ context windows being and frameworks to vector store and RAG
augmented with context retrieved from an solutions — to provide managed evaluation
external vector store. services, test suites, and adversarial data sets.
Figure 69: Reference architecture and workflow
In its new 10-minute guide, Patronus takes Equipped with the results of an analysis,
developers through a workflow showcasing there are a number of steps developers can
how to evaluate a MongoDB Atlas-based take to improve the performance of a RAG
retrieval system. The guide focuses on system. These include exploring different
evaluating hallucination and answers indexes, modifying document chunking sizes,
relevance against an SEC 10-K filing, re-engineering prompts, and for the most
simulating a financial analyst querying the domain-specific apps, fine-tuning the
document for analysis and insights. The embedding model itself. Review the
workflow is built using: 10-minute guide for a more detailed
explanation of each of these steps.
“With MongoDB, developers can store data of any structure and then expose that data to OLTP, text search,
and vector search processing using a single query API and driver. With this unification, developers have all
of the core data services they need to build AI-powered apps that rely on working with live, operational
data.”
Tiffany Peng, VP of Engineering at Gradient
Gradient's Accelerator Block Boosts RAG Model Performance and Accuracy with
Pre-Built Infrastructure
Gradient’s newest Accelerator Block focuses Together, Atlas Vector Search and
on enhancing the performance and accuracy LlamaIndex feed foundation models with
of a model through retrieval augmented up-to-date, proprietary enterprise data in
generation (RAG). The Accelerator Block uses real-time. Gradient designed the Accelerator
Gradient’s state-of-the-art LLMs and Block for RAG to improve development
embeddings, MongoDB Atlas Vector Search velocity up to 10x by removing the need for
for storing, indexing, and retrieving infrastructure, setup, or in-depth knowledge
high-dimensional vector data, and around retrieval architectures. It also
LlamaIndex for data integration. incorporates best practices in document
chunking, re-rankers, and advanced retrieval
strategies.
One AI is a company that aims to democratize and deliver AI as a service for businesses. Their
mission is to integrate AI into everyday life by transforming natural language into structured,
actionable data. This is achieved through their easy-to-use APIs, which package leading AI
capabilities from across the ecosystem.
The One AI APIs let developers analyze, process, and transform language input in their
code. No training data or NLP/ML knowledge are required.
“The MongoDB document model really allows The company also benefits from the regular
us to spread our wings and freely explore new new releases from MongoDB, such as Atlas
capabilities for the AI, such as new predictions, Vector Search, which Ben sees as a highly
new insights, and new output data points.” Ben valuable addition to the platform’s toolkit.
adds, “With any other platform, we would have Ben explains: “The ability to have that
to constantly go back to the underlying vectorized language representation in the
infrastructure and maintain it. Now, we can same database as other representations,
add, expand, and explore new capabilities on a which you can then access via a single query
continuous basis.” interface, solves a core problem for us as an
API company."
Amit Ben, CEO at One AI
Founded in 2011, Kovai is an enterprise software company that offers multiple products in both the
enterprise and B2B SaaS arena. Since its founding, the company has grown to nearly 300 employees
serving over 2,500 customers.
● MongoDB Vector Search offers architectural ● Atlas Vector Search enables Kovai to store
simplicity, making it easier for Kovai to both knowledge base articles and their
optimize the technical architecture needed to embeddings together in MongoDB
implement their AI assistant, "Eddy." This collections. This eliminates the need for data
simplicity likely streamlines development syncing between multiple databases, which
efforts and reduces complexity in not only simplifies operations but also
integrating the search functionality into their reduces potential inaccuracies in answers
system. provided by the assistant. Operational
efficiency is crucial for a seamless user
● MongoDB Vector Search delivers faster
experience.
query response times at scale, ensuring a
positive user experience for Kovai's
customers interacting with the AI assistant.
Figure 72: Reference architecture
Faster, Simpler, More Efficient: How Kovai Leverages MongoDB Atlas Vector Search
The release of MongoDB Atlas Vector Search Specifically, the team has seen the average
provided a solution with three key time taken to return three, five, and 10
advantages for the engineers: chunks between two and four milliseconds,
Robust Intelligence safeguards organizations from AI's risks. Their end-to-end platform continuously
validates models, protecting them with an AI Firewall. This empowers confident AI adoption for any
model type, from basic to generative. Trusted by leaders like JPMorgan Chase, Robust Intelligence is
your key to unlocking AI's potential.
“By incorporating MongoDBʼs Atlas Vector Search into the AI validation process,
customers can confidently use their databases to enhance LLM responses
knowing that sensitive information will remain secure. The integration provides
seamless protection against a comprehensive set of security, ethical, and
operational risks.”
Yaron Singer
CEO and co-founder at Robus Intelligence
Figure 73: High level architecture
Fireworks AI and MongoDB are now partnering to make innovating with generative AI faster, more
efficient, and more secure. Fireworks AI was founded in late 2022 by industry veterans from Meta’s
PyTorch team, where they focused on performance optimization, improving the developer
experience, and running AI apps at scale. It’s this expertise that Fireworks AI brings to its production
AI platform, curating and optimizing the industry's leading open models. Benchmarking by the
company shows gen AI models running on Fireworks AI deliver up to 4x faster inference speeds than
alternative platforms, with up to 8x higher throughput and scale.
Models are one part of the application stack. But for developers to unlock the power of gen AI, they
also need to bring enterprise data to those models. That’s why Fireworks AI has partnered with
MongoDB, addressing one of the toughest challenges to adopting AI. With MongoDB Atlas,
developers can securely unify operational data, unstructured data, and vector embeddings to safely
build consistent, correct, and differentiated AI applications and experiences. Fireworks AI and
MongoDB provide a solution for developers who want to leverage highly curated and optimized
open-source models, and combine these with their organization’s own proprietary data — and to do
it all with unparalleled speed and security.
Lightning-fast models from Fireworks AI: Enabling speed, efficiency, and value
MAAP Partner
Figure 74: Bringing your data to LLMs
Getting started: The Fireworks tutorial showcases how to bring your own data to LLMs
with retrieval-augmented generation (RAG) and MongoDB Atlas
With Fireworks AI and MongoDB Atlas, apps provide better intelligence and strategies.
run in isolated environments ensuring uptime Or, organize and classify a product
and privacy, protected by sophisticated catalog using product images and text.
security controls that meet the toughest
● Images to structured data extraction:
regulatory standards:
Extract meaning from images to produce
● As one of the top open-source model API structured data that can be processed
providers, Fireworks AI serves 140 billion and searched in a range of vision apps —
tokens per day (and growing). from stock photos, to fashion, to object
detection, to medical diagnostics.
● With Atlas, you run your apps on a proven
platform that serves tens of thousands of ● Alert intelligence: Process large amounts
customers, from high-growth startups to of data in real-time to automatically
the largest enterprises and governments. detect and alert on instances of fraud,
cybersecurity threats, and more.
Together, the Fireworks AI and MongoDB
joint solution enables: Getting started with Fireworks AI and
● Retrieval-augmented generation (RAG) MongoDB Atlas: review the Optimizing RAG
or Q&A from a vast pool of documents: with MongoDB Atlas and Fireworks AI
Ingest a large number of documents to tutorial, which shows you how to build a
produce summaries and structured data
that can then power conversational AI. movie recommendation app.
● Classification through
semantic/similarity search: Classify and
analyze concepts and emotions from sales
calls, video conferences, and more to
LangChain and LangGraph are open-source frameworks for building context-aware reasoning
applications and reliable agents. LangSmith is a unified developer platform for building, testing, and
monitoring LLM applications, whether you use LangChain or not.
Join 1M+ builders who standardize their development using LangChain’s frameworks and platform
together or separately to accelerate their AI application development.
MAAP Partner
© 2024 MongoDB, Inc. All rights reserved.
Augment the power of LLMs with your data
The most powerful LLM applications and Systems composing LangChain, MongoDB,
agents integrate with public data (via LLMs and an LLM together enable you to extract
and search engines), external systems (via valuable insights from complex and recent
tools), and most importantly, your data, highlighting the potential of these
organization’s private documents and data. technologies in developing intelligent
applications. For more information, please
MongoDB’s deep integration with
refer to this detailed blog.
LangChain’s Python and TypeScript
frameworks provides the powerful semantic
search capabilities of MongoDB Atlas’s
vector search engine over your private data.
MAAP Partner
© 2024 MongoDB, Inc. All rights reserved.
Figure 75: LlamaCloud indexes data from various sources and enables advanced retrieval
algorithms. This can be connected to the open-source LlamaIndex for agentic workflows, Q&A,
chatbot etc
Nomic Embed v1.5 is a truly open source text embedder for the big data era. Out of the box, this
model supports a 8192 token context length, resizable embedding dimensions, and binary
quantization, all while outperforming similar models such as OpenAI’s Ada-002 and
text-embedding-3-small on both short and long context tasks.
Truly Open Source Long Context
Nomic Embed provides open-source model Nomic Embed supports a 8192 token context
weights and training code under the Apache-2 length making it well-suited for real-world
license, with curated training data available on applications with large PDFs and text
the Nomic website. This ensures full documents.
reproducibility and auditability.
MAAP Partner
MongoDB Atlas
PDF Search is a use-case that combines the ● Large PDFs can be chunked and
capabilities of Nomic Embed and MongoDB ingested into MongoDB Atlas via
Atlas for an accessible, high-throughput stream processing, while
solution. Nomic Embed simplifies the ● Nomic Embed can quickly produce
process of embedding PDF previews directly long-context embeddings from the
into your application, while MongoDB Atlas processed text.
provides a powerful and scalable NoSQL ● MongoDB Vector Search integrates
database to store and index your PDFs for semantic search on Nomic
efficient searching. This combination allows embeddings with traditional database
you to quickly build a user-friendly search queries for multi-faceted downstream
experience for your PDFs without worrying analysis.
about complex infrastructure management.
Founded in San Francisco in 2022, Together AI is the AI Acceleration Cloud for building and running
generative AI (gen AI). Over 150,000 developers, and organizations like Zomato, The Washington
Post and DuckDuckGo run gen AI in production using Together’s Platform. The company has raised
over $200 million, counting NVIDIA, Salesforce Ventures, Kleiner Perkins, Lux, and NEA as investors.
“We prioritized integrating with MongoDB because of its relevance and importance in the
AI stack.”
Vipul Ved Prakash
Founder and CEO at Together AI
By integrating MongoDB Atlas with Together ● Reduced Complexity and Cost MongoDB
AI’s inference and embedding capabilities, Atlas and Together AI simplify RAG app
developers can build AI applications that development by making it easy to keep
deliver fast, real-time insights. This joint solution embeddings up to date. Together AI also
enables AI models to leverage RAG to provides high-performance inference at
recommend accurate, data-driven results that significantly lower costs than closed-source
meet user criteria while maintaining accuracy. solutions.
MAAP Partner
Figure 77: Together AI framework
Arcee AI meets you where you are on your AI journey, always giving you full ownership of your
models and data. We provide companies with cutting-edge out-of-the-box models, state-of-the-art
custom models, and/or our easy-to-use model training platform. You get the freedom to deploy the
models to any environment–including SaaS, dedicated SaaS, on-premise, or VPC–with guaranteed
data privacy and security. Our solutions cater to various use cases from customer service to
software development, offering flexible pricing options and dedicated support. Our featured
products include:
SuperNova & Other Top-Tier Models Advanced Model Refinement
High-performance general use models Our best-in-class post-training pipeline,
out-of-the-box, like SuperNova, our distilled incorporating synthetic dataset generation,
version of Llama-405B that outperforms SFT, reward modeling, sparse auto-encoding,
leading models in various benchmarks. model merging/infusion, and DPO to produce
state-of-the-art models.
MAAP Partner
© 2024 MongoDB, Inc. All rights reserved.
Figure 78: MongoDB Atlas and Arcee AI integrated in AWS for real-time, AI-driven
responses using RAG.
The integration of MongoDB Atlas and Arcee ● Scalability: MongoDB Atlas ensures
AI, as depicted in the diagram, enables flexible, cloud-native scaling to handle
efficient data retrieval and AI-driven growing data needs without
responses through a seamless pipeline. compromising performance.
MongoDB's scalable cloud infrastructure ● AI-Augmented Retrieval: Arcee AI
ensures fast access to data using RAG
enhances retrieved data with
indexes, allowing Arcee AI to augment the
advanced generative processing for
retrieval with powerful generative
more accurate and meaningful
capabilities. This combination enhances the
speed and accuracy of responses, making it responses.
ideal for real-time applications like ● Real-Time Performance: The
interactive AI systems. MongoDB's reliability combined architecture delivers fast,
and Arcee AI's advanced inference reliable query handling suitable for
capabilities complement each other, interactive and time-sensitive
resulting in a unified model response. applications.
Overall, this architecture enables dynamic,
scalable, and intelligent query handling in
cloud environments.
Major retail brands have long been using various forms of AI, for example statistical analysis and
machine learning models, to better serve their customers. But with its high barriers to entry, one key
channel has been slower to embrace the technology. By connecting large and small brands with
customers, e-commerce marketplaces such as Amazon, Mercado Libre, and Shopify are among the
fastest growing retail routes to market. Since 2016, GoBots has been working to extend the benefits
of AI to any retailer on any marketplace. It uses AI, analytics, and MongoDB Atlas to make
e-commerce easier, more convenient, and smarter for brands serving Latin America.
GoBots increases engagement and conversion rates for over 600 clients across Latin America,
including Adidas, Bosch, Canon, Chevrolet, Dell, Electrolux, Hering, HP, Nike, and Samsung.
The solution makes the benefits of AI available to any retailer, whether large or small. With the
GoBots natural language understanding (NLU) model, retailers automate customer interactions such
as answering questions and resolving issues through intelligent assistants. At the same time, they
leverage data analytics to offer personalized customer experiences.
By using GoBots AI for ecommerce with MongoDB Atlas, customers have grown sales conversions by
40% and reduced time to customer response by 72%.
With the power of MongoDB’s developer data platform and flexibility of MongoDB’s
document model, GoBots builds higher-performing AI-powered applications faster:
● MongoDB Atlas provides a single data platform enrich historical questions with outputs
that serves multiple operational and AI use cases. generated by different models and compare the
This includes user data and product catalogs as results. This means that they are not blocked
well as a store for AI model inferences, outputs of behind complex schema changes that would
multiple AI models for experimentation and otherwise slow down the pace of harnessing new
evaluation purposes, a data source for fine-tuning data in their models for training and inference.
models, and for vector search.
● The question-answer pairs output by the
● GoBots is evaluating the use of Atlas Triggers for company’s NLU models and LLMs are complex
invoking AI model API calls in an event-driven data structures with many nested entities and
manner as the underlying data changes. arrays. Being able to persist these directly to the
database without first having to transform them
● The flexibility provided by MongoDB’s document
into a tabular structure improves developer
model allows the development team to continually
productivity and reduces application latency.
Figure 79: GoBots question processing architecture
GoBots’ custom NLU models are built using question-answer generation with all
the Rasa framework with a neural network inferences also stored in MongoDB. If the
trained on over 150 million question-answer models are able to generate an answer with
examples and more than 50 bots — high confidence, the GoBots service will
specialists in different segments — to respond directly to the customer in real time.
understand more specific questions. In case of a low confidence response, the
models flag the question to a customer
Models are fine tuned with data from the service representative who receives a
retailer's own product catalog and website pre-generated suggested response.
corpus. The model runtime is powered by a
PyTorch microservice on Google Cloud. The With all question-answer pairs from the
larger GoBots platform is built with Kotlin different models written to the MongoDB
and orchestrated by Kubernetes, providing Atlas database, the data is used to further
the company with cloud freedom as its tune the natural language models while also
business expands and evolves. guiding model evaluations. The company has
also recently started using Atlas Vector
The GoBots AI assistants kick into action as Search to identify and retrieve semantically
soon as a customer asks a question on the similar answers to past questions. The search
marketplace site, with the questions stored results power a co-pilot-like experience for
in MongoDB Atlas. GoBots’ natural language customer service representatives and
models are programmatically called via a provide in-context training to its fleet of
REST API to perform tasks like named entity LLMs.
recognition (NER), user intent detection, and
Story Tools Studio harnesses cutting-edge generative AI (gen AI) technologies to craft immersive,
personalized, and infinite storytelling experiences. Their flagship game Myth Maker AI leverages
MUSE (Modular User Story Engine), an internally developed AI-powered, expert-guided story
generator that blends a growing collection of advanced AI technology with creative artistry to weave
real time narratives.
MUSE (Modular User Story Engine) combines professionally crafted stories with user-empowered
experiences. Players make intentional choices that guide the story with AI adapting to each decision
in real time, providing a unique and personalized journey. MUSE separates the story from game
mechanics, allowing the development of multiple game types. Its use of AI creates more agile teams
with fewer dependencies.
“By selecting MongoDB, we were able to create a prototype of our game in just 48 hours. It is only
with MongoDB that we can release new features to production multiple times per day. We
couldnʼt achieve any of this with a relational database.”
Roy Altman
Founder and CEO at Story Tools Studio
The engineering team has used MongoDB Atlas By running MongoDB in Atlas, Story Tools
from the very start of the company. MongoDB Studio’s engineering team:
stores all of the data used in the platform:
● Is free to focus on AI-driven gaming
— user data, scripts, characters, worlds, coins,
experiences, and not on the grind of
and prompts are all richly structured objects
managing a database
stored natively in MongoDB. The games are
● Was able to scale seamlessly and
built in React and Javascript.
automatically as the team graduated from
Beyond gameplay, the company’s developers their closed beta into public beta
are now exploring MongoDB’s ACID ● Manage player demands with dozens of
transactional integrity to support in-game new players every day – every 24 hours they
monetization, alongside in-app intelligence to are organically adding dozens of new players
further improve the gaming experience through with tens of gigabytes of new data streaming
player analytics. into the platform
Figure 80: Story Tools modular user experience engine
When a player starts a game in Myth Maker the team is working on generating enhanced
AI, they are presented with the option to 3D assets and video from text prompts. With
choose their starting hero character. Under the pace of AI advancement, the creativity
the covers, MUSE calls the GPT4 API, which of the team, and the input from game
takes the player’s selection and writes a fully testers, Story Tools Studio has the flexibility
customized adventure premise. From that to continuously deploy new features with
initial personalized script, MUSE MongoDB’s dynamic and flexible document
programmatically calls specialized AI models data model. This enables Story Tools Studio
to collaboratively generate an immersive, to build a truly innovative, artistic platform,
multimodal gaming experience using opening up a whole new world of
images, animation, audio, and soon, video experiences for both creators and audiences
and 3D. alike.
Of the many use cases set to be transformed by generative AI (gen AI), the bleeding edge of this
revolution is underway with software development. Developers are using gen AI to improve
productivity by writing higher-quality code faster. Tasks include autocompleting code, writing docs,
generating tests, and answering natural language queries across a code base. How does this translate
to adoption? A recent survey showed 44% of new code being committed was written by an AI code
assistant.
Codeium is one of the leaders in the fast-growing AI code assistant space. Its AI toolkit is used by
hundreds of thousands of developers for more than 70 languages across more than 40 IDEs including
Visual Studio Code, the JetBrains suite, Eclipse, and Jupyter Notebooks. The company describes its
toolkit as “the modern coding superpower,” reflected by its recent $65 million Series B funding round
and five-star reviews across extension marketplaces. Codeium was developed by a team of
researchers and engineers to build on the industry-wide momentum around large language models,
specifically for code. They realized that their specialized generative models, when deployed on their
world-class optimized deep learning serving software, could provide users with top-quality AI-based
products at the lowest possible costs.
Codeium has recently trained its models on “MongoDB is wildly popular across the
MongoDB code, libraries, and documentation. developer community. This is because
Now developers building apps with MongoDB
Atlas integrates the fully managed
can install the Codeium extension on the IDE of
their choice and enjoy rapid code completion database services that provide a unified
and codebase-aware chat and search. developer experience across
Developers can stay in the flow while they build, transactional, analytical, and generative
coding at the speed of thought, knowing that AI apps.”
Codeium has ingested MongoDB best practices
Anshul Ramachandran
and documentation.
Head of Enterprise & Partnerships at Codeium
Getting Started with MongoDB and Codeium
MongoDB APIs are incredibly powerful, but In its announcement blog post and YouTube
due to the breadth and richness of the APIs, video, the Codeium team shows how to build
it is possible for developers to be spending an app in VSCode with MongoDB serving as
more time than necessary looking through the data layer. Developers can ask questions
API documentation or using the APIs on how to read and write to the database,
inefficiently for the task at hand. An AI get code competition suggestions, explore
assistant, if trained properly, can effectively specific functions and syntax, handle errors,
assist the developer in retrieval and usage and more. This was all done at no cost using
quality of these APIs. Unlike other AI code the MongoDB Atlas free tier and Codeium
assistants, we at Codeium build our LLMs 100% free.
from scratch and own the underlying data
layer. This means we accelerate and optimize You can get started today by registering for
the developer experience in unique and novel MongoDB Atlas and then downloading the
ways unmatched by others. Codeium extension.
Jina AI has swiftly risen as a leader in multimodal AI, focusing on prompt engineering and
embedding models. With its commitment to open-source and open research, Jina AI is bridging the
gap between advanced AI theory and the real world AI-powered applications being built by
developers and data scientists. Over 400,000 users are registered to use the Jina AI platform.
Jina AI’s work in embedding models has caught significant industry interest. As many developers
now know, embeddings are essential to generative AI (gen AI). Embedding models are sophisticated
algorithms that transform and embed data of any structure into multi-dimensional numerical
encodings called vectors. These vectors give data semantic meaning by capturing its patterns and
relationships. This means we can analyze and search for unstructured data in the same way we’ve
always been able to with structured business data. Considering that over 80% of the data we create
every day is unstructured, we start to appreciate how transformational embeddings — when
combined with a powerful solution such as MongoDB Atlas Vector Search — are for gen AI.
“Our Embedding API is natively integrated with key technologies within the gen AI
developer stack including MongoDB Atlas, LangChain, LlamaIndex, Dify, and
Haystack. MongoDB Atlas unifies application data and vector embeddings in a
single platform, keeping both fully synced. Atlas Triggers keeps embeddings fresh
by calling our Embeddings API whenever data is inserted or updated in the
database. This integrated approach makes developers more productive as they
build new, cutting-edge AI-powered apps for the business.”
Dr. Han Xiao
Founder and CEO at Jina AI
Figure 81: Jina AI’s world-class embedding models improve search and RAG systems.
Jina AI's jina-embeddings-v2 is the first In our published Jina Embeddings v2 and
open-source 8K text embedding model. Its MongoDB Atlas article we show developers
8K token length provides deeper context how to get started in bringing vector
comprehension, significantly enhancing embeddings into their apps. The article
accuracy and relevance for tasks like covers:
retrieval-augmented generation (RAG) and
semantic search. Jina AI’s embeddings offer 1. Creating a MongoDB Atlas instance
enhanced data indexing and search and loading it with your data. (The
capabilities, along with bilingual support. The article uses a sample Airbnb reviews
embedding models are focused on singular data set.)
languages and language pairs, ensuring 2. Creating embeddings for the data
state-of-the-art performance on set using the Jina Embeddings API.
language-specific benchmarks. Currently, 3. Storing and indexing the
Jina Embeddings v2 includes bilingual embeddings with Atlas Vector
German-English and Chinese-English models, Search.
with other bilingual models in the works. 4. Implementing semantic search using
the embeddings.
Jina AI’s embedding models excel in
classification, reranking, retrieval, and
summarization, making them suitable for
diverse applications, especially those that
are cross-lingual. Recent examples from
multinational enterprise customers include
the automation of sales sequences, skills
matching in HR applications, and payment
reconciliation with fraud detection.
Their enterprise platform is built for efficient and secure scalability, self-hosted on existing data
infrastructure (whether in the cloud or on-prem), so that data does not need to be shared externally.
Superduper offers different AI app and workflow templates ready-to-install on MongoDB and
highly configurable with minimal development effort. Because the application templates are based
on their open-source development framework, you can fully own the codebases and control any level
of implementation detail. Current use-cases include Multi-Modal Vector Search & RAG, Document
Extraction & Analysis, Anomaly Detection, Visual Object Detection, and Image and Video Search.
“We integrate MongoDB as a core backend database for our platform, using the
PyMongo driver for app connectivity and Atlas Vector Search for vector
embedding storage and querying. Having MongoDB Ventures on board enables us
to work even more closely with their teams, optimizing our product while
engaging with MongoDBʼs vast developer community.”
Duncan Blythe
Co-Founder & CTO of Superduper.io
Figure 82: AI application lifecycle with Superduper.io
● Ability to switch between and combine ● Scalable model hosting, low latency
different tools and providers, even in the streaming inference and model
same workflow training/fine-tuning
4149.AI boosts team productivity with a dedicated AI teammate. In a successful private beta, nearly
1,000 teams leveraged this agent to streamline goal tracking and tasks. It analyzes team
communication, identifies roadblocks, and takes action in Slack discussions, meetings, calls, reports,
emails, and task trackers.
● The ability to store summaries and chat ● Being part of MongoDB’s AI Innovators
history alongside vector embeddings in program provides 4149.AI with access to
the same database accelerates developer technical support and free Atlas credits,
velocity and the release of new features. helping them quickly experiment using the
native AI capabilities available in the
● The hybrid search capability of MongoDB
MongoDB developer data platform.
Atlas allows pre-filtering data with
keyword-based Atlas Search before
semantically searching vectors, which helps
retrieve relevant information faster.
4149.AI helps teams get more work done by The engineers at 4149.AI evaluated multiple
providing them with their very own vector stores before deciding on Atlas Vector
AI-powered teammate. During the Search. The ability to store summaries and
company’s private beta program, the chat history alongside vector embeddings in
autonomous AI agent has been used by the same database accelerates developer
close to 1,000 teams to help them track velocity and the release of new features. It
goals and priorities. It does this by building also simplifies the technology stack by
an understanding of team dynamics and eliminating unnecessary data movement.
unblocking key tasks. It participates in slack
threads, joins meetings, transcribes calls, Looking forward 4149.AI has an aggressive
generates summaries from reports and roadmap for its products as it starts to more
whitepapers, responds to emails, updates fully exploit the chain-of-thought and
issue trackers, and more. multimodal capabilities provided by the
most advanced language models. This will
4149.AI uses a custom-built AI-agent enable the AI co-worker to handle more
framework leveraging a combination of creative tasks requiring deep reasoning such
embedding models and LLMs from OpenAI as conducting market research, monitoring
and AI21 Labs, with text generation and the competitive landscape, and helping
entity extraction managed by Langchain. identify new candidates for job vacancies.
The models process project documentation The goal for these AI teammates is for them
and team interactions, persisting summaries to eventually be able to take the initiative in
and associated vector embeddings into what to do next rather than rely on someone
Atlas Vector Search. There is even a no-code to manually assign them a task.
way for people to customize and expand the
functionality of their AI teammate. Over
time, the accumulated context generated for
each team means more and more tasks can
be offloaded to their AI-powered co-worker.
In the rapidly evolving digital economy, Zelta.AI stands as a beacon for product managers
navigating the sea of customer feedback. Born out of the need to synthesize diverse feedback into
coherent development plans, Zelta.AI is revolutionizing the way businesses prioritize their product
roadmaps
● MongoDB provides Zelta with the flexibility to provide customers with real-time
to constantly experiment with new features. dashboards and reporting of trends in
They can add fields and evolve the data product feedback.
model as needed without any of the
● Looking forward, as Zelta plans on creating
expensive schema migration pains imposed
its own custom models, MongoDB will prove
by relational databases.
invaluable as a source of labeled data for
● Zelta makes heavy use of the MongoDB supervised model training.
aggregation pipeline for application-driven
intelligence. Without having to ETL data out
of MongoDB, they can analyze data in place
Figure 83: Zelta leverages LLMs to process unstructured data and returns actionable insights for
product teams
The company’s engineering team uses a “We also make heavy use of the MongoDB
combination of fine-tuned OpenAI GPT-4, aggregation pipeline for application-driven
Cohere, and Anthropic models to extract, intelligence. Without having to ETL data out
classify, and encode source data into trends of MongoDB, we can analyze data in place
and sentiment around specific topics and to provide customers with real-time
features. MongoDB Atlas is used as the data dashboards and reporting of trends in
storage layer for source metadata and product feedback. This helps them make
model outputs. product decisions faster, making our service
more valuable to them.”
“The flexibility MongoDB provides us has
been unbelievable. My development team Mick Cunningham
can constantly experiment with new CTO and Co-Founder at Zelta AI
features, just adding fields and evolving the
data model as needed without any of the
expensive schema migration pains imposed
by relational databases.”
Mick Cunningham
CTO and Co-Founder at Zelta AI
Crewmate is a no-code builder for embedded AI-powered communities. The company’s builder
provides customizable communities for brands to deploy directly onto their websites. Crewmate is
already used today across companies in consumer packaged goods (CPG), B2B SaaS, gaming,
Web3, and more.
● MongoDB Atlas provides integrations with stored in MongoDB, without the need to
the fast-evolving AI ecosystem. Crewmate extract, transform, and load (ETL) it into a
leverages this capability to easily integrate separate data warehouse or data lake.
with other AI models, such as OpenAI's
● Crewmate utilizes Atlas Vector Search, a
ada-002 and potentially other models like
feature provided by MongoDB Atlas, to
Llama in the future.
power context-aware semantic search.
● Crewmate utilizes MongoDB's Query API to This enables users visiting a brand's website
process, aggregate, and analyze user to automatically access relevant content
engagement data. This allows brands to such as social media posts, forum
track community outreach efforts and discussions, job postings, and special offers.
conversions directly from the app data
Personalized Community Content with Atlas Vector Search
Potion enables salespeople to personalize prospecting videos at scale. Already over 7,500 sales
professionals at companies including SAP, AppsFlyer, CaptivateIQ, and Opensense are using
SendPotion to increase response rates, book more meetings, and build customer trust.
“We use the MongoDB database to store metadata for all the videos, including the
source content for personalization, such as the contact list and calls to action. For
every new contact entry created in MongoDB, a video is generated for it using our
AI models, and a link to that video is stored back in the database. MongoDB also
powers all of our application analytics and intelligence. With the insights we
generate from MongoDB, we can see how users interact with the service,
capturing feedback loops, response rates, video watchtimes, and more. This data
is used to continuously train and tune our models in Sagemaker."
Kanad Bahalkar
Co-Founder & CEO at Potion
Scaling Potion with MongoDB Atlas
On selecting MongoDB Kanad says, “I had To further enrich the SendPotion service,
prior experience of MongoDB and knew Kanad is planning to use more of the
developer features within MongoDB Atlas.
how easy and fast it was to get started for
This includes Atlas Vector Search to power
both modeling and querying the data. Atlas
AI-driven semantic search and RAG for
provides the best-managed database users who are exploring recommendations
experience out there, meaning we can across video libraries. The engineering team
safely offload running the database to is also planning on using Atlas Triggers to
MongoDB. This ease-of-use, speed, and enable event-driven processing of new
video content.
efficiency are all critical as we build and
scale the business." Potion is a member of the MongoDB AI
Innovators program. Asked about the value
of the program, Kanad responds, “Access to
free credits helped support rapid build and
experimentation on top of MongoDB,
coupled with access to technical guidance
and support."
● Artificial Nerds chose MongoDB for its ● By eliminating the need for a separate
flexible schema, which allows them to store search engine and ETL, MongoDB Atlas
richly structured conversation history, reduces the complexity of development and
messages, and user data. This flexibility is management. This allows developers to focus
crucial for a company focused on on building their application without worrying
AI-powered products, as it enables them to about maintaining separate data stores.
adapt and evolve their data structures as
needed to support their evolving suite of
products and services.
AI-Fueled Search and Innovation: Artificial Nerds Speeds Up with MongoDB Atlas
By adopting Atlas Search, the company Artificial Nerds is growing fast, with revenues
streamlined its search capabilities, expanding 8% every month. The company
integrating a powerful full-text index directly continues to push the boundaries of
onto its database collections. This eliminated customer service by experimenting with new
the need for separate search engines and models including the Llama 2 LLM and
ETL mechanisms, reducing cognitive multilingual sentence transformers hosted in
overhead. Similarly, the release of Atlas Hugging Face. Being part of the MongoDB
Vector Search further enhanced efficiency by AI Innovators program helps Artificial Nerds
replacing a standalone vector database with stay abreast of all of the latest MongoDB
MongoDB Atlas, resulting in improved product enhancements and provides the
developer productivity and a 4x reduction in company with free Atlas credits to build new
latency for a better customer experience. features.
Algomo uses generative AI to help companies offer their best service to both their customers and
employees across more than 100 languages. The company’s name is a portmanteau of the words
Algorithm (originating from Arabic) and Homo, (human in Latin). It reflects the two core design
principles underlying Algomo’s products:
● Alamo chose MongoDB due to its flexible ● Alamo's engineers are considering Atlas
document data model, allowing them to Vector Search as a replacement for their
store customer data alongside conversation current standalone vector database. This
history and messages, ensuring long-term move not only reduces costs but also
memory for context and continuity in support simplifies their codebase by eliminating the
interactions. need to synchronize data across two
separate systems.
● MongoDB Atlas as a fully managed cloud
service relieves Alamo's team from
operational heavy lifting, enabling them to
focus on building conversational
experiences rather than managing
infrastructure.
Figure 84: Algomo Bot
With Algomo, customers can get a Customers can instantly turn their support
ChatGPT-powered bot up on their site in less articles, past conversations, slack channels,
than 3 minutes. More than just a bot, Algomo Notion pages, Google Docs, and content on
also provides a complete conversational their public website into personalized
platform. This includes Question-Answering answers. Algomo vectorizes customer
text generators and autonomous agents that content, using that alongside OpenAI’s
triage and orchestrate support processes, ChatGPT. The company uses RAG (Retrieval
escalating to human support staff for live Augmented Generation) prompting to inject
chat as needed. It works across any relevant context to LLM prompts and
communication channel from web and Chain-Of-Thought prompting to increase
Google Chat to Intercom, Slack, WhatsApp, answer accuracy. A fine-tuned
and more. implementation of BERT is also used to
classify user intent and retrieve custom
FAQs.
The conversation about leveraging MongoDB within industries doesn't end here. We
invite you to delve deeper into MongoDB's capabilities and offerings to discover
how you can build the future of AI applications.
Contact us at [email protected]