Databricks Ai Security Framework Dasf Whitepaper v4 Final
Databricks Ai Security Framework Dasf Whitepaper v4 Final
Databricks AI
Security Framework
(DASF)
Version 1.0
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 1
Table of Contents
Executive Summary 3
1 Introduction 5
1.1 Intended audience 6
1.2 How to use this document 7
4 Conclusion 66
6 Acknowledgments 70
7 Appendix: Glossary 72
8 License 84
Authors
Resources and
The Databricks Security team created the Databricks AI Security Framework (DASF)
Further Reading
to address the evolving risks associated with the widespread integration of AI globally.
Unlike approaches that focus solely on securing models or endpoints, the DASF adopts a
Acknowledgments
comprehensive strategy to mitigate cyber risks in AI systems. Based on real-world evidence
Appendix: indicating that attackers employ simple tactics to compromise ML-driven systems, the
Glossary
DASF offers actionable defensive control recommendations. These recommendations can
License
be updated as new risks emerge and additional controls become available. The framework’s
development involved a thorough review of multiple risk management frameworks,
recommendations, whitepapers, policies and AI security acts.
The DASF is designed for collaboration between business, IT, data, AI and security teams
throughout the AI lifecycle. It addresses the evolving nature of data science from a
research-oriented to a project-based discipline, facilitating structured conversations on
security threats and mitigations without needing deep expertise crossover. We believe
the document will be valuable to security teams, ML practitioners and governance officers,
providing insights into how ML impacts system security, applying security engineering
principles to ML, and offering a detailed guide for understanding the security and
compliance of specific ML systems.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 3
The DASF walks its readers through the 12 foundational components of a generic data-
centric AI system: raw data, data prep, datasets, data and AI governance, machine
learning algorithms, evaluation, machine learning models, model management, model
serving and inference, inference response, machine learning operations, and data and AI
platform security. Databricks identified 55 technical security risks that arise across these
⟶ Executive components and dedicated a chapter describing the specific component, the associated
Summary
risks and the available controls we recommend you leverage. We also provide a guide to
Introduction each AI and ML mitigation control — its shared responsibility between Databricks and your
organization, and the associated Databricks technical documentation available to learn how
Risks in AI System
Components
to enable said control.
The framework concludes with Databricks’ final recommendations on how to manage and
Understanding
Databricks Data deploy AI models safely and securely, which are consistent with the core tenets of machine
Intelligence Platform
AI Risk Mitigation learning adoption: identify the ML business use case, determine the ML deployment model,
Controls
select the most pertinent risks, enumerate threats for each risk and choose which controls
Conclusion to implement. We also provide further reading to enhance your knowledge of the AI field and
the frameworks we reviewed as part of our analysis. While we strive for accuracy, given the
Resources and
Further Reading
evolving nature of AI, please feel free to contact us with any feedback or suggestions. Your
input is valuable to us. If you want to participate in one of our AI Security workshops, please
Acknowledgments contact [email protected]. If you are curious about how Databricks approaches security,
please visit our Security and Trust Center.
Appendix:
Glossary
License
Learn more ⟶
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 4
01
Introduction
Machine learning (ML) and generative AI (GenAI) are revolutionizing the future of work.
Organizations understand that AI is helping to build innovation, maintain competitiveness
and improve the productivity of their employees. Equally, organizations understand that
Executive
Summary
their data provides a competitive advantage for their artificial intelligence (AI) applications.
Leveraging these technologies presents opportunities but also potential risks. There is a risk
⟶ Introduction of security and privacy breaches, as the data sent to an external large language model (LLM)
could be leaked or summarized. Several organizations have even banned the use of ChatGPT
Risks in AI System
Components due to sensitive enterprise data being sent by users. Organizations are also concerned about
potential hazards such as data loss, data confidentiality, model theft, and risks of ensuring
Understanding
Databricks Data
existing and evolving compliance and regulation when they use their data for ML and GenAI.
Intelligence Platform
AI Risk Mitigation
Without the proper access controls, users can use generative AI models to find confidential
Controls
data they shouldn’t have access to. If the models are customer-facing, one organization
might accidentally receive data related to a different organization. Or a skilled attacker can
Conclusion
extract data they shouldn’t have access to. Without the auditability and traceability of these
Resources and models and their data, organizations face compliance risks.
Further Reading
AI adoption also brings a crucial regulatory dimension, emphasizing the need for thoughtful
Acknowledgments
oversight and responsible governance. In October 2023, President Biden issued an Executive
Order on safe, secure and trustworthy artificial intelligence, emphasizing the responsible
Appendix:
Glossary development and use of AI technologies. The National Institute of Standards and Technology
(NIST) recently published its Artificial Intelligence Risk Management Framework (AI RMF) to
License
help federal agencies manage and secure their information systems. It provides a structured
process for identifying, assessing and mitigating cybersecurity risks. Gartner’s 2023 Security
Leader’s Guide to Data Security report1 predicts that “at least one global company will
see its AI deployment banned by a regulator for noncompliance with data protection or AI
governance legislation by 2027.” With ownership accountability and an ever-evolving legal
and regulatory landscape, data, IT and security leaders are still unclear on how to take
advantage of generative AI for their organization while mitigating any perceived risks.
The Databricks Security team developed the Databricks AI Security Framework (DASF) to
help organizations understand how AI can be safely realized and risks mitigated as the global
community incorporates AI into more systems.
DATABRICKS 1
Gartner, Security Leader’s Guide to Data Security, Andrew Bales. September 7, 2023.
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 5
The DASF takes a holistic approach to mitigating AI security risks instead of focusing only
on the security of models or model endpoints. Abundant real-world evidence suggests
that attackers use simple tactics to subvert ML-driven systems. That is why, with the DASF,
we propose actionable defensive control recommendations. These recommendations
are subject to change as new risks are identified and new controls are made available. We
Executive reviewed many risk management frameworks, recommendations, whitepapers, policies and
Summary
acts on AI security. We encourage the audience to review such material, including some of
⟶ Introduction the material linked in the resources section of this document. Your feedback is welcome.
Risks in AI System
Components 1.1 Intended audience
Understanding The Databricks AI Security Framework is intended to be used by data and AI teams
Databricks Data
Intelligence Platform collaborating with their security teams across the AI/ML lifecycle. Traditionally, the skill sets
AI Risk Mitigation
Controls of data scientists, data engineers, security teams, governance officers and DevSecOps
engineering teams did not overlap. The communication gap between data scientists and
Conclusion
these teams was manageable, given the research-oriented nature of data science and its
Resources and primary focus on delivering information to executives. However, as data science transforms
Further Reading
into a project-based discipline, it becomes crucial for these teams to collaborate.
Acknowledgments The guidance in this document provides a way for disciplines to have structured
conversations on these new threats and mitigations without requiring security engineers to
Appendix:
Glossary become data scientists or vice versa. We mostly did this work for our customers to ensure
the security and compliance of production ML use cases on the Databricks Data Intelligence
License
Platform. That said, we believe that what we have produced will be helpful to three major
audience groups:
Security teams (CISOs, security leaders, DevSecOPs, SREs) can use the DASF
to understand how ML will impact the security of systems they may be asked to
secure, as well as to understand some of the basic mechanisms of ML.
Governance leaders, risk officers and policymakers can use the DASF as
DATABRICKS a detailed guide into a risk mindset to learn more about the security and
AI SECURITY
FRAMEWORK compliance of specific ML systems.
(DASF)
VERSION 1.0 6
If you are new to GenAI, you can build foundational knowledge, including large language
models (LLMs), with four short videos in this Generative AI Fundamentals course created by
Databricks. In this free training, you will learn what generative AI is, what the main generative AI
applications are, and their capabilities and potential applications across various domains. It will
also cover the limits and risks of generative AI technologies, including ethical considerations.
Executive
Summary
1.2 How to use this document
⟶ Introduction
The Databricks AI Security Framework is designed for collaborative use throughout the AI
Risks in AI System
Components lifecycle by data and AI teams and their security counterparts referenced above. The DASF is
meant to foster closer collaboration between these teams and improve the overall security
Understanding of AI systems. The concepts in this document are applicable for all teams, even if they do not
Databricks Data
Intelligence Platform use Databricks to build their use cases. That said, we will refer to documentation or features
AI Risk Mitigation
Controls in Databricks terminology where it allows us to simplify our language or make this document
more actionable for our direct customers. We hope those who do not use Databricks will be
Conclusion
able to follow along without issue.
Resources and
Further Reading First, we suggest that organizations find out what type of AI models are being built or being
used. As a guideline, we define model types broadly as the following:
Acknowledgments
External models (third-party services). These are models that are hosted
outside of Databricks. Endpoints that serve external models can be centrally
DATABRICKS
AI SECURITY
governed and customers can establish rate limits and access control for them.
FRAMEWORK Examples include foundation models such as OpenAI’s GPT-4, Anthropic’s
(DASF)
VERSION 1.0 Claude and others. 7
Second, we recommend that organizations identify where in their organization AI systems are
being built, the process, and who is responsible. The modern AI system lifecycle often involves
diverse stakeholders, including business stakeholders, subject matter experts, governance
officers, data engineers, data scientists, research scientists, application developers,
administrators, AI security engineers, DevSecOps engineers and MLSecOps engineers.
Executive
Summary We recommend that those responsible for AI systems begin by reviewing the 12 foundational
components of a generic data-centric AI system and the types of AI models, as outlined in
⟶ Introduction
Section 2: Risks in AI System Components. This section details security risk considerations
Risks in AI System
and potential mitigation controls for each component, helping organizations reduce overall
Components
risk in their AI system development and deployment processes. Each security risk is mapped
to a set of mitigation controls that are ranked in prioritized order, starting with the perimeter
Understanding
Databricks Data security to data security. These guidelines apply to providers of all AI systems, whether built
Intelligence Platform
AI Risk Mitigation from scratch or using third-party tools and services, and encompass both predictive ML
Controls
models and generative AI models.
Conclusion
To further refine risk identification, we categorize risks by model type: predictive ML models,
Resources and
RAG-LLMs, fine-tuned LLMs, pretrained LLMs, foundation models and external models. Once
Further Reading
the relevant risks are identified, teams can determine which controls are applicable from
the comprehensive list in Section 3: Understanding Databricks Data Intelligence Platform
Acknowledgments
AI Risk Mitigation Controls. Each control is tagged as “Out-of-the-box,” “Configuration”
Appendix: or “Implementation,” helping teams estimate the effort involved in the implementation of
Glossary
the control on the Databricks Data Intelligence Platform, with reference links to relevant
License
documentation provided.
Our experience shows that implementing these guidelines helps customers build secure and
functional AI systems.
When I think about what makes a good accelerator, it’s all about making things smoother,
more efficient and fostering innovation. The DASF is a proven and effective tool for
security teams to help their partners get the most out of AI. Additionally, it lines up with
established risk frameworks like NIST, so it’s not just speeding things up – it’s setting a
solid foundation in security work.
DATABRICKS
AI SECURITY
FRAMEWORK Riyaz Poonawala
(DASF) Vice President of Information Security
VERSION 1.0 8
02
Risks in AI System Components
The DASF starts with a generic AI system in terms of its constituent components and works
through generic system risks. By understanding the components, how they work together
and the risk analysis of such architecture, an organization concerned about security can
Executive
Summary
get a jump start on determining risks in its specific AI system. The Databricks Security team
considered these risks and built mitigation controls into our Databricks Data Intelligence
Introduction Platform. We mapped the respective Databricks Platform control and link to Databricks
product documentation for each risk.
⟶ Risks in AI System
Components
AI System Components
Understanding
Databricks Data
Intelligence Platform Governance
AI Risk Mitigation
Controls
2 3 3
4 Serving
Data Prep Datasets Develop and ævaluate Model Infrastructure
Conclusion
4 1T
ETL
Training 5 Custom models
9
Clean data
AlgoritÊm
Inference requests
Validation External models
Resources and Exploratory data 2
Prompt/RAG
analytics (EDA) 6 4
Further Reading 7 FineÝtuninÚ
Featurization Test Evaluation and pretrained
Joins, aggregations, model
transformations, etc. Model serving
Featur
Acknowledgments extraction
4
4 8
2 Model
1 Catalog Management
Inference response
Mode
l AI Gateway
1T assets
Appendix: Raw Data
Features
Glossary Your data for RAG
Indexes Vector searcÊ and feature
/
function lookup
Models
Monitor 5
License Logs 1T
New ML and RLHF data
1 6
11 Operations and Platform 12
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 9
Data operations (#1-#4 in Figure 1) include ingesting and transforming data and
ensuring data security and governance. Good ML models depend on reliable data
pipelines and secure DataOps infrastructure.
⟶ Risks in AI System
Model deployment and serving (#9 and #10 in Figure 1) consists of securely
Components building model images, isolating and securely serving models, automated scaling,
rate limiting, and monitoring deployed models. Additionally, it includes feature
Understanding
Databricks Data
and function serving, a high-availability, low-latency service for structured data
Intelligence Platform in retrieval augmented generation (RAG) applications, as well as features that are
AI Risk Mitigation
Controls required for other applications, such as models served outside of the platform or
any other application that requires features based on data in the catalog.
Conclusion
Operations and platform (#11 and #12 in Figure 1) include platform vulnerability
Resources and
Further Reading management and patching, model isolation and controls to the system, and
authorized access to models with security in the architecture. Also included is
Acknowledgments operational tooling for CI/CD. It ensures the complete lifecycle meets the required
standards by keeping the distinct execution environments — development,
Appendix: staging and production — for secure MLOps.
Glossary
License
In our analysis of AI systems, we identified 55 technical security risks across the 12
components based on the AI model types deployed by our customers (namely, predictive ML
models, generative foundation models and external models as described above), customer
questions and questionnaires, security reviews of customer deployments, in-person CISO
workshops, and customer surveys about AI risks. In the table below, we outline these basic
components that align with steps in any AI system and highlight the types of security risks
our team identified.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 10
SY S T E M SY S T E M P OT E N T I A L
S TAG E COMPONENTS (F I G U R E 1) S E C U R I T Y R I S KS
19 specific risks:
1 Raw data → 1.1 Insufficient access controls →
2 Data preparation → 1.2 Missing data classification →
Data
3 Datasets → 1.3 Poor data quality →
operations
4 C
atalog and 1.4 Ineffective storage and encryption →
governance → 1.5 Lack of data versioning →
Executive 1.6 Insufficient data lineage →
Summary
1.7 Lack of data trustworthiness →
1.8 Data legal →
Introduction 1.9 Stale data →
1.10 Lack of data access logs →
⟶ Risks in AI System
2.1 Preprocessing integrity →
Components
2.2 Feature manipulation →
2.3 Raw data criteria →
Understanding
2.4 Adversarial partitions →
Databricks Data
Intelligence Platform
AI Risk Mitigation 3.1 Data poisoning →
Controls 3.2 Ineffective storage and encryption →
3.3 Label flipping →
Conclusion 4.1 L
ack of traceability and transparency of
model assets →
4.2 Lack of end-to-end ML lifecycle →
Resources and
Further Reading
14 specific risks:
Acknowledgments
5 ML algorithm → 5.1 Lack of tracking and reproducibility of experiments ⟶
6 Evaluation → 5.2 Model drift ⟶
Appendix:
Model
7 Model build → 5.3 Hyperparameters stealing ⟶
Glossary operations
8 Model management → 5.4 Malicious libraries ⟶
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 11
SY S T E M SY S T E M P OT E N T I A L
S TAG E COMPONENTS (F I G U R E 1) S E C U R I T Y R I S KS
15 specific risks:
9 M
odel Serving — 9.1 Prompt inject →
inference requests → 9.2 Model inversion →
Model 10 odel Serving —
M 9.3 Model breakout →
deployment inference responses →
Looped input →
and serving 9.4
7 specific risks:
Conclusion 11 ML operations → 11.1 L
ack of MLOps — repeatable enforced standards →
12 ML platform →
Operations 12.1 L
ack of vulnerability management →
Resources and
Further Reading
and 12.2 L
ack of penetration testing and bug bounty →
platform 12.3 L
ack of incident response →
12.4 U
nauthorized privileged access →
Acknowledgments
12.5 P
oor SDLC →
12.6 L
ack of compliance →
Appendix:
Glossary
License The 12 foundational components of a generic data-centric AI/ML model and risk
considerations are discussed in detail below.
Note: We are aware of nascent risks such as energy-latency attacks, rowhammer attacks,
side channel attacks, evasion attacks, functional adversarial attacks and other adversarial
examples, but these are out of scope for this version of the framework. We may reconsider
these and any new novel risks in later versions if we see them becoming material.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 12
2.1 Raw Data
Data is the most important aspect of AI systems because it provides the foundation that all
AI functionality is built on. Raw data includes enterprise data, metadata and operational data.
It can be semi-structured or unstructured such as images, sensor data, documents. This
data can be batch data or streaming data. Data security is paramount and equally important
for ensuring the security of machine learning algorithms and any technical deployment
Executive
Summary particulars. Securing raw data is a challenge in its own right, and all data collections in an AI
system are subject to the usual data security challenges and some new ones. A fully trained
Introduction
machine learning (ML) system, whether online or offline, will inevitably encounter new input
⟶ Risks in AI System data during normal operations or retraining processes. Fine-tuning and pretraining of LLMs
Components
further increases these risks by allowing customizations with potentially sensitive data.
Understanding
Databricks Data
Intelligence Platform RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
AI Risk Mitigation
Controls
RAW DATA 1.1
Data operations →
Pre-trained LLMs: | Foundational models: | External models:
criticality. As data volumes grow exponentially, Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
prioritizing sensitive information protection, risk Pre-trained LLMs: | Foundational models: | External models:
reduction and data quality becomes imperative.
Classification facilitates the implementation of
appropriate security measures and governance
policies by evaluating data’s risk and value. A
robust classification strategy strengthens data
governance, mitigates risks, and ensures data
integrity and security on a scalable level.
Data operations →
⟶ Risks in AI System
Components
Data operations →
Acknowledgments
regulations require organizations to have a clear Pre-trained LLMs: | Foundational models: | External models:
AI and machine learning for anomaly detection. Pre-trained LLMs: | Foundational models: | External models:
Data operations →
Executive
Summary
Introduction D
elete records from datasets and retrain models to
Data legal DASF 12
Resources and
Further Reading
RAW DATA 1.9
help protect against such risks, maintaining Pre-trained LLMs: | Foundational models: | External models:
Data operations →
Machine learning algorithms require raw input data to be transformed into a representational
form they can understand. This data preparation step can impact the security and
explainability of an ML system, as data plays a crucial role in security. Data preparation
Executive includes the following tasks:
Summary
Introduction
Understanding
2 | Preprocessing data includes tasks like numerical transformations,
Databricks Data aggregating data, encoding text or image data, and creating new features.
Intelligence Platform
|
AI Risk Mitigation
Controls 3 Combining data includes tasks like joining tables or merging datasets.
Conclusion
4 | Label data includes tasks like identifying raw data (images, text files, videos,
and so on) and adding one or more meaningful and informative labels to
License
Companies need not sacrifice security for AI innovation. The Databricks AI Security
Framework is a comprehensive tool to enable AI adoption securely. It not only maps
AI security concerns to the AI development pipeline, but makes them actionable for
Databricks customers with practical controls. We're pleased to have contributed to the
development of this valuable community resource.
DATABRICKS
AI SECURITY
FRAMEWORK Hyrum Anderson
(DASF) CTO
VERSION 1.0 16
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Appendix:
Glossary
DATA PREP 2.2
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 17
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
AI Risk Mitigation
Pre-trained LLMs: | Foundational models: | External models:
Controls
The DASF is a very important, foundational document. I think it will go far in helping to
bridge the knowledge gap between ML and security experts.
DATABRICKS
AI SECURITY
FRAMEWORK Diana Kelley
Protect AI
(DASF) CISO
VERSION 1.0 18
2.3 Datasets
Prepared data must be grouped into different datasets: a training set, a validation set and a
testing set. The training set is used as input to the machine learning algorithm. The validation
set is used to tune hyperparameters and to monitor the machine learning algorithm for
overfitting. The test set is used after learning is complete to evaluate performance.
When creating these groupings, special care must be taken to avoid predisposing the
Executive ML algorithm to future attacks, such as adversarial partitions. In particular, the training
Summary
set deeply influences an ML system’s future behavior. Manipulating the training data
Introduction
represents a direct and potent means of compromising ML systems. By injecting malicious or
adversarial samples into the training set, attackers can subtly influence the model’s behavior,
⟶ Risks in AI System potentially leading to misclassification, performance degradation or even security breaches.
Components
These approaches, often called “data poisoning” or “backdoor attacks,” pose a significant
Understanding
Databricks Data threat to the robustness and reliability of ML systems deployed in various critical domains.
Intelligence Platform
AI Risk Mitigation Dataset security concerns with foundation models include the potential for leaks of sensitive
Controls
information. Fine-tuning and pretraining of LLMs further increases these risks as it allows
Conclusion
customizations with sensitive data.
Resources and
Further Reading RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
DATASETS 3.2
License
Data catalog and governance is a comprehensive approach that comprises the principles,
practices and tools to manage an organization’s data assets throughout their lifecycle.
Managing governance for data and AI assets enables centralized access control, auditing,
lineage, data, and model discovery capabilities, and allows organizations to limit the risk of
data or model duplication, improper use of classified data for training, loss of provenance,
and model theft.
Executive
Summary
Additionally, if sensitive information in datasets is inadequately secured, breaches and leaks
can expose personally identifiable information (PII), financial data and even trade secrets,
Introduction
and cause potential legal repercussions, reputational damage and financial losses.
⟶ Risks in AI System
Components Proper data catalog governance allows for audit trails and tracing the origin and
transformations of data used to train AI models. This transparency encourages trust and
Understanding
Databricks Data
accountability, reduces risk of biases, and improves AI outcomes.
Intelligence Platform
AI Risk Mitigation
Controls
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Conclusion
GOVERNANCE 4.1
and tracking of data and model-related Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
processes Pre-trained LLMs: | Foundational models: | External models:
Data operations →
GOVERNANCE 4.2
Data operations →
Applicable AI deployment model:
DATABRICKS
AI SECURITY Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
FRAMEWORK Pre-trained LLMs: | Foundational models: | External models:
(DASF)
VERSION 1.0 21
2.5 Machine Learning Algorithms
Machine learning algorithms primarily fall into two broad categories: offline and online. Offline
Introduction
systems are trained on a fixed dataset, “frozen” and subsequently used for predictions with
⟶ Risks in AI System new data. This approach is particularly common for classification tasks. Conversely, online
Components
systems continuously learn and adapt through iterative training with new data.
Understanding
Databricks Data From a security perspective, offline systems possess certain advantages. Their fixed, static
Intelligence Platform
AI Risk Mitigation nature reduces the attack surface and minimizes exposure to data-borne vulnerabilities
Controls
over time. In contrast, online systems are constantly exposed to new data, potentially increasing
their susceptibility to poisoning attacks, adversarial inputs and manipulation of learning
Conclusion
processes. Therefore, the choice between offline and online learning algorithms should be made
Resources and carefully, considering the ML system’s specific security requirements and operating environment.
Further Reading
Acknowledgments RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
ALGORITHMS 5.2
⟶ Risks in AI System U
pstream data changes: occur due to Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
Components alterations in data collection or processing Pre-trained LLMs: | Foundational models: | External models:
ALGORITHMS 5.3
Conclusion
Hyperparameters stealing DASF 20 T
rack ML training runs in the model development
process, including parameter settings, securely
Hyperparameters in machine learning are
Resources and DASF 43 U
se access control lists via workspace access controls
often deemed confidential due to their
Further Reading
commercial value and role in proprietary DASF 42 D
ata-centric MLOps and LLMOps employing separate
learning processes. If attackers gain access model lifecycle stages by UC schema
to these hyperparameters, they may steal or
Acknowledgments manipulate them — altering, concealing or even
adding hyperparameters. Such unauthorized Applicable AI deployment model:
interventions can harm the ML system, Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
Appendix: compromising performance and reliability or
revealing sensitive algorithmic strategies.
Pre-trained LLMs: | Foundational models: | External models:
Glossary
Model operations →
License
ALGORITHMS 5.4
Model operations →
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 23
2.6 Evaluation
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Executive
Summary
EVALUATION 6.1
Introduction S
SO with IdP and MFA to limit who can access your data
Evaluation data poisoning DASF 1
and AI platform
Upstream attacks against data, where the data
is tampered with before it is used for machine DASF 2 S
ync users and groups to inherit your organizational roles
⟶ Risks in AI System
learning, significantly complicate the training to access data
Components
and evaluation of ML models. Poisoning of the DASF 3 R
estrict access using IP access lists to restrict the IP
evaluation data impacts the model validation addresses that can authenticate to your data and AI
and testing process. These attacks can corrupt platform
Understanding
or alter the data in a way that skews the training
Databricks Data R
estrict access using private link as strong controls that
process, leading to unreliable models. DASF 4
Intelligence Platform limit the source for inbound requests
AI Risk Mitigation Model operations →
Controls DASF 5 C
ontrol access to data and other objects for permissions
model across all data assets to protect data and sources
DASF 7 E
nforce data quality checks on batch and streaming
Conclusion datasets for data sanity checks, and automatically detect
anomalies before they make it to the datasets
DASF 11 C
apture and view data lineage to capture the lineage all
Resources and the way to the original raw data sources
Further Reading
DASF 45 E
valuate models to capture performance insights for
language models
Acknowledgments DASF 44 T
rigger actions in response to a specific event via
automated jobs to notify human-in-the-loop (HITL)
DASF 49 A
utomate LLM evaluation
Appendix:
Glossary DASF 42 D
ata-centric MLOps and LLMOps unit and integration
testing
DASF
EVALUATION 6.2
A machine learning model is a program that can find patterns or make decisions from a
previously unseen dataset. During training, the machine learning algorithm is optimized to
find certain patterns or outputs from the dataset, depending on the task. The output of this
process — often a computer program with specific rules and data structures — is called a
Executive
Summary machine learning model.
Introduction Deploying a fully trained machine learning model to production introduces several critical
risks to address. Notably, some risks discussed in the previous section on evaluation risks,
⟶ Risks in AI System
such as overfitting, directly apply here. Open source or commercial models, not trained
Components
within your organization, carry the same risks with the added challenge that your organization
Understanding lacks control over the model’s development and training.
Databricks Data
Intelligence Platform
AI Risk Mitigation Additionally, external models may be Trojan horse backdoors or harboring other uncontrolled
Controls
risks, depriving you of the competitive advantage of leveraging your own data and potentially
Conclusion exposing your data to unauthorized access. Therefore, it is crucial to carefully consider and
mitigate these potential risks before deploying any pretrained model to production.
Resources and
Further Reading
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Acknowledgments
MODEL 7.1
Appendix:
Glossary Backdoor machine learning/ DASF 1 S
SO with IdP and MFA to limit who can access your data
and AI platform
Trojaned model
DASF 43 U
se access control lists to limit who can bring models
There are inherent risks when using public ML/
License and limit the use of public models
LLM models or outsourcing their training, akin to
the dangers associated with executable (.exe) DASF 42 D
ata-centric MLOps and LLMOps promote models as
files. A malicious third party handling the training code using CI/CD. Scan third-party models continuously
process could tamper with the data or deliver to identify hidden cybersecurity risks and threats such
a “Trojan model” that intentionally misclassifies as malware, vulnerabilities and integrity issues to detect
specific inputs. Additionally, open source models possible signs of malicious activity, including malware,
may contain hidden malicious code that can tampering and backdoors. See resources section for
exfiltrate sensitive data upon deployment. These third-party tools.
risks are pertinent in both external models and
DASF 23 R
egister, version, approve, promote and deploy models
outsourced model development scenarios,
and scan models for malicious code when using third-
necessitating scrutiny and verification of models
party models or libraries
before use.
DASF 19 M
anage end-to-end machine learning lifecycle
Model operations →
DASF 5 C
ontrol access to data and other objects
DASF 34 R
un models in multiple layers of isolation. Models are
considered untrusted code: deploy models and custom
LLMs with multiple layers of isolation.
MODEL 7.2
Conclusion
MODEL 7.3
MODEL 7.4
Organizations can increase trust by creating a centralized place for model management:
Introduction
development, tracking, discovering, governing, encrypting and accessing models with proper
⟶ Risks in AI System
security controls. Doing so reduces the risk of model theft, improper reuse and model
Components
inversion. Transparency is also added by appropriate levels of information based on the
stage of the AI lifecycle and tailored to the role or knowledge of practitioners or individuals
Understanding
Databricks Data interacting with the AI system. By promoting higher levels of understanding, transparency
Intelligence Platform
AI Risk Mitigation increases confidence in the AI system.
Controls
Conclusion RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Resources and
MODEL MANAGEMENT 8.1
Further Reading
Companies need not sacrifice security for AI innovation. The Databricks AI Security
Framework is a comprehensive tool supporting the adoption of secure AI. We are grateful
for Databricks’ partnership in the journey to trustworthy AI and this tool makes AI security
practical and actionable for Databricks customers.
DATABRICKS
AI SECURITY
FRAMEWORK Robert Booker
(DASF) Chief Strategy Officer
VERSION 1.0 27
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Resources and
Further Reading Applicable AI deployment model:
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 28
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Conclusion
Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
Resources and
Further Reading
Acknowledgments
2.9 Model Serving and Inference Requests
Appendix: Model Serving exposes your machine learning models as scalable REST API endpoints for
Glossary
inference and provides a highly available and low-latency service for deploying models.
License
Deploying a fully trained machine learning model introduces significant risks, including
adversarial inputs, data poisoning, privacy concerns, access control issues, model
vulnerabilities and versioning challenges. Using third-party or SaaS models amplifies
these risks and introduces further limitations like lack of customization, model mismatch,
ownership concerns and data privacy risks. Careful evaluation and mitigation strategies are
necessary to securely and responsibly deploy fully trained models in production.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 29
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 30
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
License
Additional controls to consider:
Robust Intelligence AI Firewall Prompt Injection rule: Flags
malicious user input that might direct the LLM to perform an
action unintended by the model creator.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 31
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Controls
License
DATABRICKS
AI SECURITY Hasan Yasar
Carnegie Mellon
FRAMEWORK Technical Director, Teaching Professor
(DASF) University Continuous Deployment of Capability | Software Engineering Institute
VERSION 1.0 32
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Resources and Designing robust prompts can help mitigate attacks such
Further Reading as jailbreaking.
DASF 45 E
valuate models for custom evaluation metrics
License
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 33
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
DASF 45 E
valuate models for custom evaluation metrics
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 34
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 35
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Introduction DASF 46 C
reate embeddings to securely integrate data objects
with sensitive data that goes into LLMs
⟶ Risks in AI System
DASF 49 A
utomate LLM evaluation to evaluate RAG applications
with LLM-as-a-judge and get out-of-the-box metrics
Components
like toxicity, latency, tokens and more to quickly and
efficiently compare and contrast various LLMs to navigate
your RAG application requirements
Understanding
Databricks Data
Intelligence Platform
Additional controls to consider:
AI Risk Mitigation
Use guardrails to define and enforce assurance for LLM
Controls
applications. Please see the resources section for a collection of
third-party tools.
Conclusion
Applicable AI deployment model:
Model deployment and serving → Pre-trained LLMs: | Foundational models: | External models:
Conclusion
Resources and
Further Reading
2.10 Model Serving and Inference Response
Acknowledgments
While the technical intricacies of the algorithm may seem like the most vulnerable point
Appendix:
Glossary for malicious actors seeking to compromise the integrity and reliability of the ML system,
an equally effective, and often overlooked, attack vector lies in how it generates output
License (inference response). The inference response represents the real-world manifestation of
the model’s learned knowledge and forms the basis for its decision-making capabilities.
Consequently, compromising the inference response directly can have devastating
consequences, undermining the system’s integrity and reliability.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 37
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
License
The DASF is the first-ever framework that would allow businesses to mitigate AI/ML
risks at scale versus approaches that operate in silos — collectivism at best for
responsible AI/ML.
DATABRICKS
AI SECURITY
FRAMEWORK Ebrima N. Ceesay, PhD, CISSP
Capital One Financial
(DASF) Senior Distinguished Engineer
VERSION 1.0 38
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
License
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 40
2.11 Machine Learning Operations (MLOps)
MLOps is a useful approach for creating quality AI solutions. It is a core function of machine
learning engineering, focused on streamlining the process of taking machine learning
models to production and then maintaining and monitoring them. By adopting an MLOps
Executive approach, data scientists and machine learning engineers can collaborate and increase the
Summary
pace of model development and production by implementing continuous integration and
Introduction continuous deployment (CI/CD) practices with proper monitoring, validation and governance
of ML models with a “security in the process” mindset. Organizations without MLOps will risk
⟶ Risks in AI System
missing some of the controls we discussed above or not applying them consistently at scale
Components
to manage thousands of models.
Understanding
Databricks Data
Intelligence Platform
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
AI Risk Mitigation
Controls
OPERATIONS 11.1
Conclusion
Lack of MLOps — repeatable DASF 45 E
valuate models to capture performance insights for
language models
enforced standards
Resources and DASF 44 T
rigger actions in response to a specific event to trigger
Further Reading Operationalizing an ML solution requires joining
automated jobs to keep human-in-the-loop (HITL)
data from predictions, monitoring and feature
tables with other relevant data. DASF 42 D
ata-centric MLOps and LLMOps. MLOps best practices:
separate environments by workspace and schema,
Acknowledgments Duplicating data, moving AI assets, and driving
promote models with code, MLOps Stacks for repeatable
governance and tracking across these stages
ML infra across environments.
may represent roadblocks to practitioners
who would rather shortcut security controls to
Appendix:
deliver their solution. Many organizations will Applicable AI deployment model:
Glossary
find that the simplest way to securely combine
ML solutions, input data and feature tables is to
Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 41
2.12 Data and AI Platform Security
Abundant real-world evidence suggests that actual attackers use simple tactics to subvert
ML-driven systems. The choice of platform used for building and deploying AI models can
have inherent risks and rewards.
Executive
Summary
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
Introduction
PLATFORM 12.1
⟶ Risks in AI System
Lack of vulnerability management DASF 38 P
latform security — vulnerability management to build,
Components
deploy and monitor AI/ML models on a platform that
Detecting and promptly addressing software takes responsibility seriously and shares remediation
vulnerabilities in systems that support data and timeline commitments
Understanding AI/ML operations is a critical responsibility for
Databricks Data software and service providers. Attackers do
Intelligence Platform not necessarily need to target AI/ML algorithms Applicable AI deployment model:
AI Risk Mitigation directly; compromising the layers underlying AI/ Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
Controls ML systems is often easier. Therefore, adhering Pre-trained LLMs: | Foundational models: | External models:
to traditional security threat mitigation practices,
such as a secure software development lifecycle,
Conclusion is essential across all software layers.
Resources and
Further Reading
PLATFORM 12.2
generally more prone to attacks than the AI/ML Pre-trained LLMs: | Foundational models: | External models:
components.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 42
RISK/DESCRIPTION M I T I G AT I O N C O N T R O L S
PLATFORM 12.3
Executive
Summary
PLATFORM 12.4
Conclusion
PLATFORM 12.5
Resources and
Further Reading
Poor security in the software DASF 41 P
latform security — secure SDLC
development lifecycle
Acknowledgments Software platform security is an important part of Applicable AI deployment model:
any progressive security program. ML hackers have Predictive ML models: | RAG-LLMs: | Fine-tuned LLMs:
shown that they don’t need to know sophisticated AI/ Pre-trained LLMs: | Foundational models: | External models:
Appendix: ML concepts to compromise a system. Hackers have
Glossary busied themselves with exposing and exploiting bugs
in a platform where AI is built, as those systems are
well known to them. The security of AI depends on the
License platform’s security.
PLATFORM 12.6
In this section, we delve into the comprehensive risk mitigation controls available in the
Databricks Data Intelligence Platform for Artificial Intelligence (AI) and Machine Learning
Executive
Summary (ML). As organizations increasingly harness the power of AI, a nuanced understanding of
these robust controls becomes imperative to ensure data integrity, security and regulatory
Introduction compliance throughout the data lifecycle.
Risks in AI System
Components
3.1 The Databricks Data Intelligence Platform
⟶ Understanding
Databricks Data Databricks is the data and AI company with origins in academia and the open source
Intelligence Platform
AI Risk Mitigation community. Databricks was founded in 2013 by the original creators of Apache Spark™, Delta
Controls
Lake and MLflow. We pioneered the concept of the lakehouse to combine and unify the best of
Conclusion
data warehouses and data lakes. Databricks made this vision a reality in 2020; since then, it has
seen tremendous adoption as a category. Today, 74% of global CIOs report having a lakehouse
Resources and in their estate, and almost all of the remainder intend to have one within the next three years.
Further Reading
In November 2023, we announced the Databricks Data Intelligence Platform. It’s built on a
Acknowledgments
lakehouse to provide an open, unified foundation for all data and governance. We built the
Appendix:
Data Intelligence Platform to allow every employee in every organization to find success
Glossary
with data and AI. The Data Intelligence Engine, at the heart of the platform, understands
the semantics of your data and how it flows across all of your workloads. This allows for
License
new methods of optimization, as well as for technical and nontechnical users to use natural
language to discover and use data and AI in the context of your business.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 44
Executive
Summary
Introduction
Data Intelligence Platform
Risks in AI System
Mlsaic AI Delta Lie alew Wlrkflow Dataricks §ª
Create, tune and Automate data Optimize job cost Build with
Components serve custom LLMs quality based on past runs text-to-SQL
⟶ Understanding
Data ntellience Xnine
Databricks Data Use nenerative Ad to understand the semantics o your data
Intelligence Platform
AI Risk Mitigation
Controls
Unity Catallg
Unifed security, novernance and cataloninn
Delta Lake
Conclusion
Appendix:
Glossary
License
The Databricks Data Intelligence Platform combines AI assets — from data and features to
models — into one catalog, ensuring full visibility and fine-grained control throughout the
AI workflow. We provide automatic lineage tracking, centralized governance, and seamless
cross-workspace collaboration for simplified MLOps and enhanced productivity. Furthermore,
we give customers complete control and ownership of their data and models with privacy
controls to maintain compliance as well as efficiency and granular models on their data, fine-
tuned at lower costs.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 45
Executive
Summary Databricks Mosaic AI
Introduction Databricks provides a scalable, collaborative platform that empowers ML teams to prepare
and process data, streamline cross-team collaboration, and standardize the full ML lifecycle
Risks in AI System
Components from experimentation to production, including generative AI and large language models
(LLMs). You can both build models from scratch and tune existing models on your data
⟶ Understanding
Databricks Data
to maintain privacy and control. However, it’s not just about building and serving models.
Intelligence Platform
AI Risk Mitigation
Databricks Mosaic AI covers the end-to-end AI workflow to help you deploy and manage
Controls
models all the way through production. Our AI offerings include:
Conclusion
Resources and
Further Reading
1 | End-to-end RAG (retrieval augmented generation) to build high-quality
conversational agents on your data, leveraging the Mosaic AI Vector Search
Acknowledgments
(Public Preview) for increased relevance and accuracy.
5 | End-to-end MLOps based on the popular MLflow open source project, with
all data produced automatically actionable, tracked and monitorable in the
lakehouse.
6 | Improve visibility and proactively detect anomalies in your entire data and
AI workflow, reducing risks, time to value, and high operational costs with
Databricks Lakehouse Monitoring (Public Preview).
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 46
Executive
Summary Databricks Unity Catalog
Introduction Databricks Unity Catalog is the industry’s first unified governance solution for data and AI
on the lakehouse. With Unity Catalog, organizations can seamlessly govern their structured
Risks in AI System
Components and unstructured data, machine learning models, notebooks, dashboards, and files on
any cloud or platform. Data scientists, analysts and engineers can use Unity Catalog to
⟶ Understanding
Databricks Data
securely discover, access and collaborate on trusted data and AI assets, leveraging AI to
Intelligence Platform
AI Risk Mitigation
boost productivity and unlock the full potential of the lakehouse environment. This unified
Controls
approach to governance accelerates data and AI initiatives while ensuring regulatory
compliance in a simplified manner. Unity Catalog provides:
Conclusion
Resources and
Further Reading
1 | Access control for data and AI: Unity Catalog is the only governance
Acknowledgments solution for data and AI. The foundational capabilities of Unity Catalog are in
governance and access control of all your data and AI assets. This simplified
Appendix: governance experience works across workspaces and clouds helps you
Glossary
manage your entire data estates. Discover and classify structured and
unstructured data, ML models, notebooks, dashboards and arbitrary files on
License
any cloud. Consolidate, map and query data from various platforms, including
MySQL, PostgreSQL, Amazon Redshift, Snowflake, Azure SQL, Azure Synapse
and Google’s BigQuery in one place. Accelerate your data and AI initiatives
with a single point of access for data exploration. Boost productivity by
securely searching, understanding and extracting insights from your data and
AI using natural language.
2 | Open data sharing and collaboration: Easily share data and AI assets
across clouds, regions and platforms with open source Delta Sharing, natively
integrated within Unity Catalog. Securely collaborate with anyone, anywhere
to unlock new revenue streams and drive business value without relying on
proprietary formats, complex ETL processes or costly data replication.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 47
3 | Centralized data search and discovery: Quickly find, understand and
reference data from across your data estate, boosting productivity. Data
search in Unity Catalog is secure by default, limiting search results based on
access privileges of the users and adding an additional layer of security for
privacy considerations.
4 | A
utomated lineage for all workloads: Build better understanding of
your data estate with automated lineage, tags and auto-generated data
insights. Create a unified, transparent view of your entire data ecosystem
Executive
Summary with automated and granular lineage for all workloads in SQL, R, Python and
Scala, and across all asset types — tables, files, notebooks, workflows and
Introduction dashboards. Lineage can be retrieved via REST APIs to support integrations
with our catalog partners.
Risks in AI System
Components 5 | Security and compliance: Ability to define access policies at scale for
all data and AI assets such as files, tables, ML models, notebooks and
⟶ Understanding
Databricks Data
dashboards and to audit the access patterns.
Intelligence Platform
AI Risk Mitigation
Controls
Conclusion
Resources and
Further Reading
Acknowledgments
Appendix:
Glossary
We use the phrase “hybrid PaaS” because our lakehouse architecture is split into two
separate planes to simplify your permissions, avoid data duplication and reduce risk. The
control plane is the management plane where Databricks runs the workspace application
and manages notebooks, configuration and clusters. The compute plane handles your
data processing. Customers deploy a compute plane (virtual network and compute) in a
cloud service provider account (such as AWS, Azure or GCP) that the customer owns. With
serverless deployments, the compute plane exists in the customer’s Databricks account
rather than their cloud service provider account. Customers get the benefits of PaaS with
DATABRICKS
AI SECURITY the option to keep their data processing clusters locally within their environment.
FRAMEWORK
(DASF)
VERSION 1.0 48
The phrase “general-purpose data-agnostic” means that, unlike a pure SaaS, Databricks
doesn’t know what data your teams process with the Databricks Platform. The actual code,
business logic, model artifacts, SaaS, open source models, choice of LLMs, and datasets are
provided by your teams. You won’t find recommendations like “truncate user IDs” or “hash
feature names” because we don’t know what data you’re analyzing and what models you are
deploying.
If you’re new to Databricks or the lakehouse architecture, start with an overview of the
architecture and a review of common security questions before you hop into specific
Executive
Summary
recommendations. You’ll see those in our Security and Trust Center and the Security and
Trust Overview Whitepaper.
Introduction
Risks in AI System
Components
⟶ Understanding
Databricks Data
Intelligence Platform
AI Risk Mitigation
Controls
Conclusion
Data and AI are your most valuable assets and always have to be protected — that’s why
Acknowledgments
security is built into every layer of the Databricks Data Intelligence Platform. Databricks
Appendix: Security is based on three core principles: Trust, Technology and Transparency.
Glossary
License
Risks in AI System
Components At Databricks, we strive to continuously innovate and advance our product offerings to
simplify the ability to build AI-powered solutions on the Databricks Data Intelligence Platform
⟶ Understanding
safely. We believe there is no greater accelerant to delivering ML to production than building
Databricks Data
Intelligence Platform on a unified, data-centric AI platform. On Databricks, data and models can be managed
AI Risk Mitigation
Controls and governed in a single governance solution with Unity Catalog. With Mosaic AI Model
Serving, we streamlined the complexities associated with infrastructure for real-time model
Conclusion
deployment, providing a scalable and user-friendly solution. For long-term efficiency and
Resources and performance stability in ML production, Databricks Lakehouse Monitoring plays a pivotal role.
Further Reading
This tool ensures continuous performance monitoring, contributing to sustained excellence
in machine learning operations. These components collectively form the data pipelines of an
Acknowledgments
ML solution, all of which can be orchestrated using Databricks Workflows.
Appendix:
Glossary Perhaps the most significant recent change in the machine learning landscape has been
the rapid advancement of generative AI. Generative models such as large language models
License
(LLMs) and image generation models have revolutionized the field, unlocking previously
unattainable levels of natural language and image generation. However, their arrival also
introduces new challenges and decisions to be made in the context of MLOps.
With all these developments in mind, below is a list of the necessary mitigation controls
for organizations to address AI security risks. This mitigation guidance incorporates new
Databricks features such as Models in Unity Catalog, Model Serving, and Lakehouse
Monitoring into our MLOps architecture recommendations.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 50
D E S C R I P T I O N O F C O N T R O L I M P L E M E N TAT I O N
CONTROL/RISK
O N DATA B R I C K S P L AT F O R M
RAW DATA 1.1 DATA PREP 2.1 Implementing single sign-on with an identity
DATA PREP 2.2 DATA PREP 2.3 provider’s (IdP) multi-factor authentication is
Configuration
DATA PREP 2.4 DATASETS 3.1 critical for secure authentication. It adds an extra
EVALUATION 6.1 MODEL 7.1 layer of security, ensuring that only authorized
users access the Databricks Platform. PRODUCT REFERENCE
MODEL 7.2 MODEL MANAGEMENT 8.2
Executive
MODEL SERVING — INFERENCE REQUESTS 9.2
Summary
MODEL SERVING — INFERENCE REQUESTS 9.5
⟶ Understanding
Databricks Data
Intelligence Platform
DASF 2 Sync users and groups
AI Risk Mitigation
Controls RISKS DESCRIPTION CONTROL CATEGORY
RAW DATA 1.1 DATA PREP 2.1 Synchronizing users and groups from your
Conclusion DATA PREP 2.2 DATA PREP 2.3 identity provider (IdP) with Databricks using
Configuration
DATA PREP 2.4 DATASETS 3.1 the SCIM standard facilitates consistent and
EVALUATION 6.1 MODEL 7.2 automated user provisioning for enhancing
security. PRODUCT REFERENCE
Resources and MODEL MANAGEMENT 8.2
Further Reading MODEL MANAGEMENT 8.4
RAW DATA 1.1 DATA PREP 2.1 Configure IP access lists to restrict authentication
DATA PREP 2.2 DATA PREP 2.3 to Databricks from specific IP ranges, such as
Configuration
DATA PREP 2.4 DATASETS 3.1 VPNs or office networks, and strengthen network
EVALUATION 6.1 MODEL 7.2 security by preventing unauthorized access from
untrusted locations. PRODUCT REFERENCE
MODEL MANAGEMENT 8.2
RAW DATA 1.1 DATA PREP 2.1 Use AWS PrivateLink, Azure Private Link or GCP
DATA PREP 2.2 DATA PREP 2.3 Private Service Connect to create a private
Configuration
DATA PREP 2.4 DATASETS 3.1 network route between the customer and the
Executive EVALUATION 6.1 MODEL 7.2 Databricks control plane or the control plane and
Summary the customer’s compute plane environments to PRODUCT REFERENCE
MODEL MANAGEMENT 8.2
enhance data security by avoiding public internet
MODEL MANAGEMENT 8.4
AI Risk Mitigation
Controls
Resources and RAW DATA 1.1 RAW DATA 1.4 Implementing Unity Catalog for unified
Further Reading DATA PREP 2.1 DATASETS 3.1 permissions management and assets simplifies
Implementation
DATASETS 3.2 DATASETS 3.3 access control and enhances security.
GOVERNANCE 4.1 EVALUATION 6.1
Acknowledgments PRODUCT REFERENCE
MODEL 7.1 MODEL 7.2
RAW DATA 1.2 Tags are attributes containing keys and optional
values that you can apply to different securable
Implementation
objects in Unity Catalog. Organizing securable
objects with tags in Unity Catalog aids in
efficient data management, data discovery PRODUCT REFERENCE
RAW DATA 1.3 RAW DATA 1.9 Databricks Delta Live Tables (DLT) simplifies ETL
DATA PREP 2.1 DATASETS 3.1 development with declarative pipelines that
Implementation
GOVERNANCE 4.1 EVALUATION 6.1 integrate quality control checks and performance
Executive monitoring.
Summary PRODUCT REFERENCE
Acknowledgments RAW DATA 1.4 DATASETS 3.2 Databricks supports TLS 1.2+ encryption to
DATASETS 3.3 protect customer data during transit. This applies
Out-of-the-box
to data transfer between the customer and the
Appendix: Databricks control plane and within the compute
Glossary plane. Customers can also secure inter-cluster PRODUCT REFERENCE
RAW DATA 1.5 RAW DATA 1.7 Store data in a lakehouse architecture using Delta
tables. Delta tables can be versioned to revert
Implementation
any user’s or malicious actor’s poisoning of data.
Data can be stored in a lakehouse architecture
in the customer’s cloud account. Both raw data PRODUCT REFERENCE
RAW DATA 1.6 DATA PREP 2.1 Unity Catalog tracks and visualizes real-time data
Executive DATASETS 3.1 GOVERNANCE 4.1 lineage across all languages to the column level,
Out-of-the-box
Summary EVALUATION 6.1 providing a traceable history of an object from
notebooks, workflows, models and dashboards.
This enhances transparency and compliance, PRODUCT REFERENCE
Introduction with accessibility provided through the Catalog
Explorer. AWS | Azure | GCP
Risks in AI System
Components
DASF 12 Delete records from datasets
⟶ Understanding
RISKS DESCRIPTION CONTROL CATEGORY
Databricks Data
Intelligence Platform RAW DATA 1.8 Data governance in Delta Lake, the lakehouse
AI Risk Mitigation
storage layer, utilizes its atomicity, consistency,
Controls Implementation
isolation, durability (ACID) properties for effective
data management. This includes the capability to
remove data based on specific predicates from PRODUCT REFERENCE
Conclusion
a Delta Table, including the complete removal
of data’s history, supporting compliance with AWS | Azure | GCP
regulations like GDPR and CCPA.
Resources and
Further Reading
Appendix: RAW DATA 1.9 Use Databricks for near real-time data ingestion,
Glossary
processing, machine learning, and AI for
Implementation
streaming data.
RAW DATA 1.10 DATASETS 3.1 Databricks auditing, enhanced by Unity Catalog’s
events, delivers fine-grained visibility into data
Implementation
access and user activities. This is vital for robust
data governance and security, especially in
regulated industries. It enables organizations to PRODUCT REFERENCE
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 54
D E S C R I P T I O N O F C O N T R O L I M P L E M E N TAT I O N
CONTROL/RISK
O N DATA B R I C K S P L AT F O R M
DATA PREP 2.1 Iteratively explore, share and prep data for
the machine learning lifecycle by creating
Implementation
reproducible, editable and shareable datasets,
Executive tables and visualizations. Within Databricks this
Summary EDA process can be accelerated with Mosaic PRODUCT REFERENCE
⟶ Understanding
Databricks Data DASF 16 Secure model features
Intelligence Platform
AI Risk Mitigation
Controls RISKS DESCRIPTION CONTROL CATEGORY
DATA PREP 2.1 DATA PREP 2.2 Databricks Feature Store is a centralized
DATASETS 3.1 GOVERNANCE 4.1 repository that enables data scientists to find and
Conclusion Implementation
ALGORITHMS 5.2 share features and also ensures that the same
MODEL SERVING — INFERENCE REQUESTS 9.10 code used to compute the feature values is used
for model training and inference. Unity Catalog’s PRODUCT REFERENCE
Resources and
capabilities, such as security, lineage, table
Further Reading
history, tagging and cross-workspace access, AWS | Azure | GCP
are automatically available to the feature table to
reduce the risk of malicious actors manipulating
Acknowledgments
the features that feed into ML training.
Appendix:
Glossary
DASF 17 Track and reproduce the training data used for ML model training
DATA PREP 2.4 DATASETS 3.1 MLflow with Delta Lake tracks the training data
GOVERNANCE 4.1 ALGORITHMS 5.2 used for ML model training. It also enables the
Configuration
identification of specific ML models and runs
derived from particular datasets for regulatory
and auditable attribution. PRODUCT REFERENCE
DATABRICKS
AWS | Azure | GCP
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 55
D E S C R I P T I O N O F C O N T R O L I M P L E M E N TAT I O N
CONTROL/RISK
O N DATA B R I C K S P L AT F O R M
Introduction
DASF 20 Track ML training runs
Conclusion
DASF 21 Monitor data and AI system from a single pane of glass
License
DASF 22 Build models with all representative, accurate and relevant data sources
Introduction
DASF 24 Control access to models and model assets
Resources and
Further Reading
DASF 25 Use retrieval augmented generation (RAG) with large language models (LLMs)
Acknowledgments
RISKS DESCRIPTION CONTROL CATEGORY
MODEL SERVING — INFERENCE REQUESTS 9.8 Data is your competitive advantage. Use it
to customize large AI models to beat your
Implementation
competition. Produce new model variants with
tailored LLM response style and structure via
fine-tuning. PRODUCT REFERENCE
DATABRICKS
Fine-tune your own LLM with open models to
own your IP.
AWS | Azure | N/A
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 57
D E S C R I P T I O N O F C O N T R O L I M P L E M E N TAT I O N
CONTROL/RISK
O N DATA B R I C K S P L AT F O R M
RAW DATA 1.8 MODEL 7.3 Data is your competitive advantage. Use it
MODEL SERVING — INFERENCE REQUESTS 9.8 to customize large AI models to beat your
Implementation
competition by pretraining models with your
data, imbuing the model with domain-specific
knowledge, vocabulary and semantics. Pretrain PRODUCT REFERENCE
Risks in AI System MODEL MANAGEMENT 8.1 Model aliases in machine learning workflows
Components MODEL MANAGEMENT 8.3 allow you to assign a mutable, named reference
Implementation
MODEL SERVING — INFERENCE REQUESTS 9.5 to a specific version of a registered model.
MODEL SERVING — INFERENCE REQUESTS 9.6 This functionality is beneficial for tracking and
⟶ Understanding managing different stages of a model’s lifecycle, PRODUCT REFERENCE
MODEL SERVING — INFERENCE RESPONSE 10.3
Databricks Data indicating the current deployment status of any
Intelligence Platform
MODEL SERVING — INFERENCE RESPONSE 10.4
given model version. AWS | Azure | GCP
AI Risk Mitigation
Controls
Resources and RAW DATA 1.8 MODEL MANAGEMENT 8.1 The lakehouse forms the foundation of a data-
Further Reading MODEL MANAGEMENT 8.3 centric AI platform. Key to this is the ability to
Implementation
manage both data and AI assets from a unified
governance solution on the lakehouse. Databricks
Acknowledgments Unity Catalog enables this by providing PRODUCT REFERENCE
MODEL 7.2 MODEL MANAGEMENT 8.2 Databricks Secrets stores your credentials
and references them in notebooks, scripts,
Implementation
configuration properties and jobs.
Risks in AI System
Components
DASF 35 Track model performance
⟶ Understanding
RISKS DESCRIPTION CONTROL CATEGORY
Databricks Data
Intelligence Platform MODEL SERVING — INFERENCE RESPONSE 10.1 Databricks Lakehouse Monitoring provides
AI Risk Mitigation
performance metrics and data quality
Controls Implementation
statistics across all account tables. It tracks the
performance of machine learning models and
model serving endpoints by observing inference PRODUCT REFERENCE
Conclusion
tables with model inputs and predictions.
AWS | Azure | N/A
Resources and
Further Reading
DASF 36 Set up monitoring alerts
Acknowledgments
RISKS DESCRIPTION CONTROL CATEGORY
RAW DATA 1.3 Databricks SQL alerts can monitor the metrics
Appendix: MODEL SERVING — INFERENCE RESPONSE 10.1 table for security-based conditions, ensuring
Implementation
Glossary data integrity and timely response to potential
issues:
PRODUCT REFERENCE
License S
tatistic range Alert: Triggers when a specific
statistic, such as the fraction of missing
values, exceeds a predetermined threshold
AWS | Azure | N/A
D
ata distribution shift alert: Activates
upon shifts in data distribution, as indicated
by the drift metrics table
B
aseline divergence alert: Alerts if data
significantly diverges from a baseline,
suggesting potential needs for data analysis
or model retraining, particularly
in InferenceLog analysis
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 60
D E S C R I P T I O N O F C O N T R O L I M P L E M E N TAT I O N
CONTROL/RISK
O N DATA B R I C K S P L AT F O R M
MODEL SERVING — INFERENCE REQUESTS 9.1 Databricks inference tables automatically record
MODEL SERVING — INFERENCE REQUESTS 9.2 incoming requests and outgoing responses
Implementation
MODEL SERVING — INFERENCE REQUESTS 9.3 to model serving endpoints, storing them as
MODEL SERVING — INFERENCE REQUESTS 9.4 a Unity Catalog Delta table. This table can be
used to monitor, debug and enhance ML models. PRODUCT REFERENCE
MODEL SERVING — INFERENCE REQUESTS 9.5
By coupling inference tables with Lakehouse
MODEL SERVING — INFERENCE REQUESTS 9.6
R
etraining dataset creation: Building
datasets for the next iteration of your models
Risks in AI System
Components Q
uality monitoring: Keeping track of
production data and model performance
D
iagnostics and debugging: Investigating and
⟶ Understanding
resolving issues with suspicious inferences
Databricks Data
Intelligence Platform M
islabeled dataidentification: Compiling
AI Risk Mitigation data that needs relabeling
Controls
Executive
the option to disable this access. Additionally, AWS | Azure | GCP
staff activity within these environments is
Summary recorded in customer audit logs. Accessing these
areas requires multi-factor authentication, and
employees must connect to the Databricks VPN.
Introduction
Conclusion
C
ode peer reviews
AWS | Azure | GCP
S
tatic and dynamic scans for code and
containers, including dependencies
Resources and F
eature-level security reviews
Further Reading
A
nnual software engineering security training
C
ross-organizational collaborations between
Acknowledgments security, product management, product
security and security champions
DATA PREP 2.2 DATA PREP 2.3 MLOps enhances efficiency, scalability, security
DATA PREP 2.4 GOVERNANCE 4.2 and risk reduction in machine learning projects.
Implementation
ALGORITHMS 5.1 ALGORITHMS 5.3 Databricks integrates with MLflow, focusing on
enterprise reliability, security and scalability for
EVALUATION 6.1 MODEL 7.1
managing the machine learning lifecycle. The PRODUCT REFERENCE
MODEL 7.2 MODEL 7.3
latest update to MLflow introduces new LLMOps
MODEL MANAGEMENT 8.3
features for better management and deployment AWS | Azure | GCP
OPERATIONS 11.1 of large language models (LLMs). This includes
integrations with Hugging Face Transformers,
OpenAI and the external models in Mosaic AI
Model Serving.
DATA PREP 2.3 ALGORITHMS 5.3 Databricks access control lists (ACLs) enable
Executive MODEL 7.1 you to configure permissions for accessing and
Implementation
Summary interacting with workspace objects, including
folders, notebooks, experiments, models,
clusters, pools, jobs, Delta Live Tables pipelines, PRODUCT REFERENCE
Introduction alerts, dashboards, queries and SQL warehouses.
AWS | Azure | GCP
Risks in AI System
Components
DASF 44 Triggering actions in response to a specific event
C
I workflow trigger: Validate your model
AWS | Azure | GCP
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 63
D E S C R I P T I O N O F C O N T R O L I M P L E M E N TAT I O N
CONTROL/RISK
O N DATA B R I C K S P L AT F O R M
MODEL SERVING — INFERENCE REQUESTS 9.1 Mosaic AI Vector Search is a vector database
MODEL SERVING — INFERENCE REQUESTS 9.2 that is built into the Databricks Data Intelligence
Implementation
MODEL SERVING — INFERENCE REQUESTS 9.5 Platform and integrated with its governance
MODEL SERVING — INFERENCE REQUESTS 9.6 and productivity tools. A vector database is a
Executive database that is optimized to store and retrieve PRODUCT REFERENCE
MODEL SERVING — INFERENCE REQUESTS 9.7
Summary embeddings. Embeddings are mathematical
MODEL SERVING — INFERENCE REQUESTS 9.8
Resources and
Further Reading
DASF 47 Compare LLM outputs on set prompts
Acknowledgments
RISKS DESCRIPTION CONTROL CATEGORY
Appendix:
Glossary
AWS | Azure | GCP
DATA PREP 2.1 MODEL 7.4 Databricks’ Git Repository integration supports
effective code and third-party libraries
Out-of-the-box
management, enhancing customer control over
their development environment.
PRODUCT REFERENCE
Datasets
Resources and
Further Reading Stakeholders 1
Compliance Use case
identified
Applications
Acknowledgments
Finetuned LLMs 2
Pretrained LLMs Deploèmenï
model
identified
Rag with LLMs
External Models
Figure 2: Implementation guidance of DASF controls on the Databricks Data Intelligence Platform..
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 66
1 | Identify the AI business use case: Always keep your business goals in mind.
Make sure there is a well-defined use case with stakeholders you are trying to
secure adequately, whether already implemented or in planning phases. This
will help inform which AI system components are of greatest business value
for any given business use case
Executive
2 | Determine the AI deployment model: Choose an appropriate model (e.g.,
Summary predictive ML models, Foundation Model APIs, RAG LLMs, fine-tuned LLMs and
pretrained LLMs, as described in Section: 1.2 How to use this document) to
Introduction
determine how shared responsibilities (especially for securing each component)
Risks in AI System are split across the 12 ML/GenAI components between your organization, the
Components
Databricks Data Intelligence Platform and any partners involved.
Understanding
Databricks Data
3 | Select the most pertinent risks: From our documented list of 55 security
Intelligence Platform risks, pinpoint the ones most relevant to your organization based on the
AI Risk Mitigation
Controls outcome of step #2. Identify the specific threats linked to each risk and the
targeted ML/GenAI component for every threat.
⟶ Conclusion
4 | Choose and implement controls: Select controls that align with your
Resources and organization’s risk appetite. These controls are defined generically for
Further Reading
compatibility with any data platform. Our framework also provides guidelines on
Acknowledgments tailoring these controls specifically for the Databricks Data Intelligence Platform
with specific Databricks product references by cloud. You use these controls
Appendix: alongside your organization’s policies and have the right assurance in place.
Glossary
License
Databricks stands uniquely positioned as a secure, unified, data-centric platform for both
MLOps and LLMOps by taking a defense-in-depth approach to helping organizations
implement security across all AI system components. Red teaming and testing can help
iteratively improve and mitigate discovered weaknesses of models. As we embrace the
ongoing wave of AI advancements, it’s clear that employing a robust, secure MLOps
strategy will remain central to unlocking AI’s full potential. With firm, secure MLOps
foundations in place, organizations will be able to maximize their AI investments to drive
innovation and deliver business value.
A lot of care has been taken to make this whitepaper accurate; however, as AI is an evolving
field, please reach out to us if you have any feedback. If you’re interested in participating
in one of our AI Security workshops, please contact [email protected].
DATABRICKS
AI SECURITY If you are curious about how Databricks approaches security, please visit our
FRAMEWORK
(DASF) Security and Trust Center.
VERSION 1.0 67
05
Resources and Further Reading
We have discussed many different capabilities in this document, with documentation links
where possible. Organizations that prioritize high security can learn more than what is in this
document. Here are additional resources to dive deeper:
Executive
Summary
Conclusion
License
Review the security features in the Security and Trust Center, along with the overall
documentation about the Databricks security and compliance programs.
The Security and Trust Overview Whitepaper provides an outline of the Databricks
architecture and platform security practices.
Acknowledgments
Model scanners: HiddenLayer Model Scanner ⟶ | fickling ⟶ | ModelScan ⟶
AI Risk Database ⟶ | NB Defense ⟶
Appendix:
Glossary Model validation tools: Robust Intelligence continuous validation ⟶
Vigil LLM security scanner ⟶ | Garak automated scanning ⟶ | HiddenLayer MLDR ⟶
License
Citadel Lens ⟶
The information in this document does not constitute or imply endorsement or recommendation of any third-party organization,
product or service by Databricks. Links and references to websites and third-party materials are provided for informational
purposes only and do not represent endorsement or recommendation of such resources over others.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 69
06
Acknowledgments
This whitepaper would not be possible without the insight and guidance provided by our
reviewers and contributors at Databricks and externally. Additionally, we extend our appreciation
to the frameworks that inspired our research (MITRE, OWASP, NIST, BIML, etc.), as they have
Executive
Summary
played a pivotal role in shaping the foundation of the Databricks AI Security Framework.
DATABRICKS
Risks in AI System
Components
Understanding
Databricks Data
Intelligence Platform Matei Zaharia Fermín Serna Omar Khawaja Arun Pamulapati David Wells
Chief Technology Chief Security Vice President, Senior Staff Security Staff Security
AI Risk Mitigation
Officer and Co-Founder Officer Field CISO Field Engineer Field Engineer
Controls
Conclusion
Resources and Kelly Albano Erika Ehrli Abhi Arikapudi David Veuve Tim Lortz
Further Reading Product Marketing Senior Director Senior Director Head of Security Lead Specialist
Manager Product Marketing Security Engineering Field Engineering Solutions Architect
⟶ Acknowledgments
Appendix:
Glossary Joseph Bradley Arthur Dooner Veronica Gomes Jeffrey Hirschey Aliaksandra Nita
Principal ML Specialist Solutions Architect Senior Product Counsel Senior Technical
Product Specialist Solutions Architect Program Manager
License
NAVY FEDERAL
ROBUST INTELLIGENCE CREDIT UNION
Neil Archibald Hyrum Anderson Alie Fordyce Adam Swanda Riyaz Poonawala
Senior Staff Chief Technology Product Policy AI Security Researcher Vice President
Security Engineer Officer — Threat Intelligence Information Security
Ebrima N. Ceesay, PhD, CISSP Christopher Sestito Abigail Maines Hiep Dang
Senior Distinguished Engineer Co-founder & CEO CRO VP of Strategic
Tech Alliances
HITRUST
DATABRICKS
AI SECURITY Robert Booker Jeremy Huval
FRAMEWORK EVP Strategy Chief Innovation
(DASF) Research and Innovation Center of Excellence Officer
and Chief Strategy Officer
VERSION 1.0 70
07
Appendix: Glossary
A
Adversarial examples: Modified testing samples that induce misclassification of a machine
Executive
Summary learning model at deployment time.
AI governance: The actions to ensure stakeholder needs, conditions and options are
Introduction
evaluated to determine balanced, agreed-upon enterprise objectives; setting direction
Risks in AI System through prioritization and decision-making; and monitoring performance and compliance
Components
against agreed-upon directions and objectives. AI governance may include policies on the
Understanding nature of AI applications developed and deployed versus those limited or withheld.
Databricks Data
Intelligence Platform
AI Risk Mitigation Artificial intelligence (AI): A multidisciplinary field of computer science that aims to create
Controls
systems capable of emulating and surpassing human-level intelligence.
Conclusion
B
Resources and
Further Reading
Bug bounty program: A program that offers monetary rewards to ethical hackers for
Acknowledgments
successfully discovering and reporting a vulnerability or bug to the application’s developer.
Bug bounty programs allow companies to leverage the hacker community to improve their
⟶ Appendix: systems’ security posture over time.
Glossary
License
C
Compute plane: Where your data is processed in Databricks Platform architecture.
Concept drift: A situation where statistical properties of the target variable change and the
very concept of what you are trying to predict changes as well. For example, the definition
of what is considered a fraudulent transaction could change over time as new ways are
developed to conduct such illegal transactions. This type of change will result in concept drift.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 71
Continuous integration and continuous delivery (or continuous deployment) (CI/CD):
CI is a modern software development practice in which incremental code changes are
made frequently and reliably. CI/CD is common to software development, but it is becoming
increasingly necessary to data engineering and data science. By automating the building,
testing and deployment of code, development teams are able to deliver releases more
frequently and reliably than with the manual processes still common to data engineering and
data science teams.
Executive
Control plane: The back-end services that Databricks manages in your Databricks account.
Summary
Notebook commands and many other workspace configurations are stored in the control
plane and encrypted at rest.
Introduction
Risks in AI System
Components
D
Data classification: A crucial part of data governance that involves organizing and
Understanding
Databricks Data categorizing data based on its sensitivity, value and criticality.
Intelligence Platform
AI Risk Mitigation
Controls Data drift: The features used to train a model are selected from the input data. When
statistical properties of this input data change, it will have a downstream impact on the
Conclusion
model’s quality. For example, data changes due to seasonality, personal preference changes,
Data Intelligence Platform: A new era of data platform that employs AI models to deeply
understand the semantics of enterprise data. It builds the foundation of the data lakehouse —
a unified system to query and manage all data across the enterprise — but automatically
analyzes both the data (contents and metadata) and how it is used (queries, reports, lineage,
etc.) to add new capabilities.
Data lake: A central location that holds a large amount of data in its native, raw format.
Compared to a hierarchical data warehouse, which stores data in files or folders, a data lake
uses a flat architecture and object storage to store the data. With object storage, data is
stored with metadata tags and a unique identifier, which makes it easier to locate and retrieve
DATABRICKS
AI SECURITY data across regions and improves performance. By leveraging inexpensive object storage
FRAMEWORK
(DASF) and open formats, data lakes enable many applications to take advantage of the data.
VERSION 1.0 72
Data lakehouse: A new, open data management architecture that combines the flexibility,
cost-efficiency and scale of data lakes with the data management and ACID transactions of
data warehouses, enabling business intelligence (BI) and machine learning (ML) on all data.
Data lineage: A powerful tool that helps organizations ensure data quality and
trustworthiness by providing a better understanding of data sources and consumption. It
captures relevant metadata and events throughout the data’s lifecycle, providing an end-to-
end view of how data flows across an organization’s data estate.
Executive
Summary
Data partitioning: A partition is composed of a subset of rows in a table that share the same
value for a predefined subset of columns called the partitioning columns. Data partitioning
Introduction can speed up queries against the table as well as data manipulation.
Risks in AI System Data pipeline: A data pipeline implements the steps required to move data from source
Components
systems, transform that data based on requirements, and store the data in a target system. A
Understanding
data pipeline includes all the processes necessary to turn raw data into prepared data that
Databricks Data
Intelligence Platform
users can consume. For example, a data pipeline might prepare data so data analysts and
AI Risk Mitigation
Controls
data scientists can extract value from the data through analysis and reporting. An extract,
transform and load (ETL) workflow is a common example of a data pipeline.
Conclusion
Data poisoning: Attacks in which a part of the training data is under the control of the
Resources and adversary.
Further Reading
Data preparation (data prep): The set of preprocessing operations performed in the
Acknowledgments
early stages of a data processing pipeline, i.e., data transformations at the structural and
syntactical levels.
⟶ Appendix:
Glossary
Data privacy: Attacks against machine learning models to extract sensitive information
License about training data.
Data streaming: Data that is continuously and/or incrementally flowing from a variety of
sources to a destination to be processed and analyzed in near real-time. This unlocks a new
world of use cases around real-time ETL, real-time analytics, real-time ML and real-time
operational applications that in turn enable faster decision-making.
Databricks Delta Live Tables: A declarative framework for building reliable, maintainable and
testable data processing pipelines. You define the transformations to perform on your data
and Delta Live Tables manages task orchestration, cluster management, monitoring, data
quality and error handling.
Databricks Feature Store: A centralized repository that enables data scientists to find and
share features and also ensures that the same code used to compute the feature values is
DATABRICKS
AI SECURITY used for model training and inference.
FRAMEWORK
(DASF)
VERSION 1.0 73
Databricks Secrets: Sometimes accessing data requires that you authenticate to external
data sources through Java Database Connectivity (JDBC). Databricks Secrets stores your
credentials so you can reference them in notebooks and jobs instead of directly entering
your credentials into a notebook.
Databricks SQL: The collection of services that bring data warehousing capabilities and
performance to your existing data lakes. Databricks SQL supports open formats and
standard ANSI SQL. An in-platform SQL editor and dashboarding tools allow team members
Executive
to collaborate with other Databricks users directly in the workspace. Databricks SQL also
Summary
integrates with a variety of tools so that analysts can author queries and dashboards in their
favorite environments without adjusting to a new platform.
Introduction
Resources and Delta Lake: The optimized storage layer that provides the foundation for storing data and
Further Reading
tables in the Databricks lakehouse. Delta Lake is open source software that extends Parquet
Acknowledgments
data files with a file-based transaction log for ACID transactions and scalable metadata
handling. Delta Lake is fully compatible with Apache Spark™ APIs, and was developed for
⟶ Appendix: tight integration with Structured Streaming, allowing you to easily use a single copy of data
Glossary
for both batch and streaming operations and providing incremental processing at scale.
License
Denial of service (DoS): An attack meant to shut down access to information systems,
devices or other network resources, making them inaccessible to their intended users. DoS
attacks accomplish this by flooding the target with traffic, or sending it information that
triggers a crash. In both instances, the DoS attack deprives legitimate users (i.e., employees,
members or account holders) of the service or resource they expected due to the actions of
a malicious cyberthreat actor.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 74
E
Embeddings: Mathematical representations of the semantic content of data, typically
text or image data. Embeddings are generated by a large language model and are a key
component of many GenAI applications that depend on finding documents or images that
are similar to each other. Examples are RAG systems, recommender systems, and image and
video recognition.
Executive Exploratory data analysis (EDA): Methods for exploring datasets to summarize their
Summary
main characteristics and identify any problems with the data. Using statistical methods
and visualizations, you can learn about a dataset to determine its readiness for analysis
Introduction
and inform what techniques to apply for data preparation. EDA can also influence which
Risks in AI System algorithms you choose to apply for training ML models.
Components
Extract, transform and load (ETL): The foundational process in data engineering of
Conclusion combining data from multiple sources into a large, central repository called a data
warehouse. ETL uses a set of business rules to clean and organize raw data and prepare it for
Resources and
Further Reading storage, data analytics and machine learning (ML).
Acknowledgments
F
⟶ Appendix:
Glossary Feature engineering: The process of extracting features (characteristics, properties,
attributes) from raw data to develop machine learning models.
License
Foundation Model: A general purpose machine learning model trained on vast quantities of
data and fine-tuned for more specific language understanding and generation tasks.
G
Generative: Type of machine learning methods that learn the data distribution and can
generate new examples from distribution.
Generative AI: Also known as GenAI, this is a form of machine learning that uses large
quantities of data to train models to produce content.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 75
H
Hardened runtime: Databricks handles the actual base system image (e.g., AMI) by
leveraging Ubuntu with a hardening configuration based on CIS. As a part of the Databricks
Threat and Vulnerability Management program, we perform weekly scanning of the AMIs as
they are making their way from dev to production.
Human-in-the-loop (HITL): The process of machine learning that allows people to validate
Executive a machine learning model’s predictions as right or wrong at the time of training and inference
Summary
with intervention.
Introduction Hyperparameter: A parameter whose value is set before the machine learning process
begins. In contrast, the values of other parameters are derived via training.
Risks in AI System
Components
Understanding
I
Databricks Data
Intelligence Platform
Identity provider (IdP): A service that stores and manages digital identities. Companies use
AI Risk Mitigation
Controls
these services to allow their employees or users to connect with the resources they need.
They provide a way to manage access, adding or removing privileges, while security remains
Conclusion
tight.
Resources and
Further Reading Inference: The stage of ML in which a model is applied to a task by running data points into a
machine learning model to calculate an output such as a single numerical score. For example,
Acknowledgments
a classifier model produces the classification of a test sample.
⟶ Appendix: Inference tables: A table that automatically captures incoming requests and outgoing
Glossary
responses for a model serving endpoint and logs them as a table.
License
Insider risk: An insider is any person who has or had authorized access to or knowledge of
an organization’s resources, including personnel, facilities, information, equipment, networks
and systems. Should an individual choose to act against the organization, with their privileged
access and their extensive knowledge, they are well positioned to cause serious damage.
IP access list (IP ACL): Enables you to restrict access to your AI system based on a user’s
IP address. For example, you can configure IP access lists to allow users to connect only
through existing corporate networks with a secure perimeter. If the internal VPN network is
authorized, users who are remote or traveling can use the VPN to connect to the corporate
network. If a user attempts to connect to the AI system from an insecure network, like from a
coffee shop, access is blocked.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 76
J
Jailbreaking: An attack that employs prompt injection to specifically circumvent the safety
and moderation features placed on LLMs by their creators.
L
Label-flipping (LF) attacks: A targeted poisoning attack where the attackers poison their
Executive
training data by flipping the labels of some examples from one class (e.g., the source class)
Summary
to another (e.g., the target class).
Introduction
Lakehouse Monitoring: Databricks Lakehouse Monitoring lets you monitor the statistical
Risks in AI System properties and quality of the data in all of the tables in your account. You can also use
Components
it to track the performance of machine learning models and model serving endpoints by
monitoring inference tables that contain model inputs and predictions.
Understanding
Databricks Data
Intelligence Platform
Large language model (LLM): A model trained on massive datasets to achieve advanced
AI Risk Mitigation
Controls
language processing capabilities based on deep learning neural networks.
Conclusion LLM-as-a-judge: A scalable and explainable way to approximate human preferences, which
are otherwise very expensive to obtain. Evaluating large language model (LLM) based chat
Resources and
Further Reading assistants is challenging due to their broad capabilities and the inadequacy of existing
benchmarks in measuring human preferences. Use LLMs as judges to evaluate these models
Acknowledgments
on more open-ended questions.
⟶ Appendix: LLM hallucination: A phenomenon wherein a large language model (LLM) — often a
Glossary
generative AI chatbot or computer vision tool — perceives patterns or objects that are
License nonexistent or imperceptible to human observers, creating outputs that are nonsensical or
altogether inaccurate.
M
Machine learning (ML): A form of AI that learns from existing data and makes predictions
without being explicitly programmed.
Machine learning algorithms: Pieces of code that help people explore, analyze and find
meaning in complex datasets. Each algorithm is a finite set of unambiguous step-by-step
instructions that a machine can follow to achieve a certain goal. In a machine learning model,
the goal is to establish or discover patterns that people can use to make predictions or
categorize information.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 77
Machine learning models: Process of using mathematical models of data to help a
computer learn without direct instruction. Machine learning uses algorithms to identify
patterns within data, and those patterns are then used to create a data model that can make
predictions. For example, in natural language processing, machine learning models can parse
and correctly recognize the intent behind previously unheard sentences or combinations of
words. In image recognition, a machine learning model can be taught to recognize objects —
such as cars or dogs. A machine learning model can perform such tasks by having it “trained”
with a large dataset. During training, the machine learning algorithm is optimized to find
Executive
Summary
certain patterns or outputs from the dataset, depending on the task. The output of this
process — often a computer program with specific rules and data structures — is called a
Introduction machine learning model.
Risks in AI System Machine learning operations (MLOps): The practice of creating new machine learning (ML)
Components
models and running them through a repeatable, automated workflow that deploys them to
Understanding
production. An MLOps pipeline provides a variety of services to data science processes,
Databricks Data
Intelligence Platform
including model version control, continuous integration and continuous delivery (CI/CD),
AI Risk Mitigation
Controls
model catalogs for models in production, infrastructure management, monitoring of live
model performance, security, and governance. MLOps is a collaborative function, often
Conclusion comprising data scientists, devops engineers, security teams and IT.
Resources and Malicious libraries: Software components that were intentionally designed to cause harm
Further Reading
to computer systems or the data they process. Such packages can be distributed through
Acknowledgments
various means, including phishing emails, compromised websites or even legitimate
software repositories.
⟶ Appendix:
Glossary Metadata: Data that annotates other data and AI assets. It generally includes the
permissions that govern access to them with descriptive information, possibly including their
License
data descriptions, data about data ownership, access paths, access rights and data volatility.
MLflow Model Registry: A centralized model store, set of APIs, and UI to collaboratively
manage the full lifecycle of an MLflow model. It provides model lineage (which MLflow
experiment and run produced the model), model versioning, model aliasing, model tagging
and annotations.
Model drift: The decay of models’ predictive power as a result of the changes in real-
DATABRICKS world environments.
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 78
Model inference: The use of a trained model on new data to create a result.
Model inversion: In machine learning models, private assets like training data, features and
hyperparameters, which are typically confidential, can potentially be recovered by attackers
through a process known as model inversion. This technique involves reconstructing private
elements without direct access, compromising the model’s security.
Model Zoo: A repository or library that contains pretrained models for various machine
Understanding
Databricks Data
learning tasks. These models are trained on large datasets and are ready to be deployed or
Intelligence Platform
AI Risk Mitigation
fine-tuned for specific tasks.
Controls
Mosaic AI AutoML: Helps you automatically apply machine learning to a dataset. You
Conclusion provide the dataset and identify the prediction target, while AutoML prepares the dataset
for model training. AutoML then performs and records a set of trials that creates, tunes and
Resources and
Further Reading evaluates multiple models. After model evaluation, AutoML displays the results and provides
a Python notebook with the source code for each trial run so you can review, reproduce and
Acknowledgments
modify the code. AutoML also calculates summary statistics on your dataset and saves this
information in a notebook that you can review later.
⟶ Appendix:
Glossary
Mosaic AI Model Serving: A unified service for deploying, governing, querying and
License monitoring models fine-tuned or pre-deployed by Databricks like Llama 2, MosaicML MPT or
BGE, or from any other model provider like Azure OpenAI, AWS Bedrock, AWS SageMaker and
Anthropic. Model Serving provides a highly available and low-latency service for deploying
models. The service automatically scales up or down to meet demand changes, saving
infrastructure costs while optimizing latency performance.
Mosaic AI Vector Search: A vector database that is built into the Databricks Data
Intelligence Platform and integrated with its governance and productivity tools. A vector
database is a database that is optimized to store and retrieve embeddings. Embeddings are
mathematical representations of the semantic content of data, typically text or image data.
Embeddings are generated by a large language model and are a key component of many
GenAI applications that depend on finding documents or images that are similar to each
other. Examples are RAG systems, recommender systems, and image and video recognition.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 79
Model theft: Theft of a system’s knowledge through direct observation of its input and
output observations, akin to reverse engineering. This can lead to unauthorized access,
copying or exfiltration of proprietary models, resulting in economic losses, eroded
competitive advantage and exposure of sensitive information.
N
Notebook: A common tool in data science and machine learning for developing code and
Executive presenting results.
Summary
Introduction
O
Risks in AI System
Components
Offline system: ML systems that are trained up, “frozen,” and then operated using new data
on the frozen trained system.
Understanding
Databricks Data Online system: An ML system is said to be “online” when it continues to learn during
Intelligence Platform
AI Risk Mitigation operational use, modifying its behavior over time.
Controls
Ontology: A formally defined vocabulary for a particular domain of interest used to capture
Conclusion
knowledge about that (restricted) domain of interest. Adversaries may discover the ontology
Resources and
of a machine learning model’s output space — for example, the types of objects a model can
Further Reading
detect. The adversary may discover the ontology by repeated queries to the model, forcing
it to enumerate its output space. Or the ontology may be discovered in a configuration file or
Acknowledgments
in documentation about the model.
⟶ Appendix:
Glossary
P
License
Penetration testing (pen testing): A security exercise where a cybersecurity expert
attempts to find and exploit vulnerabilities in a computer system through a combination of
an in-house offensive security team, qualified third-party penetration testers and a year-
round public bug bounty program. The purpose of this simulated attack is to identify any
weak spots in a system’s defenses that attackers could take advantage of.
Pretrained LLM: Training an LLM from scratch using your own data for better domain
performance.
Private link: Enables private connectivity between users and their Databricks workspaces
and between clusters on the compute plane and core services on the control plane within
the Databricks workspace infrastructure.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 80
Prompt injection
Direct: A direct prompt injection occurs when a user injects text that is intended to alter
the behavior of the LLM
Indirect: When a user might modify or exfiltrate resources (e.g., documents, web pages)
that will be ingested by the GenAI model at runtime via the RAG process.
R
Executive Red teaming: NIST defines cybersecurity red teaming as “a group of people authorized
Summary
and organized to emulate a potential adversary’s attack or exploitation capabilities
Introduction against an enterprise’s security posture. The Red Team’s objective is to improve enterprise
cybersecurity by demonstrating the impacts of successful attacks and by demonstrating
Risks in AI System
Components what works for the defenders (i.e., the Blue Team) in an operational environment.” (CNSS
2015 [80]) Traditional red teaming might combine physical and cyberattack elements,
Understanding attack multiple systems, and aim to evaluate the overall security posture of an organization.
Databricks Data
Intelligence Platform Penetration testing (pen testing), in contrast, tests the security of a specific application
AI Risk Mitigation
Controls or system. In AI discourse, red teaming has come to mean something closer to pen testing,
where the model may be rapidly or continuously tested by a set of evaluators and under
Conclusion
conditions other than normal operation.
Resources and
Further Reading Reinforcement learning from human feedback (RLHF): A method of training AI models
where human feedback is used as a source of reinforcement signals. Instead of relying solely
Acknowledgments on predefined reward functions, RLHF incorporates feedback from humans to guide the
learning process.
⟶ Appendix:
Glossary
Resource control: A capability in which the attacker has control over the resources
Retrieval augmented generation (RAG): An architectural approach that can improve the
efficacy of large language model (LLM) applications by leveraging custom data. This is done
by retrieving data/documents relevant to a question or task and providing them as context
for the LLM.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 81
S
Serverless compute: An architectural design that follows infrastructure as a service (IaaS)
and platform as a service (PaaS), and which primarily requires the customer to provide
the necessary business logic for execution. Meanwhile, the service provider takes care of
infrastructure management. Compared to other platform architectures like PaaS, serverless
provides a considerably quicker path to realizing value and typically offers better cost
efficiency and performance.
Executive
Summary
Single-sign on (SSO): A user authentication tool that enables users to securely access
multiple applications and services using just one set of credentials.
Introduction
Software development lifecycle (SDLC): A structured process that enables the production
Risks in AI System
Components of high-quality, low-cost software, in the shortest possible production time. The goal of the
SDLC is to produce superior software that meets and exceeds all customer expectations
Understanding
Databricks Data and demands. The SDLC defines and outlines a detailed plan with stages, or phases, that
Intelligence Platform
AI Risk Mitigation each encompasses their own process and deliverables. Adherence to the SDLC enhances
Controls
development speed and minimizes project risks and costs associated with alternative
methods of production.
Conclusion
Source code control: A capability in which the attacker has control over the source code of
Resources and
Further Reading the machine learning algorithm.
Acknowledgments System for Cross-domain Identity Management (SCIM): An open standard designed to
manage user identity information. SCIM provides a defined schema for representing users
⟶ Appendix:
and groups, and a RESTful API to run CRUD operations on those user and group resources.
Glossary
The goal of SCIM is to securely automate the exchange of user identity data between your
License company’s cloud applications and any service providers, such as enterprise SaaS applications.
T
Train proxy: The ability of an attacker to extract training data of a generative model by
prompting the model on specific inputs.
Train proxy via replication: Adversaries may replicate a private model. By repeatedly
querying the victim’s ML Model Inference API Access, the adversary can collect the target
model’s inferences into a dataset. The inferences are used as labels for training a separate
model offline that will mimic the behavior and performance of the target model.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 82
Trojan: A malicious code/logic inserted into the code of a software or hardware system,
typically without the knowledge and consent of the organization that owns/develops the
system, and which is difficult to detect and may appear harmless, but can alter the intended
function of the system upon a signal from an attacker to cause a malicious behavior desired
by the attacker. For Trojan attacks to be effective, the trigger must be rare in the normal
operating environment so that it does not affect the normal effectiveness of the AI and raise
the suspicions of human users.
Executive
Trojan horse backdoor: In the context of adversarial machine learning, the term “backdoor”
Summary
describes a malicious module injected into the ML model that introduces some secret and
unwanted behavior. This behavior can then be triggered by specific inputs, as defined by
Introduction
the attacker.
Risks in AI System
Components
U
Understanding
Databricks Data Unity Catalog (UC): A unified governance solution for data and AI assets on the Databricks
Intelligence Platform
AI Risk Mitigation Data Intelligence Platform. It provides centralized access control, auditing, lineage and data
Controls
discovery capabilities across Databricks workspaces.
Conclusion
Resources and
V
Further Reading
Vulnerability management: An information security continuous monitoring (ISCM) process
Acknowledgments of identifying, evaluating, treating and reporting on security vulnerabilities in systems and the
software that runs on them. This, implemented alongside other security tactics, is vital for
⟶ Appendix:
Glossary organizations to prioritize possible threats and minimizing their “attack surface.”
License
W
Watering hole attacks: A form of cyberattack that targets groups of users by infecting
websites that they commonly visit to gain access to the victim’s computer and network.
Webhooks: Enable you to listen for Model Registry events so your integrations can
automatically trigger actions. You can use webhooks to automate and integrate your machine
learning pipeline with existing CI/CD tools and workflows. For example, you can trigger CI
builds when a new model version is created or notify your team members through Slack each
time a model transition to production is requested.
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF)
VERSION 1.0 83
08
License
This work is licensed under the Creative Commons Attribution-Share Alike 4.0 License.
Executive
Summary
Creative Commons
171 Second Street, Suite 300
Introduction San Francisco, California, 94105
USA
Risks in AI System
Components
Understanding
Databricks Data
Intelligence Platform
AI Risk Mitigation
Controls
Conclusion
Resources and
Further Reading
Acknowledgments
About Databricks Security & Trust Center
Databricks is the data and AI company Your data security is our priority
Appendix:
Glossary
Learn more ⟶ Learn more ⟶
⟶ License
Databricks is the data and AI company. More than 10,000 organizations worldwide —
including Comcast, Condé Nast, Grammarly and over 50% of the Fortune 500 — rely on the
Databricks Data Intelligence Platform to unify and democratize data, analytics and AI.
Databricks is headquartered in San Francisco, with offices around the globe, and was
founded by the original creators of Lakehouse, Apache Spark™, Delta Lake and MLflow.
Evaluate Databricks for yourself. Visit us at databricks.com and try Databricks free!
DATABRICKS
AI SECURITY
FRAMEWORK
(DASF) © Databricks 2024. All rights reserved. Apache, Apache Spark, Spark and the Spark logo are trademarks of the Apache Software Foundation.
VERSION 1.0 Privacy Policy | Terms of Use 84