Azure Interview Questions
Azure Interview Questions
Create a Linked Service for source data store which is SQL Server
Database
Assume that we have a cars dataset
Create a Linked Service for destination data store which is Azure
Data Lake Store
Create a dataset for Data Saving
Create the pipeline and add copy activity
Schedule the pipeline by adding a trigger
What is Blob?
Blob is a service for storing large amounts of unstructured data that
can be accessed from anywhere in the world via HTTP or HTTPS.”
Blob stands for ” Binary Large Object “. It’s designed to store large
amounts of unstructured text or binary data like virtual hard disks,
videos, images or even log files.
The data can be exposed to the public or stored privately. It scales up
or down as your needs change. We no longer manage it, we only pay
for what we use.
Control flows and scale
To support the diverse integration flows and patterns in the modern
data warehouse, Data Factory enables flexible data pipeline modeling.
This entails full control flow programming paradigms, which include
conditional execution, branching in data pipelines, and the ability to
explicitly pass parameters within and across these flows. Control flow
also encompasses transforming data through activity dispatch to
external execution engines and data flow capabilities, including data
movement at scale, via the Copy activity.
Data Factory provides freedom to model any flow style that's required
for data integration and that can be dispatched on demand or
repeatedly on a schedule. A few common flows that this model
enables are:
Control flows:
Activities can be chained together in a sequence within a pipeline.
Activities can be branched within a pipeline.
Parameters:
Parameters can be defined at the pipeline level and arguments can be
passed while you invoke the pipeline on demand or from a trigger.
Activities can consume the arguments that are passed to the pipeline.
Custom state passing:
Activity outputs, including state, can be consumed by a subsequent
activity in the pipeline.
Looping containers:
The foreach activity will iterate over a specified collection of
activities in a loop.
Trigger-based flows:
Pipelines can be triggered on demand, by wall-clock time, or in
response to driven by event grid topics
Delta flows:
Parameters can be used to define your high-water mark for delta copy
while moving dimension or reference tables from a relational store,
either on-premises or in the cloud, to load the data into the lake.