10 Essential Azure Data Factory Interview Questions

Ajith Nagarajan
3 min readOct 26, 2022

--

  1. Is Azure Data Factory ETL or ELT ?

Azure data factory is a managed cloud service that’s build for these complex hybrid extract transform load (ETL) and extract load transform(ELT) and data integration project

With Azure Data Factory, it is fast and easy to build code-free or code-centric ETL and ELT processes.

2. what is global parameters in azure data factory ?

Global parameters are constants across a data factory that can be consumed by a pipeline in any expression. They’re useful when you have multiple pipelines with identical parameter names and values

3. What is parameters in azure data factory ?

You can use parameters to pass external values into pipelines, datasets, linked services, and data flows. Once the parameter has been passed into the resource, it cannot be changed. By parameterizing resources, you can reuse them with different values each time

EX : “name”: “@pipeline().parameters.password”

4. What is the use of variable in azure data factory ?

Variables in the adf pipeline provide the functionality to temporary hold the values. They are used for similar reason like we do use variables in the programming language. Variables are available inside the pipeline and it can be set from inside the pipeline.

5. What are the advantages of Microsoft Azure?

Scalability on Demand,
Cost-efficiency,
Data Backup and Disaster Recovery,
Secure Identity & Access Management (IAM),
Data Security

6. Can i pass parameters to a pipeline run ?

Users can pass parameters into their pipelines when triggering a new run of a pipeline through the API.

7. What does lookup activity do in Azure Data Factory?

Lookup activity can retrieve a dataset from any of the data sources supported by data factory and Synapse pipelines. You can use it to dynamically determine which objects to operate on in a subsequent activity, instead of hard coding the object name. Some object examples are files and tables.

Lookup activity reads and returns the content of a configuration file or table. It also returns the result of executing a query or stored procedure. The output can be a singleton value or an array of attributes, which can be consumed in a subsequent copy, transformation, or control flow activities like ForEach activity.

8. How does Azure optimize costs?

Shut down unused resources
Set up budgets and allocate costs to teams and projects
Right-size underused resources
Clean up the unwanted datas

9. Is rerun the particular activity in azure data factory ?

Azure Data Factory now allows you to rerun activities inside your pipelines. You can rerun the entire pipeline or choose to rerun downstream from a particular activity inside your data factory pipelines. You can also view the rerun history for all your pipeline runs inside the data factory.

10. How much is the limit on the number of integration runtime?

There is no hard limit on the number of integration runtime instances you can have in a data factory. There is, however, a limit on the number of VM cores that the integration runtime can use per subscription for SSIS package execution

There you have it, then. I sincerely hope that you were able to benefit in some manner from reading the questions and answers. I will continue to post additional questions about Azure Storage Accounts, Logic Apps and the rest of the resources in Azure’s ecosystem.

Please feel free to leave a comment or your feedback if you enjoyed what you read in the meantime :)

--

--

No responses yet