
For more information, check Copy data from On-premises data store to an Azure data store using Azure Data Factory.Self-Hosted Integration Runtime should be installed on the on-premises machine where the SQL Server instance is hosted. Q7: If we need to copy data from an on-premises SQL Server instance using Data Factory, which Integration Runtime should be used and where should it be installed? Q6: When copying data from or to an Azure SQL Database using Data Factory, what is the firewall option that we should enable to allow the Data Factory to access that database?Īllow Azure services and resources to access this server firewall option. Self-Hosted Integration Runtime: used for copying data from or to an on-premises data store or networks with access controlĪzure SSIS Integration Runtime: used to run SSIS packages in the Data Factory Mention these supported types with a brief description for eachĪzure Integration Runtime: used for copying data from or to data stores accessed publicly via the internet Q5: Data Factory supports three types of Integration Runtimes. For more information, check Copy data between Azure data stores using Azure Data Factory.Integration Runtime is a secure compute infrastructure that is used by Data Factory to provide the data integration capabilities across the different network environments and make sure that these activities will be executed in the closest possible region to the data store. Q4: What is Data Factory Integration Runtime? When ingesting data from a SQL Server instance, the dataset points to the name of the table that contains the target data or the query that returns data from different tables. For example, when ingesting data from a SQL Server instance, the linked service contains the name for the SQL Server instance and the credentials used to connect to that instance.ĭataset is a reference to the data store that is described by the linked service. Linked Service is a description of the connection string that is used to connect to the data stores. Q3: What is the difference between the Dataset and Linked Service in Data Factory? Trigger: Specify when the pipeline will be executedĬontrol flow: Controls the execution flow of the pipeline activities Linked Service: A descriptive connection string for the data sources used in the pipeline activities Mapping Data Flow: A data transformation UI logicĭataset: A pointer to the data used in the pipeline activities Pipeline: The activities logical containerĪctivity: An execution step in the Data Factory pipeline that can be used for data ingestion and transformation Q2: Data Factory consists of a number of components. For more information, check Starting your journey with Microsoft Azure Data Factory.It is ETL, or ELT tool for data ingestion in most Big Data solutions. Also, ADF can be used for transforming the ingested data to meet your business requirements. Q1: Briefly describe the purpose of the ADF ServiceĪDF is used mainly to orchestrate the data copying between different relational and non-relational data sources, hosted in the cloud or locally in your datacenters. In this article, we will discuss a number of questions about the Azure Data Factory service that you may be asked when applying to an Azure Data Engineer role.
