I have a SQL Server instance installed on a Virtual Machine, said VM is connected to an Azure VNet. My goal is to consecutively get data out of this SQL Server and store it in an Azure Data Lake.
Is it possible to use an Azure Data Factory + Azure Integration Runtime (NOT a self-hosted integration runtime!) to do this?
If I create a new Data Factory (including Private Endpoints so the factory gets an IP address of the VNet) and use the standard Integration Runtime "AutoResolveIntegrationRuntime", it's possible to specify a SQL server name and credentials, however, the connection test fails:
After installing a self-hosted integration runtime on the VM and using this IR, it appears to work:
What exactly is the integration runtime doing that wouldn't be possible without it? In the end, couldn't the Azure Integration Runtime (hosted on Azure) just open up an ODBC connection, especially since all resources are in the same network?