Latest Microsoft Dynamics 365 Blogs | CloudFronts - Page 2

Integration of CSV file from ADLS to Azure SQL DB using ADF

Go through the blog to understand how to integrate a CSV file stored in your ADLS to Azure SQL DB.

Share Story :

Azure Logic App for creating an API for SQL DB.

Read this blog to understand how one can create logic App using Azure portal to create an API for SQL DB.

Share Story :

Triggering A Logic App

Read this blog to understand about Triggers in Logic App. Learn how you can work with Simple Timer Invoke and how to trigger Logic App containing a start Block as ‘When an HTTP request is received’

Share Story :

Modern Enterprise BI: Part 1

Power BI has some new features and Future Promises for Modern Enterprise applications in Business.

Share Story :

How to Pin Entire Report Page to Dashboard in Power BI

Dashboard is created to get a brief overview of your report by pinning visuals to an Existing dashboard or to a new dashboard. But sometimes it might be required to pin all the visuals of your page in the report to your dashboard. This blog will guide you through how this can be achieved.

Share Story :

How to publish your Power BI report to CRM Dashboard

Instead of viewing dashboard and report on Power BI Web Service, we can directly view it in CRM. We need to publish the Power BI dashboard to our CRM Environment. This Blog will guide you through it can be done.

Share Story :

How to Install and Locate new Plugin in XRM

This Blog will show you how you can Install and locate your Plugins in XRM new interface.

Share Story :

How to Connect your Azure Data Factory with Github

Introduction: In order to move your Pipeline from one environment to other you require your Pipeline to be saved on Github. This Blog will guide you through how to connect your Azure Data Factory Pipeline with your Github account. Step 1: Create a new Azure Data Factory and Tick the Enable GIT checkbox.     Step 2: Create a new Repository in your Github Account.     Step 3: Copy the URL from of the newly created repository from the address bar.     Step 4: Enter the GIT URL, the repository name which you recently created, branch name (usually master) and your root folder (usually same as your repository name) and click on Create.     Step 5: Now you must authorize your GIT Account on ADF. You will be promoted to Login with your Github Account. Do the necessary and click on Authorize Azure Data Factory.     Step 6: Now you must select the Working Branch. We are using the default master branch. Click Save and then you can proceed to create your Pipeline.     Conclusion: Each Time you publish the changes in ADF they are simultaneously stored in the GIT repository’s branch.

Share Story :

How to Upsert Records in SQL(Sink) through ADF?

Introduction We are performing Integration of Accounts from CRM to SQL using ADF Copy activity pipeline. We want to upsert the accounts instead of inserting duplicate records again. Step 1: Auto create the Table named “accounts” in SQL Server during the first Integration run by selecting the Auto create table option.   Step 2: Create a custom data type named “AccountType” using following query.   CREATE TYPE AccountType AS TABLE( accountid uniqueidentifier, transactioncurrencyid uniqueidentifier, address1_city nvarchar(MAX), createdon datetime2(7), accountnumber nvarchar(MAX), name nvarchar(MAX), address1_country nvarchar(MAX), address1_composite nvarchar(MAX), telephone1 nvarchar(MAX), emailaddress1 nvarchar(MAX), websiteurl nvarchar(MAX), primarycontactid uniqueidentifier ) Step 3: Create a Stored Procedure named “spUpsertAccounts”.   CREATE PROCEDURE spUpsertAccounts @account AccountType READONLY AS BEGIN   MERGE dbo.accounts AS target_sqldb USING @account AS source_tblstg ON (target_sqldb.accountid = source_tblstg.accountid) WHEN MATCHED THEN UPDATE SET accountid = source_tblstg.accountid, transactioncurrencyid = source_tblstg.transactioncurrencyid, address1_city = source_tblstg.address1_city, createdon = source_tblstg.createdon, accountnumber = source_tblstg.accountnumber, name = source_tblstg.name, address1_country = source_tblstg.address1_country, address1_composite = source_tblstg.address1_composite, telephone1 = source_tblstg.telephone1, emailaddress1 = source_tblstg.emailaddress1, websiteurl = source_tblstg.websiteurl, primarycontactid = source_tblstg.primarycontactid     WHEN NOT MATCHED THEN INSERT ( accountid, transactioncurrencyid, address1_city, createdon, accountnumber, name, address1_country, address1_composite, telephone1, emailaddress1, websiteurl, primarycontactid ) VALUES ( source_tblstg.accountid, source_tblstg.transactioncurrencyid, source_tblstg.address1_city, source_tblstg.createdon, source_tblstg.accountnumber, source_tblstg.name, source_tblstg.address1_country, source_tblstg.address1_composite, source_tblstg.telephone1, source_tblstg.emailaddress1, source_tblstg.websiteurl, source_tblstg.primarycontactid ); END Step 4: Enter the Stored Procedure Name, Table Type and Table type parameter as shown in the image below:     Step 5: Publish all the changes and debug your Pipeline.  You can Verify the results in SQL Server “accounts” table.  

Share Story :

SEARCH BLOGS:

FOLLOW CLOUDFRONTS BLOG :


Secured By miniOrange