Dynamics CRM Archives - Page 11 of 15 - - Page 11

Category Archives: Dynamics CRM

Filter a Lookup Field on New Form Using PowerApps.

Introduction: In this blog, we will learn how to filter a Lookup Field on the New Form. Use Case: We have a requirement where there is a Lookup Field on the form, which should only show those values whose Status is Active. Steps: 1. This is the form of Quote Product. We want to filter the Existing Product field. 2. To filter the field, click on the Combo Box.         Items property:          Set ExistingProduct.Items =  Filter(<Data Source>,Status = <Field Name>.Active) For eg: ExistingProductQPEditForm.Items = Filter([@Products],Status = ‘Status (Products)’.Active)   Conclusion: Hope the above Blog helps you to filter a Lookup Field on the New Form.

Share Story :

Nested Filters in PowerApps.

Introduction: In this blog, we will learn how to use Nested filters in PowerApps. Use Case: We have a Gallery(GalleryQuoteLineDetail) where we need to put filter which equals Quote Product selected in another Gallery. Once we get those Quote Products, we need to filter the Gallery(GalleryQuoteLineDetail) based on ‘Transaction Type’ Field,  which is a field on Quote Product. The ‘Transaction Type’ should be equal to “Project Contract”. Solution: Steps to be followed: 1. Below is the CDS Data Source, we want to filter. 2. To filter the Gallery Based on another Gallery, Use the below Formula: Items property of Gallery:      Set: GalleryQuoteLineDetail.Items: Filter( ‘Data Source’,Condition)      For eg: GalleryQuoteLineDetail.Items : Filter(‘Quote Line Detail’,’Quote Line’.’Quote Product’ =  GalleryQuoteLine.Selected.’Quote Product’)   3. To filter the Gallery Based on “Transaction Type”, Transaction Type is of Option Set Data Type, Use the below Formula: Items property of Gallery:      Set: GalleryQuoteLineDetail.Items : Filter( ‘Quote Line Detail’,  ‘Transaction Type’ in “Project Contract” ) 4. To Combine both the Filters in a single Formula, use the first filter as ‘Data Source’ of the second filter: Following is the formula:  Filter( Filter (‘Quote Line Detail’, ‘Quote Line’.’Quote Product’ =  GalleryQuoteLine.Selected.’Quote Product’ ), ‘Transaction Type’ in “Project Contract” )   Conclusion: Hope this Blog helps you to combine multiple Filters into single Filter.

Share Story :

Schedule Serverless CRON Job to Pull data from REST APIs (Part – 1)

REST API is an Application Program Interface that uses HTTP request to GET, PUT, POST and DELETE data, it’s an architecture style approach to communicate with third party application. In order to integrate our module with third party applications or in order to pull data from third party application to our database for analysis purpose the REST APIs are useful. For the analysis purpose we can consume REST API with SSIS using third party connector, but the problem with this approach is that, it will be required On Premise server for package deployment and job schedule. The alternate approach is to use serverless CRON expression. Since, mid 2018 the serverless development methodology is ditching the traditional development. In this blog we are going to see how to consume REST API with Serverless CRON expression. Here, we are using Zoho People API for integrating HR modules data to Azure DB. With Zoho People API, you can extract employee’s data and form data in XML or JSON format to develop new applications or integrate with your existing business applications. Zoho People API is independent of programming languages which helps you to develop applications in any programming languages (reference). Implementation Authentication token generation In order to access Zoho People API, Zoho People authentication token is required. The token can be generated using Browser mode and API mode. For the API mode each request has Username or Email and Password needs to include in the POST body Another approach is to register your app with Zoho by going to zoho.com/developerconsole and Add Client ID, Once added it can be used to generate new access_token. We can check the response by use using POST request, Open any REST Client and send request by using any REST API method, here we are send request to get leave data and we are using VS Code REST Client. If the request is valid then the we can get response in following format containing the data requested. 3. As we are getting proper response, the next step is to create CRON expression (Serverless approach) to pull the data from REST API, here we are going to use App services on Azure to create Timer Trigger Function App. The CRON expression is a time-based job schedular. which has six field to define time precision in following format. {second} {minute} {hour} {day} {month} {day-of-week} Each field can have one of the following types of values: Type Example When triggered A specific value “0 5 * * * *” at hh:05:00 where hh is every hour (once an hour) All values (*) “0 * 5 * * *” at 5:mm:00 every day, where mm is every minute of the hour (60 times a day) A range (- operator) “5-7 * * * * *” at hh:mm:05, hh:mm:06, and hh:mm:07 where hh:mm is every minute of every hour (3 times a minute) A set of values (, operator) “5,8,10 * * * * *” at hh:mm:05, hh:mm:08, and hh:mm:10 where hh:mm is every minute of every hour (3 times a minute) An interval value (/ operator) “0 */5 * * * *” at hh:05:00, hh:10:00, hh:15:00, and so on through hh:55:00 where hh is every hour (12 times an hour) 4. The next step is to create function app, for development purpose we are going to use Visual Studio Community 2019/2017. In Visual Studio create a project by selecting File > New > Project Select Visual C# (here we are going to use C# for development you can choose Php, Python or F#) any of your choice. Under the Visual C# select the Azure Functions and click on Next The next step is select a new Azure Function Application, Select the Timer trigger function choose your Azure Function version (v1, v2 or v3). Select Authorization level as Anonymous as we don’t want to include API Key for the function and kept the other settings as it is. After clicking on Create button, it will create the Azure Function App Solution for us, the directory structure is as follow. The next step is to publish the function on Azure, it will deploy your function on IIS and Azure. That we will see in next part of this blog.    

Share Story :

Schedule Serverless CRON Job to Pull data from REST APIs (Part – 2)

In this blog we’ll see how t publish the and deploy the Azure function on App service. With Continuing with the previous part of this blog Right Click on the solution and click on Publish, it will ask to select the proper subscription, resource group. Select the option relevant to you. Following are the Publishing option available for us. Publish on Azure App Service Azure Virtual Machines File system Local webserver (IIS) Note:- Make sure you have sign in with proper credential and cloud explorer is connected. Once we publish the function, we can check Azure function by going to the https://portal.azure.com/ then search for the Function App. We can see our function app in read only mode, since we have deployed it from visual studio environment. The next step is to create the development logic for the data pull from the Zoho People. Make sure to add all reference libraries required for the development, we can manage the NuGet package manager by going Right clicking on the solution and select manage NuGet package. Make sure to add all required reference to your script file. using System; using Microsoft.Azure.WebJobs; using Microsoft.Extensions.Logging; using Newtonsoft.Json; using System.Data.SqlClient; using System.Net.Http; using System.Net.Http.Headers; using System.Collections.Generic; using System.Linq; As our destination is Azure DB so, we need to create the environment variable to store ADO.NET connection string in settings.json file or by clicking on Azure App Service Settings Then click on Add Setting and then add the setting name like in following diagram we are creating setting for the Azure DB connection. Click on ok and provide the value for the setting. We can get azure connection string by going to the Azure Portal search the DB name and then click on the connection string and copy the ADO.NET connection string, replace the {your_password} with your server password. After adding the environment variable, we can check this in settings.json file The Next step is to implement the code logic and debug the code on local. That we’ll see in final part of this blog.

Share Story :

Schedule Serverless CRON Job to Pull data from REST APIs (Part – 3)

In this part we’ll see how to Debug the code on local. Once we done with the configuration settings the next step is to call the REST API in our code, here we are using C# for the development. Once we done with the code logic, we can run the code on local by adding the breakpoint The Azure Function app will launch in a local Functions host, and your trigger will be available locally on http://localhost:7071/api/Function1 default listening port is 7071, we can also change the port number by going to the project setting and setting Application arguments to the project settings. host start –pause-on-error –port 7079 We can also launch multiple apps at the same time, we can do this by adding multiple breakpoint at the same time and then start the function without debugging and start another function through debugger and attached the debugger to function app by, Select Debug > Attach to process. Search for fun The greyed-out function is in running state and other is running through debugger. Select the other and click Attach and then we can open the browser and run the both function app different port gets assigned to each function. Publish the App on the Azure and make sure to add the all the parameters on the configuration by going to the Configuration on the Function App.   Note:- For every changes on app we need to build and publish the app.

Share Story :

[SOLVED] Root Web Page is not getting created in D365 Portals

Introduction In this blog, we will look into how to resolve the “Root Web Pages are not getting created in D365 Portal” issue. Whenever we create a new web page in a portal its root web page automatically gets created and this root page renders entity forms, entity list or web form on the portal. In one of our environments, root web pages were not getting created automatically. We followed the steps below to resolve this issue. Steps Navigate to Customizations -> Customize the system => Default solution will open Select SDK message processing steps Change the view to “All” Filter by primary object type begins with “web page” This would show 10 sdk message processing steps, select all and activate these steps. After this root web page will get created successfully.

Share Story :

Adding “Change Password” in Profile Navigation of D365 Portal.

Introduction This blog explains how we can add “Change Password” in the Profile navigation of the D365 Portal. Steps to be followed: Go to Weblink sets 🡪 Profile Navigation Open Profile Navigation record 3. Click on the “Links” tab. Add “+ New Web Link” 4. Add below details in the form Name: Change Password Web Link Set: Profile Navigation Publishing State: Published External URL: <portal url>/en-US/Account/Manage/ChangePassword Display Order: 2 5. Save the record. 6. Now you can see the “Change Password” option in the profile navigation of the portal. NOTE: if you don’t see the changes clear the cache of the portal and check again. – Click on “Change Password” and it will navigate you to the page below. – You can change your password from here.

Share Story :

Customize the business card scanner control

To edit the mappings of Business card scanner control, refer below steps. Select the Settings icon, and then select Advanced Settings. The Business Management page opens. On the navigation bar, select Settings, and then under Customization, select Customization. Select Customize the System. Under Components in the solution explorer, expand Entities, and then expand the Contact or Lead Select Forms. Open the form of type Quick Create. Locate and select the Business Card field on the form, and select Change Properties in the Edit In the Field Properties dialog box, select AI Builder Business Card control on the Control Select the property that you want to change the mapping for, and select the Edit icon to change the value the field maps to. When done, select OK.

Share Story :

Clear DateTime Field using MS Flows

Introduction:   We had a requirement, where using MS flows, we had to set the field named Outreach which was a datetime field as blank. For a  date time field,null expression cannot be directly set and so we could not clear the existing Date in this field When I tried to set null, it failed with an error “Minimum value for date can be 01/01/1753”.   Solution: We used the below work around. Add a step Convert Time Zone, add base time as Null. Since we are not actually converting from any time zone, we can keep source and destination time zone as same. For the record on which we want to set date as null, configure run after as below. For the date, now set converted time instead of null Conclusion: The date will be cleared and set to blank.

Share Story :

Sort on Related Column in a View

Introduction:   When we create a view, We can configure the sort but the sorting is available only for the columns which are a part of the entity on which the view is created. Eg: In the active accounts view, there is a column ” Email” which is a field from Contact Entity. If we see the columns available for sorting we cannot see Email. However, in case you want to sort on a column from the related entity we can do it using the view designer and Fetch Xml builder tools from the XRM Toolbox. Solution: On load of My active accounts, the default sort is on Account Name Open View Designer on XRM Toolbox Click on Edit Query to open Fetch XML Builder We can see the order attribute is on name Now, we can edit and add order attribute to emailaddress1 After you click on Return FetchXML, you will return back to view designer. Then click on edit XML and remove the disable sorting on the attribute. You can now save and Publish.   Conclusion: The View My active accounts will be by default sorted by email address in ascending order.

Share Story :

SEARCH BLOGS:

FOLLOW CLOUDFRONTS BLOG :


Secured By miniOrange