Category Archives: Dynamics 365
Bank Integration and Reconciliation using D365 F&O
Timely vendor invoice processing and vendor payments means good supplier relationships and operational efficiency. Manual processing of vendor invoice and payments involve risks of amount error, duplicate payment. Challenges: Solution : This will allow automation for invoice processing and no scope for manual intervention for payment processing and record reconciliation. Conclusion: Apt vendor invoice management is essential in building and sustaining a company’s operational capabilities and financial balance. This translates into streamlining payment operations, avoiding expensive delays, and strengthening supplier relationships. With the rise of automation and digital solutions, managing procurement and payments has become more efficient and error-free. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Share Story :
Clear Tax GST integration with D365 F&O
In order to operate and prosper Companies need to complete several compliances. Legal compliances are crucial as non-compliance attracts financial penalties, interest charges, and additional tax assessments. For businesses, tax compliance is crucial for maintaining a good reputation and building trust with customers, suppliers, and investors Critical Issue: Manual data upload in GST portal for GST return filing. Generating E-Invoices and E-Way bill manually. Challenges: Risk of errors in manual processing. Delays in data synchronization impacting compliance. Solution : Finance clear tax integration for D365 helps to manage e-way bill, e-invoicing through integration with GSP portal for GST. It automates the following : •Generate e-invoice, e-way bill. •Fetch IRN Number, QR Code & E-way bill number. •Cancel e-invoice, e-way bill. Conclusion: Ensuring tax compliance involves understanding your tax obligations, keeping accurate records, and staying informed about changes in tax laws. In addition, by enabling automation in compliances, Companies can achieve and maintain data accuracy, scalability, and enhanced reporting and real time updates. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Share Story :
Creating and Accessing Blob Storage with Azure Data Factory: A Complete Guide
Introduction: This guide will walk you through creating and accessing Azure Blob Storage and integrating it with Azure Data Factory to automate data pipelines. From setting up a storage account and managing containers to configuring pipelines and transferring data to an Azure SQL Database, this step-by-step tutorial ensures you gain a comprehensive understanding of the process. Steps: 3. Click on + Create to initiate the creation of a new storage account. 4. Fill in the required fields like subscription, resource group, and region. Review all the settings before proceeding. 5. Create a Storage Account 6. Once the storage account is created, go to the resource by clicking on Go to Resource. 7. In the storage account, navigate to the Containers section and click + Container to create a new container for storing your files. 8. Click on the container you just created to access its contents. 9. Upload the desired JSON file into the container by clicking on Upload and selecting the file from your local system. 10. Ensure that the uploaded file is now listed in the container. 11. Go back to the Azure Portal and search for Azure Data Factory to open the ADF service. 12. From the ADF home screen, go to Author > Datasets. Click + New Dataset to create a new dataset for your Blob Storage. 13. Select the Azure Blob Storage dataset type, as you are working with data stored in Blob Storage. 14. Choose the data format that matches the file you uploaded, such as JSON, and click Continue. 15. Enter the necessary details for your dataset, including the file path and format settings. Select the appropriate Authentication type and specify the Storage account where the Blob Storage resides. Click Create to finalize the dataset creation. 16. Verify the settings and click OK to confirm the dataset configuration. 17. Navigate to the Pipelines section and click + New Pipeline to create a pipeline that will define your data flow. 18. Pipeline gets created successfully as shown below. 19. In the pipeline, select the dataset type as Azure SQL Database and click Continue to set up the SQL Database dataset. 20. Provide the necessary Linked Service details for your SQL database and click Create. 21. After configuring both the source and target datasets, and the pipeline, publish all the elements to save your work. 22. Once the pipeline is running successfully, you can verify its functionality by querying the destination database to ensure data is being transferred properly. a. Go to the SQL Database and select the relevant database. b. Select the database on which we have perform a query. c. Log in with your credentials. d. Write a simple test query to verify data has been transferred from Blob Storage to the SQL Database. Execute the query and confirm that the expected output is returned. Conclusion: Integrating Azure Blob Storage with Azure Data Factory is a powerful way to manage and automate data workflows in the cloud. This guide walks you through creating a storage account, configuring containers, uploading data, and designing a pipeline to process and transfer data to Azure SQL Database. By following these steps, you can efficiently handle large-scale data integration and ensure seamless communication between your data sources and destinations. Azure Data Factory not only simplifies the process of orchestrating data pipelines but also provides robust options for monitoring and optimizing workflows. Whether you are managing JSON files, processing transactional data, or setting up complex ETL processes, Azure’s ecosystem offers a reliable and scalable solution. Start exploring these tools today to unlock new possibilities in data-driven operations! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Share Story :
Building Custom Solutions with Low-Code Plugins: Part 1- Overview
Low-code development has revolutionized the way businesses build software applications. By providing a visual, drag-and-drop interface, low-code platforms enable developers to quickly create complex applications without writing much code. However, even with the power of low-code platforms, there may be times when you need to extend their capabilities to meet specific business requirements. This is where low-code plugins come into play. Low-code plugins are small pieces of software that can be added to a low-code platform to extend its functionality. In this blog post, we will discuss the benefits of using low-code plugins, the steps involved in creating them, and some tips for successful development. Benefits of Using Low-Code Plugins Low-code plugins offer a number of benefits for businesses, including: Steps in Creating a Low-Code Plugin The process of creating a low-code plugin typically involves the following steps: Tips for Successful Low-Code Plugin Development Here are some tips for developing successful low-code plugins: Example Use Cases Low-code plugins can be used to solve a variety of business problems. Here are some examples: Conclusion Low-code plugins offer a powerful way to extend the capabilities of low-code platforms and create custom solutions that meet specific business needs. By following the steps outlined in this blog post and incorporating the tips for successful development, you can effectively leverage low-code plugins to drive innovation and achieve your business objectives. Later we will see working of the Low-Code Plugin in Dynamics 365 CRM with an example. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Share Story :
Taking a deep dive into the physical and financial postings in Dynamics 365 F&O.
In Dynamics 365 Finance & Operations (D365F&O), the concepts of physical and financial posting are at the core of inventory and transaction management. Understanding how these two processes work and their impact on inventory valuation and ledger updates is crucial for maintaining accurate financial records and operational efficiency. The Physical and Financial posting Checkboxes are in the Item Model Group. The pathway for which is: Inventory Management>Set Up> Inventory> Item Model Group. So, what is Physical Posting? Physical Posting refers to recording the movement or status change of Inventory Items without affecting the Financial Ledger. If this option is cleared, packing slips, product receipts, and production orders that are reported as finished are not posted in the ledger, regardless of the settings in the parameter setup pages. These transactions track physical inventory levels and ensure operational accuracy. Examples of physical postings include: Physical postings are essential for operational teams to track stock levels and manage logistics effectively. However, they do not impact the financial statements until a corresponding financial posting occurs. What is Financial Posting? Financial Posting occurs when a Transaction affects the company’s General Ledger, impacting financial accounts such as Cost of Goods Sold (COGS) and Accounts Payable/Receivable. If this option is cleared, the way accounting entries are handled changes significantly to simplify the process. When a purchase order is invoice-updated, the value of the items is posted only to the item consumption account and not to the inventory receipt account. Similarly, when a sales order is invoice-updated, no entries are made in either the item consumption account or the issue account. This option is especially helpful for service items, where posting item consumption during sales order invoicing isn’t necessary. By clearing this option, the journal lines for these items do not generate any ledger postings, keeping your financial records clean and focused without unnecessary complexities. Examples of financial postings include: Financial postings ensure that all inventory transactions are accurately reflected in financial records, enabling proper accounting and compliance with regulatory standards. Key Differences Between Physical and Financial Posting Aspect Physical Posting Financial Posting Impact Tracks inventory movement/status. Updates financial accounts. Ledger Update No impact on the general ledger. Impacts general ledger accounts. Use Case Operational purposes (e.g., stock tracking). Financial reporting and accounting. Examples Product receipts, stock transfers. Invoices, COGS postings, sales revenue. Configuring Posting in D365F&O D365F&O allows businesses to control how physical and financial postings are handled using parameters and setups. Here’s how you can configure them: To encapsulate, Physical and financial postings in D365F&O are fundamental to achieving a seamless connection between operational processes and financial reporting. They ensure that inventory movements are accurately tracked and that financial records reflect real-time business activities. By configuring these setups correctly, organizations can enhance their decision-making capabilities, reduce errors, and maintain compliance with accounting standards. Moreover, understanding the nuances of these postings allows businesses to streamline operations. For example, leveraging features like item model groups or automated posting parameters ensures that teams can focus on strategic growth rather than manual corrections. This integration of operational and financial data also supports better collaboration between departments, paving the way for improved efficiency and transparency. Ultimately, D365F&O empowers businesses to not only track their inventory effectively but also align their financial records with operational realities, creating a robust framework for sustainable growth and success. That’s it for this blog. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Share Story :
Easy JavaScript Examples for Dynamics 365 CRM – Repository
Are you tired of spending hours searching for the right JavaScript functions to use in Dynamics 365 CRM? If so, you’re not alone. Developers often struggle to find commonly used functions scattered across different sources, making it frustrating to build quick solutions or bug fixing. What if you had a single repository containing all the essential JavaScript functions at your fingertips? That’s exactly what this blog offers, a one-stop resource where you’ll find everything you need, from retrieving field values to automating actions on forms. With these functions in one place, you can save time, eliminate guesswork, and focus on creating impactful solutions for your Dynamics 365 projects. As a Microsoft Certified Trainer (MCT) and Microsoft Certified Professional, I’ve spent my career deploying Dynamics 365 solutions for organizations across the globe. My hands-on experience in architecting and implementing complex solutions has given me deep insight into the challenges developers face—one of the most common being finding and applying the right JavaScript functions efficiently. Let’s explore the most commonly used JavaScript functions for quick reference and seamless development Best Practices: Always check if a field or control is null before interacting with it. Keep JavaScript functions modular and reusable. Avoid using deprecated APIs, always follow the latest Microsoft documentation. Conclusion: JavaScript is a game-changer when it comes to customizing Dynamics 365 CRM, and having a go-to repository for commonly used functions can save you significant time and effort. With these functions at your fingertips, you’ll be better equipped to build dynamic forms, automate processes, and enhance the overall user experience. And smoother operations for your business. Now that you’ve explored these essential JavaScript functions, why not take your Dynamics 365 knowledge even further? Check out this blog on error handling in Dynamics 365 plugins to strengthen your expertise in server-side customizations as well. Bookmark this repo, and let’s make development faster and easier together!
Share Story :
Step-by-Step Guide to Deploying Extensions from Sandbox to Production in Dynamics 365 F&O
Are you struggling with deploying extensions from the sandbox to production in Dynamics 365 Finance and Operations? I’m going to show you how to do it step by step for a smooth and error-free transition! In this guide, we will walk through the process of moving an extension from a sandbox environment to production in Dynamics 365 Finance and Operations (F&O). This process involves exporting the extension from the sandbox, importing it into production, and ensuring that it functions as expected in the live environment. 3. Mark as Release and Sign Off: Once the extension is confirmed to be deployed, click on “Mark as Release” and ensure that you sign off on the extension. Signing off is crucial, as failure to do so will prevent the extension from appearing in the production environment. 4. Navigate to LCS and Open the Production Environment: Go to Lifecycle Services (LCS) and access the production environment. 5. Update Environment: Click on the “Maintain” and “Update Environment” option and select the sandbox environment. Wait for a few seconds to allow the system to load all the signed-off extensions that have been deployed in the sandbox. 6. Select the Latest Extension: Your extension will appear at the top of the list as the most recent signed-off version. Click on it and schedule the deployment for the desired time to update the production environment with the extension. Validate the Deployment – After deployment, test the Production environment to ensure the new logic works as expected. – Run key processes to validate the customizations and confirm no errors. Resolve Issues if Necessary – If there are any deployment errors or issues, review the error logs and fix them before re-deploying. Following the outlined process ensures that your extension is successfully migrated from sandbox to production, minimizing deployment errors and maintaining the integrity of your customizations in Dynamics 365 Finance and Operations. By adhering to these steps, you can confidently deploy your extensions to the production environment with minimal disruptions, ensuring a smooth transition and operational efficiency. Conclusion Successfully migrating an extension from the sandbox to production in Dynamics 365 Finance and Operations is a critical step in ensuring your customizations are properly implemented in the live environment. By following the steps outlined above, you can seamlessly transition your extension while maintaining control over the deployment process.If you’re facing challenges deploying an extension in the sandbox, check out our previous blog for a step-by-step guide. Here’s the link: Step-by-Step Guide: Deploying a Package from the Asset Library to UAT – CloudFronts Happy deploying!
Share Story :
BizTalk vs. Azure Logic Apps: Choosing the Right Integration Platform
Integration platforms are critical to modern business operations, allowing different applications, data, and systems to communicate effectively. While both serve the purpose of integration, they cater to different needs and scenarios. In this blog, we’ll compare BizTalk and Azure Logic Apps, helping you choose the right platform for your business. Outline 1. Opening Section: 2. Introduction: 3. Core Content: Key Differences Between BizTalk and Azure Logic Apps: When to Choose BizTalk Server: When to Choose Azure Logic Apps: 4. Conclusion and CTA: In conclusion, BizTalk Server and Azure Logic Apps cater to different integration needs. While BizTalk excels in enterprise-grade, on-premises scenarios, Azure Logic Apps shines in cloud-native, modern workflows. Choosing the right platform depends on your organization’s integration requirements, scalability goals, and budget. CTA: If you’re still unsure which platform aligns best with your needs, our team of integration experts can help. Contact us for a detailed assessment and tailored recommendations for your business integration journey. Let’s streamline your operations and drive growth together
Share Story :
Resolving SSL/TLS Secure Channel Trust Errors in Dynamics 365 Finance and Operations
Have you ever encountered the error:“The underlying connection was closed: Could not establish trust relationship for the SSL/TLS secure channel”while deploying from Visual Studio to Finance and Operations. This error is often linked to an expired or invalid SSL certificate in your environment. This also shows up as an expired SSL Certificate warning when you open your Finance and Operations environment from the browser. Certificates are critical for securing communication channels, and an expired certificate can disrupt services and integrations.In this blog, we’ll explore the cause of the error and provide steps to resolve it. References Eugene Dmytriienko – Onpremise Certificate Rotation Said Nikjou – Rotate Secrets via LCS MS Docs – Certificate Rotation Configuration In a new cloud hosted environment, the SSL Certificate stays valid for one year by default.Post that, it expires at which point it is essential to renew the SSL Certificate. For Cloud Hosted environments, it is really simple to do via the LCS. Go to LCS and open the environment which has the expired SSL. Click on Maintain and then “Rotate Secrets” In the pop-up menu, select the change type as “Rotate SSL Certificates” After that the environment will go into servicing and the status will reflect “Rotating Secrets” This entire process should take less than 15 minutes. The documentation suggests secrets rotation should show up in the enviroment history however in my attempt it didn’t so I’m not sure if that’s reliable or not or whether that is only for Tier 2 and above environments though that doesn’t make much sense. Anyways, once this is done we can see that the SSL error has been resolved. Conclusion SSL certificates are the backbone of secure communication in Dynamics 365 Finance and Operations environments.An expired certificate can disrupt critical functionalities, but with proper certificate management, such issues can be avoided.Regularly monitor your SSL certificate validity to ensure uninterrupted operations. We hope you found this article useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfronts.com
Share Story :
How to change a Bill of material (BOM) item on a production order
In manufacturing industry, there are scenarios where there is need to change some bill of materials items due to various reasons. i.e. due to lack of availability of original material, or there may be increase in price of original material. & once production order is scheduled then it is not possible to change raw materials. As a result, Microsoft has added a new feature that allows users to switch out BOM items for other items on planned or estimated production orders. This feature is useful for production manager to keep production running as well as to business personnel to keep business running in absence of or lack of required material, provided that alternate material is available. Prerequisites for altering BOM items on production orders: – system must meet the following requirements: Also note that, one can change BOM items only for production orders that have a status of Estimated or Scheduled. 1.Now, will consider finish good L0001 for this example. Finish good number L0001, as per original BOM has raw materials M0001, M0002, M0003, M0004, M0005, M0006 & P0002. 2.Create production order for item L0001. 3.Change status to Estimate. 4.Now check estimation from Manage cost -> View calculation details -> original raw materials are considered for production order. 5.Now to change BOM item go to Production order -> Change BOM item 6.Now select raw material which needs to be replaced “From item” & “To item”. & “To item quantity” specify required quantity of new item. Or if “To item quantity” indicates how much of the new item is needed. 7.In Production order lines Tab, select required production order / Orders in which needs to replace BOM item. Click Ok. 8.Now if we check estimation from Manage cost -> View calculation details -> New raw material is considered for production order. M0007 is considered for production order instead of M0006. & consumption is same as original material as “To quantity” was kept as 0. Conclusion – By activating “Change BOM item” feature on system version 10.0.38 or later, can change BOM item on Estimated or scheduled production orders. We hope you found this article useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfronts.com