Dynamics 365 Archives - Page 7 of 88 - - Page 7

Category Archives: Dynamics 365

Phases of Quality Control in Business Central – 2

In this continuation of our journey through the GMP-compliant quality control module in Business Central, we’ll be diving into key components that come into play post-MRP (Material Requirements Planning/Planning worksheet) run. The MRP identifies what’s required to meet demand, but ensuring that all materials align with quality standards demands a closer look at each phase—from defining specifications to the actual receipt of goods. This blog will cover four essential steps: setting up a Specification Master for Quality, creating a Purchase Indent to formalize demand, generating a Purchase Order to confirm procurement, and finally, processing the Goods Receipt Note (GRN), where quality checks ensure that only materials meeting specified standards are accepted. Each of these steps is integral in maintaining a seamless, controlled flow of materials that meet stringent quality requirements. Through these processes, we not only streamline procurement but also build quality control measures right into the purchasing workflow. Let’s explore how Business Central’s quality control module facilitates this alignment with Good Manufacturing Practices (GMP), helping to achieve a compliant, efficient supply chain. Specification Master In the Specifications Master is a centralized repository of quality parameters. It defines acceptance criteria for raw materials, intermediates, and finished products, ensuring compliance with regulatory standards and maintaining consistent product quality throughout procurement, production, and delivery processes. Purchase Indent Purchase indent can be created from the planning worksheet where system calculates the shortages. The EOPA no. which is tagged in the sales order will also be tagged against the shortage raw material in the planning worksheet. The quantity to purchase can be adjusted on purchase indent. Once necessary information is filled in, the purchase indent can be converted to purchase order. Multiple purchase order can be created from 1 purchase indent till the indent quantity is exhausted. Purchase Order In the purchase order page, location should be where the materials will be received. Let’s call it as quarantine. The locations will be according to the type of the material. Eg. If it is a raw material the location will be RM quarantine. While generating GRN, system creates the posted purchase receipt document and the inspection datasheet. (if the item is QC enabled) Inspection datasheet- It is a page where sampling is performed. The quality control process starts from the inspection datasheet page. Conclusion:  Incorporating quality control in procurement ensures GMP compliance from the start. Business Central’s module streamlines this with defined specifications, structured purchasing, and enforced quality checks, creating a seamless and compliant supply chain. We will be continuing the quality control of purchased goods in the next blog. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Correction of Inventory Cost

Inventory valuation is important for any manufacturing and trading business. The stakeholders would be Cost Accountants, CFOs and investors. Further, Inventory cost is the major budget element. Recently, we had a client raise the issue of cost price of inventory items not getting correctly calculated because of the following factors: Steps to be followed: 2. Go to Adjustment=>Fixed Price 3. It will show inventory transactions as per the selection criteria, with their quantity and cost price. 4. Click on Fixed Price and put the approved cost price per unit for the item variant.Click on the post button. 5. There will be a posting entry in Closing & Adjustment tab. 6. Run recalculation for the item as on the cost price date. 7. Review the Inventory Aging report or Inventory transaction report. The updated price must be reflected. 8. Recalculation can be run together after updating individual item’s cost price. 9. This must be done prior to the running inventory month close.   Conclusion : This process should be an integral part of inventory valuation. Correct inventory valuation would ensure correct cost of goods sold (COGS), gross profit (GP) and cost value of asset in Balance Sheet. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Error Handling Techniques in Dynamics 365 Plugins

Posted On February 12, 2025 by Vidit Gholam Posted in Tagged in

Have You Ever Struggled with Debugging Errors in Dynamics 365 Plugins? If you’ve been working with Dynamics 365 plugins, you’ve likely encountered scenarios where your plugin failed unexpectedly. Debugging these failures can be a challenge, especially in production environments where attaching a debugger is not always an option. How do you ensure that errors are logged effectively? How do you prevent the plugin from breaking critical business processes? In this blog, I will walk you through the best error-handling techniques for Dynamics 365 plugins, ensuring that you can capture, log, and handle errors gracefully. Why Trust Me? As a Microsoft Certified Trainer and Dynamics 365 Consultant, I have extensive experience working with Dynamics 365 CRM, Power Platform, and Azure. Over the years, I have encountered and resolved numerous plugin errors in live environments. Through my blogs and speaking engagements, I have shared valuable insights on building robust and scalable solutions in Dynamics 365. This expertise allows me to provide you with practical and effective error-handling strategies that you can implement immediately. Understanding Plugin Execution and Error Scenarios Before diving into error handling techniques, let’s briefly understand the plugin execution model. Plugins in Dynamics 365 execute in the sandbox (isolated) mode or full-trust (non-isolated) mode and can be synchronous or asynchronous. Common error scenarios in plugins include: Now, let’s explore how to handle these errors effectively. 1.) Using Try-Catch Blocks for Exception Handling The simplest and most effective way to handle errors is by wrapping your plugin logic inside a try-catch block. Why This Works: 2.) Using ITracingService for Logging Dynamics 365 provides the ITracingService to log debug messages, which is particularly useful in sandboxed plugins where direct debugging is not possible. Benefits: 3.) Logging Errors to a Custom Entity For persistent logging, consider storing error details in a custom entity (e.g., Plugin Error Log). Why This Helps: 4. Using Secure Configuration for External API Calls If your plugin interacts with external APIs, store credentials in the secure configuration rather than hardcoding them. Benefits: 5. Handling Recursion and Infinite Loops Dynamics 365 allows detecting recursive plugin execution using Depth in IPluginExecutionContext. Why? Conclusion Error handling in Dynamics 365 plugins is crucial for maintaining stability and ensuring seamless business operations. By implementing try-catch blocks, using tracing services, logging errors to a custom entity, managing secure configurations, and handling recursion, you can build robust and maintainable plugins. I encourage you to apply these techniques to your plugins and explore additional monitoring tools like Application Insights for even better observability. Have you faced any plugin debugging challenges? Share your experiences in the comments below! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Bank Integration and Reconciliation using D365 F&O

Timely vendor invoice processing and vendor payments means good supplier relationships and operational efficiency. Manual processing of vendor invoice and payments involve risks of amount error, duplicate payment. Challenges:   Solution : This will allow automation for invoice processing and no scope for manual intervention for payment processing and record reconciliation. Conclusion: Apt vendor invoice management is essential in building and sustaining a company’s operational capabilities and financial balance. This translates into streamlining payment operations, avoiding expensive delays, and strengthening supplier relationships. With the rise of automation and digital solutions, managing procurement and payments has become more efficient and error-free. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com. ‍

Share Story :

Clear Tax GST integration with D365 F&O

In order to operate and prosper Companies need to complete several compliances. Legal compliances are crucial as non-compliance attracts financial penalties, interest charges, and additional tax assessments. For businesses, tax compliance is crucial for maintaining a good reputation and building trust with customers, suppliers, and investors Critical Issue:   Manual data upload in GST portal for GST return filing. Generating E-Invoices and E-Way bill manually. Challenges:  Risk of errors in manual processing. Delays in data synchronization impacting compliance. Solution : Finance clear tax integration for D365 helps to manage e-way bill, e-invoicing through integration with GSP portal for GST. It automates the following : •Generate e-invoice, e-way bill. •Fetch IRN Number, QR Code & E-way bill number. •Cancel e-invoice, e-way bill. Conclusion: Ensuring tax compliance involves understanding your tax obligations, keeping accurate records, and staying informed about changes in tax laws. In addition, by enabling automation in compliances, Companies can achieve and maintain data accuracy, scalability, and enhanced reporting and real time updates. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Creating and Accessing Blob Storage with Azure Data Factory: A Complete Guide

Introduction: This guide will walk you through creating and accessing Azure Blob Storage and integrating it with Azure Data Factory to automate data pipelines. From setting up a storage account and managing containers to configuring pipelines and transferring data to an Azure SQL Database, this step-by-step tutorial ensures you gain a comprehensive understanding of the process. Steps: 3. Click on + Create to initiate the creation of a new storage account. 4. Fill in the required fields like subscription, resource group, and region. Review all the settings before proceeding. 5. Create a Storage Account 6. Once the storage account is created, go to the resource by clicking on Go to Resource. 7. In the storage account, navigate to the Containers section and click + Container to create a new container for storing your files. 8. Click on the container you just created to access its contents. 9. Upload the desired JSON file into the container by clicking on Upload and selecting the file from your local system. 10. Ensure that the uploaded file is now listed in the container. 11. Go back to the Azure Portal and search for Azure Data Factory to open the ADF service. 12. From the ADF home screen, go to Author > Datasets. Click + New Dataset to create a new dataset for your Blob Storage. 13. Select the Azure Blob Storage dataset type, as you are working with data stored in Blob Storage. 14. Choose the data format that matches the file you uploaded, such as JSON, and click Continue. 15. Enter the necessary details for your dataset, including the file path and format settings. Select the appropriate Authentication type and specify the Storage account where the Blob Storage resides. Click Create to finalize the dataset creation. 16. Verify the settings and click OK to confirm the dataset configuration. 17. Navigate to the Pipelines section and click + New Pipeline to create a pipeline that will define your data flow. 18. Pipeline gets created successfully as shown below. 19. In the pipeline, select the dataset type as Azure SQL Database and click Continue to set up the SQL Database dataset. 20. Provide the necessary Linked Service details for your SQL database and click Create. 21. After configuring both the source and target datasets, and the pipeline, publish all the elements to save your work. 22. Once the pipeline is running successfully, you can verify its functionality by querying the destination database to ensure data is being transferred properly. a. Go to the SQL Database and select the relevant database. b. Select the database on which we have perform a query. c. Log in with your credentials. d. Write a simple test query to verify data has been transferred from Blob Storage to the SQL Database. Execute the query and confirm that the expected output is returned. Conclusion: Integrating Azure Blob Storage with Azure Data Factory is a powerful way to manage and automate data workflows in the cloud. This guide walks you through creating a storage account, configuring containers, uploading data, and designing a pipeline to process and transfer data to Azure SQL Database. By following these steps, you can efficiently handle large-scale data integration and ensure seamless communication between your data sources and destinations. Azure Data Factory not only simplifies the process of orchestrating data pipelines but also provides robust options for monitoring and optimizing workflows. Whether you are managing JSON files, processing transactional data, or setting up complex ETL processes, Azure’s ecosystem offers a reliable and scalable solution. Start exploring these tools today to unlock new possibilities in data-driven operations! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Building Custom Solutions with Low-Code Plugins: Part 1- Overview

Low-code development has revolutionized the way businesses build software applications. By providing a visual, drag-and-drop interface, low-code platforms enable developers to quickly create complex applications without writing much code. However, even with the power of low-code platforms, there may be times when you need to extend their capabilities to meet specific business requirements. This is where low-code plugins come into play. Low-code plugins are small pieces of software that can be added to a low-code platform to extend its functionality. In this blog post, we will discuss the benefits of using low-code plugins, the steps involved in creating them, and some tips for successful development. Benefits of Using Low-Code Plugins Low-code plugins offer a number of benefits for businesses, including: Steps in Creating a Low-Code Plugin The process of creating a low-code plugin typically involves the following steps: Tips for Successful Low-Code Plugin Development Here are some tips for developing successful low-code plugins: Example Use Cases Low-code plugins can be used to solve a variety of business problems. Here are some examples: Conclusion Low-code plugins offer a powerful way to extend the capabilities of low-code platforms and create custom solutions that meet specific business needs. By following the steps outlined in this blog post and incorporating the tips for successful development, you can effectively leverage low-code plugins to drive innovation and achieve your business objectives. Later we will see working of the Low-Code Plugin in Dynamics 365 CRM with an example. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Taking a deep dive into the physical and financial postings in Dynamics 365 F&O.

In Dynamics 365 Finance & Operations (D365F&O), the concepts of physical and financial posting are at the core of inventory and transaction management. Understanding how these two processes work and their impact on inventory valuation and ledger updates is crucial for maintaining accurate financial records and operational efficiency. The Physical and Financial posting Checkboxes are in the Item Model Group. The pathway for which is: Inventory Management>Set Up> Inventory> Item Model Group. So, what is Physical Posting? Physical Posting refers to recording the movement or status change of Inventory Items without affecting the Financial Ledger. If this option is cleared, packing slips, product receipts, and production orders that are reported as finished are not posted in the ledger, regardless of the settings in the parameter setup pages. These transactions track physical inventory levels and ensure operational accuracy. Examples of physical postings include: Physical postings are essential for operational teams to track stock levels and manage logistics effectively. However, they do not impact the financial statements until a corresponding financial posting occurs. What is Financial Posting? Financial Posting occurs when a Transaction affects the company’s General Ledger, impacting financial accounts such as Cost of Goods Sold (COGS) and Accounts Payable/Receivable. If this option is cleared, the way accounting entries are handled changes significantly to simplify the process. When a purchase order is invoice-updated, the value of the items is posted only to the item consumption account and not to the inventory receipt account. Similarly, when a sales order is invoice-updated, no entries are made in either the item consumption account or the issue account. This option is especially helpful for service items, where posting item consumption during sales order invoicing isn’t necessary. By clearing this option, the journal lines for these items do not generate any ledger postings, keeping your financial records clean and focused without unnecessary complexities. Examples of financial postings include: Financial postings ensure that all inventory transactions are accurately reflected in financial records, enabling proper accounting and compliance with regulatory standards. Key Differences Between Physical and Financial Posting Aspect Physical Posting Financial Posting Impact Tracks inventory movement/status. Updates financial accounts. Ledger Update No impact on the general ledger. Impacts general ledger accounts. Use Case Operational purposes (e.g., stock tracking). Financial reporting and accounting. Examples Product receipts, stock transfers. Invoices, COGS postings, sales revenue. Configuring Posting in D365F&O D365F&O allows businesses to control how physical and financial postings are handled using parameters and setups. Here’s how you can configure them: To encapsulate, Physical and financial postings in D365F&O are fundamental to achieving a seamless connection between operational processes and financial reporting. They ensure that inventory movements are accurately tracked and that financial records reflect real-time business activities. By configuring these setups correctly, organizations can enhance their decision-making capabilities, reduce errors, and maintain compliance with accounting standards. Moreover, understanding the nuances of these postings allows businesses to streamline operations. For example, leveraging features like item model groups or automated posting parameters ensures that teams can focus on strategic growth rather than manual corrections. This integration of operational and financial data also supports better collaboration between departments, paving the way for improved efficiency and transparency. Ultimately, D365F&O empowers businesses to not only track their inventory effectively but also align their financial records with operational realities, creating a robust framework for sustainable growth and success. That’s it for this blog. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.

Share Story :

Easy JavaScript Examples for Dynamics 365 CRM – Repository

Are you tired of spending hours searching for the right JavaScript functions to use in Dynamics 365 CRM? If so, you’re not alone. Developers often struggle to find commonly used functions scattered across different sources, making it frustrating to build quick solutions or bug fixing. What if you had a single repository containing all the essential JavaScript functions at your fingertips? That’s exactly what this blog offers, a one-stop resource where you’ll find everything you need, from retrieving field values to automating actions on forms. With these functions in one place, you can save time, eliminate guesswork, and focus on creating impactful solutions for your Dynamics 365 projects. As a Microsoft Certified Trainer (MCT) and Microsoft Certified Professional, I’ve spent my career deploying Dynamics 365 solutions for organizations across the globe. My hands-on experience in architecting and implementing complex solutions has given me deep insight into the challenges developers face—one of the most common being finding and applying the right JavaScript functions efficiently. Let’s explore the most commonly used JavaScript functions for quick reference and seamless development Best Practices: Always check if a field or control is null before interacting with it. Keep JavaScript functions modular and reusable. Avoid using deprecated APIs, always follow the latest Microsoft documentation. Conclusion: JavaScript is a game-changer when it comes to customizing Dynamics 365 CRM, and having a go-to repository for commonly used functions can save you significant time and effort. With these functions at your fingertips, you’ll be better equipped to build dynamic forms, automate processes, and enhance the overall user experience. And smoother operations for your business. Now that you’ve explored these essential JavaScript functions, why not take your Dynamics 365 knowledge even further? Check out this blog on error handling in Dynamics 365 plugins to strengthen your expertise in server-side customizations as well. Bookmark this repo, and let’s make development faster and easier together!

Share Story :

Step-by-Step Guide to Deploying Extensions from Sandbox to Production in Dynamics 365 F&O

Are you struggling with deploying extensions from the sandbox to production in Dynamics 365 Finance and Operations? I’m going to show you how to do it step by step for a smooth and error-free transition! In this guide, we will walk through the process of moving an extension from a sandbox environment to production in Dynamics 365 Finance and Operations (F&O). This process involves exporting the extension from the sandbox, importing it into production, and ensuring that it functions as expected in the live environment. 3. Mark as Release and Sign Off: Once the extension is confirmed to be deployed, click on “Mark as Release” and ensure that you sign off on the extension. Signing off is crucial, as failure to do so will prevent the extension from appearing in the production environment. 4. Navigate to LCS and Open the Production Environment: Go to Lifecycle Services (LCS) and access the production environment. 5. Update Environment: Click on the “Maintain” and “Update Environment” option and select the sandbox environment. Wait for a few seconds to allow the system to load all the signed-off extensions that have been deployed in the sandbox. 6. Select the Latest Extension: Your extension will appear at the top of the list as the most recent signed-off version. Click on it and schedule the deployment for the desired time to update the production environment with the extension. Validate the Deployment – After deployment, test the Production environment to ensure the new logic works as expected. – Run key processes to validate the customizations and confirm no errors. Resolve Issues if Necessary – If there are any deployment errors or issues, review the error logs and fix them before re-deploying. Following the outlined process ensures that your extension is successfully migrated from sandbox to production, minimizing deployment errors and maintaining the integrity of your customizations in Dynamics 365 Finance and Operations. By adhering to these steps, you can confidently deploy your extensions to the production environment with minimal disruptions, ensuring a smooth transition and operational efficiency. Conclusion Successfully migrating an extension from the sandbox to production in Dynamics 365 Finance and Operations is a critical step in ensuring your customizations are properly implemented in the live environment. By following the steps outlined above, you can seamlessly transition your extension while maintaining control over the deployment process.If you’re facing challenges deploying an extension in the sandbox, check out our previous blog for a step-by-step guide. Here’s the link: Step-by-Step Guide: Deploying a Package from the Asset Library to UAT – CloudFronts Happy deploying!

Share Story :

SEARCH BLOGS:

FOLLOW CLOUDFRONTS BLOG :


Secured By miniOrange