Tag Archives: Dynamics 365
Phases of Quality Control in Business Central – 3
Welcome back to our series on navigating the GMP-compliant quality control module in Business Central! In our previous blog, we took you through the process up to the Goods Receipt Note (GRN), laying the foundation for efficient and compliant quality management. In this blog, we’ll dive deeper into an equally important aspect of the process: the quality control of raw materials and packing materials. Ensuring that your raw materials meet the necessary standards is crucial for maintaining product integrity and compliance with regulatory requirements. Let’s explore how Business Central helps streamline this critical step in the manufacturing process. Previously, we discussed the process of posting a purchase order in Business Central, which triggers several behind-the-scenes actions. When the purchase order is posted, the system generates a posted purchase receipt along with an inspection datasheet document. This seamless integration ensures that both the material tracking and quality control processes are aligned. In the background, the system also handles the transfer of items between locations. For instance, the raw material (RM) item gets posted to the location specified in the purchase order. If the item is Quality Control (QC) enabled, an inspection datasheet is automatically created upon posting. The system then transfers the item from the purchase order location to an “undertest” location, where the quality control checks are carried out before the materials are accepted into stock. Inspection Datasheet When a Goods Receipt Note (GRN) is created, an inspection datasheet is automatically generated. This datasheet pulls details from the posted purchase order, such as product information, quantities, and other relevant data. The document type for this datasheet is classified as “Purchase” to indicate that it pertains to a purchased item from a vendor. Users have the ability to edit the sample quantity on the inspection datasheet. This allows for flexibility in determining how much of the received goods will be inspected or tested When an inspection datasheet is generated from the Goods Receipt Note (GRN), the Specification ID specified on the Purchase Order (PO) for each item is automatically transferred to the datasheet. The Specification ID links to a detailed set of standards or criteria that are predefined for the item (e.g., testing methods, acceptable ranges for quality attributes). The user performs the required testing on the received goods, and after testing, the user records the test results (e.g., pass/fail, measured values) in the specification table on the inspection datasheet. After all the data is filled and verified, the user posts the inspection datasheet. Posting the datasheet signifies that the inspection process is complete, and the items are ready for further processing or acceptance. Once posted, the system creates a final, official version of the inspection datasheet, capturing all test results and any other relevant data entered during the QC process. Along with the posted datasheet, the system generates an Inspection Receipt. This receipt serves as confirmation that the goods have passed or failed inspection, and it also indicate the status (e.g., approved or rejected) Inspection Receipt On the Inspection Receipt page, the user will review the test results and specifications from the inspection datasheet. a) Based on these results, the user decides whether to accept or reject the lot. Accept: If the results meet the required specifications. Reject: If the results fail to meet the specifications. b) After making the acceptance or rejection decision, the user will enter the location and bin information for the lot to be transferred. Undertest Location: Initially, the lot is in a holding or undertest location. Accepted Lot: If the lot is accepted, the user will move it to an appropriate approved location (e.g., RM-approve for raw materials). Rejected Lot: If the lot is rejected, the user will move it to a rejected location (e.g., RM-reject for raw materials). The bins will vary based on whether the lot is accepted or rejected and its type .Once posted, the system creates a final, posted inspection receipt. This document becomes part of the system’s records, confirming the final status of the lot. The lot is moved to its designated location (approved or rejected), and inventory records are updated accordingly. A transfer entry will be created in the Item Ledger to reflect that the material has been moved to an approved/Rejected location (e.g., RM-approve/reject). c) Posted inspection receipt On the posted inspection receipt page, the user can initiate the generation of the COA report. The Certificate of analysis(COA) report contains detailed test results, pass/fail statuses, specifications, and approval information, providing a formal certificate of compliance. Conclusion: In this blog, I’ve highlighted how a streamlined Quality Control (QC) process ensures that only materials meeting your standards are accepted into inventory. From automated inspection datasheets to real-time inventory updates and generating Certificates of Analysis (COA), you can be confident in the quality and compliance of every batch. Why It Matters for Your Business: a) Ensure Consistent Quality: Only accept materials that meet your standards. b) Save Time: Automation reduces manual work and errors. c) Stay Compliant: Easy access to COAs for audits and regulatory checks. d) Build Trust: Your customers will appreciate your commitment to quality. Ready to optimize your QC process and improve efficiency? We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Phases of Quality Control in Business Central – 2
In this continuation of our journey through the GMP-compliant quality control module in Business Central, we’ll be diving into key components that come into play post-MRP (Material Requirements Planning/Planning worksheet) run. The MRP identifies what’s required to meet demand, but ensuring that all materials align with quality standards demands a closer look at each phase—from defining specifications to the actual receipt of goods. This blog will cover four essential steps: setting up a Specification Master for Quality, creating a Purchase Indent to formalize demand, generating a Purchase Order to confirm procurement, and finally, processing the Goods Receipt Note (GRN), where quality checks ensure that only materials meeting specified standards are accepted. Each of these steps is integral in maintaining a seamless, controlled flow of materials that meet stringent quality requirements. Through these processes, we not only streamline procurement but also build quality control measures right into the purchasing workflow. Let’s explore how Business Central’s quality control module facilitates this alignment with Good Manufacturing Practices (GMP), helping to achieve a compliant, efficient supply chain. Specification Master In the Specifications Master is a centralized repository of quality parameters. It defines acceptance criteria for raw materials, intermediates, and finished products, ensuring compliance with regulatory standards and maintaining consistent product quality throughout procurement, production, and delivery processes. Purchase Indent Purchase indent can be created from the planning worksheet where system calculates the shortages. The EOPA no. which is tagged in the sales order will also be tagged against the shortage raw material in the planning worksheet. The quantity to purchase can be adjusted on purchase indent. Once necessary information is filled in, the purchase indent can be converted to purchase order. Multiple purchase order can be created from 1 purchase indent till the indent quantity is exhausted. Purchase Order In the purchase order page, location should be where the materials will be received. Let’s call it as quarantine. The locations will be according to the type of the material. Eg. If it is a raw material the location will be RM quarantine. While generating GRN, system creates the posted purchase receipt document and the inspection datasheet. (if the item is QC enabled) Inspection datasheet- It is a page where sampling is performed. The quality control process starts from the inspection datasheet page. Conclusion: Incorporating quality control in procurement ensures GMP compliance from the start. Business Central’s module streamlines this with defined specifications, structured purchasing, and enforced quality checks, creating a seamless and compliant supply chain. We will be continuing the quality control of purchased goods in the next blog. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
How to configure Device License in Business Central
A Device License in Business Central is a type of user license that provides access to the application on a specific device, rather than a user-based license. This is particularly useful for scenarios where a shared device, such as a point of sale (POS) terminal, warehouse scanner, or shared workstation, needs to access Business Central without requiring individual user licenses. Why Device Licenses are Important Device licenses are particularly useful for organizations that have multiple employees using the same device at different times. Examples of such use cases include: Steps to achieve the goal
Error Handling Techniques in Dynamics 365 Plugins
Have You Ever Struggled with Debugging Errors in Dynamics 365 Plugins? If you’ve been working with Dynamics 365 plugins, you’ve likely encountered scenarios where your plugin failed unexpectedly. Debugging these failures can be a challenge, especially in production environments where attaching a debugger is not always an option. How do you ensure that errors are logged effectively? How do you prevent the plugin from breaking critical business processes? In this blog, I will walk you through the best error-handling techniques for Dynamics 365 plugins, ensuring that you can capture, log, and handle errors gracefully. Why Trust Me? As a Microsoft Certified Trainer and Dynamics 365 Consultant, I have extensive experience working with Dynamics 365 CRM, Power Platform, and Azure. Over the years, I have encountered and resolved numerous plugin errors in live environments. Through my blogs and speaking engagements, I have shared valuable insights on building robust and scalable solutions in Dynamics 365. This expertise allows me to provide you with practical and effective error-handling strategies that you can implement immediately. Understanding Plugin Execution and Error Scenarios Before diving into error handling techniques, let’s briefly understand the plugin execution model. Plugins in Dynamics 365 execute in the sandbox (isolated) mode or full-trust (non-isolated) mode and can be synchronous or asynchronous. Common error scenarios in plugins include: Now, let’s explore how to handle these errors effectively. 1.) Using Try-Catch Blocks for Exception Handling The simplest and most effective way to handle errors is by wrapping your plugin logic inside a try-catch block. Why This Works: 2.) Using ITracingService for Logging Dynamics 365 provides the ITracingService to log debug messages, which is particularly useful in sandboxed plugins where direct debugging is not possible. Benefits: 3.) Logging Errors to a Custom Entity For persistent logging, consider storing error details in a custom entity (e.g., Plugin Error Log). Why This Helps: 4. Using Secure Configuration for External API Calls If your plugin interacts with external APIs, store credentials in the secure configuration rather than hardcoding them. Benefits: 5. Handling Recursion and Infinite Loops Dynamics 365 allows detecting recursive plugin execution using Depth in IPluginExecutionContext. Why? Conclusion Error handling in Dynamics 365 plugins is crucial for maintaining stability and ensuring seamless business operations. By implementing try-catch blocks, using tracing services, logging errors to a custom entity, managing secure configurations, and handling recursion, you can build robust and maintainable plugins. I encourage you to apply these techniques to your plugins and explore additional monitoring tools like Application Insights for even better observability. Have you faced any plugin debugging challenges? Share your experiences in the comments below! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Creating and Accessing Blob Storage with Azure Data Factory: A Complete Guide
Introduction: This guide will walk you through creating and accessing Azure Blob Storage and integrating it with Azure Data Factory to automate data pipelines. From setting up a storage account and managing containers to configuring pipelines and transferring data to an Azure SQL Database, this step-by-step tutorial ensures you gain a comprehensive understanding of the process. Steps: 3. Click on + Create to initiate the creation of a new storage account. 4. Fill in the required fields like subscription, resource group, and region. Review all the settings before proceeding. 5. Create a Storage Account 6. Once the storage account is created, go to the resource by clicking on Go to Resource. 7. In the storage account, navigate to the Containers section and click + Container to create a new container for storing your files. 8. Click on the container you just created to access its contents. 9. Upload the desired JSON file into the container by clicking on Upload and selecting the file from your local system. 10. Ensure that the uploaded file is now listed in the container. 11. Go back to the Azure Portal and search for Azure Data Factory to open the ADF service. 12. From the ADF home screen, go to Author > Datasets. Click + New Dataset to create a new dataset for your Blob Storage. 13. Select the Azure Blob Storage dataset type, as you are working with data stored in Blob Storage. 14. Choose the data format that matches the file you uploaded, such as JSON, and click Continue. 15. Enter the necessary details for your dataset, including the file path and format settings. Select the appropriate Authentication type and specify the Storage account where the Blob Storage resides. Click Create to finalize the dataset creation. 16. Verify the settings and click OK to confirm the dataset configuration. 17. Navigate to the Pipelines section and click + New Pipeline to create a pipeline that will define your data flow. 18. Pipeline gets created successfully as shown below. 19. In the pipeline, select the dataset type as Azure SQL Database and click Continue to set up the SQL Database dataset. 20. Provide the necessary Linked Service details for your SQL database and click Create. 21. After configuring both the source and target datasets, and the pipeline, publish all the elements to save your work. 22. Once the pipeline is running successfully, you can verify its functionality by querying the destination database to ensure data is being transferred properly. a. Go to the SQL Database and select the relevant database. b. Select the database on which we have perform a query. c. Log in with your credentials. d. Write a simple test query to verify data has been transferred from Blob Storage to the SQL Database. Execute the query and confirm that the expected output is returned. Conclusion: Integrating Azure Blob Storage with Azure Data Factory is a powerful way to manage and automate data workflows in the cloud. This guide walks you through creating a storage account, configuring containers, uploading data, and designing a pipeline to process and transfer data to Azure SQL Database. By following these steps, you can efficiently handle large-scale data integration and ensure seamless communication between your data sources and destinations. Azure Data Factory not only simplifies the process of orchestrating data pipelines but also provides robust options for monitoring and optimizing workflows. Whether you are managing JSON files, processing transactional data, or setting up complex ETL processes, Azure’s ecosystem offers a reliable and scalable solution. Start exploring these tools today to unlock new possibilities in data-driven operations! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Building Better Forms: Mastering Form Components in Dynamics 365
In today’s ever-evolving app development landscape, delivering an exceptional user experience is critical. Power Apps offers various tools to help developers create intuitive and efficient applications, and one of the standout features is the Form Component. This feature simplifies the design and usability of forms, making applications more scalable and maintainable. What Are Form Components? Form Components in Power Apps are modular elements that can be created once and reused across multiple forms or applications. By utilizing these components, developers can maintain consistency in design, functionality, and behavior. Essentially, they act as reusable building blocks for forms, streamlining the development process and enhancing the user experience. A common use case for Form Components is displaying entity-specific forms, such as a Quote Lookup field. Let’s explore how to implement a Form Component for this scenario. Implementing a Form Component for the Quote Lookup Field Imagine you have a requirement to display the form of a specific entity, such as a Quote, using the Quote Lookup field. Follow these steps to set it up: After selecting the form in the Component, the Lookup field will display like this Save and Publish: After adding the Form Component, click ‘Save’ and then ‘Publish’ to apply your changes. Key Considerations Once the setup is complete, your Quote Lookup field will display the desired form seamlessly. Here’s how it will look: With these steps, you can enhance the functionality of your forms and deliver a better user experience in your Dynamics 365 applications. Happy developing! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Taking a deep dive into the physical and financial postings in Dynamics 365 F&O.
In Dynamics 365 Finance & Operations (D365F&O), the concepts of physical and financial posting are at the core of inventory and transaction management. Understanding how these two processes work and their impact on inventory valuation and ledger updates is crucial for maintaining accurate financial records and operational efficiency. The Physical and Financial posting Checkboxes are in the Item Model Group. The pathway for which is: Inventory Management>Set Up> Inventory> Item Model Group. So, what is Physical Posting? Physical Posting refers to recording the movement or status change of Inventory Items without affecting the Financial Ledger. If this option is cleared, packing slips, product receipts, and production orders that are reported as finished are not posted in the ledger, regardless of the settings in the parameter setup pages. These transactions track physical inventory levels and ensure operational accuracy. Examples of physical postings include: Physical postings are essential for operational teams to track stock levels and manage logistics effectively. However, they do not impact the financial statements until a corresponding financial posting occurs. What is Financial Posting? Financial Posting occurs when a Transaction affects the company’s General Ledger, impacting financial accounts such as Cost of Goods Sold (COGS) and Accounts Payable/Receivable. If this option is cleared, the way accounting entries are handled changes significantly to simplify the process. When a purchase order is invoice-updated, the value of the items is posted only to the item consumption account and not to the inventory receipt account. Similarly, when a sales order is invoice-updated, no entries are made in either the item consumption account or the issue account. This option is especially helpful for service items, where posting item consumption during sales order invoicing isn’t necessary. By clearing this option, the journal lines for these items do not generate any ledger postings, keeping your financial records clean and focused without unnecessary complexities. Examples of financial postings include: Financial postings ensure that all inventory transactions are accurately reflected in financial records, enabling proper accounting and compliance with regulatory standards. Key Differences Between Physical and Financial Posting Aspect Physical Posting Financial Posting Impact Tracks inventory movement/status. Updates financial accounts. Ledger Update No impact on the general ledger. Impacts general ledger accounts. Use Case Operational purposes (e.g., stock tracking). Financial reporting and accounting. Examples Product receipts, stock transfers. Invoices, COGS postings, sales revenue. Configuring Posting in D365F&O D365F&O allows businesses to control how physical and financial postings are handled using parameters and setups. Here’s how you can configure them: To encapsulate, Physical and financial postings in D365F&O are fundamental to achieving a seamless connection between operational processes and financial reporting. They ensure that inventory movements are accurately tracked and that financial records reflect real-time business activities. By configuring these setups correctly, organizations can enhance their decision-making capabilities, reduce errors, and maintain compliance with accounting standards. Moreover, understanding the nuances of these postings allows businesses to streamline operations. For example, leveraging features like item model groups or automated posting parameters ensures that teams can focus on strategic growth rather than manual corrections. This integration of operational and financial data also supports better collaboration between departments, paving the way for improved efficiency and transparency. Ultimately, D365F&O empowers businesses to not only track their inventory effectively but also align their financial records with operational realities, creating a robust framework for sustainable growth and success. That’s it for this blog. We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfonts.com.
Easy JavaScript Examples for Dynamics 365 CRM – Repository
Are you tired of spending hours searching for the right JavaScript functions to use in Dynamics 365 CRM? If so, you’re not alone. Developers often struggle to find commonly used functions scattered across different sources, making it frustrating to build quick solutions or bug fixing. What if you had a single repository containing all the essential JavaScript functions at your fingertips? That’s exactly what this blog offers, a one-stop resource where you’ll find everything you need, from retrieving field values to automating actions on forms. With these functions in one place, you can save time, eliminate guesswork, and focus on creating impactful solutions for your Dynamics 365 projects. As a Microsoft Certified Trainer (MCT) and Microsoft Certified Professional, I’ve spent my career deploying Dynamics 365 solutions for organizations across the globe. My hands-on experience in architecting and implementing complex solutions has given me deep insight into the challenges developers face—one of the most common being finding and applying the right JavaScript functions efficiently. Let’s explore the most commonly used JavaScript functions for quick reference and seamless development Best Practices: Always check if a field or control is null before interacting with it. Keep JavaScript functions modular and reusable. Avoid using deprecated APIs, always follow the latest Microsoft documentation. Conclusion: JavaScript is a game-changer when it comes to customizing Dynamics 365 CRM, and having a go-to repository for commonly used functions can save you significant time and effort. With these functions at your fingertips, you’ll be better equipped to build dynamic forms, automate processes, and enhance the overall user experience. And smoother operations for your business. Now that you’ve explored these essential JavaScript functions, why not take your Dynamics 365 knowledge even further? Check out this blog on error handling in Dynamics 365 plugins to strengthen your expertise in server-side customizations as well. Bookmark this repo, and let’s make development faster and easier together!
Add Tooltip for Column Headers in Power BI: A Step-by-Step Guide
Introduction Tooltips are a powerful feature in Power BI, offering additional context and information for visuals. However, tooltips are not natively supported for column headers in Power BI. This means users cannot view detailed information about column headers directly. Fortunately, there’s a workaround to address this limitation, which we’ll explore in this blog post. The method we are going to use is the Action buttons. We will add an action button to column headers, and the action button will have a tooltip that shows header details. Please find a step-by-step guide for the same – 1. First, we need to insert a button. Here I have used a blank button since I want to show the tooltip when the user hovers through a particular column area, but you can use any of the buttons based on your choice. 2. Now, go to the button visual settings, turn on the action button, set the type to bookmark with None as an option, and specify the tooltip. Conclusion In this blog post, we discussed adding tooltips to the column headers to give users more context. we hope you found this blog post helpful! If you have any questions or want to discuss further, please contact us at transform@cloudfronts.com.
Seamless Integration between D365 Project Operations & Business Central
Are you an organization who has sophisticated Project needs, however your financial requirements are simple? You may be in a position where you are having Project Operations and Business Central deployed or you are considering going with Project Operations for your sophisticated Project needs, and Business Central because your finances are simple, and you feel D365 Finance & Operations would be an over kill for your requirements. Once you get into this position, you realise that Project operations and Business Central are two disintegrated systems. Microsoft does not provide this integration out of the box. While it does for Finance & Operations, it has left PO and BC integration for the partners to figure out. Why a Single Source of Truth Matters For businesses managing complex projects, real-time, accurate insights into both project progress and financials aren’t a luxury—they’re essential. Without an integrated approach, processes slow down, errors multiply, and operational agility takes a hit. Our Perspective: By integrating D365 Project Operations with Business Central, companies can seamlessly connect project and financial data. This integration reduces errors, unifies workflows, and enables decision-makers to act with confidence. Why Integrated Systems Are Essential As projects scale, cost tracking, billing, and resource management become harder to manage. An integration of D365 Project Operations and Business Central creates a cohesive environment where data flows naturally, helping teams move past the blockers of siloed systems. Here’s How Integration Changes the Game – Unified, Real-Time Data Visibility Imagine having access to all of your financial and project data in one location. Team leads and finance managers can save time, no longer needing to cross-check numbers across systems. Data flows seamlessly between Project Operations and Business Central, enabling accurate budgeting and invoicing. – Automation Cuts Out Tedious Processes Say goodbye to manual reconciliation. Updates made in Project Operations sync directly with Business Central—eliminating double entries and minimizing errors. This enhances accuracy and efficiency, freeing your team members to focus on other things rather than data entry. – End-to-End Project Lifecycle Management This integration supports each project phase, from budgeting and invoicing to reporting. Full visibility means greater accountability, and everyone—from managers to teams—has the insights needed to make informed, timely decisions. – Insights That Drive Better Planning and Execution With integrated analytics and Power BI, you’re not just gathering data; you’re transforming it into actionable insights. View project profitability and resource utilization in one place, enabling better project planning and seamless operations. – Stock scenarios are covered If you are an engineering company running heavy, long term deployment projects, you are probably worried how will Project operations cover the scenarios where I need to have stock consumption on my project tasks. We have you covered here, too, so don’t worry. With the Project operations and Business Central integration we have also figured out how the stocks entries need to flow from Project Operations to Business Central because your stock movements happen in Business Central This will ensure accurate stock consumption against the projects without worrying about what goes on in the background. – WIP tracking With the actuals being passed onto Business Central, your Finance team will have WIP postings in place to give an accurate picture of the progress on the Project. Why Partner with Us? Having guided numerous businesses through D365 implementations, we know how to bridge gaps between project and financial management to unlock greater flexibility and efficiency. Our team is here to tailor this integration to your unique business needs. Ready to See the Difference? Think about your current project and financial workflows. Imagine the time saved and clarity gained by integrating them. Ready to explore what D365 integration can do for your business? Reach out to us at transform@cloudfronts.com for a free consultation, and let’s work toward operational excellence together.
