Blog Archives -

Category Archives: Blog

Taming the Chaos: A Guide to Dimension Correction in Business Central

We’ve all been there. You’re closing out the month, and you spot it: a General Journal line where the “Department” dimension is set to “Sales” but should have been “Marketing.” Or perhaps a purchase invoice was posted with an incorrect “Project” code. In the world of accounting and Microsoft Dynamics 365 Business Central, dimensions are the lifeblood of meaningful reporting, and even a single mistake can ripple through your financial statements, leading to misguided decisions and frantic period-end corrections. Fortunately, Microsoft Dynamics 365 Business Central offers a powerful, built-in safety net: the Dimension Correction feature. This isn’t just a handy tool, it’s a game-changer for financial integrity and auditor peace of mind. What Are Dimensions, and Why Do Mistakes Happen? Before diving into corrections, let’s quickly recap. Dimensions in Business Central are tags like Department, Project, Cost Center, or Region. Instead of creating separate G/L accounts for every possible combination, dimensions allow you to slice and dice your financial data, delivering incredible analytical power. Common Reasons These Errors Occur: In the past, fixing mistakes meant reversing entries, posting manual journals, and leaving a messy audit trail. Not anymore. Enter the Hero: The Dimension Correction Feature The Dimension Correction feature allows you to change dimensions on already posted entries without creating new transactions or affecting original amounts. It simply updates the dimensional context of the existing entry. Key Benefits of Dimension Correction How to Perform a Dimension Correction: A Step-by-Step Guide Let’s walk through correcting a simple example. Scenario: A telephone expense was incorrectly posted to the SALES department. It should have been posted to the MARKETING department. Step 1: Locate the Posted Entry Step 2: Initiate the Dimension Correction Step 3: Make the Correction Step 4: Verify the Change To conclude, The Dimension Correction feature transforms a once-tedious, error-prone process into a controlled, efficient, and auditable task. It empowers your finance team to maintain the integrity of your financial data without complex accounting workarounds. By understanding how to use this feature and following simple best practices, you ensure that your dimensions-and therefore your management reports – are always accurate, reliable, and ready to guide your business forward.

Share Story :

Don’t Just Delete, TRUNCATE: A Deep Dive into Blazing-Fast Data Clearing in Business Central

If you’ve worked with data in Business Central, you’ve undoubtedly used the DELETE or DELETEALL commands. They get the job done, but when you’re dealing with massive datasets ike clearing out old ledger entries, archived sales orders, or temporary import tables they can feel painfully slow. There’s a better, faster way. Let’s talk about the TRUNCATE TABLE command, the unsung hero of high-performance data purging. What is TRUNCATE TABLE? In simple terms, TRUNCATE TABLE is a SQL command that instantly removes all rows from a table. Unlike DELETE, it doesn’t log individual row deletions in the transaction log. It’s a bulk operation that de-allocates the data pages used by the table, which is why it’s so incredibly fast. In the context of Business Central, you can execute this command directly from an AL codeunit. Yes, it’s that simple. Calling the .TruncateTable() method on a record variable targets its corresponding table and empties it completely. TRUNCATE TABLE vs. DELETE/DELETEALL: What’s the Difference? This is the crucial part. Choosing the right tool is key to performance and data integrity. Feature TRUNCATE TABLE DELETE / DELETEALL Performance Extremely Fast. Operates at the data page level. Slow. Logs every single row deletion individually. Transaction Log Minimal logging. Fills the log with a single “deallocated page” entry. Heavy logging. Fills the log with an entry for every row deleted. Where Clause No. It’s all or nothing. You cannot add a filter. Yes. You can use SETFILTER or SETRANGE to delete specific records. Table Triggers Does not fire. No OnBeforeDelete or OnAfterDelete triggers are executed. Fires for each row that is deleted. Referential Integrity Can fail if a FOREIGN KEY constraint exists. Respects and checks constraints, potentially failing on related records. Resets Identity Seed Yes. The next record inserted will have the first ID in the series (e.g., 1). No. The identity seed continues from where it left off. Transaction Rollback Can be rolled back if used inside a transaction, but it’s still minimally logged. Can be rolled back, as all individual deletions are logged. When Should You Use TRUNCATE TABLE? Given its power and limitations, TRUNCATE TABLE is perfect for specific scenarios: A Real-World Business Central Example Imagine you have a custom “Data Import Staging” table. Every night, a job imports thousands of items from an external system. The first step is always to clear the staging area. The Slow Way (using DELETEALL): The Blazing-Fast Way (using TRUNCATE TABLE): The performance difference can be staggering, turning a minutes-long operation into one that completes in under a second. Critical Warnings and Best Practices With great power comes great responsibility. The limitations of TRUNCATE TABLE are not just footnotes—they are critical considerations. NO FILTERS! This is the biggest “gotcha.” You cannot use SETRANGE before calling TruncateTable(). The method will ignore any filters and always delete everything. Double and triple-check your code to ensure you are targeting the correct table. Bypasses Business Logic: Because table triggers do not fire, any essential business logic in the OnDelete trigger will be skipped. Do not use TRUNCATE TABLE on tables where the delete triggers perform critical actions (e.g., posting, ledger entry creation, validation). Using it on main transaction tables like “G/L Entry” or “Sales Line” is almost always a bad idea. Foreign Key Constraints: If another table has a foreign key constraint pointing to the table you’re trying to truncate, the command will fail with an error. DELETEALL would also fail in this case, but the error message might be different. To Conclude, TRUNCATE TABLE is a powerful tool that should be in every Business Central developer’s arsenal. When used correctly, it can dramatically improve the performance of data maintenance tasks. The Rule of Thumb: Use DELETEALL when you need to respect business logic, delete specific records, or work with tables that have complex relationships. Use TRUNCATE TABLE when you need to quickly and completely empty a large, standalone table where bypassing business logic is safe and acceptable. Embrace TRUNCATE TABLE for the right jobs and watch your large-scale data operations fly. Reference: https://yzhums.com/67343/, We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfronts.com

Share Story :

Configuring OAuth 2.0 Authentication in Power Automate

In today’s automated world, businesses depend on secure, streamlined connections between systems to improve efficiency. Power Automate, a robust tool for building workflows between various services, allows seamless integration of applications and APIs. However, when working with third-party services, ensuring that data access is secure and well-managed is critical. This is where OAuth 2.0, a secure and standard protocol for authorization, comes into play. Are you struggling to configure OAuth 2.0 authentication in your Power Automate flows? If you are considering automating workflows that interact with secured APIs, this article is for you. I will walk you through configuring OAuth 2.0 in Power Automate, so you can ensure the safety of your automation while keeping your services accessible. Why OAuth 2.0? OAuth 2.0 is the industry-standard protocol for authorization. It allows users to grant third-party applications limited access to their resources without exposing passwords. By using OAuth 2.0 in Power Automate, you ensure that the services and APIs you connect to are secure, and that tokens are used to access data on behalf of the user. How OAuth 2.0 Enhances Security OAuth 2.0 significantly improves security by eliminating the need to share sensitive credentials. Instead, access is granted through tokens, which are time-limited and easily revocable. OAuth 2.0 is widely used by many companies, including Microsoft, Google, and Salesforce, to integrate applications securely. Step-by-Step Guide to Configuring OAuth 2.0 in Power Automate 1. Set Up OAuth 2.0 Credentials Before configuring OAuth 2.0 in Power Automate, you need to set up OAuth 2.0 credentials in the platform you’re working with. For example, if you’re using Microsoft Graph API or any third-party service, follow these steps: 2. Initialize OAuth 2.0 Variables in Power Automate Now that you have your client ID and client secret, it’s time to configure them in Power Automate. Set up the variables: 3. Configuring the OAuth 2.0 Connection in Power Automate With the client credentials set, it’s time to establish the connection to the service using OAuth 2.0. 4. Use OAuth Token to Access Secure Data Now that you have the OAuth token, you can use it to authenticate your requests to third-party APIs. 5. Best Practices for OAuth 2.0 in Power Automate To conclude, OAuth 2.0 authentication provides a secure and effective way to authorize third-party applications in Power Automate. By following the steps outlined in this guide, you can set up OAuth 2.0 authentication, ensure data security, and integrate third-party services into your automation workflows with ease. If you’re ready to secure your Power Automate workflows with OAuth 2.0, follow the steps outlined in this post and start integrating APIs in a secure manner today. For more tips and detailed guides, check out our other blog posts on Power Automate and API integration. Need help with the OAuth 2.0 integration? Feel free to reach out for assistance! We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfronts.com

Share Story :

Mitigating Implementation Risks Through a Structured Business Assessment

The landscape of digital transformation has never been more complex. Rapid market shifts, rising customer demands, and tightening budgets have made technology decisions more consequential than ever. The challenge isn’t adopting new tools it’s leading transformation by ensuring that every investment is grounded in clarity, alignment, and predictability. At CloudFront’s, we understand this. That’s why our Business Assessment Engagement model has become a proven first step toward successful, low-risk technology implementations. What Is a Business Assessment? A Business Assessment is a structured, short-term engagement conducted before signing a full implementation Statement of Work (SoW). It is designed to create complete visibility into your current business processes, desired future state, and the potential risks that could impact your project. Typically spanning 3 – 4 weeks, this engagement brings together functional and technical stakeholders from both your organization and CloudFront’s. Whenever feasible, we conduct this assessment onsite, ensuring close collaboration and a deep understanding of your business landscape. During this engagement, our experts: The result is a detailed Business Requirements Study (BRS) a comprehensive document that translates assessment insights into an actionable implementation roadmap. This BRS becomes the foundation for a precise and mutually agreed Statement of Work, ensuring every phase of your digital transformation is built on validated insights and shared understanding. Why a Business Assessment Matters For enterprise technology leaders, the Business Assessment approach delivers tangible benefits: Ultimately, this process transforms uncertainty into informed decision-making, enabling IT leaders to confidently advance from planning to execution. Proven Success with CloudFront’s At CloudFront’s, we’ve seen firsthand how Business Assessment engagements set the stage for successful digital transformations. Clients who adopt this model enter implementation phases with greater predictability, stronger governance, and renewed confidence in both the technology and the partnership driving it. Recently, we partnered with one of the world’s largest U.S. based commercial vehicle manufacturers to conduct an onsite Business Requirements Study (BRS). Our team worked closely with their stakeholders to map existing systems and design a strategic roadmap for migration to Microsoft Dynamics 365 Supply Chain Management (SCM). Following the successful completion of the BRS, we are now leading Phase 1, enabling their inventory, advanced warehouse, and procurement operations to establish a strong operational foundation. In Phase 2, we will enable master planning, production, and quality management to deliver end-to-end operational efficiency, ensuring a seamless and future-ready digital ecosystem. Our clients consistently tell us that this approach not only de-risks their investment but also enhances alignment between business and IT, a crucial factor in any transformation journey. To conclude, in today’s unpredictable business landscape, a well-executed Business Assessment isn’t just a preliminary step, it’s a strategic imperative. By partnering with CloudFront’s for a Business Assessment, you’re not committing to uncertainty; you’re investing in clarity, alignment, and long-term success. If your organization is planning a digital transformation initiative, start with a Business Assessment Engagement and move forward with the confidence of knowing your path is mapped, risks are managed, and success is measurable. Ready to move from uncertainty to clarity?Connect with CloudFront’s at transform@cloudfronts.com to schedule a Business Assessment Engagement and gain a clear, actionable roadmap for your next digital transformation. Contact Us to start your assessment today.

Share Story :

FetchXML Made Simple: Power Pages Tips for Dynamic Data Retrieval

Dynamics 365 Power Apps Portals (formerly Dynamics 365 Portals) allow organizations to securely expose CRM data to external users. However, fetching and displaying CRM records in a portal page requires more than just entity lists – it often needs custom data queries. That’s where FetchXML comes in. FetchXML is Dynamics 365’s native XML-based query language used to retrieve data and it’s fully supported in Liquid templates within portals. Step 1: Pre-Outline Brief Target Audience: How this blog helps: Step 2: Blog Outline Opening:Identify the need for FetchXML in Power Pages and its importance for developers and portal managers. Core Content: Step 3: Blog Post Introduction For businesses leveraging Microsoft Power Pages, the ability to pull dynamic data from Dataverse is critical. While out-of-the-box entity lists work for simple scenarios, complex needs — such as personalized dashboards and filtered data — require custom FetchXML queries embedded in Liquid templates. In this post, we’ll walk you through how FetchXML works in Power Pages, share examples, and provide best practices so you can deliver efficient, personalized portals. Why This Matters For growing businesses, service portals need more than just static lists. As the volume of data increases, the ability to dynamically query and display relevant information becomes essential to maintain performance, improve user experience, and reduce maintenance efforts. With FetchXML in Liquid, developers can: Prerequisites Before getting started, ensure: Understanding FetchXML FetchXML is an XML-based query language for Dataverse. It allows you to: Example: Retrieve all active contacts: Using FetchXML in Power Pages (Liquid Templates) Here’s a basic implementation: This will execute the query and display results dynamically in your portal. Making FetchXML Dynamic You can make FetchXML personalized by using Liquid variables. Example: Display cases only for the logged-in user: Real-World Example: Recent Cases Dashboard] Best Practices To conclude,FetchXML in Power Pages is a powerful tool for creating customized, dynamic, and efficient portals. Start small — add a dynamic list or dashboard to your portal today. If you need expert guidance, CloudFronts can help you implement FetchXML-driven solutions tailored to your business needs. 💡 Want to learn more? Reach out to CloudFronts Technologies at transform@cloudfronts.com to explore FetchXML use cases for your portals and improve your customer experience.

Share Story :

From Portal Chaos to Power Pages Zen: My Journey Automating Client Forms

Power Pages, the modern evolution of Power Apps Portals, has redefined how organizations build secure, data-driven web experiences connected to Dynamics 365. But let’s be honest, for anyone who’s wrestled with the old portal setup, the journey from chaos to clarity isn’t always smooth. In this blog, I’ll share how I transformed a tangled web of client forms and scripts into a streamlined Power Pages experience using Dynamics 365 forms, Liquid templates, and JavaScript automation — and what I learned along the way. The Beginning of PortalsMy story began with what I thought was a simple request, automate a few client onboarding forms in Power Apps Portals.What followed? I realized I wasn’t managing a portal — I was managing chaos. That’s when I decided to rebuild everything in Power Pages, the modernized, secure, and design-friendly version of Power Apps Portals. Why Power Pages Changed Everything Power Pages offers a low-code, high-control environment that connects directly to Dataverse and Dynamics 365.Here’s what made it a game-changer for me: 1. Built-In Dataverse Power No more juggling SQL tables or external APIs.Dataverse made it simple to store, validate, and update client data directly within Dynamics 365 — cutting down my custom integration scripts by almost 60%. 2. Cleaner Authentication With Azure AD B2C integration, user sign-ins became seamless and secure.I could finally define granular access roles without needing custom web roles or Liquid conditionals scattered across pages. 3. Design That Doesn’t Break Your Brain The Power Pages Design Studio felt like moving from notepad to Figma — I could visually build layouts, insert lists, and add forms connected to Dynamics data without touching complex HTML. Automating Client Forms: My Aha Moment The real “Zen” moment came when I realized that automation in Power Pages didn’t need to be messy.Here’s how I approached it step-by-step: Used Dynamics 365 Forms in Power PagesEmbedded native forms from Dynamics instead of building them from scratch — they respected business rules and validation logic automatically. Applied Liquid Templates for Smart RenderingI used Liquid to conditionally show fields and sections, keeping client forms dynamic and user-friendly.Example: Added JavaScript AutomationFor client-side logic like field dependencies, autofill, and dynamic visibility, JavaScript did the trick. Because Power Pages supports modern script handling, I could isolate my logic cleanly instead of cluttering the HTML.Example: Leveraged Power AutomateIntegrated flows triggered on form submission to send confirmation emails, update records, and even notify the sales team instantly. I integrated Power Automate flows for backend actions: This separation of concerns (frontend in JS/Liquid, backend in Flows) made everything more maintainable. Design Meets Logic: Keeping It Clean One of my key lessons – separate design from logic.Power Pages Studio handled the look and feel, while all the conditional logic stayed in: This modular approach made my site easier to maintain and upgrade later. Security & Permissions Simplified Earlier, managing web roles in Portals was like untangling a spider web.Now with Power Pages: The result? A cleaner, safer, and more scalable structure. The End Result: From Chaos to Zen After weeks of trial, testing, and caffeine, my new Power Pages site was: What once required hours of manual fixes now runs seamlessly, freeing me to focus on building rather than babysitting. Happy Developing!We hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfronts.com

Share Story :

GST Implementation Made Easy in Dynamics 365 Business Central

For any Indian business running on Microsoft Dynamics 365 Business Central, tax compliance isn’t optional, it’s foundational. The Goods and Services Tax (GST) framework is complex and manually managing it is a high-risk gamble. This guide isn’t just a list of steps; it’s your definitive blueprint for configuring Business Central’s powerful Indian localization features to handle GST seamlessly. We will transform your ERP from a standard ledger into a fully automated, compliance-ready machine. Ready to banish tax-related data entry errors and audit anxiety? Let’s dive in and set up the system correctly, from defining your GSTINs to mastering the G/L posting matrix. Microsoft Dynamics 365 Business Central offers robust localization features for India, including comprehensive support for the Goods and Services Tax (GST). Properly configuring GST is essential for calculating, recording, and settling taxes on all your inward and outward supplies, ensuring compliance with Indian tax laws. This guide provides a straightforward, step-by-step process for setting up GST in Business Central, based on Microsoft’s best practices. Phase 1: Laying the Foundation (Tax Periods & Registration) The initial phase involves setting up the legal and temporal frameworks for your GST configuration. Step 1: Define Tax Accounting Periods (GST Calendar) The GST regime operates on a specific timeline, and you need to define this within Business Central. Step 2: Establish Your GST Registration Numbers (GSTINs) Your Goods and Service Tax Payer Identification Number (GSTIN) is critical for identifying your tax entity and the state you operate in. Phase 2: Core Configuration (G/L Accounts and Masters) This phase links the statutory requirements with your company’s general ledger structure. Step 3: Configure GST Groups and HSN/SAC Codes These setups classify your goods and services for accurate rate calculation. Step 4: Define the GST Posting Setup (The Accounting Link) This is perhaps the most crucial step, as it determines which General Ledger (G/L) accounts are used to post GST amounts. Step 5: Set Up GST Rates With your Groups and HSN/SAC codes defined, you now specify the actual tax percentages. Phase 3: Master Data Integration (Connecting the Dots) The final phase ensures that your business entities and locations are linked to the defined GST rules. Step 6: Update Company and Location Information Your company’s primary details must be GST-compliant. Step 7: Configure Customer and Vendor Master Data For every trading partner, you must define their GST status and registration details. To conclude, by following these seven steps, your Indian company’s Business Central environment will be fully configured to handle GST calculations automatically. This setup allows the system to determine the correct tax component (CGST, SGST, or IGST), apply the right rate, and post the amounts to the designated G/L accounts, simplifying your day-to-day transactions and preparing you for GST settlements and reporting. I Hope you found this blog useful, and if you would like to discuss anything, you can reach out to us at transform@cloudfronts.com.

Share Story :

Advantages and Future Scope of the Unified Databricks Architecture – Part 2

Following our unified data architecture implementation using Databricks Unity Catalog, the next step focuses on understanding the advantages and future potential of this Lakehouse-driven ecosystem. The architecture consolidates data from multiple business systems and transforms it into an AI-powered data foundation that will support advanced analytics, automation, and conversational insights. Key Advantages Centralized Governance:Unity Catalog provides complete visibility into data lineage, security, and schema control — eliminating silos. Dynamic and Scalable Data Loading:A single Databricks notebook can dynamically load and transform data from multiple systems, simplifying maintenance. Enhanced Collaboration:Teams across domains can access shared data securely while maintaining compliance and data accuracy. Improved BI and Reporting:More than 30 Power BI reports are being migrated to the Gold layer for unified reporting. AI & Automation Ready:The architecture supports seamless integration with GenAI tools like Genie for natural language Q&A and predictive insights. Future Aspects In the next phase, we aim to:– Integrate Genie for conversational analytics.– Enable real-time insights through streaming pipelines.– Extend the Lakehouse to additional business sources.– Automate AI-based report generation and anomaly detection. For example, business users will soon be able to ask questions like:“How many hours did a specific resource submit in CRM time entries last week?”Databricks will process this query dynamically, returning instant, AI-driven insights. To conclude, the unified Databricks architecture is more than a data pipeline — it’s the foundation for AI-powered decision-making. By merging governance, automation, and intelligence, CloudFronts is building the next generation of data-first, AI-ready enterprise solutions.

Share Story :

Unified Data Architecture with Databricks Unity Catalog – Part 1

At CloudFronts Technologies, we are implementing a Unified Data Architecture powered by Databricks Unity Catalog to bring together data from multiple business systems into one governed, AI-ready platform. This solution integrates five major systems — Zoho People, Zoho Books, Business Central, Dynamics 365 CRM, and QuickBooks — using Azure Logic Apps, Blob Storage, and Databricks to build a centralized Lakehouse foundation. Objective To design a multi-source data architecture that supports:– Centralized data storage via Unity Catalog.– Automated ingestion through Azure Logic Apps.– Dynamic data loading and transformation in Databricks.– Future-ready integration for AI and BI analytics. Architecture Overview Data Flow Summary:1. Azure Logic Apps extract data from each of the five sources via APIs.2. Data is stored in Azure Blob Storage containers.3. Blob containers are mounted to Databricks for unified access.4. A dynamic Databricks notebook reads and processes data from all sources. Each data source operates independently while following a governed and modular design, making the solution scalable and easily maintainable. Role of Unity Catalog Unity Catalog enables lineage, and secure access across teams. Each layer — Bronze (raw), Silver (refined), and Gold (business-ready) — is managed under Unity Catalog, ensuring clear visibility into data flow and ownership. This ensures that as data grows, governance and performance remain consistent across all environments. Implementation Preview:In the upcoming blog, I will demonstrate the end-to-end implementation of one Power BI report using this unified Databricks architecture. This will include connecting the gold layer dataset from Databricks to Power BI, building dynamic visuals, and showcasing how the unified data foundation simplifies report creation and maintenance across multiple systems. To conclude, this architecture lays the foundation for a unified, governed, and scalable data ecosystem. By combining Azure Logic Apps, Blob Storage, and Databricks Unity Catalog, we are enabling a single source of truth that supports analytics, automation, and future AI innovations.

Share Story :

Optimum Window Partners with CloudFronts for Managed Services Agreement (MSA) Renewal 

We are delighted to announce that the largest US-based manufacturer of Fire-Rated and Architectural steel windows, Optimum Windows is partnering with CloudFronts for Managed Services Agreement (MSA) renewal.  Optimum Window, established in 1985, is a family-owned business based out of Ellenville, in upstate NY. Since then, Optimum Window has become the largest and most diversified manufacturer of Fire-Rated and Architectural steel windows in the United States and has continued its growth with a series of custom high-tech metal window and door systems designed for commercial, high-end residential & landmark applications. Learn more about Optimum Window at https://optimumwindow.com/   Optimum Window’s partnership with CloudFronts began with the implementation of a CRM system with custom enhancements that automates their end-to-end sales & order processes.  Under this MSA, CloudFronts will provide support & maintenance services for the system based on Microsoft Dynamics 365 Sales.  About CloudFronts  CloudFronts is a global AI- First Microsoft Solutions & Databricks Partner for Business Applications, Data & AI, helping teams and organizations worldwide solve their complex business challenges with Microsoft Cloud, AI, and Azure Integration Services. We have a global presence with offices in U.S, Singapore & India.    Since its inception in 2012, CloudFronts has successfully served over 200+ small and medium-sized clients all over the world, such as North America, Europe, Australia, MENA, Maldives & India, with diverse experiences in sectors ranging from Professional Services, Financial Services, Manufacturing, Retail, Logistics/SCM, and Non-profits.     Please feel free to connect with us at transform@cloudfronts.com 

Share Story :

SEARCH BLOGS:

FOLLOW CLOUDFRONTS BLOG :


Secured By miniOrange