In this post I will give an up-to-date overview of D365 Finance and Operations integration options, mentioning their methods of integration and advantages and disadvantages of every integration method compared to each other. I will also give you information and advices from my first hand experiences using them.
Topics
OData
Odata is an open source protocol to serve and consume interoperable data using common query operations with RESTful APIs. D365FO exposes all its public data entities as OData endpoints which can then be accessed using following URI format :
https://<d365fourl>/data/<dataentity>
OData provides a quick, codeless data integration method with many data query options and CRUD operations. You can use its open standard query string language to query your data and do data manipulations using standard Odata CRUD commands, all using just simple and open standard REST calls. If you would like to call a custom method within your data entity, this is also supported by exposing custom actions with your OData entities, so the commands are also extendable to a certain point.
You can definitely use OData for your own integration projects, but there are also many OData-Ready software available today, and these can directly be connected to D365FO Odata endpoints. Microsoft Power BI also supports OData connection and you can connect Power BI using OData if you feel lazy with setting up faster data integration possibilities methods, like Entity Store and Data Lake integrations.
Although it looks like the optimum way of data integration with D365FO, there are some drawbacks involved. OData queries and data operations are executed really slow and data reading may take ages if you try to retrieve a large entity. OData is mainly designed for simple CRUD operations and simpler queries. If you need to execute complex queries, like complex joins and lookups for example, you may start to hit its limits.. Although you can add some custom actions to extend available OData commands with your own ones, complex operations and business logic unfortunately do not go very well with it. It may be required to place this complex logic on consumer application if you decided to integrate it using OData.
There is also a rather new feature of D365FO to throttle calls to OData endpoints by giving priorities to them, to avoid system lockdowns that might be caused by frequent OData calls. You can read more about it from the link below :
Remember you can also use OData endpoints in Azure API manager, just like custom services which I will mention soon..
ADVANTAGES
- Open and standard data integration protocol with many data inquiry and CRUD commands out of the box
- Support for extending with custom data actions
- OData-ready software can directly be integrated using existing endpoints
DISADVANTAGES
- Slow…
- Not suitable to be used with complex business logic and queries
More info: https://docs.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/data-entities/odata
Custom Services
This is by far the most flexible and customizable way to integrate with D365FO and the one I use mostly in my integration projects. Custom services are created with standard D365FO X++ code and can be used for both data based and operation based (posting stuff etc.) integrations, limit being just your imagination.
AIF services used in Dynamics AX are now upgraded to custom services in FO and they are automatically published into SOAP and JSON REST endpoints when deployed. However, management and monitoring tools used for old WCF AIF services are now disappeared, leaving just naked, not configurable services endpoints.
When you deploy a custom service, following endpoints are created and you can call them remotely using standard AAD OAuth authorization and SOAP/REST HTTP calls:
SOAP:
https://<d365fourl>/soap/services/<servicename>?wsdl
JSON:
https://<d365fourl>/api/services/<servicegroupname>/<servicename>/<operation_name>
Data integration in custom services is done via data contracts which are then converted to XML or JSON depending on the endpoint. Performance of doing so is quite fast, compared to other data integration options since everything is handled with low level .NET commands. You can in theory transfer big amounts of data this way however I do not recommend doing so. With large data transfers you may hit some not-foreseen limitations and timeout problems on the way. But it is still possible to program some kind of paging mechanism in your service class to send small chunks of larger data to skip these limitations as well.
If you need to see an example, I have used a custom service in my previous Android integration blog post, and X++ code of this service is available in my github page:
The throttling mechanism we mentioned in OData topic can also be used for custom service endpoints.
If you need to do management and monitoring of your custom service, or configure some custom authorization methods instead of AAD, there is a possibility connect it with Azure API Manager and empower its monitoring and authoring abilities, as described in my another blog post :
ADVANTAGES
- Highly flexible and customisable way of integration
- Best performance compared to other integration methods
- Support for both operation and data based integrations
- Both SOAP and JSON endpoints are available upon deployment
- Possibility to use them in Azure API Manager
DISADVANTAGES
- Lack of management and monitoring tools in standard (possible with Azure API Manager)
- Development requires an experienced X++ developer
- Not suitable out-of-the-box for large data transfers
More info: https://docs.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/data-entities/custom-services
DMF (Data management framework)
DMF allows you to import and export big data in D365FO using various source/target data files and services. This is the recommended way of data integration in D365FO if you need to transfer large amounts of data from other systems using supported file formats and services. It supports also incremental data load/export with change tracking and a data staging feature for post data validations and processing.

Import and export projects are processed using “data entities”, which are normalized views of D365FO data tables and can be added programatically if necessary. There is a dedicated Data management workspace which you can setup data import and export projects, import actual data or see available data entities and check integration mappings in graphical schemas. However useful it is, I cannot tell you that the DMF UI is an user friendly one. Some parts of the “advanced” and the “standard” views are confusing to users, even to developers. Also field mapping and Unicode setup parts are confusing and do not always work as expected.
There is support for every old file format which was supported by Dynamics AX before, and some new formats and SQL Server DB connection options are available, but there is still no support for modern data transfer file formats like JSON, Parquet, ARRF, Avro etc.. Also as package options you can only use ZIP compression and there is no support for GZIP which is widely used for data transfer today. To transfer the data inside a file, CSV files are the recommended option. As you can see in the analysis below from luminousmen.com, a CSV file is quite an outdated file format and falls far behind on every aspect compared to modern data transfer file formats.
https://luminousmen.com/post/big-data-file-formats
DMF has REST APIs and Azure actions to import and export data from outside D365FO. You can find Azure actions to be used in Logic Apps under the “DataManagementDefinitionGroups” section of “Execute action” :

You can find example code (and logic apps) for data import and export using DMF in the Dynamics-AX-Integration GitHub page. However; If you expect a straightforward way of using these API commands to export or import your data, you will be disappointed.. For example for importing a CSV file, you need to create a ZIP file package containing the file you want to import and add fixed Manifest and Header XML files in it,.. then place it in a temporary container DMF provides for you, import it using another action and then use an infinite loop for checking if the import is completed successfully or not, and then try to get the actual error checking various DMF actions to inform the user about the problem that occured..
There is currently no other direct file import or export method available, you need to zip it with those fixed Manifest and header files (which you also need to store in a fixed location) and setup workflows in the example code to import them. It would be very nice to have one straightforward option in the future though, that receives the data file as input and consolidated error and log messages as output..
Another thing to mention is running import jobs in parallel; if you call two APIs in parallel importing into the same data entity, you get errors and problems with it, although it is possible to execute such action using async DMF API methods, or async processings in logic apps etc. Also the table you need to export should not have any locks on it, or you run into some nasty problems.
ADVANTAGES
- Suitable for big data integration
- Incremental data import/export possibility
- Staging tables for post data validation
DISADVANTAGES
- No support for modern big data formats
- UI is not user friendly and confusing
- DMF API’s and Actions for integration from outside D365FO are not straightforward and practical
- Table lock problems, parallel executions will fail for the same data entity although async import like this is possible
- Performance is disappointing
Recurring integrations
Recurring integrations is a D365FO data integration platform based on DMF and data entities, providing automated data exchange possibilities between third party data providers. Once setup, Recurring Integration platform REST API’s can be used by third party integrators to import/export data from and to D365FO.
Recurring integrations feature can be enabled from a single button click in a DMF project and later can be managed/monitored from D365FO RI admin forms. REST API’s then used to push and pull data generated by this recurring integration, which can then be picked up in an ordered and scheduled way. These API’s used for recurring integrations are fixed and you cannot extend them for custom functionality.

ADVANTAGES
- Complete, stable integration platform with useful features like importing data in a sequence and skipping empty files etc.
- Fully automated and scheduled way of integration
- Unlike DMF API, both data file import and package import options are available in the API
DISADVANTAGES
- APIs are fixed and not customizable
More info : https://docs.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/data-entities/recurring-integrations
Business events
D365FO business events feature allows you to send notifications of FO business events to Azure event handlers and trigger based workflow providers. It comes with many out-of-the-box FO business events plus provides you the ability to create new ones with X++ programming. You can send notifications from these events to Azure endpoints like Service bus, Event hub, Event grid and Logic apps (or Power Automate), also to custom HTTPS endpoints using standard http methods. It is also possible to subscribe a business event directly from Logic apps or power automate using “When a business event occurs” trigger, as shown below:

It is also possible to attach small amounts of data to the notifications using message payload, however I would advise you to be really careful with that since attaching large amounts of data to your payload will not only create all sorts of problems, it will also ruin the lightweight operation expected from a notification service.
If you need more information, I have a complete blog post about business events, describing also how to create a new event using X++:
ADVANTAGES
- Realtime integration with event based Azure services and custom HTTPS based services
- Trigger methods available for logic apps and power automate
- Possibility to create custom events
DISADVANTAGES
- Not suitable for large data transfers
- Creating custom events requires skilled X++ developers
More info: https://docs.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/business-events/home-page
Entity store
Many get surprised when they hear D365FO is not really a SQL Server based software and its database logic and metadata are not directly accessible via SQL server. D365FO, in other names Axapta or AX, stores its metadata, record id mechanism, data type information and data relations in its own kernel and uses SQL server only as a bulk storage, which is then not directly readable by third party applications and analytics software. Dinosaurs like me would explain the reason for that as the ancient Damgaard Axapta was designed to be a “multi-database” application and supported Oracle DB as well as SQL server as a database options. So in that time, it was a decision to keep these metadata in the kernel instead and make it easier to switch databases if needed.
So; to be able to publish normalized, third party readable SQL server data from D365FO, we need to use “data entities” and “aggregate measures” for data normalization and export their content to another SQL server database using the “Entity store” and BYOD functionalities. Entity store option publishes aggregate measures (start schema of data entities)(cubes, sort of) of D365FO in a separate SQL server database (AXDW) which can later be read by Azure analytical software (Power BI) to run analytical reports.

Entity store data can be updated manually or using batch jobs. As necessary, there is also an option to develop custom aggregate measures using D365FO development tools. As you can guess this integration is read-only and in fact currently only available to MS Power BI. The reason is, in D365FO production environments, this AXDW databases are not accessible to end users, and you cannot get authorization for any other third party app if you need to use the same database for your own purposes.. BYOD on the other hand, which we will mention in the next topic, makes this possible by exposing data to a user database instead.
ADVANTAGES
- Faster way of providing analytical data to Azure reporting tools
- Batch update possibility
DISADVANTAGES
- Read-only
- Power BI only
Bring your own database (BYOD)
BYOD allows you to export entity data to a user created SQL Server database instead of the Entity store AXDW, which is currently not allowed to be accessed by end users in production environments. Using BYOD you can export D365FO data entities to your database manually or with batch update or change tracking options using the standard DMF export projects page. To start using BYOD, you need to configure your SQL Server database in the “Data management” workspace “Configure Entity export to database” tile:

Then you can use the Publish button to create SQL tables for selected data entities in the target database. After these steps are completed, you can export the data to your SQL database using DMF export jobs and selecting your BYOD SQL Database from the list as a target.
Advantages
- Data can be exported to and accessed from user created SQL server databases
- Exporting table data using query filtering is possible
- Incremental updates with change tracking are supported.
- Batch update possibility
Disadvantages
- One way, export-only integration
Azure Data Lake
A feature still in public preview, Azure data lake publishing provides a more practical, cheaper and faster storage option for your big analytical and integration databases which you used to access using Entity store or BYOD options before. It can publish data from entity store, data entities and standard FO tables to a “Data Lake Storage Gen2” storage, in a “near real-time” speed using the new D365FO “Data feed service”. This new data feed service works faster and saves you the hassle of maintaining DMF export jobs and batches.
Publishing your data into a data lake storage brings in many benefits. It costs much less to store data in Data Lake compared to a SQL Server data space in Azure. Plus data lake allows you to combine your FO data with data from other applications, and also CDS inside the same data lake space and run consolidated reports with them. Data lake supports Common Data Model folder structure which can be understood by many Cloud applications and services, and it is also possible move the data from and to CDS using CDS Data lake integrations. Usage of Azure data transformation tools like Azure Data Factory is also possible.
All in all, it brings you a couple of steps closer to Azure cloud integration and provides you the future proof “near-real time” data synchronization.
In the time of writing, Azure data lake options is still in public preview and needs to be enabled from D365FO “Feature management” window first. Then you will be able to choose tables to be published to data lake from “Export to Data Lake” form and push Entity Store to Azure data lake from the Entity store window:

ADVANTAGES
- Reduced data storage costs compared to SQL Server storage used by BYOD or Entity Store options
- Near-realtime data integration with automatic data feeds
- Improved integration with many Azure cloud apps and services using Common Data Model
- Possibility of using Azure data transformation tools like Data Factory
Disadvantages
- One way-only integration
Dual Write
Dual write, one of the newest integration features of D365FO, provides a near-real time, tightly coupled and synchronous data transfer between D365FO and CDS. CDS sits at the heart of Microsoft Power Platform and used by many MS applications (also called model-driven apps). Dual write will eventually make it possible to synchronously integrate D365FO with other software in Dynamics 365 package and power platform components.
Any data written to D365FO dual -write data entities is written directly into its CDS couple in near real-time speed, and similarly the other way around. Real time data synchronization process has “Running” and “Paused” mode options, and can cache write operations in offline or paused mode to be executed afterwards.
At the time of writing, dual write setup can be done via LCS and from the Data management module section below. However, I will not get into details of the setup since it is subject to change in short term. Once setup, you first need to do an initial sync for the entities, then you can start using them in the real-time sync mode :

Dual write feature is just out of its “public preview” and released for general availability in PU33. In my own tests in the current release the “near real-time” part had some delays and glitches. Another thing to mention is the data schema differences between CDS and D365FO. CDS is using “Common Data Model”, which has quite simple schemas for its master data tables, on the other hand D365FO can have quite complex data structures like trade agreement and pricing modules or directory and party structures for storing contact information and addresses. So it is quite challenging to keep two different data structures in harmony with each other, and MS is working on some unified data models to achieve that. In the release version some of the out-of-the-box mappings of these complex areas are not complete or do not match with CDS; But Dual write is just a very new feature which MS still invests on and it will surely improve in time.
ADVANTAGES
- Tightly coupled near real-time integration with power platform and model-driven apps
- Online and offline modes
- Queries and filters for data integration
- Ability to customise field-entity mappings
- Pausing possibility
DISADVANTAGES
- Room for improvement in “near-real time” performance
- Mappings of some data models not complete or do not match each other yet
Consuming external web services
In addition to the standard integration features of D365FO, there is also a possibility to integrate by consuming external web services (or Azure services) within D365FO. To be able to do that, we need to use .NET libraries and functions to call the external web service within D365FO.
In the past I was able to do that using only X++ code and standard CLR-Interop features of AX, but nowadays, web service technologies are evolved and the libraries and functions to call them get much more complex. So I recommend you to build a seperate .NET library for your external call to avoid hitting some limitations you might experience with CLR-Interop (like async features and so) with using your new .NET service call functions. In D365FO, you can easily add .NET libraries to your VS solution and reference them to your X++ projects in the same solution.
The web service you are trying to access might also be operating in async web service mode. Async web services are special ones that do not wait for its operations to complete, but respond immediately with an 202 Accepted response and queue your request for later processing. Afterwards you need to inquiry if your operation is completed or not from your caller code. So they have quite different method of calling compared to legacy web services.
Also the authorization methods nowadays are a bit more complicated, like we know the OAuth from Azure services. You may write a .NET library using Microsoft sample code with OAuth but it may go outdated afterwards and you may need to update your library reflecting the changes (happened quite often to me). Therefore, as you see things nowadays are a bit more complicated with web services and I not only recommend you doing your REST-Web service calls in .NET libraries but also recommend using a good, maintained open source service call NuGet package to do so, like “Service stack”, “Restsharp” or “Flurl”.
Once your library is complete and connected to your X++ class, you can use standard X++ calls to retrieve data from the external service, or put it in a batch and do scheduled calls.
Unfortunately, I do not plan to post an example library for that topic in my blog because the methods used for doing web service calls evolve rapidly and such examples and blog posts get outdated in no time. But further questions, you can leave in comments below and I will try to answer them.
ADVANTAGES
- High performance, direct calls with external services using low level .NET commands
- Flexibility
DISADVANTAGES
- Requires highly skilled X++/.NET engineering
More info (outdated): https://docs.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/data-entities/consume-external-web-service
Nice post. Do you plan to support/update it?
if yes
Can you also add a link to throttling page for OData https://docs.microsoft.com/en-us/dynamics365/fin-ops-core/dev-itpro/data-entities/priority-based-throttling ?
and probably add one more point in your list
Custom x++ periodic batch jobs (where you can read/write from the external services, like Azure SQL, webservices, etc..). I think there are several examples in the standard system like Periodic import of exchange rates..
Sure!
Indeed a very helpful post
Awesome overview, thanks!
And yes, dual-write and data lake are still maturing technologies, so you will have a full-time job updating details 🙂
Suggest to add Virtual Entities, even if they are simply oData endpoints wrapped up nicely
… and, for the historical record, Axapta could also run on Office Access db :-O
Thank you, yes it is quite difficult these days to update blog posts, Business Events post for example has more updates than post itself and today it needs an update again because MS removed Logic App IDs as links.
Virtual entities stays on the CE side so I did not include it here.
I remember running Navision on Access DB for demos but personally never saw Axapta on Access DB 🙂 Interesting fact, thanks.
Thank you for the post. One question, can recurring integration update data? I think it can only insert new records. Thanks.
Hi, RI is also based on DMF platform and data entities so if you can update records from a DMF project for that entity, you can also do it via recurring integrations. You can also directly create a recurring integration from a DMF project by clicking the menu item on the toolbar.
Very useful post, I have a client who is experiencing difficulties with using the BYOD option, data from power BI reports went out of sync for a number of days, would you recommend using data lakes instead now ?
If applicable for the customer scenario, that would be the best
Great post! Maybe you can help me to clarify how to approach the following.
We have this integration scenario: a third-party developed mobile app and cloud-based middleware that collect sales orders and process payments. They store this data in the middleware and need to send it to D365FO. At the same time, from D365FO we need to send them Parameters, Master data and some transactional data aggregated like sales stats, customer balance, collection balance, etc.
Currently we are using a file-based integration based on the Recurring Integrations Scheduler(RIS). But both the customer and the mobile app provider have some issues with this setup and want to get rid off it.
Taking into account that we might have different sizes of daily inbound/outbound loads, where master data might represent a few hundreds, few thousands for sales orders and aggregated data could be several hundreds of thousands, which approach would you recommend using?
-File-based integration based on periodic import/export of files from/to Azure storage and calling external Azure services from X++ to download/upload these files. Not using RIS nor DMF. For small and large loads.
-Azure Integration Platform (Service Bus queues) and read/write messages from X++. For small and large loads.
-Hybrid: OData/Custom services for small loads and either file-based or Azure Service Bus queues for large ones.
-The current one, based on RIS.
Any clue would be much appreciated.
Thanks!
Thanks. I understand the problem you have,. file based integrations with DMF unfortunately works like this and not quite performance friendly options but they can handle large amounts of data and do bulk data transfers with data staging/error tracking abilities. Other options like OData have their own shortcomings like lack of staging functionality, error handling and data consistency checks.
It would be a good idea for you to check custom FO integration solutions done by various ISVs which are designed to remedy shortcomings of DMF and OData. Otherwise, if possible, a complete customized solution using custom services on FO side and Business events/Azure functions and services on cloud side will perform much better and be more flexible but of course you need a lot of development work to achieve this.
Very nice post. I would add some sort of information with examples and information regarding DW and common issues which might happen when you decide which integration to choose. I also consider ADF and Azure service bus and Synapse as a complex solutions to build integration.
Thanks. I would prefer keeping it on basics like that otherwise I need to update the post at least every week.
Hi Sertan,
Do OData export occur as batch? What I mean is, if I am executing an OData Request to return 100 records, does FinOps processes each record at a time or does it executes as a batch?
That being said, if lets say 1 record takes 10 seconds, will processing an export of 100 records via OData take 100*10 seconds?
Hi, OData executes as a REST service with request and respond logic. And it is using FO data entities to send normalized data.
I think from your description you mean ‘batch’ as sending bulk data from SQL. Unfortunately FO does not do that and processes OData internally to collect the data line by line. This is how virtual fields and data entity “postLoad” code works for each line, those are not executed inside SQL server but FO.
Similar processes can be utilized to record sure subcontract prices.
Great post indeed. However, i am always really disappointed of the lack in the FO On-Prem docs as well as the limitations that Microsoft intends to set on on-prem.
i would have appreciated if you clearly mentioned if there is any way to apply for on-prem for for every option.
Hi Sertan,
Thanks for the detailed explanation on the approaches of integration. I have a use case where we have to send the customer/product/order data from third party to D365 F&O. Also we have to send the order status updates back to third party from D365 F&O. We will have to do this both in realtime and in batches. For real time data transfer, is “custom services” the only option?. Can you please suggest us on which approaches we can use for this? Also can we use Power Automate to do this data exchange? Can we have connections to third party and D365 F&O in power automate?
Hi, for real time custom services is one of the options and OData is another. So basically for real time you need to send a request and receive a response to confirm it is processed, which is done via RESTful services. Both DMF and service calls can be done via power automate, MS has example logic app in Github for DMF but you need to tweak it a bit as written in project comments.
Very helpful post for beginners in integrations.
Thanks for this kind of valuable information shared with us.