Salesforce and Database Account Aggregation
home
Aggregates accounts from Salesforce and a database into a CSV file. This basic pattern can be modified to collect from more or different sources, and to produce formats other than CSV. You can trigger this manually or programmatically with an HTTP call.
Accounts are sorted such that the accounts only in Salesforce appear first, followed by accounts only in the database, and lastly by accounts found in both systems. The custom sort or merge logic can be easily modified to present the data as needed. This template also serves as a base for building APIs using the Anypoint Platform. A database table schema is included to make testing this template easier.
Aggregation templates can be easily extended to return a multitude of data in mobile friendly form to power your mobile initiatives by providing easily consumable data structures from otherwise complex backend systems.
Note: Any references in the video to DataMapper have been updated in the template with DataWeave transformations.
License Agreement
This template is subject to the conditions of the MuleSoft License Agreement.
Review the terms of the license before downloading and using this template. You can use this template for free with the Mule Enterprise Edition, CloudHub, or as a trial in Anypoint Studio.
Use Case
I want to aggregate accounts from Salesforce and database instances and compare them to see which accounts can only be found in one of the two and which accounts are in both instances.
This template has these features:
- Generates the result in a CSV report sent by email to addresses you configure.
- Extracts data from two systems, aggregates the data, compares the values of fields for the objects, and generates a report on the differences.
- Gets two accounts, one from Salesforce and other from a database instance. The template then compares by the name on the accounts, and generates a CSV file that shows accounts in A, accounts in B, and accounts in A and B. The report is emailed to a configured group of email addresses.
Considerations
To make this template run, there are certain preconditions that must be considered. All of them deal with the preparations in both, that must be made in order for all to run smoothly. Failing to do so could lead to unexpected behavior of the template.
Note: This template illustrates the aggregation use case between SalesForce and a database, thus it requires a database instance to work.
The template comes package with a MySQL script to create the DB table that uses.
It is your responsibility to use that script to create the table in an available schema and change the configuration accordingly.
The SQL script file can be found in src/main/resources/sfdc2jdbc.sql.
This template is customized for MySQL. To use it with different SQL implementation:
- Update the SQL script dialect to the desired one.
- Replace MySQL driver library dependency to desired one in the pom.xml file.
- Update the database connector configuration.
- Update the database properties in the
mule.*.properties
file.
Database Considerations
To get this template to work:
This template may use date time or timestamp fields from the database to do comparisons and take further actions.
While the template handles the time zone by sending all such fields in a neutral time zone, it cannot handle time offsets.
We define time offsets as the time difference that may surface between date time and timestamp fields from different systems due to a differences in the system's internal clock.
Take this in consideration and take the actions needed to avoid the time offset.
As a Data Destination
There are no considerations with using a database as a data destination.
Salesforce Considerations
Here's what you need to know about Salesforce to get this template to work.
FAQ
- Where can I check that the field configuration for my Salesforce instance is the right one? See: Salesforce: Checking Field Accessibility for a Particular Field
- Can I modify the Field Access Settings? How? See: Salesforce: Modifying Field Access Settings
As a Data Source
If the user who configured the template for the source system does not have at least read only permissions for the fields that are fetched, then an InvalidFieldFault API fault displays.
java.lang.RuntimeException: [InvalidFieldFault [ApiQueryFault [ApiFault
exceptionCode='INVALID_FIELD'
exceptionMessage='
Account.Phone, Account.Rating, Account.RecordTypeId, Account.ShippingCity
^
ERROR at Row:1:Column:486
No such column 'RecordTypeId' on entity 'Account'. If you are attempting to use a
custom field, be sure to append the '__c' after the custom field name. Reference
your WSDL or the describe call for the appropriate names.'
]
row='1'
column='486'
]
]
Run it!
Simple steps to get Salesforce and database Account Aggregation running.
Running On Premises
In this section we help you run your template on your computer.
Where to Download Anypoint Studio and the Mule Runtime
If you are a newcomer to Mule, here is where to get the tools.
Importing a Template into Studio
In Studio, click the Exchange X icon in the upper left of the taskbar, log in with your
Anypoint Platform credentials, search for the template, and click Open.
Running on Studio
After you import your template into Anypoint Studio, follow these steps to run it:
- Locate the properties file
mule.dev.properties
, in src/main/resources. - Complete all the properties required as per the examples in the "Properties to Configure" section.
- Right click the template project folder.
- Hover your mouse over
Run as
- Click
Mule Application (configure)
- Inside the dialog, select Environment and set the variable
mule.env
to the valuedev
- Click
Run
Running on Mule Standalone
Complete all properties in one of the property files, for example in mule.prod.properties and run your app with the corresponding environment variable. To follow the example, this is mule.env=prod
.
After this, to trigger the use case, browse to the HTTP endpoint with the port you configured in your file, for example 9090
, then browse to http://localhost:9090/generatereport
. This creates a CSV report and sends it to the email addresses you set.
Running on CloudHub
While creating your application on CloudHub (or you can do it later as a next step), go to Runtime Manager > Manage Application > Properties to set the environment variables listed in "Properties to Configure" as well as the mule.env.
After your application is running, if you chose as a domain name sfdc2db-account-aggregation
to trigger the use case, browse to http://sfdc2db-account-aggregation.cloudhub.io/generatereport
, which invokes the application and sends the report to the email addresses you configured.
Deploying your Anypoint Template on CloudHub
Studio provides an easy way to deploy your template directly to CloudHub, for the specific steps to do so check this
Properties to Configure
To use this template, configure properties (credentials, configurations, etc.) in the properties file or in CloudHub from Runtime Manager > Manage Application > Properties. The sections that follow list example values.
Application Configuration
HTTP Connector Configuration
- http.port
9090
SalesForce Connector Configuration
- sfdc.username
bob.dylan@org
- sfdc.password
DylanPassword123
- sfdc.securityToken
avsfwCUl7apQs56Xq2AKi3X
Database Connector Configuration
- db.host
localhost
- db.port
3306
- db.user
joan.baez
- db.password
JoanBaez456
- db.databasename
template-sfdc2db-user-broadcast
SMTP Services Configuration
- smtp.host
smtp.example.com
- smtp.port
587
- smtp.user
exampleuser@example.com
- smtp.password
ExamplePassword456
Email Details
- mail.from
exampleuser@example.com
- mail.to
woody.guthrie@example.com
- mail.subject
Accounts Report
- mail.body
Find attached your Accounts Report
- attachment.name
accounts_report.csv
API Calls
Salesforce imposes limits on the number of API calls that can be made. However, we make an API call to Salesforce only once during aggregation.
Customize It!
This brief guide intends to give a high level idea of how this template is built and how you can change it according to your needs.
As Mule applications are based on XML files, this page describes the XML files used with this template.
More files are available such as test classes and Mule application files, but to keep it simple, we focus on these XML files:
- config.xml
- businessLogic.xml
- endpoints.xml
- errorHandling.xml
config.xml
Configuration for connectors and configuration properties are set in this file. Even change the configuration here, all parameters that can be modified are in properties file, which is the recommended place to make your changes. However if you want to do core changes to the logic, you need to modify this file.
In the Studio visual editor, the properties are on the Global Element tab.
businessLogic.xml
The functional aspect of the template is implemented in this XML directed by a flow responsible for conducting the aggregation of data, comparing records, and finally formating the output, in this case as a report.
Using the Scatter-Gather component, we query the data in different systems. After that the aggregation implements a DataWeave 2 script using the Transform component.
Aggregated results are sorted by:
- Accounts only in Salesforce.
- Accounts only in the database.
- Accounts in both Salesforce and the database.
These are transformed to a CSV format report. The final report is sent to the email addresses you configured in the mule.*.properties file.
endpoints.xml
The file where you find the endpoint to start the aggregation. This template uses an HTTP Listener to trigger the use case.
Trigger Flow
HTTP Listener - Start Report Generation
${http.port}
is set as a property to be defined either on a property file or in CloudHub environment variables.The path configured by default is
generatereport
and you are free to change for the one you prefer.The host name for all endpoints in your CloudHub configuration should be defined as
localhost
. CloudHub routes requests from your application domain URL to the endpoint.
errorHandling.xml
This is the right place to handle how your integration reacts depending on the different exceptions.
This file provides error handling that is referenced by the main flow in the business logic.