Are you looking for a way to export data from your Salesforce account? We have you covered. This article will take you through each step involved in using Salesforce Data Loader to export your data easily.
You will be looking at the following aspects:
- Introduction to Salesforce Data Loader
- What Salesforce Data Loader is Used for?
- How to Export Data Using Data Loader?
- Scheduling Exports Using Data Loader
Introduction to Salesforce Data Loader
Salesforce Data Loader Export is a very handy tool that is used by everyone that is dealing with Salesforce including an admin, developer, consultant or a data analyst. Data Loader is known to be a client application for the bulk import or export of data. We can use it to perform basic data operations including:
- insert (used to put new data into Salesforce)
- update (used to update the existing data)
- delete (used to delete the extra data or data no longer needed)
- export Salesforce records (used to export data in CSV file format)
While importing data, Data Loader needs files (containing records in the range of 50k to 5 million) in CSV format or a working database connection. For records up to 50k, you can use a Data Import Wizard.
For exporting data, Data Loader simply outputs (can be in bulk as well) the data files in CSV format. Unlike other features of Salesforce, users can install the Data Loader in a Mac or Windows environment, using either the graphical interface or command line.
Hevo- A Convenient Alternative to Export your Data
Hevo Data provides a hassle-free and a fully managed solution using its No-Code Data Pipelines. It helps you replicate the data effortlessly from Salesforce without any intervention. Hevo’s pre-built integration with Salesforce (among 100+ Sources) will take full charge of the data replication process, allowing you to focus on key business activities.
Check out what makes Hevo amazing:
- Minimal Learning: Hevo with its simple and interactive UI, is extremely simple for new customers to work on and perform operations.
- Completely Managed Platform: Hevo is fully-managed. You need not invest time and effort to maintain or monitor the infrastructure involved in executing codes.
- Data Transformation: It provides a simple interface to perfect, modify, and enrich the data you want to export.
- Secure: Hevo has a fault-tolerant architecture that ensures that the data is handled in a secure, consistent manner with zero data loss.
- Hevo Is Built To Scale: As the number of sources and the volume of your data grows, Hevo scales horizontally, handling millions of records per minute with very little latency.
- Live Support: The Hevo team is available round the clock to extend exceptional support to its customers through chat, email, and support call.
What Salesforce Data Loader is Used for?
Well, imagine, you want to upload a huge amount of data to your Salesforce environment, upto five million records or more than that (you might need the help of a Salesforce representative in that case). Usually, a normal amount of records i.e. a few hundred thousand can be uploaded using the insert function or import CSV. However, for records of about more than five million, you will need to work with a Salesforce partner or you could also visit AppExchange to choose a suitable product partner.
Apart from basic applications, Salesforce Data Loader also provides the following exciting features:
- Intelligent Data Mapping
- Scheduled Tasks
- Detailed Success and Error Logs
- Import and Export Directly from Multiple Repositories
- Support for Both Windows and Mac Computers
Intelligent Data Mapping
It lets you use shortcuts and other filters for faster searches and efficient work. The Data Loader will populate the list at the bottom of the window based on the similarity of column and field names. You can use manual field-matching as well.
One of the key options that Salesforce provides is the import/export feature. Data Loader usually uses a user interface for this manual work but there are instances when you need these tasks to be automated, for example, you want something to be imported/exported at 1:00 am every day. These scenarios can be implemented using Data Loader CLI (Command Load Interface) and Windows task scheduler. It is a very useful feature that Data Loader provides. We will discuss it in further detail later in this article.
Detailed Success and Error Logs
Data Loader has all the logs covered. The log file named sdl.log contains a detailed chronological list of the log entries. Following log entries are usually required:
INFO: These logs are required for procedural items such as login, logout, update, delete etc.
ERROR: These logs are required to check any type of errors popping up e.g. missing values etc.
The log file sdl.log can be viewed by entering %TEMP%sdl.log in either the Run dialog box or Windows explorer address bar. In case of Mac, you need to view log file using the terminal, enter and open $TMPDIR/sdl.log.
Import and Export Directly from Multiple Repositories
Using Salesforce Data Loader, all team members can export and import data from SFTP, Dropbox, FTP etc.
Support for Windows and Mac
This feature is helpful especially when companies switch from Windows to Mac. The compatibility of Data Loader with both makes it easy for companies to make a switch and they don’t have to worry about Data Loader support.
Let’s move into some further details.
How to Export Data Using Data Loader?
It is a simple process to export data using Salesforce Data Loader. Following steps are required to achieve this functionality:
- For Windows users, search Windows and open Data Loader. For Mac users, search Applications and open Data Loader.
- Click Export. For archived activity and deleted records, you should click Export All.
- Select Password Authentication.
- Enter your username and password by using https://login.salesforce.com for the Salesforce login.
- Once verified, click Next.
- Choose an object, for example, from the “Select Salesforce Object” list, select Account (Account).
Note: If the object is not listed, select show all objects to see all the objects that can be accessed. For object description, see the SOAP API Developer Guide.
- Choose a target for the extraction text box, e.g. Accounts Export.csv and click Browse.
- Navigate to the desktop, click Save and then click Next.
- The next step is to create a required SOQL by using the query input fields. For example, select Id and Name in the query fields and click Finish. We have discussed SOQL in detail in our previous article and how one can write different queries in SOQL. After you click Finish, the operations window appears, mentioning the number of completed extractions.
- Next, export the extracted records into the CSV file. To view it in excel, click on View Extraction. In CSV viewer, click open on the external program ( this is for Windows users). The file should open automatically in excel.
Note: Data Loader does not support exporting attachments. As an alternate method, a weekly export feature in an online application can be used to do the same. Also, if you select compound fields for export in the Data Loader, it will result in error messages. To export values, use individual field components.
Scheduling Exports Using Data Loader
Let’s briefly discuss the scheduling functionality of Data Loader. For automated batch operations, such as scheduled loads or extractions, we need to run Data Loader from the Windows command-line. Please note that the Data Loader command-line interface is supported for Windows only.
You need to include an encrypted password in the configuration file to run a batch operation.
From the command line, navigate to the bin directory and open the process.bat file. Following parameters are needed to pass to the file.
- config.properties file contained in the directory.
- process-conf.xml file containing the name of the batch process. process-conf.xml file is used to configure batch processing.
We need to set the name of the process in the bean element’s id attribute: (e.g. <bean id=”myProcessName”>).
Parameters can be changed at run time by giving param=value as program arguments.
E.g. Adding process.operation=insert to the command will change the configuration during run time.
The process.bat file supports running JAR with up to two parameters. The format for two parameters is:
java -cp ..dataloader-xx.0.0-uber.jar -Dsalesforce.config.dir=<configdir> com.salesforce.dataloader.process.ProcessRunner <operation> parameterName1=ParemeterValue1 parameterName2=ParemeterValue2
The purpose of this part is to just give you an introductory overview of the scheduling process. The process of automation does not end here. This is just an initial setup. For more details, you can visit this page.
When we talk about the limitations of Data Loader, one that really stands out is the limitation to be able to upload only 5000 to 5,000,000 records. One can perform insert/update/delete operations, essentially on an unlimited number of rows, although limited by the lowest of the following conditions: the number of API calls per day, Salesforce.com storage limitations and local storage device’s available capacity.
The purpose of this blog post was to give you an idea about Data Loader’s basic functionalities, how some tasks can be scheduled or automated and also what are some of its limitations. Some technical aspects were discussed to describe the possibilities of functions that can be performed using this amazing tool.
Do you want to share your thoughts? Have any other queries? Get in touch with us in the comments section below.