You need to sign in to do that
Don't have an account?

how to download EventLogFile
I need to create a user report that shows user login, logout, duration of login for each instance.
Can I use EventLogfile to get this information, if so How can I access to this EventLogFile?
I run a query against this object, and saw it was a txt/csv file.
How can I download it from workbench?
This link from trailhead is about EventLogFIle but did not provide technical guide to download the file.
https://trailhead.salesforce.com/en/modules/event_monitoring/units/event_monitoring_intro
Can I use EventLogfile to get this information, if so How can I access to this EventLogFile?
I run a query against this object, and saw it was a txt/csv file.
How can I download it from workbench?
This link from trailhead is about EventLogFIle but did not provide technical guide to download the file.
https://trailhead.salesforce.com/en/modules/event_monitoring/units/event_monitoring_intro
You can do this using scripting.
Event Monitoring, new in the Winter '15 release, enables use cases like adoption, user audit, troubleshooting, and performance profiling using an easy to download, file based API to extract Salesforce app log data.
The most important part is making it easy to download the data so that you can integrate it with your analytics platform.
To help make it easy, I created a simple bash shell script to download these CSV (comma separated value) files to your local drive. It works best on Mac and Linux but can be made to work with Windows with a little elbow grease. You can try these scripts out at http://bit.ly/elfScripts. These scripts do require a separate JSON library called jqto parse the JSON that's returned by the REST API.
It's not difficult to build these scripts using other languages such as Ruby, Perl, or Python. What's important is the data flow.
I prompt the user to enter their username and password (which is masked). This information can just as easily be stored in environment variables or encrypted so that you can automate the download on a daily basis using CRON or launchd schedulers. Once we have the credentials, we can log in using oAuth and get the access token. Then we can query the event log files to get the Ids necessary to download the files and store the event type and log date in order to properly name the download directory and files. Using jq, we can parse the id, event type, and date in order to create the directory and file names We create the directories to store the files. In this case, we download the raw data and then convert the timestamp to something our analytics platform will understand better.
Then we can iterate through each download, renaming it to the Event Type + Log Date so that we easily refer back to it later on. I also transform the Timestamp field to make it easier to import into an analytics platform like Project Wave from Salesforce Analytics Cloud. Downloading event log files is quick and efficient. You can try these scripts out at http://bit.ly/elfScripts.
Give it a try!!!!!!!!!!!!!
Hope this helps.
Kindly mark this as solved if the reply was helpful.
Thanks,
Nagendra