Historic event data can be downloaded from your event archive.
A complete archive of your event data is maintained on Amazon S3. It is updated overnight with the events from the previous day and contains separate folders for your DEV and LIVE environments, each contains a compressed CSV files for each days’ events.
Events will not necessarily be stored in sequential order inside an events archive file, the next event in a sequence may be in a different file.
You will find the login credentials and folder name for your Amazon S3 folder on the Game Details page for your game:
There are numerous free and paid tools that can be used to connect to your Amazon S3 folder and download your archived data. A couple of the more popular tools are S3 Browser and S3 Command.
Download S3 Browser from http://s3browser.com/
To download using S3 browser:
Add a new account using the Access Key ID and Secret Access Key provided on your Game Details page:
Add an External Bucket using the Buckets > Add External Bucket option and paste in the bucket and S3 Folder name provided on your Game Details page:
If you fail to do this you will get a permissions error as connecting to the root of the deltaDNA archive is not permitted.
Connect to your Archive folders (notice the split between DEV and LIVE):
Select the file(s) you are interested in and download click download.
There is a GZIP compressed CSV for each day, for both DEV and LIVE environments:
S3 Command is a command line tool that can be downloaded from http://s3tools.org/s3cmd. This may be preferable if you wish to download your archives from the command line, or use a script to download archives automatically.
There are numerous other tools, for multiple platforms, available online for accessing and downloading S3 archives.
The platform performs data enrichment and processing as part of a process when an event received. This process must be completed before an event can be exported into an archive file for S3.
As well as the processing time, the archive process runs continually, and cannot start a new job until the last one is finished.
During peak times this can translate as a delays between 3 and 6 hours being the norm for the archive’s cadence.
The following query can be run in data mining to give an insight on the currency latency between events being received, and their processing and readiness for export into archival dumps.
select eventDate - current_date as Days_ago,
date_part('hour', eventTimestamp) as hour,
where eventDate >= current_date-5
group by 1,2
order by 1,2