Data export
Last updated
Was this helpful?
Last updated
Was this helpful?
The Data Export Service automatically exports data related to your operations into a standard format at a chosen interval and stores it in a configured external location or provides through one email notification a temporary link to access it.
The data export is designed for BI purposes, to export your data in operational applications or to trigger processes with a near real time latency, contact your implementation manager or customer support manager.
To enable the data export you should notify it to your customer success manager and with them choose the right configuration for your export. After finalizing your configuration, we will validate and deploy the solution for you.
Base configuration parameters
site_id
List of uuids
A list of one or more site identifiers for which you want to export data.
export_type
work-orders-states
or work-orders-timeline
Indicates the type of data to export.
file_format
json
The output format for the exported data files.
Note: we only support Jsonline.
interval
once, daily, weekly, monthly, ..
The frequency at which the export job will run.
connector_type
email
/ cloud_storage
connector configuration
Depending on the connector
You can have a one shot export or configure a regular data export with a defined frequency.
Daily
Weekly
Monthly
The export will be generated and sent at the chosen interval.@
The connector_type section determines how and where your exported data is delivered. Two primary connector types are supported:
Email with temporary link
Your cloud bucket
Below are the configuration details that differ based on your chosen connector type.
This connector allows you to receive at your email address one temporary link to the structured data export file data export generated.
Email connector Config
emails
List of strings (valid email addresses)
Recipients to be notified when data is available.
This connector allows you to receive directly into your cloud bucket the data export you decided, both the raw data and the images.
Cloud Storage connector Config
storage_uri
String (e.g., gs://customer-cloud-bucket-export/folder/customer-external/
)
Prefixed URI of the storage location where the exported files will be saved on the customer cloud.
Note: Must be a valid Google Cloud Storage (GCS) bucket path. Example: gs://my-bucket/deepomatic-data
export_images
True
/ False
Determines whether images associated with the exported data should also be included in the export.
storage_uri
We already have a separation between different sites, but we advise to have one extra layer to separate different environments.
Separate the production / testing sites in different folders
gs://customer-cloud-bucket-export/deepomatic/testing/
gs://customer-cloud-bucket-export/deepomatic/production/
export_images
When enabled we will transfer images related to your export_type from our infrastructure to yours. (See image export section)
GCP setup:
Given that we will transfer the data from our infrastructure to yours, to enable the Deepomatic GCP Cloud Storage Connector:
We will provide you one Deepomatic Principal email (a Deepomatic GCP Service Account):
customer-<customer_name>-data-export@deepomatic-160015.iam.gserviceaccount.com
To this GSA you will need to assign these permissions on the bucket specified in the external_storage_uri:
storage.objects.create
storage.objects.list
storage.objects.get
storage.buckets.get
Structure of the data in the target bucket
The new data exports and images will be added to your bucket following this structure :
TBD - WIP
TBD - WIP
Each generated file contains structured data on all the work orders created or updated during the last interval time frame, including all the analysis performed on them.
This format can be used for:
book keeping
field ops analysis, AI and FTR analysis
field operator experience analysis
The work-order-timeline file is in the jsonline format.
Work order level fields
id
UUID (string)
Unique identifier of the work order
name
string
Descriptive name of the work order
site_id
UUID (string)
Identifier of the related site
create_date
ISO 8601 string
Creation timestamp with timezone
update_date
ISO 8601 string
Last updated timestamp with timezone
metadata
object
Additional information about the work order
progress_score
number
Optional progress indicator (e.g., between 0 and 1).
work_items
array
List of items to be processed. Each contains an id
, name
, and input
.
A work item relates to one taskgroup and has the same name
Work item
id
UUID (string)
Identifier of the related work_item
name
string
Name of the work item
input
array
List of input, images, and their analysis results
input
id
UUID (string)
Input identifier (image id)
metadata
object
May contain metadata of the image like geolocation or timestamp.
data_conformity
array of objects
List of conformity checks on raw input data. Can be empty.
job_conformity
array of objects
List of job-specific conformity checks. Can be empty.
data_conformity fields
check
string
The data conformity check.
code
string
One code per data conformity check.
level
string
Error, information or warning. Error is blocking, Warning and information are not blocking.
job_conformity fields
task_name
string
Name of the task being validated.
analyzed_value
string / number / bool
Output of the automatic analysis (optional).
corrected_value
string / number / bool
Manually corrected value (optional).
corrected_date
ISO 8601 string
When correction was made (nullable).
message
string
Explanation, error, or note about the result.
is_conformity_validated
boolean
if this task is considered validate.
This format is the same as the timeline data format, except that to simplify the analysis, there is only one item in the input array. The input exported is always the last one created.
The images analyzed during each interval timeframe will be uploaded into the customer target cloud storage.
The name of each image is composed of the input uuid and the file extension. Both information can be found in the data file in the fields input
and file_extension
.
In the data file, within each input
you have the id
and file_extension
fields. The absolute path to the images is therefore:
See The timeline format is the default one.
See
Determines how the structured data and images will be sent to you. See
See