Bulk Read Rows
Catalyst allows you to perform bulk read jobs on a specific table present in the Data Store.
In the SDK snippet below, the Bulk Read job can read thousands of records from a specific table and generate a CSV file containing the results of the read operation, if the job is successful.The table is referred to by its unique Table ID. The datastore_service reference used below is already defined in the component instance page.
Method Used | Description |
---|---|
bulk_read.create_job({ criteria, page, select_columns }) | Create a new bulk read job. |
bulk_read.get_status(job ID) | Get a bulk read job's status. |
bulk_read.get_result(job ID) | Get a bulk read job's result. |
Copy the SDK snippet below to perform a bulk read job on a particular table.
copy
#Bulk read
bulk_read = datastore_service.table("sampleTable").bulk_read()
#Create bulk read job
bulk_read_Job = bulk_read.create_job({
"criteria": {
"group_operator": 'or',
"group": [
{
"column_name": 'Department',
"comparator": 'equal',
"value": 'Marketing'
},
{
"column_name": 'EmpId',
"comparator": 'greater_than',
"value": '1000'
},
{
"column_name": 'EmpName',
"comparator": 'starts_with',
"value": 'S'
}
]
},
"page": 1,
"select_columns": ['EmpId', 'EmpName', 'Department']
})
#Get bulk read status
status = bulk_read.get_status(bulk_read_Job['job_id'])
#Get bulk read result
result = bulk_read.get_result(bulk_read_Job['job_id'])
Note: A maximum of 200,000 rows can be read simultaneously.
Last Updated 2024-09-20 18:35:48 +0530 +0530
Yes
No
Send your feedback to us
Skip
Submit