Large dataset processing using Mule ESB from datab

2019-09-15 04:28发布

问题:

I have a large set of records to be processed eg: 100,000 records. My use case has 4 steps:

  1. pick the records from database table using jdbc inbound adapter
  2. convert the record to xml format
  3. post the message to the queue
  4. then update the same record with some status flag as it has been processed so that it will not be picked again

I don't want to pick all the records from table at one shot for processing: is there a way so that pick in some batches and don't want to update table for one record? Is there any option of bulk/batch update at batch level?

Or is there any other better way to approach this use case? Any suggestions highly appreciated.

回答1:

I would write the SQL select query to return only N records (like LIMIT 100 or equivalent) with a where clause that excludes the already processed records based on the status flag.



标签: jdbc mule esb