batch processing - Take extremely long time in kettle pentaho job -


good day,

i have kettle pentaho file run batch job.

basically, files contain of 2 main steps, first step, read input file (txt file) , store inside table1. second step, same first step, read same input file , store inside table2.

this batch working fine until put in 20mb input file.it require more 7hours finish job.

below test case have done:

15360 records, 1.4mb, 2 minutes , 20 seconds (140 seconds total). 30720 records, 2.8mb , 7 minutes , 30 seconds (450 seconds total) 61440 records, 5.5mb, 26 minutes , 55 seconds (1615 seconds total). 250000 records, 20mb, 7 hours , 30 minutes 

in log, found there steps occupied of time consuming. follow: 1. text file input. 2. select values. 3. modified java script value. enter image description here

both main steps contain 3 kettle pentaho function. 20mb input file, first step take around 7 minutes, second step take more 7 hours.

try @ in quite long time, still cant find out problem.

kindly advise.

there might multiple reasons (i assume). first of all, try optimize steps "select values" , "modified javascript". of performance tuning tips given in here.

also may try increase java memory in pan.sh. check image below:

enter image description here

change javamaxmem somevalue higher 1024.

hope these changes might :)


Comments

Popular posts from this blog

apache - PHP Soap issue while content length is larger -

asynchronous - Python asyncio task got bad yield -

javascript - Complete OpenIDConnect auth when requesting via Ajax -