I need to pull record from Oracle db and push to MS SQL server using job scheduling.
can you please any one explain how to create job scheduling in Pentaho ETL 8.1 Community Edition
You can use kitchen to run a job. Create a batch file that call kitchen and that runs the job. Then you can schedule with a cron job or a Windows task
Thank you Mr.johan
Hi Premkumar Manipillai,
There is two ways to create a job scheduling in Pentaho Data Integration.
echo "NAME OF JOB $fecha"
/home/pdi-7.1/data-integration/./kitchen.sh -file="/PATH_TO_JOB/NAME_OF_JOB.kjb" -level="Basic" -logfile="PATH_TO_LOG/NAME_OF_JOB_$date.log"
If you need an explanation on Windows, plase let me know
Thank you Mr.Ricardo,
I am using Pentaho ETL 8.0 community Edition in windows machine and job scheduling running using .bat file successfully :-)
Now i need to run the same in jBoss server. For this, what i have to do.
My doubt is:
1.Need to place full pentaho "data-integration" tool folder(1.33 GB) in jBoss server?.
2.We can schedule the batch file using windows scheduler form jBoss server ?
can you please guide me.
Thank you very much!
We can schedule the batch file using windows scheduler
Hi again Premkumar ,
My first question..... why in jboss server? you simply can use this without a server (PDI works by itself)
2.- you can schendule without Jboss Server
You can open a terminal (Windows+R) and execute the command taskschd.msc
i think i asked question wrongly.
My requirement is : Need to run this same "oracletosql.bat" file from online server (not jBoss).
For this, what are the steps i have to do. Please guide me.
what are the supporting files i have to upload to server to run this?
should i upload "data-integration" folder to server?
Thanks for your support & Sorry for disturb to you :-)
Yes, the easiest way is to copy the data integration folder to the server.
Thank you very much Mr.Johan
Johan Hammink Is right!
Yes Mr.Ricardo, i did the same.Its working good and i tested upto 1 lakhs record at a time, fetch from oracle and insert into ms sql server.
Now i am facing another problem,
i need to handle huge(e.g. 50 lakhs+) fetch from oracle and insert/update to ms sql server. So, How to handle huge record in pentaho?
Please help me Mr.Ricardo
Create a new question when the question is not about "How to schedule a job"
There are a lot of things to improve performance.
- a table output step is much faster then a insert/update
- You can change the number of copies to start of the output step
- Perhaps it is faster first to dump the Oracle table to a csv file and then use a Bulk Insert into the MsSQL
Retrieving data ...