I'm using Pentaho Data Integration to create a transformation from xlsx files to mysql, but I can't import data from large files with Excel 2007 xlsx(apache POI Straiming)
. It gives me out of memory errors.
相关问题
- Error setting up initial repository in Pentaho Ket
- Error connecting to database: (using class org.gjt
- Filling data gaps in a stream in Pentaho Data Inte
- Limit no. of rows in mongodb input
- java调用kettle,main方法不报错,tomcat下运行报错,但不影响kettle的
相关文章
- java调用kettle,main方法不报错,tomcat下运行报错,但不影响kettle的
- Pentaho DI - JSON Nested File Output
- Running pentaho spoon UI in intellij
- Problems connecting Pentaho Kettle/Spoon to Heroku
- 从BI服务器ETL JAR执行(ETL JAR execution from BI server)
- 使用的pentaho釜Excel工作表的动态命名(Dynamic naming of excel s
- Pentaho的帧大小(17727647)比最大长度(16384000)更大!(Pentaho Fr
- 如何在不断线程读取JMS消息,并根据其在另一个线程JMSMessageID按achnowledge呢
Did you try this option ?
Advanced settings -> Generation mode -> Less memory consumed for large excel(Event mode
(You need to check "Read excel2007 file format" first)
I would recommend you to increase jvm memory allocation before running the transformation. By default, pentaho data integration aka kettle comes with low memory allocation which would cause issues with running ETLs involving large files. You would need to modify the
-Xmx
value so that it specifies a larger upper memory limit in spoon.bat accordingly.If you are using spoon in windows and edit spoon.bat in the line show below.
If you are using kitchen or pan, edit in those pan.bat or kitchen.bat accordingly. If you are using in linux, change in .sh files.