Hi,

I am tryng to read a file from HDFS in Hadoop from Pentaho DI and I am getting problems:

- Pentaho DI (open source) in a local machine Win 7 version 6.1
- HDFS in a Virtual Machine Cloudera Quick Start 5.4

data-integration\plugins\pentaho-big-data-plugin\hadoop-configurations
1. core-site.xml
2. hbase-site.xml
3. hdfs-site.xml
4. hive-site.xml
5. mapred-site.xml
6. yarn-site.xml

I din't have hdfs-site.xml and copied it from MV the directory: /etc/hive/conf

I have indicated my Hadoop Distribution (cdh55) in Tools, I have created a cluster in 'view', when I test the cluster it works properly.
Now I use Hadoop File Input step, I can set the cluster and find the file,but there is a problem in read process:

at org.pentaho.commons.launcher.Launcher.main (Launcher.java:92)
at java.lang.reflect.Method.invoke (null:-1)
at sun.reflect.DelegatingMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke0 (null:-2)
at org.pentaho.di.ui.spoon.Spoon.main (Spoon.java:662)
at org.pentaho.di.ui.spoon.Spoon.start (Spoon.java:9269)
at org.pentaho.di.ui.spoon.Spoon.waitForDispose (Spoon.java:7989)
at org.pentaho.di.ui.spoon.Spoon.readAndDispatch (Spoon.java:1347)
at org.eclipse.swt.widgets.Display.readAndDispatch (null:-1)
at org.eclipse.swt.widgets.Display.runDeferredEvents (null:-1)
at org.eclipse.swt.widgets.Widget.sendEvent (null:-1)
at org.eclipse.swt.widgets.EventTable.sendEvent (null:-1)
at org.eclipse.jface.action.ActionContributionItem$5.handleEvent (ActionContributionItem.java:402)
at org.eclipse.jface.action.ActionContributionItem.access$2 (ActionContributionItem.java:490)
at org.eclipse.jface.action.ActionContributionItem.handleWidgetSelection (ActionContributionItem.java:545)
at org.eclipse.jface.action.Action.runWithEvent (Action.java:498)
at org.pentaho.ui.xul.jface.tags.JfaceMenuitem$1.run (JfaceMenuitem.java:106)
at org.pentaho.ui.xul.jface.tags.JfaceMenuitem.access$100 (JfaceMenuitem.java:43)
at org.pentaho.ui.xul.impl.AbstractXulComponent.invoke (AbstractXulComponent.java:141)
at org.pentaho.ui.xul.impl.AbstractXulComponent.invoke (AbstractXulComponent.java:157)
at org.pentaho.ui.xul.impl.AbstractXulDomContainer.invoke (AbstractXulDomContainer.java:313)
at java.lang.reflect.Method.invoke (null:-1)
at sun.reflect.DelegatingMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke0 (null:-2)
at org.pentaho.di.ui.spoon.trans.TransGraph.editStep (TransGraph.java:2129)
at org.pentaho.di.ui.spoon.trans.TransGraph.editStep (TransGraph.java:3072)
at org.pentaho.di.ui.spoon.Spoon.editStep (Spoon.java:8783)
at org.pentaho.di.ui.spoon.delegates.SpoonStepsDelegate.editStep (SpoonStepsDelegate.java:125)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.open (HadoopFileInputDialog.java:575)
at org.eclipse.swt.widgets.Display.readAndDispatch (null:-1)
at org.eclipse.swt.widgets.Display.runDeferredEvents (null:-1)
at org.eclipse.swt.widgets.Widget.sendEvent (null:-1)
at org.eclipse.swt.widgets.EventTable.sendEvent (null:-1)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog$3.handleEvent (HadoopFileInputDialog.java:482)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.access$200 (HadoopFileInputDialog.java:125)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.first (HadoopFileInputDialog.java:2634)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.getFirst (HadoopFileInputDialog.java:2722)
at org.pentaho.di.trans.steps.textfileinput.TextFileInput.getLine (TextFileInput.java:97)
at org.pentaho.di.trans.steps.textfileinput.TextFileInput.getLine (TextFileInput.java:127)
at java.io.InputStreamReader.read (null:-1)
at sun.nio.cs.StreamDecoder.read (null:-1)
at sun.nio.cs.StreamDecoder.read0 (null:-1)
at sun.nio.cs.StreamDecoder.read (null:-1)
at sun.nio.cs.StreamDecoder.implRead (null:-1)
at sun.nio.cs.StreamDecoder.readBytes (null:-1)
at org.pentaho.di.core.compress.CompressionInputStream.read (CompressionInputStream.java:68)
at org.apache.commons.vfs2.util.MonitorInputStream.read (MonitorInputStream.java:99)
at java.io.BufferedInputStream.read (null:-1)
at java.io.BufferedInputStream.read1 (null:-1)
at java.io.DataInputStream.read (null:-1)
at org.apache.hadoop.hdfs.DFSInputStream.read (DFSInputStream.java:903)
at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy (DFSInputStream.java:851)
at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo (DFSInputStream.java:624)
at org.apache.hadoop.hdfs.BlockReaderFactory.build (BlockReaderFactory.java:374)
at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp (BlockReaderFactory.java:753)
at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer (BlockReaderFactory.java:838)
at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer (DFSClient.java:3492)
at org.apache.hadoop.net.NetUtils.connect (NetUtils.java:530)
at org.apache.hadoop.net.SocketIOWithTimeout.connect (SocketIOWithTimeout.java:192)
at sun.nio.ch.SocketChannelImpl.connect (null:-1)
at sun.nio.ch.Net.checkAddress (null:-1)


at org.pentaho.commons.launcher.Launcher.main (Launcher.java:92)
at java.lang.reflect.Method.invoke (null:-1)
at sun.reflect.DelegatingMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke0 (null:-2)
at org.pentaho.di.ui.spoon.Spoon.main (Spoon.java:662)
at org.pentaho.di.ui.spoon.Spoon.start (Spoon.java:9269)
at org.pentaho.di.ui.spoon.Spoon.waitForDispose (Spoon.java:7989)
at org.pentaho.di.ui.spoon.Spoon.readAndDispatch (Spoon.java:1347)
at org.eclipse.swt.widgets.Display.readAndDispatch (null:-1)
at org.eclipse.swt.widgets.Display.runDeferredEvents (null:-1)
at org.eclipse.swt.widgets.Widget.sendEvent (null:-1)
at org.eclipse.swt.widgets.EventTable.sendEvent (null:-1)
at org.eclipse.jface.action.ActionContributionItem$5.handleEvent (ActionContributionItem.java:402)
at org.eclipse.jface.action.ActionContributionItem.access$2 (ActionContributionItem.java:490)
at org.eclipse.jface.action.ActionContributionItem.handleWidgetSelection (ActionContributionItem.java:545)
at org.eclipse.jface.action.Action.runWithEvent (Action.java:498)
at org.pentaho.ui.xul.jface.tags.JfaceMenuitem$1.run (JfaceMenuitem.java:106)
at org.pentaho.ui.xul.jface.tags.JfaceMenuitem.access$100 (JfaceMenuitem.java:43)
at org.pentaho.ui.xul.impl.AbstractXulComponent.invoke (AbstractXulComponent.java:141)
at org.pentaho.ui.xul.impl.AbstractXulComponent.invoke (AbstractXulComponent.java:157)
at org.pentaho.ui.xul.impl.AbstractXulDomContainer.invoke (AbstractXulDomContainer.java:313)
at java.lang.reflect.Method.invoke (null:-1)
at sun.reflect.DelegatingMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke (null:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke0 (null:-2)
at org.pentaho.di.ui.spoon.trans.TransGraph.editStep (TransGraph.java:2129)
at org.pentaho.di.ui.spoon.trans.TransGraph.editStep (TransGraph.java:3072)
at org.pentaho.di.ui.spoon.Spoon.editStep (Spoon.java:8783)
at org.pentaho.di.ui.spoon.delegates.SpoonStepsDelegate.editStep (SpoonStepsDelegate.java:125)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.open (HadoopFileInputDialog.java:575)
at org.eclipse.swt.widgets.Display.readAndDispatch (null:-1)
at org.eclipse.swt.widgets.Display.runDeferredEvents (null:-1)
at org.eclipse.swt.widgets.Widget.sendEvent (null:-1)
at org.eclipse.swt.widgets.EventTable.sendEvent (null:-1)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog$3.handleEvent (HadoopFileInputDialog.java:482)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.access$200 (HadoopFileInputDialog.java:125)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.first (HadoopFileInputDialog.java:2634)
at org.pentaho.big.data.kettle.plugins.hdfs.trans.HadoopFileInputDialog.getFirst (HadoopFileInputDialog.java:2722)
at org.pentaho.di.trans.steps.textfileinput.TextFileInput.getLine (TextFileInput.java:97)
at org.pentaho.di.trans.steps.textfileinput.TextFileInput.getLine (TextFileInput.java:127)
at java.io.InputStreamReader.read (null:-1)
at sun.nio.cs.StreamDecoder.read (null:-1)
at sun.nio.cs.StreamDecoder.read0 (null:-1)
at sun.nio.cs.StreamDecoder.read (null:-1)
at sun.nio.cs.StreamDecoder.implRead (null:-1)
at sun.nio.cs.StreamDecoder.readBytes (null:-1)
at org.pentaho.di.core.compress.CompressionInputStream.read (CompressionInputStream.java:68)
at org.apache.commons.vfs2.util.MonitorInputStream.read (MonitorInputStream.java:99)
at java.io.BufferedInputStream.read (null:-1)
at java.io.BufferedInputStream.read1 (null:-1)
at java.io.DataInputStream.read (null:-1)
at org.apache.hadoop.hdfs.DFSInputStream.read (DFSInputStream.java:903)
at org.apache.hadoop.hdfs.DFSInputStream.readWithStrategy (DFSInputStream.java:851)
at org.apache.hadoop.hdfs.DFSInputStream.blockSeekTo (DFSInputStream.java:624)
at org.apache.hadoop.hdfs.BlockReaderFactory.build (BlockReaderFactory.java:374)
at org.apache.hadoop.hdfs.BlockReaderFactory.getRemoteBlockReaderFromTcp (BlockReaderFactory.java:753)
at org.apache.hadoop.hdfs.BlockReaderFactory.nextTcpPeer (BlockReaderFactory.java:838)
at org.apache.hadoop.hdfs.DFSClient.newConnectedPeer (DFSClient.java:3492)
at org.apache.hadoop.net.NetUtils.connect (NetUtils.java:530)
at org.apache.hadoop.net.SocketIOWithTimeout.connect (SocketIOWithTimeout.java:192)
at sun.nio.ch.SocketChannelImpl.connect (null:-1)
at sun.nio.ch.Net.checkAddress (null:-1)


I can Access to the Hive databse in te same Cloudera MV using jdbc for Hive in ‘generic database’ connection.

Can anybody hepl me?, is there any error in my process? Any advice will be greatly apreciatted.