PDA

View Full Version : Connecting to mapR haoop cluster using PDI



giri.sgs143
06-30-2013, 02:32 PM
Hello Friends,

i have one mapR hadoop cluster, i would like to connect and process the files present in HDFS.

when i try read some file in HDFS, i am facing the below exception.

2013/06/30 23:38:05 - FileInputList - ERROR (version 4.4.1-GA, build 18329 from 2013-04-17 12.31.49 by buildguy) : org.pentaho.di.core.exception.KettleFileException:
2013/06/30 23:38:05 - FileInputList - ERROR (version 4.4.1-GA, build 18329 from 2013-04-17 12.31.49 by buildguy) :
2013/06/30 23:38:05 - FileInputList - ERROR (version 4.4.1-GA, build 18329 from 2013-04-17 12.31.49 by buildguy) : Unable to get VFS File object for filename 'maprfs://192.168.233.128:7221/user/hello.txt' : Could not resolve file "maprfs://192.168.233.128:7221/user/hello.txt".
2013/06/30 23:38:05 - FileInputList - ERROR (version 4.4.1-GA, build 18329 from 2013-04-17 12.31.49 by buildguy) :
2013/06/30 23:38:05 - FileInputList - ERROR (version 4.4.1-GA, build 18329 from 2013-04-17 12.31.49 by buildguy) : at org.pentaho.di.core.vfs.KettleVFS.getFileObject(KettleVFS.java:161)

can someone please suggest me what might be the issue ?

Regards,
Giri

mattb_pdi
06-30-2013, 10:23 PM
Is your Hadoop configuration set to "mapr"? To check (and change if necessary), edit data-integration/plugins/pentaho-big-data-plugin/plugin.properties and make sure the "active.hadoop.configuration" property is set to "mapr". If you have to change it, restart Spoon.

If this doesn't fix your issue, please paste in the full stack trace and I will take a look :)