Hitachi Vantara Pentaho Community Forums
Results 1 to 3 of 3

Thread: Issue While connecting to Hortonworks Sandbox 2.0 From Pentaho Trial Edition 5.0.2

  1. #1
    Join Date
    Jan 2014
    Posts
    3

    Unhappy Issue While connecting to Hortonworks Sandbox 2.0 From Pentaho Trial Edition 5.0.2

    Hello All,
    I have just started evaluating the Pentoho for our Business Integration purpose. We are having a Hortonworks Sandbox 2.0 Environment which is up and running successfully.

    I Created a job to copy a file from local disk to HDFS. When I run the job, it's failing.

    The logger prints the below exception.

    2014/01/08 11:32:24 - Hadoop Copy Files - ERROR (version 5.0.2, build 1 from 2013-12-04_15-52-25 by buildguy) : Can not copy file/folder [file:///E:/Data/FifteenGigaText.txt] to [hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test]. Exception : [
    2014/01/08 11:32:24 - Hadoop Copy Files -
    2014/01/08 11:32:24 - Hadoop Copy Files - Unable to get VFS File object for filename 'hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test' : Could not resolve file "hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test".
    2014/01/08 11:32:24 - Hadoop Copy Files -
    2014/01/08 11:32:24 - Hadoop Copy Files - ]
    2014/01/08 11:32:24 - Hadoop Copy Files - ERROR (version 5.0.2, build 1 from 2013-12-04_15-52-25 by buildguy) : org.pentaho.di.core.exception.KettleFileException:
    2014/01/08 11:32:24 - Hadoop Copy Files -
    2014/01/08 11:32:24 - Hadoop Copy Files - Unable to get VFS File object for filename 'hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test' : Could not resolve file "hdfs://[[sandbox environment ip goes here]]:8020/user/hue/test".
    2014/01/08 11:32:24 - Hadoop Copy Files -
    2014/01/08 11:32:24 - Hadoop Copy Files -
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.core.vfs.KettleVFS.getFileObject(KettleVFS.java:162)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.core.vfs.KettleVFS.getFileObject(KettleVFS.java:105)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.entries.copyfiles.JobEntryCopyFiles.ProcessFileFolder(JobEntryCopyFiles.java:378)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.entries.copyfiles.JobEntryCopyFiles.execute(JobEntryCopyFiles.java:326)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.execute(Job.java:678)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.execute(Job.java:815)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.execute(Job.java:500)
    2014/01/08 11:32:24 - Hadoop Copy Files - at org.pentaho.di.job.Job.run(Job.java:407)
    2014/01/08 11:32:24 - HadoopFileCopy - Finished job entry [Hadoop Copy Files] (result=[false])
    2014/01/08 11:32:24 - HadoopFileCopy - Job execution finished
    2014/01/08 11:32:24 - Spoon - Job has ended.

    When i try to google it up, I landed on the page where pentaho talked about the various hadoop distros configuration and compatibility with PDI.

    http://wiki.pentaho.com/display/BAD/...ro+and+Version

    In this page, it's clearly stated that HDP 2.x is not supported by Pentaho unless we make some manual effort.

    Update : I found a Jira http://jira.pentaho.com/browse/PDI-10807 Ticket available for the same reason what I 've asked here for.... The ticket status states it is closed. But no idea what it talks , and what i have to do......

    Can Any one guide me through this to get it done.

    Thanks In Advance

    Regards,
    VAP.
    Last edited by VAP; 01-08-2014 at 05:45 AM. Reason: Found Jira regarding the same.

  2. #2
    Join Date
    Sep 2012
    Posts
    71

    Default

    The HDP 2.0 configuration (aka "shim") is not yet released (see the wiki page from your post, the associated Jira case has not yet been validated/closed). The subtask for the initial creation (without validation) is PDI-10807, and although the sub-task is closed, the shim is not considered released until it is available in either a Pentaho suite release, Service Pack, the Support Portal, or the PDI Marketplace.

    Having said that, it should be coming shortly, stay tuned! If you are an EE customer, please contact Support, they may be able to get you an evaluation version. If you are using the Community Edition, you can find the latest (unsupported) version on our Continuous Integration server at http://ci.pentaho.com/view/Big%20Dat...K-SNAPSHOT.zip
    Last edited by mattb_pdi; 01-08-2014 at 11:35 PM.

  3. #3
    Join Date
    Jan 2014
    Posts
    3

    Red face Resolved....

    Quote Originally Posted by mattb_pdi View Post
    The HDP 2.0 configuration (aka "shim") is not yet released (see the wiki page from your post, the associated Jira case has not yet been validated/closed). The subtask for the initial creation (without validation) is PDI-10807, and although the sub-task is closed, the shim is not considered released until it is available in either a Pentaho suite release, Service Pack, the Support Portal, or the PDI Marketplace.

    Having said that, it should be coming shortly, stay tuned! If you are an EE customer, please contact Support, they may be able to get you an evaluation version. If you are using the Community Edition, you can find the latest (unsupported) version on our Continuous Integration server at http://ci.pentaho.com/view/Big%20Dat...K-SNAPSHOT.zip
    Thanks for the reply matt.

    Actaully I ended up creating the zip from the source code available on GIT , and it worked well though....

    But it has taken 2 hours to complete the entire build process.

    Anyway day went good. Thanks once again.

    Regards,
    VAP.

Tags for this Thread

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •  
Privacy Policy | Legal Notices | Safe Harbor Privacy Policy

Copyright © 2005 - 2019 Hitachi Vantara Corporation. All Rights Reserved.