>> It's better if you explicitly specify on the command line the ZK entry
point.

Can you give an example?

On Thu, Feb 1, 2018 at 12:54 PM, Christopher <ctubb...@apache.org> wrote:

> Normally, you'd set up Accumulo to use the HDFS volume in your
> accumulo-site.xml file for your servers by setting the instance.volumes
> field (in your case to the value of 'hdfs://haz0-m:8020/accumulo' or
> similar).
>
> The shell typically connects to either ZooKeeper using client
> configuration files or command-line options as its entry point. Run it with
> the '--help' or '-?' options to see the available options.
>
> If it has read permission for your accumulo-site.xml file and the Accumulo
> conf directory where this file is located is on its class path, the shell
> may fall back to using your hdfs-site.xml or your accumulo-site.xml to try
> to figure out things using HDFS... but that's mostly a
> backwards-compatible/legacy mode. It's better if you explicitly specify on
> the command line the ZK entry point.
>
> On Thu, Feb 1, 2018 at 10:59 AM Geoffry Roberts <threadedb...@gmail.com>
> wrote:
>
>> Thanks Adam, that worked.  Accumulo starts but when I try the shell I get:
>>
>> ERROR: unable obtain instance id at file:/accumulo/instance_id
>>
>> $ hadoop fs -ls /
>>
>>
>> Shows the id file and the Hadoop configuration directory is on the
>> Accumulo class path according to accumulo-site.xml.
>>
>> Is the shell looking in the local file system or in hdfs?  I never had
>> this problem until I started up with Google.
>>
>> Thanks
>>
>> On Wed, Jan 31, 2018 at 5:06 PM, Adam J. Shook <adamjsh...@gmail.com>
>> wrote:
>>
>>> Yes, it does use RPC to talk to HDFS.  You will need to update the value
>>> of instance.volumes in accumulo-site.xml to reference this address,
>>> haz0-m:8020, instead of the default localhost:9000.
>>>
>>> --Adam
>>>
>>> On Wed, Jan 31, 2018 at 4:45 PM, Geoffry Roberts <threadedb...@gmail.com
>>> > wrote:
>>>
>>>> I have a situation where Accumulo cannot find Hadoop.
>>>>
>>>> Hadoop is running and I can access hdfs from the cli.
>>>> Zookeeper also says it is ok and I can log in using the client.
>>>> Accumulo init is failing with a connection refused for localhost:9000.
>>>>
>>>> netstat shows nothing listening on 9000.
>>>>
>>>> Now the plot thickens...
>>>>
>>>> The Hadoop I am running is Google's Dataproc and the Hadoop
>>>> installation is not my own.  I have already found a number of differences.
>>>>
>>>> Here's my question:  Does Accumulo use RPC to talk to Hadoop? I ask
>>>> because of things like this:
>>>>
>>>> From hfs-site.xml
>>>>
>>>>   <property>
>>>>
>>>>     <name>dfs.namenode.rpc-address</name>
>>>>
>>>>     <value>haz0-m:8020</value>
>>>>
>>>>     <description>
>>>>
>>>>       RPC address that handles all clients requests. If empty then
>>>> we'll get
>>>>
>>>>       thevalue from fs.default.name.The value of this property will
>>>> take the
>>>>
>>>>       form of hdfs://nn-host1:rpc-port.
>>>>
>>>>     </description>
>>>>
>>>>   </property>
>>>>
>>>> Or does it use something else?
>>>>
>>>> Thanks
>>>> --
>>>> There are ways and there are ways,
>>>>
>>>> Geoffry Roberts
>>>>
>>>
>>>
>>
>>
>> --
>> There are ways and there are ways,
>>
>> Geoffry Roberts
>>
>


-- 
There are ways and there are ways,

Geoffry Roberts

Reply via email to