What are you trying to accomplish here? You’d really need to have more nodes in 
your HDFS cluster to test anything realistic. I am not sure what happens when 
the cluster is really just 1 datanode - not sure the replication code within 
HDFS really handles that case.


Thanks,
Hari

On Wed, Oct 1, 2014 at 6:04 AM, Ed Judge <ejud...@gmail.com> wrote:

> Looks like they are up.  I see the following on one of the nodes but both 
> look generally the same (1 live datanode).
> [hadoop@localhost bin]$ hdfs dfsadmin -report
> 14/10/01 12:51:56 WARN util.NativeCodeLoader: Unable to load native-hadoop 
> library for your platform... using builtin-java classes where applicable
> Configured Capacity: 40797364224 (38.00 GB)
> Present Capacity: 37030862848 (34.49 GB)
> DFS Remaining: 37030830080 (34.49 GB)
> DFS Used: 32768 (32 KB)
> DFS Used%: 0.00%
> Under replicated blocks: 0
> Blocks with corrupt replicas: 0
> Missing blocks: 0
> -------------------------------------------------
> Datanodes available: 1 (1 total, 0 dead)
> Live datanodes:
> Name: 127.0.0.1:50010 (localhost)
> Hostname: localhost
> Decommission Status : Normal
> Configured Capacity: 40797364224 (38.00 GB)
> DFS Used: 32768 (32 KB)
> Non DFS Used: 3766501376 (3.51 GB)
> DFS Remaining: 37030830080 (34.49 GB)
> DFS Used%: 0.00%
> DFS Remaining%: 90.77%
> Configured Cache Capacity: 0 (0 B)
> Cache Used: 0 (0 B)
> Cache Remaining: 0 (0 B)
> Cache Used%: 100.00%
> Cache Remaining%: 0.00%
> Last contact: Wed Oct 01 12:51:57 UTC 2014
> I don’t know how to demonstrate that they are accessible except to telnet 
> into each of them.  Right now that test shows that both nodes accept the 
> connection to port 50010.
> Is there some other test I can perform?
> Thanks,
> -Ed
> On Oct 1, 2014, at 12:31 AM, Hari Shreedharan <hshreedha...@cloudera.com> 
> wrote:
>> Looks like one data node is inaccessible or down - so the HDFS client has 
>> black listed it and the writes are failing as blocks are allocated to that 
>> one.
>> 
>> Thanks,
>> Hari
>> 
>> 
>> On Tue, Sep 30, 2014 at 7:33 PM, Ed Judge <ejud...@gmail.com> wrote:
>> 
>> I’ve pulled over all of the Hadoop jar files for my flume instance to use.  
>> I am seeing some slightly different errors now.  Basically I have 2 
>> identically configured hadoop instances on the same subnet.  Running flume 
>> on those same instances and pointing flume at the local hadoop/hdfs instance 
>> works fine and the files get written.  However, when I point it to the 
>> adjacent hadoop/hdfs instance I get many exceptions/errors (show below) and 
>> the files never get written.  Here is my HDFS sink configuration on 
>> 10.0.0.14:
>> 
>> # Describe the sink
>> a1.sinks.k1.type = hdfs
>> a1.sinks.k1.hdfs.path = hdfs://10.0.0.16:9000/tmp/
>> a1.sinks.k1.hdfs.filePrefix = twitter
>> a1.sinks.k1.hdfs.fileSuffix = .ds
>> a1.sinks.k1.hdfs.rollInterval = 0
>> a1.sinks.k1.hdfs.rollSize = 10
>> a1.sinks.k1.hdfs.rollCount = 0
>> a1.sinks.k1.hdfs.fileType = DataStream
>> #a1.sinks.k1.serializer = TEXT
>> a1.sinks.k1.channel = c1
>> 
>> Any idea why this is not working?
>> 
>> Thanks.
>> 
>> 01 Oct 2014 01:59:45,098 INFO  
>> [SinkRunner-PollingRunner-DefaultSinkProcessor] 
>> (org.apache.flume.sink.hdfs.HDFSDataStream.configure:58)  - Serializer = 
>> TEXT, UseRawLocalFileSystem = false
>> 01 Oct 2014 01:59:45,385 INFO  
>> [SinkRunner-PollingRunner-DefaultSinkProcessor] 
>> (org.apache.flume.sink.hdfs.BucketWriter.open:261)  - Creating 
>> hdfs://10.0.0.16:9000/tmp//twitter.1412128785099.ds.tmp
>> 01 Oct 2014 01:59:45,997 INFO  [Twitter4J Async Dispatcher[0]] 
>> (org.apache.flume.source.twitter.TwitterSource.onStatus:178)  - Processed 
>> 100 docs
>> 01 Oct 2014 01:59:47,754 INFO  [Twitter4J Async Dispatcher[0]] 
>> (org.apache.flume.source.twitter.TwitterSource.onStatus:178)  - Processed 
>> 200 docs
>> 01 Oct 2014 01:59:49,379 INFO  [Thread-7] 
>> (org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream:1378)
>>   - Exception in createBlockOutputStream
>> java.io.EOFException: Premature EOF: no length prefix available
>>      at 
>> org.apache.hadoop.hdfs.protocolPB.PBHelper.vintPrefixed(PBHelper.java:1987)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1346)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1272)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:525)
>> 01 Oct 2014 01:59:49,390 INFO  [Thread-7] 
>> (org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream:1275)
>>   - Abandoning BP-1768727495-127.0.0.1-1412117897373:blk_1073743575_2751
>> 01 Oct 2014 01:59:49,398 INFO  [Thread-7] 
>> (org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream:1278)
>>   - Excluding datanode 127.0.0.1:50010
>> 01 Oct 2014 01:59:49,431 WARN  [Thread-7] 
>> (org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run:627)  - 
>> DataStreamer Exception
>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
>> /tmp/twitter.1412128785099.ds.tmp could only be replicated to 0 nodes 
>> instead of minReplication (=1).  There are 1 datanode(s) running and 1 
>> node(s) are excluded in this operation.
>>      at 
>> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1430)
>>      at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2684)
>>      at 
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:584)
>>      at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:440)
>>      at 
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>>      at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>      at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)
>> 
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1410)
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1363)
>>      at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>>      at com.sun.proxy.$Proxy18.addBlock(Unknown Source)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:190)
>>      at 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103)
>>      at com.sun.proxy.$Proxy18.addBlock(Unknown Source)
>>      at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:361)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1439)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1261)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:525)
>> 01 Oct 2014 01:59:49,437 WARN  [hdfs-k1-call-runner-2] 
>> (org.apache.hadoop.hdfs.DFSOutputStream.flushOrSync:1950)  - Error while 
>> syncing
>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
>> /tmp/twitter.1412128785099.ds.tmp could only be replicated to 0 nodes 
>> instead of minReplication (=1).  There are 1 datanode(s) running and 1 
>> node(s) are excluded in this operation.
>>      at 
>> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1430)
>>      at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2684)
>>      at 
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:584)
>>      at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:440)
>>      at 
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
>>      at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
>>      at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
>>      at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
>>      at java.security.AccessController.doPrivileged(Native Method)
>>      at javax.security.auth.Subject.doAs(Subject.java:415)
>>      at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
>>      at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)
>> 
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1410)
>>      at org.apache.hadoop.ipc.Client.call(Client.java:1363)
>>      at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
>>      at com.sun.proxy.$Proxy18.addBlock(Unknown Source)
>>      at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>      at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>      at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>      at 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:190)
>>      at 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103)
>>      at com.sun.proxy.$Proxy18.addBlock(Unknown Source)
>>      at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:361)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1439)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1261)
>>      at 
>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:525)
>> 01 Oct 2014 01:59:49,439 WARN  
>> [SinkRunner-PollingRunner-DefaultSinkProcessor] 
>> (org.apache.flume.sink.hdfs.HDFSEventSink.process:463)  - HDFS IO error
>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
>> /tmp/twitter.1412128785099.ds.tmp could only be replicated to 0 nodes 
>> instead of minReplication (=1).  There are 1 datanode(s) running and 1 
>> node(s) are excluded in this operation.
>> 
>> On Sep 30, 2014, at 3:18 PM, Hari Shreedharan <hshreedha...@cloudera.com> 
>> wrote:
>> 
>>> You'd need to add the jars that hadoop itself depends on. Flume pulls it in 
>>> if Hadoop is installed on that machine, else you'd need to manually 
>>> download it and install it. If you are using Hadoop 2.x, install the RPM 
>>> provided by Bigtop.
>>> 
>>> On Tue, Sep 30, 2014 at 12:12 PM, Ed Judge <ejud...@gmail.com> wrote:
>>> I added commons-configuration and there is now another missing dependency.  
>>> What do you mean by “all of Hadoop’s dependencies”?
>>> 
>>> 
>>> On Sep 30, 2014, at 2:51 PM, Hari Shreedharan <hshreedha...@cloudera.com> 
>>> wrote:
>>> 
>>>> You actually need to add of all Hadoop’s dependencies to Flume classpath. 
>>>> Looks like Apache Commons Configuration is missing in classpath.
>>>> 
>>>> Thanks,
>>>> Hari
>>>> 
>>>> 
>>>> On Tue, Sep 30, 2014 at 11:48 AM, Ed Judge <ejud...@gmail.com> wrote:
>>>> 
>>>> Thank you.  I am using hadoop 2.5 which I think uses 
>>>> protobuf-java-2.5.0.jar.
>>>> 
>>>> I am getting the following error even after adding those 2 jar files to my 
>>>> flume-ng classpath:
>>>> 
>>>> 30 Sep 2014 18:27:03,269 INFO  [lifecycleSupervisor-1-0] 
>>>> (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start:61)
>>>>   - Configuration provider starting
>>>> 30 Sep 2014 18:27:03,278 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:133)
>>>>   - Reloading configuration file:./src.conf
>>>> 30 Sep 2014 18:27:03,288 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,289 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:930)
>>>>   - Added sinks: k1 Agent: a1
>>>> 30 Sep 2014 18:27:03,289 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,292 WARN  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration.<init>:101)  - Configuration 
>>>> property ignored: i# = Describe the sink
>>>> 30 Sep 2014 18:27:03,292 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,292 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,293 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,293 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,293 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,293 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,293 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty:1016)
>>>>   - Processing:k1
>>>> 30 Sep 2014 18:27:03,312 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.conf.FlumeConfiguration.validateConfiguration:140)  - 
>>>> Post-validation flume configuration contains configuration for agents: [a1]
>>>> 30 Sep 2014 18:27:03,312 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:150)  - 
>>>> Creating channels
>>>> 30 Sep 2014 18:27:03,329 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.channel.DefaultChannelFactory.create:40)  - Creating 
>>>> instance of channel c1 type memory
>>>> 30 Sep 2014 18:27:03,351 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.node.AbstractConfigurationProvider.loadChannels:205)  - 
>>>> Created channel c1
>>>> 30 Sep 2014 18:27:03,352 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.source.DefaultSourceFactory.create:39)  - Creating 
>>>> instance of source r1, type org.apache.flume.source.twitter.TwitterSource
>>>> 30 Sep 2014 18:27:03,363 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.source.twitter.TwitterSource.configure:110)  - Consumer 
>>>> Key:        'tobhMtidckJoe1tByXDmI4pW3'
>>>> 30 Sep 2014 18:27:03,363 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.source.twitter.TwitterSource.configure:111)  - Consumer 
>>>> Secret:     '6eZKRpd6JvGT3Dg9jtd9fG9UMEhBzGxoLhLUGP1dqzkKznrXuQ'
>>>> 30 Sep 2014 18:27:03,363 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.source.twitter.TwitterSource.configure:112)  - Access 
>>>> Token:        '1588514408-o36mOSbXYCVacQ3p6Knsf6Kho17iCwNYLZyA9V5'
>>>> 30 Sep 2014 18:27:03,364 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.source.twitter.TwitterSource.configure:113)  - Access 
>>>> Token Secret: 'vBtp7wKsi2BOQqZSBpSBQSgZcc93oHea38T9OdckDCLKn'
>>>> 30 Sep 2014 18:27:03,825 INFO  [conf-file-poller-0] 
>>>> (org.apache.flume.sink.DefaultSinkFactory.create:40)  - Creating instance 
>>>> of sink: k1, type: hdfs
>>>> 30 Sep 2014 18:27:03,874 ERROR [conf-file-poller-0] 
>>>> (org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run:145)
>>>>   - Failed to start agent because dependencies were not found in 
>>>> classpath. Error follows.
>>>> java.lang.NoClassDefFoundError: 
>>>> org/apache/commons/configuration/Configuration
>>>>    at 
>>>> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.<init>(DefaultMetricsSystem.java:38)
>>>>    at 
>>>> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.<clinit>(DefaultMetricsSystem.java:36)
>>>>    at 
>>>> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.create(UserGroupInformation.java:106)
>>>>    at 
>>>> org.apache.hadoop.security.UserGroupInformation.<clinit>(UserGroupInformation.java:208)
>>>>    at 
>>>> org.apache.flume.sink.hdfs.HDFSEventSink.authenticate(HDFSEventSink.java:553)
>>>>    at 
>>>> org.apache.flume.sink.hdfs.HDFSEventSink.configure(HDFSEventSink.java:272)
>>>>    at org.apache.flume.conf.Configurables.configure(Configurables.java:41)
>>>>    at 
>>>> org.apache.flume.node.AbstractConfigurationProvider.loadSinks(AbstractConfigurationProvider.java:418)
>>>>    at 
>>>> org.apache.flume.node.AbstractConfigurationProvider.getConfiguration(AbstractConfigurationProvider.java:103)
>>>>    at 
>>>> org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run(PollingPropertiesFileConfigurationProvider.java:140)
>>>>    at 
>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>>>>    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304)
>>>>    at 
>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
>>>>    at 
>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>    at 
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>    at 
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>>    at java.lang.Thread.run(Thread.java:745)
>>>> Caused by: java.lang.ClassNotFoundException: 
>>>> org.apache.commons.configuration.Configuration
>>>>    at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>>>>    at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>>>    at java.security.AccessController.doPrivileged(Native Method)
>>>>    at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>>>    at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>>>    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>>>    at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>>>    ... 17 more
>>>> 30 Sep 2014 18:27:33,491 INFO  [agent-shutdown-hook] 
>>>> (org.apache.flume.lifecycle.LifecycleSupervisor.stop:79)  - Stopping 
>>>> lifecycle supervisor 10
>>>> 30 Sep 2014 18:27:33,493 INFO  [agent-shutdown-hook] 
>>>> (org.apache.flume.node.PollingPropertiesFileConfigurationProvider.stop:83) 
>>>>  - Configuration provider stopping
>>>> [vagrant@localhost 6]$ 
>>>> 
>>>> Is there another jar file I need?
>>>> 
>>>> Thanks.
>>>> 
>>>> On Sep 29, 2014, at 9:04 PM, shengyi.pan <shengyi....@gmail.com> wrote:
>>>> 
>>>>> you need hadoop-common-x.x.x.jar and hadoop-hdfs-x.x.x.jar under your 
>>>>> flume-ng classpath, and the dependent hadoop jar version must match your 
>>>>> hadoop system.
>>>>>  
>>>>> if sink to hadoop-2.0.0,  you should use "protobuf-java-2.4.1.jar" 
>>>>> (defaultly, flume-1.5.0 uses "protobuf-java-2.5.0.jar", the jar file is 
>>>>> under flume lib directory ), because the pb interface of hdfs-2.0 is 
>>>>> compiled wtih protobuf-2.4, while using protobuf-2.5 the flume-ng will 
>>>>> fail to start....
>>>>>  
>>>>>  
>>>>>  
>>>>>  
>>>>> 2014-09-30
>>>>> shengyi.pan
>>>>> 发件人:Ed Judge <ejud...@gmail.com>
>>>>> 发送时间:2014-09-29 22:38
>>>>> 主题:HDFS sink to a remote HDFS node
>>>>> 收件人:"user@flume.apache.org"<user@flume.apache.org>
>>>>> 抄送:
>>>>>  
>>>>> I am trying to run the flume-ng agent on one node with an HDFS sink 
>>>>> pointing to an HDFS filesystem on another node.
>>>>> Is this possible?  What packages/jar files are needed on the flume agent 
>>>>> node for this to work?  Secondary goal is to install only what is needed 
>>>>> on the flume-ng node.
>>>>> 
>>>>> # Describe the sink
>>>>> a1.sinks.k1.type = hdfs
>>>>> a1.sinks.k1.hdfs.path = hdfs://<remote IP address>/tmp/
>>>>> 
>>>>> 
>>>>> Thanks,
>>>>> Ed
>>>> 
>>>> 
>>> 
>>> 
>> 
>> 

Reply via email to