2014-02-14 58 views
1

我想水槽與HDFS整合和我的水槽的配置文件是水槽和HDFS集成,HDFS IO錯誤

hdfs-agent.sources= netcat-collect 
hdfs-agent.sinks = hdfs-write 
hdfs-agent.channels= memoryChannel 

hdfs-agent.sources.netcat-collect.type = netcat 
hdfs-agent.sources.netcat-collect.bind = localhost 
hdfs-agent.sources.netcat-collect.port = 11111 

hdfs-agent.sinks.hdfs-write.type = FILE_ROLL 
hdfs-agent.sinks.hdfs-write.hdfs.path = hdfs://127.0.0.1:50020/user/oracle/flume 
hdfs-agent.sinks.hdfs-write.rollInterval = 30 
hdfs-agent.sinks.hdfs-write.hdfs.writeFormat=Text 
hdfs-agent.sinks.hdfs-write.hdfs.fileType=DataStream 

hdfs-agent.channels.memoryChannel.type = memory 
hdfs-agent.channels.memoryChannel.capacity=10000 
hdfs-agent.sources.netcat-collect.channels=memoryChannel 
hdfs-agent.sinks.hdfs-write.channel=memoryChannel. 

而我的核心站點文件

<configuration> 
    <property> 
     <name>fs.default.name</name> 
     <value>hdfs://localhost</value> 
    </property> 
</configuration> 

當我嘗試運行水槽代理,它正在啓動,它能夠從nc命令中讀取,但在寫入hdfs時,我得到以下例外。我試圖用安全模式開始使用hadoop dfsadmin -safemode leave,但我仍然有以下例外。

2014-02-14 10:31:53,785 (SinkRunner-PollingRunner-DefaultSinkProcessor) [INFO - org.apache.flume.sink.hdfs.BucketWriter.open(BucketWriter.java:219)] Creating hdfs://127.0.0.1:50020/user/oracle/flume/FlumeData.1392354113707.tmp 
2014-02-14 10:31:54,011 (SinkRunner-PollingRunner-DefaultSinkProcessor) [WARN - org.apache.flume.sink.hdfs.HDFSEventSink.process(HDFSEventSink.java:418)] HDFS IO error 
java.io.IOException: Call to /127.0.0.1:50020 failed on local exception: java.io.EOFException 
     at org.apache.hadoop.ipc.Client.wrapException(Client.java:1089) 
     at org.apache.hadoop.ipc.Client.call(Client.java:1057) 
     at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:226) 
     at $Proxy5.getProtocolVersion(Unknown Source) 
     at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:369) 
     at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:111) 
     at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:213) 
     at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:180) 
     at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89) 
     at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1489) 
     at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) 
     at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1523) 
     at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1505) 
     at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:227) 
     at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175) 
     at org.apache.flume.sink.hdfs.BucketWriter$1.call(BucketWriter.java:226) 
     at org.apache.flume.sink.hdfs.BucketWriter$1.call(BucketWriter.java:220) 
     at org.apache.flume.sink.hdfs.BucketWriter$8$1.run(BucketWriter.java:536) 
     at org.apache.flume.sink.hdfs.BucketWriter.runPrivileged(BucketWriter.java:160) 
     at org.apache.flume.sink.hdfs.BucketWriter.access$1000(BucketWriter.java:56) 
     at org.apache.flume.sink.hdfs.BucketWriter$8.call(BucketWriter.java:533) 
     at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) 
     at java.util.concurrent.FutureTask.run(FutureTask.java:138) 
     at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) 
     at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) 
     at java.lang.Thread.run(Thread.java:662) 
Caused by: java.io.EOFException 
     at java.io.DataInputStream.readInt(DataInputStream.java:375) 
     at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:781) 
     at org.apache.hadoop.ipc.Client$Connection.run(Client.java:689) 

請讓我知道是否在任何屬性文件中配置了錯誤的東西,以便它可以工作。

也請讓我知道如果我使用的是正確的端口爲這個

我的目標是整合水槽和Hadoop。 我對Hadoop的

回答

2

單節點服務器設置您必須提供端口號與fs.default.name

例子:

<configuration> 
    <property> 
     <name>fs.default.name</name> 
     <value>hdfs://localhost:9001</value> 
    </property> 
</configuration> 

是編輯水槽的配置文件,如下

hdfs-agent.sources= netcat-collect 
hdfs-agent.sinks = hdfs-write 
hdfs-agent.channels= memoryChannel 

hdfs-agent.sources.netcat-collect.type = netcat 
hdfs-agent.sources.netcat-collect.bind = localhost 
hdfs-agent.sources.netcat-collect.port = 11111 

hdfs-agent.sinks.hdfs-write.type = hdfs 
hdfs-agent.sinks.hdfs-write.hdfs.path = hdfs://127.0.0.1:9001/user/oracle/flume 
hdfs-agent.sinks.hdfs-write.rollInterval = 30 
hdfs-agent.sinks.hdfs-write.hdfs.writeFormat=Text 
hdfs-agent.sinks.hdfs-write.hdfs.fileType=DataStream 

hdfs-agent.channels.memoryChannel.type = memory 
hdfs-agent.channels.memoryChannel.capacity=10000 
hdfs-agent.sources.netcat-collect.channels=memoryChannel 
hdfs-agent.sinks.hdfs-write.channel=memoryChannel 

變化:

hdfs-agent.sinks.hdfs-write.type = hdfs(sink type as hdfs) 
hdfs-agent.sinks.hdfs-write.hdfs.path = hdfs://127.0.0.1:9001/user/oracle/flume(port number) 
hdfs-agent.sinks.hdfs-write.channel=memoryChannel(Removed the dot symbol after memoryChannel)