linux 命令之ln

    xiaoxiao2025-07-24  15

    ln是用于创建链接,分为软链接和硬链接

    一、创建软链接

    [root@hadoop001 opt]# ll total 80 -rw-r--r-- 1 root root 73128 May 26 22:20 hadoop-root-datanode- hadoop001.log drwxr-xr-x 2 root root 4096 Mar 4 21:42 module drwxr-xr-x 3 root root 4096 Mar 5 21:10 software [root@hadoop001 opt]# ln -s hadoop-root-datanode-hadoop001.log link2019 [root@hadoop001 opt]# ll total 80 -rw-r--r-- 1 root root 73128 May 26 22:20 hadoop-root-datanode- hadoop001.log lrwxrwxrwx 1 root root 34 May 26 22:21 link2019 -> hadoop-root- datanode-hadoop001.log drwxr-xr-x 2 root root 4096 Mar 4 21:42 module drwxr-xr-x 3 root root 4096 Mar 5 21:10 software

    试一下:

    [root@hadoop001 opt]# cat link2019 2019-03-23 13:37:06,815 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = localhost/127.0.0.1 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.1 STARTUP_MSG: classpath = /opt/software/hadoop- 2.7.1/etc/hadoop:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-configuration- 1.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jetty- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- codec-1.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- beanutils-core-1.8.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/stax-api-1.0-2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/avro- 1.7.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/curator-client- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/activation- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/httpcore-4.2.5.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/hadoop-auth-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/slf4j-log4j12- 1.7.10.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jaxb-api- 2.2.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/xmlenc- 0.52.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/gson- 2.2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jaxb-impl- 2.2.3-1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jersey- json-1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/xz- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/paranamer- 2.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/mockito-all- 1.8.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/apacheds- i18n-2.0.0-M15.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jsr305-3.0.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-compress- 1.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- httpclient-3.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/servlet-api-2.5.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/htrace-core-3.1.0- incubating.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/hadoop- annotations-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jersey-server-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/curator-recipes- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- digester-1.8.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-net-3.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-cli-1.2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/api-asn1-api-1.0.0- M20.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jsp-api- 2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/java- xmlbuilder-0.4.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jettison-1.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/zookeeper-3.4.6.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jetty-util-6.1.26.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/snappy-java- 1.0.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/slf4j-api- 1.7.10.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jets3t- 0.9.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons-math3 -3.1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- lang-2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- beanutils-1.7.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/junit- 4.11.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/httpclient- 4.2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/apacheds- kerberos-codec-2.0.0-M15.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jersey-core-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/hamcrest-core-1.3.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/asm-3.2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jackson-jaxrs- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jsch- 0.1.42.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jackson- core-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-collections- 3.2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/curator- framework-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- nfs-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- common-2.7.1-tests.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xmlenc- 0.52.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/netty-all- 4.0.23.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-daemon- 1.0.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jsr305- 3.0.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/servlet-api- 2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/htrace-core-3.1.0 -incubating.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jersey- server-1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xml-apis- 1.3.04.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-cli- 1.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xercesImpl- 2.9.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jetty-util- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-lang- 2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jersey-core- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jackson-core-asl- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs-2.7.1- tests.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs-nfs- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jetty- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/javax.inject- 1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-codec- 1.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-guice- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/stax-api-1.0- 2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/activation- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guice- 3.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jaxb-api- 2.2.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-client- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jaxb-impl-2.2.3- 1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-json- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/aopalliance- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/xz- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/leveldbjni-all- 1.8.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jsr305- 3.0.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons- compress-1.4.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/lib/zookeeper-3.4.6- tests.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/servlet-api- 2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-xc- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-server- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-cli- 1.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jettison- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/zookeeper- 3.4.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jetty-util- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guice-servlet- 3.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-lang- 2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-core- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-jaxrs- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-core- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons- collections-3.2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-server-resourcemanager-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-client- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- applicationhistoryservice-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-server-nodemanager- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn- registry-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-applications-unmanaged-am-launcher-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-applications-distributedshell- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- web-proxy-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-api-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-server-tests-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/javax.inject-1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/jersey-guice- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/avro- 1.7.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/commons- io-2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jackson- mapper-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/guice-3.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/xz-1.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/leveldbjni-all- 1.8.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/commons- compress-1.4.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/hadoop-annotations- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jersey- server-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/snappy-java- 1.0.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/guice- servlet-3.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/junit- 4.11.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jersey- core-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/hamcrest-core- 1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jackson- core-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-shuffle-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-common- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-examples-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-app- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-core-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-hs-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.7.1- tests.jar:/opt/software/hadoop-2.7.1/contrib/capacity- scheduler/*.jar:/opt/software/hadoop-2.7.1/contrib/capacity- scheduler/*.jar:/opt/software/hadoop-2.7.1/contrib/capacity-scheduler/*.jar STARTUP_MSG: build = https://git-wip-us.apache.org/repos/asf/hadoop.git -r 15ecc87ccf4a0228f35af08fc56de536e6ce657a; compiled by 'jenkins' on 2015 -06-29T06:04Z STARTUP_MSG: java = 1.8.0_65 ************************************************************/ 2019-03-23 13:37:06,824 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT] 2019-03-23 13:37:07,194 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin- java classes where applicable 2019-03-23 13:37:07,431 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2019-03-23 13:37:07,514 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2019-03-23 13:37:07,514 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2019-03-23 13:37:07,520 INFO org.apache.hadoop.hdfs.server.datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2019-03-23 13:37:07,523 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is localhost 2019-03-23 13:37:07,530 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2019-03-23 13:37:07,559 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:50010 2019-03-23 13:37:07,561 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is 1048576 bytes/s 2019-03-23 13:37:07,561 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Number threads for balancing is 5 2019-03-23 13:37:07,647 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2019-03-23 13:37:07,655 INFO org.apache.hadoop.security.authentication.server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2019-03-23 13:37:07,660 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2019-03-23 13:37:07,666 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2019-03-23 13:37:07,668 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2019-03-23 13:37:07,668 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2019-03-23 13:37:07,668 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2019-03-23 13:37:07,681 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 35875 2019-03-23 13:37:07,681 INFO org.mortbay.log: jetty-6.1.26 2019-03-23 13:37:07,843 INFO org.mortbay.log: Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:35875 2019-03-23 13:37:07,948 INFO org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Listening HTTP traffic on /0.0.0.0:50075 2019-03-23 13:37:08,109 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnUserName = root 2019-03-23 13:37:08,109 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: supergroup = supergroup 2019-03-23 13:37:09,190 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue 2019-03-23 13:37:09,205 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 50020 2019-03-23 13:37:09,227 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:50020 2019-03-23 13:37:09,240 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null 2019-03-23 13:37:09,261 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting BPOfferServices for nameservices: <default> 2019-03-23 13:37:09,276 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool <registering> (Datanode Uuid unassigned) service to hadoop001/127.0.0.1:9000 starting to offer service 2019-03-23 13:37:09,280 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2019-03-23 13:37:09,281 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 50020: starting 2019-03-23 13:37:09,756 INFO org.apache.hadoop.hdfs.server.common.Storage: Lock on /opt/software/hadoop-2.7.1/data/tmp/dfs/data/in_use.lock acquired by nodename 1552@localhost 2019-03-23 13:37:09,758 WARN org.apache.hadoop.hdfs.server.common.Storage: java.io.IOException: Incompatible clusterIDs in /opt/software/hadoop- 2.7.1/data/tmp/dfs/data: namenode clusterID = CID-a135bed3-efab-4cc6-b1a4- 5ef7b1d2057b; datanode clusterID = CID-bfd504a2-6851-4fe6-b980-c71f1808b2ac 2019-03-23 13:37:09,758 FATAL org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for Block pool <registering> (Datanode Uuid unassigned) service to hadoop001/127.0.0.1:9000. Exiting. java.io.IOException: All specified directories are failed to load. at org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead (DataStorage.java:477) at org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage (DataNode.java:1361) at org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool (DataNode.java:1326) at org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespace Info(BPOfferService.java:316) at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandsha ke(BPServiceActor.java:223) at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run (BPServiceActor.java:801) at java.lang.Thread.run(Thread.java:745) 2019-03-23 13:37:09,760 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service for: Block pool <registering> (Datanode Uuid unassigned) service to hadoop001/127.0.0.1:9000 2019-03-23 13:37:09,761 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool <registering> (Datanode Uuid unassigned) 2019-03-23 13:37:11,762 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode 2019-03-23 13:37:11,763 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 0 2019-03-23 13:37:11,765 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at localhost/127.0.0.1 ************************************************************/ 2019-03-23 13:44:36,059 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = localhost/127.0.0.1 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.1 STARTUP_MSG: classpath = /opt/software/hadoop- 2.7.1/etc/hadoop:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-configuration- 1.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jetty- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- codec-1.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- beanutils-core-1.8.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/stax-api-1.0-2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/avro- 1.7.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/curator-client- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/activation- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/httpcore-4.2.5.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/hadoop-auth-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/slf4j-log4j12- 1.7.10.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jaxb-api- 2.2.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/xmlenc- 0.52.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/gson- 2.2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jaxb-impl- 2.2.3-1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jersey- json-1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/xz- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/paranamer- 2.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/mockito-all- 1.8.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/apacheds- i18n-2.0.0-M15.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jsr305-3.0.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-compress- 1.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- httpclient-3.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/servlet-api-2.5.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/htrace-core-3.1.0- incubating.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/hadoop- annotations-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jersey-server-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/curator-recipes- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- digester-1.8.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-net-3.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-cli-1.2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/api-asn1-api-1.0.0- M20.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jsp-api- 2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/java- xmlbuilder-0.4.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jettison-1.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/zookeeper-3.4.6.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jetty-util-6.1.26.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/snappy-java- 1.0.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/slf4j-api- 1.7.10.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jets3t- 0.9.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons-math3 -3.1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- lang-2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- beanutils-1.7.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/junit- 4.11.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/httpclient- 4.2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/apacheds- kerberos-codec-2.0.0-M15.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jersey-core-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/hamcrest-core-1.3.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/asm-3.2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jackson-jaxrs- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jsch- 0.1.42.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jackson- core-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-collections- 3.2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/curator- framework-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- nfs-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- common-2.7.1-tests.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xmlenc- 0.52.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/netty-all- 4.0.23.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-daemon- 1.0.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jsr305- 3.0.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/servlet-api- 2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/htrace-core-3.1.0 -incubating.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jersey- server-1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xml-apis- 1.3.04.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-cli- 1.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xercesImpl- 2.9.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jetty-util- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-lang- 2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jersey-core- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jackson-core-asl- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs-2.7.1- tests.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs-nfs- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jetty- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/javax.inject- 1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-codec- 1.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-guice- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/stax-api-1.0- 2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/activation- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guice- 3.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jaxb-api- 2.2.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-client- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jaxb-impl-2.2.3- 1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-json- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/aopalliance- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/xz- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/leveldbjni-all- 1.8.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jsr305- 3.0.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons- compress-1.4.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/lib/zookeeper-3.4.6- tests.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/servlet-api- 2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-xc- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-server- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-cli- 1.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jettison- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/zookeeper- 3.4.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jetty-util- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guice-servlet- 3.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-lang- 2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-core- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-jaxrs- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-core- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons- collections-3.2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-server-resourcemanager-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-client- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- applicationhistoryservice-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-server-nodemanager- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn- registry-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-applications-unmanaged-am-launcher-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-applications-distributedshell- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- web-proxy-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-api-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-server-tests-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/javax.inject-1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/jersey-guice- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/avro- 1.7.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/commons- io-2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jackson- mapper-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/guice-3.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/xz-1.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/leveldbjni-all- 1.8.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/commons- compress-1.4.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/hadoop-annotations- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jersey- server-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/snappy-java- 1.0.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/guice- servlet-3.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/junit- 4.11.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jersey- core-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/hamcrest-core- 1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jackson- core-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-shuffle-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-common- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-examples-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-app- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-core-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-hs-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.7.1- tests.jar:/opt/software/hadoop-2.7.1/contrib/capacity- scheduler/*.jar:/opt/software/hadoop-2.7.1/contrib/capacity- scheduler/*.jar:/opt/software/hadoop-2.7.1/contrib/capacity-scheduler/*.jar STARTUP_MSG: build = https://git-wip-us.apache.org/repos/asf/hadoop.git -r 15ecc87ccf4a0228f35af08fc56de536e6ce657a; compiled by 'jenkins' on 2015 -06-29T06:04Z STARTUP_MSG: java = 1.8.0_65 ************************************************************/ 2019-03-23 13:44:36,079 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT] 2019-03-23 13:44:36,472 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin- java classes where applicable 2019-03-23 13:44:36,698 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2019-03-23 13:44:36,765 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2019-03-23 13:44:36,765 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2019-03-23 13:44:36,770 INFO org.apache.hadoop.hdfs.server.datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2019-03-23 13:44:36,771 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is localhost 2019-03-23 13:44:36,777 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2019-03-23 13:44:36,798 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:50010 2019-03-23 13:44:36,800 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is 1048576 bytes/s 2019-03-23 13:44:36,801 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Number threads for balancing is 5 2019-03-23 13:44:36,871 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2019-03-23 13:44:36,878 INFO org.apache.hadoop.security.authentication.server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2019-03-23 13:44:36,883 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2019-03-23 13:44:36,889 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2019-03-23 13:44:36,890 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2019-03-23 13:44:36,891 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2019-03-23 13:44:36,891 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2019-03-23 13:44:36,901 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 42776 2019-03-23 13:44:36,901 INFO org.mortbay.log: jetty-6.1.26 2019-03-23 13:44:37,060 INFO org.mortbay.log: Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:42776 2019-03-23 13:44:37,152 INFO org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Listening HTTP traffic on /0.0.0.0:50075 2019-03-23 13:44:37,334 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnUserName = root 2019-03-23 13:44:37,334 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: supergroup = supergroup 2019-03-23 13:44:37,396 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue 2019-03-23 13:44:37,412 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 50020 2019-03-23 13:44:37,435 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:50020 2019-03-23 13:44:37,447 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null 2019-03-23 13:44:37,468 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting BPOfferServices for nameservices: <default> 2019-03-23 13:44:37,482 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool <registering> (Datanode Uuid unassigned) service to hadoop001/127.0.0.1:9000 starting to offer service 2019-03-23 13:44:37,494 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2019-03-23 13:44:37,501 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 50020: starting 2019-03-23 13:44:37,910 INFO org.apache.hadoop.hdfs.server.common.Storage: Lock on /opt/software/hadoop-2.7.1/data/tmp/dfs/data/in_use.lock acquired by nodename 1813@localhost 2019-03-23 13:44:37,950 INFO org.apache.hadoop.hdfs.server.common.Storage: Analyzing storage directories for bpid BP-1244105038-127.0.0.1- 1553319341262 2019-03-23 13:44:37,950 INFO org.apache.hadoop.hdfs.server.common.Storage: Locking is disabled for /opt/software/hadoop- 2.7.1/data/tmp/dfs/data/current/BP-1244105038-127.0.0.1-1553319341262 2019-03-23 13:44:37,950 INFO org.apache.hadoop.hdfs.server.common.Storage: Block pool storage directory /opt/software/hadoop- 2.7.1/data/tmp/dfs/data/current/BP-1244105038-127.0.0.1-1553319341262 is not formatted for BP-1244105038-127.0.0.1-1553319341262 2019-03-23 13:44:37,950 INFO org.apache.hadoop.hdfs.server.common.Storage: Formatting ... 2019-03-23 13:44:37,951 INFO org.apache.hadoop.hdfs.server.common.Storage: Formatting block pool BP-1244105038-127.0.0.1-1553319341262 directory /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current/BP-1244105038- 127.0.0.1-1553319341262/current 2019-03-23 13:44:37,953 INFO org.apache.hadoop.hdfs.server.common.Storage: Restored 0 block files from trash. 2019-03-23 13:44:37,955 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Setting up storage: nsid=1341947409;bpid=BP-1244105038-127.0.0.1-1553319341262;lv=- 56;nsInfo=lv=-63;cid=CID-a135bed3-efab-4cc6-b1a4- 5ef7b1d2057b;nsid=1341947409;c=0;bpid=BP-1244105038-127.0.0.1- 1553319341262;dnuuid=fc4fe5b9-f39d-462d-adbf-73e2a2270bd7 2019-03-23 13:44:37,994 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added new volume: DS-091cd237-2e30-437c-b919-867e6a8b5fcf 2019-03-23 13:44:37,994 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added volume - /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current, StorageType: DISK 2019-03-23 13:44:38,000 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Registered FSDatasetState MBean 2019-03-23 13:44:38,000 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding block pool BP-1244105038-127.0.0.1-1553319341262 2019-03-23 13:44:38,002 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Scanning block pool BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current... 2019-03-23 13:44:38,015 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time taken to scan block pool BP-1244105038-127.0.0.1-1553319341262 on /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current: 14ms 2019-03-23 13:44:38,016 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to scan all replicas for block pool BP-1244105038-127.0.0.1- 1553319341262: 15ms 2019-03-23 13:44:38,016 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding replicas to map for block pool BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current... 2019-03-23 13:44:38,016 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time to add replicas to map for block pool BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current: 0ms 2019-03-23 13:44:38,016 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to add all replicas to map: 0ms 2019-03-23 13:44:38,152 INFO org.apache.hadoop.hdfs.server.datanode.VolumeScanner: Now scanning bpid BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop- 2.7.1/data/tmp/dfs/data 2019-03-23 13:44:38,155 INFO org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Periodic Directory Tree Verification scan starting at 1553335933155 with interval 21600000 2019-03-23 13:44:38,157 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool BP-1244105038- 127.0.0.1-1553319341262 (Datanode Uuid null) service to hadoop001/127.0.0.1:9000 beginning handshake with NN 2019-03-23 13:44:38,174 INFO org.apache.hadoop.hdfs.server.datanode.VolumeScanner: VolumeScanner (/opt/software/hadoop-2.7.1/data/tmp/dfs/data, DS-091cd237-2e30-437c-b919- 867e6a8b5fcf): finished scanning block pool BP-1244105038-127.0.0.1- 1553319341262 2019-03-23 13:44:38,207 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool BP- 1244105038-127.0.0.1-1553319341262 (Datanode Uuid null) service to hadoop001/127.0.0.1:9000 successfully registered with NN 2019-03-23 13:44:38,208 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode hadoop001/127.0.0.1:9000 using DELETEREPORT_INTERVAL of 300000 msec BLOCKREPORT_INTERVAL of 21600000msec CACHEREPORT_INTERVAL of 10000msec Initial delay: 0msec; heartBeatInterval=3000 2019-03-23 13:44:38,250 INFO org.apache.hadoop.hdfs.server.datanode.VolumeScanner: VolumeScanner (/opt/software/hadoop-2.7.1/data/tmp/dfs/data, DS-091cd237-2e30-437c-b919- 867e6a8b5fcf): no suitable block pools found to scan. Waiting 1814399902 ms. 2019-03-23 13:44:38,318 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool BP- 1244105038-127.0.0.1-1553319341262 (Datanode Uuid fc4fe5b9-f39d-462d-adbf- 73e2a2270bd7) service to hadoop001/127.0.0.1:9000 trying to claim ACTIVE state with txid=2 2019-03-23 13:44:38,319 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE Namenode Block pool BP-1244105038-127.0.0.1-1553319341262 (Datanode Uuid fc4fe5b9-f39d-462d-adbf-73e2a2270bd7) service to hadoop001/127.0.0.1:9000 2019-03-23 13:44:38,391 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Successfully sent block report 0x13ff7e48735, containing 1 storage report(s), of which we sent 1. The reports had 0 total blocks and used 1 RPC(s). This took 8 msec to generate and 64 msecs for RPC and NN processing. Got back one command: FinalizeCommand/5. 2019-03-23 13:44:38,391 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Got finalize command for block pool BP-1244105038-127.0.0.1-1553319341262 2019-03-23 14:59:42,118 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: RECEIVED SIGNAL 15: SIGTERM 2019-03-23 14:59:42,123 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at localhost/127.0.0.1 ************************************************************/ 2019-03-23 15:06:02,355 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = localhost/127.0.0.1 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.1 STARTUP_MSG: classpath = /opt/software/hadoop- 2.7.1/etc/hadoop:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-configuration- 1.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jetty- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- codec-1.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- beanutils-core-1.8.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/stax-api-1.0-2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/avro- 1.7.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/curator-client- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/activation- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/httpcore-4.2.5.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/hadoop-auth-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/slf4j-log4j12- 1.7.10.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jaxb-api- 2.2.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/xmlenc- 0.52.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/gson- 2.2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jaxb-impl- 2.2.3-1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jersey- json-1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/xz- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/paranamer- 2.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/mockito-all- 1.8.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/apacheds- i18n-2.0.0-M15.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jsr305-3.0.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-compress- 1.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- httpclient-3.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/servlet-api-2.5.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/htrace-core-3.1.0- incubating.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/hadoop- annotations-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jersey-server-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/curator-recipes- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- digester-1.8.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-net-3.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-cli-1.2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/api-asn1-api-1.0.0- M20.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jsp-api- 2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/java- xmlbuilder-0.4.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jettison-1.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/zookeeper-3.4.6.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jetty-util-6.1.26.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/snappy-java- 1.0.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/slf4j-api- 1.7.10.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jets3t- 0.9.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons-math3 -3.1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- lang-2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/commons- beanutils-1.7.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/junit- 4.11.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/httpclient- 4.2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/apacheds- kerberos-codec-2.0.0-M15.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jersey-core-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/hamcrest-core-1.3.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/asm-3.2.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/jackson-jaxrs- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jsch- 0.1.42.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/jackson- core-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/common/lib/commons-collections- 3.2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/lib/curator- framework-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- nfs-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/common/hadoop- common-2.7.1-tests.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xmlenc- 0.52.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/netty-all- 4.0.23.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/opt/software/hadoop- 2.7.1/share/hadoop/hdfs/lib/commons-daemon- 1.0.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jsr305- 3.0.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/servlet-api- 2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/htrace-core-3.1.0 -incubating.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jersey- server-1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xml-apis- 1.3.04.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-cli- 1.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/xercesImpl- 2.9.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jetty-util- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/commons-lang- 2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jersey-core- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/lib/jackson-core-asl- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs-2.7.1- tests.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs-nfs- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/hdfs/hadoop-hdfs- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jetty- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/javax.inject- 1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-codec- 1.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-guice- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/stax-api-1.0- 2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-logging- 1.1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-io- 2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/activation- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-mapper- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guice- 3.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jaxb-api- 2.2.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-client- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jaxb-impl-2.2.3- 1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-json- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/aopalliance- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/xz- 1.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/leveldbjni-all- 1.8.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jsr305- 3.0.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons- compress-1.4.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/lib/zookeeper-3.4.6- tests.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/servlet-api- 2.5.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-xc- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-server- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/log4j- 1.2.17.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-cli- 1.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jettison- 1.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/zookeeper- 3.4.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jetty-util- 6.1.26.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guice-servlet- 3.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons-lang- 2.6.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/guava- 11.0.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jersey-core- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-jaxrs- 1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/jackson-core- asl-1.9.13.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/lib/commons- collections-3.2.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-server-resourcemanager-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-client- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- applicationhistoryservice-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-server-nodemanager- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn- registry-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-applications-unmanaged-am-launcher-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/yarn/hadoop-yarn-applications-distributedshell- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop-yarn-server- web-proxy-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-api-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-common-2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/yarn/hadoop- yarn-server-tests-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/netty- 3.6.2.Final.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/javax.inject-1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/jersey-guice- 1.9.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/avro- 1.7.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/commons- io-2.4.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jackson- mapper-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/guice-3.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/xz-1.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/leveldbjni-all- 1.8.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/commons- compress-1.4.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/hadoop-annotations- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jersey- server-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/snappy-java- 1.0.4.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/guice- servlet-3.0.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/protobuf-java- 2.5.0.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/junit- 4.11.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jersey- core-1.9.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/lib/hamcrest-core- 1.3.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/asm- 3.2.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/lib/jackson- core-asl-1.9.13.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-shuffle-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-common- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-examples-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-app- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-core-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins- 2.7.1.jar:/opt/software/hadoop-2.7.1/share/hadoop/mapreduce/hadoop- mapreduce-client-hs-2.7.1.jar:/opt/software/hadoop- 2.7.1/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.7.1- tests.jar:/contrib/capacity-scheduler/*.jar:/opt/software/hadoop- 2.7.1/contrib/capacity-scheduler/*.jar:/opt/software/hadoop- 2.7.1/contrib/capacity-scheduler/*.jar STARTUP_MSG: build = https://git-wip-us.apache.org/repos/asf/hadoop.git -r 15ecc87ccf4a0228f35af08fc56de536e6ce657a; compiled by 'jenkins' on 2015 -06-29T06:04Z STARTUP_MSG: java = 1.8.0_65 ************************************************************/ 2019-03-23 15:06:02,363 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT] 2019-03-23 15:06:02,740 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin- java classes where applicable 2019-03-23 15:06:02,961 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2019-03-23 15:06:03,024 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2019-03-23 15:06:03,024 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2019-03-23 15:06:03,029 INFO org.apache.hadoop.hdfs.server.datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2019-03-23 15:06:03,031 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is localhost 2019-03-23 15:06:03,036 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2019-03-23 15:06:03,057 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:50010 2019-03-23 15:06:03,059 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is 1048576 bytes/s 2019-03-23 15:06:03,059 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Number threads for balancing is 5 2019-03-23 15:06:03,130 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2019-03-23 15:06:03,137 INFO org.apache.hadoop.security.authentication.server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2019-03-23 15:06:03,144 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2019-03-23 15:06:03,150 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2019-03-23 15:06:03,152 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2019-03-23 15:06:03,152 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2019-03-23 15:06:03,152 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2019-03-23 15:06:03,164 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 44010 2019-03-23 15:06:03,164 INFO org.mortbay.log: jetty-6.1.26 2019-03-23 15:06:03,312 INFO org.mortbay.log: Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:44010 2019-03-23 15:06:03,419 INFO org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Listening HTTP traffic on /0.0.0.0:50075 2019-03-23 15:06:03,572 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnUserName = root 2019-03-23 15:06:03,572 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: supergroup = supergroup 2019-03-23 15:06:03,652 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue class java.util.concurrent.LinkedBlockingQueue 2019-03-23 15:06:03,683 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 50020 2019-03-23 15:06:03,708 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:50020 2019-03-23 15:06:03,719 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null 2019-03-23 15:06:03,741 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting BPOfferServices for nameservices: <default> 2019-03-23 15:06:03,751 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool <registering> (Datanode Uuid unassigned) service to hadoop001/127.0.0.1:9000 starting to offer service 2019-03-23 15:06:03,755 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2019-03-23 15:06:03,765 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 50020: starting 2019-03-23 15:06:04,263 INFO org.apache.hadoop.hdfs.server.common.Storage: Lock on /opt/software/hadoop-2.7.1/data/tmp/dfs/data/in_use.lock acquired by nodename 2365@localhost 2019-03-23 15:06:04,309 INFO org.apache.hadoop.hdfs.server.common.Storage: Analyzing storage directories for bpid BP-1244105038-127.0.0.1- 1553319341262 2019-03-23 15:06:04,309 INFO org.apache.hadoop.hdfs.server.common.Storage: Locking is disabled for /opt/software/hadoop- 2.7.1/data/tmp/dfs/data/current/BP-1244105038-127.0.0.1-1553319341262 2019-03-23 15:06:04,310 INFO org.apache.hadoop.hdfs.server.common.Storage: Restored 0 block files from trash. 2019-03-23 15:06:04,311 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Setting up storage: nsid=1341947409;bpid=BP-1244105038-127.0.0.1-1553319341262;lv=- 56;nsInfo=lv=-63;cid=CID-a135bed3-efab-4cc6-b1a4- 5ef7b1d2057b;nsid=1341947409;c=0;bpid=BP-1244105038-127.0.0.1- 1553319341262;dnuuid=fc4fe5b9-f39d-462d-adbf-73e2a2270bd7 2019-03-23 15:06:04,353 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added new volume: DS-091cd237-2e30-437c-b919-867e6a8b5fcf 2019-03-23 15:06:04,353 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Added volume - /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current, StorageType: DISK 2019-03-23 15:06:04,361 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Registered FSDatasetState MBean 2019-03-23 15:06:04,362 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding block pool BP-1244105038-127.0.0.1-1553319341262 2019-03-23 15:06:04,363 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Scanning block pool BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current... 2019-03-23 15:06:04,375 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Cached dfsUsed found for /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current/BP- 1244105038-127.0.0.1-1553319341262/current: 28672 2019-03-23 15:06:04,377 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time taken to scan block pool BP-1244105038-127.0.0.1-1553319341262 on /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current: 14ms 2019-03-23 15:06:04,377 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to scan all replicas for block pool BP-1244105038-127.0.0.1- 1553319341262: 15ms 2019-03-23 15:06:04,377 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding replicas to map for block pool BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current... 2019-03-23 15:06:04,378 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Time to add replicas to map for block pool BP-1244105038-127.0.0.1-1553319341262 on volume /opt/software/hadoop-2.7.1/data/tmp/dfs/data/current: 0ms 2019-03-23 15:06:04,378 INFO org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Total time to add all replicas to map: 1ms 2019-03-23 15:06:04,564 INFO org.apache.hadoop.hdfs.server.datanode.VolumeScanner: VolumeScanner (/opt/software/hadoop-2.7.1/data/tmp/dfs/data, DS-091cd237-2e30-437c-b919- 867e6a8b5fcf): no suitable block pools found to scan. Waiting 1809513588 ms. 2019-03-23 15:06:04,566 INFO org.apache.hadoop.hdfs.server.datanode.DirectoryScanner: Periodic Directory Tree Verification scan starting at 1553338118566 with interval 21600000 2019-03-23 15:06:04,568 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool BP-1244105038- 127.0.0.1-1553319341262 (Datanode Uuid null) service to hadoop001/127.0.0.1:9000 beginning handshake with NN 2019-03-23 15:06:04,607 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool BP- 1244105038-127.0.0.1-1553319341262 (Datanode Uuid null) service to hadoop001/127.0.0.1:9000 successfully registered with NN 2019-03-23 15:06:04,607 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode hadoop001/127.0.0.1:9000 using DELETEREPORT_INTERVAL of 300000 msec BLOCKREPORT_INTERVAL of 21600000msec CACHEREPORT_INTERVAL of 10000msec Initial delay: 0msec; heartBeatInterval=3000 2019-03-23 15:06:04,690 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool BP- 1244105038-127.0.0.1-1553319341262 (Datanode Uuid fc4fe5b9-f39d-462d-adbf- 73e2a2270bd7) service to hadoop001/127.0.0.1:9000 trying to claim ACTIVE state with txid=3 2019-03-23 15:06:04,690 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE Namenode Block pool BP-1244105038-127.0.0.1-1553319341262 (Datanode Uuid fc4fe5b9-f39d-462d-adbf-73e2a2270bd7) service to hadoop001/127.0.0.1:9000 2019-03-23 15:06:04,753 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Successfully sent block report 0x5b1aa041d3b, containing 1 storage report(s), of which we sent 1. The reports had 0 total blocks and used 1 RPC(s). This took 3 msec to generate and 60 msecs for RPC and NN processing. Got back one command: FinalizeCommand/5. 2019-03-23 15:06:04,753 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Got finalize command for block pool BP-1244105038-127.0.0.1-1553319341262 2019-03-23 15:22:51,793 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-1244105038- 127.0.0.1-1553319341262:blk_1073741825_1001 src: /127.0.0.1:33314 dest: /127.0.0.1:50010 2019-03-23 15:22:51,946 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /127.0.0.1:33314, dest: /127.0.0.1:50010, bytes: 45, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_2021642614_1, offset: 0, srvID: fc4fe5b9-f39d-462d- adbf-73e2a2270bd7, blockid: BP-1244105038-127.0.0.1- 1553319341262:blk_1073741825_1001, duration: 118686326 2019-03-23 15:22:51,963 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP- 1244105038-127.0.0.1-1553319341262:blk_1073741825_1001, type=LAST_IN_PIPELINE, downstreams=0:[] terminating 2019-03-23 15:26:05,845 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-1244105038- 127.0.0.1-1553319341262:blk_1073741826_1002 src: /127.0.0.1:33416 dest: /127.0.0.1:50010 2019-03-23 15:26:12,872 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /127.0.0.1:33416, dest: /127.0.0.1:50010, bytes: 134217728, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_1340210517_1, offset: 0, srvID: fc4fe5b9- f39d-462d-adbf-73e2a2270bd7, blockid: BP-1244105038-127.0.0.1- 1553319341262:blk_1073741826_1002, duration: 7024381628 2019-03-23 15:26:12,872 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP- 1244105038-127.0.0.1-1553319341262:blk_1073741826_1002, type=LAST_IN_PIPELINE, downstreams=0:[] terminating 2019-03-23 15:26:12,963 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving BP-1244105038- 127.0.0.1-1553319341262:blk_1073741827_1003 src: /127.0.0.1:33430 dest: /127.0.0.1:50010 2019-03-23 15:26:15,344 INFO org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /127.0.0.1:33430, dest: /127.0.0.1:50010, bytes: 76389079, op: HDFS_WRITE, cliID: DFSClient_NONMAPREDUCE_1340210517_1, offset: 0, srvID: fc4fe5b9- f39d-462d-adbf-73e2a2270bd7, blockid: BP-1244105038-127.0.0.1- 1553319341262:blk_1073741827_1003, duration: 2374456847 2019-03-23 15:26:15,344 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder: BP- 1244105038-127.0.0.1-1553319341262:blk_1073741827_1003, type=LAST_IN_PIPELINE, downstreams=0:[] terminating 2019-03-23 16:10:27,767 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: IOException in offerService java.io.IOException: Failed on local exception: java.io.IOException: Connection reset by peer; Host Details : local host is: "localhost/127.0.0.1"; destination host is: "hadoop001":9000; at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:773) at org.apache.hadoop.ipc.Client.call(Client.java:1480) at org.apache.hadoop.ipc.Client.call(Client.java:1407) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke (ProtobufRpcEngine.java:229) at com.sun.proxy.$Proxy14.sendHeartbeat(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolClientSideTranslatorPB.se ndHeartbeat(DatanodeProtocolClientSideTranslatorPB.java:153) at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.sendHeartBeat (BPServiceActor.java:553) at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService (BPServiceActor.java:653) at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run (BPServiceActor.java:823) at java.lang.Thread.run(Thread.java:745) Caused by: java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223) at sun.nio.ch.IOUtil.read(IOUtil.java:197) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380) at org.apache.hadoop.net.SocketInputStream$Reader.performIO (SocketInputStream.java:57) at org.apache.hadoop.net.SocketIOWithTimeout.doIO (SocketIOWithTimeout.java:142) at org.apache.hadoop.net.SocketInputStream.read (SocketInputStream.java:161) at org.apache.hadoop.net.SocketInputStream.read (SocketInputStream.java:131) at java.io.FilterInputStream.read(FilterInputStream.java:133) at java.io.FilterInputStream.read(FilterInputStream.java:133) at org.apache.hadoop.ipc.Client$Connection$PingInputStream.read (Client.java:515) at java.io.BufferedInputStream.fill(BufferedInputStream.java:246) at java.io.BufferedInputStream.read(BufferedInputStream.java:265) at java.io.DataInputStream.readInt(DataInputStream.java:387) at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse (Client.java:1079) at org.apache.hadoop.ipc.Client$Connection.run(Client.java:974) 2019-03-23 16:10:31,749 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: hadoop001/127.0.0.1:9000. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS) 2019-03-23 16:10:32,562 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: RECEIVED SIGNAL 15: SIGTERM 2019-03-23 16:10:32,564 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at localhost/127.0.0.1 ************************************************************/

    二、创建硬链接

    [root@hadoop001 opt]# ln hadoop-root-datanode-hadoop001.log ln_hadoop [root@hadoop001 opt]# ll total 152 -rw-r--r-- 2 root root 73128 May 26 22:20 hadoop-root-datanode- hadoop001.log lrwxrwxrwx 1 root root 34 May 26 22:21 link2019 -> hadoop-root- datanode-hadoop001.log -rw-r--r-- 2 root root 73128 May 26 22:20 ln_hadoop drwxr-xr-x 2 root root 4096 Mar 4 21:42 module drwxr-xr-x 3 root root 4096 Mar 5 21:10 software

    总结一下:我们可以看出来,软链接相当于快捷方式,硬链接相当于副本

    最新回复(0)