in normal account.
i created directories.
/usr/local/hadoop-2.7.3/data/dfs/namenode /usr/local/hadoop-2.7.3/data/dfs/namesecondary /usr/local/hadoop-2.7.3/data/dfs/datanode /usr/local/hadoop-2.7.3/data/yarn/nm-local-dir /usr/local/hadoop-2.7.3/data/yarn/system/rmstore and typed commands
bin/hdfs namenode –format sudo sbin/start-all.sh jps then
in normal account, see jps.
in root account, see jps, datanode, secondarynamenode, nodemanager , resourcemanager.
i have 2 questions.
- why can see jps in normal account?
- why namenode not started?
thanks reading. , if me, appreciate you.
namenode log file
2017-04-06 01:16:15,217 info org.apache.hadoop.hdfs.server.namenode.namenode: registered unix signal handlers [term, hup, int] 2017-04-06 01:16:15,220 info org.apache.hadoop.hdfs.server.namenode.namenode: createnamenode [] 2017-04-06 01:16:15,680 info org.apache.hadoop.metrics2.impl.metricsconfig: loaded properties hadoop-metrics2.properties 2017-04-06 01:16:15,843 info org.apache.hadoop.metrics2.impl.metricssystemimpl: scheduled snapshot period @ 10 second(s). 2017-04-06 01:16:15,843 info org.apache.hadoop.metrics2.impl.metricssystemimpl: namenode metrics system started 2017-04-06 01:16:15,845 info org.apache.hadoop.hdfs.server.namenode.namenode: fs.defaultfs hdfs://localhost:9010 2017-04-06 01:16:15,846 info org.apache.hadoop.hdfs.server.namenode.namenode: clients use localhost:9010 access namenode/service. 2017-04-06 01:16:16,070 info org.apache.hadoop.hdfs.dfsutil: starting web-server hdfs at: http://localhost:50070 2017-04-06 01:16:16,152 info org.mortbay.log: logging org.slf4j.impl.log4jloggeradapter(org.mortbay.log) via org.mortbay.log.slf4jlog 2017-04-06 01:16:16,158 info org.apache.hadoop.security.authentication.server.authenticationfilter: unable initialize filesignersecretprovider, falling use random secrets. 2017-04-06 01:16:16,165 info org.apache.hadoop.http.httprequestlog: http request log http.requests.namenode not defined 2017-04-06 01:16:16,169 info org.apache.hadoop.http.httpserver2: added global filter 'safety' (class=org.apache.hadoop.http.httpserver2$quotinginputfilter) 2017-04-06 01:16:16,171 info org.apache.hadoop.http.httpserver2: added filter static_user_filter (class=org.apache.hadoop.http.lib.staticuserwebfilter$staticuserfilter) context hdfs 2017-04-06 01:16:16,171 info org.apache.hadoop.http.httpserver2: added filter static_user_filter (class=org.apache.hadoop.http.lib.staticuserwebfilter$staticuserfilter) context logs 2017-04-06 01:16:16,171 info org.apache.hadoop.http.httpserver2: added filter static_user_filter (class=org.apache.hadoop.http.lib.staticuserwebfilter$staticuserfilter) context static 2017-04-06 01:16:16,300 info org.apache.hadoop.http.httpserver2: added filter 'org.apache.hadoop.hdfs.web.authfilter' (class=org.apache.hadoop.hdfs.web.authfilter) 2017-04-06 01:16:16,303 info org.apache.hadoop.http.httpserver2: addjerseyresourcepackage: packagename=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, pathspec=/webhdfs/v1/* 2017-04-06 01:16:16,330 info org.apache.hadoop.http.httpserver2: jetty bound port 50070 2017-04-06 01:16:16,330 info org.mortbay.log: jetty-6.1.26 2017-04-06 01:16:16,581 info org.mortbay.log: started httpserver2$selectchannelconnectorwithsafestartup@localhost:50070 2017-04-06 01:16:16,612 warn org.apache.hadoop.hdfs.server.common.util: path /usr/local/hadoop-2.7.3/data/dfs/namenode should specified uri in configuration files. please update hdfs configuration. 2017-04-06 01:16:16,612 warn org.apache.hadoop.hdfs.server.common.util: path /usr/local/hadoop-2.7.3/data/dfs/namenode should specified uri in configuration files. please update hdfs configuration. 2017-04-06 01:16:16,613 warn org.apache.hadoop.hdfs.server.namenode.fsnamesystem: 1 image storage directory (dfs.namenode.name.dir) configured. beware of data loss due lack of redundant storage directories! 2017-04-06 01:16:16,613 warn org.apache.hadoop.hdfs.server.namenode.fsnamesystem: 1 namespace edits storage directory (dfs.namenode.edits.dir) configured. beware of data loss due lack of redundant storage directories! 2017-04-06 01:16:16,617 warn org.apache.hadoop.hdfs.server.common.util: path /usr/local/hadoop-2.7.3/data/dfs/namenode should specified uri in configuration files. please update hdfs configuration. 2017-04-06 01:16:16,617 warn org.apache.hadoop.hdfs.server.common.util: path /usr/local/hadoop-2.7.3/data/dfs/namenode should specified uri in configuration files. please update hdfs configuration. 2017-04-06 01:16:16,639 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: no keyprovider found. 2017-04-06 01:16:16,639 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: fslock fair:true 2017-04-06 01:16:16,668 info org.apache.hadoop.hdfs.server.blockmanagement.datanodemanager: dfs.block.invalidate.limit=1000 2017-04-06 01:16:16,668 info org.apache.hadoop.hdfs.server.blockmanagement.datanodemanager: dfs.namenode.datanode.registration.ip-hostname-check=true 2017-04-06 01:16:16,669 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: dfs.namenode.startup.delay.block.deletion.sec set 000:00:00:00.000 2017-04-06 01:16:16,669 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: block deletion start around 2017 apr 06 01:16:16 2017-04-06 01:16:16,670 info org.apache.hadoop.util.gset: computing capacity map blocksmap 2017-04-06 01:16:16,670 info org.apache.hadoop.util.gset: vm type = 64-bit 2017-04-06 01:16:16,671 info org.apache.hadoop.util.gset: 2.0% max memory 966.7 mb = 19.3 mb 2017-04-06 01:16:16,671 info org.apache.hadoop.util.gset: capacity = 2^21 = 2097152 entries 2017-04-06 01:16:16,690 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: dfs.block.access.token.enable=false 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: defaultreplication = 1 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: maxreplication = 512 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: minreplication = 1 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: maxreplicationstreams = 2 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: replicationrecheckinterval = 3000 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: encryptdatatransfer = false 2017-04-06 01:16:16,691 info org.apache.hadoop.hdfs.server.blockmanagement.blockmanager: maxnumblockstolog = 1000 2017-04-06 01:16:16,706 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: fsowner = root (auth:simple) 2017-04-06 01:16:16,707 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: supergroup = supergroup 2017-04-06 01:16:16,707 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: ispermissionenabled = true 2017-04-06 01:16:16,707 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: ha enabled: false 2017-04-06 01:16:16,708 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: append enabled: true 2017-04-06 01:16:16,963 info org.apache.hadoop.util.gset: computing capacity map inodemap 2017-04-06 01:16:16,963 info org.apache.hadoop.util.gset: vm type = 64-bit 2017-04-06 01:16:16,970 info org.apache.hadoop.util.gset: 1.0% max memory 966.7 mb = 9.7 mb 2017-04-06 01:16:16,970 info org.apache.hadoop.util.gset: capacity = 2^20 = 1048576 entries 2017-04-06 01:16:16,971 info org.apache.hadoop.hdfs.server.namenode.fsdirectory: acls enabled? false 2017-04-06 01:16:16,971 info org.apache.hadoop.hdfs.server.namenode.fsdirectory: xattrs enabled? true 2017-04-06 01:16:16,971 info org.apache.hadoop.hdfs.server.namenode.fsdirectory: maximum size of xattr: 16384 2017-04-06 01:16:16,971 info org.apache.hadoop.hdfs.server.namenode.namenode: caching file names occuring more 10 times 2017-04-06 01:16:16,977 info org.apache.hadoop.util.gset: computing capacity map cachedblocks 2017-04-06 01:16:16,977 info org.apache.hadoop.util.gset: vm type = 64-bit 2017-04-06 01:16:16,977 info org.apache.hadoop.util.gset: 0.25% max memory 966.7 mb = 2.4 mb 2017-04-06 01:16:16,977 info org.apache.hadoop.util.gset: capacity = 2^18 = 262144 entries 2017-04-06 01:16:16,978 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033 2017-04-06 01:16:16,978 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: dfs.namenode.safemode.min.datanodes = 0 2017-04-06 01:16:16,978 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: dfs.namenode.safemode.extension = 30000 2017-04-06 01:16:16,980 info org.apache.hadoop.hdfs.server.namenode.top.metrics.topmetrics: nntop conf: dfs.namenode.top.window.num.buckets = 10 2017-04-06 01:16:16,980 info org.apache.hadoop.hdfs.server.namenode.top.metrics.topmetrics: nntop conf: dfs.namenode.top.num.users = 10 2017-04-06 01:16:16,980 info org.apache.hadoop.hdfs.server.namenode.top.metrics.topmetrics: nntop conf: dfs.namenode.top.windows.minutes = 1,5,25 2017-04-06 01:16:16,983 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: retry cache on namenode enabled 2017-04-06 01:16:16,983 info org.apache.hadoop.hdfs.server.namenode.fsnamesystem: retry cache use 0.03 of total heap , retry cache entry expiry time 600000 millis 2017-04-06 01:16:16,984 info org.apache.hadoop.util.gset: computing capacity map namenoderetrycache 2017-04-06 01:16:16,984 info org.apache.hadoop.util.gset: vm type = 64-bit 2017-04-06 01:16:16,984 info org.apache.hadoop.util.gset: 0.029999999329447746% max memory 966.7 mb = 297.0 kb 2017-04-06 01:16:16,984 info org.apache.hadoop.util.gset: capacity = 2^15 = 32768 entries 2017-04-06 01:16:17,005 info org.apache.hadoop.hdfs.server.common.storage: lock on /usr/local/hadoop-2.7.3/data/dfs/namenode/in_use.lock acquired nodename 5360@localhost 2017-04-06 01:16:17,007 warn org.apache.hadoop.hdfs.server.namenode.fsnamesystem: encountered exception loading fsimage java.io.ioexception: namenode not formatted. at org.apache.hadoop.hdfs.server.namenode.fsimage.recovertransitionread(fsimage.java:225) at org.apache.hadoop.hdfs.server.namenode.fsnamesystem.loadfsimage(fsnamesystem.java:975) at org.apache.hadoop.hdfs.server.namenode.fsnamesystem.loadfromdisk(fsnamesystem.java:681) at org.apache.hadoop.hdfs.server.namenode.namenode.loadnamesystem(namenode.java:585) at org.apache.hadoop.hdfs.server.namenode.namenode.initialize(namenode.java:645) at org.apache.hadoop.hdfs.server.namenode.namenode.<init>(namenode.java:812) at org.apache.hadoop.hdfs.server.namenode.namenode.<init>(namenode.java:796) at org.apache.hadoop.hdfs.server.namenode.namenode.createnamenode(namenode.java:1493) at org.apache.hadoop.hdfs.server.namenode.namenode.main(namenode.java:1559) 2017-04-06 01:16:17,032 info org.mortbay.log: stopped httpserver2$selectchannelconnectorwithsafestartup@localhost:50070 2017-04-06 01:16:17,035 warn org.apache.hadoop.http.httpserver2: httpserver acceptor: isrunning false. rechecking. 2017-04-06 01:16:17,035 warn org.apache.hadoop.http.httpserver2: httpserver acceptor: isrunning false 2017-04-06 01:16:17,035 info org.apache.hadoop.metrics2.impl.metricssystemimpl: stopping namenode metrics system... 2017-04-06 01:16:17,035 info org.apache.hadoop.metrics2.impl.metricssystemimpl: namenode metrics system stopped. 2017-04-06 01:16:17,035 info org.apache.hadoop.metrics2.impl.metricssystemimpl: namenode metrics system shutdown complete. 2017-04-06 01:16:17,035 error org.apache.hadoop.hdfs.server.namenode.namenode: failed start namenode. java.io.ioexception: namenode not formatted. at org.apache.hadoop.hdfs.server.namenode.fsimage.recovertransitionread(fsimage.java:225) at org.apache.hadoop.hdfs.server.namenode.fsnamesystem.loadfsimage(fsnamesystem.java:975) at org.apache.hadoop.hdfs.server.namenode.fsnamesystem.loadfromdisk(fsnamesystem.java:681) at org.apache.hadoop.hdfs.server.namenode.namenode.loadnamesystem(namenode.java:585) at org.apache.hadoop.hdfs.server.namenode.namenode.initialize(namenode.java:645) at org.apache.hadoop.hdfs.server.namenode.namenode.<init>(namenode.java:812) at org.apache.hadoop.hdfs.server.namenode.namenode.<init>(namenode.java:796) at org.apache.hadoop.hdfs.server.namenode.namenode.createnamenode(namenode.java:1493) at org.apache.hadoop.hdfs.server.namenode.namenode.main(namenode.java:1559) 2017-04-06 01:16:17,036 info org.apache.hadoop.util.exitutil: exiting status 1 2017-04-06 01:16:17,040 info org.apache.hadoop.hdfs.server.namenode.namenode: shutdown_msg:
why can see jps in normal account?
as have started daemons sudo, root user owns processes. command jps reports jvms has access permissions. normal account has no access processes owned root.
why namenode not started?
java.io.ioexception: namenode not formatted. namenode not yet formatted. possible have missed provide y when format command prompted (y/n).
Comments
Post a Comment