用户登录
用户注册

分享至

hadoop for windows

  • 作者: 我想做你的狗
  • 来源: 51数据库
  • 2020-10-06
环境及软件准备:
win7(64位)

cygwin 1.7.9-1

jdk-6u25-windows-x64.zip

hadoop-0.20.2.tar.gz

1.安装jdk,并置java环境变量包括:JAVA_HOME,PATH,CLASSPATH

2.安装Hadoop,版本为0.20.2,我是直接放到/home目录下,并解压
tar –zxvf
hadoop-0.20.2.tar.gz

3.配置Hadoop,需要修改hadoop的配置文件,它们位于conf子目录下,分别是hadoop-env.sh、core-site.xml、hdfs-site.xml
和mapred-site.xml

(1) 修改hadoop-env.sh:

只需要将JAVA_HOME 修改成JDK 的安装目录即可

export
JAVA_HOME=/cygdrive/d/java/jdk1.6.0_25

(注意:路径不能是windows 风格的目录d:\java\jdk1.6.0_25,而是LINUX
风格/cygdrive/d/java/jdk1.6.0_25)

(2) 修改core-site.xml:(指定namenode)





fs.default.name

hdfs://localhost:9000




(3)修改hdfs-site.xml(指定副本为1)





dfs.replication

1




(4) 修改mapred-site.xml (指定jobtracker)





mapred.job.tracker

localhost:9001




4.验证安装是否成功,并运行Hadoop

(1) 验证安装

$ bin/hadoop

Usage: hadoop [--config confdir] COMMAND

where COMMAND is one of:

namenode -format format the DFS filesystem

secondarynamenode run the DFS secondary namenode

namenode run the DFS namenode

datanode run a DFS datanode

dfsadmin run a DFS admin client

mradmin run a Map-Reduce admin client

fsck run a DFS filesystem checking utility

fs run a generic filesystem user client

balancer run a cluster balancing utility

jobtracker run the MapReduce job Tracker node

pipes run a Pipes job

tasktracker run a MapReduce task Tracker node

job manipulate MapReduce jobs

queue get information regarding JobQueues

version print the version

jar run a jar file

distcp copy file or directories recursively

archive -archiveName NAME * create a hadoop archive

daemonlog get/set the log level for each daemon

or

CLASSNAME run the class named CLASSNAME

Most commands print help when invoked w/o parameters.
(2) 格式化并启动Hadoop

bi




  额
软件
前端设计
程序设计
Java相关