Install Hadoop1.2.1 in Ubuntu12.04

来源:互联网 发布:几米漫画推荐知乎 编辑:程序博客网 时间:2024/05/18 13:45

Install Hadoop1.2.1 in Ubuntu12.04

  • Install Hadoop121 in Ubuntu1204
      • 配置Java环境变量
    • 禁用ipv6
    • 配置 SSH
    • 下载解压hadoop-121
    • 配置Hadoop
    • Hadoop的运行
      • 运行Hadoop前要删除临时文件
      • 打开SSH
      • 格式化HDFS文件系统
      • 启动Hadoop环境
      • 执行Hadoop自带例子
      • 停止Hadoop守护进程
  • 进一步阅读
    • most relevant
    • less relevant

## 1.安装VirtualBox虚拟机这里用的版本是VirtualBox-4.3.20-96997-Win.exe。该虚拟机是开源的,安装文件只有100M多,与VMware相比,系统资源消耗得少。##2.安装Ubuntu在VirtualBox虚拟机上安装64位的Ubuntu12.04,具体版本 ubuntu-12.04.3-desktop-amd64。## 3.安装Java### 下载解压安装我这里的版本是jdk-7u51-linux-x64.tar.gz,将其解压到安装目录,安装后的目录如下:
root@jin-VirtualBox:~# ls /usr/java/jdk1.7.0_51/bin        jre      README.html                         THIRDPARTYLICENSEREADME.txtCOPYRIGHT  lib      releasedb         LICENSE  src.zipinclude    man      THIRDPARTYLICENSEREADME-JAVAFX.txt

配置Java环境变量

打开文件 /etc/profile ,在文件结尾处添加以下几行与上一步安装Java的目录相关的内容

JAVA_HOME=/usr/java/jdk1.7.0_51CLASSPATH=.:$JAVA_HOME/lib/tools.jar:$JAVA_HOME/lib/dt.jarPATH=$JAVA_HOME/bin:$PATHexport JAVA_HOME CLASSPATH PATH

以后每次可能用到JavaVM之前,要检查Java是否可用,可以通过打印JRE版本(java -version)和JDK版本(javac -version)的命令来达到这种效果

执行

java -version

如果打印出JRE版本信息,说明JRE环境变量有效,即类似下面的情况

root@jin-VirtualBox:/usr/local/hadoop# java -versionjava version "1.7.0_51"Java(TM) SE Runtime Environment (build 1.7.0_51-b13)Java HotSpot(TM) 64-Bit Server VM (build 24.51-b03, mixed mode)

执行

javac -version

如果打印出JDK的版本信息,说明JDK环境变量设置有效,即类似下面的情况

root@jin-VirtualBox:/usr/local/hadoop# javac -version1.7.0_51

其中任何一个的版本信息打印异常,就执行以下命令

source profile

然后再次检查Java是否可用,如果仍不能打印出Java版本信息,可能是/etc/profile 没有设置好,或者Java没有正确安装。查找原因,再次检查,直到其可用为止。

4.禁用ipv6

打开 /etc/sysctl.conf 文件,在文件末尾添加如下内容并保存

net.ipv6.conf.all.disable_ipv6 = 1net.ipv6.conf.default.disable_ipv6 = 1net.ipv6.conf.lo.disable_ipv6 = 1

重启Ubuntu系统,执行如下命令

cat$/proc/sys/net/ipv6/conf/all/disable_ipv6

如果打印 1 ,说明设置成功,ipv6已被禁用。

5.配置 SSH

  • 生成秘钥对

    root@jin-VirtualBox:/usr/local/hadoop# ssh-keygen -t rsa

  • 然后一直按键,就会按默认的选项将生成的秘钥对保存在 ~/.ssh/id_rsa 文件中。

  • 进入 .ssh目录,执行如下命令

    root@jin-VirtualBox:~/.ssh# cp id_rsa.pub authorized_keys

  • 然后执行如下命令

    ssh localhost

    如果不用输入密码,说明配置成功。
    配置SSH过程的屏幕输出记录如下:

root@jin-VirtualBox:/usr/local/hadoop# ssh-keygen -t rsaGenerating public/private rsa key pair.Enter file in which to save the key (/root/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /root/.ssh/id_rsa.Your public key has been saved in /root/.ssh/id_rsa.pub.The key fingerprint is:67:cc:ea:e5:a3:60:47:7f:cd:94:04:13:7d:dd:58:40 root@jin-VirtualBoxThe key's randomart image is:+--[ RSA 2048]----+|            ++E++||             oo +||              .. ||         o   . . ||        S =   o  ||       . =   +   ||      o o o . o  ||     . + o..     ||        o...     |+-----------------+root@jin-VirtualBox:/usr/local/hadoop# cd ~/.ssh/root@jin-VirtualBox:~/.ssh# lsid_rsa  id_rsa.pub  known_hostsroot@jin-VirtualBox:~/.ssh# cp id_rsa.pub authorized_keysroot@jin-VirtualBox:~/.ssh# cd root@jin-VirtualBox:~# ssh localhostWelcome to Ubuntu 12.04.3 LTS (GNU/Linux 3.8.0-29-generic x86_64) * Documentation:  https://help.ubuntu.com/388 packages can be updated.212 updates are security updates.New release '14.04.2 LTS' available.Run 'do-release-upgrade' to upgrade to it.Last login: Sun Mar 15 09:05:03 2015 from localhost

6.下载解压hadoop-1.2.1

这里给出个下载各个版本的Hadoop的安装文件的网址http://archive.apache.org/dist/hadoop/core/ ,我这里下载的是 hadoop-1.2.1.tar.gz。

下载后解压到安装目录,我的安装目录如下所示

root@jin-VirtualBox:/usr/local/hadoop# lsbin          google-chrome_amd64.deb       hadoop-tools-1.2.1.jar  logsbuild.xml    hadoop-ant-1.2.1.jar          input                   NOTICE.txtc++          hadoop-client-1.2.1.jar       ivy                     README.txtCHANGES.txt  hadoop-core-1.2.1.jar         ivy.xml                 sbinconf         hadoop-examples-1.2.1.jar     lib                     sharecontrib      hadoop-minicluster-1.2.1.jar  libexec                 srcdocs         hadoop-test-1.2.1.jar         LICENSE.txt             webapps

7.配置Hadoop

打开文件 /usr/local/hadoop/conf/hadoop-env.sh ,在文件末尾添加Java目录,内容如下

export JAVA_HOME=/usr/java/jdk1.7.0_51

由于我要安装为分布式(Pseudo-Distributed)的Hadoop平台,所以需要配置conf/core-site.xml、conf/hdfs-site.xml和conf/mapred-site.xml,这三个文件都在Hadoop安装目录下。下面分别是配置后的这三个文件的内容

  • conf/core-site.xml:
<?xml version="1.0"?><?xml-stylesheet type="text/xsl" href="configuration.xsl"?><!-- Put site-specific property overrides in this file. --><configuration>     <property>         <name>fs.default.name</name>         <value>hdfs://localhost:9000</value>     </property></configuration>
  • conf/hdfs-site.xml:
<?xml version="1.0"?><?xml-stylesheet type="text/xsl" href="configuration.xsl"?><!-- Put site-specific property overrides in this file. --><configuration>     <property>         <name>dfs.replication</name>         <value>1</value>     </property></configuration>
  • conf/mapred-site.xml
<?xml version="1.0"?><?xml-stylesheet type="text/xsl" href="configuration.xsl"?><!-- Put site-specific property overrides in this file. --><configuration>     <property>         <name>mapred.job.tracker</name>         <value>localhost:9001</value>     </property></configuration>

8.Hadoop的运行

运行Hadoop前要删除临时文件

如果不是第一次运行,需要删除 /tmp/* 临时文件,否则一些进程如 datanode 可能无法启动。
如下所示,/tmp/ 目录下有之前启动Hadoop所产生的一些文件,将其删除即可

root@jin-VirtualBox:/usr/local/hadoop# ls /tmp/hadoop-root                        hsperfdata_roothadoop-root-datanode.pid           Jetty_0_0_0_0_50030_job____yn7qmkhadoop-root-jobtracker.pid         Jetty_0_0_0_0_50060_task____.2vcltfhadoop-root-namenode.pid           Jetty_0_0_0_0_50070_hdfs____w2cu08hadoop-root-secondarynamenode.pid  Jetty_0_0_0_0_50075_datanode____hwtdwqhadoop-root-tasktracker.pid        Jetty_0_0_0_0_50090_secondary____y6aanvroot@jin-VirtualBox:/usr/local/hadoop# rm -rf /tmp/*

打开SSH

执行命令

ssh localhost

屏幕输出:

root@jin-VirtualBox:~# ssh localhostWelcome to Ubuntu 12.04.3 LTS (GNU/Linux 3.8.0-29-generic x86_64) * Documentation:  https://help.ubuntu.com/388 packages can be updated.212 updates are security updates.New release '14.04.2 LTS' available.Run 'do-release-upgrade' to upgrade to it.Last login: Sun Mar 15 09:05:03 2015 from localhost

格式化HDFS文件系统

执行命令:

root@jin-VirtualBox:/usr/local/hadoop# bin/hadoop namenode -format

屏幕输出:

root@jin-VirtualBox:/usr/local/hadoop# bin/hadoop namenode -format15/03/15 09:21:44 INFO namenode.NameNode: STARTUP_MSG: /************************************************************STARTUP_MSG: Starting NameNodeSTARTUP_MSG:   host = jin-VirtualBox/127.0.1.1STARTUP_MSG:   args = [-format]STARTUP_MSG:   version = 1.2.1STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2 -r 1503152; compiled by 'mattf' on Mon Jul 22 15:23:09 PDT 2013STARTUP_MSG:   java = 1.7.0_51************************************************************/15/03/15 09:21:45 INFO util.GSet: Computing capacity for map BlocksMap15/03/15 09:21:45 INFO util.GSet: VM type       = 64-bit15/03/15 09:21:45 INFO util.GSet: 2.0% max memory = 101364531215/03/15 09:21:45 INFO util.GSet: capacity      = 2^21 = 2097152 entries15/03/15 09:21:45 INFO util.GSet: recommended=2097152, actual=209715215/03/15 09:21:45 INFO namenode.FSNamesystem: fsOwner=root15/03/15 09:21:45 INFO namenode.FSNamesystem: supergroup=supergroup15/03/15 09:21:45 INFO namenode.FSNamesystem: isPermissionEnabled=true15/03/15 09:21:45 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=10015/03/15 09:21:45 INFO namenode.FSNamesystem: isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)15/03/15 09:21:45 INFO namenode.FSEditLog: dfs.namenode.edits.toleration.length = 015/03/15 09:21:45 INFO namenode.NameNode: Caching file names occuring more than 10 times 15/03/15 09:21:46 INFO common.Storage: Image file /tmp/hadoop-root/dfs/name/current/fsimage of size 110 bytes saved in 0 seconds.15/03/15 09:21:46 INFO namenode.FSEditLog: closing edit log: position=4, editlog=/tmp/hadoop-root/dfs/name/current/edits15/03/15 09:21:46 INFO namenode.FSEditLog: close success: truncate to 4, editlog=/tmp/hadoop-root/dfs/name/current/edits15/03/15 09:21:46 INFO common.Storage: Storage directory /tmp/hadoop-root/dfs/name has been successfully formatted.15/03/15 09:21:46 INFO namenode.NameNode: SHUTDOWN_MSG: /************************************************************SHUTDOWN_MSG: Shutting down NameNode at jin-VirtualBox/127.0.1.1************************************************************/

启动Hadoop环境

执行命令:

root@jin-VirtualBox:/usr/local/hadoop# bin/start-all.sh

屏幕输出:

root@jin-VirtualBox:/usr/local/hadoop# bin/start-all.shstarting namenode, logging to /usr/local/hadoop/libexec/../logs/hadoop-root-namenode-jin-VirtualBox.outroot@localhost's password: localhost: starting datanode, logging to /usr/local/hadoop/libexec/../logs/hadoop-root-datanode-jin-VirtualBox.outroot@localhost's password: localhost: starting secondarynamenode, logging to /usr/local/hadoop/libexec/../logs/hadoop-root-secondarynamenode-jin-VirtualBox.outstarting jobtracker, logging to /usr/local/hadoop/libexec/../logs/hadoop-root-jobtracker-jin-VirtualBox.outroot@localhost's password: localhost: starting tasktracker, logging to /usr/local/hadoop/libexec/../logs/hadoop-root-tasktracker-jin-VirtualBox.out

然后可以用 jps 命令查看Hadoop进程启动情况,如下

root@jin-VirtualBox:/usr/local/hadoop# jps24194 JobTracker24430 TaskTracker23854 DataNode24111 SecondaryNameNode24557 Jps23618 NameNode

jps 命令的打印输出可以看到一共有6个进程,这六个进程缺一不可,否则就意味着启动Hadoop失败,意味着前边步骤有误。

Hadoop守护进程的日志目录是${HADOOP_LOG_DIR} ,即(默认是 ${HADOOP_HOME}/logs).
可以通过浏览器查看NameNode和JobTracker,默认情况下她们的地址:
NameNode - http://localhost:50070/
JobTracker - http://localhost:50030/
下面第一张截图为JobTracker的WEB接口页面
这里写图片描述

下面一张截图是NameNode的WEB接口页面
这里写图片描述

执行Hadoop自带例子

  • 将conf文件夹的内容拷贝到分布式文件系统 的input文件,作为样例程序的输入:
bin/hadoop fs -put conf input
  • 执行样例程序:
 bin/hadoop jar hadoop-examples-*.jar grep input output 'dfs[a-z.]+'

我们执行的样例程序就是 hadoop-examples-1.2.1.jar,这是个编译好的jar包;参数input是输入,它位于的伪分布式文件系统中,供Hadoop程序调用,在本地文件是不能直接看到的;参数output是输出,它也位于分布式文件系统中,不能直接在本地系统看到。

Hadoop程序执行过程的屏幕输出内容如下:

root@jin-VirtualBox:/usr/local/hadoop# bin/hadoop jar hadoop-examples-*.jar grep input output 'dfs[a-z.]+' 15/03/15 09:23:15 INFO util.NativeCodeLoader: Loaded the native-hadoop library15/03/15 09:23:15 WARN snappy.LoadSnappy: Snappy native library not loaded15/03/15 09:23:15 INFO mapred.FileInputFormat: Total input paths to process : 1715/03/15 09:23:16 INFO mapred.JobClient: Running job: job_201503150922_000115/03/15 09:23:17 INFO mapred.JobClient:  map 0% reduce 0%15/03/15 09:23:52 INFO mapred.JobClient:  map 11% reduce 0%15/03/15 09:24:23 INFO mapred.JobClient:  map 23% reduce 0%15/03/15 09:24:34 INFO mapred.JobClient:  map 23% reduce 7%15/03/15 09:24:38 INFO mapred.JobClient:  map 35% reduce 7%15/03/15 09:24:47 INFO mapred.JobClient:  map 35% reduce 11%15/03/15 09:24:50 INFO mapred.JobClient:  map 41% reduce 11%15/03/15 09:24:52 INFO mapred.JobClient:  map 47% reduce 11%15/03/15 09:24:56 INFO mapred.JobClient:  map 47% reduce 15%15/03/15 09:25:02 INFO mapred.JobClient:  map 58% reduce 15%15/03/15 09:25:12 INFO mapred.JobClient:  map 58% reduce 19%15/03/15 09:25:15 INFO mapred.JobClient:  map 70% reduce 19%15/03/15 09:25:21 INFO mapred.JobClient:  map 82% reduce 19%15/03/15 09:25:27 INFO mapred.JobClient:  map 94% reduce 27%15/03/15 09:25:31 INFO mapred.JobClient:  map 100% reduce 27%15/03/15 09:25:36 INFO mapred.JobClient:  map 100% reduce 31%15/03/15 09:25:40 INFO mapred.JobClient:  map 100% reduce 100%15/03/15 09:25:43 INFO mapred.JobClient: Job complete: job_201503150922_000115/03/15 09:25:45 INFO mapred.JobClient: Counters: 3015/03/15 09:25:45 INFO mapred.JobClient:   Job Counters 15/03/15 09:25:45 INFO mapred.JobClient:     Launched reduce tasks=115/03/15 09:25:45 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=23015215/03/15 09:25:45 INFO mapred.JobClient:     Total time spent by all reduces waiting after reserving slots (ms)=015/03/15 09:25:45 INFO mapred.JobClient:     Total time spent by all maps waiting after reserving slots (ms)=015/03/15 09:25:45 INFO mapred.JobClient:     Launched map tasks=1715/03/15 09:25:45 INFO mapred.JobClient:     Data-local map tasks=1715/03/15 09:25:45 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=9892115/03/15 09:25:45 INFO mapred.JobClient:   File Input Format Counters 15/03/15 09:25:45 INFO mapred.JobClient:     Bytes Read=3425115/03/15 09:25:45 INFO mapred.JobClient:   File Output Format Counters 15/03/15 09:25:45 INFO mapred.JobClient:     Bytes Written=18015/03/15 09:25:45 INFO mapred.JobClient:   FileSystemCounters15/03/15 09:25:45 INFO mapred.JobClient:     FILE_BYTES_READ=8215/03/15 09:25:45 INFO mapred.JobClient:     HDFS_BYTES_READ=3608515/03/15 09:25:45 INFO mapred.JobClient:     FILE_BYTES_WRITTEN=98554615/03/15 09:25:45 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=18015/03/15 09:25:45 INFO mapred.JobClient:   Map-Reduce Framework15/03/15 09:25:45 INFO mapred.JobClient:     Map output materialized bytes=17815/03/15 09:25:45 INFO mapred.JobClient:     Map input records=95915/03/15 09:25:45 INFO mapred.JobClient:     Reduce shuffle bytes=17815/03/15 09:25:45 INFO mapred.JobClient:     Spilled Records=615/03/15 09:25:45 INFO mapred.JobClient:     Map output bytes=7015/03/15 09:25:45 INFO mapred.JobClient:     Total committed heap usage (bytes)=236340838415/03/15 09:25:45 INFO mapred.JobClient:     CPU time spent (ms)=1502015/03/15 09:25:45 INFO mapred.JobClient:     Map input bytes=3425115/03/15 09:25:45 INFO mapred.JobClient:     SPLIT_RAW_BYTES=183415/03/15 09:25:45 INFO mapred.JobClient:     Combine input records=315/03/15 09:25:45 INFO mapred.JobClient:     Reduce input records=315/03/15 09:25:45 INFO mapred.JobClient:     Reduce input groups=315/03/15 09:25:45 INFO mapred.JobClient:     Combine output records=315/03/15 09:25:45 INFO mapred.JobClient:     Physical memory (bytes) snapshot=292294656015/03/15 09:25:45 INFO mapred.JobClient:     Reduce output records=315/03/15 09:25:45 INFO mapred.JobClient:     Virtual memory (bytes) snapshot=1754535116815/03/15 09:25:45 INFO mapred.JobClient:     Map output records=315/03/15 09:25:46 INFO mapred.FileInputFormat: Total input paths to process : 115/03/15 09:25:49 INFO mapred.JobClient: Running job: job_201503150922_000215/03/15 09:25:50 INFO mapred.JobClient:  map 0% reduce 0%15/03/15 09:25:59 INFO mapred.JobClient:  map 100% reduce 0%15/03/15 09:26:13 INFO mapred.JobClient:  map 100% reduce 100%15/03/15 09:26:15 INFO mapred.JobClient: Job complete: job_201503150922_000215/03/15 09:26:15 INFO mapred.JobClient: Counters: 3015/03/15 09:26:15 INFO mapred.JobClient:   Job Counters 15/03/15 09:26:15 INFO mapred.JobClient:     Launched reduce tasks=115/03/15 09:26:15 INFO mapred.JobClient:     SLOTS_MILLIS_MAPS=958515/03/15 09:26:15 INFO mapred.JobClient:     Total time spent by all reduces waiting after reserving slots (ms)=015/03/15 09:26:15 INFO mapred.JobClient:     Total time spent by all maps waiting after reserving slots (ms)=015/03/15 09:26:15 INFO mapred.JobClient:     Launched map tasks=115/03/15 09:26:15 INFO mapred.JobClient:     Data-local map tasks=115/03/15 09:26:15 INFO mapred.JobClient:     SLOTS_MILLIS_REDUCES=1374415/03/15 09:26:15 INFO mapred.JobClient:   File Input Format Counters 15/03/15 09:26:15 INFO mapred.JobClient:     Bytes Read=18015/03/15 09:26:15 INFO mapred.JobClient:   File Output Format Counters 15/03/15 09:26:15 INFO mapred.JobClient:     Bytes Written=5215/03/15 09:26:15 INFO mapred.JobClient:   FileSystemCounters15/03/15 09:26:15 INFO mapred.JobClient:     FILE_BYTES_READ=8215/03/15 09:26:15 INFO mapred.JobClient:     HDFS_BYTES_READ=29515/03/15 09:26:15 INFO mapred.JobClient:     FILE_BYTES_WRITTEN=10794115/03/15 09:26:15 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=5215/03/15 09:26:15 INFO mapred.JobClient:   Map-Reduce Framework15/03/15 09:26:15 INFO mapred.JobClient:     Map output materialized bytes=8215/03/15 09:26:15 INFO mapred.JobClient:     Map input records=315/03/15 09:26:15 INFO mapred.JobClient:     Reduce shuffle bytes=8215/03/15 09:26:15 INFO mapred.JobClient:     Spilled Records=615/03/15 09:26:15 INFO mapred.JobClient:     Map output bytes=7015/03/15 09:26:15 INFO mapred.JobClient:     Total committed heap usage (bytes)=12327731215/03/15 09:26:15 INFO mapred.JobClient:     CPU time spent (ms)=208015/03/15 09:26:15 INFO mapred.JobClient:     Map input bytes=9415/03/15 09:26:15 INFO mapred.JobClient:     SPLIT_RAW_BYTES=11515/03/15 09:26:15 INFO mapred.JobClient:     Combine input records=015/03/15 09:26:15 INFO mapred.JobClient:     Reduce input records=315/03/15 09:26:15 INFO mapred.JobClient:     Reduce input groups=115/03/15 09:26:15 INFO mapred.JobClient:     Combine output records=015/03/15 09:26:15 INFO mapred.JobClient:     Physical memory (bytes) snapshot=24789811215/03/15 09:26:15 INFO mapred.JobClient:     Reduce output records=315/03/15 09:26:15 INFO mapred.JobClient:     Virtual memory (bytes) snapshot=195495526415/03/15 09:26:15 INFO mapred.JobClient:     Map output records=3
  • 查看Hadoop程序的输出文件内容
    有两种方法,一个是在本地文件系统上查看,一个是直接在分布式文件系统上查看。

    • 把程序输出从分布式文件系统拷贝到本地系统,然后在本地系统查看文件内容

      bin/hadoop fs -get output output
      cat output/*

    • 直接在分布式文件系统查看

      bin/hadoop fs -cat output/*

    文件内容如下:

    cat: output/_logs: Is a directory
    1 dfs.replication
    1 dfs.server.namenode.
    1 dfsadmin

停止Hadoop守护进程

每次使用Hadoop结束后不要忘了关闭Hadoop程序,命令:

root@jin-VirtualBox:/usr/local/hadoop# bin/stop-all.sh

屏幕输出:

root@jin-VirtualBox:/usr/local/hadoop# bin/stop-all.shstopping jobtrackerroot@localhost's password: localhost: stopping tasktrackerstopping namenoderoot@localhost's password: localhost: stopping datanoderoot@localhost's password: localhost: stopping secondarynamenode

进一步阅读

most relevant

  • Cluster Setup
    http://hadoop.apache.org/docs/r1.2.1/cluster_setup.html
  • Single Node Setup
    http://hadoop.apache.org/docs/r1.2.1/single_node_setup.html
  • HADOOP TUTORIALS
    http://hadooptutorials.co.in/index.html
  • INSTALL HADOOP ON UBUNTU
    http://hadooptutorials.co.in/tutorials/hadoop/install-hadoop-on-ubuntu.html#

less relevant

  • 用MapReduce实现矩阵乘法
    http://blog.fens.me/hadoop-mapreduce-matrix/
  • MapReduce实现大矩阵乘法
    http://blog.csdn.net/xyilu/article/details/9066973
  • PageRank算法并行实现
    http://blog.fens.me/algorithm-pagerank-mapreduce/
  • Ubuntu上搭建Hadoop环境(单机模式+伪分布模式)
    http://blog.csdn.net/hitwengqi/article/details/8008203
  • 运行Hadoop遇到的问题
    http://www.cnblogs.com/liangzh/archive/2012/04/06/2434602.html
  • hadoop 配置中的几个小笔记
    http://blog.csdn.net/shomy_liu/article/details/43192231
  • hadoop-2.6.0集群环境搭建
    http://blog.csdn.net/fteworld/article/details/41944597
  • Hadoop-2.6.0环境搭建精简极致指导
    http://www.linuxidc.com/Linux/2015-01/111258.htm
1 0
原创粉丝点击