hadoop挂载多硬盘,ZZ-- multiple disks per node
hadoop挂载多硬盘 ...multiple disks per node
http://blog.sina.com.cn/s/blog_b88e09dd01013rd4.html
Ubuntu - 硬盘分区、格式化、自动挂载配置 | Hard disk add new partition, format, auto mount in ubuntu
http://aofengblog.blog.163.com/blog/static/6317021201101502540117/
http://my.oschina.net/leejun2005/blog/290073
proper-care-and-feeding-of-drives-in-a-hadoop-cluster-a-conversation-with-stackiqs-dr-bruno
http://hortonworks.com/blog/proper-care-and-feeding-of-drives-in-a-hadoop-cluster-a-conversation-with-stackiqs-dr-bruno/
Utilizing-multiple-hard-disks-for-hadoop-HDFS
http://lucene.472066.n3.nabble.com/Utilizing-multiple-hard-disks-for-hadoop-HDFS-td3553851.html
=================
First, Hadoop requires at least two locations for storing it’s files: mapred.local.dir, where MapReduce stores intermediary files,
and dfs.data.dir, where HDFS stores the HDFS data (there are other locations as well, like hadoop.tmp.dir, where Hadoop and components stores its temporary data).
Both of them can cover multiple partitions.
While the two locations can be placed on physically different partitions, Cloudera recommends to configure them across the same set of partitions to maximize disk-level parallelism (this might not be an issue if the number of disk is much larger than the number of cores).
==
- 設定新磁碟的權限為777(也許不用這麼高),讓Hadoop擁有讀寫的權限
sudo chmod -c 777 /media/diskName - 修改conf資料夾內的hdfs-site.xml文件,在<configuration></configuration>之間加入property
The hadoop.tmp.dir property does not accept multiple paths and you should avoid using it in production - its more of a utility property that acts as a default base path for other properties.