Apache Hadoop

Apache Hadoop project develops open-source software for reliable, scalable, distributed computing.

The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage.

Hadoop modules:

Install

   1 cd ~/tmp
   2 wget https://archive.apache.org/dist/hadoop/core/hadoop-3.3.1/hadoop-3.3.1.tar.gz
   3 tar tvzf hadoop-3.3.1.tar.gz
   4 tar xvzf hadoop-3.3.1.tar.gz

~/tmp/hadoop-3.3.1/etc/hadoop/hadoop-env.sh

   1 export JAVA_HOME=/home/vitor/jdk-11.0.10+9

~/tmp/hadoop-3.3.1/etc/hadoop/core-site.xml

   1 <?xml version="1.0" encoding="UTF-8"?>
   2 <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
   3 <configuration>
   4   <property>
   5     <name>fs.default.name</name>
   6     <value>hdfs://master:9000</value>
   7   </property>
   8 </configuration>

~/tmp/hadoop-3.3.1/etc/hadoop/hdfs-site.xml

   1 <?xml version="1.0" encoding="UTF-8"?>
   2 <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
   3 <configuration>
   4   <property>
   5     <name>dfs.namenode.name.dir</name>
   6     <value>/tmp/nameNode</value>
   7   </property>
   8   <property>
   9     <name>dfs.namenode.data.dir</name>
  10     <value>/tmp/dataNode</value>
  11   </property>
  12   <property>
  13     <name>dfs.replication</name>
  14     <value>2</value>
  15    </property>
  16 </configuration>

~/tmp/hadoop-3.3.1/etc/hadoop/mapred-site.xml

   1 <?xml version="1.0"?>
   2 <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
   3 <configuration>
   4   <property>
   5     <name>mapreduce.framework.name</name>
   6     <value>yarn</value>
   7   </property>
   8 </configuration>

~/tmp/hadoop-3.3.1/etc/hadoop/yarn-site.xml

   1 <?xml version="1.0"?>
   2 <configuration>
   3   <property>
   4     <name>yarn.acl.enable</name>
   5     <value>0</value>
   6   </property>
   7   <property>
   8     <name>yarn.resourcemanager.hostname</name>
   9     <value>master</value>
  10   </property>
  11   <property>
  12     <name>yarn.nodemanager.aux-services</name>
  13     <value>mapreduce_shuffle</value>
  14   </property>
  15 </configuration>

~/tmp/hadoop-3.3.1/etc/hadoop/slaves

   1 localhost

~/.bashrc

   1 export HADOOP_HOME=/home/vitor/tmp/hadoop-3.3.1
   2 export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

Setup HDFS

   1 ls $HADOOP_HOME/bin/
   2 hdfs dfsadmin -report 
   3 # 2021-07-23 14:07:17,010 WARN fs.FileSystem: Failed to initialize fileystem hdfs://master:9000: 
   4 # java.lang.IllegalArgumentException: java.net.UnknownHostException: master
   5 # report: java.net.UnknownHostException: master
   6 # add 127.0.0.1 master to /etc/hosts
   7 hdfs namenode -format master
   8 hdfs --daemon start namenode
   9 hdfs --daemon start datanode
  10 yarn --daemon start resourcemanager
  11 yarn --daemon start nodemanager
  12 yarn --daemon start proxyserver
  13 mapred --daemon start historyserver
  14 hdfs dfsadmin -report
  15 # http://localhost:9870/
  16 # http://localhost:9870/dfshealth.html#tab-overview
  17 # http://localhost:9870/explorer.html#
  18 # http://localhost:8088/
  19 # http://localhost:8088/cluster
  20 # http://localhost:19888/
  21 hadoop fs -ls /
  22 hadoop fs -ls /tmp 
  23 hadoop fs -mkdir /test
  24 hadoop fs -ls /

ApacheHadoop (last edited 2023-05-26 13:15:20 by 127)