Eu apenas executei isso em uma VM (Kubuntu 16.04) com a versão hadoop 2.7.3.
Pré-requisitos:
- Ubuntu OS
- Usuário não raiz com privilégios de sudo
- Java instalado
Etapas:
-
Faça o download da versão hadoop de aqui por
Verifiqueaintegridadedodownload:
Instaleohadoop:
Descomprimaemova:
tar-xzvfhadoop-2.7.3.tar.gz&&sudomvhadoop-2.7.3/usr/local/hadoop
localizeojavanoseusistemacom:
readlink-f/usr/bin/java|sed"s:bin/java::" # result /usr/lib/jvm/java-8-openjdk-amd64/jre/
-
edite o
hadoop-env.sh
filesudo nano /usr/local/hadoop/etc/hadoop/hadoop-env.sh
:-
Opção 1 (valor estático):
#export JAVA_HOME=${JAVA_HOME} export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64/jre/
-
Opção 2 (alterações no valor dinâmico com alteração no java no sistema):
#export JAVA_HOME=${JAVA_HOME} export JAVA_HOME=$(readlink -f /usr/bin/java | sed "s:bin/java::")
-
-
Executar hadoop:
-
/usr/local/hadoop/bin/hadoop
-
Saída:
Usage: hadoop [--config confdir] [COMMAND | CLASSNAME] CLASSNAME run the class named CLASSNAME or where COMMAND is one of: fs run a generic filesystem user client version print the version jar <jar> run a jar file note: please use "yarn jar" to launch YARN applications, not this command. checknative [-a|-h] check native hadoop and compression libraries availability distcp <srcurl> <desturl> copy file or directories recursively archive -archiveName NAME -p <parent path> <src>* <dest> create a hadoop archive classpath prints the class path needed to get the credential interact with credential providers Hadoop jar and the required libraries daemonlog get/set the log level for each daemon
-
-
-
Teste:
mkdir ~/input cp /usr/local/hadoop/etc/hadoop/*.xml ~/input /usr/local/hadoop/bin/hadoop jar /usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.3.jar grep ~/input ~/grep_example 'principal[.]*';
-
Verifique se há erros no resultado:
File System Counters FILE: Number of bytes read=1247674 FILE: Number of bytes written=2324248 FILE: Number of read operations=0 FILE: Number of large read operations=0 FILE: Number of write operations=0 Map-Reduce Framework Map input records=2 Map output records=2 Map output bytes=37 Map output materialized bytes=47 Input split bytes=114 Combine input records=0 Combine output records=0 Reduce input groups=2 Reduce shuffle bytes=47 Reduce input records=2 Reduce output records=2 Spilled Records=4 Shuffled Maps =1 Failed Shuffles=0 Merged Map outputs=1 GC time elapsed (ms)=61 Total committed heap usage (bytes)=263520256 Shuffle Errors BAD_ID=0 CONNECTION=0 IO_ERROR=0 WRONG_LENGTH=0 WRONG_MAP=0 WRONG_REDUCE=0 File Input Format Counters Bytes Read=151 File Output Format Counters Bytes Written=37
-
Fonte:
Adicione o caminho para .bashrc:
# hadoop executable
export PATH="/usr/local/hadoop/bin:$PATH"