Copyright 2018 ABSA Group Limited
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
(mostly) Windows related remarks
-
Have Spark with Hadoop support, still need to have Hadoop as its own separate installation
-
Installing Hadoop on Windows or How to Install Hadoop on Windows
-
Environment variables:
CATALINA_BASE
CATALINA_HOME
HADOOP
HADOOP_BIN
HADOOP_CONF_DIR
HADOOP_HOME
JAVA_HOME
-
use slashes in paths except when specifying the jar file for Standardization/Conformance
-
Enceladus REST uri is with
/api
suffix after the war file - e.g.http://localhost:8080/rest_api/api
Hadoop | Startup |
---|
Spin up of Hadoop fails with:
'-classpath' is not recognized as an internal or external command, operable program or batch file.
The paths for Java and Hadoop env variables ought not to have any space (or other white characters) within. Eg. use JAVA_HOME=C:\Progra~1\Java\jdk1.8.0_191
instead of JAVA_HOME=C:\Program Files\Java\jdk1.8.0_191
Hadoop | Startup |
---|
Spin up of Hadoop fails with:
FATAL org.apache.hadoop.hdfs.server.namenode.NameNode:
Exception in namenode join java.net.BindException: Port in use: 0.0.0.0:50070`
Set Hadoop dfs.http.address in core-site.xml and dfs.namenode.http-address in hdfs-site.xml to different values. Eg.:
<property>
<name>dfs.http.address</name>
<value>localhost:50170</value>
</property>
<property>
<name>dfs.namenode.http-address</name>
<value>localhost:50070</value>
<description>
The address and the base port where the dfs namenode web ui will listen on.
</description>
</property>
Hdfs | Command-line |
---|
On a freshly formatted node upon ls
:
hdfs dfs -ls ls: `.': No such file or directory
The default directory for the connected user is /user/%USERNAME%
where %USERNAME%
is of the connected user. But this
directory does not exist by default. Execute: hdfs dfs -mkdir /user/%USERNAME%
to create the home directory - if that
is needed and/or use absolute path to access the files, e.g. hdfs dfs -ls /