Top
Interstage Big DataParallel Processing ServerV1.0.1 User's Guide
FUJITSU Software

Appendix C Hadoop Configuration Parameters

This appendix explains the configuration parameters for Hadoop provided by this product.

Some parameters are configured automatically when setting up this product. This includes parameters for Hadoop to use a DFS and parameters for optimizing parallel distributed processing of Hadoop jobs.

This appendix shows parameters that are changed from default values and that are added to OS setup files within the setup files (excluding pdfs-site.xml) provided by the open source software Apache Hadoop.

List of setup files

Setup file

Description

Directory

hadoop-env.sh

Setup file for defining environment variables used in Hadoop

/etc/hadoop

core-site.xml

Common setup file for Hadoop

/etc/hadoop

mapred-site.xml

MapReduce-related setup file

/etc/hadoop

pdfs-site.xml

DFS-related setup file

/etc/hadoop

sysctl.conf

OS setup file specifying kernel parameters

/etc

limits.conf

OS setup file for limiting system resources

/etc/security

Note

  • This appendix does not provide detailed explanation of all the parameters in setup files. If necessary, refer to a resource such as the Apache Hadoop project website.

  • If tuning the configuration settings, make changes immediately following the setup. This excludes when changes are made to the number of slave servers (see "12.1.7 Changing Hadoop Configuration Parameters").
    Note that if any changes need to be made to the Hadoop configuration parameters after operations begin, settings will need to be reviewed for each master server, slave server, and development server.