This appendix explains the configuration parameters for Hadoop provided by this product.
Some parameters are configured automatically when setting up this product. This includes parameters for Hadoop to use a DFS and parameters for optimizing parallel distributed processing of Hadoop jobs.
This appendix shows parameters that are changed from default values and that are added to OS setup files within the setup files (excluding pdfs-site.xml) provided by the open source software Apache Hadoop.
Setup file | Description | Directory |
---|---|---|
Setup file for defining environment variables used in Hadoop | /etc/hadoop | |
Common setup file for Hadoop | /etc/hadoop | |
MapReduce-related setup file | /etc/hadoop | |
DFS-related setup file | /etc/hadoop | |
OS setup file specifying kernel parameters | /etc | |
OS setup file for limiting system resources | /etc/security |
Note
This appendix does not provide detailed explanation of all the parameters in setup files. If necessary, refer to a resource such as the Apache Hadoop project website.
If tuning the configuration settings, make changes immediately following the setup. This excludes when changes are made to the number of slave servers (see "12.1.7 Changing Hadoop Configuration Parameters").
Note that if any changes need to be made to the Hadoop configuration parameters after operations begin, settings will need to be reviewed for each master server, slave server, and development server.