Hadoop Installation Guide
Description : bashrc is a shell script file that Bash runs when it is started interactively. we can put any command or script in this file which you want to type from the command prompt. we can insert commands here to set up the shell for use in our particular environment, or to customize things as per our preferences.
Hadoop is a framework which is written in Java which runs the applications on large clusters of commodity hardware(General Purpose Hardware) which is very similar to the Google File System (GFS). Hadoop Distributed File System is a highly fault-tolerant system which is design to used on low budget hardware.It provides high throughput access to application data and is suitable for applications that have large data sets.
We uses the dedicated user for installing the hadoop. But this is not recomended but it helps to separate the Hadoop installation from other software applications and user accounts running on the same machine which isolate security, permission and backup.
The sudo command is used to elevate your permissions for a linux command.we can use the sudo command to run a command as any other user although it is commonly used to run a command as the root user.If we have multiple users on our computer then we probably don't want all of the users to be administrators because administrators can do things like installing or uninstalling the sytem applications or can change key system settings.
Hadoop requires SSH access to manage its nodes which include remote machines and local machine if we want to use Hadoop on it For our single-node setup of Hadoop, we need to configure SSH access to
localhost for the
hadoopusr.Ssh-keygen -t rsa -P “ ”
~/.ssh/authorized_keys are keys of other computers that you connected to/trust, not our own key.we have to add the public key of our computer to the
authorized_keys file of the computer we want to access using SSH Keys.
-p parent directory