EXCEPTIONAL TRAINING COURSES FOR IT PROFESSIONALS
LONDON UK | ONSITE | ONLINE
In this Hadoop architecture and administration training course, you gain the skills to install, configure, and manage the Apache Hadoop platform and its associated ecosystem, and build a Hadoop solution that satisfies your business requirements.
IT professionals looking to learn about how to architect and administer Apache Hadoop and clusters for Big Data
Installing the Hadoop Distributed File System (HDFS)
Defining key design assumptions and architecture
Configuring and setting up the file system
Issuing commands from the console
Reading and writing files
Setting the stage for MapReduce
Reviewing the MapReduce approach
Introducing the computing daemons
Dissecting a MapReduce job
Planning the architecture
Selecting appropriate hardware
Designing a scalable cluster
Building the cluster
Installing Hadoop daemons
Optimising the network architecture
Setting basic configuration parameters
Configuring block allocation, redundancy and replication
Installing and setting up the MapReduce environment
Delivering redundant load balancing via Rack Awareness
Creating a fault–tolerant file system
Isolating single points of failure
Maintaining High Availability
Triggering manual failover
Automating failover with Zookeeper
Leveraging NameNode Federation
Extending HDFS resources
Managing the namespace volumes
Critiquing the YARN architecture
Identifying the new daemons
Setting quotas to constrain HDFS utilisation
Prioritising access to MapReduce using schedulers
Starting and stopping Hadoop daemons
Monitoring HDFS status
Adding and removing data nodes
Managing MapReduce jobs
Tracking progress with monitoring tools
Commissioning and decommissioning compute nodes
Employing the standard built–in tools
Managing and debugging processes using JVM metrics
Performing Hadoop status checks
Tuning with supplementary tools
Assessing performance with Ganglia
Benchmarking to ensure continued performance
Simplifying information access
Enabling SQL–like querying with Hive
Installing Pig to create MapReduce jobs
Integrating additional elements of the ecosystem
Imposing a tabular view on HDFS with HBase
Configuring Oozie to schedule workflows
Facilitating generic input/output
Moving bulk data into and out of Hadoop
Transmitting HDFS data over HTTP with WebHDFS
Acquiring application–specific data
Collecting multi–sourced log files with Flume
Importing and exporting relational information with Sqoop
Planning for Backup, Recovery and Security
Coping with inevitable hardware failures
Securing your Hadoop cluster
See why people choose JBI
Call our Tech Team : 0800 028 6400 request info
Bring a JBI course to your office
and train a whole team onsite
0800 028 6400 or request quote
Get in touch
0800 028 6400
The training was to target a team of developers, with a mix of Java, C# and some C++. The team made heavy use of Java SE with some elements of...
The Graduate Programme provided a gateway into technology within investment banking. Graduates (Computer Science, Engineering, Maths, Physics...
This organisation needed their Supply Chain department to get fully involved with Microsoft’s Power BI reporting product as soon as possible....
Our inspiring and up-to-date courses gain exceptional feedback
"great tips to help reduce build times"
"we got access to exclusive content"
"intensive course meant less time off"
"what an inspiring trainer !"
"colleagues at 2 sites joined via web"
"I passed my exam the next day"
Microsoft recently released the public road map that gives the holistic approach to Integration and the key product offerings that contribute...
Scrum is one of the most popular frameworks for implementing agile development, and many people are doing a scrum course to understand how to...