Artificial Intelligence
Articles
eBooks
Interview Questions
Videos
Keras
Articles
Create Model using both Sequential and Functional API in Keras
Deep dive into Keras - Convolution Neural Network (CNN)
Deep Dive into Modules provided by Keras Library
Detail understanding of Keras Layers
Detailed understanding of Keras applications
Detailed understanding of the Keras Model compilation process
How Keras help in Deep Learning and Architecture of Keras Library
How to write a simple MPL based Artificial Neural Network to perform regression prediction.
Keras Backend Implementations and overview of Deep Learning
Overview of Deep Learning Library Keras and How to install Keras Library on your machine
Write a simple Long Short-Term Memory (LSTM) based RNN to do sequence analysis
eBooks
Interview Questions
Videos
Create Model using both Sequential and Functional API in Keras
Deep dive into Keras - Convolution Neural Network (CNN)
Deep Dive into Modules provided by Keras Library
Detail understanding of Keras Layers
Detailed understanding of Keras applications
Detailed understanding of the Keras Model compilation process
How Keras help in Deep Learning and Architecture of Keras Library
How to write a simple MPL based Artificial Neural Network to perform regression prediction.
Keras Backend Implementations and overview of Deep Learning
Overview of Deep Learning Library Keras and How to install Keras Library on your machine
Write a simple Long Short-Term Memory (LSTM) based RNN to do sequence analysis
Tensor Flow
Articles
Concept of Agents and Environments in AI
Hidden Layer Perceptron in TensorFlow
Multi-layer Perceptron in TensorFlow
Concept of Fuzzy Logic Systems
Deep Dive into TensorFlow Playground
Difference between TensorFlow and Keras
Difference between TensorFlow and PyTorch
Difference between TensorFlow and Theano
How to Install TensorFlow Through pip in Windows
Idea of Intelligence and components of Intelligence
Implementation of Neural Network in TensorFlow
Linear Regression in TensorFlow
Machine Learning and Deep Learning
How to Install TensorFlow through Anaconda
Introduction of Convolutional Neural Network in TensorFlow
Long short-term memory (LSTM) RNN in TensorFlow
What are Artificial Neural Networks?
Working of Convolutional Neural Network
Advantages and Disadvantages of TensorFlow
Architecture of TensorFlow explained
AI - Popular Search Algorithms
Artificial Intelligence - Research Areas
Artificial Neural Network in TensorFlow
CIFAR-10 and CIFAR-100 Dataset in TensorFlow
Classification of Neural Network in TensorFlow
TensorFlow Single and Multiple GPU
TensorFlow Security and TensorFlow Vs Caffe
Style Transferring in TensorFlow
Single Layer Perceptron in TensorFlow
Robotics in Artificial Intelligence
Recurrent Neural Network (RNN) in TensorFlow
eBooks
Interview Questions
Videos
Concept of Agents and Environments in AI
Hidden Layer Perceptron in TensorFlow
Multi-layer Perceptron in TensorFlow
Concept of Fuzzy Logic Systems
Deep Dive into TensorFlow Playground
Difference between TensorFlow and Keras
Difference between TensorFlow and PyTorch
Difference between TensorFlow and Theano
How to Install TensorFlow Through pip in Windows
Idea of Intelligence and components of Intelligence
Implementation of Neural Network in TensorFlow
Linear Regression in TensorFlow
Machine Learning and Deep Learning
How to Install TensorFlow through Anaconda
Introduction of Convolutional Neural Network in TensorFlow
Long short-term memory (LSTM) RNN in TensorFlow
What are Artificial Neural Networks?
Working of Convolutional Neural Network
Advantages and Disadvantages of TensorFlow
Architecture of TensorFlow explained
AI - Popular Search Algorithms
Artificial Intelligence - Research Areas
Artificial Neural Network in TensorFlow
CIFAR-10 and CIFAR-100 Dataset in TensorFlow
Classification of Neural Network in TensorFlow
TensorFlow Single and Multiple GPU
TensorFlow Security and TensorFlow Vs Caffe
Style Transferring in TensorFlow
Single Layer Perceptron in TensorFlow
Robotics in Artificial Intelligence
Recurrent Neural Network (RNN) in TensorFlow
Data Science Introduction and How to set up python
How to use Hibernate Query Language
Handling Arrays and Strings in PHP
Cookies and Sessions Handling in PHP
Cookies and Sessions Handling in PHP
Concept of Agents and Environments in AI
Hidden Layer Perceptron in TensorFlow
Multi-layer Perceptron in TensorFlow
Concept of Fuzzy Logic Systems
Deep Dive into TensorFlow Playground
Difference between TensorFlow and Keras
Difference between TensorFlow and PyTorch
Difference between TensorFlow and Theano
How to Install TensorFlow Through pip in Windows
Idea of Intelligence and components of Intelligence
Implementation of Neural Network in TensorFlow
Linear Regression in TensorFlow
Machine Learning and Deep Learning
How to Install TensorFlow through Anaconda
Introduction of Convolutional Neural Network in TensorFlow
Long short-term memory (LSTM) RNN in TensorFlow
What are Artificial Neural Networks?
Working of Convolutional Neural Network
Advantages and Disadvantages of TensorFlow
Architecture of TensorFlow explained
AI - Popular Search Algorithms
Artificial Intelligence - Research Areas
Artificial Neural Network in TensorFlow
CIFAR-10 and CIFAR-100 Dataset in TensorFlow
Classification of Neural Network in TensorFlow
Create Model using both Sequential and Functional API in Keras
Deep dive into Keras - Convolution Neural Network (CNN)
Deep Dive into Modules provided by Keras Library
Detail understanding of Keras Layers
Detailed understanding of Keras applications
Detailed understanding of the Keras Model compilation process
How Keras help in Deep Learning and Architecture of Keras Library
How to write a simple MPL based Artificial Neural Network to perform regression prediction.
Keras Backend Implementations and overview of Deep Learning
Overview of Deep Learning Library Keras and How to install Keras Library on your machine
Write a simple Long Short-Term Memory (LSTM) based RNN to do sequence analysis
Model Evaluation and Model Prediction in Keras
How to create our own Customized Layer in Keras Library
TensorFlow Single and Multiple GPU
TensorFlow Security and TensorFlow Vs Caffe
Style Transferring in TensorFlow
Single Layer Perceptron in TensorFlow
Robotics in Artificial Intelligence
Recurrent Neural Network (RNN) in TensorFlow
Overview of Artificial Intelligence and its Application
Natural Language Processing in AI
Test Article Main differences between Selenium RC and Selenium WebDriver - Dont Delete
Top Artificial Intelligence Interview Questions and Answers
Top Oracle DBA Interview Questions and Answers
Basics of Splunk and Installation of Splunk Environment
Microsoft Azure Solutions Architect Certification Exam Questions (AZ-300 & AZ-301)
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
BI and Visualization
Articles
eBooks
Videos
Cognos Analytics
Articles
Perform Report Operations in IBM Cognos
Introduction to IBM Cognos and its Components and Services
How to open, create, save, run and print report in Cognos
How to open, create and save Analysis in Analysis Studio in Cognos
How to create report in Report Studio
How to create List and CrossTab Report in Cognos
How to create a package using Cognos
Filters and Custom Calculations in Cognos
Data Warehouse Schemas, ETL and Reporting Tools
Cognos Studios and other capabilities
eBooks
Videos
Perform Report Operations in IBM Cognos
Introduction to IBM Cognos and its Components and Services
How to open, create, save, run and print report in Cognos
How to open, create and save Analysis in Analysis Studio in Cognos
How to create report in Report Studio
How to create List and CrossTab Report in Cognos
How to create a package using Cognos
Filters and Custom Calculations in Cognos
Data Warehouse Schemas, ETL and Reporting Tools
Cognos Studios and other capabilities
Cognos - Relationships in Metadata Model
Top Tableau Desktop Interview Questions and Answers
Top Tableau Server Interview Questions and Answers
Top Power BI Interview Questions and Answers
Top Cognos TM1 Interview Questions and Answers
Cognos TM1
eBooks
Interview Questions
Videos
Microsoft Excel
Articles
How to Merge & Wrap Cells, Borders and Shades and Apply Formatting in Excel
BackStage View and Explore Window in Excel
Creating Formulas, Copying Formulas in Excel
Data Sorting and Using Ranges in Excel
Data Tables and Pivot Tables in Excel
Excel Fill Handle and Excel If Function
Freeze Panes and Conditional Format in Excel
Header and Footer, Page Break and Set Background in Excel
How to add Graphics and Perform Cross-Referencing in Excel
How to Create and Copy Worksheet in Excel
How to Create Worksheets in Excel
How to enter values and move around in Excel
How to Insert Comments and Add Text Box in Excel
How to Open, Close, Delete and Hide Worksheet in Excel
How to Perform Copy & Paste, Find & Replace in Excel
Using Styles, Themes and Templates in Excel
Using Functions and Built-in Functions in Excel
Translate Worksheet and workbook Security in Excel
Simple Charts and Pivot Charts in Excel
Sheet Options, Adjust Margins and Page Orientation in Excel
Printing Worksheets and Email workbooks in Excel
Perform Spell Check, Zoom In-Out and Use Special Symbols in Excel
How to use COUNT, COUNTIF, and COUNTIFS Function and Advanced If in Excel
How to Undo Changes, Setting Cell and Fonts, Text Decoration in Excel
How to Select, Insert, Delete and Move Data in Excel
How to Rotate Cells, Setting Colors and Text Alignment in Excel
eBooks
Interview Questions
Videos
How to Merge & Wrap Cells, Borders and Shades and Apply Formatting in Excel
BackStage View and Explore Window in Excel
Creating Formulas, Copying Formulas in Excel
Data Sorting and Using Ranges in Excel
Data Tables and Pivot Tables in Excel
Excel Fill Handle and Excel If Function
Freeze Panes and Conditional Format in Excel
Header and Footer, Page Break and Set Background in Excel
How to add Graphics and Perform Cross-Referencing in Excel
How to Create and Copy Worksheet in Excel
How to Create Worksheets in Excel
How to enter values and move around in Excel
How to Insert Comments and Add Text Box in Excel
How to Open, Close, Delete and Hide Worksheet in Excel
How to Perform Copy & Paste, Find & Replace in Excel
Using Styles, Themes and Templates in Excel
Using Functions and Built-in Functions in Excel
Translate Worksheet and workbook Security in Excel
Simple Charts and Pivot Charts in Excel
Sheet Options, Adjust Margins and Page Orientation in Excel
Printing Worksheets and Email workbooks in Excel
Perform Spell Check, Zoom In-Out and Use Special Symbols in Excel
How to use COUNT, COUNTIF, and COUNTIFS Function and Advanced If in Excel
How to Undo Changes, Setting Cell and Fonts, Text Decoration in Excel
How to Select, Insert, Delete and Move Data in Excel
How to Rotate Cells, Setting Colors and Text Alignment in Excel
OBIEEE
Articles
Concept of Testing Repository in OBIEE
Understanding Schemas in OBIEE
Overview of Oracle Business Intelligence Edition (OBIEE)
Multiple Logical Table Sources, Calculation Measures and Dimension Hierarchies
Level-Based Measures and Aggregates in OBIEE
Deep Dive into Repositories in OBIEE
eBooks
Interview Questions
Videos
Concept of Testing Repository in OBIEE
Understanding Schemas in OBIEE
Overview of Oracle Business Intelligence Edition (OBIEE)
Multiple Logical Table Sources, Calculation Measures and Dimension Hierarchies
Level-Based Measures and Aggregates in OBIEE
Deep Dive into Repositories in OBIEE
Pentaho
Articles
User interfaces available in Pentaho and their navigation
Overview of Pentaho and How to install Pentaho on your system
How to use the Pentaho Reporting Designer
How to use Grouping in Pentaho
How to use Functions in Reports in Pentaho
How to create Chart Report in Pentaho
eBooks
Interview Questions
Videos
User interfaces available in Pentaho and their navigation
Overview of Pentaho and How to install Pentaho on your system
How to use the Pentaho Reporting Designer
How to use Grouping in Pentaho
How to use Functions in Reports in Pentaho
How to create Chart Report in Pentaho
Power BI
Articles
Visualization Options in Power BI
Power BI Data Sources and How to connect with them
Power BI - Supported Data Sources
Power BI - Comparison with Other BI Tools
Overview of Power BI Embedded, Power BI Gateway and Power BI Report Server
Overview of Business Intelligence (BI) and Power BI
How to use various DAX functions in Power BI
How to Share Power BI Dashboard
How to Integrate Excel in Power BI
How to Download and Install Power BI Desktop
eBooks
Interview Questions
Videos
Visualization Options in Power BI
Power BI Data Sources and How to connect with them
Power BI - Supported Data Sources
Power BI - Comparison with Other BI Tools
Overview of Power BI Embedded, Power BI Gateway and Power BI Report Server
Overview of Business Intelligence (BI) and Power BI
How to use various DAX functions in Power BI
How to Share Power BI Dashboard
How to Integrate Excel in Power BI
How to Download and Install Power BI Desktop
Qlik View
Articles
List Box and Multi Box in QlikView
Navigation Options in QlikView
Overview of Data files (QVD) in QlikView
Processing Web Files in QlikView
Resident Load, Preceding Load and Incremental Load in QlikView
How to create Cross Tables in QlikView
How to create Pie Chart in QlikView
Straight Tables and Pivot Tables in QlikView
Database Connection in QlikView
Dimensions and Measures in QlikView
Usage of Keep Command in QlikView
Using Peek and RangeSum Function in QlikView
Handling Delimited Files in QlikView
Handling Excel Files in QlikView
How to create Bar Chart in QlikView
Using Match and Rank Function in QlikView
Overview of QlikView and How to install QlikView on your machine
Inline Data and Scripting in QlikView
Data Transformation in QlikView
Creating Dashboard in QlikView
Concept of Star Schema and Synthetic Key in QlikView
Concatenation and Master Calendar in QlikView
Column Manipulation in QlikView
eBooks
Interview Questions
Videos
List Box and Multi Box in QlikView
Navigation Options in QlikView
Overview of Data files (QVD) in QlikView
Processing Web Files in QlikView
Resident Load, Preceding Load and Incremental Load in QlikView
How to create Cross Tables in QlikView
How to create Pie Chart in QlikView
Straight Tables and Pivot Tables in QlikView
Database Connection in QlikView
Dimensions and Measures in QlikView
Usage of Keep Command in QlikView
Using Peek and RangeSum Function in QlikView
Handling Delimited Files in QlikView
Handling Excel Files in QlikView
How to create Bar Chart in QlikView
Using Match and Rank Function in QlikView
Overview of QlikView and How to install QlikView on your machine
Inline Data and Scripting in QlikView
Data Transformation in QlikView
Creating Dashboard in QlikView
Concept of Star Schema and Synthetic Key in QlikView
Concatenation and Master Calendar in QlikView
Column Manipulation in QlikView
Circular Reference in QlikView
QLikSense
Articles
Navigating in Qlik Sense Selections
Qlik Sense Conditional Functions
Qlik Sense Counter and Exponential and Logarithmic Functions
Qlik Sense Developer: Roles and Responsibilities
Overview of Gauge Chart in Qlik Sense
Qlik Sense Advantages and Limitations
Qlik Sense Architecture Components
Qlik Sense Capabilities for people, Groups and Organizations
What is Qlik Sense Pivot Table?
Ways of Qlik Sense Collaboration
Qlik Sense Formatting Functions
Qlik Sense distribution and Trigonometric and HyperBolic Functions
Qlik Sense Mapping and Logical Functions
Qlik Sense Financial Functions
Types of Qlik Sense Aggregation Functions
Tableau vs Qlik Sense vs Power BI
Significance of Text and Image in Qlik Sense
Set Analysis and Set Expressions in Qlik Sense
QlikView Vs Qlik Sense: Overview
Using a Scatter Plot in Qlik Sense
Types of Operators in Qlik Sense
Qlik Sense System Requirements
Treemap Visualization in Qlik Sense
Qlik Sense Interpretation Functions
Modulo Functions in Qlik Sense
Key Performance Indicators (KPI) in Qlik Sense
Introduction to Qlik Sense Mashup
How to Manage Content and Resources in Qlik Management Console
How to Interact With Qlik Sense Visualizations?
How to Interact with Qlik Sense interface
How to create Qlik Sense Application
General Numeric Functions in Qlik Sense
Concept of Social Engineering Attacks and Cross-Site Scripting
Components of Qlik Sense Desktop
eBooks
Interview Questions
Videos
Navigating in Qlik Sense Selections
Qlik Sense Conditional Functions
Qlik Sense Counter and Exponential and Logarithmic Functions
Qlik Sense Developer: Roles and Responsibilities
Overview of Gauge Chart in Qlik Sense
Qlik Sense Advantages and Limitations
Qlik Sense Architecture Components
Qlik Sense Capabilities for people, Groups and Organizations
What is Qlik Sense Pivot Table?
Ways of Qlik Sense Collaboration
Qlik Sense Formatting Functions
Qlik Sense distribution and Trigonometric and HyperBolic Functions
Qlik Sense Mapping and Logical Functions
Qlik Sense Financial Functions
Types of Qlik Sense Aggregation Functions
Tableau vs Qlik Sense vs Power BI
Significance of Text and Image in Qlik Sense
Set Analysis and Set Expressions in Qlik Sense
QlikView Vs Qlik Sense: Overview
Using a Scatter Plot in Qlik Sense
Types of Operators in Qlik Sense
Qlik Sense System Requirements
Treemap Visualization in Qlik Sense
Qlik Sense Interpretation Functions
Modulo Functions in Qlik Sense
Key Performance Indicators (KPI) in Qlik Sense
Introduction to Qlik Sense Mashup
How to Manage Content and Resources in Qlik Management Console
How to Interact With Qlik Sense Visualizations?
How to Interact with Qlik Sense interface
How to create Qlik Sense Application
General Numeric Functions in Qlik Sense
Concept of Social Engineering Attacks and Cross-Site Scripting
Components of Qlik Sense Desktop
SSAS
eBooks
Interview Questions
Videos
SSIS
eBooks
Interview Questions
Videos
SSRS
eBooks
Interview Questions
Videos
Tableau Desktop
Articles
How to create Pareto Chart in Tableau
How to create Gantt Chart in Tableau
How to create Dual Axis Chart, Box Plot and Heat Map in Tableau
How to create Crosstab and Motion Chart in Tableau
How to create Bump and Bubble Chart in Tableau
How to create Bar, Line and Pie Chart in Tableau
How to Build Hierarchy and Groups in Tableau
Filter Operations and Extract Filters in Tableau
Different Tools of Tableau and Tableau Architecture
Understanding Tableau Navigation and Data Terminology
Understanding Tableau Desktop Workspace
Data Window, Data Types, Data Aggregation and File Types in Tableau
Top 10 Data Visualization Tools
Tableau Quick and Context Filters
Perform Table Calculations in Tableau
Perform Data Sorting in Tableau
Perform Calculation and Operators and Functions in Tableau
Overview of Tableau and Data Visualization
How to perform Numeric, String and Date Calculations in Tableau
How to Join Data in Tableau using multiple sources
Condition Filters, Data Source and Top Filters in Tableau
Comparison of Tableau and Power BI
How to install Tableau on your system
How to create Waterfall, Bullet and Area Chart in Tableau
eBooks
Interview Questions
Videos
How to create Pareto Chart in Tableau
How to create Gantt Chart in Tableau
How to create Dual Axis Chart, Box Plot and Heat Map in Tableau
How to create Crosstab and Motion Chart in Tableau
How to create Bump and Bubble Chart in Tableau
How to create Bar, Line and Pie Chart in Tableau
How to Build Hierarchy and Groups in Tableau
Filter Operations and Extract Filters in Tableau
Different Tools of Tableau and Tableau Architecture
Understanding Tableau Navigation and Data Terminology
Understanding Tableau Desktop Workspace
Data Window, Data Types, Data Aggregation and File Types in Tableau
Top 10 Data Visualization Tools
Tableau Quick and Context Filters
Perform Table Calculations in Tableau
Perform Data Sorting in Tableau
Perform Calculation and Operators and Functions in Tableau
Overview of Tableau and Data Visualization
How to perform Numeric, String and Date Calculations in Tableau
How to Join Data in Tableau using multiple sources
Condition Filters, Data Source and Top Filters in Tableau
Comparison of Tableau and Power BI
How to install Tableau on your system
How to create Waterfall, Bullet and Area Chart in Tableau
Tableau Server
TIBCO BW
eBooks
Videos
How to Merge & Wrap Cells, Borders and Shades and Apply Formatting in Excel
How to Clone Repository in Git
Navigating in Qlik Sense Selections
Qlik Sense Conditional Functions
Qlik Sense Counter and Exponential and Logarithmic Functions
Qlik Sense Developer: Roles and Responsibilities
Overview of Gauge Chart in Qlik Sense
Qlik Sense Advantages and Limitations
Qlik Sense Architecture Components
Qlik Sense Capabilities for people, Groups and Organizations
What is Qlik Sense Pivot Table?
Ways of Qlik Sense Collaboration
Qlik Sense Formatting Functions
Qlik Sense distribution and Trigonometric and HyperBolic Functions
Qlik Sense Mapping and Logical Functions
Qlik Sense Financial Functions
Types of Qlik Sense Aggregation Functions
Tableau vs Qlik Sense vs Power BI
Significance of Text and Image in Qlik Sense
Set Analysis and Set Expressions in Qlik Sense
QlikView Vs Qlik Sense: Overview
Using a Scatter Plot in Qlik Sense
Types of Operators in Qlik Sense
Qlik Sense System Requirements
List Box and Multi Box in QlikView
Navigation Options in QlikView
Overview of Data files (QVD) in QlikView
Processing Web Files in QlikView
Resident Load, Preceding Load and Incremental Load in QlikView
How to create Cross Tables in QlikView
How to create Pie Chart in QlikView
Straight Tables and Pivot Tables in QlikView
Database Connection in QlikView
Dimensions and Measures in QlikView
Usage of Keep Command in QlikView
Using Peek and RangeSum Function in QlikView
Handling Delimited Files in QlikView
Handling Excel Files in QlikView
How to create Bar Chart in QlikView
How to create Pareto Chart in Tableau
BackStage View and Explore Window in Excel
Creating Formulas, Copying Formulas in Excel
Treemap Visualization in Qlik Sense
Overview of SSRS and its Architecture
Overview of SSIS and why SSIS is required
Introduction to TIBCO Business Works (TIBCO BW)
Overview of Tableau Server and How to install it
Overview of SSAS and its Architecture
Using Match and Rank Function in QlikView
Data Sorting and Using Ranges in Excel
Data Tables and Pivot Tables in Excel
Excel Fill Handle and Excel If Function
Freeze Panes and Conditional Format in Excel
Overview of QlikView and How to install QlikView on your machine
Header and Footer, Page Break and Set Background in Excel
Inline Data and Scripting in QlikView
How to add Graphics and Perform Cross-Referencing in Excel
How to Create and Copy Worksheet in Excel
How to Create Worksheets in Excel
How to enter values and move around in Excel
How to Insert Comments and Add Text Box in Excel
How to Open, Close, Delete and Hide Worksheet in Excel
How to Perform Copy & Paste, Find & Replace in Excel
Qlik Sense Interpretation Functions
Setting Up Distributed Servers in Tableau Server
Concept of Testing Repository in OBIEE
Understanding Schemas in OBIEE
Modulo Functions in Qlik Sense
Key Performance Indicators (KPI) in Qlik Sense
Introduction to Qlik Sense Mashup
Data Transformation in QlikView
Creating Dashboard in QlikView
Concept of Star Schema and Synthetic Key in QlikView
How to Manage Content and Resources in Qlik Management Console
Concatenation and Master Calendar in QlikView
Column Manipulation in QlikView
How to Interact With Qlik Sense Visualizations?
How to Interact with Qlik Sense interface
Circular Reference in QlikView
Aggregate Functions in QlikView
Perform Report Operations in IBM Cognos
Overview of Oracle Business Intelligence Edition (OBIEE)
Introduction to IBM Cognos and its Components and Services
How to open, create, save, run and print report in Cognos
Multiple Logical Table Sources, Calculation Measures and Dimension Hierarchies
How to open, create and save Analysis in Analysis Studio in Cognos
Level-Based Measures and Aggregates in OBIEE
How to create report in Report Studio
Deep Dive into Repositories in OBIEE
Concept of Data Warehouse and Dimension Modelling
How to create List and CrossTab Report in Cognos
How to create a package using Cognos
How to create Qlik Sense Application
General Numeric Functions in Qlik Sense
Concept of Social Engineering Attacks and Cross-Site Scripting
Business and Presentation Layer of OBIEE explained
Filters and Custom Calculations in Cognos
Components of Qlik Sense Desktop
BI Tools for giant Data Visualization
Data Warehouse Schemas, ETL and Reporting Tools
Aggregation Functions in Qlik Sense
How to create Gantt Chart in Tableau
How to create Dual Axis Chart, Box Plot and Heat Map in Tableau
How to create Crosstab and Motion Chart in Tableau
How to create Bump and Bubble Chart in Tableau
How to create Bar, Line and Pie Chart in Tableau
Introduction to Cognos TM1 Perspective
How to Setup TM1 Application Server
How to Build Hierarchy and Groups in Tableau
How to Configure Security in TM1
Concept of Dimensions in Cognos TM1
Filter Operations and Extract Filters in Tableau
Cognos TM1 Installation and Configuration
Different Tools of Tableau and Tableau Architecture
Understanding Tableau Navigation and Data Terminology
Understanding Tableau Desktop Workspace
Data Window, Data Types, Data Aggregation and File Types in Tableau
Top 10 Data Visualization Tools
Tableau Quick and Context Filters
Perform Table Calculations in Tableau
Perform Data Sorting in Tableau
Perform Calculation and Operators and Functions in Tableau
Overview of Tableau and Data Visualization
How to perform Numeric, String and Date Calculations in Tableau
How to Join Data in Tableau using multiple sources
Condition Filters, Data Source and Top Filters in Tableau
Comparison of Tableau and Power BI
How to install Tableau on your system
How to create Waterfall, Bullet and Area Chart in Tableau
How to create Tree Maps and Heat Maps in Tableau
How to create Scatter Plot and Histogram Chart in Tableau
User interfaces available in Pentaho and their navigation
Overview of Pentaho and How to install Pentaho on your system
How to use the Pentaho Reporting Designer
How to use Grouping in Pentaho
How to use Functions in Reports in Pentaho
How to create Chart Report in Pentaho
How to add Page Footer Fields in Pentaho
Formatting Report Elements in Pentaho Reporting Designer
Using Styles, Themes and Templates in Excel
Using Functions and Built-in Functions in Excel
Translate Worksheet and workbook Security in Excel
Simple Charts and Pivot Charts in Excel
Sheet Options, Adjust Margins and Page Orientation in Excel
Printing Worksheets and Email workbooks in Excel
Perform Spell Check, Zoom In-Out and Use Special Symbols in Excel
How to use COUNT, COUNTIF, and COUNTIFS Function and Advanced If in Excel
How to Undo Changes, Setting Cell and Fonts, Text Decoration in Excel
How to Select, Insert, Delete and Move Data in Excel
How to Rotate Cells, Setting Colors and Text Alignment in Excel
How to Perform Data Validation and Data Filtering in Excel
Cognos Studios and other capabilities
Cognos - Relationships in Metadata Model
Visualization Options in Power BI
Power BI Data Sources and How to connect with them
Power BI - Supported Data Sources
Power BI - Comparison with Other BI Tools
Overview of Power BI Embedded, Power BI Gateway and Power BI Report Server
Overview of Business Intelligence (BI) and Power BI
How to use various DAX functions in Power BI
How to Share Power BI Dashboard
How to Integrate Excel in Power BI
How to Download and Install Power BI Desktop
How to create Power BI Dashboard and Reports
Top Qlik Sense Interview Questions and Answers
Top Microsoft BI Interview Questions and Answers
Top TIBCO Spotfire Interview Questions and Answers
Top OBIEE Interview Questions and Answers
Top Tableau Desktop Interview Questions and Answers
Top Tableau Server Interview Questions and Answers
Top Qlik View Interview Questions and Answers
Top TIBCO Business Works Interview Questions and Answers
Top Oracle Hyperion Interview Questions and Answers
Top Power BI Interview Questions and Answers
Top Pentaho Interview Questions and Answers
Top Cognos TM1 Interview Questions and Answers
Top IBM DataStage Interview Questions and Answers
Top IBM Cognos Analytics Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
Big Data
eBooks
Videos
Aapche Cassandra
Articles
Deep dive into Cassandra Query Language Collections and user defined data types.
Deep dive into Cassandra Shell Commands
How to Create and Alter Tables in Apache Cassandra
How to Create and Drop Indexes in Apache Cassandra
How to create, alter and drop Keyspaces in Cassandra
How to Drop and Truncate Tables in Apache Cassandra
How to set up Both cqlsh and Java environments to work with Cassandra
How to Perform CRUD ( Create , Read , Update and Delete ) Operations in Table in Apache Cassandra
Introduction to Apache Cassandra, History and Architecture
Overview of How Cassandra Stores its data
Overview of important class in Cassandra and introduction of Cassandra query shell language
eBooks
Interview Questions
Videos
Deep dive into Cassandra Query Language Collections and user defined data types.
Deep dive into Cassandra Shell Commands
How to Create and Alter Tables in Apache Cassandra
How to Create and Drop Indexes in Apache Cassandra
How to create, alter and drop Keyspaces in Cassandra
How to Drop and Truncate Tables in Apache Cassandra
How to set up Both cqlsh and Java environments to work with Cassandra
How to Perform CRUD ( Create , Read , Update and Delete ) Operations in Table in Apache Cassandra
Introduction to Apache Cassandra, History and Architecture
Overview of How Cassandra Stores its data
Overview of important class in Cassandra and introduction of Cassandra query shell language
Apache NiFi
Articles
How to Monitor System statistics using Apache NiFi
Concept of Logging in Apache NiFi
Basic Concepts of Apache NiFi and its Installation
Deep Dive into Apache Nifi – Flow Files, Queues, Process Groups and Labels
Deep dive into Apache NiFi-Processors
Detailed understanding of Apache NiFi -Templates
How to Administer Apache NiFi and Create Flows in Apache NiFi
Understanding Apache NiFi API’s with request and response example
Understanding Apache Nifi Processors Categorization and its relationship
Introduction to Apache NiFi, its History, Features and Architecture
eBooks
Interview Questions
Videos
How to Monitor System statistics using Apache NiFi
Concept of Logging in Apache NiFi
Basic Concepts of Apache NiFi and its Installation
Deep Dive into Apache Nifi – Flow Files, Queues, Process Groups and Labels
Deep dive into Apache NiFi-Processors
Detailed understanding of Apache NiFi -Templates
How to Administer Apache NiFi and Create Flows in Apache NiFi
Understanding Apache NiFi API’s with request and response example
Understanding Apache Nifi Processors Categorization and its relationship
Introduction to Apache NiFi, its History, Features and Architecture
Apache Oozie
eBooks
Interview Questions
Videos
Apache Pig
Articles
Explanation of Apache Pig Group and Cogroup Operators
Detailed Study of Architecture of Apache Pig
Deep Dive into Pig Latin Diagnostic Operators
Deep Dive into Apache Pig Functions: Load & Store, Bag & Tuple, String, Date-time, Math
Apache Pig Basics, Features and Comparison with MapReduce, Hive & SQL and History of Apache Pig
Explanation of Shell and Utility Commands provided by Apache Grunt Shell
How to Install Apache Pig and Configure Pig
How to Load data to Apache Pig from Hadoop File System
How to run Apache Pig Scripts in Batch Mode
How to Store data in Apache Pig using Store Operator
How to use Cross Operator and Union Operator in Pig Latin
How to use Split and Filter Operator in Apache Pig Latin
How to use the Join Operators in Pig Latin
How to use Distinct, For Each, Order By, Limit Operators and Eval Functions in Apache Pig
eBooks
Interview Questions
Videos
Explanation of Apache Pig Group and Cogroup Operators
Detailed Study of Architecture of Apache Pig
Deep Dive into Pig Latin Diagnostic Operators
Deep Dive into Apache Pig Functions: Load & Store, Bag & Tuple, String, Date-time, Math
Apache Pig Basics, Features and Comparison with MapReduce, Hive & SQL and History of Apache Pig
Explanation of Shell and Utility Commands provided by Apache Grunt Shell
How to Install Apache Pig and Configure Pig
How to Load data to Apache Pig from Hadoop File System
How to run Apache Pig Scripts in Batch Mode
How to Store data in Apache Pig using Store Operator
How to use Cross Operator and Union Operator in Pig Latin
How to use Split and Filter Operator in Apache Pig Latin
How to use the Join Operators in Pig Latin
How to use Distinct, For Each, Order By, Limit Operators and Eval Functions in Apache Pig
Apache Spark
Articles
Overview of Scala programming language and How to install Scala on your system
How to Install Apache Spark on your system
How to perform pattern matching in Scala and use of Regex expressions
How to use Functions in Scala programming Language
How to use Collections in Scala
How to use Arrays in Scala Programming Language
How to perform Exception Handling in Scala Language
How to Deploy Spark Application on Cluster
Extractor Object in Scala and how to perform pattern matching using extractors
Details of Data Types and Basic Literals in Scala
Detailed understanding of Operators in Scala Language
Deep dive into File Handling in Scala
Deep dive into Advanced programming in Spark
Basics of Scala Programming Language
Concept of String Manipulation in Scala
Conditional statements and Loop control structures in Scala
Concept of Resilient Distributed Datasets (RDD) in Apache Spark
How to use Classes and Objects in Scala programming
Overview of Apache Spark Framework
Spark Core and implementation of RDD transformations and actions in RDD programming
eBooks
Interview Questions
Videos
Overview of Scala programming language and How to install Scala on your system
How to Install Apache Spark on your system
How to perform pattern matching in Scala and use of Regex expressions
How to use Functions in Scala programming Language
How to use Collections in Scala
How to use Arrays in Scala Programming Language
How to perform Exception Handling in Scala Language
How to Deploy Spark Application on Cluster
Extractor Object in Scala and how to perform pattern matching using extractors
Details of Data Types and Basic Literals in Scala
Detailed understanding of Operators in Scala Language
Deep dive into File Handling in Scala
Deep dive into Advanced programming in Spark
Basics of Scala Programming Language
Concept of String Manipulation in Scala
Conditional statements and Loop control structures in Scala
Concept of Resilient Distributed Datasets (RDD) in Apache Spark
How to use Classes and Objects in Scala programming
Overview of Apache Spark Framework
Spark Core and implementation of RDD transformations and actions in RDD programming
Detailed understanding of Scala Access Modifiers
Apache Sqoop
Articles
How to use the Apache Sqoop Eval and Codegen tool
How to list out the databases and tables of a particular database using Sqoop
How to import data and tables from MySQL to Hadoop HDFS
How to export data back from Hadoop HDFS to RDBMS and Create and maintain the Sqoop jobs
Introduction, Installation and Configuration of Apache Sqoop
eBooks
Interview Questions
Videos
How to use the Apache Sqoop Eval and Codegen tool
How to list out the databases and tables of a particular database using Sqoop
How to import data and tables from MySQL to Hadoop HDFS
How to export data back from Hadoop HDFS to RDBMS and Create and maintain the Sqoop jobs
Introduction, Installation and Configuration of Apache Sqoop
Apache Storm
Articles
Application of Apache Storm Framework in Yahoo Finance
Concept of Cluster Architecture in Apache Storm
Introduction to Apache Storm and Core Concepts of Apache Storm
How to Install Apache Storm framework on your machine
How to implement Mobile Call log Analyzer using Apache Storm
How Apache Storm is used in Twitter
eBooks
Interview Questions
Videos
Application of Apache Storm Framework in Yahoo Finance
Concept of Cluster Architecture in Apache Storm
Introduction to Apache Storm and Core Concepts of Apache Storm
How to Install Apache Storm framework on your machine
How to implement Mobile Call log Analyzer using Apache Storm
How Apache Storm is used in Twitter
Detailed understanding of Workflow of Apache Storm
Hadoop and MapReduce
Articles
Concept of Combiners in Hadoop MapReduce
Concept of MapReduce in BigData
Detailed understanding of Hadoop Architecture and Hadoop Distributed File System (HDFS)
Concept of Partitioner in MapReduce and its implementation using example
Deep dive into Hadoop administration
Deep dive into the MapReduce API
Detailed understanding of Hadoop Distributed File System (HDFS)
Phases of MapReduce Data flow and detailed understanding of Mapreduce API
Overview of YARN and its components and benefits of YARN
Overview of Big Data and Hadoop, Big Data technologies
Implementation of Word Count program using Hadoop MapReduce
Operation of MapReduce in Hadoop framework using Java
Implementation of Character Count program using Hadoop MapReduce
How to set up Hadoop Multi-Node Cluster on a distributed environment
How to perform operations in Hadoop and commands used in Hadoop
How to install Hadoop on your system
eBooks
Videos
Concept of Combiners in Hadoop MapReduce
Concept of MapReduce in BigData
Detailed understanding of Hadoop Architecture and Hadoop Distributed File System (HDFS)
Concept of Partitioner in MapReduce and its implementation using example
Deep dive into Hadoop administration
Deep dive into the MapReduce API
Detailed understanding of Hadoop Distributed File System (HDFS)
Phases of MapReduce Data flow and detailed understanding of Mapreduce API
Overview of YARN and its components and benefits of YARN
Overview of Big Data and Hadoop, Big Data technologies
Implementation of Word Count program using Hadoop MapReduce
Operation of MapReduce in Hadoop framework using Java
Implementation of Character Count program using Hadoop MapReduce
How to set up Hadoop Multi-Node Cluster on a distributed environment
How to perform operations in Hadoop and commands used in Hadoop
How to install Hadoop on your system
How to install Hadoop Framework on your system
How the MapReduce Algorithm works using example
HBase
Articles
Deep dive into HBase architecture
Deep dive into Java Client API for HBase and its associated classes
How to create and List Table in HBase shell
How to create data in an HBase table
How to delete data in Table in HBase
How to enable and disable a Table using HBase shell
How to install HBase and configure on your system
How to make changes to an existing Table and describe it in HBase
How to read data from Table in HBase
How to start HBase interactive shell and how HBase general commands works
How to Stop HBase using Java API
How to update data in Table using HBase Shell
How to verify the existence of a Table and How to Drop a Table in HBase
Overview of HBase, its Advantages, Features and history
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
eBooks
Interview Questions
Videos
Deep dive into HBase architecture
Deep dive into Java Client API for HBase and its associated classes
How to create and List Table in HBase shell
How to create data in an HBase table
How to delete data in Table in HBase
How to enable and disable a Table using HBase shell
How to install HBase and configure on your system
How to make changes to an existing Table and describe it in HBase
How to read data from Table in HBase
How to start HBase interactive shell and how HBase general commands works
How to Stop HBase using Java API
How to update data in Table using HBase Shell
How to verify the existence of a Table and How to Drop a Table in HBase
Overview of HBase, its Advantages, Features and history
Top Apache HBase Interview Questions and Answers
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Hive and Impala
Articles
Concept of Partitioning of table in Hive
Detailed understanding of built-in functions available in Hive
Different Data Types in Hive which are involved in creation of table.
How to Alter the attributes of a table and delete a Table in Hive
How to create a table in Hive and how to insert data into it
How to create and drop a database in Hive
How to create and manage Views and Create and Drop an index in Hive
How to install Hive on your system
How to perform Join operations in Hive Query Language (HQL)
How to use the select statement in Hive Query Language
Introduction to Impala, its features, advantages and disadvantages
How to start Impala Shell and the various options of the shell
How to select a database using Command and select database using Hue Browser in Impala
How to perform changes on a given table and how to delete table in Impala
How to fetch the data from one or more tables in a database and fetch description in Impala
How to download, install and set up Impala in your system
How to create a table in the required database in Impala
How to Create, Alter and Drop a View in Impala
Explanation of Union Clause, With Clause and Distinct Operator in Impala
Explanation of Limit Clause and Offset Clause in Impala
Data Types in Impala Query Language
Detailed understanding of Architecture of Impala
Explanation of Order by Clause, Group by Clause and Having Clause in Impala
How to add new records into an existing table in a database using INSERT in Impala
How to create a database in Impala
eBooks
Videos
Concept of Partitioning of table in Hive
Detailed understanding of built-in functions available in Hive
Different Data Types in Hive which are involved in creation of table.
How to Alter the attributes of a table and delete a Table in Hive
How to create a table in Hive and how to insert data into it
How to create and drop a database in Hive
How to create and manage Views and Create and Drop an index in Hive
How to install Hive on your system
How to perform Join operations in Hive Query Language (HQL)
How to use the select statement in Hive Query Language
Introduction to Impala, its features, advantages and disadvantages
How to start Impala Shell and the various options of the shell
How to select a database using Command and select database using Hue Browser in Impala
How to perform changes on a given table and how to delete table in Impala
How to fetch the data from one or more tables in a database and fetch description in Impala
How to download, install and set up Impala in your system
How to create a table in the required database in Impala
How to Create, Alter and Drop a View in Impala
Explanation of Union Clause, With Clause and Distinct Operator in Impala
Explanation of Limit Clause and Offset Clause in Impala
Data Types in Impala Query Language
Detailed understanding of Architecture of Impala
Explanation of Order by Clause, Group by Clause and Having Clause in Impala
How to add new records into an existing table in a database using INSERT in Impala
How to create a database in Impala
How to drop a database in Impala
Top Apache Impala Interview Questions and Answers
Top Apache Hive Interview Questions and Answers
MongoDB
Articles
Advanced Indexing in MongoDB and Limitation of Indexing in MongoDB
Concept of Capped Collections and Auto-Increment Sequence in MongoDB
Concept of Map Reduce in MongoDB
Concept of Relationships and Database References in MongoDB
Concept of Sharding process and How to create a backup in MongoDB
Data Modelling in MongoDB and How to create and Drop database in MongoDB
Deep dive into Covered Queries in MongoDB and Analyzing queries
Deep dive into Replication process in MongoDB
How to Create and Drop a collection using MongoDB
How to Insert, Update, Delete and Query Document in MongoDB Collection
How to Install MongoDB on your system
How to limit records using MongoDB ad use projection in MongoDB
How to Set up MongoDB JDBC driver
How to sort records in MongoDB and concept of Indexing and Aggregation in MongoDB
How to use Regex Expressions and Text Search in MongoDB
MongoDB Administration using RockMongo and concept of GridFS in MongoDB
Overview of MongoDB, its history and purpose of building MongoDB
Understand NoSQL Databases and MongoDB advantages over Relational DBMS
eBooks
Interview Questions
Videos
Advanced Indexing in MongoDB and Limitation of Indexing in MongoDB
Concept of Capped Collections and Auto-Increment Sequence in MongoDB
Concept of Map Reduce in MongoDB
Concept of Relationships and Database References in MongoDB
Concept of Sharding process and How to create a backup in MongoDB
Data Modelling in MongoDB and How to create and Drop database in MongoDB
Deep dive into Covered Queries in MongoDB and Analyzing queries
Deep dive into Replication process in MongoDB
How to Create and Drop a collection using MongoDB
How to Insert, Update, Delete and Query Document in MongoDB Collection
How to Install MongoDB on your system
How to limit records using MongoDB ad use projection in MongoDB
How to Set up MongoDB JDBC driver
How to sort records in MongoDB and concept of Indexing and Aggregation in MongoDB
How to use Regex Expressions and Text Search in MongoDB
MongoDB Administration using RockMongo and concept of GridFS in MongoDB
Overview of MongoDB, its history and purpose of building MongoDB
Understand NoSQL Databases and MongoDB advantages over Relational DBMS
Splunk
Articles
Deep dive into Splunk Search processing Language (SPL)
How to perform Basic Search in Splunk
How to perform searching using fields in Splunk
How to perform Time Range search in Splunk
How to share and export the search result in Splunk
A Deep Dive into Splunk Web Interface
eBooks
Videos
Deep dive into Splunk Search processing Language (SPL)
How to perform Basic Search in Splunk
How to perform searching using fields in Splunk
How to perform Time Range search in Splunk
How to share and export the search result in Splunk
Top Splunk SIEM Interview Questions and Answers
Top Splunk Interview Questions and Answers
A Deep Dive into Splunk Web Interface
Application of Apache Storm Framework in Yahoo Finance
Deep dive into Cassandra Query Language Collections and user defined data types.
Deep dive into Cassandra Shell Commands
How to Create and Alter Tables in Apache Cassandra
How to Create and Drop Indexes in Apache Cassandra
How to create, alter and drop Keyspaces in Cassandra
How to Drop and Truncate Tables in Apache Cassandra
Concept of Combiners in Hadoop MapReduce
Concept of MapReduce in BigData
Detailed understanding of Hadoop Architecture and Hadoop Distributed File System (HDFS)
Concept of Partitioner in MapReduce and its implementation using example
Deep dive into Hadoop administration
Deep dive into the MapReduce API
Detailed understanding of Hadoop Distributed File System (HDFS)
How to set up Both cqlsh and Java environments to work with Cassandra
How to Perform CRUD ( Create , Read , Update and Delete ) Operations in Table in Apache Cassandra
Explanation of Apache Pig Group and Cogroup Operators
Detailed Study of Architecture of Apache Pig
Deep Dive into Pig Latin Diagnostic Operators
Deep Dive into Apache Pig Functions: Load & Store, Bag & Tuple, String, Date-time, Math
Apache Pig Basics, Features and Comparison with MapReduce, Hive & SQL and History of Apache Pig
Phases of MapReduce Data flow and detailed understanding of Mapreduce API
Overview of YARN and its components and benefits of YARN
Overview of Big Data and Hadoop, Big Data technologies
Implementation of Word Count program using Hadoop MapReduce
Operation of MapReduce in Hadoop framework using Java
Implementation of Character Count program using Hadoop MapReduce
How to set up Hadoop Multi-Node Cluster on a distributed environment
How to perform operations in Hadoop and commands used in Hadoop
How to install Hadoop on your system
How to install Hadoop Framework on your system
How the MapReduce Algorithm works using example
Introduction to Apache Cassandra, History and Architecture
Overview of How Cassandra Stores its data
Overview of important class in Cassandra and introduction of Cassandra query shell language
Concept of Partitioning of table in Hive
Detailed understanding of built-in functions available in Hive
Different Data Types in Hive which are involved in creation of table.
How to Alter the attributes of a table and delete a Table in Hive
How to create a table in Hive and how to insert data into it
How to create and drop a database in Hive
How to create and manage Views and Create and Drop an index in Hive
How to install Hive on your system
How to perform Join operations in Hive Query Language (HQL)
How to use the select statement in Hive Query Language
Deep dive into Splunk Search processing Language (SPL)
Explanation of Shell and Utility Commands provided by Apache Grunt Shell
How to perform Basic Search in Splunk
How to Install Apache Pig and Configure Pig
How to perform searching using fields in Splunk
How to Load data to Apache Pig from Hadoop File System
How to perform Time Range search in Splunk
How to run Apache Pig Scripts in Batch Mode
How to Store data in Apache Pig using Store Operator
How to share and export the search result in Splunk
How to use Cross Operator and Union Operator in Pig Latin
How to use Split and Filter Operator in Apache Pig Latin
How to use the Join Operators in Pig Latin
How to use Distinct, For Each, Order By, Limit Operators and Eval Functions in Apache Pig
User Defined Functions in Apache Pig Latin
Advanced Indexing in MongoDB and Limitation of Indexing in MongoDB
Concept of Capped Collections and Auto-Increment Sequence in MongoDB
Concept of Map Reduce in MongoDB
Concept of Relationships and Database References in MongoDB
Concept of Sharding process and How to create a backup in MongoDB
Data Modelling in MongoDB and How to create and Drop database in MongoDB
Deep dive into Covered Queries in MongoDB and Analyzing queries
Deep dive into Replication process in MongoDB
How to Create and Drop a collection using MongoDB
How to Insert, Update, Delete and Query Document in MongoDB Collection
How to Install MongoDB on your system
How to limit records using MongoDB ad use projection in MongoDB
Deep dive into HBase architecture
Deep dive into Java Client API for HBase and its associated classes
How to create and List Table in HBase shell
How to create data in an HBase table
How to delete data in Table in HBase
How to enable and disable a Table using HBase shell
How to Set up MongoDB JDBC driver
How to install HBase and configure on your system
How to make changes to an existing Table and describe it in HBase
How to read data from Table in HBase
How to sort records in MongoDB and concept of Indexing and Aggregation in MongoDB
How to start HBase interactive shell and how HBase general commands works
How to Stop HBase using Java API
How to update data in Table using HBase Shell
How to verify the existence of a Table and How to Drop a Table in HBase
Overview of HBase, its Advantages, Features and history
How to use Regex Expressions and Text Search in MongoDB
Overview of Scala programming language and How to install Scala on your system
MongoDB Administration using RockMongo and concept of GridFS in MongoDB
Overview of MongoDB, its history and purpose of building MongoDB
Understand NoSQL Databases and MongoDB advantages over Relational DBMS
How to Monitor System statistics using Apache NiFi
Concept of Logging in Apache NiFi
How to use the Apache Sqoop Eval and Codegen tool
How to list out the databases and tables of a particular database using Sqoop
How to import data and tables from MySQL to Hadoop HDFS
How to export data back from Hadoop HDFS to RDBMS and Create and maintain the Sqoop jobs
How to Install Apache Spark on your system
How to perform pattern matching in Scala and use of Regex expressions
How to use Functions in Scala programming Language
How to use Collections in Scala
How to use Arrays in Scala Programming Language
How to perform Exception Handling in Scala Language
How to Deploy Spark Application on Cluster
Extractor Object in Scala and how to perform pattern matching using extractors
Details of Data Types and Basic Literals in Scala
Detailed understanding of Operators in Scala Language
Deep dive into File Handling in Scala
Deep dive into Advanced programming in Spark
Basics of Scala Programming Language
Concept of String Manipulation in Scala
Conditional statements and Loop control structures in Scala
Introduction to Impala, its features, advantages and disadvantages
How to start Impala Shell and the various options of the shell
How to select a database using Command and select database using Hue Browser in Impala
How to perform changes on a given table and how to delete table in Impala
How to fetch the data from one or more tables in a database and fetch description in Impala
How to download, install and set up Impala in your system
How to create a table in the required database in Impala
How to Create, Alter and Drop a View in Impala
Explanation of Union Clause, With Clause and Distinct Operator in Impala
Explanation of Limit Clause and Offset Clause in Impala
Data Types in Impala Query Language
Concept of Resilient Distributed Datasets (RDD) in Apache Spark
How to use Classes and Objects in Scala programming
Overview of Apache Spark Framework
Spark Core and implementation of RDD transformations and actions in RDD programming
Introduction, Installation and Configuration of Apache Sqoop
Basic Concepts of Apache NiFi and its Installation
Deep Dive into Apache Nifi – Flow Files, Queues, Process Groups and Labels
Deep dive into Apache NiFi-Processors
Detailed understanding of Apache NiFi -Templates
How to Administer Apache NiFi and Create Flows in Apache NiFi
Understanding Apache NiFi API’s with request and response example
Understanding Apache Nifi Processors Categorization and its relationship
Detailed understanding of Architecture of Impala
Explanation of Order by Clause, Group by Clause and Having Clause in Impala
How to add new records into an existing table in a database using INSERT in Impala
How to create a database in Impala
How to drop a database in Impala
Detailed understanding of Scala Access Modifiers
How to use Variables in Scala with the help of example
Introduction to Apache NiFi, its History, Features and Architecture
Concept of Cluster Architecture in Apache Storm
Introduction to Apache Storm and Core Concepts of Apache Storm
How to Install Apache Storm framework on your machine
How to implement Mobile Call log Analyzer using Apache Storm
How Apache Storm is used in Twitter
Detailed understanding of Workflow of Apache Storm
Deep Dive into Trident – an extension of Apache Storm
Top Splunk SIEM Interview Questions and Answers
Top Big Data Hadoop Interview Questions and Answers
Top MongoDB Interview Questions and Answers
Top Scala Interview Questions and Answers
Top Splunk Interview Questions and Answers
Top Hadoop Administration Interview Questions and Answers
Top Apache Sqoop Interview Questions and Answers
Top Apache NiFi Interview Questions and Answers
Top Apache Impala Interview Questions and Answers
Top Apache HBase Interview Questions and Answers
Top Apache Flume Interview Questions and Answers
Top Apache Spark Interview Questions and Answers
Top Apache Pig Interview Questions and Answers
Top Apache Cassandra Interview Questions and Answers
Top Apache Hive Interview Questions and Answers
Top Apache Oozie Interview Questions and Answers
Top Apache Storm Interview Questions and Answers
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Deep Dive into Apache NiFi User Interface
Deep dive into built-in operators of Hive
Concept of Atomic Operations in MongoDB
A Deep Dive into Splunk Web Interface
Deep Dive into Apache Oozie Workflow
How to Configure Oozie Workflow using Property File
Concept of Coordinators applications using Apache Oozie
Basics of Apache Oozie and Oozie Editors
Deep Dive into Oozie Bundle System and CLI & Extensions
Process of Data Ingestion in Splunk Environment
Deep Dive into Apache NiFi User Interface
Deep dive into HBase Scan, Count and Truncate command and how to achieve security in HBase
Basics of Splunk and Installation of Splunk Environment
Top Apache Oozie Interview Questions and Answers You must Prepare Gaurav
Blockchain
Articles
Introduction to Ethereum and Smart Contracts
Ethereum - Interacting with Deployed Contract
Ethereum – Attaching Wallet to Ganache Blockchain
Ethereum - Creating Contract Users
Concept of Blockchain Double Spending and Bitcoin Cash
Bitcoin Forks and SegWit and BlockChain Merkel Tree
Comparison between Blockchain and Database
Basic Components of Bitcoin and Blockchain Proof of Work
Ethereum - Solidity for Contract Writing
Ethereum - Ganache for Blockchain
eBooks
Interview Questions
Videos
BlockChain and Ethereum
Articles
eBooks
Interview Questions
Videos
Introduction to Ethereum and Smart Contracts
Ethereum - Interacting with Deployed Contract
Ethereum – Attaching Wallet to Ganache Blockchain
Ethereum - Creating Contract Users
Concept of Blockchain Double Spending and Bitcoin Cash
Bitcoin Forks and SegWit and BlockChain Merkel Tree
Comparison between Blockchain and Database
Basic Components of Bitcoin and Blockchain Proof of Work
Ethereum - Solidity for Contract Writing
Ethereum - Ganache for Blockchain
Overview and History of Blockchain
Overview of Bitcoin and Key Concepts of Bitcoin
Top BlockChain Interview Questions and Answers
Top Ethereum Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
Cloud Computing
Articles
eBooks
Interview Questions
Videos
AWS
Articles
How to Use Amazon Machine Learning
How to use Amazon KCL and set up Amazon EMR
How to Set Up Amazon RDS (Relational Database Service)
How to Configure AWS Direct Connect
How to Configure Amazon Simple Storage Service (S3)
How to Configure Amazon Route 53
How AWS CloudFront Delivers the Content
Amazon Elastic Block Storage (EBS) and Storage Gateway
How to use Simple Workflow Service (SWF) and Amazon WorkMail
Understanding of AWS Management Console
How to Set Up AWS Data Pipeline
eBooks
Videos
How to Use Amazon Machine Learning
How to use Amazon KCL and set up Amazon EMR
How to Set Up Amazon RDS (Relational Database Service)
How to Configure AWS Direct Connect
How to Configure Amazon Simple Storage Service (S3)
How to Configure Amazon Route 53
How AWS CloudFront Delivers the Content
Amazon Elastic Block Storage (EBS) and Storage Gateway
How to use Simple Workflow Service (SWF) and Amazon WorkMail
Understanding of AWS Management Console
How to Set Up AWS Data Pipeline
How to Create Amazon Workspaces
Top Azure Developer Interview Questions and Answers
Top Amazon Web Services (AWS) Interview Questions and Answers
Azure
Articles
How to configure Azure Cloud Service
How to configure Azure Load Balancer
How to Configure Azure Storage Security
How to create Azure Mobile App
Overview of Microsoft Azure and Cloud Computing
Creating App Service Plan in Azure Portal
Azure Virtual Machines and Compute Service
Azure Virtual Machine Scale Set and Auto Scaling
Azure Table, Queue and Disk Storage
Azure Storage Monitoring and Resource Tool
Azure Storage Building Blocks and Storage Account
Azure Storage account and Blob service configuration
Azure SQL Managed Instance and SQL Stretch Database
Azure SQL Database and its Configuration
Azure Network Service and Azure Virtual Network
Azure Media Service and Database Service
Azure Backup and Virtual Machine Security
Azure Availability Zones and Sets and VNet Connectivity
Azure App Service Monitoring and Azure CDN
eBooks
Interview Questions
Videos
How to configure Azure Cloud Service
How to configure Azure Load Balancer
How to Configure Azure Storage Security
How to create Azure Mobile App
Overview of Microsoft Azure and Cloud Computing
Creating App Service Plan in Azure Portal
Azure Virtual Machines and Compute Service
Azure Virtual Machine Scale Set and Auto Scaling
Azure Table, Queue and Disk Storage
Azure Storage Monitoring and Resource Tool
Azure Storage Building Blocks and Storage Account
Azure Storage account and Blob service configuration
Azure SQL Managed Instance and SQL Stretch Database
Azure SQL Database and its Configuration
Azure Network Service and Azure Virtual Network
Azure Media Service and Database Service
Azure Backup and Virtual Machine Security
Azure Availability Zones and Sets and VNet Connectivity
Azure App Service Monitoring and Azure CDN
Azure App Service Backup and Security
How to Use Amazon Machine Learning
How to use Amazon KCL and set up Amazon EMR
How to Set Up Amazon RDS (Relational Database Service)
How to Configure AWS Direct Connect
How to Configure Amazon Simple Storage Service (S3)
How to Configure Amazon Route 53
How AWS CloudFront Delivers the Content
Amazon Elastic Block Storage (EBS) and Storage Gateway
How to use Simple Workflow Service (SWF) and Amazon WorkMail
Understanding of AWS Management Console
How to configure Azure Cloud Service
How to configure Azure Load Balancer
How to Configure Azure Storage Security
How to create Azure Mobile App
Overview of Microsoft Azure and Cloud Computing
Creating App Service Plan in Azure Portal
Azure Virtual Machines and Compute Service
Azure Virtual Machine Scale Set and Auto Scaling
Azure Table, Queue and Disk Storage
Azure Storage Monitoring and Resource Tool
Azure Storage Building Blocks and Storage Account
Azure Storage account and Blob service configuration
How to Set Up AWS Data Pipeline
How to Create Amazon Workspaces
Azure SQL Managed Instance and SQL Stretch Database
Azure SQL Database and its Configuration
Azure Network Service and Azure Virtual Network
Azure Media Service and Database Service
Azure Backup and Virtual Machine Security
Azure Availability Zones and Sets and VNet Connectivity
Azure App Service Monitoring and Azure CDN
Azure App Service Backup and Security
Azure API Apps and API Management
Top Azure Developer Interview Questions and Answers
Top Azure Architect Interview Questions and Answers
Top Amazon Web Services (AWS) Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Migration of 3-tier e-commerce web application using Amazon web Services (AWS)
Cyber Security
Articles
eBooks
Interview Questions
Videos
Ethical Hacking
Articles
Concept of Enumeration in Ethical Hacking
Concept of Exploitation in Ethical Hacking
Concept of Social Engineering Attacks and Cross-Site Scripting
Concept of SQL Injection Attack
Concept of TCP/IP Hijacking and Trojan Attacks
DDOS Attacks in Ethical Hacking
Ethical Hacking - Fingerprinting
Ethical Hacking - Footprinting
Processes in Ethical Hacking and Reconnaissance
eBooks
Interview Questions
Videos
Concept of Enumeration in Ethical Hacking
Concept of Exploitation in Ethical Hacking
Concept of Social Engineering Attacks and Cross-Site Scripting
Concept of SQL Injection Attack
Concept of TCP/IP Hijacking and Trojan Attacks
DDOS Attacks in Ethical Hacking
Ethical Hacking - Fingerprinting
Ethical Hacking - Footprinting
Processes in Ethical Hacking and Reconnaissance
Concept of Enumeration in Ethical Hacking
Concept of Exploitation in Ethical Hacking
Concept of Social Engineering Attacks and Cross-Site Scripting
Concept of SQL Injection Attack
Concept of TCP/IP Hijacking and Trojan Attacks
DDOS Attacks in Ethical Hacking
Ethical Hacking - Fingerprinting
Ethical Hacking - Footprinting
Processes in Ethical Hacking and Reconnaissance
Data Science
Articles
Regression Analysis in Machine learning
Regression vs Classification in Machine Learning
Simple Linear Regression in Machine Learning
Naïve Bayes Classifier Algorithm
Support Vector Machine Algorithm
Logistic Regression in Machine Learning
Linear Regression in Machine Learning
K-Nearest Neighbor (KNN) Algorithm for Machine Learning
eBooks
Interview Questions
Videos
Machine Learning
Python with Data Science
Articles
Processing JSON Data in Python and Matplotlib
Processing Unstructured Data and rectilinear regression and Chi-Square Test in Python
P-Value and Correlation in Python
Python - Data Science Introduction
Relational Databases in Python
Perform Data Cleansing in Python
Performing Data Wrangling in Python
Introduction to Pandas, NumPy and SciPy Libraries
How to Read HTML Pages in Python
How to interact with MongoDB in Python
Box Plots and Scatter Plots and Heat Maps in Python
Bubble Charts and 3D Charts in Python
Data Aggregation and binomial distribution in Python
How to create Geographical Maps and Graphs in Python
Measuring Central Tendency and Variance in Python
eBooks
Interview Questions
Videos
Processing JSON Data in Python and Matplotlib
Processing Unstructured Data and rectilinear regression and Chi-Square Test in Python
P-Value and Correlation in Python
Python - Data Science Introduction
Relational Databases in Python
Perform Data Cleansing in Python
Performing Data Wrangling in Python
Introduction to Pandas, NumPy and SciPy Libraries
How to Read HTML Pages in Python
How to interact with MongoDB in Python
Box Plots and Scatter Plots and Heat Maps in Python
Bubble Charts and 3D Charts in Python
Data Aggregation and binomial distribution in Python
How to create Geographical Maps and Graphs in Python
Measuring Central Tendency and Variance in Python
Normal, Binomial and Poisson distribution in Python
R Language
Articles
Arrays and Factors in R Language
Binomial Distribution and Poisson Regression in R
Analysis of Covariance in R Language
Decision making and Loops in R Language
Handling Excel and Binary Files in R
Handling XML Files in R Language
How to create Line Graphs in R
How to create Scatterplots in R
How to create Histograms and Box Plots in R
Random Forest and Survival Analysis in R
Operators and Variables in R Language
Normal Distribution in R Language
Multiple and Logistic Regression in R
eBooks
Interview Questions
Videos
Arrays and Factors in R Language
Binomial Distribution and Poisson Regression in R
Analysis of Covariance in R Language
Decision making and Loops in R Language
Handling Excel and Binary Files in R
Handling XML Files in R Language
How to create Line Graphs in R
How to create Scatterplots in R
How to create Histograms and Box Plots in R
Random Forest and Survival Analysis in R
Operators and Variables in R Language
Normal Distribution in R Language
Multiple and Logistic Regression in R
SAS
Articles
One Way Anova and Hypothesis Testing
Overview of SAS and its Features
SAS - Basic Syntax and Program Structure
How to Perform Standard Deviation in SAS
How to perform Correlation Analysis in SAS
How to perform Bland Altman Analysis
SAS Applications and Loops and Decision Making
SAS Intelligence Platform Architecture
Strings Manipulation and Arrays in SAS
How to Format Data Sets in SAS
How to create Scatter Plots in SAS
How to create Pie Charts in SAS
How to create Histogram and Simulations in SAS
How to create Box Plots in SAS
How to Create Bar Charts in SAS
How to Concatenate Data Sets in SAS
How to calculate Arithmetic Mean and Handling Data and Time
Frequency Distributions and Cross Tabulations in SAS
eBooks
Interview Questions
Videos
One Way Anova and Hypothesis Testing
Overview of SAS and its Features
SAS - Basic Syntax and Program Structure
How to Perform Standard Deviation in SAS
How to perform Correlation Analysis in SAS
How to perform Bland Altman Analysis
SAS Applications and Loops and Decision Making
SAS Intelligence Platform Architecture
Strings Manipulation and Arrays in SAS
How to Format Data Sets in SAS
How to create Scatter Plots in SAS
How to create Pie Charts in SAS
How to create Histogram and Simulations in SAS
How to create Box Plots in SAS
How to Create Bar Charts in SAS
How to Concatenate Data Sets in SAS
How to calculate Arithmetic Mean and Handling Data and Time
Frequency Distributions and Cross Tabulations in SAS
Fishers Exact Tests and Repeated Measure Analysis in SAS
Regression Analysis in Machine learning
Regression vs Classification in Machine Learning
Simple Linear Regression in Machine Learning
Naïve Bayes Classifier Algorithm
Support Vector Machine Algorithm
Logistic Regression in Machine Learning
Linear Regression in Machine Learning
K-Nearest Neighbor (KNN) Algorithm for Machine Learning
Difference between Supervised and Unsupervised Learning
Classification Algorithm in Machine Learning
How to get datasets for Machine Learning
Processing JSON Data in Python and Matplotlib
Processing Unstructured Data and rectilinear regression and Chi-Square Test in Python
P-Value and Correlation in Python
Python - Data Science Introduction
Relational Databases in Python
One Way Anova and Hypothesis Testing
Overview of SAS and its Features
SAS - Basic Syntax and Program Structure
How to Perform Standard Deviation in SAS
How to perform Correlation Analysis in SAS
How to perform Bland Altman Analysis
SAS Applications and Loops and Decision Making
SAS Intelligence Platform Architecture
Strings Manipulation and Arrays in SAS
Perform Data Cleansing in Python
Performing Data Wrangling in Python
Introduction to Pandas, NumPy and SciPy Libraries
How to Read HTML Pages in Python
How to interact with MongoDB in Python
Box Plots and Scatter Plots and Heat Maps in Python
Arrays and Factors in R Language
Binomial Distribution and Poisson Regression in R
Bubble Charts and 3D Charts in Python
Analysis of Covariance in R Language
Difference between Artificial intelligence and Machine learning
Data Preprocessing in Machine learning
Data Aggregation and binomial distribution in Python
How to create Geographical Maps and Graphs in Python
Measuring Central Tendency and Variance in Python
Introduction to Machine Learning
Normal, Binomial and Poisson distribution in Python
Installing Anaconda and Python
Applications of Machine learning
Handle Date and Time in Python
Decision making and Loops in R Language
Handling Excel and Binary Files in R
Handling XML Files in R Language
How to create Line Graphs in R
How to create Scatterplots in R
How to create Histograms and Box Plots in R
How to Format Data Sets in SAS
How to create Scatter Plots in SAS
How to create Pie Charts in SAS
How to create Histogram and Simulations in SAS
How to create Box Plots in SAS
How to Create Bar Charts in SAS
How to Concatenate Data Sets in SAS
How to calculate Arithmetic Mean and Handling Data and Time
Frequency Distributions and Cross Tabulations in SAS
Fishers Exact Tests and Repeated Measure Analysis in SAS
Advantages and Disadvantages of SAS Programming Language
Random Forest and Survival Analysis in R
Operators and Variables in R Language
Normal Distribution in R Language
Multiple and Logistic Regression in R
Linear Regression in R Language
Top Data Science Interview Questions and Answers
Top Machine Learning Interview Questions and Answers
Top SAS Interview Questions and Answers
Top Python Interview Questions and Answers
Data Warehousing and ETL
Articles
eBooks
Interview Questions
Videos
ETL Testing
eBooks
Interview Questions
Videos
Informatica
Articles
Aggregator Transformation in Informatica
Concept of Informatica (Big Data Management) BDM
Informatica Master Data Management (MDM) Process
Lookup and Normalizer Transformation in Informatica
Performance Tuning and Partitioning in Informatica
Rank Transformation in Informatica
Router and Joiner Transformation in Informatica
Source Qualifier Transformation in Informatica
Transaction Control Transformation in Informatica
Sequence Generator Transformation in Informatica
eBooks
Interview Questions
Videos
Aggregator Transformation in Informatica
Concept of Informatica (Big Data Management) BDM
Informatica Master Data Management (MDM) Process
Lookup and Normalizer Transformation in Informatica
Performance Tuning and Partitioning in Informatica
Rank Transformation in Informatica
Router and Joiner Transformation in Informatica
Source Qualifier Transformation in Informatica
Transaction Control Transformation in Informatica
Sequence Generator Transformation in Informatica
Concept of Informatica IDQ (Informatica Data Quality)
Concept of ETL Pipeline and Files
Overview of ELT Testing and its Architecture
Aggregator Transformation in Informatica
Concept of Informatica (Big Data Management) BDM
Informatica Master Data Management (MDM) Process
Lookup and Normalizer Transformation in Informatica
Performance Tuning and Partitioning in Informatica
Rank Transformation in Informatica
Router and Joiner Transformation in Informatica
Source Qualifier Transformation in Informatica
Transaction Control Transformation in Informatica
Sequence Generator Transformation in Informatica
Concept of Informatica IDQ (Informatica Data Quality)
Installation of Informatica PowerCenter
Comparison between ETL and ELT
Detailed understanding of ETL (Extraction, Transformation and Loading) Testing
Databases
Articles
eBooks
Interview Questions
Videos
MS-SQL Server
Articles
Backup and restore a database in SQL Server
Concept of Primary Key in SQL Server
CRUD Operations of Data in MS SQL Server
How to Enable, Disable and Drop a Foreign Key
Popular Functions in MS SQL Server
SQL Server BETWEEN Condition (Operator)
SQL Server Comparison Operator
Create and Delete Table in MS SQL Server
SQL Server DISTINCT and GROUP BY Clause
eBooks
Interview Questions
Videos
Backup and restore a database in SQL Server
Concept of Primary Key in SQL Server
CRUD Operations of Data in MS SQL Server
How to Enable, Disable and Drop a Foreign Key
Popular Functions in MS SQL Server
SQL Server BETWEEN Condition (Operator)
SQL Server Comparison Operator
Create and Delete Table in MS SQL Server
SQL Server DISTINCT and GROUP BY Clause
Oracle DBA
Overview of Oracle Tablespace Group
Overview of Oracle Database and its Architecture
Oracle ALTER USER and DROP USER
Introduction to Oracle Data Pump Import and Export tool
Introduction to Oracle CREATE USER statement
How to use the Oracle STARTUP command to start out an Oracle Database instance
How to shut down the Oracle Database
How to Manage Tablespaces in Oracle
How To List Users within the Oracle Database
How to Grant SELECT Object Privilege on One or More Tables to a User and Unlock a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Create User Profiles in Oracle
How to Create Oracle Database Links
How to Alter and Drop Roles in Oracle
Oracle PL-SQL
eBooks
Interview Questions
Videos
Date and Time Handling in PL-SQL
Constants and Literals and Operators in PL-SQL
Conditions and Loops in PL-SQL
Backup and restore a database in SQL Server
Concept of Primary Key in SQL Server
CRUD Operations of Data in MS SQL Server
How to Enable, Disable and Drop a Foreign Key
Popular Functions in MS SQL Server
SQL Server BETWEEN Condition (Operator)
SQL Server Comparison Operator
Create and Delete Table in MS SQL Server
SQL Server DISTINCT and GROUP BY Clause
SQL Server NOT Condition (Operator)
Overview of Oracle Tablespace Group
Overview of Oracle Database and its Architecture
Oracle ALTER USER and DROP USER
Introduction to Oracle Data Pump Import and Export tool
Introduction to Oracle CREATE USER statement
How to use the Oracle STARTUP command to start out an Oracle Database instance
How to shut down the Oracle Database
How to Manage Tablespaces in Oracle
How To List Users within the Oracle Database
How to Grant SELECT Object Privilege on One or More Tables to a User and Unlock a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Grant All Privileges to a User in Oracle
How to Create User Profiles in Oracle
How to Create Oracle Database Links
How to Alter and Drop Roles in Oracle
How to Alter and Drop Oracle Database Link
Introduction to PL-SQL and Environment setup
Top Oracle PL-SQL Interview Questions and Answers
Top MS-SQL Server Interview Questions and Answers
Best Approach for Storing data to AWS DynamoDB and S3 – AWS Implementation
Maintain High Availability in AWS with anticipated Additional Load
DevOps
eBooks
Interview Questions
Videos
Ansible
Articles
Overview of YAML and Ad-hoc commands in Ansible
A Detailed comparison of Ansible and Puppet
A Detailed comparison of Ansible Vs Chef
Detailed understanding of concept of Playbooks in Ansible
Deep dive into Pip module in Ansible
How to perform troubleshooting in Ansible
How to use variables in playbooks in Ansible and concept of exception handling
Overview of Ansible, its History and How to set-up Ansible on your machine
eBooks
Interview Questions
Videos
Overview of YAML and Ad-hoc commands in Ansible
A Detailed comparison of Ansible and Puppet
A Detailed comparison of Ansible Vs Chef
Detailed understanding of concept of Playbooks in Ansible
Deep dive into Pip module in Ansible
How to perform troubleshooting in Ansible
How to use variables in playbooks in Ansible and concept of exception handling
Overview of Ansible, its History and How to set-up Ansible on your machine
Chef
Articles
Chef-Client as Daemon and Chef-Shell
Concept of Libraries , Definition and setting environment variable
Concept of Lightweight Resource Provider and Blueprints in Chef
Concept of Templates and Dynamically Configuring Recipes
Dealing with Files and Software packages and Community Cookbooks
Execute Cookbook on Node and run Chef-Client
Detailed understanding of Resources in Chef
How to Set up Chef on your system
How to set up Test Kitchen Workflow
How to write Cross-Platform Cookbooks
Overview of Chef and its Architecture
Plain Ruby with Chef DSL and Ruby Gems with Recipes
Testing Cookbook with Test Kitchen
Roles in Chef and perform environment specific configuration
eBooks
Interview Questions
Videos
Chef-Client as Daemon and Chef-Shell
Concept of Libraries , Definition and setting environment variable
Concept of Lightweight Resource Provider and Blueprints in Chef
Concept of Templates and Dynamically Configuring Recipes
Dealing with Files and Software packages and Community Cookbooks
Execute Cookbook on Node and run Chef-Client
Detailed understanding of Resources in Chef
How to Set up Chef on your system
How to set up Test Kitchen Workflow
How to write Cross-Platform Cookbooks
Overview of Chef and its Architecture
Plain Ruby with Chef DSL and Ruby Gems with Recipes
Testing Cookbook with Test Kitchen
Roles in Chef and perform environment specific configuration
Docker
Articles
Concept of Docker Cloud Service
Deep dive into Docker Architecture
Concept of public repositories in Docker
Concept of Container Linking and Storage in Docker
Building a web Server Docker File
Working with Docker Toolbox and how to use the Jenkins Docker image from Docker Hub
Overview of Docker and its features
Managing ports and private registries in Docker
Instruction commands in Docker
How to work with Containers in Docker
How to Set-up MongoDB in Docker
How to set up Node.js in Docker
How to set up Kubernetes in Docker
How to set up ASP.net in Docker
How to perform Continuous integration using Jenkins in Docker
How to install Docker on Windows
eBooks
Interview Questions
Videos
Concept of Docker Cloud Service
Deep dive into Docker Architecture
Concept of public repositories in Docker
Concept of Container Linking and Storage in Docker
Building a web Server Docker File
Working with Docker Toolbox and how to use the Jenkins Docker image from Docker Hub
Overview of Docker and its features
Managing ports and private registries in Docker
Instruction commands in Docker
How to work with Containers in Docker
How to Set-up MongoDB in Docker
How to set up Node.js in Docker
How to set up Kubernetes in Docker
How to set up ASP.net in Docker
How to perform Continuous integration using Jenkins in Docker
How to install Docker on Windows
How to install docker on Linux
Git and GitHub
eBooks
Interview Questions
Videos
Concept of Git Index and Git Head
Comparison of Git with SVN and Mercurial
Deep dive into Git Branching Model
Git Repository and How to Fork Repository
Git Terminology and General Tools
How to Clone Repository in Git
Working with Remote Repository
Version Control System and its Types
Overview of GitHub and Comparison of Git and GitHub
Overview of Git and its features
Merging Branches and Resolve conflicts in Git
How to use Git via the command line
How to switch branches without committing the current branch in Git
How to perform Rebasing in Git
How to Install Git on Linux (Ubuntu) and Mac
Jenkins
Articles
Deep dive into Metrics and Trends for builds
Server maintenance and Plugins Management in Jenkins
Perform Continuous Deployment using Jenkins
Overview of Jenkins, its History and Architecture
How to take Back-up in Jenkins using Backup plugin
How to set up Git and Maven Plugin in Jenkins
How to set up Distributed build and Automated deployment in Jenkins
How to set up Build jobs in Jenkins
How to run Remote tests using Jenkins
How to perform Notification, Reporting and Code Analysis
How to perform Junit Testing in Jenkins
How to perform Automation Testing in Jenkins
How to install Jenkins on your system
Comparison of Jenkins with Ansible and Hudson Frameworks
Comparison of Jenkins with Bamboo and TeamCity
eBooks
Interview Questions
Videos
Deep dive into Metrics and Trends for builds
Server maintenance and Plugins Management in Jenkins
Perform Continuous Deployment using Jenkins
Overview of Jenkins, its History and Architecture
How to take Back-up in Jenkins using Backup plugin
How to set up Git and Maven Plugin in Jenkins
How to set up Distributed build and Automated deployment in Jenkins
How to set up Build jobs in Jenkins
How to run Remote tests using Jenkins
How to perform Notification, Reporting and Code Analysis
How to perform Junit Testing in Jenkins
How to perform Automation Testing in Jenkins
How to install Jenkins on your system
Comparison of Jenkins with Ansible and Hudson Frameworks
Comparison of Jenkins with Bamboo and TeamCity
Comparison of Jenkins with GoCD and Maven Tools
Kubernetes
Articles
eBooks
Interview Questions
Videos
How to setup Kubernetes on your machine
How to Set up Kubernetes Dashboard
How to manage Deployments and Concept of Kubernetes Volume
How to achieve Autoscaling in Kubernetes cluster
Deep dive into Kubectl command line utility
Create an Application for Kubernetes deployment
Concept of Secrets, Network Policy and Kubernetes API
Concept of Replication Controller and Replica Sets
Concept of Node, Service and Pod in Kubernetes
Concept of Images and creating a Job in Kubernetes
Namespace, Labels and Selectors in Kubernetes
Overview of Kubernetes and its Architecture and components
Maven
Articles
Introduction to Maven and How to Set up Maven Environment
How to manage Maven Project in NetBeans and IntelliJ IDEA
How to manage a web-based project using Maven
How to import Maven Project in Eclipse IDE
How to create documentation of Application in Maven
How to automate the Deployment process in Maven
Deep dive into Build Automation
Creating Java Project in Maven
Concept of Project Object Model (POM) in Maven
Concept of Maven Repositories and Plugins in Maven
eBooks
Interview Questions
Videos
Introduction to Maven and How to Set up Maven Environment
How to manage Maven Project in NetBeans and IntelliJ IDEA
How to manage a web-based project using Maven
How to import Maven Project in Eclipse IDE
How to create documentation of Application in Maven
How to automate the Deployment process in Maven
Deep dive into Build Automation
Creating Java Project in Maven
Concept of Project Object Model (POM) in Maven
Concept of Maven Repositories and Plugins in Maven
Nagios
Articles
Look into Nagios Features, applications, Hosts and services and Commands
Overview of Nagios, its architecture and Nagios products
Ports and protocols and Add-ons and plugins in Nagios
Detailed understanding of Checks and States in Nagios
How to run Nagios plugins on other machines remotely using NRPE
eBooks
Interview Questions
Videos
Look into Nagios Features, applications, Hosts and services and Commands
Overview of Nagios, its architecture and Nagios products
Ports and protocols and Add-ons and plugins in Nagios
Detailed understanding of Checks and States in Nagios
How to run Nagios plugins on other machines remotely using NRPE
Puppet
Articles
Implementation of Live working demo project in Puppet
How to Set-up and configure Puppet Master
How to install and configure r10k tool and validate puppet setup
How to install and configure puppet on your machine
How to define Functions and Custom functions in Puppet
Concept of Templating in Puppet
Concept of Type and Provider in Puppet
How to create custom environment in Puppet
Detailed understanding of architecture of puppet and its components and application of puppet
Detail understanding of environment conf file in puppet
Deep Dive into Resources in Puppet
Concept of Resource Abstraction Layer (RAL) in Puppet
Concept of File Server in Puppet
Concept of Facter and Facts in Puppet
Understanding Puppet Manifest files and How to write a manifest file in Puppet
Overview of Puppet and its components and concept of configuration management
How to Set-up Puppet agent and How to sign and check for SSL Ceritficate
eBooks
Interview Questions
Videos
How to use RESTful APIs in Puppet
Implementation of Live working demo project in Puppet
How to Set-up and configure Puppet Master
How to install and configure r10k tool and validate puppet setup
How to install and configure puppet on your machine
How to define Functions and Custom functions in Puppet
Concept of Templating in Puppet
Concept of Type and Provider in Puppet
How to create custom environment in Puppet
Detailed understanding of architecture of puppet and its components and application of puppet
Detail understanding of environment conf file in puppet
Deep Dive into Resources in Puppet
Concept of Resource Abstraction Layer (RAL) in Puppet
Concept of File Server in Puppet
Concept of Facter and Facts in Puppet
Understanding Puppet Manifest files and How to write a manifest file in Puppet
Overview of Puppet and its components and concept of configuration management
How to Set-up Puppet agent and How to sign and check for SSL Ceritficate
Concept of Git Index and Git Head
Comparison of Git with SVN and Mercurial
Deep dive into Git Branching Model
Git Repository and How to Fork Repository
Git Terminology and General Tools
How to Clone Repository in Git
Deep dive into Metrics and Trends for builds
Concept of Docker Cloud Service
Server maintenance and Plugins Management in Jenkins
How to use RESTful APIs in Puppet
Implementation of Live working demo project in Puppet
Perform Continuous Deployment using Jenkins
Overview of Jenkins, its History and Architecture
How to take Back-up in Jenkins using Backup plugin
How to set up Git and Maven Plugin in Jenkins
How to set up Distributed build and Automated deployment in Jenkins
How to set up Build jobs in Jenkins
How to run Remote tests using Jenkins
How to perform Notification, Reporting and Code Analysis
How to perform Junit Testing in Jenkins
How to perform Automation Testing in Jenkins
How to install Jenkins on your system
Deep dive into Docker Architecture
Concept of public repositories in Docker
Concept of Container Linking and Storage in Docker
Building a web Server Docker File
Comparison of Jenkins with Ansible and Hudson Frameworks
How to setup Kubernetes on your machine
How to Set up Kubernetes Dashboard
How to manage Deployments and Concept of Kubernetes Volume
How to achieve Autoscaling in Kubernetes cluster
Deep dive into Kubectl command line utility
Create an Application for Kubernetes deployment
Concept of Secrets, Network Policy and Kubernetes API
Concept of Replication Controller and Replica Sets
Concept of Node, Service and Pod in Kubernetes
Concept of Images and creating a Job in Kubernetes
How to Set-up and configure Puppet Master
How to install and configure r10k tool and validate puppet setup
How to install and configure puppet on your machine
How to define Functions and Custom functions in Puppet
Concept of Templating in Puppet
Concept of Type and Provider in Puppet
How to create custom environment in Puppet
Detailed understanding of architecture of puppet and its components and application of puppet
Detail understanding of environment conf file in puppet
Deep Dive into Resources in Puppet
Concept of Resource Abstraction Layer (RAL) in Puppet
Concept of File Server in Puppet
Concept of Facter and Facts in Puppet
Working with Docker Toolbox and how to use the Jenkins Docker image from Docker Hub
Overview of Docker and its features
Managing ports and private registries in Docker
Instruction commands in Docker
How to work with Containers in Docker
How to Set-up MongoDB in Docker
How to set up Node.js in Docker
How to set up Kubernetes in Docker
How to set up ASP.net in Docker
How to perform Continuous integration using Jenkins in Docker
How to install Docker on Windows
How to install docker on Linux
Namespace, Labels and Selectors in Kubernetes
Comparison of Jenkins with Bamboo and TeamCity
Comparison of Jenkins with GoCD and Maven Tools
Comparison of Jenkins with Travis CI and Circle CI
Working with Remote Repository
Version Control System and its Types
Overview of GitHub and Comparison of Git and GitHub
Overview of Git and its features
Merging Branches and Resolve conflicts in Git
How to use Git via the command line
How to switch branches without committing the current branch in Git
How to perform Rebasing in Git
How to Install Git on Linux (Ubuntu) and Mac
How to create a new Blank Repository and commit code in it
Overview of Kubernetes and its Architecture and components
Monitor processes in Kubernetes
Introduction to Maven and How to Set up Maven Environment
How to manage Maven Project in NetBeans and IntelliJ IDEA
How to manage a web-based project using Maven
How to import Maven Project in Eclipse IDE
How to create documentation of Application in Maven
How to automate the Deployment process in Maven
Deep dive into Build Automation
Creating Java Project in Maven
Concept of Project Object Model (POM) in Maven
Concept of Maven Repositories and Plugins in Maven
Concept of Dependency Management in Maven
Understanding Puppet Manifest files and How to write a manifest file in Puppet
Overview of Puppet and its components and concept of configuration management
Look into Nagios Features, applications, Hosts and services and Commands
Overview of Nagios, its architecture and Nagios products
Ports and protocols and Add-ons and plugins in Nagios
Detailed understanding of Checks and States in Nagios
How to run Nagios plugins on other machines remotely using NRPE
Chef-Client as Daemon and Chef-Shell
Concept of Libraries , Definition and setting environment variable
Concept of Lightweight Resource Provider and Blueprints in Chef
Concept of Templates and Dynamically Configuring Recipes
Dealing with Files and Software packages and Community Cookbooks
Execute Cookbook on Node and run Chef-Client
Detailed understanding of Resources in Chef
How to Set up Chef on your system
How to set up Test Kitchen Workflow
How to write Cross-Platform Cookbooks
Overview of Chef and its Architecture
Plain Ruby with Chef DSL and Ruby Gems with Recipes
Testing Cookbook with Test Kitchen
How to Set-up Puppet agent and How to sign and check for SSL Ceritficate
Roles in Chef and perform environment specific configuration
Overview of YAML and Ad-hoc commands in Ansible
A Detailed comparison of Ansible and Puppet
A Detailed comparison of Ansible Vs Chef
Detailed understanding of concept of Playbooks in Ansible
Deep dive into Pip module in Ansible
How to perform troubleshooting in Ansible
How to use variables in playbooks in Ansible and concept of exception handling
Overview of Ansible, its History and How to set-up Ansible on your machine
Concept of Advanced Execution with Ansible
Popular DevOps and DevOps Automation Tools
Comparison between DevOps and Agile methodologies
Concept of DevOps Pipeline and Who are DevOps Engineers
Overview of DevOps and its Architecture
DevOps Training Certification and Azure and AWS DevOps
How to set-up Nagios on Ubuntu
Top Docker Interview Questions and Answers
Top Ansible Interview Questions and Answers
Top Chef Interview Questions and Answers
Top Git and GitHub Interview Questions and Answers
Top DevOps Interview Questions and Answers
Top Puppet Interview Questions and Answers
Top Nagios Interview Questions and Answers
Top Kubernetes Interview Questions and Answers
Digital Marketing
Articles
Understanding Mobile marketing
Understanding Google Analytics
Online Marketing - Web Analytics
Why can we need an SEO Friendly Website?
Concept of Pay Per Click (PPC) and Conversion Rate Optimization (CRO) explained
Online Marketing - Impact, Pros & Cons
Online Marketing - Blogs, banners and forums
Introduction to Online Marketing
Digital Marketing using Twitter and LinkedIn
Digital Marketing using Social Media and YouTube
Digital Marketing using Facebook and Pinterest
Digital Marketing using Content marketing and Email Marketing
eBooks
Interview Questions
Videos
SEO and SMM
Articles
Social Media Marketing using Blogs
Social Media Marketing using Facebook
Social Media Marketing using Google Plus
Social Media Marketing using Linkedin
Social Media Marketing using Pinterest
Social Media Marketing using Twitter
Social Media Marketing using Video
Social Media Analysis and Monitoring Social Media Accounts
SMM - Image Optimization and Social Bookmarking
eBooks
Interview Questions
Videos
Social Media Marketing using Blogs
Social Media Marketing using Facebook
Social Media Marketing using Google Plus
Social Media Marketing using Linkedin
Social Media Marketing using Pinterest
Social Media Marketing using Twitter
Social Media Marketing using Video
Social Media Analysis and Monitoring Social Media Accounts
SMM - Image Optimization and Social Bookmarking
Social Media Marketing using Blogs
Social Media Marketing using Facebook
Social Media Marketing using Google Plus
Social Media Marketing using Linkedin
Social Media Marketing using Pinterest
Social Media Marketing using Twitter
Social Media Marketing using Video
Understanding Mobile marketing
Understanding Google Analytics
Online Marketing - Web Analytics
Why can we need an SEO Friendly Website?
Concept of Pay Per Click (PPC) and Conversion Rate Optimization (CRO) explained
Online Marketing - Impact, Pros & Cons
Online Marketing - Blogs, banners and forums
Introduction to Online Marketing
Digital Marketing using Twitter and LinkedIn
Digital Marketing using Social Media and YouTube
Digital Marketing using Facebook and Pinterest
Digital Marketing using Content marketing and Email Marketing
Overview of Digital Marketing and SEO
Social Media Analysis and Monitoring Social Media Accounts
SMM - Image Optimization and Social Bookmarking
SEO Strategy to Optimize Keywords and Metatags
Affiliate Marketing and Email Marketing
Frontend Development
Articles
eBooks
Interview Questions
Videos
Angular JS
Articles
Create Angular Application and Angular MVC Architecture
Custom Directives in Angular JS
Dependency Injection in Angular JS
Directives and Filters in Angular JS
Embedding Html Pages within HTML page
Expressions and Controllers in Angular JS
How to create Forms in Angular JS
How to create Single Page Application via multiple views
Internationalization in Angular JS
Services Architecture in Angular JS
Spring Angular CRUD Application
Spring Angular Login & Logout Application
Spring Angular Search Field Application
Tables and HTML DOM in Angular JS
Using Directives and Expressions in Angular JS
eBooks
Interview Questions
Videos
Create Angular Application and Angular MVC Architecture
Custom Directives in Angular JS
Dependency Injection in Angular JS
Directives and Filters in Angular JS
Embedding Html Pages within HTML page
Expressions and Controllers in Angular JS
How to create Forms in Angular JS
How to create Single Page Application via multiple views
Internationalization in Angular JS
Services Architecture in Angular JS
Spring Angular CRUD Application
Spring Angular Login & Logout Application
Spring Angular Search Field Application
Tables and HTML DOM in Angular JS
Using Directives and Expressions in Angular JS
React JS
Articles
Comparison Between AngularJS and ReactJS
How to implement flux pattern in React Applications
How to Animate elements using React
Error Handling using Error Boundaries
Environment Setup for React JS
Component Life Cycle Methods in React JS
Comparison between ReactJS and React Native
Overview of ReactJS and its Features
Overview of React Redux with an example
How to set up Router for an app
Using Refs and Keys in React JS
eBooks
Interview Questions
Videos
Comparison Between AngularJS and ReactJS
How to implement flux pattern in React Applications
How to Animate elements using React
Error Handling using Error Boundaries
Environment Setup for React JS
Component Life Cycle Methods in React JS
Comparison between ReactJS and React Native
Overview of ReactJS and its Features
Overview of React Redux with an example
How to set up Router for an app
Using Refs and Keys in React JS
Create Angular Application and Angular MVC Architecture
Custom Directives in Angular JS
Dependency Injection in Angular JS
Directives and Filters in Angular JS
Embedding Html Pages within HTML page
Expressions and Controllers in Angular JS
How to create Forms in Angular JS
How to create Single Page Application via multiple views
Internationalization in Angular JS
Services Architecture in Angular JS
Spring Angular CRUD Application
Spring Angular Login & Logout Application
Spring Angular Search Field Application
Tables and HTML DOM in Angular JS
Using Directives and Expressions in Angular JS
How to Setup AngularJS Environment
Comparison Between AngularJS and ReactJS
How to implement flux pattern in React Applications
How to Animate elements using React
Error Handling using Error Boundaries
Environment Setup for React JS
Component Life Cycle Methods in React JS
Comparison between ReactJS and React Native
Overview of ReactJS and its Features
Overview of React Redux with an example
How to set up Router for an app
Using Refs and Keys in React JS
Understanding ReactJS Components
Top React JS Interview Questions and Answers
IOT
Articles
IoT project of controlling home light using WiFi Node MCU, and Relay module
IoT project of Sonar system using Ultrasonic Sensor HC-SR04 and Arduino device
IoT project of Temperature and Pressure measurement using Pressure sensor BMP180 and Arduino device
IoT (Internet of Things) Project: Google Firebase controlling LED with NodeMCU
IoT link Communication Protocol
IoT Decision Framework and Architecture
IoT in Energy and Biometrics Domain
IoT in Security Camera and Smart Home
IoT in Smart Agriculture and Healthcare Domain
IoT Network Layer and Session Layer Protocols
IoT – Platform and Thing Worx in IoT
IoT Project Google Firebase controlling LED using Android App
IoT Project: Google Firebase using NodeMCU ESP8266
IoT project of controlling home light using WiFi Node MCU, and Relay module
Overview of Internet of Things (IoT)
CISCO Virtualized Packet Zone and Salesforce in IoT
Embedded Devices (System) in (IoT) and IoT Ecosystem
GE Predix Platform and Eclipse IoT
How is IoT transforming businesses and IoT in transportation
eBooks
Interview Questions
Videos
IoT project of controlling home light using WiFi Node MCU, and Relay module
IoT project of Sonar system using Ultrasonic Sensor HC-SR04 and Arduino device
IoT project of Temperature and Pressure measurement using Pressure sensor BMP180 and Arduino device
IoT (Internet of Things) Project: Google Firebase controlling LED with NodeMCU
IoT link Communication Protocol
IoT Decision Framework and Architecture
IoT in Energy and Biometrics Domain
IoT in Security Camera and Smart Home
IoT in Smart Agriculture and Healthcare Domain
IoT Network Layer and Session Layer Protocols
IoT – Platform and Thing Worx in IoT
IoT Project Google Firebase controlling LED using Android App
IoT Project: Google Firebase using NodeMCU ESP8266
IoT project of controlling home light using WiFi Node MCU, and Relay module
Overview of Internet of Things (IoT)
CISCO Virtualized Packet Zone and Salesforce in IoT
Embedded Devices (System) in (IoT) and IoT Ecosystem
GE Predix Platform and Eclipse IoT
How is IoT transforming businesses and IoT in transportation
Internet of Things – Contiki and Security Flaws
Internet of Things – Security and Identity Protection
Top Internet of Things (IoT) Interview Questions and Answers
Mobile Development
Articles
eBooks
Interview Questions
Videos
Operating Systems
Articles
eBooks
Interview Questions
Videos
Programming and Frameworks
Articles
Cookies in Laravel based web applications
Encryption and Hashing in Laravel
How to create Blade Templates Layout
How to Create Façade in Laravel
How to perform Redirections and connect to Database
Installation Process of Laravel
Introduction to Laravel and its History
Laravel vs CodeIgniter and Laravel Vs Symphony
Laravel vs Django and Laravel vs WordPress
Middleware Mechanism in Laravel
Process of Authentication and Authorization in Laravel
Responses in Laravel web applications
Understanding Release Process in Laravel
How to setup Check/Money Order payment method in Magento 2
Dynamic Content Handling in PHP
eBooks
Interview Questions
Videos
Hibernate and Spring
Articles
How to use Node Package Manager and REPL Terminal
Handling GET and POST Request in NodeJS
Using Sessions and POJO Classes in Hibernate
Transaction Management in Spring
Overview and Architecture of Spring Framework
ORM Overview and Overview of Hibernate
IoC Containers, AOP and JDBC Framework in Spring
Injecting Inner Beans and Collections in Spring
How to use Criteria Queries in Hibernate
How to perform Java Based Configuration in Spring
How to Install Hibernate and its Configuration
eBooks
Interview Questions
Videos
How to use Node Package Manager and REPL Terminal
Handling GET and POST Request in NodeJS
Using Sessions and POJO Classes in Hibernate
Transaction Management in Spring
Overview and Architecture of Spring Framework
ORM Overview and Overview of Hibernate
IoC Containers, AOP and JDBC Framework in Spring
Injecting Inner Beans and Collections in Spring
How to use Criteria Queries in Hibernate
How to perform Java Based Configuration in Spring
How to Install Hibernate and its Configuration
Java
Articles
Variables and Keywords in Java
Transaction Management and Batch Processing in JDBC
StringBuffer and StringBuilder Class in Java
String Vs StringBuffer Vs StringBuilder
Stream API Improvement in Java 9
Static Binding and Dynamic Binding and Final Keyword
Serialization and Reflection in Java
Properties class and Generics in Java
Method Parameter Reflection in Java
Java StringJoiner and ArrayList Vs Vector
Java Queue and Deque Interface
Java Parallel Array Sorting and Type Inference
Java Networking and Socket Programming
Java Nested Interface and Method Overloading and Overriding
Java Method References and Functional Interfaces
Java Garbage Collection and Java Runtime Class
Java forEach loop and Collectors
Java Comments and Naming Conventions
Java 9 Process API Improvement
Java 9 Module System and Control Panel
Java 9 Anonymous Inner Classes Improvement and SafeVarargs Annotation
Introduction to Java and History of Java
Inter-thread communication and Deadlock in Java
How to write the Hello World Java program
How to create Immutable class in Java
Features of Java and C++ Vs Java
ExceptionHandling with MethodOverriding in Java
Difference between JDK, JRE, and JVM
Deep Dive into Threads in Java
Deep Dive into LinkedList in Java
Deep dive into LinkedHashMap and TreeMap
Deep dive into HashSet , LinkedHashSet and TreeSet
Deep Dive into HashMap in Java
Deep Dive into ArrayList in Java
Conditional Statements in Java
Concept of Method Overloading and Method Overriding in Java
Concept of Inheritance and Aggregation in Java
Comparable and Comparator interface in Java
eBooks
Interview Questions
Videos
Variables and Keywords in Java
Transaction Management and Batch Processing in JDBC
StringBuffer and StringBuilder Class in Java
String Vs StringBuffer Vs StringBuilder
Stream API Improvement in Java 9
Static Binding and Dynamic Binding and Final Keyword
Serialization and Reflection in Java
Properties class and Generics in Java
Method Parameter Reflection in Java
Java StringJoiner and ArrayList Vs Vector
Java Queue and Deque Interface
Java Parallel Array Sorting and Type Inference
Java Networking and Socket Programming
Java Nested Interface and Method Overloading and Overriding
Java Method References and Functional Interfaces
Java Garbage Collection and Java Runtime Class
Java forEach loop and Collectors
Java Comments and Naming Conventions
Java 9 Process API Improvement
Java 9 Module System and Control Panel
Java 9 Anonymous Inner Classes Improvement and SafeVarargs Annotation
Introduction to Java and History of Java
Inter-thread communication and Deadlock in Java
How to write the Hello World Java program
How to create Immutable class in Java
Features of Java and C++ Vs Java
ExceptionHandling with MethodOverriding in Java
Difference between JDK, JRE, and JVM
Deep Dive into Threads in Java
Deep Dive into LinkedList in Java
Deep dive into LinkedHashMap and TreeMap
Deep dive into HashSet , LinkedHashSet and TreeSet
Deep Dive into HashMap in Java
Deep Dive into ArrayList in Java
Conditional Statements in Java
Concept of Method Overloading and Method Overriding in Java
Concept of Inheritance and Aggregation in Java
Comparable and Comparator interface in Java
JSP
eBooks
Interview Questions
Videos
Laravel
Articles
Understanding Release Process in Laravel
Responses in Laravel web applications
Process of Authentication and Authorization in Laravel
Middleware Mechanism in Laravel
Laravel vs Django and Laravel vs WordPress
Laravel vs CodeIgniter and Laravel Vs Symphony
Introduction to Laravel and its History
Installation Process of Laravel
How to perform Redirections and connect to Database
How to Create Façade in Laravel
How to create Blade Templates Layout
Encryption and Hashing in Laravel
Cookies in Laravel based web applications
Contracts and CSRF Protection in Laravel
Available Validation Rules of Laravel
eBooks
Interview Questions
Videos
Understanding Release Process in Laravel
Responses in Laravel web applications
Process of Authentication and Authorization in Laravel
Middleware Mechanism in Laravel
Laravel vs Django and Laravel vs WordPress
Laravel vs CodeIgniter and Laravel Vs Symphony
Introduction to Laravel and its History
Installation Process of Laravel
How to perform Redirections and connect to Database
How to Create Façade in Laravel
How to create Blade Templates Layout
Encryption and Hashing in Laravel
Cookies in Laravel based web applications
Contracts and CSRF Protection in Laravel
Available Validation Rules of Laravel
Magento
Articles
Architecture of Magento 2 and Product Overview
How to use the multi language feature of Magento
How to Setup System Theme, Page Title, Layout and New Pages in Magento
How to Setup Shipping Rates and Payment Plans in Magento
How to setup shipping methods in Magento 2
How to Setup Paypal Payment and Google checkout in Magento
How to Setup Newsletter in Magento
How to Setup Google Analytics Youtube Videos and Facebook Likes in Magento
How to setup Check/Money Order payment method in Magento 2
How to set up Zero Subtotal Checkout payment method in Magento 2
How to set up the tax rules, tax rates, and tax zones in Magento 2
How to set up Purchase Order (PO) payment method in Magento 2
How to set up Order Emails in Magento 2
How to set up multiple websites, stores, and store views in Magento 2
How to Set up Contact, Categories, Products and Inventory in Magento
How to set up Cash on Delivery (COD) payment method in Magento 2
How to set up Bank Transfer payment method in Magento 2
How to set up Authorize.net method in Magento 2
How to Manage Tax Classes in Magento
How to Install Magento on your system
How to install Magento 2 using Composer
How to install Magento 2 on windows
Ways for Site Optimization in Magento
Store Configuration in Magento 2
Search Engine Optimization in Magento 2
Products and their Types in Magento 2
Overview of Magento and its Features
Orders Life Cycle in Magento 2
Ways for Site Optimization in Magento
Basic Configuration in Magento 2
Create and Manage CMS (Content Management System) in Magento 2
How to add the product on Home page in Magento 2
How to configure and Manage the Inventory in Magento 2
How to create Attribute Sets in Magento 2
How to create Product Attributes in Magento 2
How to create Product Category in Magento 2
eBooks
Interview Questions
Videos
Architecture of Magento 2 and Product Overview
How to use the multi language feature of Magento
How to Setup System Theme, Page Title, Layout and New Pages in Magento
How to Setup Shipping Rates and Payment Plans in Magento
How to setup shipping methods in Magento 2
How to Setup Paypal Payment and Google checkout in Magento
How to Setup Newsletter in Magento
How to Setup Google Analytics Youtube Videos and Facebook Likes in Magento
How to setup Check/Money Order payment method in Magento 2
How to set up Zero Subtotal Checkout payment method in Magento 2
How to set up the tax rules, tax rates, and tax zones in Magento 2
How to set up Purchase Order (PO) payment method in Magento 2
How to set up Order Emails in Magento 2
How to set up multiple websites, stores, and store views in Magento 2
How to Set up Contact, Categories, Products and Inventory in Magento
How to set up Cash on Delivery (COD) payment method in Magento 2
How to set up Bank Transfer payment method in Magento 2
How to set up Authorize.net method in Magento 2
How to Manage Tax Classes in Magento
How to Install Magento on your system
How to install Magento 2 using Composer
How to install Magento 2 on windows
Ways for Site Optimization in Magento
Store Configuration in Magento 2
Search Engine Optimization in Magento 2
Products and their Types in Magento 2
Overview of Magento and its Features
Orders Life Cycle in Magento 2
Ways for Site Optimization in Magento
Basic Configuration in Magento 2
Create and Manage CMS (Content Management System) in Magento 2
How to add the product on Home page in Magento 2
How to configure and Manage the Inventory in Magento 2
How to create Attribute Sets in Magento 2
How to create Product Attributes in Magento 2
How to create Product Category in Magento 2
NodeJS
Articles
Scaffolding and Middleware in ExpressJS
Overview of expressJS, installation and Request-response model
NodeJS environment setup and Creating First Application
How to scale application in NodeJS and concept of packaging
Event Driven Programming in NodeJS
Cookies Management, Routing and Template Engine in ExpressJS
eBooks
Interview Questions
Videos
Scaffolding and Middleware in ExpressJS
Overview of expressJS, installation and Request-response model
NodeJS environment setup and Creating First Application
How to scale application in NodeJS and concept of packaging
Event Driven Programming in NodeJS
Cookies Management, Routing and Template Engine in ExpressJS
PHP
Articles
Variable Types and Constant Types in PHP
Operations in MySQL DB using PHP
Object Oriented Programming in PHP
Login with Facebook and Paypal Integration in PHP
Dynamic Content Handling in PHP
How to Install PHP on your system
How to access information from DB using PHP and AJAX
Error and Exception Handling in PHP
CRUD operations in MySQL DB using PHP
eBooks
Interview Questions
Videos
Variable Types and Constant Types in PHP
Operations in MySQL DB using PHP
Object Oriented Programming in PHP
Login with Facebook and Paypal Integration in PHP
Dynamic Content Handling in PHP
How to Install PHP on your system
How to access information from DB using PHP and AJAX
Error and Exception Handling in PHP
CRUD operations in MySQL DB using PHP
Python
Articles
Variable Types and Basic Operators in Python
Time Series, Geographical and Graph Data in Python
Sending Email using SMTP in Python
Processing CSV, JSON and XLS Data in Python
MySQL Database Access in Python
Multithreaded Programming in Python
Introduction to Python and Installing Python
How to draw different Charts in Python
Handling Relational and NoSQL Databases in Python
Handling Date and Time in Python
Extension Programming with C in Python
Data Wrangling and Data Aggregations in Python
Data Science Libraries in Python
eBooks
Interview Questions
Videos
Variable Types and Basic Operators in Python
Time Series, Geographical and Graph Data in Python
Sending Email using SMTP in Python
Processing CSV, JSON and XLS Data in Python
MySQL Database Access in Python
Multithreaded Programming in Python
Introduction to Python and Installing Python
How to draw different Charts in Python
Handling Relational and NoSQL Databases in Python
Handling Date and Time in Python
Extension Programming with C in Python
Data Wrangling and Data Aggregations in Python
Data Science Libraries in Python
Servlet
eBooks
Interview Questions
Videos
Spring Boot
Articles
How to write a Scheduler on the Spring applications and CORS Support
Service Components in Spring Boot
Tracing Micro Service Logs in Spring Boot
How to perform Bootstrapping on a Spring Boot application
How to use Spring Boot JDBC driver connection to connect the database
How to write a unit test case by using Mockito and Web Controller
Spring Boot - Code Structure and Build Systems
Spring Boot - Enabling Swagger2
Spring Boot - Google Cloud Platform
Spring Boot - Rest Controller Unit Test
Spring Boot - Securing Web Applications
Spring Boot - Tomcat Deployment
Spring Boot Architecture and Why Spring Boot is used
Spring Boot Security mechanisms and OAuth2 with JWT
Spring Vs Spring Boot Vs Spring MVC
Application Properties in Spring Boot
How to implement the SMS sending and making voice calls by using Spring Boot with Twilio
Building RESTful Web Services using Spring Boot
Consuming RESTful Web Services by using jQuery AJAX
Create a Web Application in Spring Boot using Thymeleaf
Creating Servlet Filter using Spring Boot
Exception Handling in Spring Boot
File Handling using Spring Boot
How to add the Google OAuth2 Sign-In by using Spring Boot application with Gradle build
How to build a Eureka Server using Spring Boot
How to build an interactive web application by using Spring Boot with Web sockets
How to Build Spring Boot Admin Server and Client
How to Create Applications that consume Restful Web Services
How to Create Spring Cloud Configuration Server
How to Configure Flyway Database in your Spring Boot application
How to create a Docker Image using Maven and Gradle
How to create a Spring Boot Application using Maven and Gradle
How to create Zuul Proxy Server application in Spring Boot
How to implement the Apache Kafka in Spring Boot application
eBooks
Interview Questions
Videos
How to write a Scheduler on the Spring applications and CORS Support
Service Components in Spring Boot
Tracing Micro Service Logs in Spring Boot
How to perform Bootstrapping on a Spring Boot application
How to use Spring Boot JDBC driver connection to connect the database
How to write a unit test case by using Mockito and Web Controller
Spring Boot - Code Structure and Build Systems
Spring Boot - Enabling Swagger2
Spring Boot - Google Cloud Platform
Spring Boot - Rest Controller Unit Test
Spring Boot - Securing Web Applications
Spring Boot - Tomcat Deployment
Spring Boot Architecture and Why Spring Boot is used
Spring Boot Security mechanisms and OAuth2 with JWT
Spring Vs Spring Boot Vs Spring MVC
Application Properties in Spring Boot
How to implement the SMS sending and making voice calls by using Spring Boot with Twilio
Building RESTful Web Services using Spring Boot
Consuming RESTful Web Services by using jQuery AJAX
Create a Web Application in Spring Boot using Thymeleaf
Creating Servlet Filter using Spring Boot
Exception Handling in Spring Boot
File Handling using Spring Boot
How to add the Google OAuth2 Sign-In by using Spring Boot application with Gradle build
How to build a Eureka Server using Spring Boot
How to build an interactive web application by using Spring Boot with Web sockets
How to Build Spring Boot Admin Server and Client
How to Create Applications that consume Restful Web Services
How to Create Spring Cloud Configuration Server
How to Configure Flyway Database in your Spring Boot application
How to create a Docker Image using Maven and Gradle
How to create a Spring Boot Application using Maven and Gradle
How to create Zuul Proxy Server application in Spring Boot
How to implement the Apache Kafka in Spring Boot application
Variable Types and Basic Operators in Python
Time Series, Geographical and Graph Data in Python
Sending Email using SMTP in Python
Processing CSV, JSON and XLS Data in Python
MySQL Database Access in Python
Multithreaded Programming in Python
Introduction to Python and Installing Python
How to draw different Charts in Python
Handling Relational and NoSQL Databases in Python
Handling Date and Time in Python
Extension Programming with C in Python
Data Wrangling and Data Aggregations in Python
Data Science Libraries in Python
Calendar and Date and Time in Python
Scaffolding and Middleware in ExpressJS
Overview of expressJS, installation and Request-response model
NodeJS environment setup and Creating First Application
How to use Node Package Manager and REPL Terminal
How to scale application in NodeJS and concept of packaging
Handling GET and POST Request in NodeJS
Event Driven Programming in NodeJS
Cookies Management, Routing and Template Engine in ExpressJS
Concept of Callbacks and Streams in NodeJS
Comparison of NodeJS with other programming languages
Using Sessions and POJO Classes in Hibernate
Transaction Management in Spring
Overview and Architecture of Spring Framework
ORM Overview and Overview of Hibernate
IoC Containers, AOP and JDBC Framework in Spring
Architecture of Magento 2 and Product Overview
Injecting Inner Beans and Collections in Spring
How to use Criteria Queries in Hibernate
How to perform Java Based Configuration in Spring
How to Install Hibernate and its Configuration
Environment Setup for Spring Framework
Variables and Keywords in Java
Transaction Management and Batch Processing in JDBC
StringBuffer and StringBuilder Class in Java
String Vs StringBuffer Vs StringBuilder
Stream API Improvement in Java 9
Static Binding and Dynamic Binding and Final Keyword
Serialization and Reflection in Java
Properties class and Generics in Java
Method Parameter Reflection in Java
Java StringJoiner and ArrayList Vs Vector
Java Queue and Deque Interface
Java Parallel Array Sorting and Type Inference
Java Networking and Socket Programming
Java Nested Interface and Method Overloading and Overriding
Java Method References and Functional Interfaces
Java Garbage Collection and Java Runtime Class
Java forEach loop and Collectors
Java Comments and Naming Conventions
Java 9 Process API Improvement
Java 9 Module System and Control Panel
Java 9 Anonymous Inner Classes Improvement and SafeVarargs Annotation
Introduction to Java and History of Java
Inter-thread communication and Deadlock in Java
How to write the Hello World Java program
How to create Immutable class in Java
Features of Java and C++ Vs Java
ExceptionHandling with MethodOverriding in Java
Difference between JDK, JRE, and JVM
Deep Dive into Threads in Java
Deep Dive into LinkedList in Java
Deep dive into LinkedHashMap and TreeMap
Deep dive into HashSet , LinkedHashSet and TreeSet
Deep Dive into HashMap in Java
Deep Dive into ArrayList in Java
Conditional Statements in Java
Concept of Method Overloading and Method Overriding in Java
Concept of Inheritance and Aggregation in Java
Comparable and Comparator interface in Java
Call by Value and Call by Reference in Java
Cookies in Laravel based web applications
Encryption and Hashing in Laravel
How to create Blade Templates Layout
How to Create Façade in Laravel
How to perform Redirections and connect to Database
Installation Process of Laravel
Introduction to Laravel and its History
Laravel vs CodeIgniter and Laravel Vs Symphony
Laravel vs Django and Laravel vs WordPress
Middleware Mechanism in Laravel
Process of Authentication and Authorization in Laravel
Responses in Laravel web applications
Understanding Release Process in Laravel
How to setup Check/Money Order payment method in Magento 2
Dynamic Content Handling in PHP
Object Oriented Programming in PHP
How to use the multi language feature of Magento
How to Setup System Theme, Page Title, Layout and New Pages in Magento
How to Setup Shipping Rates and Payment Plans in Magento
How to setup shipping methods in Magento 2
How to Setup Paypal Payment and Google checkout in Magento
How to Setup Newsletter in Magento
How to Setup Google Analytics Youtube Videos and Facebook Likes in Magento
How to setup Check/Money Order payment method in Magento 2
How to set up Zero Subtotal Checkout payment method in Magento 2
How to set up the tax rules, tax rates, and tax zones in Magento 2
How to set up Purchase Order (PO) payment method in Magento 2
How to set up Order Emails in Magento 2
How to set up multiple websites, stores, and store views in Magento 2
How to Set up Contact, Categories, Products and Inventory in Magento
How to set up Cash on Delivery (COD) payment method in Magento 2
How to set up Bank Transfer payment method in Magento 2
How to set up Authorize.net method in Magento 2
How to Manage Tax Classes in Magento
How to Install Magento on your system
How to install Magento 2 using Composer
How to install Magento 2 on windows
Ways for Site Optimization in Magento
Store Configuration in Magento 2
Search Engine Optimization in Magento 2
Products and their Types in Magento 2
Overview of Magento and its Features
Orders Life Cycle in Magento 2
Ways for Site Optimization in Magento
Basic Configuration in Magento 2
Create and Manage CMS (Content Management System) in Magento 2
How to add the product on Home page in Magento 2
How to configure and Manage the Inventory in Magento 2
How to create Attribute Sets in Magento 2
How to create Product Attributes in Magento 2
How to create Product Category in Magento 2
How to generate Order Report in Magento 2
How to create Product in Magento 2
Variable Types and Constant Types in PHP
Operations in MySQL DB using PHP
Object Oriented Programming in PHP
Login with Facebook and Paypal Integration in PHP
Dynamic Content Handling in PHP
How to Install PHP on your system
How to access information from DB using PHP and AJAX
Error and Exception Handling in PHP
CRUD operations in MySQL DB using PHP
Handling Arrays and Strings in PHP
Standard Tag Library (JSTL) in JSP
Page Redirecting and Hits Counter and Auto Refresh
Overview of Java Server Pages and its Architecture
How to Access Database with JSP
Servlets - Server HTTP Response
Servlets - Page Redirection and Auto Refresh
Internationalization in Servlets
Handling Date and Time using Servlets
Exception Handling in Servlets
Overview of Servlets and setup of Environment
How to write a Scheduler on the Spring applications and CORS Support
Service Components in Spring Boot
Tracing Micro Service Logs in Spring Boot
How to perform Bootstrapping on a Spring Boot application
How to use Spring Boot JDBC driver connection to connect the database
How to write a unit test case by using Mockito and Web Controller
Spring Boot - Code Structure and Build Systems
Spring Boot - Enabling Swagger2
Spring Boot - Google Cloud Platform
Spring Boot - Rest Controller Unit Test
Spring Boot - Securing Web Applications
Spring Boot - Tomcat Deployment
Spring Boot Architecture and Why Spring Boot is used
Spring Boot Security mechanisms and OAuth2 with JWT
Spring Vs Spring Boot Vs Spring MVC
Application Properties in Spring Boot
How to implement the SMS sending and making voice calls by using Spring Boot with Twilio
Building RESTful Web Services using Spring Boot
Consuming RESTful Web Services by using jQuery AJAX
Create a Web Application in Spring Boot using Thymeleaf
Creating Servlet Filter using Spring Boot
Exception Handling in Spring Boot
File Handling using Spring Boot
How to add the Google OAuth2 Sign-In by using Spring Boot application with Gradle build
How to build a Eureka Server using Spring Boot
How to build an interactive web application by using Spring Boot with Web sockets
How to Build Spring Boot Admin Server and Client
How to Create Applications that consume Restful Web Services
How to Create Spring Cloud Configuration Server
How to Configure Flyway Database in your Spring Boot application
How to create a Docker Image using Maven and Gradle
How to create a Spring Boot Application using Maven and Gradle
How to create Zuul Proxy Server application in Spring Boot
How to implement the Apache Kafka in Spring Boot application
How to implement the Internationalization in Spring Boot
How to implement the Hystrix in a Spring Boot application
Login with Facebook and Paypal Integration in PHP
Understanding Release Process in Laravel
Responses in Laravel web applications
Process of Authentication and Authorization in Laravel
Middleware Mechanism in Laravel
Laravel vs Django and Laravel vs WordPress
Laravel vs CodeIgniter and Laravel Vs Symphony
Introduction to Laravel and its History
Installation Process of Laravel
How to perform Redirections and connect to Database
How to Create Façade in Laravel
How to create Blade Templates Layout
Encryption and Hashing in Laravel
Cookies in Laravel based web applications
Contracts and CSRF Protection in Laravel
Available Validation Rules of Laravel
Artisan Console for interaction in Laravel
Application Structure of Laravel
Expression Language (EL) in JSP
Expression Language (EL) in JSP
Project Management and Methodologies
Articles
eBooks
Interview Questions
Videos
Robotic Process Automation
eBooks
Interview Questions
Videos
RPA-UiPath
Articles
eBooks
Interview Questions
Videos
Working of RPA and its Services
Understanding User Interface Components
UiPath Studio - Workflow Design
RPA Use Cases and Applications
RPA Life Cycle and Implementation
Recording using UiPath in Detail
Keyboard Shortcuts and Customization in UiPath Studio
Key Basics of UiPath and the related concepts
Installation of UiPath on your local system
How to work with Automation Projects in UiPath and their Debugging methods
How to deal and work with variables and arguments in UiPath
Data Scraping and Screen Scraping in UiPath
Comparison of RPA and AI, Test Automation and Traditional Automation
Architecture and Components of RPA
Advantages and drawbacks of RPA
Top Robotic Process Automation (RPA) with UiPath Interview Questions and Answers
Salesforce
Articles
Different Levels of Data Access in Salesforce
Variables & Formulas in Salesforce
Using Records, Fields and Tables in Salesforce
Using Forms and List Controllers in Salesforce
Creating Static Resources in Salesforce
Standard and Custom Objects in Salesforce platform
Overview of Salesforce and its architecture
Master Detail Relationship in Salesforce
Lookup Relationship in Salesforce
How to Import Data in Salesforce
How to Export Data from Salesforce
How to Define Sharing Rules in Salesforce
How to create Visual force Pages in Salesforce
How to create Reports and Dashboards in Salesforce
Get Started with Salesforce - Environment
How to Create a Role Hierarchy in Salesforce
eBooks
Interview Questions
Videos
Apex Programming
Articles
Classes and Methods in Apex programming language
Concept of Objects and Interfaces in Apex programming language
Database Methods and process of executing the Apex class in Salesforce
Deployment in Salesforce using Sandbox
Enterprise Application Development Example
How to Perform Debugging in Apex
How to perform the various Database Modification Functionalities in Salesforce
How to perform Unit Testing in Apex
Overview of Apex Programming and its environment
Search Functionality using SOSL and SOQL
Understand Batch Processing in Salesforce Apex
Understanding deciding, Loops and Collections in Apex
Understanding Governor Limits in Salesforce Apex
Understanding the info Types and variables in Apex programming language
Understanding the environment for Salesforce Apex development
Understanding the String Manipulation, Arrays and Constants in Apex programming language
eBooks
Interview Questions
Videos
Classes and Methods in Apex programming language
Concept of Objects and Interfaces in Apex programming language
Database Methods and process of executing the Apex class in Salesforce
Deployment in Salesforce using Sandbox
Enterprise Application Development Example
How to Perform Debugging in Apex
How to perform the various Database Modification Functionalities in Salesforce
How to perform Unit Testing in Apex
Overview of Apex Programming and its environment
Search Functionality using SOSL and SOQL
Understand Batch Processing in Salesforce Apex
Understanding deciding, Loops and Collections in Apex
Understanding Governor Limits in Salesforce Apex
Understanding the info Types and variables in Apex programming language
Understanding the environment for Salesforce Apex development
Understanding the String Manipulation, Arrays and Constants in Apex programming language
Different Levels of Data Access in Salesforce
Variables & Formulas in Salesforce
Using Records, Fields and Tables in Salesforce
Using Forms and List Controllers in Salesforce
Creating Static Resources in Salesforce
Standard and Custom Objects in Salesforce platform
Overview of Salesforce and its architecture
Master Detail Relationship in Salesforce
Lookup Relationship in Salesforce
How to Import Data in Salesforce
How to Export Data from Salesforce
How to Define Sharing Rules in Salesforce
How to create Visual force Pages in Salesforce
How to create Reports and Dashboards in Salesforce
Get Started with Salesforce - Environment
How to Create a Role Hierarchy in Salesforce
Classes and Methods in Apex programming language
Concept of Objects and Interfaces in Apex programming language
Database Methods and process of executing the Apex class in Salesforce
Deployment in Salesforce using Sandbox
Enterprise Application Development Example
How to Perform Debugging in Apex
How to perform the various Database Modification Functionalities in Salesforce
How to perform Unit Testing in Apex
Overview of Apex Programming and its environment
Search Functionality using SOSL and SOQL
Understand Batch Processing in Salesforce Apex
Understanding deciding, Loops and Collections in Apex
Understanding Governor Limits in Salesforce Apex
Understanding the info Types and variables in Apex programming language
Understanding the environment for Salesforce Apex development
Understanding the String Manipulation, Arrays and Constants in Apex programming language
Using Formula Fields in Salesforce
SAP
Articles
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
eBooks
Interview Questions
Videos
SAP Business Object
Articles
Using Filters in SAP BO Analysis
Sheets and Sharing Workspaces in SAP BO Analysis
Perform Conditional Formatting in SAP BO Analysis
Overview of SAP Business Object Analysis
How to create a Workspace in SAP Business Objects
How to Connect to SAP BW in SAP Business Objects
Export Options in SAP BO Analysis
Concept of Sub Analysis in SAP BO
eBooks
Interview Questions
Videos
Using Filters in SAP BO Analysis
Sheets and Sharing Workspaces in SAP BO Analysis
Perform Conditional Formatting in SAP BO Analysis
Overview of SAP Business Object Analysis
How to create a Workspace in SAP Business Objects
How to Connect to SAP BW in SAP Business Objects
Export Options in SAP BO Analysis
Concept of Sub Analysis in SAP BO
Calculations in SAP BO Analysis
SAP Hana
Articles
Alert Monitoring and Logging in SAP Hana
Authentications and Authorization Methods in SAP HANA
DXC Replication Method and CTL Method and MDX provider in SAP Hana
Excel Integration with SAP Hana and Bi 4.0 Connectivity to Hana Views
User Administration & Role Management and Security Overview in SAP Hana
Usage of SQL Script in SAP Hana
SQL Triggers, Synonym and Data Profiling in SAP Hana
SQL Overview and Data Types in SAP Hana
SQL Functions and Operators in SAP Hana
SQL Expressions, Stored Procedures and Sequences in SAP Hana
Packages and Attribute and Analytic View in SAP Hana
Modeling and Schemas in SAP HANA
Log Based and ETL Based Replication in SAP Hana
License Management and Auditing in SAP Hana
Information Modeler and System Monitor in SAP HANA
High Availability and Backup and Recovery in SAP Hana
Export and Import Options in Sap Hana
eBooks
Videos
Alert Monitoring and Logging in SAP Hana
Authentications and Authorization Methods in SAP HANA
DXC Replication Method and CTL Method and MDX provider in SAP Hana
Excel Integration with SAP Hana and Bi 4.0 Connectivity to Hana Views
User Administration & Role Management and Security Overview in SAP Hana
Usage of SQL Script in SAP Hana
SQL Triggers, Synonym and Data Profiling in SAP Hana
SQL Overview and Data Types in SAP Hana
SQL Functions and Operators in SAP Hana
SQL Expressions, Stored Procedures and Sequences in SAP Hana
Packages and Attribute and Analytic View in SAP Hana
Modeling and Schemas in SAP HANA
Log Based and ETL Based Replication in SAP Hana
License Management and Auditing in SAP Hana
Information Modeler and System Monitor in SAP HANA
High Availability and Backup and Recovery in SAP Hana
Export and Import Options in Sap Hana
Data Replication Overview in SAP Hana
Analytic Privileges and Information Composer in SAP Hana
SAP Hana Adminstration
Articles
SAP HANA Admin Studio and System Management
Overview of SAP HANA Administration
SAP HANA License Management and Multitenant DB Container Management
Smart Data Access and Integration with Hadoop
How to Start, Stop and Monitor a HANA System
HANA XS Application Service and Data Provisioning in SAP Hana
Data Compression and Solman Integration in SAP Hana
eBooks
Interview Questions
Videos
SAP HANA Admin Studio and System Management
Overview of SAP HANA Administration
SAP HANA License Management and Multitenant DB Container Management
Smart Data Access and Integration with Hadoop
How to Start, Stop and Monitor a HANA System
HANA XS Application Service and Data Provisioning in SAP Hana
Data Compression and Solman Integration in SAP Hana
SAP Hana Finance
Articles
Profitability Analysis and Management Accounting in SAP Simple Finance
Overview of SAP Hana and SAP Hana Finance
Migration and Manual Reposting of Costs in SAP Simple Finance
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Ledger Management in SAP Simple Finance
Reporting Options and G/L Accounting in SAP Simple Finance
Universal Journal and Document Number in SAP Simple Finance
SAP Simple Finance Architecture and Deployment Options
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
eBooks
Interview Questions
Videos
Profitability Analysis and Management Accounting in SAP Simple Finance
Overview of SAP Hana and SAP Hana Finance
Migration and Manual Reposting of Costs in SAP Simple Finance
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Ledger Management in SAP Simple Finance
Reporting Options and G/L Accounting in SAP Simple Finance
Universal Journal and Document Number in SAP Simple Finance
SAP Simple Finance Architecture and Deployment Options
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
SAP Hana Logistics
Articles
Supply Chain Planning and Integrated Business Planning in SAP Hana Logistics
Overview of SAP Hana Simple Logistics
MRP Procedures and Key Features in SAP Simple Logistics
MIGO Transactions in SAP Simple Logistics
Manufacturing Process in SAP Simple Logistics
Invoice Management and Operational Procurement in SAP Simple Logistics
How to Manage Business Partner in SAP Simple Logistics
How to Execute MRP Live planning
How to Create Business Partner in SAP HANA Logistics
Fiori UX and Deployment and Procurement Types in SAP Hana Logistics
Execute Discrete Production in SAP Hana Logistics
Contract Management and Perform Procurement Transfer Stock in SAP Hana Logistics
eBooks
Interview Questions
Videos
Supply Chain Planning and Integrated Business Planning in SAP Hana Logistics
Overview of SAP Hana Simple Logistics
MRP Procedures and Key Features in SAP Simple Logistics
MIGO Transactions in SAP Simple Logistics
Manufacturing Process in SAP Simple Logistics
Invoice Management and Operational Procurement in SAP Simple Logistics
How to Manage Business Partner in SAP Simple Logistics
How to Execute MRP Live planning
How to Create Business Partner in SAP HANA Logistics
Fiori UX and Deployment and Procurement Types in SAP Hana Logistics
Execute Discrete Production in SAP Hana Logistics
Contract Management and Perform Procurement Transfer Stock in SAP Hana Logistics
Concept of Simplification Item in SAP Simple Logistics
SAP UDT & IDT
Articles
Building Data Foundation in SAP IDT
Building Query in Query Panel, Publishing in SAP IDT
Business Layer Properties in SAP IDT
Dealing with Published Universes in SAP IDT
Deploying Universe in SAP Universe Designer
Format Editor Overview in SAP IDT
How to create universe in SAP IDT
How to use Table Browser and Derived Tables in SAP Universal Designer
Joins In Data Foundation in SAP IDT
Managing Connections in SAP IDT
Managing Resources in Repository, Qualifiers and Owners
OLAP Data Sources in SAP Universe Designer
Overview of SAP Universe Designer
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
Concept of Calculated Measures and Aggregate Awareness
Business Layer View in SAP IDT
eBooks
Interview Questions
Videos
Building Data Foundation in SAP IDT
Building Query in Query Panel, Publishing in SAP IDT
Business Layer Properties in SAP IDT
Dealing with Published Universes in SAP IDT
Deploying Universe in SAP Universe Designer
Format Editor Overview in SAP IDT
How to create universe in SAP IDT
How to use Table Browser and Derived Tables in SAP Universal Designer
Joins In Data Foundation in SAP IDT
Managing Connections in SAP IDT
Managing Resources in Repository, Qualifiers and Owners
OLAP Data Sources in SAP Universe Designer
Overview of SAP Universe Designer
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
Concept of Calculated Measures and Aggregate Awareness
Business Layer View in SAP IDT
Sap Webi
Articles
Working with Reports in SAP Webi
Sending Documents in SAP Web Intelligence
Query Filters and Filters Type in SAP Webi
Queries using Bex and Analysis View in SAP Webi
How to use Formulas and Variables in SAP Webi
How to use Breaks, Sorts and Ranking Data in SAP Webi
How to Create SAP Webi documents
How to achieve Conditional Formatting in SAP Webi
eBooks
Interview Questions
Videos
Working with Reports in SAP Webi
Sending Documents in SAP Web Intelligence
Query Filters and Filters Type in SAP Webi
Queries using Bex and Analysis View in SAP Webi
How to use Formulas and Variables in SAP Webi
How to use Breaks, Sorts and Ranking Data in SAP Webi
How to Create SAP Webi documents
How to achieve Conditional Formatting in SAP Webi
SAP HANA Admin Studio and System Management
Overview of SAP HANA Administration
SAP HANA License Management and Multitenant DB Container Management
Smart Data Access and Integration with Hadoop
Building Data Foundation in SAP IDT
Building Query in Query Panel, Publishing in SAP IDT
Business Layer Properties in SAP IDT
Dealing with Published Universes in SAP IDT
Deploying Universe in SAP Universe Designer
Format Editor Overview in SAP IDT
How to create universe in SAP IDT
How to use Table Browser and Derived Tables in SAP Universal Designer
Joins In Data Foundation in SAP IDT
Managing Connections in SAP IDT
Managing Resources in Repository, Qualifiers and Owners
OLAP Data Sources in SAP Universe Designer
Overview of SAP Universe Designer
unv Universe in SAP Business Object
Using Formula Bar and Universe Operations in SAP Universe Designer
Using LOVs and Create, Edit and Save a Universe
Concept of Calculated Measures and Aggregate Awareness
Business Layer View in SAP IDT
Profitability Analysis and Management Accounting in SAP Simple Finance
Overview of SAP Hana and SAP Hana Finance
Migration and Manual Reposting of Costs in SAP Simple Finance
How to Display Financial Tables in SAP Simple Finance
Concept of Period Lock Transaction in SAP Simple Finance
Concept of Asset Scrapping in SAP Simple Finance
Create Default Account Assignment in SAP Simple Finance
How to Create a Primary Cost in G-L Account
Ledger Management in SAP Simple Finance
Reporting Options and G/L Accounting in SAP Simple Finance
Universal Journal and Document Number in SAP Simple Finance
SAP Simple Finance Architecture and Deployment Options
Alert Monitoring and Logging in SAP Hana
Authentications and Authorization Methods in SAP HANA
DXC Replication Method and CTL Method and MDX provider in SAP Hana
Excel Integration with SAP Hana and Bi 4.0 Connectivity to Hana Views
Working with Reports in SAP Webi
Sending Documents in SAP Web Intelligence
Query Filters and Filters Type in SAP Webi
Queries using Bex and Analysis View in SAP Webi
How to use Formulas and Variables in SAP Webi
How to use Breaks, Sorts and Ranking Data in SAP Webi
How to Create SAP Webi documents
How to achieve Conditional Formatting in SAP Webi
Filtering Report Data in SAP Webi
Drill Options in Reports and Sharing Reports in SAP Webi
Supply Chain Planning and Integrated Business Planning in SAP Hana Logistics
Overview of SAP Hana Simple Logistics
MRP Procedures and Key Features in SAP Simple Logistics
MIGO Transactions in SAP Simple Logistics
Manufacturing Process in SAP Simple Logistics
Invoice Management and Operational Procurement in SAP Simple Logistics
How to Manage Business Partner in SAP Simple Logistics
How to Execute MRP Live planning
How to Create Business Partner in SAP HANA Logistics
Fiori UX and Deployment and Procurement Types in SAP Hana Logistics
Execute Discrete Production in SAP Hana Logistics
Contract Management and Perform Procurement Transfer Stock in SAP Hana Logistics
Concept of Simplification Item in SAP Simple Logistics
Analyze Sales Orders in SAP Simple Logistics
User Administration & Role Management and Security Overview in SAP Hana
Usage of SQL Script in SAP Hana
SQL Triggers, Synonym and Data Profiling in SAP Hana
SQL Overview and Data Types in SAP Hana
SQL Functions and Operators in SAP Hana
SQL Expressions, Stored Procedures and Sequences in SAP Hana
Packages and Attribute and Analytic View in SAP Hana
Modeling and Schemas in SAP HANA
Log Based and ETL Based Replication in SAP Hana
License Management and Auditing in SAP Hana
Information Modeler and System Monitor in SAP HANA
High Availability and Backup and Recovery in SAP Hana
Export and Import Options in Sap Hana
Data Replication Overview in SAP Hana
Analytic Privileges and Information Composer in SAP Hana
Using Filters in SAP BO Analysis
Sheets and Sharing Workspaces in SAP BO Analysis
Perform Conditional Formatting in SAP BO Analysis
Overview of SAP Business Object Analysis
How to create a Workspace in SAP Business Objects
Asset Accounting in SAP Simple Finance
Concept of Integrated Business Planning and Integration of Simple Finance with other Modules
How to Connect to SAP BW in SAP Business Objects
Export Options in SAP BO Analysis
Concept of Sub Analysis in SAP BO
Calculations in SAP BO Analysis
Aggregations and Hierarchies in SAP BO Analysis
SAP IDT - Overview and User Interface
Creating Parameters and Schemas in SAP Universe Designer
How to Start, Stop and Monitor a HANA System
HANA XS Application Service and Data Provisioning in SAP Hana
Data Compression and Solman Integration in SAP Hana
Authentication Methods supported by SAP HANA
Auditing Activities in SAP Hana
Top SAP S4 HANA Logistics Interview Questions and Answers
Top SAP S4 HANA Finance Interview Questions and Answers
Top SAP HANA Interview Questions and Answers
Software Testing
Articles
eBooks
Interview Questions
Videos
Selenium WebDriver
Articles
How to run your Selenium Test Scripts on IE Browser
How to run your Selenium Test Scripts on Firefox Browser
Comparison of Selenium vs QTP and Selenium Tool Suite
How to run your Selenium Test Scripts on Safari Browser
Overview of Selenium WebDriver
Overview of Selenium, its features and limitations
Scrolling an internet page in Selenium WebDriver
Selenium IDE- Locating Strategies by Identifier and By Id
Selenium IDE- Locating Strategies by Name, XPath , CSS and DOM
How to run your Selenium Test Scripts on Chrome Browser
How to Handle Alerts in Selenium WebDriver
Selenium WebDriver - Navigation and Web Element Commands
How to handle radio buttons and checkbox in selenium web driver
Selenium WebDriver - Browser Commands
Selenium WebDriver- Locating Strategies and Handling Drop-downs
Comparison between Selenium WebDriver and Selenium RC
Creating Test Cases Manually in Selenium IDE
How to create Login test suit in Selenium IDE
How to create your First Selenium Automation Test Script
Selenium IDE- Commands (Selenese)
Using Assertions in Selenium WebDriver
Overview of Selenium Integrated Development Environment (IDE)
eBooks
Interview Questions
Videos
How to run your Selenium Test Scripts on IE Browser
How to run your Selenium Test Scripts on Firefox Browser
Comparison of Selenium vs QTP and Selenium Tool Suite
How to run your Selenium Test Scripts on Safari Browser
Overview of Selenium WebDriver
Overview of Selenium, its features and limitations
Scrolling an internet page in Selenium WebDriver
Selenium IDE- Locating Strategies by Identifier and By Id
Selenium IDE- Locating Strategies by Name, XPath , CSS and DOM
How to run your Selenium Test Scripts on Chrome Browser
How to Handle Alerts in Selenium WebDriver
Selenium WebDriver - Navigation and Web Element Commands
How to handle radio buttons and checkbox in selenium web driver
Selenium WebDriver - Browser Commands
Selenium WebDriver- Locating Strategies and Handling Drop-downs
Comparison between Selenium WebDriver and Selenium RC
Creating Test Cases Manually in Selenium IDE
How to create Login test suit in Selenium IDE
How to create your First Selenium Automation Test Script
Selenium IDE- Commands (Selenese)
Using Assertions in Selenium WebDriver
Overview of Selenium Integrated Development Environment (IDE)
Selenium with Maven
Articles
Execute Selenium code through Maven and TestNG
How to Configure Selenium using NUnit in Visual Studio
How to Configure Selenium with Visual Studio in C#
How to handle or download dependency Jar using Maven
Write a Selenium test script using C#
Selenium Test Script using NUnit
How to write a Selenium test script using C#
eBooks
Interview Questions
Videos
Execute Selenium code through Maven and TestNG
How to Configure Selenium using NUnit in Visual Studio
How to Configure Selenium with Visual Studio in C#
How to handle or download dependency Jar using Maven
Write a Selenium test script using C#
Selenium Test Script using NUnit
How to write a Selenium test script using C#
Test NG
Articles
How to Run test cases in TestNG without java compiler
Overview of TestNG and its Features
Importance of XML file in TestNG Configuration
How to use TestNG Annotation Attributes
How to Run test cases with Regex in TestNG
How to install TestNG Framework and Configuration in Eclipse
How to enable and disable test cases in TestNG
eBooks
Interview Questions
Videos
How to Run test cases in TestNG without java compiler
Overview of TestNG and its Features
Importance of XML file in TestNG Configuration
How to use TestNG Annotation Attributes
How to Run test cases with Regex in TestNG
How to install TestNG Framework and Configuration in Eclipse
How to enable and disable test cases in TestNG
How to run your Selenium Test Scripts on IE Browser
How to run your Selenium Test Scripts on Firefox Browser
Comparison of Selenium vs QTP and Selenium Tool Suite
How to run your Selenium Test Scripts on Safari Browser
Overview of Selenium WebDriver
Overview of Selenium, its features and limitations
Scrolling an internet page in Selenium WebDriver
Selenium IDE- Locating Strategies by Identifier and By Id
Selenium IDE- Locating Strategies by Name, XPath , CSS and DOM
How to run your Selenium Test Scripts on Chrome Browser
How to Handle Alerts in Selenium WebDriver
Selenium WebDriver - Navigation and Web Element Commands
How to handle radio buttons and checkbox in selenium web driver
Selenium WebDriver - Browser Commands
Selenium WebDriver- Locating Strategies and Handling Drop-downs
Comparison between Selenium WebDriver and Selenium RC
Creating Test Cases Manually in Selenium IDE
How to create Login test suit in Selenium IDE
How to create your First Selenium Automation Test Script
Execute Selenium code through Maven and TestNG
How to Configure Selenium using NUnit in Visual Studio
How to Configure Selenium with Visual Studio in C#
How to handle or download dependency Jar using Maven
Write a Selenium test script using C#
Selenium Test Script using NUnit
How to write a Selenium test script using C#
Write and Execute the Selenium test script
Using Maven with Selenium TestNG
Selenium IDE- Commands (Selenese)
Using Assertions in Selenium WebDriver
Overview of Selenium Integrated Development Environment (IDE)
How to Run test cases in TestNG without java compiler
Overview of TestNG and its Features
Importance of XML file in TestNG Configuration
How to use TestNG Annotation Attributes
How to Run test cases with Regex in TestNG
How to install TestNG Framework and Configuration in Eclipse
How to enable and disable test cases in TestNG
How to create TestNG Listeners
Table of Contents
Top Big Data Hadoop Interview Questions and Answers
Which companies use Hadoop?
Yahoo! (it is the biggest contributor to the creation of Hadoop; its search engine uses Hadoop); Facebook (developed Hive for analysis); Amazon; Netflix; Adobe; eBay; Spotify; Twitter; and Adobe.
What is the use of RecordReader in Hadoop?
Though InputSplit defines a slice of work, it does not describe how to access it. Here is where the RecordReader class comes into the picture, which takes the byte-oriented data from its source and converts it into record-oriented key–value pairs such that it is fit for the Mapper task to read it. Meanwhile, InputFormat defines this Hadoop RecordReader instance.
What is Speculative Execution in Hadoop?
One limitation of Hadoop is that by distributing the tasks on several nodes, there are chances that few slow nodes limit the rest of the program. There are various reasons for the tasks to be slow, which are sometimes not easy to detect. Instead of identifying and fixing the slow-running tasks, Hadoop tries to detect when the task runs slower than expected and then launches other equivalent tasks as backup. This backup mechanism in Hadoop is speculative execution.
It creates a duplicate task on another disk. The same input can be processed multiple times in parallel. When most tasks in a job comes to completion, the speculative execution mechanism schedules duplicate copies of the remaining tasks (which are slower) across the nodes that are free currently. When these tasks are finished, it is intimated to the JobTracker. If other copies are executing speculatively, Hadoop notifies the TaskTrackers to quit those tasks and reject their output.
Speculative execution is by default true in Hadoop. To disable it, we can set mapred.map.tasks.speculative.execution and mapred.reduce.tasks.speculative.execution
JobConf options to false.
What is the different vendor-specific distributions of Hadoop?
The different vendor-specific distributions of Hadoop are Cloudera, MAPR, Amazon EMR, Microsoft Azure, IBM InfoSphere, and Hortonworks (Cloudera).
What are the different Hadoop configuration files?
The different Hadoop configuration files include:
- hadoop-env.sh
- mapred-site.xml
- core-site.xml
- yarn-site.xml
- hdfs-site.xml
- Master and Slaves
What are the three modes in which Hadoop can run?
The three modes in which Hadoop can run are:
- Standalone mode: This is the default mode. It uses the local File System and a single Java process to run the Hadoop services.
- Pseudo-distributed mode: This uses a single-node Hadoop deployment to execute all Hadoop services.
- Fully-distributed mode: This uses separate nodes to run Hadoop master and slave services.
What are the differences between regular File System and HDFS?
- Regular File System: In regular File System, data is maintained in a single system. If the machine crashes, data recovery is challenging due to low fault tolerance. Seek time is more and hence it takes more time to process the data.
- HDFS: Data is distributed and maintained on multiple systems. If a Data Node crashes, data can still be recovered from other nodes in the cluster. Time taken to read data is comparatively more, as there is local data read to the disc and coordination of data from multiple systems.
What are the two types of metadata that a NameNode server holds?
The two types of metadata that a NameNode server holds are:
- Metadata in Disk – This contains the edit log and the FSImage
- Metadata in RAM – This contains the information about DataNodes
What is the difference between a federation and high availability?
If you have an input file of 350 MB, how many input splits would HDFS create and what would be the size of each input split?
By default, each block in HDFS is divided into 128 MB. The size of all the blocks, except the last block, will be 128 MB. For an input file of 350 MB, there are three input splits in total. The size of each split is 128 MB, 128MB, and 94 MB.
Why is HDFS fault-tolerant?
HDFS is fault-tolerant because it replicates data on different DataNodes. By default, a block of data is replicated on three DataNodes. The data blocks are stored in different DataNodes. If one node crashes, the data can still be retrieved from other DataNodes.
Explain the architecture of HDFS.
The architecture of HDFS is as shown:
For an HDFS service, we have a NameNode that has the master process running on one of the machines and DataNodes, which are the slave nodes.
NameNode
NameNode is the master service that hosts metadata in disk and RAM. It holds information about the various DataNodes, their location, the size of each block, etc.
DataNode
DataNodes hold the actual data blocks and send block reports to the NameNode every 10 seconds. The DataNode stores and retrieves the blocks when the NameNode asks. It reads and writes the client’s request and performs block creation, deletion, and replication based on instructions from the NameNode.
- Data that is written to HDFS is split into blocks, depending on its size. The blocks are randomly distributed across the nodes. With the auto-replication feature, these blocks are auto-replicated across multiple machines with the condition that no two identical blocks can sit on the same machine.
- As soon as the cluster comes up, the DataNodes start sending their heartbeats to the NameNodes every three seconds. The NameNode stores this information; in other words, it starts building metadata in RAM, which contains information about the DataNodes available in the beginning. This metadata is maintained in RAM, as well as in the disk.
How does rack awareness work in HDFS?
HDFS Rack Awareness refers to the knowledge of different DataNodes and how it is distributed across the racks of a Hadoop Cluster.
By default, each block of data is replicated three times on various DataNodes present on different racks. Two identical blocks cannot be placed on the same DataNode. When a cluster is “rack-aware,” all the replicas of a block cannot be placed on the same rack. If a DataNode crashes, you can retrieve the data block from different DataNodes.
How can you restart NameNode and all the daemons in Hadoop?
The following commands will help you restart NameNode and all the daemons:
You can stop the NameNode with ./sbin /Hadoop-daemon.sh stop NameNode command and then start the NameNode using ./sbin/Hadoop-daemon.sh start NameNode command.
You can stop all the daemons with ./sbin /stop-all.sh command and then start the daemons using the ./sbin/start-all.sh command.
Which command will help you find the status of blocks and FileSystem health?
To check the status of the blocks, use the command:
hdfs fsck <path> -files -blocks
To check the health status of FileSystem, use the command:
hdfs fsck / -files –blocks –locations > dfs-fsck.log
What would happen if you store too many small files in a cluster on HDFS?
Storing several small files on HDFS generates a lot of metadata files. To store these metadata in the RAM is a challenge as each file, block, or directory takes 150 bytes for metadata. Thus, the cumulative size of all the metadata will be too large.
How do you copy data from the local system onto HDFS?
The following command will copy data from the local file system onto HDFS:
hadoop fs –copyFromLocal [source] [destination]
Example: hadoop fs –copyFromLocal /tmp/data.csv /user/test/data.csv
In the above syntax, the source is the local path and destination is the HDFS path. Copy from the local system using a -f option (flag option), which allows you to write the same file or a new file to HDFS.
When do you use the dfsadmin -refreshNodes and rmadmin -refreshNodes commands?
The commands below are used to refresh the node information while commissioning, or when the decommissioning of nodes is completed.
dfsadmin -refreshNodes
This is used to run the HDFS client and it refreshes node configuration for the NameNode.
rmadmin -refreshNodes
This is used to perform administrative tasks for ResourceManager.
Is there any way to change the replication of files on HDFS after they are already written to HDFS?
Yes, the following are ways to change the replication of files on HDFS:
We can change the dfs.replication value to a particular number in the $HADOOP_HOME/conf/hadoop-site.xml file, which will start replicating to the factor of that number for any new content that comes in.
If you want to change the replication factor for a particular file or directory, use:
$HADOOP_HOME/bin/Hadoop dfs –setrep –w4 /path of the file
Example: $HADOOP_HOME/bin/Hadoop dfs –setrep –w4 /user/temp/test.csv
Who takes care of replication consistency in a Hadoop cluster and what do under/over replicated blocks mean?
In a cluster, it is always the NameNode that takes care of the replication consistency. The fsck command provides information regarding the over and under-replicated block.
Under-replicated blocks:
These are the blocks that do not meet their target replication for the files they belong to. HDFS will automatically create new replicas of under-replicated blocks until they meet the target replication.
Consider a cluster with three nodes and replication set to three. At any point, if one of the NameNodes crashes, the blocks would be under-replicated. It means that there was a replication factor set, but there are not enough replicas as per the replication factor. If the NameNode does not get information about the replicas, it will wait for a limited amount of time and then start the re-replication of missing blocks from the available nodes.
Over-replicated blocks:
These are the blocks that exceed their target replication for the files they belong to. Usually, over-replication is not a problem, and HDFS will automatically delete excess replicas.
Consider a case of three nodes running with the replication of three, and one of the nodes goes down due to a network failure. Within a few minutes, the NameNode re-replicates the data, and then the failed node is back with its set of blocks. This is an over-replication situation, and the NameNode will delete a set of blocks from one of the nodes.
What is the distributed cache in MapReduce?
A distributed cache is a mechanism wherein the data coming from the disk can be cached and made available for all worker nodes. When a MapReduce program is running, instead of reading the data from the disk every time, it would pick up the data from the distributed cache to benefit the MapReduce processing.
To copy the file to HDFS, you can use the command:
hdfs dfs-put /user/Simplilearn/lib/jar_file.jar
To set up the application’s JobConf, use the command:
DistributedCache.addFileToClasspath(newpath(“/user/Simplilearn/lib/jar_file.jar”), conf)
Then, add it to the driver class.
What role do RecordReader, Combiner, and Partitioner play in a MapReduce operation?
RecordReader
This communicates with the InputSplit and converts the data into key-value pairs suitable for the mapper to read.
Combiner
This is an optional phase; it is like a mini reducer. The combiner receives data from the map tasks, works on it, and then passes its output to the reducer phase.
Partitioner
The partitioner decides how many reduced tasks would be used to summarize the data. It also confirms how outputs from combiners are sent to the reducer, and controls the partitioning of keys of the intermediate map outputs.
Why is MapReduce slower in processing data in comparison to other processing frameworks?
This is quite a common question in Hadoop interviews; let us understand why MapReduce is slower in comparison to the other processing frameworks:
MapReduce is slower because:
- It is batch-oriented when it comes to processing data. Here, no matter what, you would have to provide the mapper and reducer functions to work on data.
- During processing, whenever the mapper function delivers an output, it will be written to HDFS and the underlying disks. This data will be shuffled and sorted, and then be picked up for the reducing phase. The entire process of writing data to HDFS and retrieving it from HDFS makes MapReduce a lengthier process.
- In addition to the above reasons, MapReduce also uses Java language, which is difficult to program as it has multiple lines of code.
Is it possible to change the number of mappers to be created in a MapReduce job?
By default, you cannot change the number of mappers, because it is equal to the number of input splits. However, there are different ways in which you can either set a property or customize the code to change the number of mappers.
For example, if you have a 1GB file that is split into eight blocks (of 128MB each), there will only be only eight mappers running on the cluster. However, there are different ways in which you can either set a property or customize the code to change the number of mappers.
Name some Hadoop-specific data types that are used in a MapReduce program.
This is an important question, as you would need to know the different data types if you are getting into the field of Big Data.
For every data type in Java, you have an equivalent in Hadoop. Therefore, the following are some Hadoop-specific data types that you could use in your MapReduce program:
- IntWritable
- FloatWritable
- LongWritable
- DoubleWritable
- BooleanWritable
- ArrayWritable
- MapWritable
- ObjectWritable
How is identity mapper different from chain mapper?
What are the major configuration parameters required in a MapReduce program?
We need to have the following configuration parameters:
- Input location of the job in HDFS
- Output location of the job in HDFS
- Input and output formats
- Classes containing a map and reduce functions
- JAR file for mapper, reducer and driver classes
What do you mean by map-side join and reduce-side join in MapReduce?
What is the role of the OutputCommitter class in a MapReduce job?
As the name indicates, OutputCommitter describes the commit of task output for a MapReduce job.
Example: org.apache.hadoop.mapreduce.OutputCommitter
public abstract class OutputCommitter extends OutputCommitter
MapReduce relies on the OutputCommitter for the following:
- Set up the job initialization
- Cleaning up the job after the job completion
- Set up the task’s temporary output
- Check whether a task needs a commit
- Committing the task output
- Discard the task commit
Explain the process of spilling in MapReduce.
Spilling is a process of copying the data from memory buffer to disk when the buffer usage reaches a specific threshold size. This happens when there is not enough memory to fit all of the mapper output. By default, a background thread starts spilling the content from memory to disk after 80 percent of the buffer size is filled.
For a 100 MB size buffer, the spilling will start after the content of the buffer reaches a size of 80 MB.
How can you set the mappers and reducers for a MapReduce job?
The number of mappers and reducers can be set in the command line using:
-D mapred.map.tasks=5 –D mapred.reduce.tasks=2
In the code, one can configure JobConf variables:
job.setNumMapTasks(5); // 5 mappers
job.setNumReduceTasks(2); // 2 reducers
What happens when a node running a map task fails before sending the output to the reducer?
If this ever happens, map tasks will be assigned to a new node, and the entire task will be rerun to re-create the map output. In Hadoop v2, the YARN framework has a temporary daemon called application master, which takes care of the execution of the application. If a task on a particular node failed due to the unavailability of a node, it is the role of the application master to have this task scheduled on another node.
Can we write the output of MapReduce in different formats?
Yes. Hadoop supports various input and output formats, such as:
- TextOutputFormat – This is the default output format and it writes records as lines of text.
- SequenceFileOutputFormat – This is used to write sequence files when the output files need to be fed into another MapReduce job as input files.
- MapFileOutputFormat – This is used to write the output as map files.
- SequenceFileAsBinaryOutputFormat – This is another variant of SequenceFileInputFormat. It writes keys and values to a sequence file in binary format.
- DBOutputFormat – This is used for writing to relational databases and HBase. This format also sends the reduce output to a SQL table.
What benefits did YARN bring in Hadoop 2.0 and how did it solve the issues of MapReduce v1?
In Hadoop v1, MapReduce performed both data processing and resource management; there was only one master process for the processing layer known as JobTracker. JobTracker was responsible for resource tracking and job scheduling.
Managing jobs using a single JobTracker and utilization of computational resources was inefficient in MapReduce 1. As a result, JobTracker was overburdened due to handling, job scheduling, and resource management. Some of the issues were scalability, availability issue, and resource utilization. In addition to these issues, the other problem was that non-MapReduce jobs couldn’t run in v1.
To overcome this issue, Hadoop 2 introduced YARN as the processing layer. In YARN, there is a processing master called ResourceManager. In Hadoop v2, you have ResourceManager running in high availability mode. There are node managers running on multiple machines, and a temporary daemon called application master. Here, the ResourceManager is only handling the client connections and taking care of tracking the resources.
In Hadoop v2, the following features are available:
- Scalability – You can have a cluster size of more than 10,000 nodes and you can run more than 100,000 concurrent tasks.
- Compatibility – The applications developed for Hadoop v1 run on YARN without any disruption or availability issues.
- Resource utilization – YARN allows the dynamic allocation of cluster resources to improve resource utilization.
- Multitenancy – YARN can use open-source and proprietary data access engines, as well as perform real-time analysis and run ad-hoc queries.
Explain how YARN allocates resources to an application with the help of its architecture.
There is a client/application/API which talks to ResourceManager. The ResourceManager manages the resource allocation in the cluster. It has two internal components, scheduler, and application manager. The ResourceManager is aware of the resources that are available with every node manager. The scheduler allocates resources to various running applications when they are running in parallel. It schedules resources based on the requirements of the applications. It does not monitor or track the status of the applications.
Applications Manager is what accepts job submissions. It monitors and restarts the application masters in case of failures. Application Master manages the resource needs of individual applications. It interacts with the scheduler to acquire the required resources, and with NodeManager to execute and monitor tasks, which tracks the jobs running. It monitors each container’s resource utilization.
A container is a collection of resources, such as RAM, CPU, or network bandwidth. It provides the rights to an application to use a specific amount of resources.
Let us have a look at the architecture of YARN:
Whenever a job submission happens, ResourceManager requests the NodeManager to hold some resources for processing. NodeManager then guarantees the container that would be available for processing. Next, the ResourceManager starts a temporary daemon called application master to take care of the execution. The App Master, which the applications manager launches, will run in one of the containers. The other containers will be utilized for execution. This is briefly how YARN takes care of the allocation.
Which of the following has replaced JobTracker from MapReduce v1?
- NodeManager
- ApplicationManager
- ResourceManager
- Scheduler
The answer is ResourceManager. It is the name of the master process in Hadoop v2.
Write the YARN commands to check the status of an application and kill an application.
The commands are as follows:
- a) To check the status of an application:
yarn application -status ApplicationID
- b) To kill or terminate an application:
yarn application -kill ApplicationID
Can we have more than one ResourceManager in a YARN-based cluster?
Yes, Hadoop v2 allows us to have more than one ResourceManager. You can have a high availability YARN cluster where you can have an active ResourceManager and a standby ResourceManager, where the ZooKeeper handles the coordination.
There can only be one active ResourceManager at a time. If an active ResourceManager fails, then the standby ResourceManager comes to the rescue.
What are the different schedulers available in YARN?
The different schedulers available in YARN are:
- FIFO scheduler – This places application in a queue and runs them in the order of submission (first in, first out). It is not desirable, as a long-running application might block the small running applications
- Capacity scheduler – A separate dedicated queue allows the small job to start as soon as it is submitted. The large job finishes later compared to using the FIFO scheduler
- Fair scheduler – There is no need to reserve a set amount of capacity since it will dynamically balance resources between all the running jobs
What happens if a ResourceManager fails while executing an application in a high availability cluster?
In a high availability cluster, there are two ResourceManagers: one active and the other standby. If a ResourceManager fails in the case of a high availability cluster, the standby will be elected as active and instructs the ApplicationMaster to abort. The ResourceManager recovers its running state by taking advantage of the container statuses sent from all node managers.
In a cluster of 10 DataNodes, each having 16 GB RAM and 10 cores, what would be the total processing capacity of the cluster?
Every node in a Hadoop cluster will have one or multiple processes running, which would need RAM. The machine itself, which has a Linux file system, would have its own processes that need a specific amount of RAM usage. Therefore, if you have 10 DataNodes, you need to allocate at least 20 to 30 percent towards the overheads, Cloudera-based services, etc. You could have 11 or 12 GB and six or seven cores available on every machine for processing. Multiply that by 10, and that’s your processing capacity.
What happens if requested memory or CPU cores go beyond the size of container allocation?
If an application starts demanding more memory or more CPU cores that cannot fit into a container allocation, your application will fail. This happens because the requested memory is more than the maximum container size.
Now that you have learned about HDFS, MapReduce, and YARN, let us move to the next section. We’ll go over questions about Hive, Pig, HBase, and Sqoop.
Before moving into the Hive interview questions, let us summarize what Hive is all about. Facebook adopted the Hive to overcome MapReduce’s limitations. MapReduce proved to be difficult for users as they found it challenging to code because not all of them were well-versed with the coding languages. Users required a language similar to SQL, which was well-known to all the users. This gave rise to Hive.
Hive is a data warehouse system used to query and analyze large datasets stored in HDFS. It uses a query language called HiveQL, which is similar to SQL. Hive works on structured data. Let us now have a look at a few Hive questions.
What are the different components of a Hive architecture?
The different components of the Hive are:
- User Interface: This calls the execute interface to the driver and creates a session for the query. Then, it sends the query to the compiler to generate an execution plan for it
- Metastore: This stores the metadata information and sends it to the compiler for the execution of a query
- Compiler: This generates the execution plan. It has a DAG of stages, where each stage is either a metadata operation, a map, or reduces a job or operation on HDFS
- Execution Engine: This acts as a bridge between the Hive and Hadoop to process the query. Execution Engine communicates bidirectionally with Metastore to perform operations, such as create or drop tables.
What is a partition in Hive and why is partitioning required in Hive
Partition is a process for grouping similar types of data together based on columns or partition keys. Each table can have one or more partition keys to identify a particular partition.
Partitioning provides granularity in a Hive table. It reduces the query latency by scanning only relevant partitioned data instead of the entire data set. We can partition the transaction data for a bank based on month — January, February, etc. Any operation regarding a particular month, say February, will only have to scan the February partition, rather than the entire table data.
Why does Hive not store metadata information in HDFS?
We know that the Hive’s data is stored in HDFS. However, the metadata is either stored locally or it is stored in RDBMS. The metadata is not stored in HDFS, because HDFS read/write operations are time-consuming. As such, Hive stores metadata information in the metastore using RDBMS instead of HDFS. This allows us to achieve low latency and is faster.
What are the components used in Hive query processors?
The components used in Hive query processors are:
- Parser
- Semantic Analyzer
- Execution Engine
- User-Defined Functions
- Logical Plan Generation
- Physical Plan Generation
- Optimizer
- Operators
- Type checking
Suppose there are several small CSV files present in /user/input directory in HDFS and you want to create a single Hive table from these files. The data in these files have the following fields: {registration_no, name, email, address}. What will be your approach to solve this, and where will you create a single Hive table for multiple smaller files without degrading the performance of the system?
Using SequenceFile format and grouping these small files together to form a single sequence file can solve this problem. Below are the steps:
Write a query to insert a new column(new_col INT) into a hive table (h_table) at a position before an existing column (x_col).
The following query will insert a new column:
ALTER TABLE h_table
CHANGE COLUMN new_col INT
BEFORE x_col
What are the key differences between Hive and Pig?
Pig is a scripting language that is used for data processing. Let us have a look at a few questions involving Pig:
What are the different ways of executing a Pig script?
The different ways of executing a Pig script are as follows:
- Grunt shell
- Script file
- Embedded script
What are the major components of a Pig execution environment?
The major components of a Pig execution environment are:
- Pig Scripts: They are written in Pig Latin using built-in operators and UDFs, and submitted to the execution environment.
- Parser: Completes type checking and checks the syntax of the script. The output of the parser is a Directed Acyclic Graph (DAG).
- Optimizer: Performs optimization using merge, transform, split, etc. Optimizer aims to reduce the amount of data in the pipeline.
- Compiler: Converts the optimized code into MapReduce jobs automatically.
- Execution Engine: MapReduce jobs are submitted to execution engines to generate the desired results.
Explain the different complex data types in Pig.
Pig has three complex data types, which are primarily Tuple, Bag, and Map.
Tuple
A tuple is an ordered set of fields that can contain different data types for each field. It is represented by braces ().
Example: (1,3)
Bag
A bag is a set of tuples represented by curly braces {}.
Example: {(1,4), (3,5), (4,6)}
Map
A map is a set of key-value pairs used to represent data elements. It is represented in square brackets [ ].
Example: [key#value, key1#value1,….]
What are the various diagnostic operators available in Apache Pig?
Pig has Dump, Describe, Explain, and Illustrate as the various diagnostic operators.
Dump
The dump operator runs the Pig Latin scripts and displays the results on the screen.
Load the data using the “load” operator into Pig.
Display the results using the “dump” operator.
Describe
Describe operator is used to view the schema of a relation.
Load the data using “load” operator into Pig
View the schema of a relation using “describe” operator
Explain
Explain operator displays the physical, logical and MapReduce execution plans.
Load the data using “load” operator into Pig
Display the logical, physical and MapReduce execution plans using “explain” operator
Illustrate
Illustrate operator gives the step-by-step execution of a sequence of statements.
Load the data using “load” operator into Pig
Show the step-by-step execution of a sequence of statements using “illustrate” operator
State the usage of the group, order by, and distinct keywords in Pig scripts.
The group statement collects various records with the same key and groups the data in one or more relations.
Example: Group_data = GROUP Relation_name BY AGE
The order statement is used to display the contents of relation in sorted order based on one or more fields.
Example: Relation_2 = ORDER Relation_name1 BY (ASC|DSC)
Distinct statement removes duplicate records and is implemented only on entire records, and not on individual records.
Example: Relation_2 = DISTINCT Relation_name1
What are the relational operators in Pig?
The relational operators in Pig are as follows:
COGROUP
It joins two or more tables and then performs GROUP operation on the joined table result.
CROSS
This is used to compute the cross product (cartesian product) of two or more relations.
FOREACH
This will iterate through the tuples of a relation, generating a data transformation.
JOIN
This is used to join two or more tables in a relation.
LIMIT
This will limit the number of output tuples.
SPLIT
This will split the relation into two or more relations.
UNION
It will merge the contents of two relations.
ORDER
This is used to sort a relation based on one or more fields.
What is the use of having filters in Apache Pig?
FilterOperator is used to select the required tuples from a relation based on a condition. It also allows you to remove unwanted records from the data file.
Example: Filter the products with a whole quantity that is greater than 1000
A = LOAD ‘/user/Hadoop/phone_sales’ USING PigStorage(‘,’) AS (year:int, product:chararray, quantity:int);
B = FILTER A BY quantity > 1000
Suppose there’s a file called “test.txt” having 150 records in HDFS. Write the PIG command to retrieve the first 10 records of the file.
To do this, we need to use the limit operator to retrieve the first 10 records from a file.
Load the data in Pig:
test_data = LOAD “/user/test.txt” USING PigStorage(‘,’) as (field1, field2,….);
Limit the data to first 10 records:
Limit_test_data = LIMIT test_data 10;
Now let’s have a look at questions from HBase. HBase is a NoSQL database that runs on top of Hadoop. It is a four-dimensional database in comparison to RDBMS databases, which are usually two-dimensional.
What are the key components of HBase?
This is one of the most common interview questions.
Region Server
Region server contains HBase tables that are divided horizontally into “Regions” based on their key values. It runs on every node and decides the size of the region. Each region server is a worker node that handles read, writes, updates, and delete request from clients.
HMaster
This assigns regions to RegionServers for load balancing, and monitors and manages the Hadoop cluster. Whenever a client wants to change the schema and any of the metadata operations, HMaster is used.
ZooKeeper
This provides a distributed coordination service to maintain server state in the cluster. It looks into which servers are alive and available, and provides server failure notifications. Region servers send their statuses to ZooKeeper indicating if they are ready to reading and write operations.
Explain what row key and column families in HBase is.
The row key is a primary key for an HBase table. It also allows logical grouping of cells and ensures that all cells with the same row key are located on the same server.
Column families consist of a group of columns that are defined during table creation, and each column family has certain column qualifiers that a delimiter separates.
Why do we need to disable a table in HBase and how do you do it?
The HBase table is disabled to allow modifications to its settings. When a table is disabled, it cannot be accessed through the scan command.
To disable the employee table, use the command:
disable ‘employee_table’
To check if the table is disabled, use the command:
is_disabled ‘employee_table’
Write the code needed to open a connection in HBase.
The following code is used to open a connection in HBase:
Configuration myConf = HBaseConfiguration.create();
HTableInterface usersTable = new HTable(myConf, “users”);
What does replication mean in terms of HBase?
The replication feature in HBase provides a mechanism to copy data between clusters. This feature can be used as a disaster recovery solution that provides high availability for HBase.
The following commands alter the hbase1 table and set the replication_scope to 1. A replication_scope of 0 indicates that the table is not replicated.
disable ‘hbase1’
alter ‘hbase1’, {NAME => ‘family_name’, REPLICATION_SCOPE => ‘1’}
enable ‘hbase1’
Can you import/export in an HBase table?
Yes, it is possible to import and export tables from one HBase cluster to another.
HBase export utility:
hbase org.apache.hadoop.hbase.mapreduce.Export “table name” “target export location”
Example: hbase org.apache.hadoop.hbase.mapreduce.Export “employee_table” “/export/employee_table”
HBase import utility:
create ‘emp_table_import’, {NAME => ‘myfam’, VERSIONS => 10}
hbase org.apache.hadoop.hbase.mapreduce.Import “table name” “target import location”
Example: create ‘emp_table_import’, {NAME => ‘myfam’, VERSIONS => 10}
hbase org.apache.hadoop.hbase.mapreduce.Import “emp_table_import” “/export/employee_table”
What is compaction in HBase?
Compaction is the process of merging HBase files into a single file. This is done to reduce the amount of memory required to store the files and the number of disk seeks needed. Once the files are merged, the original files are deleted.
How does Bloom filter work?
The HBase Bloom filter is a mechanism to test whether an HFile contains a specific row or row-col cell. The Bloom filter is named after its creator, Burton Howard Bloom. It is a data structure that predicts whether a given element is a member of a set of data. These filters provide an in-memory index structure that reduces disk reads and determines the probability of finding a row in a particular file.
Does HBase have any concept of the namespace?
A namespace is a logical grouping of tables, analogous to a database in RDBMS. You can create the HBase namespace to the schema of the RDBMS database.
To create a namespace, use the command:
create_namespace ‘namespace name’
To list all the tables that are members of the namespace, use the command: list_namespace_tables ‘default’
To list all the namespaces, use the command:
list_namespace
- How does the Write Ahead Log (WAL) help when a RegionServer crashes?
If a RegionServer hosting a MemStore crash, the data that existed in memory, but not yet persisted, is lost. HBase recovers against that by writing to the WAL before the write completes. The HBase cluster keeps a WAL to record changes as they happen. If HBase goes down, replaying the WAL will recover data that was not yet flushed from the MemStore to the HFile.
What are the default file formats to import data using Sqoop?
The default file formats are Delimited Text File Format and SequenceFile Format. Let us understand each of them individually:
Delimited Text File Format
This is the default import format and can be specified explicitly using the –as-textfile argument. This argument will write string-based representations of each record to the output files, with delimiter characters between individual columns and rows.
1,here is a message,2010-05-01
2,strive to learn,2010-01-01
3,third message,2009-11-12
SequenceFile Format
SequenceFile is a binary format that stores individual records in custom record-specific data types. These data types are manifested as Java classes. Sqoop will automatically generate
these data types for you. This format supports the exact storage of all data in binary representations and is appropriate for storing binary data.
What is the importance of the eval tool in Sqoop?
The Sqoop eval tool allows users to execute user-defined queries against respective database servers and preview the result in the console.
Explain how does Sqoop imports and exports data between RDBMS and HDFS with its architecture.
- Introspect database to gather metadata (primary key information)
- Sqoop divides the input dataset into splits and uses individual map tasks to push the splits to
HDFS
- Introspect database to gather metadata (primary key information)
- Sqoop divides the input dataset into splits and uses individual map tasks to push the splits to RDBMS. Sqoop will export Hadoop files back to RDBMS tables.
Write the HBase command to list the contents and update the column families of a table.
The following code is used to list the contents of an HBase table:
scan ‘table_name’
Example: scan ‘employee_table’
To update column families in the table, use the following command:
alter ‘table_name’, ‘column_family_name’
Example: alter ‘employee_table’, ‘emp_address’
What are catalog tables in HBase?
The catalog has two tables: hbasemeta and -ROOT-
The catalog table hbase:meta exists as an HBase table and is filtered out of the HBase shell’s list command. It keeps a list of all the regions in the system and the location of hbase:meta is
stored in ZooKeeper. The -ROOT- table keeps track of the location of the .META table.
What is hotspotting in HBase and how can it be avoided?
In HBase, all read and write requests should be uniformly distributed across all of the regions in the RegionServers. Hotspotting occurs when a given region serviced by a single RegionServer receives most or all of the read or write requests.
Hotspotting can be avoided by designing the row key in such a way that data is written should go to multiple regions across the cluster. Below are the techniques to do so:
- Salting
- Hashing
- Reversing the key
Suppose you have a database “test_db” in MySQL. Write the command to connect this database and import tables to Sqoop.
The following commands show how to import the test_db database and test_demo table, and how to present it to Sqoop.
Explain how to export a table back to RDBMS with an example.
Suppose there is a “departments” table in “retail_db” that is already imported into Sqoop and you need to export this table back to RDBMS.
- i) Create a new “dept” table to export in RDBMS (MySQL)
2. ii) Export “departments” table to the “dept” table
How will you update the columns that are already exported? Write a Sqoop command to show all the databases in the MySQL server?
To update a column of a table which is already exported, we use the command:
–update-key parameter
The following is an example:
sqoop export –connect
jdbc:mysql://localhost/dbname –username root
–password cloudera –export-dir /input/dir
–table test_demo –-fields-terminated-by “,”
–update-key column_name
What is Codegen in Sqoop?
The Codegen tool in Sqoop generates the Data Access Object (DAO) Java classes that encapsulate and interpret imported records.
The following example generates Java code for an “employee” table in the “testdb” database.
$ sqoop codegen \
–connect jdbc:mysql://localhost/testdb \
–username root \
–table employee
Can Sqoop be used to convert data in different formats? If not, which tools can be used for this purpose?
Yes, Sqoop can be used to convert data into different formats. This depends on the different arguments that are used for importing.
What do you know about the term “Big Data”?
Big Data is a term associated with complex and large datasets. A relational database cannot handle big data, and that’s why special tools and methods are used to perform operations on a vast collection of data. Big data enables companies to understand their business better and helps them derive meaningful information from the unstructured and raw data collected on a regular basis. Big data also allows the companies to take better business decisions backed by data.
What are the five V’s of Big Data?
The five V’s of Big data is as follows:
- Volume – Volume represents the volume i.e. amount of data that is growing at a high rate i.e. data volume in Petabytes
- Velocity – Velocity is the rate at which data grows. Social media contributes a major role in the velocity of growing data.
- Variety – Variety refers to the different data types i.e. various data formats like text, audios, videos, etc.
- Veracity – Veracity refers to the uncertainty of available data. Veracity arises due to the high volume of data that brings incompleteness and inconsistency.
- Value –Value refers to turning data into value. By turning accessed big data into values, businesses may generate revenue.
5 V’s of Big Data
Note: This is one of the basic and significant questions asked in the big data interview. You can choose to explain the five V’s in detail if you see the interviewer is interested to know more. However, the names can even be mentioned if you are asked about the term “Big Data”.
Tell us how big data and Hadoop are related to each other.
Big data and Hadoop are almost synonyms terms. With the rise of big data, Hadoop, a framework that specializes in big data operations also became popular. The framework can be used by professionals to analyze big data and help businesses to make decisions.
Note: This question is commonly asked in a big data interview. You can go further to answer this question and try to explain the main components of Hadoop.
How is big data analysis helpful in increasing business revenue?
Big data analysis has become very important for the businesses. It helps businesses to differentiate themselves from others and increase the revenue. Through predictive analytics, big data analytics provides businesses customized recommendations and suggestions. Also, big data analytics enables businesses to launch new products depending on customer needs and preferences. These factors make businesses earn more revenue, and thus companies are using big data analytics. Companies may encounter a significant increase of 5-20% in revenue by implementing big data analytics. Some popular companies those are using big data analytics to increase their revenue is – Walmart, LinkedIn, Facebook, Twitter, Bank of America etc.
Explain the steps to be followed to deploy a Big Data solution.
Followings are the three steps that are followed to deploy a Big Data Solution –
- Data Ingestion
The first step for deploying a big data solution is the data ingestion i.e. extraction of data from various sources. The data source may be a CRM like Salesforce, Enterprise Resource Planning System like SAP, RDBMS like MySQL or any other log files, documents, social media feeds etc. The data can be ingested either through batch jobs or real-time streaming. The extracted data is then stored in HDFS.
Steps of Deploying Big Data Solution
- Data Storage
After data ingestion, the next step is to store the extracted data. The data either be stored in HDFS or NoSQL database (i.e. HBase). The HDFS storage works well for sequential access whereas HBase for random read/write access.
iii. Data Processing
The final step in deploying a big data solution is the data processing. The data is processed through one of the processing frameworks like Spark, MapReduce, Pig, etc.
What is the role of the JDBC driver in the Sqoop setup? Is the JDBC driver enough to connect Sqoop to the database?
JDBC driver is a standard Java API used for accessing different databases in RDBMS using Sqoop. Each database vendor is responsible for writing their own implementation that will communicate with the corresponding database with its native protocol. Each user needs to download the drivers separately and install them onto Sqoop prior to its use.
JDBC driver alone is not enough to connect Sqoop to the database. We also need connectors to interact with different databases. A connector is a pluggable piece that is used to fetch metadata and allows Sqoop to overcome the differences in SQL dialects supported by various databases, along with providing optimized data transfer.
Define respective components of HDFS and YARN
The two main components of HDFS are-
- NameNode – This is the master node for processing metadata information for data blocks within the HDFS
- DataNode/Slave node – This is the node which acts as slave node to store the data, for processing and use by the NameNode
In addition to serving the client requests, the NameNode executes either of two following roles –
- CheckpointNode – It runs on a different host from the NameNode
- BackupNode- It is a read-only NameNode which contains file system metadata information excluding the block locations
The two main components of YARN are–
- ResourceManager– This component receives processing requests and accordingly allocates to respective NodeManagers depending on processing needs.
- NodeManager– It executes tasks on each single Data Node
Why is Hadoop used for Big Data Analytics?
Since data analysis has become one of the key parameters of business, hence, enterprises are dealing with massive amount of structured, unstructured and semi-structured data. Analyzing unstructured data is quite difficult where Hadoop takes major part with its capabilities of
- Storage
- Processing
- Data collection
Moreover, Hadoop is open source and runs on commodity hardware. Hence it is a cost-benefit solution for businesses.
Do you prefer good data or good models? Why?
How to Approach: This is a tricky question but generally asked in the big data interview. It asks you to choose between good data or good models. As a candidate, you should try to answer it from your experience. Many companies want to follow a strict process of evaluating data, means they have already selected data models. In this case, having good data can be game-changing. The other way around also works as a model is chosen based on good data.
As we already mentioned, answer it from your experience. However, don’t say that having both good data and good models is important as it is hard to have both in real life projects.
Will you optimize algorithms or code to make them run faster?
How to Approach: The answer to this question should always be “Yes.” Real world performance matters and it doesn’t depend on the data or model you are using in your project.
The interviewer might also be interested to know if you have had any previous experience in code or algorithm optimization. For a beginner, it obviously depends on which projects he worked on in the past. Experienced candidates can share their experience accordingly as well. However, be honest about your work, and it is fine if you haven’t optimized code in the past. Just let the interviewer know your real experience and you will be able to crack the big data interview.
How do you approach data preparation?
How to Approach: Data preparation is one of the crucial steps in big data projects. A big data interview may involve at least one question based on data preparation. When the interviewer asks you this question, he wants to know what steps or precautions you take during data preparation.
As you already know, data preparation is required to get necessary data which can then further be used for modeling purposes. You should convey this message to the interviewer. You should also emphasize the type of model you are going to use and reasons behind choosing that particular model. Last, but not the least, you should also discuss important data preparation terms such as transforming variables, outlier values, unstructured data, identifying gaps, and others.
How would you transform unstructured data into structured data?
How to Approach: Unstructured data is very common in big data. The unstructured data should be transformed into structured data to ensure proper data analysis. You can start answering the question by briefly differentiating between the two. Once done, you can now discuss the methods you use to transform one form to another. You might also share the real-world situation where you did it. If you have recently been graduated, then you can share information related to your academic projects.
By answering this question correctly, you are signaling that you understand the types of data, both structured and unstructured, and also have the practical experience to work with these. If you give an answer to this question specifically, you will definitely be able to crack the big data interview.
Which hardware configuration is most beneficial for Hadoop jobs?
Dual processors or core machines with a configuration of 4 / 8 GB RAM and ECC memory is ideal for running Hadoop operations. However, the hardware configuration varies based on the project-specific workflow and process flow and need customization accordingly.
What happens when two users try to access the same file in the HDFS?
HDFS NameNode supports exclusive write only. Hence, only the first user will receive the grant for file access and the second user will be rejected.
How to recover a NameNode when it is down?
The following steps need to execute to make the Hadoop cluster up and running:
- Use the FsImage which is file system metadata replica to start a new NameNode.
- Configure the DataNodes and also the clients to make them acknowledge the newly started NameNode.
- Once the new NameNode completes loading the last checkpoint FsImage which has received enough block reports from the DataNodes, it will start to serve the client.
In case of large Hadoop clusters, the NameNode recovery process consumes a lot of time which turns out to be a more significant challenge in case of routine maintenance.
What do you understand by Rack Awareness in Hadoop?
It is an algorithm applied to the NameNode to decide how blocks and its replicas are placed. Depending on rack definitions network traffic is minimized between DataNodes within the same rack. For example, if we consider replication factor as 3, two copies will be placed on one rack whereas the third copy in a separate rack.
What is the difference between “HDFS Block” and “Input Split”?
The HDFS divides the input data physically into blocks for processing which is known as HDFS Block.
Input Split is a logical division of data by mapper for mapping operation.
Hadoop is one of the most popular Big Data frameworks, and if you are going for a Hadoop interview prepare yourself with these basic level interview questions for Big Data Hadoop. These questions will be helpful for you whether you are going for a Hadoop developer or Hadoop Admin interview.
What are the common input formats in Hadoop?
Below are the common input formats in Hadoop –
- Text Input Format – The default input format defined in Hadoop is the Text Input Format.
- Sequence File Input Format – To read files in a sequence, Sequence File Input Format is used.
- Key Value Input Format – The input format used for plain text files (files broken into lines) is the Key Value Input Format.
Explain some important features of Hadoop.
Hadoop supports the storage and processing of big data. It is the best solution for handling big data challenges. Some important features of Hadoop are –
- Open Source – Hadoop is an open-source framework which means it is available free of cost. Also, the users are allowed to change the source code as per their requirements.
- Distributed Processing – Hadoop supports distributed processing of data i.e. faster processing. The data in Hadoop HDFS is stored in a distributed manner and MapReduce is responsible for the parallel processing of data.
- Fault Tolerance – Hadoop is highly fault-tolerant. It creates three replicas for each block at different nodes, by default. This number can be changed according to the requirement. So, we can recover the data from another node if one node fails. The detection of node failure and recovery of data is done automatically.
- Reliability – Hadoop stores data on the cluster in a reliable manner that is independent of machine. So, the data stored in Hadoop environment is not affected by the failure of the machine.
- Scalability – Another important feature of Hadoop is the scalability. It is compatible with the other hardware and we can easily ass the new hardware to the nodes.
- High Availability – The data stored in Hadoop is available to access even after the hardware failure. In case of hardware failure, the data can be accessed from another path.
Explain the different modes in which Hadoop run.
Apache Hadoop runs in the following three modes –
- Standalone (Local) Mode – By default, Hadoop runs in a local mode i.e. on a non-distributed, single node. This mode uses the local file system to perform input and output operation. This mode does not support the use of HDFS, so it is used for debugging. No custom configuration is needed for configuration files in this mode.
- Pseudo-Distributed Mode – In the pseudo-distributed mode, Hadoop runs on a single node just like the Standalone mode. In this mode, each daemon runs in a separate Java process. As all the daemons run on a single node, there is the same node for both the Master and Slave nodes.
- Fully – Distributed Mode – In the fully-distributed mode, all the daemons run on separate individual nodes and thus forms a multi-node cluster. There are different nodes for Master and Slave nodes.
Explain the core components of Hadoop.
Hadoop is an open-source framework that is meant for storage and processing of big data in a distributed manner. The core components of Hadoop are –
- HDFS (Hadoop Distributed File System) – HDFS is the basic storage system of Hadoop. The large data files running on a cluster of commodity hardware are stored in HDFS. It can store data in a reliable manner even when hardware fails.
Core Components of Hadoop
- Hadoop MapReduce – MapReduce is the Hadoop layer that is responsible for data processing. It writes an application to process unstructured and structured data stored in HDFS. It is responsible for the parallel processing of high volume of data by dividing data into independent tasks. The processing is done in two phases Map and Reduce. The Map is the first phase of processing that specifies complex logic code and the Reduce is the second phase of processing that specifies light-weight operations.
- YARN – The processing framework in Hadoop is YARN. It is used for resource management and provides multiple data processing engines i.e. data science, real-time streaming, and batch processing.
What are the configuration parameters in a “MapReduce” program?
The main configuration parameters in “MapReduce” framework are:
- Input locations of Jobs in the distributed file system
- Output location of Jobs in the distributed file system
- The input format of data
- The output format of data
- The class which contains the map function
- The class which contains the reduce function
- JAR file which contains the mapper, reducer and the driver classes
What is a block in HDFS and what is its default size in Hadoop 1 and Hadoop 2? Can we change the block size?
Blocks are smallest continuous data storage in a hard drive. For HDFS, blocks are stored across Hadoop cluster.
- The default block size in Hadoop 1 is: 64 MB
- The default block size in Hadoop 2 is: 128 MB
Yes, we can change block size by using the parameter – dfs.block.size located in the hdfs-site.xml file.
What is Distributed Cache in a MapReduce Framework
Distributed Cache is a feature of Hadoop MapReduce framework to cache files for applications. Hadoop framework makes cached files available for every map/reduce tasks running on the data nodes. Hence, the data files can access the cache file as a local file in the designated job.
What are the three running modes of Hadoop?
The three running modes of Hadoop are as follows:
- Standalone or local: This is the default mode and does not need any configuration. In this mode, all the following components of Hadoop uses local file system and runs on a single JVM –
- NameNode
- DataNode
- ResourceManager
- NodeManager
- Pseudo-distributed:In this mode, all the master and slave Hadoop services are deployed and executed on a single node.
iii. Fully distributed: In this mode, Hadoop master and slave services are deployed and executed on separate nodes.
Explain JobTracker in Hadoop
JobTracker is a JVM process in Hadoop to submit and track MapReduce jobs.
JobTracker performs the following activities in Hadoop in a sequence –
- JobTracker receives jobs that a client application submits to the job tracker
- JobTracker notifies NameNode to determine data node
- JobTracker allocates TaskTracker nodes based on available slots.
- it submits the work on allocated TaskTracker Nodes,
- JobTracker monitors the TaskTracker nodes.
- When a task fails, JobTracker is notified and decides how to reallocate the task.
It is not easy to crack Hadoop developer interview but the preparation can do everything. If you are a fresher, learn the Hadoop concepts and prepare properly. Have a good knowledge of the different file systems, Hadoop versions, commands, system security, etc. Here are few questions that will help you pass the Hadoop developer interview.
What are the different configuration files in Hadoop?
The different configuration files in Hadoop are –
core-site.xml – This configuration file contains Hadoop core configuration settings, for example, I/O settings, very common for MapReduce and HDFS. It uses hostname a port.
mapred-site.xml – This configuration file specifies a framework name for MapReduce by setting mapreduce.framework.name
hdfs-site.xml – This configuration file contains HDFS daemons configuration settings. It also specifies default block permission and replication checking on HDFS.
yarn-site.xml – This configuration file specifies configuration settings for ResourceManager and NodeManager.
How can you achieve security in Hadoop?
Kerberos are used to achieve security in Hadoop. There are 3 steps to access a service while using Kerberos, at a high level. Each step involves a message exchange with a server.
- Authentication – The first step involves authentication of the client to the authentication server, and then provides a time-stamped TGT (Ticket-Granting Ticket) to the client.
- Authorization – In this step, the client uses received TGT to request a service ticket from the TGS (Ticket Granting Server).
- Service Request – It is the final step to achieve security in Hadoop. Then the client uses service ticket to authenticate himself to the server.
What is commodity hardware?
Commodity hardware is a low-cost system identified by less-availability and low-quality. The commodity hardware comprises of RAM as it performs a number of services that require RAM for the execution. One doesn’t require high-end hardware configuration or supercomputers to run Hadoop, it can be run on any commodity hardware.
How is NFS different from HDFS?
There are a number of distributed file systems that work in their own way. NFS (Network File System) is one of the oldest and popular distributed file storage systems whereas HDFS (Hadoop Distributed File System) is the recently used and popular one to handle big data. The main differences between NFS and HDFS are as follows –
How do Hadoop MapReduce work?
There are two phases of MapReduce operation.
- Map phase – In this phase, the input data is split by map tasks. The map tasks run in parallel. These split data is used for analysis purpose.
- Reduce phase- In this phase, the similar split data is aggregated from the entire collection and shows the result.
What is MapReduce? What is the syntax you use to run a MapReduce program?
MapReduce is a programming model in Hadoop for processing large data sets over a cluster of computers, commonly known as HDFS. It is a parallel programming model.
The syntax to run a MapReduce program is – hadoop_jar_file.jar /input_path /output_path.
What are the Port Numbers for NameNode, Task Tracker, and Job Tracker?
- NameNode – Port 50070
- Task Tracker – Port 50060
- Job Tracker – Port 50030
What are the different file permissions in HDFS for files or directory levels?
Hadoop distributed file system (HDFS) uses a specific permissions model for files and directories. Following user levels are used in HDFS –
- Owner
- Group
- Others.
For each of the user mentioned above following permissions are applicable –
- read (r)
- write (w)
- execute(x).
Above mentioned permissions work differently for files and directories.
For files –
- The r permission is for reading a file
- The w permission is for writing a file.
For directories –
- The r permission lists the contents of a specific directory.
- The w permission creates or deletes a directory.
- The X permission is for accessing a child directory.
What are the basic parameters of a Mapper?
The basic parameters of a Mapper are
- LongWritable and Text
- Text and IntWritable
The interviewer has more expectations from an experienced Hadoop developer, and thus his questions are one-level up. So, if you have gained some experience, don’t forget to cover command based, scenario-based, real-experience based questions. Here we bring some sample interview questions for experienced Hadoop developers.
How to restart all the daemons in Hadoop?
To restart all the daemons, it is required to stop all the daemons first. The Hadoop directory contains sbin directory that stores the script files to stop and start daemons in Hadoop.
Use stop daemons command /sbin/stop-all.sh to stop all the daemons and then use /sin/start-all.sh command to start all the daemons again.
What is the use of jps command in Hadoop?
The jps command is used to check if the Hadoop daemons are running properly or not. This command shows all the daemons running on a machine i.e. Datanode, Namenode, NodeManager, ResourceManager etc.
Explain the process that overwrites the replication factors in HDFS.
There are two methods to overwrite the replication factors in HDFS –
Method 1: On File Basis
In this method, the replication factor is changed on the basis of file using Hadoop FS shell. The command used for this is:
$hadoop fs – setrep –w2/my/test_file
Here, test_file is the filename that’s replication factor will be set to 2.
Method 2: On Directory Basis
In this method, the replication factor is changed on directory basis i.e. the replication factor for all the files under a given directory is modified.
$hadoop fs –setrep –w5/my/test_dir
Here, test_dir is the name of the directory, the replication factor for the directory and all the files in it will be set to 5.
What will happen with a NameNode that doesn’t have any data?
A NameNode without any data doesn’t exist in Hadoop. If there is a NameNode, it will contain some data in it or it won’t exist.
Explain NameNode recovery process.
The NameNode recovery process involves the below-mentioned steps to make Hadoop cluster running:
- In the first step in the recovery process, file system metadata replica (FsImage) starts a new NameNode.
- The next step is to configure DataNodes and Clients. These DataNodes and Clients will then acknowledge new NameNode.
- During the final step, the new NameNode starts serving the client on the completion of last checkpoint FsImage loading and receiving block reports from the DataNodes.
Note: Don’t forget to mention, this NameNode recovery process consumes a lot of time on large Hadoop clusters. Thus, it makes routine maintenance difficult. For this reason, HDFS high availability architecture is recommended to use.
How Is Hadoop CLASSPATH essential to start or stop Hadoop daemons?
CLASSPATH includes necessary directories that contain jar files to start or stop Hadoop daemons. Hence, setting CLASSPATH is essential to start or stop Hadoop daemons.
However, setting up CLASSPATH every time is not the standard that we follow. Usually CLASSPATH is written inside /etc/hadoop/hadoop-env.sh file. Hence, once we run Hadoop, it will load the CLASSPATH automatically.
Why is HDFS only suitable for large data sets and not the correct tool to use for many small files?
This is due to the performance issue of NameNode. Usually, NameNode is allocated with huge space to store metadata for the large-scale file. The metadata is supposed to be a from a single file for optimum space utilization and cost benefit. In case of small size files, NameNode does not utilize the entire space which is a performance optimization issue.
Why do we need Data Locality in Hadoop? Explain.
Datasets in HDFS store as blocks in DataNodes the Hadoop cluster. During the execution of a MapReduce job the individual Mapper processes the blocks (Input Splits). If the data does not reside in the same node where the Mapper is executing the job, the data needs to be copied from the DataNode over the network to the mapper DataNode.
Now if a MapReduce job has more than 100 Mapper and each Mapper tries to copy the data from other DataNode in the cluster simultaneously, it would cause serious network congestion which is a big performance issue of the overall system. Hence, data proximity to the computation is an effective and cost-effective solution which is technically termed as Data locality in Hadoop. It helps to increase the overall throughput of the system.
Data locality can be of three types:
- Data local – In this type data and the mapper resides on the same node. This is the closest proximity of data and the most preferred scenario.
- Rack Local – In this scenarios mapper and data reside on the same rack but on the different data nodes.
- Different Rack – In this scenario mapper and data reside on the different racks.
DFS can handle a large volume of data then why do we need Hadoop framework?
Hadoop is not only for storing large data but also to process those big data. Though DFS(Distributed File System) too can store the data, but it lacks below features-
- It is not fault tolerant
- Data movement over a network depends on bandwidth.
What is Sequencefileinputformat?
Hadoop uses a specific file format which is known as Sequence file. The sequence file stores data in a serialized key-value pair. Sequencefileinputformat is an input format to read sequence files.
What are the real-time industry applications of Hadoop?
Hadoop, well known as Apache Hadoop, is an open-source software platform for scalable and distributed computing of large volumes of data. It provides rapid, high performance, and cost-effective analysis of structured and unstructured data generated on digital platforms and within the enterprise. It is used in almost all departments and sectors today.
Here are some of the instances where Hadoop is used:
- Managing traffic on streets
- Streaming processing
- Content management and archiving e-mails
- Processing rat brain neuronal signals using a Hadoop computing cluster
- Fraud detection and prevention
- Advertisements targeting platforms are using Hadoop to capture and analyze click stream, transaction, video, and social media data
- Managing content, posts, images, and videos on social media platforms
- Analyzing customer data in real time for improving business performance
- Public sector fields such as intelligence, defense, cyber security, and scientific research
- Getting access to unstructured data such as output from medical devices, doctor’s notes, lab results, imaging reports, medical correspondence, clinical data, and financial data
How is Hadoop different from other parallel computing systems?
Hadoop is a distributed file system that lets you store and handle massive amounts of data on a cloud of machines, handling data redundancy.
The primary benefit of this is that since data is stored in several nodes, it is better to process it in a distributed manner. Each node can process the data stored on it instead of spending time on moving the data over the network.
On the contrary, in the relational database computing system, we can query data in real time, but it is not efficient to store data in tables, records, and columns when the data is huge.
Hadoop also provides a scheme to build a column database with Hadoop HBase for runtime queries on rows.
In what all modes Hadoop can be run?
Hadoop can be run in three modes:
- Standalone mode:The default mode of Hadoop, it uses local file system for input and output operations. This mode is mainly used for the debugging purpose, and it does not support the use of HDFS. Further, in this mode, there is no custom configuration required for mapred-site.xml, core-site.xml, and hdfs-site.xml files. This mode works much faster when compared to other modes.
- Pseudo-distributed mode (Single-node Cluster):In this case, you need configuration for all the three files mentioned above. In this case, all daemons are running on one node, and thus both Master and Slave nodes are the same.
Fully distributed mode (Multi-node Cluster): This is the production phase of Hadoop (what Hadoop is known for) where data is used and distributed across several nodes on a Hadoop cluster. Separate nodes are allotted as Master and Slave.
Explain the major difference between HDFS block and InputSplit.
In simple terms, a block is the physical representation of data while split is the logical representation of data present in the block. Split acts as an intermediary between the block and the mapper.
Suppose we have two blocks:
Block 1: ii nntteell
Block 2: Tecklearn
Now considering the map, it will read Block 1 from ii to ll but does not know how to process Block 2 at the same time. Here comes Split into play, which will form a logical group of Block 1 and Block 2 as a single block.
It then forms a key–value pair using InputFormat and records reader and sends map for further processing with InputSplit. If you have limited resources, you can increase the split size to limit the number of maps. For instance, if there are 10 blocks of 640 MB (64 MB each) and there are limited resources, you can assign ‘split size’ as 128 MB. This will form a logical group of 128 MB, with only 5 maps executing at a time.
However, if the ‘split size’ property is set to false, the whole file will form one InputSplit and is processed by a single map, consuming more time when the file is bigger.
What is distributed cache? What are its benefits?
Distributed cache in Hadoop is a service by MapReduce framework to cache files when needed.
Once a file is cached for a specific job, Hadoop will make it available on each DataNode both in system and in memory, where map and reduce tasks are executing. Later, you can easily access and read the cache file and populate any collection (like array, hashmap) in your code.
Benefits of using distributed cache are as follows:
- It distributes simple, read-only text/data files and/or complex types such as jars, archives, and others. These archives are then un-archived at the slave node.
- Distributed cache tracks the modification timestamps of cache files, which notify that the files should not be modified until a job is executed.
Explain the difference between NameNode, Checkpoint NameNode, and Backup Node.
NameNode is the core of HDFS that manages the metadata—the information of which file maps to which block locations and which blocks are stored on which DataNode. In simple terms, it’s the data about the data being stored. NameNode supports a directory tree-like structure consisting of all the files present in HDFS on a Hadoop cluster. It uses the following files for namespace:
- fsimage file: It keeps track of the latest Checkpoint of the namespace.
- edits file: It is a log of changes that have been made to the namespace since Checkpoint
Checkpoint NameNode has the same directory structure as NameNode and creates Checkpoints for namespace at regular intervals by downloading the fsimage, editing files, and margining them within the local directory. The new image after merging is then uploaded to NameNode. There is a similar node like Checkpoint, commonly known as the Secondary Node, but it does not support the ‘upload to NameNode’ functionality.
Backup Node provides similar functionality as Checkpoint, enforcing synchronization with NameNode. It maintains an up-to-date in-memory copy of the file system namespace and doesn’t require getting hold of changes after regular intervals. The Backup Node needs to save the current state in-memory to an image file to create a new Checkpoint
What are the most common input formats in Hadoop?
There are three most common input formats in Hadoop:
- Text Input Format: Default input format in Hadoop
- Key–Value Input Format: Used for plain text files where the files are broken into lines
- Sequence File Input Format: Used for reading files in sequence
Define DataNode. How does NameNode tackle DataNode failures?
DataNode stores data in HDFS; it is a node where actual data resides in the file system. Each DataNode sends a heartbeat message to notify that it is alive. If the NameNode does not receive a message from the DataNode for 10 minutes, the NameNode considers the DataNode to be dead or out of place and starts the replication of blocks that were hosted on that DataNode such that they are hosted on some other DataNode. A BlockReport contains a list of the all blocks on a DataNode. Now, the system starts to replicate what were stored in the dead DataNode.
The NameNode manages the replication of the data blocks from one DataNode to another. In this process, the replication data gets transferred directly between DataNodes such that the data never passes the NameNode.
What are the core methods of a Reducer?
The three core methods of a Reducer are as follows:
setup(): This method is used for configuring various parameters such as input data size and distributed cache.
public void setup (context)
reduce(): Heart of the Reducer is always called once per key with the associated reduced task.
public void reduce(Key, Value, context)
cleanup(): This method is called to clean the temporary files, only once at the end of the task.
public void cleanup (context)
What is a SequenceFile in Hadoop?
Extensively used in MapReduce I/O formats, SequenceFile is a flat file containing binary key–value pairs. The map outputs are stored as SequenceFile internally. It provides Reader, Writer, and Sorter classes. The three SequenceFile formats are as follows:
- Uncompressed key–value records
- Record compressed key–value records—only ‘values’ are compressed here
- Block compressed key–value records—both keys and values are collected in ‘blocks’ separately and compressed. The size of the ‘block’ is configurable
What is the role of a JobTracker in Hadoop?
A JobTracker’s primary function is resource management (managing the TaskTrackers), tracking resource availability, and task life cycle management (tracking the tasks’ progress and fault tolerance).
- It is a process that runs on a separate node, often not on a DataNode.
- The JobTracker communicates with the NameNode to identify data location.
- It finds the best TaskTracker nodes to execute the tasks on the given nodes.
- It monitors individual TaskTrackers and submits the overall job back to the client.
- It tracks the execution of MapReduce workloads local to the slave node.
What happens if you try to run a Hadoop job with an output directory that is already present?
It will throw an exception saying that the output file directory already exists.
To run the MapReduce job, you need to ensure that the output directory does not exist in the HDFS.
To delete the directory before running the job, we can use shell:
Hadoop fs –rmr /path/to/your/output/
Or the Java API:
FileSystem.getlocal(conf).delete(outputDir, true);
How can you debug Hadoop code?
First, we should check the list of MapReduce jobs currently running. Next, we need to see that there are no orphaned jobs running; if yes, we need to determine the location of RM logs.
- Run:
ps –ef | grep –I ResourceManager
Then, look for the log directory in the displayed result. We have to find out the job ID from the displayed list and check if there is any error message associated with that job.
- On the basis of RM logs, we need to identify the worker node that was involved in the execution of the task.
- Now, we will login to that node and run the below code:
ps –ef | grep –iNodeManager
- Then, we will examine the Node Manager log. The majority of errors come from the user-level logs for each MapReduce job.
How to configure Replication Factor in HDFS?
The hdfs-site.xml file is used to configure HDFS. Changing the dfs.replication property in hdfs-site.xml will change the default replication for all the files placed in HDFS.
We can also modify the replication factor on a per-file basis using the below:
Hadoop FS Shell:[training@localhost ~]$ hadoopfs –setrep –w 3 /my/fileConversely,
We can also change the replication factor of all the files under a directory.
[training@localhost ~]$ hadoopfs –setrep –w 3 -R /my/dir
How to compress a Mapper output not touching Reducer output?
To achieve this compression, we should set:
conf.set(“mapreduce.map.output.compress”, true)
conf.set(“mapreduce.output.fileoutputformat.compress”, false)
What is the difference between Map-side Join and Reduce-side Join?
Map-side Join at Map side is performed when data reaches the Map. We need a strict structure for defining Map-side Join.
On the other hand, Reduce-side Join (Repartitioned Join) is simpler than Map-side Join since here the input datasets need not be structured. However, it is less efficient as it will have to go through sort and shuffle phases, coming with network overheads.
How can you transfer data from Hive to HDFS?
By writing the query:
hive> insert overwrite directory ‘/’ select * from emp;
We can write our query for the data we want to import from Hive to HDFS. The output we receive will be stored in part files in the specified HDFS path.
What is fsck?
fsck stands for File System Check. It is a command used by HDFS. This command is used to check inconsistencies and if there is any problem in the file. For example, if there are any missing blocks for a file, HDFS gets notified through this command.
What are the main differences between NAS (Network-attached storage) and HDFS?
The main differences between NAS (Network-attached storage) and HDFS –
- HDFS runs on a cluster of machines while NAS runs on an individual machine. Hence, data redundancy is a common issue in HDFS. On the contrary, the replication protocol is different in case of NAS. Thus the chances of data redundancy are much less.
- Data is stored as data blocks in local drives in case of HDFS. In case of NAS, it is stored in dedicated hardware.
What is the Command to format the NameNode?
$ hdfs namenode -format
If you have some considerable experience of working in Big Data world, you will be asked a number of questions in your big data interview based on your previous experience. These questions may be simply related to your experience or scenario based. So, get prepared with these best Big data interview questions and answers –
Do you have any Big Data experience? If so, please share it with us.
How to Approach: There is no specific answer to the question as it is a subjective question and the answer depends on your previous experience. Asking this question during a big data interview, the interviewer wants to understand your previous experience and is also trying to evaluate if you are fit for the project requirement.
So, how will you approach the question? If you have previous experience, start with your duties in your past position and slowly add details to the conversation. Tell them about your contributions that made the project successful. This question is generally, the 2nd or 3rd question asked in an interview. The later questions are based on this question, so answer it carefully. You should also take care not to go overboard with a single aspect of your previous job. Keep it simple and to the point.
So, this brings us to the end of the Big Data Hadoop Interview Questions blog.This Tecklearn ‘Top Big Data Hadoop Interview Questions and Answers’ helps you with commonly asked questions if you are looking out for a job in Big Data Hadoop or Big Data Domain. If you wish to learn Big Data Hadoop and build a career in Big Data domain, then check out our interactive, Big Data Hadoop Architect Training, that comes with 24*7 support to guide you throughout your learning period.
https://www.tecklearn.com/course/bigdata-hadoop-architect-all-in-1-combo-course/
Big Data Hadoop-Architect (All in 1) Combo Training
About the Course
Tecklearn’s BigData Hadoop-Architect (All in 1) combo includes the following Courses:
- BigData Hadoop Analyst
- BigData Hadoop Developer
- BigData Hadoop Administrator
- BigData Hadoop Tester
- Big Data Security with Kerberos
Why Should you take Big Data Hadoop Combo Training?
- Average salary for a Hadoop Administrator ranges from approximately $104,528 to $141,391 per annum – Indeed.com
- Average salary for a Spark and Hadoop Developer ranges from approximately $106,366 to $127,619 per annum – Indeed.com
- Average salary for a Big Data Hadoop Analyst is $115,819– ZipRecruiter.com
What you will Learn in this Course?
Introduction
- The Case for Apache Hadoop
- Why Hadoop?
- Core Hadoop Components
- Fundamental Concepts
HDFS
- HDFS Features
- Writing and Reading Files
- NameNode Memory Considerations
- Overview of HDFS Security
- Using the Namenode Web UI
- Using the Hadoop File Shell
Getting Data into HDFS
- Ingesting Data from External Sources with Flume
- Ingesting Data from Relational Databases with Sqoop
- REST Interfaces
- Best Practices for Importing Data
YARN and MapReduce
- What Is MapReduce?
- Basic MapReduce Concepts
- YARN Cluster Architecture
- Resource Allocation
- Failure Recovery
- Using the YARN Web UI
- MapReduce Version 1
Planning Your Hadoop Cluster
- General Planning Considerations
- Choosing the Right Hardware
- Network Considerations
- Configuring Nodes
- Planning for Cluster Management
Hadoop Installation and Initial Configuration
- Deployment Types
- Installing Hadoop
- Specifying the Hadoop Configuration
- Performing Initial HDFS Configuration
- Performing Initial YARN and MapReduce Configuration
- Hadoop Logging
Installing and Configuring Hive, Impala, and Pig
- Hive
- Impala
- Pig
Hadoop Clients
- What is a Hadoop Client?
- Installing and Configuring Hadoop Clients
- Installing and Configuring Hue
- Hue Authentication and Authorization
Cloudera Manager
- The Motivation for Cloudera Manager
- Cloudera Manager Features
- Express and Enterprise Versions
- Cloudera Manager Topology
- Installing Cloudera Manager
- Installing Hadoop Using Cloudera Manager
- Performing Basic Administration Tasks Using Cloudera Manager
Advanced Cluster Configuration
- Advanced Configuration Parameters
- Configuring Hadoop Ports
- Explicitly Including and Excluding Hosts
- Configuring HDFS for Rack Awareness
- Configuring HDFS High Availability
Hadoop Security
- Why Hadoop Security Is Important
- Hadoop’s Security System Concepts
- What Kerberos Is and How it Works
- Securing a Hadoop Cluster with Kerberos
Managing and Scheduling Jobs
- Managing Running Jobs
- Scheduling Hadoop Jobs
- Configuring the Fair Scheduler
- Impala Query Scheduling
Cluster Maintenance
- Checking HDFS Status
- Copying Data Between Clusters
- Adding and Removing Cluster Nodes
- Rebalancing the Cluster
- Cluster Upgrading
Cluster Monitoring and Troubleshooting
- General System Monitoring
- Monitoring Hadoop Clusters
- Common Troubleshooting Hadoop Clusters
- Common Misconfigurations
Introduction to Pig
- What Is Pig?
- Pig’s Features
- Pig Use Cases
- Interacting with Pig
Basic Data Analysis with Pig
- Pig Latin Syntax
- Loading Data
- Simple Data Types
- Field Definitions
- Data Output
- Viewing the Schema
- Filtering and Sorting Data
- Commonly-Used Functions
Processing Complex Data with Pig
- Storage Formats
- Complex/Nested Data Types
- Grouping
- Built-In Functions for Complex Data
- Iterating Grouped Data
Multi-Dataset Operations with Pig
- Techniques for Combining Data Sets
- Joining Data Sets in Pig
- Set Operations
- Splitting Data Sets
Pig Troubleshooting and Optimization
- Troubleshooting Pig
- Logging
- Using Hadoop’s Web UI
- Data Sampling and Debugging
- Performance Overview
- Understanding the Execution Plan
- Tips for Improving the Performance of Your Pig Jobs
Introduction to Hive and Impala
- What Is Hive?
- What Is Impala?
- Schema and Data Storage
- Comparing Hive to Traditional Databases
- Hive Use Cases
Querying with Hive and Impala
- Databases and Tables
- Basic Hive and Impala Query Language Syntax
- Data Types
- Differences Between Hive and Impala Query Syntax
- Using Hue to Execute Queries
- Using the Impala Shell
Data Management
- Data Storage
- Creating Databases and Tables
- Loading Data
- Altering Databases and Tables
- Simplifying Queries with Views
- Storing Query Results
Data Storage and Performance
- Partitioning Tables
- Choosing a File Format
- Managing Metadata
- Controlling Access to Data
Relational Data Analysis with Hive and Impala
- Joining Datasets
- Common Built-In Functions
- Aggregation and Windowing
Working with Impala
- How Impala Executes Queries
- Extending Impala with User-Defined Functions
- Improving Impala Performance
Analyzing Text and Complex Data with Hive
- Complex Values in Hive
- Using Regular Expressions in Hive
- Sentiment Analysis and N-Grams
- Conclusion
Hive Optimization
- Understanding Query Performance
- Controlling Job Execution Plan
- Bucketing
- Indexing Data
Extending Hive
- SerDes
- Data Transformation with Custom Scripts
- User-Defined Functions
- Parameterized Queries
Importing Relational Data with Apache Sqoop
- Sqoop Overview
- Basic Imports and Exports
- Limiting Results
- Improving Sqoop’s Performance
- Sqoop 2
Introduction to Impala and Hive
- Introduction to Impala and Hive
- Why Use Impala and Hive?
- Comparing Hive to Traditional Databases
- Hive Use Cases
Modelling and Managing Data with Impala and Hive
- Data Storage Overview
- Creating Databases and Tables
- Loading Data into Tables
- HCatalog
- Impala Metadata Caching
Data Formats
- Selecting a File Format
- Hadoop Tool Support for File Formats
- Avro Schemas
- Using Avro with Hive and Sqoop
- Avro Schema Evolution
- Compression
Data Partitioning
- Partitioning Overview
- Partitioning in Impala and Hive
Capturing Data with Apache Flume
- What is Apache Flume?
- Basic Flume Architecture
- Flume Sources
- Flume Sinks
- Flume Channels
- Flume Configuration
Spark Basics
- What is Apache Spark?
- Using the Spark Shell
- RDDs (Resilient Distributed Datasets)
- Functional Programming in Spark
Working with RDDs in Spark
- A Closer Look at RDDs
- Key-Value Pair RDDs
- MapReduce
- Other Pair RDD Operations
Writing and Deploying Spark Applications
- Spark Applications vs. Spark Shell
- Creating the SparkContext
- Building a Spark Application (Scala and Java)
- Running a Spark Application
- The Spark Application Web UI
- Configuring Spark Properties
- Logging
Parallel Programming with Spark
- Review: Spark on a Cluster
- RDD Partitions
- Partitioning of File-based RDDs
- HDFS and Data Locality
- Executing Parallel Operations
- Stages and Tasks
Spark Caching and Persistence
- RDD Lineage
- Caching Overview
- Distributed Persistence
Common Patterns in Spark Data Processing
- Common Spark Use Cases
- Iterative Algorithms in Spark
- Graph Processing and Analysis
- Machine Learning
- Example: k-means
Preview: Spark SQL
- Spark SQL and the SQL Context
- Creating DataFrames
- Transforming and Querying DataFrames
- Saving DataFrames
- Comparing Spark SQL with Impala
Hadoop Testing
- Hadoop Application Testing
- Roles and Responsibilities of Hadoop Testing Professional
- Framework MRUnit for Testing of MapReduce Programs
- Unit Testing
- Test Execution
- Test Plan Strategy and Writing Test Cases for Testing Hadoop Application
Big Data Testing
- BigData Testing
- Unit Testing
- Integration Testing
- Functional Testing
- Non-Functional Testing
- Golden Data Set
System Testing
- Building and Set up
- Testing SetUp
- Solary Server
- Non-Functional Testing
- Longevity Testing
- Volumetric Testing
Security Testing
- Security Testing
- Non-Functional Testing
- Hadoop Cluster
- Security-Authorization RBA
- IBM Project
Automation Testing
- Query Surge Tool
Oozie
- Why Oozie
- Installation Engine
- Oozie Workflow Engine
- Oozie security
- Oozie Job Process
- Oozie terminology
- Oozie bundle
Got a question for us? Please mention it in the comments section and we will get back to you.
0 responses on "Top Big Data Hadoop Interview Questions and Answers"