Hadoop has its own file management concepts that come with many different mechanisms for data storage and retrieval. Hadoop is designed to manage large volumes of data distributed across many nodes built with commodity hardware. As such, Hadoop manages the distribution of large volumes of data using techniques designed to divide, compress, and share the data all while dealing with the possibilities of node failures and numerous processes accessing the same data simultaneously. Many of the filesystem concepts in Hadoop are exactly the same as in other systems, such as directory structures. However, other concepts, such as MapFiles and Hadoop Archive Files, are unique to Hadoop. This section covers many of the file management concepts that are unique to Hadoop.
Big Data Forensics: Learning Hadoop Investigations
Big Data Forensics: Learning Hadoop Investigations
Overview of this book
Table of Contents (15 chapters)
Big Data Forensics – Learning Hadoop Investigations
Credits
About the Author
About the Reviewers
www.PacktPub.com
Preface
Free Chapter
Starting Out with Forensic Investigations and Big Data
Understanding Hadoop Internals and Architecture
Identifying Big Data Evidence
Collecting Hadoop Distributed File System Data
Collecting Hadoop Application Data
Performing Hadoop Distributed File System Analysis
Analyzing Hadoop Application Data
Presenting Forensic Findings
Index
Customer Reviews