Hadoop is an open-source framework for distributed storage and processing of large datasets across clusters of commodity hardware. It consists of the Hadoop Distributed File System (HDFS) for storing data and the MapReduce programming model for processing data in parallel. Hadoop is commonly used for big data analytics, data processing, and batch processing of large-scale datasets.