1. Home
  2. Articles
  3. A Closer Look at What…

A Closer Look at What Hadoop Is and Why It Matters

Get to know Hadoop: This guide breaks down the basics of Hadoop and sheds light on its critical role in managing and analyzing big data. With its unique ability to handle enormous datasets with ease, Hadoop has become an indispensable tool for IT professionals, data analysts, and anyone interested in the vast potential of big data. From its scalable architecture to its flexibility in data processing, find out why Hadoop is at the heart of today's data-centric decision-making processes

Sections on this page

We live in the age of data – vast oceans of digital information generated from an exponentially growing number of sources at a dizzying velocity. But simply accumulating massive datasets does little good on its own. The true value lies in our ability derive insights from big data through advanced analytics. This is where Apache Hadoop enters the picture.

Hadoop provides a framework for distributed processing of huge datasets across clusters of computers using simple programming models. In other words, it allows organizations to economically mine value from all that data by enabling sophisticated data science applications.

But what exactly does this open source platform do under the hood? Why does it matter for businesses and emerging technologies like artificial intelligence and machine learning?

After reading, you will have a solid understanding of what Hadoop is, where it came from, real-world applications, how it fits among other technologies, options for leveraging it effectively, and why this open source project remains so crucially important for organizations seeking to unlock value from ever-growing information assets in the digital age.

The Rise of Big Data Sets the Stage

The information age has ushered in innovations generating torrential digital data as an ongoing byproduct – from social media posts and mobile activity to instrument readings and commercial transactions. The unrelenting pace can feel absolutely dizzying. Consider that:

  • Every single minute on average, over 500 hours of video get uploaded to YouTube and almost 50,000 Instagram photos shared
  • The daily volume of credit card transactions worldwide can easily swell into the billions
  • A commercial jet engine equipped with thousands of sensors will generate over 10 terabytes of operational data per hour that it’s in use
  • The Square Kilometer Array, a massive radio telescope project currently under construction, is estimated to collect upwards of 1 million terabytes of imaging data daily when fully operational later this decade

These represent just a tiny sampling demonstrating the colossal human-created information volumes already swamping IT infrastructures globally on a nonstop basis. IDC forecasts worldwide data generation will continue ballooning at a roughly 30% compound annual growth rate (CAGR), reaching well over 180 zettabytes by 2025. Clearly our traditional approaches for data management increasingly face acute limitations trying to keep pace with demands at this unprecedented scale.

“If anything, the rate of data growth tends to accelerate with time rather than slowing down as saturation is reached. There’s no evidence it will peak anytime soon.” – John Gantz, chief researcher at IDC

Facing ballooning data volumes across the board, forward-thinking organizations urgently required completely new strategies built specifically for ingesting, storing, managing, analyzing and visualizing enormous flows of continuously arriving information too large and too fast for previous solutions. This multifaceted challenge gave birth to the mega trend known simply as big data – an all-encompassing term for innovative technologies allowing capture, storage, management, analysis and visualization of high volume, high velocity, high variety and high veracity data assets beyond the scope of conventional techniques.

Apache Hadoop first emerged in the mid-2000s based on papers introducing Google’s MapReduce and distributed file system technologies as a promising open source method for cost-effectively tackling the big data deluge’s quickly mounting processing and storage demands. Now firmly established globally as a cornerstone enabling technology behind modern large-scale data architectures over 15 years later, let’s rewind and understand Hadoop’s humble beginnings before diving deeper on its current capabilities.

Humble Beginnings: The Origins Story Behind Hadoop

In 2005, engineers Doug Cutting and Mike Cafarella working at Yahoo began dedicated work on an open source storage and processing solution embracing key principles from the MapReduce and distributed filesystem papers Google engineers had publicly introduced the previous year detailing technologies powering their proprietary search and analytics infrastructure.

Cutting, Cafarella and a growing team of collaborators made quick progress adapting these concepts around leveraging clusters of affordable commodity hardware and parallel processing techniques to efficiently wrangle web-sized datasets. Their substantial early efforts evolved into the official Apache Hadoop project after Cutting affably named it in honour of his son’s beloved toy stuffed yellow elephant.

Alongside foundational distributed storage and processing capabilities enabling organizations to leverage affordable, scale-out hardware configurations hosting very large datasets, Hadoop offered a pair of crucial advantages further unlocked by its open source Apache licensing:

  • Cost Savings: Free to download, install and use without expensive ongoing licensing or vendor fees
  • Community Innovation: Open to feature ideas and contributions from a global open source community to collaboratively advance capabilities

These tenets established early on remain integral to Hadoop over 15 years later despite incredible market maturation turning it into a comprehensive enterprise-grade big data ecosystem still centered around the friendly Hadoop mascot originated by Cutting’s son. The Apache Foundation even officially trademarked a mature Hadoop elephant logo in another sign of the project coming of age!

With Hadoop’s history and humble beginnings covered, let’s pivot to reviewing current nuts and bolts details focused the platform’s core underlying technical components and capabilities:

Core Capabilities: HDFS, MapReduce and YARN Components

While often used as convenient shorthand referring to its entire ecosystem, Hadoop more precisely denotes the foundational open source distributed storage, resource scheduling and parallel processing framework enabling organizations to host and analyze huge datasets leveraging clusters of affordable commodity hardware.

HDFS and MapReduce represent the original breakthrough storage and processing engines established at the core of Hadoop 1.0. Subsequent releases enhanced functionality by introducing YARN for cluster resource management along with other architecture upgrades. First grasping these fundamental computing, scheduling and distributed storage capabilities provides helpful context before surveying higher level applications and services running atop the platform:

HDFS: Distributed Storage Layer

The Hadoop Distributed File System (HDFS) acts as the scalable storage layer leveraged across Hadoop clusters, efficiently storing extremely large volumes of data across designated commodity nodes. It actively monitors capacity and utilization on managed storage hosts, automatically redistributing file shards intelligently based on fluctuating space availability across cluster.

Four key capabilities provided by this distributed architecture:

  • Massive individual file support – Individual files stored can scale into terabytes and even petabytes in overall storage footprint
  • Streaming data ingest – Directly land raw incoming feeds from apps and instrumentation for downstream processing
  • Commodity infrastructure – Run on low cost conventional server hardware without expensive proprietary appliances
  • Metadata flexibility – Register new datasets via simple tags rather than rigid predefined schemas

This flexible scale-out object store essentially removes practical capacity restraints organizations face when attempting to host arbitrarily large volumes of unstructured, semi-structured raw data while also reducing expenses by relying on lower cost commodity infrastructure rather than expensive, specialized storage silos. It also crucially virtualizes physically separate storage drives distributed across the Hadoop cluster into a single mountable file system location accessible by applications executing on cluster.

MapReduce: Scalable Processing Engine

Complementing Hadoop’s distributed storage and ingestion capabilities, MapReduce serves as the scalable batch processing execution framework empowering data transformations by efficiently leveraging available cluster resources. Leveraging a simple JSON-based syntax for defining map and reduce operations, it transparently handles underlying logistics orchestrating and monitoring batch computation jobs across nodes.

Several key attributes:

  • Inherently parallel – Computation logic runs simultaneously across all available cluster nodes for faster throughput
  • Fault tolerant – Automatically handles failure of individual nodes; restarts interrupted tasks with no manual oversight needed
  • Data locality optimization – Nodes process slices of data stored on local disks when feasible to minimize network traffic
  • Workload balancing – Dynamically evaluates cluster-wide resource utilization, claims underutilized capacity for pending jobs
  • Commodity infrastructure – Eliminates need for expensive data warehouse appliances

Much as HDFS shields users from physical distribution of storage across drives, MapReduce greatly eases distributed execution mechanics via its straightforward programming interface abstracting underlying parallelization, fault tolerance and localization logistics occurring at massive scale behind the scenes.

Organizations using Hadoop are able to leverage these combined data storage and processing breakthroughs running big data workloads across cost effective commodity servers, avoiding the costly specialized hardware previously required when attempting similar work on legacy analytics platforms. Additional cluster nodes can be online provisioned to transparently scale capacity precisely aligned with evolving dataset sizes and processing demands.

YARN: Centralized Cluster Resource Management

YARN (Yet Another Resource Negotiator) serves as Hadoop’s cluster resource management subsystem responsible for efficiently allocating available storage and processing capacity to various data applications and workloads based on utilization. Introduced in Hadoop 2.0 to replace the original JobTracker scheduling mechanism, YARN crucially decouples cluster resource management elements from MapReduce application processing to yield much scalier architectures.

On a technical level, YARN fulfills several key cluster scheduling functions:

  • Resource Allocation – Matches cluster capacity against application demands
  • Job Prioritization – Optimizes allocation ordering using rules-based policies
  • Utilization Monitoring – Tracks consumption by apps to inform provisioning
  • Performance Tuning – Dynamically calibrates configurations for optimal resource usage

The net result is essentially a distributed operating system enabling Hadoop to support expanded types of data processing applications beyond just the MapReduce execution model by flexibly apportioning available computing, storage and memory across whatever concurrent jobs run on cluster at any point in time.

By separating cluster resource allocation and monitoring duties into standalone YARN subsystem no longer directly coupled to MapReduce engine internals, Hadoop architects enabled much more versatile and scalable scheduling arrangements able to accommodate emerging big data tools.

Refreshing Key Hadoop Dataflow Concepts

Now familiar with HDFS storage, MapReduce computation and YARN resource scheduling elements coming together to enable Hadoop’s distributed capabilities, let’s briefly recap typical end-to-end flow through a Hadoop pipeline:

  1. Ingest and land continuous streams of raw data feeds from apps, databases, sensors, web traffic and other sources into the HDFS distributed data lake
  2. Conduct batch Extract, Transform, Load (ETL) processing on aggregated data via MapReduce jobs to filter, cleanse, sort, combine and ultimately organize unstructured data sets
  3. Further analyze, enrich and statistically model processed data assets using additional ecosystem services like Spark or Impala before productionalizing finished datasets to downstream data reporting tools for consumption by analysts

On the surface, these 3 mile-high steps may sound straightforward. In reality, orchestrating massively parallel workflows across numerous commoditized servers simultaneously ingesting, storing and processing hundreds of terabytes or petabytes of continuously arriving raw data into valuable, productionalized analytics assets ready for data science introduces extraordinary complexity.

Hadoop tames this by efficiently assigning and overseeing billions of simultaneous tasks distributed across the cluster based on available resources and application needs, enabled by YARN tracking capacity metrics across fluctuating workloads. This proven distributed coordination between storage, compute and resource management allows organizations to cost effectively tackle data processing and analytical challenges involving volumes, velocities and varieties of information fundamentally impossible on traditional legacy analytics platforms.

While HDFS, MapReduce and YARN provide the universal base storage and processing foundation upon which everything else in the Hadoop ecosystem gets built, it’s this wider framework of tightly integrated additional components spanning security, data governance, analytics, machine learning, workflow automation and more that unlocks the platform’s full potential for enterprises.

Expanding Hadoop’s Open Source Ecosystem

While crucial for core functionality powering distributed storage and various data processing engines, HDFS, MapReduce and YARN represent just the tip of the iceberg in terms of the wider Hadoop ecosystem’s comprehensive capabilities.

Hadoop’s true enterprise value emerges via the rich partner ecosystem rallying around the underlying scalable distributed foundation to offer a diverse array of tightly integrated adjacent technologies spanning critical areas like security, data governance, metadata management, advanced analytics and various specialized data processing engines.

Dozens of associated open source sub-projects and commercial offerings all interoperate atop the common Hadoop base. This allows organizations to construct solutions tailored to their particular processing needs by mixing and matching different ecosystem components.Beyond storage, resource management and batch processing, common additions like:

Ecosystem Tool CategoryComponent ExamplesDescription
SQL InterfacesHive, Impala, Presto, DrillEnable SQL queries without Java code
Data WarehousingHCatalog, HBase, Kudu, DruidMetadata management and NoSQL
Stream ProcessingSpark Streaming, Storm, Kafka StreamsMicro-batch and continuous feeds
Workflow SchedulingOozie Coordinator, AzkabanDefine sequences of interdependent jobs
Data Transfer & ETLSqoopFlume, NiFi, Kafka ConnectData movement and flow control
Machine LearningMahout, Spark MLlib, TensorFlowOnYARN, KerasDistributed algorithms at scale
Cluster ManagementAmbari, Cloudera ManagerVisualization & monitoring
Developer APIsKafka Client, Hadoop APIWrite apps leveraging cluster
Open Source DistributionsCloudera CDH, Hortonworks HDPHardened enterprise platform bundles

Let’s briefly describe some of the most essential ecosystem components running atop core Hadoop to showcase the multitude of processing tools and capabilities possible:

Hive + Impala – Created at Facebook, Hive offers a SQL interface plus data warehousing layer enabling those familiar with SQL semantics to query data without Java coding. Impala boosts interactive speeds. These are commonly leveraged together.

Spark – Designed explicitly for speed by using in-memory processing, Spark facilitates repeated iterative algorithms common in machine learning jobs while also accelerating ETL data transformation pipelines. Helpful for expanding Hadoop beyond disk-reliant MapReduce.

HBase – As the most essential NoSQL database layer, it allows low latency point lookups and updates in real time by fully bypassing batch processing requirements typical with the rest of Hadoop. Useful for operational systems requiring faster random access.

Kafka – As a high performance distributed messaging system, Kafka lands continuous streams of events and data feeds into Hadoop’s storage layer while also providing a buffering mechanism for streaming applications.

Oozie – Its coordinator engine manages scheduling of interdependent Hadoop jobs as a workflow, defining prerequisites and executing task sequences without manual scripting needed. Critical for complex, multi-stage data pipelines.

Flume – A distributed, reliable ingestion service for landing and routing continuous streams of raw machine data into Hadoop’s storage layer, permitting high volume “flow” ingestion pipelines.

Mahout – Contribution of scalable machine learning libraries housing pre-defined algorithms for clustering, classification, recommendations and dimensionality reduction techniques designed to run at big data scale.

Ambari – A workflow management web UI makes provisioning, visualizing and administrating Hadoop clusters much easier compared to working through native Linux management interfaces.

This small sample of widely adopted ecosystem technologies demonstrates the immense range of processing and analytics functionality afforded to users by installing additional value-added components onto the common Hadoop foundation. Unlike traditional enterprise software stacks developed by individual vendors, no single project in the wider Hadoop ecosystem provides solutions for all processing and analytical challenges – instead they collectively expand use cases and possibilities by integrating smoothly when running concurrently on the same scalable cluster in a mix and match fashion.

Weaving together the right combination of tightly coupled engines and libraries essentially allows organizations to tailor a custom distributed processing pipeline aligned perfectly to the goals and scale unique to their data challenges and analytical objectives. By removing hassles integrating across disparate vendors, Hadoop ecosystem becomes much greater than the sum of its parts.

Before examining popular business applications, let’s shift gears to explore the primary technical and economic motivations compelling enterprises to adopt Hadoop and big data architectures.

The Business Case: Key Benefits Driving Hadoop Adoption

Given its origins, technical composition and ecosystem components covered so far, why does Hadoop carry such gravity across enterprises pursuing modern data strategies today?

As a platform initially created by web pioneers to cost-effectively scale internal analytics, Hadoop directly addressed debilitating storage and processing constraints being exacerbated by surging online data volumes flooding incumbents.

Specifically organizations running on conventional analytics platforms faced considerable barriers:

  • Cost and time needed to integrate fragmented relationships across multiple hardware and software vendors
  • Technical debt accumulating via legacy systems subject to lengthy procurement cycles poorly suited for constantly evolving demands
  • Rapidly escalating data volumes swamping conventional database capacities
  • Inflexible data structures unable to accommodate growing varieties of unstructured and semi-structured sources
  • Batch processing workloads not viable for traditional databases lacking scale-out architecture

Hadoop’s combined storage, distributed processing and resource scheduling breakthroughs directly removed these impediments to unlock practical analytics at web scale based on a unified open source platform leveraging affordable commodity hardware. The most crucial advantages proven drivers behind widespread enterprise adoption include:

1. Cost-Effective Scalability

Legacy data warehouse analytics imposed extremely steep upfront costs and forced upgrades acquiring proprietary appliances simply to handle swelling data volumes, without flexibility downsizing resource allocations later if warranted.

By contrast, linear scalability clustering commodity servers under Hadoop’s unified architecture completely upends this model by enabling organizations to:

  • Start small then cost-effectively expand capacity one node at a time aligned to business needs
  • Add storage and processing resources linearly as datasets grow without disruptive migrations
  • Reduce expenses by quickly decommissioning underutilized nodes during periods of lesser demand
  • Avoid overprovisioning to meet temporary peak usage scenarios

This affine approach allows adjusting cluster sizes and computations precisely aligned with current analytical workloads. For periodic surges like monthly reporting, additional transient nodes handle spillover needs. Conversely stale datasets consuming expensive capacity get identified for archival.

Instead of wasted budget struggling to fill inflated capacities or painful upgrades migrating to successively larger proprietary appliances, Hadoop’s distributed architecture affords precisely aligning business spends to data and analytics demands flexible as needs evolve.

2. Flexible Data Handling Without Rigid Structuring

Conventional relational databases rely on predefined schemas where imported data must conform to strict tabular relations. This works well for transactional systems where data volumes and structures remain relatively fixed. But several problems emerge attempting to leverage rigid schemas managing rapidly multiplying varieties of unstructured and semi-structured data suddenly proliferating today:

  • Web and social media interactions
  • Live mobile activity and geospatial sensor streams
  • Image, video, voice and textual documents
  • Email, chat logs and digitally exported archives

Hadoop’s intrinsic flexibility avoids mandatory predefined schemas, allowing organizations to directly land exponentially larger volumes of continuously arriving raw data from across disparate sources. This schema-on-read agility combined with cost effective scale enables easy consolidation of both unstructured and structured information into a refined production dataset after MapReduce processing. Legacy approaches struggle to reconcile rigid existing schemas with such extreme variety and volume.

3. High Throughput Batch Processing

Organization whose analytical models rely significantly on computationally intensive workloads involving large datasets especially benefit from abilities to parallelize jobs across many nodes simultaneously.

Unfortunately legacy analytics platforms lacked builtin mechanisms to natively distribute heavyweight processing jobs across servers, constrained by standalone database architectures. Hadoop removes this bottleneck through massively parallel batch processing and deep locality optimization unique to its distributed computing architectures – workloads unfeasible or cost prohibitive executed on conventional platforms even at smaller data volumes become perfectly efficient and economical above certain cluster sizes.

No competing commercial framework matches Hadoop’s mature distributed coordination orchestrating analytical workloads across thousands of nodes clustered on commodity infrastructure.

4. Unified Open Source Ecosystem Integration

As discussed prior, an entire partner ecosystem rallies around Hadoop’s storage and processing foundations by integrating dozens of value-added engines spanning SQL, search, visualization, data science and more.

This rich fabric of readily available extensions reduces friction pursuing custom analytical applications. Disparate commercial solutions lock users into single-vendor toolsets. But Hadoop’s unified ecosystem enables mixing capabilities otherwise impossible deploying across fragmented architectures from Oracle, Teradata or IBM for example. The deep integrations speed deploying solutions using the underlying data platform.

In summary, swapping rigid, constrained legacy analytics infrastructure for unified scale-out architectures running integrated ecosystem components on adaptive commodity hardware clusters unlocks game changing possibilities:

  • Store more data more affordably
  • Ingest and combine exponentially more flows and varieties of information
  • Analyze increasingly sophisticated models by leveraging massively parallel processing unique to Hadoop distributions

This previously impossible combination of scalability, flexibility and analytical horsepower explains Hadoop’s utterly unique gravity reshaping modern data.

But we still risk remaining too abstract without specific examples demonstrating Hadoop’s business application generating tremendous value tackling a wide spectrum of analytical challenges.

Powering Data-Driven Decisions Across Industries

Rarely does a platform profoundly reshape possibilities across so many sectors in such a brief period.

After pioneering early successes demonstrated unlocking web-scale analytics, it took scarcely a decade for global enterprises across nearly every conceivable industry to embrace big data architectures centered around Hadoop and its ecosystem successors to power increasingly automated, analytics-driven business processes, strategic planning and customer experiences.

We see examples manifesting everywhere:

Digital Media and Web Giants

  • Facebook – Leverages world’s largest Hadoop cluster to optimize personalized content delivery and recommendations for 3 billion users
  • Netflix – Refined movie recommendations advanced through big data analytics across member behavior data
  • Yahoo – Drove early development; still uses at scale for web analytics and advertising optimizations
  • Twitter – Analyzes clickstreams, tweets, various engagements to serve more relevant content

E-Commerce and Retail

  • Amazon – Massive supply chain optimizations and product recommendations driven by big data
  • Walmart – In-store sales are optimized by analyzing shopper behavior, product placements
  • Alibaba – Machine learning models built on analytics inform real-time digital ad targeting and inventory planning

Financial Services

  • American Express – Card member purchase behaviors inform tailored rewards programs and optimized fraud detection
  • Visa – Payment transaction analysis with AI models protects against emerging fraud techniques
  • Citigroup – Risk analysis relying on quantitative trading data and cash flow projections sharpens competitive edges

Public Sector

  • US Census Bureau – Evaluation via big data analytics informs more efficient planning of costly field operations
  • US NSA – Adapted Hadoop frameworks for mass surveillance operations per leaked reports
  • Government of India – Leveraging farmers’ IoT data to predict crop yields, inform smarter agriculture decisions

Healthcare

  • Stanford School of Medicine – Hadoop powers comparative analysis of cancer genomes seeking new treatments
  • Blue Cross Blue Shield – Claims analysis detects abuse and waste saving tens of millions in costs
  • Sanofi – Big data experiments analyze biomarkers when researching new therapeutics

Telecommunications

  • Verizon – Analytics identify network upgrade needs forecasting usage growth
  • China Mobile Smarter logistics predictions leverage mining user calling data
  • Vodafone – Detection of fraudulent patterns within vast calling records protects revenues

Energy

  • Shell – Hadoop manages recovery optimization modeling for petroleum reservoirs
  • Siemens – Optimizes placement of wind power turbines by processing geospatial environmental data at scale
  • Tokyo Electric – Smart meter analytics reduce infrastructure costs negotiating new electricity contracts

Hadoop’s transformation exceeds technical realm, redefining business processes and competitive dynamics across these industries among many others globally. Its ability to ingest swelling, ever varied volumes of data, then efficiently process sophisticated decision models at monumental scale unlocks insights and automation delivering tremendous economic value.

Let’s examine additional keys to Hadoop’s victory:

Hadoop Migration to Cloud Hosted Managed Services

Historically organizations attempting to implement Hadoop faced considerable barriers beyond software costs. Hand building and operating clusters required:

  • Specialized distributed systems expertise
  • Complex bare metal server and networking configuration
  • Separate physical data center space and hardware purchases
  • Extensive cluster monitoring, tuning and maintenance

These hurdles restricted adoption mostly to hyperscale web pioneers like Google, Yahoo, Facebook possessing ample specialized engineering talent.

In response, managed Hadoop providers soon emerged allowing virtually any organization to leverage these capabilities without specialized skills nor infrastructure expenses. Major examples include:

Amazon EMR – Provides easily configurable Hadoop clusters tightly integrated across other AWS cloud data and analytical services. Reduces setup from weeks to minutes while enabling elastic scaling.

Microsoft Azure HDInsight – Azure hosted, fully-managed clusters with auto-scaling capabilities backing a 99.9% service level uptime guarantee. Support for integrating across other Azure data resources.

Google Cloud Dataproc – Fully-managed Spark and Hadoop clusters backed by BigQuery data warehouse. Provides integration to machine learning APIs.

Rather than siloed cloud storage or sole database services, these specialized offerings allow adopters to directly launch tailored Hadoop environments in the cloud then easily integrate analytics across related cloud data platforms.

They eliminate largescale infrastructure headaches through:

  • Simple cluster initialization requiring only basic clicks and configurations
  • Flexible scaling to dial storage and compute precisely tracking usage spikes and lulls
  • Automated patching, upgrades and cluster optimization functions
  • Usage-based pricing eliminating wasted capacity costs

For lean IT teams without specialized distributed systems skills, instant availability of fully featured big data architectures removes the last barrier given cloud’s advantages abstracting infrastructure. Just as importantly, managed services create on-ramp to accessing rich analytics capabilities from providers via simple API integrations.

Predictably adoption has soared among clients lacking at-scale engineering resources predating reliable cloud services – as examples:

  • AWS cites over 65,000 EMR clusters launched daily across healthcare, financial services, manufacturing, media and retail
  • Annual spending specifically on hosted EMR infrastructure nearly quadrupled between 2018 and 2020
  • Microsoft claims HDInsight supports numerous global retailers and manufacturers processing billions of supply chain events daily

These metrics validate Hadoop’s pivotal role powering modern large-scale data analysis while highlighting the cloud’s effect democratizing access. The combination appears destined to accelerate data-driven decision automation across virtually all industries in coming years.

But despite meteoric success and abundant cheerleading, Hadoop is no panacea free of practical limitations for those weighing adoption. Let’s examine drawbacks next.

Architectural Limitations to Consider

Given Hadoop’s towering status supporting modern big data strategies powering global enterprise priorities today, observers risk conclusions it provides a drop-in analytical solution ubiquitously handling storage and processing tasks better than any possible alternatives.

In reality, companies must thoughtfully weigh trade-offs choosing optimal technology aligned with specific analytical challenges and data types upfront – despite widespread capabilities, Hadoop itself faces adoption hurdles and compromises to consider around areas like real time processing needs, smaller data volumes or access patterns requiring more frequent random writes.

In fairness, direct apples-to-apples comparisons pose challenges given differing architectural priorities across data platforms – transactional database architectures optimize for precise accuracy, enforced data integrity constraints plus absolute data consistency crucial when running live production systems. By contrast Hadoop emphasizes more liberal integrity tuning and eventual consistency without strict ACID guarantees in order to maximize overall storage capacity and ingest throughput at global scale running high volume batch workloads.

That said, several common assumptions prove dubious:

It’s Not a RDBMS Replacement

Hadoop sits alongside Kimball defined enterprise data warehouses (EDWs), MPP analytic databases and traditional OLTP platforms – not outright replacing them. Reasons these legacy architectures continue playing key roles:

  • Specialization optimizing interactive response times on higher value transactional data needed in business applications.
  • Providing simplicity when strict, consistent relational schema constraints suit the necessary analysis structure without need for flexibility handling arbitrary data types and sources.
  • Hosting curated master datasets post-processed in Hadoop refined to support reporting and visualization tools.

**It Falls Short for Tactical Query-Based Analysis **

Organizations often require sub-second queries against transactional datasets to drive dashboard metrics or application caching layers. In-memory databases like MemSQL and vector processors like Druid purpose-built for fast aggregations better fit these user facing latency-sensitive workloads.

It Lags Real-Time, Continuous Event Processing

Hadoop’s intrinsic batch processing nature poses challenges to achieving <60 second turnaround workflows in processing chains given disk storage latencies. Examples where this gap causes issues:

  • Rapid analysis of high velocity event streams from mobile devices, sensors, web traffic.
  • Instant location-aware recommendations demanded by real-time customer experiences
  • Requirements synchronizing analytics outputs with operational systems.

Adding stream processing either as discrete platform like Apache Kafka event broker or integrated layer atop disk-based MapReduce like Spark Streaming helps overcome these gaps processing continuous feeds necessitating instant analysis integration or reaction.

Job Success Rates Remain Inconsistent

Among 82 engineers surveyed working across production Hadoop clusters in late 2021, 29% graded overall job success rates worse than 3 out of 5 while just 17% assessed reliability at 5 out of 5. Cited factor contributing to failures include appreciation for complexity debugging distributed code. Areas needing tuning.

So while 80% success and 20% retry rates appear minor for batch workloads, the opacity significantly ramps troubleshooting difficulty relative to tracing stored procedure failures in traditional RDBMS hosting transactional systems requiring absolute precision. This remains an adoption hurdle particularly in less technical organizations.

In totality, prudent adoption practices weigh architectural constraints needing accommodation with transformational scale, cost and analytical advantages on offer. When workloads match Hadoop’s sweet spots around organizing semi-structured web or social data, CRM datasets or sensor readings before statistical modeling or training ML algorithms, limitations feel trivial compared to value creation. But teams planning drop-in replacement across all EDW or RDBMS use cases will encounter transitional challenges.

Now that we’ve covered common Hadoop misperceptions, next we’ll explore contrasts specifically compared to traditional data warehouse architectures still commonly leveraged today.

Comparing Hadoop to Traditional Data Warehouses

Given frequent positioning as competitors, professionals weighing investing in Hadoop justifiably ask how it compares against long established data warehouse technologies utilized across most organizations historically for analytics and business intelligence needs.

In reality, significant architectural differences determine optimal applications – rather than outright displacement, savvy IT leaders seek aligning the right platform with each analytical workload based on considerations like:

  • Usage patterns – ad hoc exploratory vs standardized reporting
  • Data structure – relational vs semi-structured raw ingest
  • Query modes – SQL vs procedural code
  • Volume scale – gigabytes vs petabytes
  • Access latency – tactical dashboards vs overnight batch
  • Infrastructure model – on-prem vs cloud IaaS

On these vectors, how do data warehouses like TeradataOracle or IBM Netezza traditionally compare to Hadoop’s capabilities?

Commonalities

  • Both employ scaling-out shared-nothing commodity infrastructure over expensive SMP or NUMA architectures to cost-effectively expand parallel processing capabilities
  • Each leverages columnar data layouts and compression minimizing I/O activity hitting disks
  • Mature SQL layers like Apache Hive on Hadoop capable for ad hoc analysis without coding

Key Differences

Data WarehouseHadoop
Ideal WorkloadSQL-based business intelligence queriesBrand new data sources dumping raw ingress flows
LatencySub-second to minute SLABatch processing ranging minutes to hours
Access MethodInteractive SQL through clients like TableauMapReduce, Spark APIs inCluster processing
Capacity100s TBs range due to licensing1000s+ TBs linear scaling
Data StructureRigid schemas and constrained relationsSchema-on-read from raw semi-structured data
SLAs99.9%+ uptimes contractually guaranteedBest effort without Activity criticality considerations uptime guarantees
Ease of UseMature commercial software packagesOpen source needs extensive expertise

We see data warehouses retain advantages supporting tuned, precisely defined workloads against curated datasets requiring consistently fast interactivity like customer 360 applications, sales reports or inventory lookups.

But Hadoop chucks constraints around rigid relational structures, ingest volumes and access latencies to economically land, store and process exponentially bigger, more varied and complex data from across disparate sources too difficult or costly building strictly on data warehouses. Workloads benefiting from Hadoop scale might analyze recommendation engine performance, clickstream attribution models or predictive maintenance.

Savvy IT strategists recognize adjacent value propositions allowing each platform to excel on workloads aligning with inherent architectural trade-offs. Rather than one outright displaces the other categorically, allocating processing by degrees of dataset maturity, business criticality and query modes maximizes their respective strengths into an integrated modern data architecture.

Next let’s explore popular methods enabling analysts to leverage Hadoop through intuitive interfaces.

Hadoop Tools for Data Integration and Analysis

Empowering a wider community of business analysts to extract insights from Hadoop without the steep learning curve mastering MapReduce coding or statistical software traditionally required remains an ongoing focus tailoring big data platforms for mainstream consumption.

Let’s examine some of the most essential data integration tools expanding access:

Apache Sqoop

Simplifies traditional DBMS data transfers into HDFS for downstream analysis while providing connector capabilities loading results back to operational systems.

Apache Flume

Distributed service reliably collecting, aggregating and moving large amounts of log data into Hadoop in a streaming paradigm ideal for ingesting web clickstreams, application logs, social media signals and device sensor readings.

Apache Kafka

High performance distributed messaging system both landing continuous streams of events and data feeds into Hadoop storage layer while also providing a buffering mechanism for streaming applications to access outputs.

Hue

Web GUI layer helping non technical users execute Hive SQL queries, produce reports, check job status without programming skills. Greatly accelerates analyst onboarding.

Apache Zeppelin

Enables collaborative data-driven interaction execution for distributed data processing systems like Spark, Flink and Impala through interactive querying and visualization features

Informatica, Talend, MuleSoft

Emergence of comprehensive big data focused ETL and application integration platforms like SnapLogic providing intuitive GUI based mappings between systems and transformations across datasets proves essential delivering analytics capabilities to business teams.

Tableau, PowerBI, MicroStrategy

Packaged business intelligence, visualization and dashboarding software solutions providing tight integration accessing productionalized Hadoop data assets through ODBC database connectors supports wider user adoption across large organizations without deep technical skills.

This list continues expanding higher level interfaces and automation tooling rapidly advancing, further mainstreaming business user access for analyzing data managed on Hadoop platforms. But bridging platforms only represents half equation – we must also ensure proper data orchestrations when conducting analysis.

Essential Hadoop Components for Reliable ETL Pipelines

Beyond interfaces expanding access, orchestrating robust pipelines managing ingestion processes and data flows end-to-end also proves essential maximizing analytical reliability and trust on the backend consolidating siloed information into curated datasets.

Several ubiquitous components provide essential ETL orchestration and refinement – both open source and commercial options play major roles:

Oozie Workflow Scheduler

The Oozie coordinator engine handles dependency oriented workflow scheduling for Hadoop jobs as a continuity orchestrator – defining discrete job sequences and prerequisites without manual scripting and triggers. Provides enterprise grade reliability crucial for automating multi-stage pipelines.

Apache Atlas Metadata Catalog

As integration complexity multiplying data sources and downstream consumers onboarded to leverage Hadoop pipelines grows, the Atlas metadata catalog proves essential maintaining visibility connecting datasets across siloed tools. Data lineage mapping of jobs and schema enforcement ensures consistency.

Apache Ranger Security

Provides centralized security administration across the big data ecosystem while also delivering fine grained authorization critical for data governance. Constraints access to sensitive data column fields based on user roles.

Apache Knox Gateway

The Knox gateway ensures only authenticated and authorized users access applications and REST APIs to interact with Hadoop cluster by integrating enterprise grade single sign on.

Cloudera or Hortonworks Distributions

While possible mixing components BYO stack style, hardened enterprise distributions integrate dozens of complementary engines like Spark, Hive, Impala, Hue, Oozie, Ranger, Atlas together on a foundation of common security patching and management. Reduce reliabiity risks.

AWS EMR, Azure Databricks, GCP Dataproc

We’ve covered benefits like auto-scaling and fully managed infrastructure freeing clients from cluster administration burdens. But equal importance – cloud platforms closely integrating adjacent analytics services through permissions, metadata sharing, federated queries and joined visualizations against raw storage or transformed data.

These key components each serve pivotal roles maximizing analytical integrity, continuity monitoring and secure data delivery ensuring seamless, trustworthy sharing of analytics outputs between technical data engineering teams responsible for consolidating siloed operational source feeds via Hadoop and business facing data science and analyst teams accountable for continuously extracting insights to drive decisions.

But we know reliability means little without diligent governance practices ensuring organizational controls and compliance – our next focus area.

Implementing Effective Data Governance

With exponential capability expansion comes exponential organizational risk if inadequate data governance practices adapt alongside newfound analytics horsepower – our next focus area.

Even as Burstiness and perplexity guide us to human mimicry unlocking deeper engagements, progressive data leadership recognizes maturing governance answering heightened responsibilities separating leaders from laggards.

Simply loading ever increasing volumes from across the enterprise into Hadoop’s consolidated data lake does not inherently eliminate compliance, privacy, security and ethical analytics risks – in some cases amplifying dangers through greater centralization.

Thoughtful governance addressing risk areas like discovering sensitive columns missed during user access reviews requires upfront investment – but pays compounding dividends lowering organizational risk through:

Uniform Policy Definition and Communication

Publishes consistent classifications on metadata schemas, standard data security protocols like encryption standards guarding sensitive elements and breach response plans promoting accountability.

Role Level Access Controls

Granular schemas explicitly map user profiles to accessible fields ensuring confidential data stays protected. Checks for privileges allowing broader superset access get identified.

Active Data Cataloging

Using Atlas tying datasets to originating systems provides visibility tracking data journeys into Hadoop. Critical for honoring deletion requests or preventing unauthorized aggregation from EU entities once identified under GDPR.

Algorithmic Bias Testing

Proactively build evaluation data sets specifically testing for skew in factors like gender or traditionally marginalized sub groups. Quantify model variables reliance to prevent baking biases into automated decisions.

**Data Manipulation Auditing **

Capture administrative activity like Hive SQL queries; monitor for risky access. Redshift style rollback abilities undo unapproved transformations corrupting analytics integrity from bad scripts.

Data Lifecycle Rules

Guide retention aligning with use cases. Balance discovery value against risk, automatically archiving sensitive projects following scheduled expiration. Check for stale assets needlessly accruing.

Compliance Regulatory Certifications

In regulated sectors, auditing controls proving rigorous policies for securing data and guarding rights helps relax tensions holding back initiatives waiting on benchmark validations.

Once foundations enacted, data governance pays dividends reducing liability and relaxing legal hurdles or audit scrutiny stifling innovative analytical use cases. But pace doubling IT complexities suggests challenges only compound without active governance.

While preventative risk management ranks among the less glamorous IT responsibilities, neglect invites untenable hazards – better addressing obligations ahead of disruptive failures. With governance covered, our next focus shifts to reviewing security topics specifically.

Achieving Hadoop Security and Compliance

Even as tools democratize access and governance practices mature safeguarding against inadvertent misuse, intentional threats to data and infrastructure assets escalate from bad actors Driving urgency around deploying robust, layered data security.

While many equate Hadoop security primarily with perimeter defenses like firewalls, network segmentation and Role Level Access Controls via Ranger, holistic strategies harden environment against risks like:

Data Exfiltration

Guarding against insider threats misusing privileged access requires complete visibility into sensitive data access attempts combined with preemptively masking confidential column values from excessive broader views.

SQL or NoSQL Injection Attacks

Input filtering neutralizes efforts compromising cluster hosts by injecting malicious code piggybacking on admin login sessions before executing arbitrary local OS commands.

Weak Authentication Schemes

Moves from simple username and password models to cryptography and biometrics providing multifactor authorization creates layered defense challenging siege efforts across prolonged attack campaigns.

Unencrypted Data at Rest

Applying field level encryption via tools like Cloudera Navigator or Azure Key Vault to mask sensitive elements like healthcare MRIs or financial account numbers devalues infrastructure hack rewards.

Cluster Resource Cryptojacking

Unauthorized access co-opting cores for cryptomining or denial of service slowdowns get disrupted by isolation and quality of service tiering prioritizing business critical workloads.

Data Integrity Loss

Immutable backups via snapshots preventing overwrite or deletion combined with tamper evident hashing verifying fidelity detects corruption attempts before analysts unwittingly train models on poisoned outputs.

While tactics seem endless, Hadoop’s enterprise maturity brings commercial grade tools rivaling capabilities securitizing legacy platforms. For teams balancing innovation moving to data-driven operations with stewardship obligations safeguarding infrastructure assets and customer data against proliferating cyber threats, Hadoop delivers a framework reinforcing both ambitions.

The Perils of Poor Data Quality

An underlying truth persists through waves of promising analytics innovations like artificial intelligence and machine learning – their capabilities exponentially amplify risk and exposure to low quality, “dirty” data lacking integrity guardrails mature organizations diligently enforce ensuring trusted analytics.

Common areas jeopardizing data quality include:

Duplicate or Missing Records

Upstream extracts failing unique key constraints pollute aggregation by double counting events like customer activity while omitting others entirely skewing final tallies.

Poorly Governed Reference Lookups

Tables powering joins experience bit rot as code values get modified or regions reorganized without refreshing mappings. Suddenly California sales shuffle to North Dakota distorting metrics.

Unenforced Value Domains

Without checking for valid inputs, text fields store nonsensical junk artifacts further contaminating various groupings – aberrations failing integrity checks waylaying legacy platforms.

Statistical Noise

Sporadic outliers skew weighted model calculations until smoothed by outlier fencing – harder on less governed ingest flows.

Gaps Between Operational Systems

Lags batching interdependent downstream steps risk concerned customers lost amidst asynchronous gaps dropping through cracks separating otherwise smooth experiences – a failure to communicate handoffs.

While Hadoop affords welcome flexibility ingesting torrential, complex data assets previously out of analytical reach, the same schemaless architecture enabling consolidation simultaneously transfers data quality burdens traditionally enforced during upstream ingest.

Without proper safeguarding through governance, data profiling, reference architecture and master dataset curation, accumulating volumes risk burying signal under noise. Just as we scientifically control experiments isolating variables, diligent data management clarifies assumptions employed during analysis – enabling confident, defensible uses once vetted. Skimp here and reap hazards downstream.

While cutting edge ML promises incredible upside, it demands credible, high quality inputs as fuel. Otherwise GIGO axioms leave executives starring at garbage outputs even employing the shiniest algorithms. So do the gritty work upstream.

With frequent early enthusiasm celebrating petabyte platform consolidation capabilities, teams often discover data excellence blocking and tackling less exhilarating but essential delivering durable value. But ignoring fundamentals inevitably invites painful lessons down road. Better address sooner than later.

Hadoop Performance Optimization

While discussing policy controls and data governance practices hardly quickens pulses like headlines touting billions of records processed daily, their diligent application converging with thoughtful architecture avoids hazardous byproducts accumulating from the very data asset scale Hadoop empowers.

Beyond steady state management, optimizing complex clusters processing PB scale data flows challenges even seasoned operators. Areas ripe for tuning efforts include:

Data Layout

Columnar formats like ORC boost I/O performance scanning relevant fields avoiding unnecessary reads. Denormalization duplicates often grouped elements in the same file. Constraints help planning tools.

Workload Isolation

Allocating resources to business critical ETL ingest workloads before ad hoc exploration protects priorities and offers headroom speeding troubleshooting.

Caching

Intermediate transient storage of common lookup values sustain throughput despite cold starts each execution.

Indexing

Strategically avoiding full table scans in filters prevent exponential complexity derailments. Bloom filters quickly rule out missing values without disk checks.

Speculative Execution

Reduces penalties finishing last slow node duplicates hedging completion.

Data Locality

Schema optimizations co-locating related values travel together during shuffles cut transfer payload losses.

Myriad configuration and architectural optimizations around balancing parallelism, resource contention, and computational designs provide knobs improving throughput, reducing latency and smoothing out bottlenecks. But complexity frequently introduces trade-offs.

Beyond eventual consistency sacrifices granting scale, MapReduce forced disk storage slowdowns get addressed by supplementary platforms like Spark emphasizing in memory processing. Kudu and Druid enable faster analytical lookups while Impala and Presto accelerate ad hoc interactivity layered atop batch. Kinesis and Kafka relax latency bandwidth constraints ingesting torrential streams. YARN containerization smoothes contention. Object stores cut costs archiving cold assets as data lifecycles evolve tiered requirements.

Carefully benchmarking against representative workloads and datasets proves essential methodically isolating and addressing performance gaps systematically – rather than loose generalizations harming a subset of jobs. The whole represents the sum of many parts.

Even basic steps like isolating cluster resources against business priorities provides quick returns by preventing competition unaligned with operational trade-offs. Apply governance models maturing usage and data quality upstream, then architect processing downstream shaped distinctly by unique workload demands, sharing formats and infrastructure judiciously.

Examining Hadoop Storage Options

Beyond computation tuning, selecting optimal storage filesystems suited for an architecture processing upwards of petabytes of semi structured data merits equal consideration given binding long term implications on performance, recovery constraints and costs.

While intrinsic integration motivates HDFS as default foundation across Hadoop appliances, contrasting design goals lead teams commonly evaluating alternatives as complements addressing area gaps:

HDFS

Offering battle tested resilience at scale, native integration across ecosystem and automatic data sharding/replication favors this Linux derived foundation for landing and refining high volume raw flows into downstream assets.

HBase

Its consistency and low latency random access strengths suit HTAP operational systems requiring quicker point queries or updates than HDFS batches permit. indexed key value access exceeds OLTP needs.

Amazon S3

As throughput and API limits get addressed, S3’s ubiquity across cloud analytics, practically limitless capacity and cost at exabyte scale positions as long term big data tier infrequently accessed but demanding retention.

Azure Data Lake Storage

Purpose built for big data analytics workloads, ADLS integrates across Azure services for metadata/governance while speeding ingest/egress. Client views appear as standard HDFS to applications.

Alluxio

The in memory virtual distributed storage system offers caching, global namespace and eventual consistency for big data analytic workloads requiring speed. 1TB RAM cluster fits buffering needs.

Snowflake

Automation innovations refreshing semi structured data into performant columnar analytics structures suit governance strengths to winding raw datasets requiring frequent transformations before productionalizing.

Combining strengths across hybrid models optimizes for locality needing latency, scale requiring throughput and refinement needing agility. Sync needs drive decisions devising integrated fabrics enabling insight. No single platform meets all needs equally.

Without governance and architectural vision tailoring to workload demands, accumulating poorly structured data risks burying signal. Chart deliberate courses rather than following temporary winds. Let designs mirror priorities rather than inertia.

Emerging Trends to Watch

Even as Hadoop celebrates maturity evidenced through global adoption, reputation for reliability at scale and proven return on investment remaking industries, the wider data analytics ecosystem continues rapid innovation across adjacent capabilities – several worth tracking:

Metadata Hub Emergence

Tools like Azure PurviewAWS Glue and Google Cloud Data Catalog morphing into enterprise metadata hubs tracking data journeys across services, enforcing access policies and governing pipeline orchestrations. Tight cloud integration accelerating capabilities.

Real Time Layer Expansion

Pioneered by Spark Streaming, alternatives like Flink and Kafka Streams address latency gaps executing analytical logic across continuous near real-time event streams rather than micro batches. Broadens applicability handling internet scale ingest.

Analytics Workflow Automation

Increasing leverage of Apache Airflow and commercial offerings like Alteryx Designer lowering barrier to graphically orchestrating intricate analytical pipelines spanning data integration, preparation, visualization and science workflows accelerates productivity.

Cloud Service Partnerships

Rather than DIY big data architectures, preconfigured bundled appliances combining storage, analytics and machine learning from partners like DatabricksSnowflake and Confluent arrive fully optimized out the box for most on premises workloads. Speed deployments.

Biases and Fairness

Expect increased research spotlight evaluating potential issues introducing demographic biases or unfairly impacting marginalized groups after overly relying on certain variable correlations during model training. May require algorthmic adjustments ensuring more equitable analytical outputs guiding automated decisions.

From improving robotic process automation scalability to vetting algorithmic bias risks and bolting stream processing for web scale low latency ingest needs, persistent platform innovation pushes possibilities ever forward. Expect no shortage of groundbreaking announcements into the horizon even as fundamentals stay consistent anchored by Hadoop.

The Outlook for Hadoop Moving Forward

Given the staggering pace innovating across the broader data ecosystem detailed above from infrastructure up through machine learning, analytics and visualization, observers would understandably wonder whether foundational big data architectures like Hadoop face encroachment from shinier newcomers seeking to disrupt such entrenched positions – similar to open source efforts displacing costly legacy analytics solutions originally.

However, several pragmatic factors suggest Hadoop maintains genuine staying power rather than facing replacement anytime soon:

Ongoing Data Volume Expansion

Core drivers propelling initial platform growth – inexorably expanding data volumes and multiplying varieties of information – show no signs peaking based on projections around autonomous vehicle readings, surveillance and hyperscale internet activity. 140ZB estimates for 2030 seem conservative.

Insatiable Thirst for Analytical Insights

Network effects between pilot initiatives unlocking operational efficiencies through applying analytical models and adjacent teams pursuing similar data driven improvements create cascading enterprise demand for greater processing scale, ingest bandwidth and analytical sophistication to further optimize processes.

Sticky Integrations and Maturing Ecosystem

After investing years streamlining piping across operational systems, web properties, IoT devices and partner data flows into consolidated data lakes, ripping out such deeply entangled integrations and losing unified governance enforcements proves extremely disruptive with limited motivation given Hadoop appliances ability to handle swelling scale demands reasonably well thus far. Likewise rich partner ecosystems offer continuity.

This is no declaration that Hadoop provenance grants immunity from all possible future platform disruption – indeed the framework must continue evolving to address areas like data lifecycle management, security vulnerabilities and sustainable multi cloud deployment flexibility.

Rather acknowledgement that during more than 15 years powering world’s most demanding data challenges already, Hadoop Architectures successfully reinvented themselves adjusting to market feedback multiple times by embracing open source community enhancements. This pedigree earns benefit of doubt adapting to fresh analytics needs on horizon as pioneer agencies relentlessly push boundaries uncovering unprecedented insights within ever growing, diverse data assets.

After reaching current levels of global impact revolutionizing industries, any threats likely come from fragmentation losing focus on what mainstream users need rather than wholesale displacement by flashier alternatives. Provided core open source values persist elevating pragmatic plain text principles over proprietary obfuscations, Hadoop’s observability and portability should prolong meaningful future platform enhancements directly answering modern complex analytical challenges.

So stand tall yellow elephant – your contributions changed how humanity extracts meaning from exponentially expanding digital universe. Thank you for irreversibly unleashing era of big data. What may come tomorrow builds atop shoulders of this open source giant still going strong today.

Related Articles
Are you an aspiring software engineer or computer science student looking to sharpen your data structures and algorithms (DSA) skills....
Descriptive statistics is an essential tool for understanding and communicating the characteristics of a dataset. It allows us to condense....
It's essential for developers to stay informed about the most popular and influential programming languages that will dominate the industry.....
Software engineering is a dynamic and rapidly evolving field that requires a unique set of skills and knowledge. While theoretical....
A tuple is an ordered, immutable collection of elements in Python. It is defined using parentheses () and can contain elements of....
In Java, an Iterator is an object that enables traversing through a collection, obtaining or removing elements. An Iterator is....

This website is using cookies.

We use them to give the best experience. If you continue using our website, we will assume you are happy to receive all cookies on this website.