AWS · DEA-C01

Operational Complexity Underestimation — AWS Data Engineer (DEA-C01)

The answer is correct but operationally expensive. The exam prefers managed services over self-managed when both meet functional requirements.

Flexible pipelines hide a coordination tax

Architecture requirement: ingest, transform, and route streaming records with custom business logic. Competing choices: managed Kinesis Data Firehose transformation versus a custom Lambda fan-out wired to multiple SQS queues with a self-managed Glue job downstream. The deciding constraint is failure-recovery burden. The exam penalizes solutions that introduce retry logic, DLQ management, and orchestration overhead when the scenario provides no evidence the team can sustain it.

34%of exam questions affected (68 of 200)

The Scenario

A team of 3 developers needs to run a containerized application with auto-scaling. You recommend Kubernetes on EC2 with kops for cluster management. The correct answer is ECS on Fargate. The scenario said "small team" and "minimize operational burden." Self-managed Kubernetes requires managing the control plane (etcd backups, API server upgrades, certificate rotation), node group updates, CNI plugin configuration, and ingress controller maintenance. ECS on Fargate eliminates all of that — AWS manages compute, scaling, and patching. The trade-off is less customization, but the scenario never asked for Kubernetes-specific features like custom operators or CRDs.

How to Spot It

  • "Minimize operational overhead," "small team," "reduce management burden" — these phrases are signals to choose the most managed option. ECS Fargate over EKS self-managed nodes. Aurora over self-managed PostgreSQL on EC2. Lambda over always-on containers for event-driven workloads.
  • EKS managed node groups reduce operational burden compared to self-managed nodes, but you still manage node AMI updates, pod scaling, and cluster upgrades. EKS with Fargate eliminates node management entirely but loses DaemonSet support and some storage options. The exam tests these operational trade-offs at each level.
  • Self-managed options (EC2, EKS self-managed, self-hosted databases, self-managed Kafka) are only correct when the scenario explicitly requires a capability that managed services cannot provide — custom kernel modules, specific OS versions, or unsupported database engines.

Decision Rules

Choose between serverless managed ETL (AWS Glue) and managed-cluster processing (Amazon EMR) when transformation logic is standard and the team has no capacity to size, launch, or maintain a cluster.

AWS GlueAmazon EMR

Whether the workload's per-invocation payload size and execution duration fit within Lambda's operational ceiling, making Lambda+SAM strictly preferable to EMR on operational overhead grounds when the constraint is eliminating persistent infrastructure and deployment drift.

AWS LambdaAWS Serverless Application Model (AWS SAM)Amazon EMR

Whether the ingestion cadence and SaaS source type justify the shard provisioning, consumer application, and failure-path management that Kinesis Data Streams introduces, or whether a managed SaaS connector eliminates all of that overhead while fully satisfying the scheduled-batch latency requirement.

Amazon AppFlowAmazon Kinesis Data StreamsAmazon S3

Whether the transformation workload's volume and complexity justify accepting EMR's cluster management burden, or whether AWS Glue's serverless execution model satisfies the performance requirement while eliminating that burden entirely.

AWS GlueAmazon EMR

Does the pipeline's shape and operational constraint justify the environment-management overhead of MWAA, or does a simple linear fixed-schedule pipeline fit serverless-native Step Functions at lower operational cost?

AWS Step FunctionsAmazon Managed Workflows for Apache Airflow (Amazon MWAA)

Which AWS data store natively delivers sub-millisecond key/value read latency for a session cache workload without requiring an additional caching or acceleration component?

Amazon MemoryDB for RedisAmazon DynamoDB

Whether to satisfy retention-compliance and retrieval-time constraints through a native S3 Lifecycle policy (declarative, zero operational overhead, built-in transition engine) or through a custom orchestration layer such as a scheduled Glue job that scans and moves objects (flexible but adds scheduling, failure-path, and maintenance complexity).

Amazon S3AWS Glue

Whether to rely on Glue crawler-based schema discovery — which auto-detects structural changes but creates ongoing operational burden through classifier tuning, partition projection maintenance, and cross-consumer version drift — or Lake Formation governed tables, which provide transactional schema evolution with centralized versioning and lower per-change management cost.

AWS GlueAWS Lake Formation

Whether to rely on DMS native DDL replication (with Oracle supplemental logging and DDL-enabled task settings) to propagate ADD COLUMN operations automatically, or to build a custom schema-diff orchestration layer outside DMS that detects and applies DDL changes as a separate operational concern.

Amazon RedshiftAWS Database Migration Service (AWS DMS)

Whether to adopt a single purpose-built durable in-memory store (MemoryDB for Redis) or a two-tier cache-plus-database architecture (ElastiCache for Redis + Aurora) given simultaneous sub-millisecond latency, persistence, and minimal operational footprint constraints.

Amazon MemoryDB for RedisAmazon Aurora

When S3 partition paths follow a fully deterministic Hive-style scheme, choose Athena partition projection or programmatic Glue batch_create_partition API calls over a scheduled Glue crawler; crawlers provide schema-discovery value only when prefixes are irregular or schema is unknown, and for predictable layouts they add scheduling coordination, re-crawl duration, IAM scope, and failure-mode complexity that delivers no incremental benefit.

AWS GlueAWS Lake FormationAmazon Athena

When a pipeline dependency graph requires native sensor operators, calendar-based scheduling windows, and multi-task DAG visualization, choose MWAA over Step Functions — the operational cost of replicating those primitives in custom state-machine logic exceeds what the team can sustain.

Amazon Managed Workflows for Apache Airflow (Amazon MWAA)AWS Step FunctionsAWS Glue

Whether to apply AWS Glue DataBrew's managed declarative ruleset framework for data quality profiling or embed custom validation logic inside an AWS Glue ETL job, given the dominant constraint of minimal operational overhead.

AWS Glue DataBrewAWS Glue

When a pipeline's dependency graph includes external sensor waits, conditional branching, and cross-task data sharing, choose the orchestrator whose native operator set satisfies those patterns without custom Lambda glue code — MWAA over Step Functions — despite MWAA's higher managed-environment baseline cost.

Amazon Managed Workflows for Apache Airflow (Amazon MWAA)AWS Step FunctionsAWS Glue

When query frequency is low and unpredictable, does the utilization-threshold heuristic favor a serverless pay-per-query engine (Athena) over a provisioned always-on cluster (Redshift) given the explicit constraint to eliminate operational overhead?

Amazon AthenaAmazon Redshift

Match query frequency to service tier: choose serverless pay-per-query (Athena over CloudTrail S3 logs) rather than an always-on managed cluster (OpenSearch Service) when audit queries are infrequent and eliminating persistent operational burden is the dominant constraint.

AWS CloudTrailAmazon AthenaAmazon OpenSearch Service

Whether to enforce least-privilege data access through per-service IAM resource policies attached independently to each analytics service, or through a centralized Lake Formation permission grant model that governs all registered data assets from a single control plane.

AWS Lake FormationAWS Identity and Access Management (IAM)

Domain Coverage

Data Ingestion and TransformationData Store ManagementData Operations and SupportData Security and Governance

Difficulty Breakdown

Medium: 32Hard: 32Easy: 4

Related Patterns