{"id":117,"date":"2026-04-12T21:20:02","date_gmt":"2026-04-12T21:20:02","guid":{"rendered":"https:\/\/www.devopsschool.com\/tutorials\/aws-data-exchange-tutorial-architecture-pricing-use-cases-and-hands-on-guide-for-analytics\/"},"modified":"2026-04-12T21:20:02","modified_gmt":"2026-04-12T21:20:02","slug":"aws-data-exchange-tutorial-architecture-pricing-use-cases-and-hands-on-guide-for-analytics","status":"publish","type":"post","link":"https:\/\/www.devopsschool.com\/tutorials\/aws-data-exchange-tutorial-architecture-pricing-use-cases-and-hands-on-guide-for-analytics\/","title":{"rendered":"AWS Data Exchange Tutorial: Architecture, Pricing, Use Cases, and Hands-On Guide for Analytics"},"content":{"rendered":"\n<h2 class=\"wp-block-heading\">Category<\/h2>\n\n\n\n<p>Analytics<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">1. Introduction<\/h2>\n\n\n\n<p>AWS Data Exchange is an AWS service that helps you <strong>find, subscribe to, and use third\u2011party datasets<\/strong> (and some AWS-provided datasets) directly in your AWS environment. It is designed for teams that need reliable access to external data for <strong>Analytics<\/strong>, machine learning, reporting, risk modeling, enrichment, or research\u2014without building one-off vendor ingestion pipelines for every provider.<\/p>\n\n\n\n<p>In simple terms: <strong>AWS Data Exchange is a data marketplace workflow built for AWS<\/strong>. You browse data products, subscribe under clear terms, and then consume the data in AWS services like Amazon S3, Amazon Athena, AWS Glue, and (for some products) Amazon Redshift\u2014using repeatable, auditable processes.<\/p>\n\n\n\n<p>Technically, AWS Data Exchange provides a managed catalog and subscription mechanism around <strong>data products<\/strong>. Providers publish data products (containing datasets, revisions, and assets). Subscribers accept terms and gain entitlement to those datasets, then use AWS Data Exchange jobs and integrations to <strong>export or access<\/strong> data in their own AWS account. This separates <strong>procurement\/entitlement<\/strong> (control plane) from <strong>consumption\/analytics<\/strong> (data plane).<\/p>\n\n\n\n<p>What problem does it solve?<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Procurement friction<\/strong>: negotiating, billing, and contracting for external datasets can be slow and inconsistent.<\/li>\n<li><strong>Operational friction<\/strong>: ad hoc SFTP drops, emailed CSVs, bespoke APIs, and custom pipelines are brittle and hard to govern.<\/li>\n<li><strong>Governance gaps<\/strong>: auditability, access control, and lineage are difficult when data arrives outside standard cloud workflows.<\/li>\n<li><strong>Time-to-value<\/strong>: data teams spend too much time acquiring data, not analyzing it.<\/li>\n<\/ul>\n\n\n\n<p>AWS Data Exchange is not a general ETL tool. It is a <strong>data subscription and delivery mechanism<\/strong> that plugs into your existing analytics stack.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">2. What is AWS Data Exchange?<\/h2>\n\n\n\n<p>AWS Data Exchange is an AWS service that enables data providers to publish <strong>data products<\/strong> and data subscribers to discover, subscribe to, and use those data products on AWS. It integrates tightly with <strong>AWS Marketplace<\/strong> for product listings, subscriptions, entitlement, and billing (the exact commerce flow depends on the product).<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Official purpose (scope)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>For subscribers (consumers)<\/strong>: discover and subscribe to third-party data products and then consume them in AWS.<\/li>\n<li><strong>For providers (publishers)<\/strong>: package datasets, manage versions (revisions), define product offers\/terms, and deliver updates through AWS-managed mechanisms.<\/li>\n<\/ul>\n\n\n\n<p>Official docs: https:\/\/docs.aws.amazon.com\/data-exchange\/<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Core capabilities<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Browse and subscribe to data products (often via AWS Marketplace).<\/li>\n<li>Work with structured publishing concepts:<\/li>\n<li><strong>Data products<\/strong><\/li>\n<li><strong>Datasets<\/strong><\/li>\n<li><strong>Revisions<\/strong> (versioned updates)<\/li>\n<li><strong>Assets<\/strong> (files or other deliverables)<\/li>\n<li>Export data to your AWS environment (commonly <strong>Amazon S3<\/strong>).<\/li>\n<li>Receive update notifications for new revisions (commonly via <strong>Amazon EventBridge<\/strong>).<\/li>\n<li>Integrate with analytics services (Athena, Glue, Redshift, EMR, SageMaker) via standard AWS data lake patterns.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Major components (conceptual model)<\/h3>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Component<\/th>\n<th>What it is<\/th>\n<th>Why it matters<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Data product<\/td>\n<td>What you subscribe to (commercial + technical packaging)<\/td>\n<td>Defines terms, pricing, and what you receive<\/td>\n<\/tr>\n<tr>\n<td>Dataset<\/td>\n<td>A logical collection of data<\/td>\n<td>Groups revisions\/assets into a manageable unit<\/td>\n<\/tr>\n<tr>\n<td>Revision<\/td>\n<td>A point-in-time version of a dataset<\/td>\n<td>Enables updates, backfills, historical snapshots<\/td>\n<\/tr>\n<tr>\n<td>Asset<\/td>\n<td>The actual deliverable item (often a file)<\/td>\n<td>The \u201cdata payload\u201d you export\/use<\/td>\n<\/tr>\n<tr>\n<td>Subscription \/ entitlement<\/td>\n<td>The rights to access the product<\/td>\n<td>Enforced by AWS-integrated entitlement controls<\/td>\n<\/tr>\n<tr>\n<td>Jobs (for some flows)<\/td>\n<td>Managed actions like exporting assets to S3<\/td>\n<td>Makes delivery repeatable and auditable<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\">Service type<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Managed data exchange and entitlement service<\/strong> (control plane), with integrations into AWS storage and analytics for the data plane.<\/li>\n<li>Closely integrated with <strong>AWS Marketplace<\/strong> for subscriptions and billing.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Regional\/global and scoping notes<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>AWS Data Exchange is generally <strong>regional<\/strong> (you choose a region in the console). Data products and exports are performed in that region.<br\/>\n<strong>Verify region availability and any region-specific behaviors in official docs<\/strong>, because not all Marketplace products or delivery methods are available in all regions.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">How it fits into the AWS ecosystem<\/h3>\n\n\n\n<p>AWS Data Exchange is typically used at the \u201cdata acquisition\u201d layer:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Discovery &amp; procurement<\/strong>: AWS Marketplace + AWS Data Exchange<\/li>\n<li><strong>Landing zone<\/strong>: Amazon S3 (often a \u201craw\/vendor\u201d bucket)<\/li>\n<li><strong>Catalog<\/strong>: AWS Glue Data Catalog (and optionally Lake Formation)<\/li>\n<li><strong>Query &amp; analytics<\/strong>: Amazon Athena, Amazon Redshift, Amazon EMR, Amazon OpenSearch Service (depending on use case)<\/li>\n<li><strong>ML<\/strong>: Amazon SageMaker<\/li>\n<li><strong>Governance &amp; security<\/strong>: IAM, KMS, CloudTrail, Config, SCPs, Lake Formation<\/li>\n<li><strong>Automation<\/strong>: EventBridge + Lambda\/Step Functions for new revision handling<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">3. Why use AWS Data Exchange?<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Business reasons<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Faster procurement<\/strong>: standardized subscription workflow, often with clear commercial terms.<\/li>\n<li><strong>Access to a broad ecosystem<\/strong>: many providers distribute datasets through AWS channels.<\/li>\n<li><strong>Predictable delivery model<\/strong>: updates published as revisions; you can build processes around them.<\/li>\n<li><strong>Reduced vendor integration effort<\/strong>: fewer custom pipelines per data vendor.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Technical reasons<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Repeatable ingestion<\/strong>: export processes can be standardized (e.g., \u201cexport revision to S3 raw zone\u201d).<\/li>\n<li><strong>Versioning via revisions<\/strong>: ingest can be incremental and traceable.<\/li>\n<li><strong>Works with common AWS analytics patterns<\/strong>: S3 + Glue + Athena\/Redshift.<\/li>\n<li><strong>Event-driven updates<\/strong>: automate ingestion when new revisions appear.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Operational reasons<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Auditability<\/strong>: subscriptions and access are tied to AWS identities and logged through AWS mechanisms.<\/li>\n<li><strong>Separation of duties<\/strong>: procurement can subscribe; engineering can operationalize exports.<\/li>\n<li><strong>Fewer brittle manual processes<\/strong>: less reliance on emailed files or unmanaged access.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Security\/compliance reasons<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Centralized IAM<\/strong>: manage who can subscribe\/export, and where data can land.<\/li>\n<li><strong>Encryption<\/strong>: encrypt data at rest in your buckets with SSE-S3 or SSE-KMS.<\/li>\n<li><strong>Logging<\/strong>: use AWS CloudTrail for governance and audit requirements.<\/li>\n<li><strong>Policy guardrails<\/strong>: enforce allowed regions, bucket policies, and KMS key usage.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Scalability\/performance reasons<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Scales with AWS-native storage and query engines (S3, Athena, Redshift).<\/li>\n<li>Supports data lake patterns that decouple storage from compute.<\/li>\n<li>Allows you to scale ingestion workflows as your number of datasets grows.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">When teams should choose AWS Data Exchange<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>You need <strong>third-party datasets<\/strong> inside AWS for Analytics\/ML.<\/li>\n<li>You want a <strong>governable subscription + ingestion workflow<\/strong> with versioning (revisions).<\/li>\n<li>You want to <strong>automate updates<\/strong> and reduce manual vendor handling.<\/li>\n<li>You already operate a data lake\/warehouse on AWS.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">When teams should not choose AWS Data Exchange<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>You only need <strong>public\/open data<\/strong> already available via direct download or AWS Open Data (you may not need subscription workflows).<\/li>\n<li>You need <strong>real-time streaming<\/strong> data ingestion (ADX is not a streaming service; you\u2019d typically use Kinesis\/MSK + provider integration).<\/li>\n<li>Your vendor only supports bespoke delivery (SFTP, private API) and is not present in AWS Data Exchange.<\/li>\n<li>Your main requirement is <strong>transformation\/ETL<\/strong> (use Glue, EMR, dbt, Step Functions, etc.).<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">4. Where is AWS Data Exchange used?<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Industries<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Financial services (market\/reference data, alternative data)<\/li>\n<li>Insurance (risk, claims enrichment, fraud signals)<\/li>\n<li>Retail\/e-commerce (demographics, mobility, pricing intelligence)<\/li>\n<li>Healthcare and life sciences (licensed datasets, research data; ensure compliance)<\/li>\n<li>Manufacturing and logistics (supply chain, geo and routing data)<\/li>\n<li>Media and advertising (audience, location, campaign enrichment)<\/li>\n<li>Energy and utilities (weather, satellite, commodity analytics)<\/li>\n<li>Public sector (licensed geospatial and economic datasets; procurement constraints apply)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Team types<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Data engineering and analytics engineering teams<\/li>\n<li>BI\/reporting teams<\/li>\n<li>ML engineering and data science teams<\/li>\n<li>Platform and cloud infrastructure teams<\/li>\n<li>Security and governance teams (data access controls, audit)<\/li>\n<li>Procurement \/ FinOps teams (subscription governance and cost controls)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Workloads<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Data enrichment pipelines (join customer events with external features)<\/li>\n<li>Risk scoring and forecasting<\/li>\n<li>Market intelligence dashboards<\/li>\n<li>Geospatial analytics (mobility, POI, mapping datasets)<\/li>\n<li>Training ML models with proprietary labeled datasets<\/li>\n<li>Compliance reporting and backtesting using historical snapshots<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Architectures<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Data lake (S3 + Glue + Athena)<\/li>\n<li>Lakehouse patterns (S3 + open table formats, if applicable to the dataset you receive)<\/li>\n<li>Data warehouse augmentation (Redshift loading, or Redshift-integrated offerings where applicable)<\/li>\n<li>MLOps pipelines (S3 landing -&gt; feature store \/ training datasets)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Real-world deployment contexts<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Production<\/strong>: automated ingestion with EventBridge notifications, strict bucket policies, encryption, and partitioning\/cost controls for Athena\/Redshift.<\/li>\n<li><strong>Dev\/test<\/strong>: smaller subscriptions (often free products), sampling workflows, schema validation, and cost-limited Athena workgroups.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">5. Top Use Cases and Scenarios<\/h2>\n\n\n\n<p>Below are realistic scenarios where AWS Data Exchange fits well. Each example assumes you are using AWS as your primary Analytics platform.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">1) Vendor dataset ingestion to an S3 data lake<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: External vendor drops monthly CSVs via SFTP; ingestion is manual and error-prone.<\/li>\n<li><strong>Why AWS Data Exchange fits<\/strong>: Versioned revisions + export jobs to S3 create a repeatable ingestion pattern.<\/li>\n<li><strong>Example<\/strong>: Subscribe to a demographics dataset and export each monthly revision to <code>s3:\/\/datalake-raw\/vendor_x\/demographics\/revision_date=...\/<\/code>.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">2) Event-driven pipeline when data updates<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Data arrives irregularly; teams miss updates and dashboards become stale.<\/li>\n<li><strong>Why it fits<\/strong>: New revisions can trigger EventBridge events, enabling automated ingestion.<\/li>\n<li><strong>Example<\/strong>: EventBridge rule triggers Lambda to export new revision assets and refresh Glue partitions.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">3) Rapid proof-of-concept with free datasets<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: You need a dataset quickly to validate a model or dashboard.<\/li>\n<li><strong>Why it fits<\/strong>: Many listings are free; subscription is quick and doesn\u2019t require vendor-specific onboarding.<\/li>\n<li><strong>Example<\/strong>: Subscribe to a free sample dataset and query it in Athena within an hour.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">4) Controlled procurement for regulated environments<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Procurement wants traceability: who subscribed, what terms, when data changed.<\/li>\n<li><strong>Why it fits<\/strong>: Central subscription workflow with AWS account-level entitlement and audit trails.<\/li>\n<li><strong>Example<\/strong>: Enforce that only a procurement role can subscribe; engineering can export but not subscribe.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">5) Multi-account data platform with centralized landing zone<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Business units need shared vendor data, but you want a single controlled landing zone.<\/li>\n<li><strong>Why it fits<\/strong>: You can standardize exports into a centralized raw bucket and share curated data downstream.<\/li>\n<li><strong>Example<\/strong>: Export to a central S3 bucket in a data account; share curated tables to analytics accounts via Lake Formation (if used).<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6) Historical backtesting using revision snapshots<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Analysts need \u201cas-of\u201d datasets to reproduce decisions made months ago.<\/li>\n<li><strong>Why it fits<\/strong>: Revisions can represent snapshots; you can store each revision under a revision-specific prefix.<\/li>\n<li><strong>Example<\/strong>: Save each revision and use Athena to query \u201cdataset as of 2024-12-01\u201d.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">7) Data enrichment for customer segmentation<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Internal customer events lack geographic or demographic context.<\/li>\n<li><strong>Why it fits<\/strong>: External datasets can be joined to internal data in the lake\/warehouse.<\/li>\n<li><strong>Example<\/strong>: Join customer ZIP\/postcode with a vendor socioeconomic dataset to improve segmentation.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">8) ML feature generation from third-party signals<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: You want additional features for churn prediction but lack external signals.<\/li>\n<li><strong>Why it fits<\/strong>: Subscribe once; revisions update features over time.<\/li>\n<li><strong>Example<\/strong>: Export updated features monthly; retrain model with the latest revision.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">9) Standardized vendor data catalog for analysts<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Analysts don\u2019t know what external data exists or how to access it.<\/li>\n<li><strong>Why it fits<\/strong>: Data products are discoverable and documented; you can maintain internal documentation pointing to products.<\/li>\n<li><strong>Example<\/strong>: Data platform team curates an internal \u201capproved external datasets\u201d list sourced from AWS Data Exchange products.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">10) Replace bespoke vendor APIs with governed access paths<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Vendor API keys are spread across teams; access is uncontrolled.<\/li>\n<li><strong>Why it fits<\/strong>: Subscription\/entitlement can be centralized, and downstream access can be managed via AWS controls.<\/li>\n<li><strong>Example<\/strong>: Central team subscribes and operationalizes access in a shared environment rather than distributing keys to many developers.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">11) Faster onboarding of new regions or environments<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: When you expand to a new AWS region, setting up data vendor pipelines takes weeks.<\/li>\n<li><strong>Why it fits<\/strong>: If the product is available in-region, you can replicate the same export workflow.<\/li>\n<li><strong>Example<\/strong>: Re-run standardized export + catalog automation in the new region.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">12) Governance-driven \u201capproved dataset\u201d pipelines<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: Security requires controls before any external data enters analytics environments.<\/li>\n<li><strong>Why it fits<\/strong>: You can land vendor data into a quarantine bucket\/prefix, scan and validate, then promote.<\/li>\n<li><strong>Example<\/strong>: Export into <code>raw-quarantine\/<\/code>, run classification\/validation, then copy to curated zones.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">6. Core Features<\/h2>\n\n\n\n<p>This section focuses on <strong>current, commonly used<\/strong> AWS Data Exchange capabilities. If a feature depends on product type or region, it\u2019s called out.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">6.1 Data product discovery and subscription (Marketplace-integrated)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Lets you browse data products and subscribe under defined terms.<\/li>\n<li><strong>Why it matters<\/strong>: Reduces friction and standardizes procurement.<\/li>\n<li><strong>Practical benefit<\/strong>: Faster onboarding, fewer vendor-specific processes.<\/li>\n<li><strong>Caveats<\/strong>: Subscription and billing mechanics may be handled through AWS Marketplace; product availability varies by region. Verify in official docs and the specific product listing.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6.2 Datasets, revisions, and assets (versioned delivery)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Structures delivered data as datasets, with revisions containing assets.<\/li>\n<li><strong>Why it matters<\/strong>: Enables repeatable ingestion and \u201cwhat changed when\u201d tracking.<\/li>\n<li><strong>Practical benefit<\/strong>: You can build pipelines that process \u201cnew revision\u201d events and store revision-specific snapshots.<\/li>\n<li><strong>Caveats<\/strong>: Asset formats and schemas are provider-defined; you should validate schemas and quality per revision.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6.3 Export workflows (commonly to Amazon S3)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Exports entitled assets into an S3 bucket\/prefix in your account.<\/li>\n<li><strong>Why it matters<\/strong>: S3 is the standard landing zone for Analytics on AWS.<\/li>\n<li><strong>Practical benefit<\/strong>: Once data is in S3, you can use Glue\/Athena\/EMR\/SageMaker easily.<\/li>\n<li><strong>Caveats<\/strong>: Ensure bucket policies, encryption settings, and region constraints align with export requirements. Some exports may require service-linked roles. Verify exact prerequisites in official docs.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6.4 Event-driven notifications (new revision)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Notifies you when a provider publishes a new revision (commonly via Amazon EventBridge).<\/li>\n<li><strong>Why it matters<\/strong>: Eliminates manual checking and enables near-automated refresh.<\/li>\n<li><strong>Practical benefit<\/strong>: Automate ingestion, re-cataloging, partition updates, and downstream refresh.<\/li>\n<li><strong>Caveats<\/strong>: Event payloads and configuration specifics should be verified in official docs for your region and product type.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6.5 Provider publishing workflows (for data sellers)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Helps providers create datasets, add revisions, attach assets, and publish products.<\/li>\n<li><strong>Why it matters<\/strong>: Makes dataset distribution scalable and manageable.<\/li>\n<li><strong>Practical benefit<\/strong>: Providers can ship updates and manage versions without bespoke delivery to every customer.<\/li>\n<li><strong>Caveats<\/strong>: Provider onboarding and commerce flows are tied to AWS Marketplace capabilities and policies.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6.6 Integration with AWS analytics services (via standard patterns)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Enables downstream consumption in Athena\/Glue\/Redshift\/EMR\/SageMaker.<\/li>\n<li><strong>Why it matters<\/strong>: AWS Data Exchange is not the query engine; it\u2019s the ingestion\/subscription layer.<\/li>\n<li><strong>Practical benefit<\/strong>: You keep your standard analytics architecture; AWS Data Exchange just supplies the data.<\/li>\n<li><strong>Caveats<\/strong>: You\u2019re responsible for table definitions, partitioning, and optimizing query\/storage formats unless the product provides optimized formats.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">6.7 Support for multiple delivery modalities (product-dependent)<\/h3>\n\n\n\n<p>AWS Data Exchange offerings may include different delivery modalities, depending on the product:\n&#8211; <strong>File-based datasets<\/strong> (commonly exported to S3)\n&#8211; <strong>Other integrated modalities<\/strong> (for example, certain products integrate with Amazon Redshift or provide API-based access)<\/p>\n\n\n\n<p>Because these vary by product and evolve over time, <strong>verify supported modalities for your chosen product<\/strong> in the product listing and official docs.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">6.8 Auditing and governance alignment (CloudTrail\/IAM)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>What it does<\/strong>: Allows you to manage access via IAM and capture actions in AWS audit trails.<\/li>\n<li><strong>Why it matters<\/strong>: External data is still sensitive and often licensed; you need traceability.<\/li>\n<li><strong>Practical benefit<\/strong>: Aligns external data access with your AWS governance model.<\/li>\n<li><strong>Caveats<\/strong>: You still must implement internal controls (tagging, bucket policies, Lake Formation permissions, retention).<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">7. Architecture and How It Works<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">7.1 High-level architecture<\/h3>\n\n\n\n<p>AWS Data Exchange has a typical pattern:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Discover\/Subscribe<\/strong>: A user subscribes to a data product (often via AWS Marketplace flow).<\/li>\n<li><strong>Entitlement<\/strong>: The subscription grants entitlement to datasets.<\/li>\n<li><strong>Delivery\/Export<\/strong>: Subscriber uses AWS Data Exchange to export assets to an S3 bucket (common pattern) or uses another supported access method (product-dependent).<\/li>\n<li><strong>Catalog and query<\/strong>: Use AWS Glue to catalog; query with Athena or load into a warehouse.<\/li>\n<li><strong>Automate updates<\/strong>: Use EventBridge to detect new revisions and orchestrate repeat exports.<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">7.2 Control flow vs data flow<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Control plane<\/strong>: subscriptions, entitlements, dataset\/revision metadata, jobs, permissions.<\/li>\n<li><strong>Data plane<\/strong>: actual bytes moved to your storage (S3) or accessed through supported integrated endpoints.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">7.3 Integrations with related services<\/h3>\n\n\n\n<p>Common integrations in Analytics stacks:\n&#8211; <strong>Amazon S3<\/strong>: landing and storage\n&#8211; <strong>AWS Glue Data Catalog<\/strong>: schema\/table metadata\n&#8211; <strong>Amazon Athena<\/strong>: serverless SQL queries over S3\n&#8211; <strong>Amazon Redshift<\/strong>: warehouse loading or integrated access (product-dependent)\n&#8211; <strong>Amazon EventBridge<\/strong>: revision notifications\n&#8211; <strong>AWS Lambda \/ Step Functions<\/strong>: automation and orchestration\n&#8211; <strong>AWS KMS<\/strong>: encryption keys for S3 SSE-KMS\n&#8211; <strong>AWS CloudTrail<\/strong>: audit\n&#8211; <strong>AWS Config \/ SCPs<\/strong>: governance guardrails<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">7.4 Security\/authentication model<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Access is controlled with <strong>IAM<\/strong>. Users\/roles need permission to subscribe, view datasets, and run export jobs.<\/li>\n<li>AWS Data Exchange may create or use a <strong>service-linked role<\/strong> to perform actions on your behalf (for example, writing into your S3 bucket). The exact role name and required trust\/permissions should be validated in official docs for your region and workflow.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">7.5 Networking model<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>AWS Data Exchange is managed by AWS; you interact via AWS console\/API endpoints in a region.<\/li>\n<li>Data consumption usually happens via AWS services (S3, Athena). For private network patterns, use:<\/li>\n<li><strong>S3 VPC Gateway Endpoint<\/strong> for private S3 access from within a VPC<\/li>\n<li>Private connectivity patterns for downstream systems<\/li>\n<li>Export itself is an AWS-managed operation; you mainly control destination buckets and encryption\/policies.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">7.6 Monitoring\/logging\/governance considerations<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>CloudTrail<\/strong>: track API calls related to AWS Data Exchange actions (subscribe\/export\/job actions).<\/li>\n<li><strong>CloudWatch<\/strong>: monitor Lambda\/Step Functions if you automate.<\/li>\n<li><strong>S3 server access logs \/ CloudTrail data events<\/strong> (optional): track object-level access to exported datasets.<\/li>\n<li><strong>Tagging<\/strong>: tag destination buckets\/prefixes and track dataset provenance (product name, revision id, subscription id) in metadata.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">7.7 Simple architecture diagram (Mermaid)<\/h3>\n\n\n\n<pre><code class=\"language-mermaid\">flowchart LR\n  U[Data Engineer] --&gt;|Subscribe| DX[AWS Data Exchange]\n  DX --&gt;|Entitlement| SUB[Subscription to Data Product]\n  SUB --&gt;|Export assets| S3[(Amazon S3 Raw Bucket)]\n  S3 --&gt; GLUE[AWS Glue Data Catalog]\n  GLUE --&gt; ATHENA[Amazon Athena]\n  ATHENA --&gt; BI[BI Tool \/ Notebooks]\n<\/code><\/pre>\n\n\n\n<h3 class=\"wp-block-heading\">7.8 Production-style architecture diagram (Mermaid)<\/h3>\n\n\n\n<pre><code class=\"language-mermaid\">flowchart TB\n  subgraph Procurement_and_Governance\n    IAM[IAM Roles &amp; SCP Guardrails]\n    CT[CloudTrail Audit Logs]\n    CMK[KMS CMK for S3 SSE-KMS]\n  end\n\n  subgraph Data_Subscription\n    DX[AWS Data Exchange]\n    MP[AWS Marketplace Listing\/Subscription Flow]\n  end\n\n  subgraph Landing_Zone\n    S3Q[(S3 Quarantine Prefix)]\n    S3R[(S3 Raw Vendor Zone - Versioned)]\n    LF[Optional: Lake Formation Governance]\n  end\n\n  subgraph Automation\n    EB[EventBridge: New Revision Event]\n    SF[Step Functions Orchestrator]\n    L1[Lambda: Export Job + Metadata]\n    L2[Lambda: Glue Catalog\/Partition Updates]\n    DQ[Data Quality Checks]\n  end\n\n  subgraph Analytics\n    GLUE[AWS Glue Data Catalog]\n    ATHENA[Amazon Athena]\n    RS[Amazon Redshift \/ Spectrum]\n    ML[SageMaker Training\/Feature Pipelines]\n    DW[Curated S3 Zone \/ Warehouse Tables]\n  end\n\n  MP --&gt; DX\n  IAM --&gt; DX\n  DX --&gt;|Export to S3| S3Q\n  EB --&gt; SF --&gt; L1 --&gt; DX\n  S3Q --&gt; DQ --&gt; S3R\n  S3R --&gt; GLUE --&gt; ATHENA --&gt; DW\n  S3R --&gt; RS\n  S3R --&gt; ML\n  CMK --&gt; S3Q\n  CMK --&gt; S3R\n  DX --&gt; CT\n<\/code><\/pre>\n\n\n\n<h2 class=\"wp-block-heading\">8. Prerequisites<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Account and billing<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>An AWS account with <strong>billing enabled<\/strong>.<\/li>\n<li>Ability to subscribe to AWS Marketplace products (some organizations restrict this).<\/li>\n<li>If you\u2019re in AWS Organizations:<\/li>\n<li>Confirm whether your org uses <strong>service control policies (SCPs)<\/strong> restricting Marketplace or AWS Data Exchange.<\/li>\n<li>Confirm whether procurement requires a centralized payer\/approval process.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">IAM permissions<\/h3>\n\n\n\n<p>For the hands-on lab (subscriber workflow), you need permissions to:\n&#8211; Use AWS Data Exchange (subscribe, view datasets, export).\n&#8211; Create\/manage an S3 bucket and objects.\n&#8211; Use AWS Glue (create database\/table or crawler) and Athena (run queries).<\/p>\n\n\n\n<p>For simplicity in a lab:\n&#8211; Use an admin role, <strong>or<\/strong> attach AWS-managed policies appropriate to your environment.<\/p>\n\n\n\n<p>For production, prefer least privilege:\n&#8211; Limit AWS Data Exchange actions to specific datasets\/products and restrict S3 destinations via bucket policy and IAM conditions.<\/p>\n\n\n\n<p><strong>Note:<\/strong> AWS-managed policy names and granular permissions can change. Verify the current recommended policies in official docs:\n&#8211; https:\/\/docs.aws.amazon.com\/data-exchange\/<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Tools<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>AWS Console access<\/li>\n<li>AWS CLI v2 (optional but useful):<\/li>\n<li>Install: https:\/\/docs.aws.amazon.com\/cli\/latest\/userguide\/install-cliv2.html<\/li>\n<li>Configure: <code>aws configure<\/code><\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Region availability<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Choose a region where <strong>AWS Data Exchange<\/strong>, <strong>S3<\/strong>, <strong>Athena<\/strong>, and <strong>Glue<\/strong> are available.<\/li>\n<li>AWS Data Exchange and specific products are not necessarily available in all regions. <strong>Verify in the console<\/strong> by switching regions and checking the service and listings.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Quotas\/limits<\/h3>\n\n\n\n<p>AWS Data Exchange has service quotas (jobs, concurrency, etc.). Quotas evolve, so:\n&#8211; Check the AWS Data Exchange Service Quotas page (in the AWS console under Service Quotas) for your region\/account.\n&#8211; Verify dataset\/asset size constraints for your chosen product in the listing and docs.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Prerequisite services<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Amazon S3 (bucket for exports)<\/li>\n<li>AWS Glue Data Catalog (to create tables or crawler)<\/li>\n<li>Amazon Athena (to query exported data)<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">9. Pricing \/ Cost<\/h2>\n\n\n\n<p>AWS Data Exchange cost has two major parts:<\/p>\n\n\n\n<p>1) <strong>The data product price<\/strong> (set by the provider, often billed through AWS Marketplace)<br\/>\n2) <strong>The downstream AWS usage costs<\/strong> (S3 storage, Athena queries, Glue crawlers, Redshift compute, data transfer, etc.)<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">9.1 Pricing dimensions (data product side)<\/h3>\n\n\n\n<p>Data products can be:\n&#8211; Free\n&#8211; Paid (subscription, contract-based, or usage-based depending on the listing and product type)<\/p>\n\n\n\n<p>The exact commercial model depends on the provider and product listing. Always review:\n&#8211; The product\u2019s pricing terms in the listing\n&#8211; Your Marketplace subscription details and invoices<\/p>\n\n\n\n<p>Official service pricing page:\n&#8211; https:\/\/aws.amazon.com\/data-exchange\/pricing\/<\/p>\n\n\n\n<p>Also consult AWS Pricing Calculator for downstream services:\n&#8211; https:\/\/calculator.aws\/#\/<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">9.2 Is there a free tier?<\/h3>\n\n\n\n<p>AWS Data Exchange itself does not have a typical \u201cfree tier\u201d like some AWS services, because <strong>the dataset pricing is provider-defined<\/strong>. However:\n&#8211; Many products are free (or have free samples).\n&#8211; Even with a free product, you still pay for S3, Athena, Glue, and any other services you use.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">9.3 Cost drivers (most common)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Data product subscription fees<\/strong> (if not free)<\/li>\n<li><strong>S3 storage<\/strong> of exported data (size \u00d7 duration, plus versioning if enabled)<\/li>\n<li><strong>S3 requests<\/strong> (PUT\/LIST\/GET) and lifecycle transitions<\/li>\n<li><strong>Athena<\/strong> query costs (per TB scanned; costs vary by region)<\/li>\n<li><strong>Glue<\/strong> crawler and job costs (DPU-hours; region-dependent)<\/li>\n<li><strong>Redshift<\/strong> compute\/storage (if you load data or query via Spectrum)<\/li>\n<li><strong>Data transfer<\/strong>:<\/li>\n<li>Intra-region data movement between AWS services is often low or no cost, but internet egress and cross-region transfers can be significant.<\/li>\n<li>If you copy exported data across regions\/accounts, data transfer and duplication costs apply.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">9.4 Hidden\/indirect costs to watch<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Query inefficiency<\/strong>: Athena scanning raw CSV can become expensive without partitioning and columnar formats.<\/li>\n<li><strong>Duplicate storage<\/strong>: storing multiple revisions forever can grow costs.<\/li>\n<li><strong>Automation sprawl<\/strong>: Lambda\/Step Functions costs are usually minor, but can increase with frequent updates and heavy orchestration.<\/li>\n<li><strong>Egress to non-AWS systems<\/strong>: exporting data out of AWS can trigger large egress charges and may violate licensing terms\u2014review the product terms.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">9.5 How to optimize cost<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Prefer <strong>partitioned<\/strong> layouts in S3 for large datasets (e.g., by date, region, or provider\u2019s natural partitions).<\/li>\n<li>Convert raw files to <strong>columnar<\/strong> formats (Parquet\/ORC) in curated layers if license allows.<\/li>\n<li>Use <strong>Athena workgroups<\/strong> with enforced limits and separate output buckets.<\/li>\n<li>Use <strong>S3 lifecycle policies<\/strong>:<\/li>\n<li>Move older revisions to cheaper storage classes if appropriate.<\/li>\n<li>Expire obsolete revisions if you don\u2019t need historical backtesting.<\/li>\n<li>Keep a clear retention policy by dataset and revision.<\/li>\n<li>For large datasets, consider a curated warehouse strategy (Redshift) when it reduces repeated scan costs.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">9.6 Example low-cost starter estimate (free product)<\/h3>\n\n\n\n<p>A realistic low-cost lab might include:\n&#8211; A free AWS Data Exchange product\n&#8211; Exporting a small dataset into S3 (tens to hundreds of MB)\n&#8211; Running a few Athena queries<\/p>\n\n\n\n<p>Costs you should expect:\n&#8211; S3 storage: small\n&#8211; Athena: depends on bytes scanned (keep queries selective; avoid <code>SELECT *<\/code> on huge files)\n&#8211; Glue crawler: optional (you can define schema manually to avoid crawler cost)<\/p>\n\n\n\n<p>Because exact prices are region-dependent and change over time, <strong>use the AWS Pricing Calculator<\/strong> and your chosen region for estimates:\n&#8211; https:\/\/calculator.aws\/#\/<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">9.7 Example production cost considerations (paid product)<\/h3>\n\n\n\n<p>For production, add:\n&#8211; Paid subscription\/contract fees (provider-defined)\n&#8211; Larger S3 footprint (raw + curated + historical revisions)\n&#8211; Regular Glue jobs to convert\/curate data\n&#8211; Regular Athena\/Redshift usage by analysts and dashboards\n&#8211; Multi-account replication (optional) and governance tooling<\/p>\n\n\n\n<p>A good FinOps practice is to model costs by dataset:\n&#8211; subscription fee + ingestion + storage + query compute + retention<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">10. Step-by-Step Hands-On Tutorial<\/h2>\n\n\n\n<p>This lab demonstrates a safe, low-cost \u201chello world\u201d workflow:\n&#8211; Subscribe to a <strong>free<\/strong> AWS Data Exchange product (file-based).\n&#8211; Export a dataset revision to <strong>Amazon S3<\/strong>.\n&#8211; Catalog and query it using <strong>AWS Glue<\/strong> and <strong>Amazon Athena<\/strong>.\n&#8211; Clean up resources.<\/p>\n\n\n\n<p>Because product listings change over time, the exact dataset you pick may differ. The steps are written so you can complete them with <strong>any free file-based data product<\/strong> available in your region.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Objective<\/h3>\n\n\n\n<p>Subscribe to a free AWS Data Exchange data product and query exported data in Athena.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Lab Overview<\/h3>\n\n\n\n<p>You will:\n1. Choose a region and create an S3 bucket for AWS Data Exchange exports.\n2. Subscribe to a free data product in AWS Data Exchange.\n3. Export a dataset revision (assets) to your S3 bucket.\n4. Create an Athena table (or use Glue) and run a query.\n5. Clean up: delete S3 objects, remove Athena\/Glue artifacts, and unsubscribe if appropriate.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Step 1: Choose a region and prepare naming<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>In the AWS Console, pick a region you will use for the lab (top-right region selector).<\/li>\n<li>Write down:\n   &#8211; Region (example: <code>us-east-1<\/code>)\n   &#8211; Bucket name you will create (must be globally unique), e.g.:<ul>\n<li><code>my-dx-lab-&lt;accountid&gt;-&lt;region&gt;<\/code><\/li>\n<\/ul>\n<\/li>\n<\/ol>\n\n\n\n<p><strong>Expected outcome:<\/strong> You have a chosen AWS region and a unique S3 bucket name plan.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Step 2: Create an S3 bucket for exports (secure-by-default)<\/h3>\n\n\n\n<p>Create an S3 bucket in the same region you will use for AWS Data Exchange.<\/p>\n\n\n\n<p><strong>Option A: Console<\/strong>\n1. Go to <strong>Amazon S3<\/strong> \u2192 <strong>Create bucket<\/strong>\n2. Bucket name: <code>my-dx-lab-...<\/code>\n3. Region: same as your AWS Data Exchange region\n4. <strong>Block Public Access<\/strong>: keep <strong>enabled<\/strong>\n5. <strong>Bucket Versioning<\/strong>: optional (recommended for real pipelines; optional for lab)\n6. <strong>Default encryption<\/strong>: enable SSE-S3 or SSE-KMS<br\/>\n   &#8211; SSE-S3 is simplest<br\/>\n   &#8211; SSE-KMS gives stronger control\/audit, but requires KMS permissions<\/p>\n\n\n\n<p><strong>Option B: AWS CLI<\/strong><\/p>\n\n\n\n<pre><code class=\"language-bash\">aws s3api create-bucket \\\n  --bucket my-dx-lab-123456789012-us-east-1 \\\n  --region us-east-1\n<\/code><\/pre>\n\n\n\n<p>Enable default encryption (SSE-S3):<\/p>\n\n\n\n<pre><code class=\"language-bash\">aws s3api put-bucket-encryption \\\n  --bucket my-dx-lab-123456789012-us-east-1 \\\n  --server-side-encryption-configuration '{\n    \"Rules\": [{\n      \"ApplyServerSideEncryptionByDefault\": {\"SSEAlgorithm\": \"AES256\"}\n    }]\n  }'\n<\/code><\/pre>\n\n\n\n<p><strong>Expected outcome:<\/strong> An S3 bucket exists, private, encrypted, ready to receive exported assets.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Step 3: Find and subscribe to a free AWS Data Exchange data product<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Open <strong>AWS Data Exchange<\/strong> in the AWS Console (in your chosen region).<\/li>\n<li>Go to <strong>Discover data products<\/strong> (wording may vary slightly).<\/li>\n<li>Filter for:\n   &#8211; <strong>Free<\/strong> products\n   &#8211; Delivery type: choose a product that clearly indicates <strong>file-based dataset<\/strong> (commonly delivered\/exported to S3)<\/li>\n<li>Open the product listing and review:\n   &#8211; Data dictionary \/ documentation\n   &#8211; Update frequency\n   &#8211; File formats (CSV\/JSON\/Parquet, etc.)\n   &#8211; Terms and conditions<\/li>\n<li>Click <strong>Subscribe<\/strong> and complete the subscription flow.<\/li>\n<\/ol>\n\n\n\n<p><strong>Expected outcome:<\/strong> You have an active subscription to a free product, granting you access to its dataset(s).<\/p>\n\n\n\n<p><strong>Verification tip:<\/strong> In AWS Data Exchange, you should now see the product under something like <strong>Subscriptions<\/strong> or <strong>Entitled data<\/strong>.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Step 4: Export a dataset revision to your S3 bucket<\/h3>\n\n\n\n<p>Now you will export one revision (the latest) to S3.<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>In AWS Data Exchange console, navigate to your subscribed product and locate:\n   &#8211; The <strong>dataset<\/strong>\n   &#8211; The latest <strong>revision<\/strong>\n   &#8211; The list of <strong>assets<\/strong> (files)<\/li>\n<li>Choose an export option such as:\n   &#8211; Export assets to Amazon S3<br\/>\n   (Exact UI labels can vary.)<\/li>\n<li>Destination:\n   &#8211; Bucket: your lab bucket\n   &#8211; Prefix: choose a structured path, for example:<ul>\n<li><code>dataexchange\/product=&lt;product-name&gt;\/dataset=&lt;dataset-id&gt;\/revision=&lt;revision-id&gt;\/<\/code><\/li>\n<\/ul>\n<\/li>\n<li>Start the export job and wait for completion.<\/li>\n<\/ol>\n\n\n\n<p><strong>Expected outcome:<\/strong> The job completes successfully, and exported files appear in your S3 bucket under the prefix.<\/p>\n\n\n\n<p><strong>Verification (S3 console):<\/strong>\n&#8211; Go to the bucket \u2192 browse to your prefix \u2192 confirm files exist.<\/p>\n\n\n\n<p><strong>Verification (CLI):<\/strong><\/p>\n\n\n\n<pre><code class=\"language-bash\">aws s3 ls s3:\/\/my-dx-lab-123456789012-us-east-1\/dataexchange\/ --recursive | head\n<\/code><\/pre>\n\n\n\n<h3 class=\"wp-block-heading\">Step 5: Create an Athena query environment (output bucket\/prefix)<\/h3>\n\n\n\n<p>Athena needs a location to write query results.<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Open <strong>Amazon Athena<\/strong> (same region).<\/li>\n<li>In <strong>Settings<\/strong>, set a query result location, e.g.:\n   &#8211; <code>s3:\/\/my-dx-lab-...\/athena-results\/<\/code><\/li>\n<\/ol>\n\n\n\n<p><strong>Expected outcome:<\/strong> Athena is configured to store query outputs in your S3 bucket.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Step 6: Create a table (Glue crawler or manual DDL)<\/h3>\n\n\n\n<p>You have two common options:<\/p>\n\n\n\n<h4 class=\"wp-block-heading\">Option A (recommended for beginners): Use a Glue crawler<\/h4>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Open <strong>AWS Glue<\/strong> \u2192 <strong>Crawlers<\/strong> \u2192 <strong>Create crawler<\/strong><\/li>\n<li>Data source:\n   &#8211; S3 path to your exported dataset prefix<\/li>\n<li>IAM role:\n   &#8211; Choose an existing role or create a new one with S3 read permissions to your bucket<\/li>\n<li>Output:\n   &#8211; Create a new database (e.g., <code>dx_lab_db<\/code>)<\/li>\n<li>Run the crawler.<\/li>\n<\/ol>\n\n\n\n<p><strong>Expected outcome:<\/strong> Glue creates one or more tables in the Data Catalog for your exported files.<\/p>\n\n\n\n<h4 class=\"wp-block-heading\">Option B: Create an external table in Athena (manual)<\/h4>\n\n\n\n<p>If the dataset is a simple CSV and you know its columns, you can write DDL yourself. Example skeleton (you must edit column names\/types to match your dataset):<\/p>\n\n\n\n<pre><code class=\"language-sql\">CREATE DATABASE IF NOT EXISTS dx_lab_db;\n\nCREATE EXTERNAL TABLE IF NOT EXISTS dx_lab_db.vendor_dataset (\n  col1 string,\n  col2 string,\n  col3 bigint\n)\nROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.OpenCSVSerde'\nWITH SERDEPROPERTIES (\n  'separatorChar' = ',',\n  'quoteChar'     = '\"',\n  'escapeChar'    = '\\\\'\n)\nLOCATION 's3:\/\/my-dx-lab-123456789012-us-east-1\/dataexchange\/product=...\/dataset=...\/revision=...\/'\nTBLPROPERTIES ('skip.header.line.count'='1');\n<\/code><\/pre>\n\n\n\n<p><strong>Expected outcome:<\/strong> You can see a table in Athena (or Glue Data Catalog) pointing to the exported S3 data.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Step 7: Query the dataset in Athena<\/h3>\n\n\n\n<p>Run a safe, low-scan query first:<\/p>\n\n\n\n<pre><code class=\"language-sql\">SELECT *\nFROM dx_lab_db.vendor_dataset\nLIMIT 10;\n<\/code><\/pre>\n\n\n\n<p>If your table is partitioned (or you created partitions), query with filters to reduce scanned data:<\/p>\n\n\n\n<pre><code class=\"language-sql\">SELECT count(*) \nFROM dx_lab_db.vendor_dataset;\n<\/code><\/pre>\n\n\n\n<p><strong>Expected outcome:<\/strong> Athena returns rows and query completes successfully. You can now use this dataset in Analytics workflows.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Validation<\/h3>\n\n\n\n<p>Use this checklist:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Subscription exists and is active in AWS Data Exchange.<\/li>\n<li>Export job completed successfully.<\/li>\n<li>S3 bucket contains exported files.<\/li>\n<li>Glue catalog has a database\/table (or Athena DDL created a table).<\/li>\n<li>Athena query returns data.<\/li>\n<\/ul>\n\n\n\n<p>Optional extra validation:\n&#8211; Confirm encryption at rest in S3:\n  &#8211; S3 object \u2192 Properties \u2192 Server-side encryption shows AES-256 or AWS-KMS.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Troubleshooting<\/h3>\n\n\n\n<p><strong>Problem: Can\u2019t find any free products<\/strong>\n&#8211; Some regions have fewer listings.\n&#8211; Try a different region where AWS Data Exchange is available.\n&#8211; Confirm your account is allowed to use AWS Marketplace and AWS Data Exchange.<\/p>\n\n\n\n<p><strong>Problem: Subscription blocked or requires approvals<\/strong>\n&#8211; Your org may restrict AWS Marketplace subscriptions.\n&#8211; Work with your AWS Organizations admin\/procurement team, or test in a sandbox account.<\/p>\n\n\n\n<p><strong>Problem: Export fails with AccessDenied to S3<\/strong>\n&#8211; Confirm bucket is in the same region you\u2019re operating in.\n&#8211; Confirm bucket policy doesn\u2019t deny the AWS Data Exchange service role.\n&#8211; If using SSE-KMS, ensure the KMS key policy allows the required principal(s).\n&#8211; Verify the service-linked role requirements in official docs for your workflow.<\/p>\n\n\n\n<p><strong>Problem: Glue crawler creates incorrect schema<\/strong>\n&#8211; Many vendor datasets have complex CSV quirks.\n&#8211; Manually define the table DDL in Athena, or adjust crawler settings and classifiers.<\/p>\n\n\n\n<p><strong>Problem: Athena returns no rows<\/strong>\n&#8211; Confirm the S3 location is correct and includes files.\n&#8211; Confirm file format settings (CSV delimiter, header skip).\n&#8211; Confirm that the files are not compressed in an unexpected format.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Cleanup<\/h3>\n\n\n\n<p>To avoid ongoing costs:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>\n<p><strong>Delete Athena query results<\/strong>\n   &#8211; Delete objects under <code>s3:\/\/...\/athena-results\/<\/code><\/p>\n<\/li>\n<li>\n<p><strong>Delete exported dataset objects<\/strong>\n   &#8211; Delete objects under <code>s3:\/\/...\/dataexchange\/...<\/code><\/p>\n<\/li>\n<li>\n<p><strong>Delete Glue resources<\/strong>\n   &#8211; Delete crawler (if created)\n   &#8211; Delete Glue tables and database (<code>dx_lab_db<\/code>) if not needed<\/p>\n<\/li>\n<li>\n<p><strong>Delete S3 bucket (optional)<\/strong>\n   &#8211; Empty the bucket first, then delete it<\/p>\n<\/li>\n<li>\n<p><strong>Unsubscribe from the data product<\/strong> (if appropriate)\n   &#8211; Go to AWS Data Exchange \u2192 Subscriptions \u2192 unsubscribe<br\/>\n   Note: Unsubscribing does not automatically delete data already exported to your S3 bucket. You must delete it yourself if required by your data handling policy and license terms.<\/p>\n<\/li>\n<\/ol>\n\n\n\n<h2 class=\"wp-block-heading\">11. Best Practices<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Architecture best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Use a <strong>multi-zone data lake layout<\/strong>:<\/li>\n<li><code>quarantine\/<\/code> (optional) \u2192 <code>raw\/<\/code> \u2192 <code>curated\/<\/code><\/li>\n<li>Store each revision under a <strong>revision-specific prefix<\/strong> to preserve provenance:<\/li>\n<li><code>raw\/vendor=&lt;name&gt;\/product=&lt;id&gt;\/revision=&lt;id&gt;\/...<\/code><\/li>\n<li>Keep metadata about each revision (revision id, publish date, provider) in a small control table (e.g., DynamoDB or a Glue table).<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">IAM\/security best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Separate roles:<\/li>\n<li>Procurement role: subscribe\/accept terms<\/li>\n<li>Data engineering role: export jobs + write to controlled S3 paths<\/li>\n<li>Analyst role: read curated datasets only<\/li>\n<li>Use least privilege:<\/li>\n<li>Restrict S3 destinations via bucket policies and IAM condition keys where possible.<\/li>\n<li>If using SSE-KMS, design KMS key policies to support:<\/li>\n<li>Export job writes<\/li>\n<li>Downstream reads (Athena\/Glue\/EMR)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Cost best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Avoid long-term retention of every revision unless required.<\/li>\n<li>Convert large text datasets to Parquet in curated zones (if license permits).<\/li>\n<li>Use Athena partitioning and column pruning.<\/li>\n<li>Use S3 lifecycle policies and storage classes intentionally.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Performance best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Prefer columnar formats for repeated analytics.<\/li>\n<li>Partition by common query dimensions (date, geography, category).<\/li>\n<li>Maintain consistent naming conventions to simplify partition discovery.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Reliability best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Build idempotent ingestion:<\/li>\n<li>If you re-export a revision, write to the same prefix and verify checksums\/manifest.<\/li>\n<li>Implement retries and alerts on job failures (especially if automating).<\/li>\n<li>Maintain a \u201clast successfully processed revision\u201d state.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Operations best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Emit operational metrics:<\/li>\n<li>number of new revisions processed<\/li>\n<li>export job duration<\/li>\n<li>bytes landed in S3<\/li>\n<li>Centralize logs:<\/li>\n<li>CloudTrail for audit<\/li>\n<li>CloudWatch for automation logs<\/li>\n<li>Run periodic access reviews of who can subscribe\/export.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Governance\/tagging\/naming best practices<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Tag S3 buckets and datasets with:<\/li>\n<li><code>data-owner<\/code>, <code>cost-center<\/code>, <code>environment<\/code>, <code>vendor<\/code>, <code>license-class<\/code>, <code>retention<\/code><\/li>\n<li>Keep a dataset register internally:<\/li>\n<li>product link, license summary, allowed uses, retention rules, PII classification<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">12. Security Considerations<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Identity and access model<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>AWS Data Exchange uses IAM for access control.<\/li>\n<li>Common security model:<\/li>\n<li>A limited set of roles can <strong>subscribe<\/strong> to products.<\/li>\n<li>A small set of roles can <strong>export<\/strong> to approved S3 locations.<\/li>\n<li>Analysts can only access curated datasets, not raw vendor drops.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Encryption<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>For S3 destinations:<\/li>\n<li>Enable <strong>default encryption<\/strong> (SSE-S3 or SSE-KMS).<\/li>\n<li>Prefer SSE-KMS when you need key-level access control and audit.<\/li>\n<li>For SSE-KMS:<\/li>\n<li>Ensure key policies allow the principals that need to write\/read.<\/li>\n<li>Use separate CMKs by environment (dev\/test\/prod) when practical.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Network exposure<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Keep exported data in <strong>private S3 buckets<\/strong> with Block Public Access enabled.<\/li>\n<li>If accessing from VPC-based compute (EMR, EC2, EKS):<\/li>\n<li>Use <strong>S3 VPC endpoints<\/strong> and restrict S3 bucket policy to your VPC endpoint if appropriate.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Secrets handling<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Avoid embedding vendor credentials in code.<\/li>\n<li>For API-based data products (where applicable), store tokens\/keys in <strong>AWS Secrets Manager<\/strong> and rotate when possible.<\/li>\n<li>Restrict who can read those secrets, and log access.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Audit\/logging<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Enable and retain <strong>CloudTrail<\/strong> logs for:<\/li>\n<li>subscription actions<\/li>\n<li>export job actions<\/li>\n<li>IAM changes<\/li>\n<li>Consider S3 object-level logging (CloudTrail data events) for sensitive datasets.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Compliance considerations<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>External datasets often come with license restrictions:<\/li>\n<li>permitted uses<\/li>\n<li>retention limits<\/li>\n<li>redistribution limits<\/li>\n<li>geography constraints<\/li>\n<li>Build compliance into your pipeline:<\/li>\n<li>retention policies via S3 lifecycle<\/li>\n<li>access control via IAM\/Lake Formation<\/li>\n<li>data classification tags<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Common security mistakes<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Exporting vendor data into a broadly accessible \u201cshared bucket\u201d without controls.<\/li>\n<li>Allowing many developers to subscribe to products directly (no procurement governance).<\/li>\n<li>Using SSE-KMS but forgetting to grant Athena\/Glue read permissions, causing broken queries.<\/li>\n<li>Copying data across regions\/accounts without checking license terms and costs.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Secure deployment recommendations<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Use separate accounts for:<\/li>\n<li>procurement\/landing (data account)<\/li>\n<li>analytics consumption (analytics account)<\/li>\n<li>Use central KMS key management and standardized bucket policies.<\/li>\n<li>Automate policy checks (AWS Config rules, security-as-code).<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">13. Limitations and Gotchas<\/h2>\n\n\n\n<p>Because AWS Data Exchange is a managed subscription\/delivery service, many \u201cgotchas\u201d are about <strong>product differences<\/strong> and <strong>operational controls<\/strong> rather than raw performance.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Known limitations \/ constraints (verify current specifics)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Regional availability<\/strong>: AWS Data Exchange and specific products are region-scoped; not all products exist in all regions.<\/li>\n<li><strong>Product modality differences<\/strong>: file-based vs other delivery modalities behave differently; not every product supports every integration.<\/li>\n<li><strong>Schema drift<\/strong>: providers may change columns\/types across revisions; you must validate and handle drift.<\/li>\n<li><strong>Large asset handling<\/strong>: very large datasets can create long export times and significant S3 footprint. Verify any export\/job quotas in your account\/region.<\/li>\n<li><strong>SSE-KMS permissions complexity<\/strong>: misconfigured KMS policies are a frequent cause of export or query failures.<\/li>\n<li><strong>Retention vs licensing<\/strong>: storing every revision forever may violate license terms; implement retention policies aligned to agreements.<\/li>\n<li><strong>Athena scan costs<\/strong>: raw CSV\/JSON exports can be expensive to query repeatedly.<\/li>\n<li><strong>Unsubscribe behavior<\/strong>: unsubscribing typically doesn\u2019t delete data already exported to your S3 bucket\u2014your data governance must handle that.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Operational gotchas<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Failing to separate \u201craw vendor data\u201d from curated datasets can lead to analysts using raw data incorrectly.<\/li>\n<li>Lack of metadata tracking (revision ids, publish time) makes audits and reproducibility difficult.<\/li>\n<li>Mixing multiple datasets\/products in one prefix without a consistent naming scheme leads to crawler\/table confusion.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Migration challenges<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>If you previously ingested vendor data by SFTP\/API, migrating to AWS Data Exchange:<\/li>\n<li>requires validating that the dataset is identical (fields, update schedule)<\/li>\n<li>may change how you detect updates (revisions vs file timestamps)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Vendor-specific nuances<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Providers differ in:<\/li>\n<li>update frequency<\/li>\n<li>completeness\/backfills<\/li>\n<li>documentation quality<\/li>\n<li>file format conventions<\/li>\n<li>Always build <strong>data quality checks<\/strong> and treat vendor data as external input.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">14. Comparison with Alternatives<\/h2>\n\n\n\n<p>AWS Data Exchange is not the only way to obtain external data for Analytics. Here\u2019s how it compares.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Key alternatives<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>AWS Marketplace (general)<\/strong>: Marketplace is broader (software, AMIs, SaaS). AWS Data Exchange focuses on data product subscription and dataset\/revision\/asset handling.<\/li>\n<li><strong>AWS Open Data Registry \/ public S3 buckets<\/strong>: great for open datasets; lacks subscription entitlements and commercial workflows.<\/li>\n<li><strong>Direct vendor delivery<\/strong> (SFTP, API, cloud storage share): flexible but operationally heavy and inconsistent.<\/li>\n<li><strong>Snowflake Marketplace \/ Databricks Marketplace<\/strong>: strong if your primary analytics platform is Snowflake\/Databricks.<\/li>\n<li><strong>Azure Data Share \/ Google Analytics Hub<\/strong>: similar concepts in other clouds; best if you operate primarily in those clouds.<\/li>\n<li><strong>Open-source ingestion<\/strong> (Airbyte, Singer taps, custom pipelines): powerful but you own reliability, schema drift handling, and governance.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Comparison table<\/h3>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Option<\/th>\n<th>Best For<\/th>\n<th>Strengths<\/th>\n<th>Weaknesses<\/th>\n<th>When to Choose<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>AWS Data Exchange<\/td>\n<td>Subscribing to third-party datasets in AWS<\/td>\n<td>Subscription + entitlement + revision model; integrates with AWS analytics<\/td>\n<td>Product availability varies; still must build catalog\/query layers<\/td>\n<td>You want governed external data acquisition inside AWS<\/td>\n<\/tr>\n<tr>\n<td>AWS Open Data Registry \/ Public datasets on S3<\/td>\n<td>Open\/public datasets<\/td>\n<td>Often free; easy access<\/td>\n<td>No commercial terms\/entitlements; variable update practices<\/td>\n<td>You only need open data and can accept public-source constraints<\/td>\n<\/tr>\n<tr>\n<td>Direct vendor SFTP\/API<\/td>\n<td>Highly customized vendor relationships<\/td>\n<td>Maximum flexibility<\/td>\n<td>High ops burden; weak standardization; auditing harder<\/td>\n<td>Vendor not on ADX or needs bespoke integration<\/td>\n<\/tr>\n<tr>\n<td>Snowflake Marketplace<\/td>\n<td>Snowflake-centric analytics<\/td>\n<td>In-warehouse sharing patterns; strong for Snowflake users<\/td>\n<td>Less native if most workloads are on AWS lake patterns<\/td>\n<td>Your analytics stack is primarily Snowflake<\/td>\n<\/tr>\n<tr>\n<td>Databricks Marketplace<\/td>\n<td>Databricks-centric analytics<\/td>\n<td>Strong for lakehouse + notebooks<\/td>\n<td>Less ideal if you\u2019re not using Databricks as primary<\/td>\n<td>Your org standardizes on Databricks<\/td>\n<\/tr>\n<tr>\n<td>Azure Data Share<\/td>\n<td>Azure-first orgs<\/td>\n<td>Native to Azure sharing patterns<\/td>\n<td>Not AWS-native<\/td>\n<td>Your workloads are primarily in Azure<\/td>\n<\/tr>\n<tr>\n<td>Google Analytics Hub<\/td>\n<td>GCP-first orgs<\/td>\n<td>Native to BigQuery sharing<\/td>\n<td>Not AWS-native<\/td>\n<td>Your workloads are primarily in GCP<\/td>\n<\/tr>\n<tr>\n<td>Airbyte\/Singer\/custom ingestion<\/td>\n<td>Engineering-heavy orgs<\/td>\n<td>Works with many sources; customizable<\/td>\n<td>You own reliability\/security\/compliance; not a marketplace<\/td>\n<td>You need custom connectors or transformations beyond marketplace data<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">15. Real-World Example<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">Enterprise example (regulated financial services)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: A bank needs licensed market\/reference datasets for Analytics and risk modeling. Procurement requires auditability and strict access control. Data must be reproducible for model validation.<\/li>\n<li><strong>Proposed architecture<\/strong>:<\/li>\n<li>Procurement role subscribes to products in AWS Data Exchange.<\/li>\n<li>Data engineering role exports each revision to an encrypted S3 raw bucket (<code>raw\/vendor=...\/revision=...\/<\/code>).<\/li>\n<li>EventBridge triggers a Step Functions workflow:<ul>\n<li>export revision<\/li>\n<li>run data quality checks<\/li>\n<li>convert to Parquet (if license allows)<\/li>\n<li>update Glue tables and partitions<\/li>\n<\/ul>\n<\/li>\n<li>Lake Formation (optional) restricts table access by business domain.<\/li>\n<li>Analysts query curated tables using Athena; risk models run in SageMaker\/EMR.<\/li>\n<li>CloudTrail retained for audit; S3 lifecycle enforces retention per license.<\/li>\n<li><strong>Why AWS Data Exchange was chosen<\/strong>:<\/li>\n<li>Standard subscription and entitlement model aligned with governance requirements.<\/li>\n<li>Revision-based updates support reproducibility and audit trails.<\/li>\n<li><strong>Expected outcomes<\/strong>:<\/li>\n<li>Faster onboarding of new datasets<\/li>\n<li>Repeatable monthly\/weekly updates<\/li>\n<li>Improved audit posture and reduced operational risk<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Startup \/ small-team example (lean product analytics)<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Problem<\/strong>: A startup wants to enrich product usage analytics with external demographic or geospatial context, but has limited engineering bandwidth.<\/li>\n<li><strong>Proposed architecture<\/strong>:<\/li>\n<li>Subscribe to one or two data products (prefer free or low-cost).<\/li>\n<li>Export to a single S3 bucket.<\/li>\n<li>Use Glue crawler to catalog and Athena to query\/join with internal events (also in S3).<\/li>\n<li>Schedule a simple monthly refresh reminder or a lightweight EventBridge+Lambda automation later.<\/li>\n<li><strong>Why AWS Data Exchange was chosen<\/strong>:<\/li>\n<li>Quick time-to-value; minimal custom vendor integration.<\/li>\n<li>Works with serverless Athena to avoid managing clusters.<\/li>\n<li><strong>Expected outcomes<\/strong>:<\/li>\n<li>Enriched dashboards within days, not weeks<\/li>\n<li>Controlled costs by staying serverless and limiting scans<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">16. FAQ<\/h2>\n\n\n\n<p>1) <strong>Is AWS Data Exchange the same as AWS Marketplace?<\/strong><br\/>\nNo. AWS Marketplace is the broader commerce\/catalog platform for software and data products. AWS Data Exchange provides the dataset\/revision\/asset model and data delivery workflows that many Marketplace data products use.<\/p>\n\n\n\n<p>2) <strong>Do I always export data to S3?<\/strong><br\/>\nNot always. Many products are file-based and export to S3, which is the most common pattern. Some products may use other delivery modalities. Check the product listing and official docs for the supported method.<\/p>\n\n\n\n<p>3) <strong>Can I query AWS Data Exchange data directly without copying?<\/strong><br\/>\nFor file-based products, you typically export to your S3 bucket first. Some offerings may support alternative access methods. Verify for your product.<\/p>\n\n\n\n<p>4) <strong>Does unsubscribing delete data already exported to my bucket?<\/strong><br\/>\nTypically, no. Data already in your S3 bucket remains until you delete it. Your license terms and governance policy should define retention and deletion requirements.<\/p>\n\n\n\n<p>5) <strong>How do I know when a dataset updates?<\/strong><br\/>\nAWS Data Exchange supports notifications for new revisions (commonly integrated with Amazon EventBridge). Verify the exact configuration steps in official docs.<\/p>\n\n\n\n<p>6) <strong>Can I automate exports when a new revision is published?<\/strong><br\/>\nYes, commonly by combining EventBridge with Lambda or Step Functions to trigger export workflows and downstream catalog updates.<\/p>\n\n\n\n<p>7) <strong>What\u2019s the difference between a dataset and a revision?<\/strong><br\/>\nA dataset is the logical container. A revision is a versioned snapshot\/update of that dataset. Revisions contain assets.<\/p>\n\n\n\n<p>8) <strong>What file formats should I expect?<\/strong><br\/>\nIt depends on the provider: CSV, JSON, Parquet, GeoJSON, compressed archives, etc. Always review the product documentation and sample data if available.<\/p>\n\n\n\n<p>9) <strong>How do I handle schema changes across revisions?<\/strong><br\/>\nImplement schema validation and drift handling. Keep revision-specific paths and consider versioned tables or views in Glue\/Athena.<\/p>\n\n\n\n<p>10) <strong>Can I share the exported data with other accounts?<\/strong><br\/>\nTechnically you can share S3 data (and Glue tables) across accounts, but you must check the data product\u2019s license terms and your organization\u2019s governance policies before sharing.<\/p>\n\n\n\n<p>11) <strong>Is AWS Data Exchange suitable for real-time streaming data?<\/strong><br\/>\nGenerally it\u2019s aimed at subscription-based dataset delivery and updates, not high-frequency streaming ingestion. Use Kinesis\/MSK for streaming patterns.<\/p>\n\n\n\n<p>12) <strong>How do I control who can subscribe to new products?<\/strong><br\/>\nUse IAM and organizational controls (SCPs) to restrict Marketplace and AWS Data Exchange subscription actions to approved roles.<\/p>\n\n\n\n<p>13) <strong>What are the biggest cost risks?<\/strong><br\/>\nAthena scanning large raw files repeatedly, storing many revisions without lifecycle policies, and cross-region\/cross-account duplication. Also the data product subscription price if it\u2019s paid.<\/p>\n\n\n\n<p>14) <strong>How do I ensure exported data is encrypted?<\/strong><br\/>\nEnable default bucket encryption (SSE-S3 or SSE-KMS). If using SSE-KMS, ensure KMS policies allow required writes\/reads.<\/p>\n\n\n\n<p>15) <strong>Is AWS Data Exchange a data quality tool?<\/strong><br\/>\nNo. It delivers data. You should implement data quality checks using Glue, Deequ, Great Expectations, or your preferred validation approach.<\/p>\n\n\n\n<p>16) <strong>Can I use AWS Data Exchange with a lakehouse table format (Iceberg\/Hudi\/Delta)?<\/strong><br\/>\nAWS Data Exchange delivers datasets; you can transform landed files into your preferred table format in curated zones if license terms permit.<\/p>\n\n\n\n<p>17) <strong>Do I need Glue to use the data?<\/strong><br\/>\nNo, but it\u2019s commonly used for cataloging. You can also define Athena tables manually or load into other systems.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">17. Top Online Resources to Learn AWS Data Exchange<\/h2>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Resource Type<\/th>\n<th>Name<\/th>\n<th>Why It Is Useful<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Official documentation<\/td>\n<td>AWS Data Exchange Docs \u2014 https:\/\/docs.aws.amazon.com\/data-exchange\/<\/td>\n<td>Authoritative reference for concepts, APIs, permissions, and workflows<\/td>\n<\/tr>\n<tr>\n<td>Official product page<\/td>\n<td>AWS Data Exchange \u2014 https:\/\/aws.amazon.com\/data-exchange\/<\/td>\n<td>High-level overview and capabilities<\/td>\n<\/tr>\n<tr>\n<td>Official pricing<\/td>\n<td>AWS Data Exchange Pricing \u2014 https:\/\/aws.amazon.com\/data-exchange\/pricing\/<\/td>\n<td>Explains pricing model and what you pay for<\/td>\n<\/tr>\n<tr>\n<td>AWS Marketplace<\/td>\n<td>AWS Marketplace \u2014 https:\/\/aws.amazon.com\/marketplace\/<\/td>\n<td>Where many ADX data products are listed and subscribed<\/td>\n<\/tr>\n<tr>\n<td>Getting started (official)<\/td>\n<td>AWS Data Exchange Getting Started (see docs index) \u2014 https:\/\/docs.aws.amazon.com\/data-exchange\/latest\/userguide\/what-is-data-exchange.html<\/td>\n<td>Step-by-step orientation for subscriber\/provider concepts<\/td>\n<\/tr>\n<tr>\n<td>API\/CLI reference (official)<\/td>\n<td>AWS Data Exchange API Reference \u2014 https:\/\/docs.aws.amazon.com\/data-exchange\/latest\/apireference\/welcome.html<\/td>\n<td>Details operations used for automation (jobs, revisions, assets)<\/td>\n<\/tr>\n<tr>\n<td>Event-driven integration<\/td>\n<td>Amazon EventBridge Docs \u2014 https:\/\/docs.aws.amazon.com\/eventbridge\/<\/td>\n<td>Used to automate new revision processing<\/td>\n<\/tr>\n<tr>\n<td>Analytics consumption<\/td>\n<td>Amazon Athena Docs \u2014 https:\/\/docs.aws.amazon.com\/athena\/<\/td>\n<td>Query exported datasets on S3<\/td>\n<\/tr>\n<tr>\n<td>Cataloging<\/td>\n<td>AWS Glue Docs \u2014 https:\/\/docs.aws.amazon.com\/glue\/<\/td>\n<td>Build tables\/catalog and ETL for curated layers<\/td>\n<\/tr>\n<tr>\n<td>Pricing calculator<\/td>\n<td>AWS Pricing Calculator \u2014 https:\/\/calculator.aws\/#\/<\/td>\n<td>Model S3\/Athena\/Glue\/Redshift costs around your dataset usage<\/td>\n<\/tr>\n<tr>\n<td>Videos (official)<\/td>\n<td>AWS YouTube Channel \u2014 https:\/\/www.youtube.com\/@amazonwebservices<\/td>\n<td>Search for \u201cAWS Data Exchange\u201d sessions and demos<\/td>\n<\/tr>\n<tr>\n<td>Samples (community\/varies)<\/td>\n<td>AWS Samples on GitHub \u2014 https:\/\/github.com\/awslabs and https:\/\/github.com\/aws-samples<\/td>\n<td>Look for ADX automation patterns; validate recency and security before use<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">18. Training and Certification Providers<\/h2>\n\n\n\n<p>Exactly the following institutes are listed as training resources. Verify current course availability and delivery mode on their websites.<\/p>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Institute<\/th>\n<th>Suitable Audience<\/th>\n<th>Likely Learning Focus<\/th>\n<th>Mode<\/th>\n<th>Website URL<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>DevOpsSchool.com<\/td>\n<td>DevOps, cloud engineers, platform teams<\/td>\n<td>AWS fundamentals, DevOps, cloud operations; may include analytics tooling<\/td>\n<td>Check website<\/td>\n<td>https:\/\/www.devopsschool.com\/<\/td>\n<\/tr>\n<tr>\n<td>ScmGalaxy.com<\/td>\n<td>Beginners to intermediate IT professionals<\/td>\n<td>DevOps\/SCM, cloud basics, operational practices<\/td>\n<td>Check website<\/td>\n<td>https:\/\/www.scmgalaxy.com\/<\/td>\n<\/tr>\n<tr>\n<td>CLoudOpsNow.in<\/td>\n<td>Cloud operations and engineering teams<\/td>\n<td>Cloud ops, automation, reliability practices<\/td>\n<td>Check website<\/td>\n<td>https:\/\/cloudopsnow.in\/<\/td>\n<\/tr>\n<tr>\n<td>SreSchool.com<\/td>\n<td>SREs, ops teams, reliability engineers<\/td>\n<td>SRE practices, monitoring, reliability engineering<\/td>\n<td>Check website<\/td>\n<td>https:\/\/sreschool.com\/<\/td>\n<\/tr>\n<tr>\n<td>AiOpsSchool.com<\/td>\n<td>Ops teams adopting automation<\/td>\n<td>AIOps concepts, automation, monitoring analytics<\/td>\n<td>Check website<\/td>\n<td>https:\/\/aiopsschool.com\/<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">19. Top Trainers<\/h2>\n\n\n\n<p>The following trainer-related sites are provided as learning resources. Verify offerings and expertise directly on each site.<\/p>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Platform\/Site<\/th>\n<th>Likely Specialization<\/th>\n<th>Suitable Audience<\/th>\n<th>Website URL<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>RajeshKumar.xyz<\/td>\n<td>Cloud\/DevOps training content<\/td>\n<td>Students, engineers seeking guided training<\/td>\n<td>https:\/\/rajeshkumar.xyz\/<\/td>\n<\/tr>\n<tr>\n<td>devopstrainer.in<\/td>\n<td>DevOps tooling and practices<\/td>\n<td>Beginners to working professionals<\/td>\n<td>https:\/\/devopstrainer.in\/<\/td>\n<\/tr>\n<tr>\n<td>devopsfreelancer.com<\/td>\n<td>Independent DevOps consulting\/training<\/td>\n<td>Teams needing practical, hands-on help<\/td>\n<td>https:\/\/devopsfreelancer.com\/<\/td>\n<\/tr>\n<tr>\n<td>devopssupport.in<\/td>\n<td>DevOps support and training<\/td>\n<td>Ops\/engineering teams needing troubleshooting guidance<\/td>\n<td>https:\/\/devopssupport.in\/<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">20. Top Consulting Companies<\/h2>\n\n\n\n<p>Exactly the following consulting companies are listed. Descriptions are general; confirm detailed capabilities directly with each company.<\/p>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Company Name<\/th>\n<th>Likely Service Area<\/th>\n<th>Where They May Help<\/th>\n<th>Consulting Use Case Examples<\/th>\n<th>Website URL<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>cotocus.com<\/td>\n<td>Cloud\/DevOps\/engineering services<\/td>\n<td>Architecture, implementation support, automation<\/td>\n<td>Set up S3 data lake landing zone, governance guardrails, ingestion automation<\/td>\n<td>https:\/\/cotocus.com\/<\/td>\n<\/tr>\n<tr>\n<td>DevOpsSchool.com<\/td>\n<td>DevOps and cloud consulting\/training<\/td>\n<td>Enablement, platform engineering support<\/td>\n<td>Build CI\/CD for data pipelines, operational best practices for analytics stacks<\/td>\n<td>https:\/\/www.devopsschool.com\/<\/td>\n<\/tr>\n<tr>\n<td>DEVOPSCONSULTING.IN<\/td>\n<td>DevOps and cloud consulting<\/td>\n<td>Ops modernization, automation, reliability<\/td>\n<td>Implement monitoring\/logging around data ingestion and analytics workloads<\/td>\n<td>https:\/\/devopsconsulting.in\/<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">21. Career and Learning Roadmap<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">What to learn before AWS Data Exchange<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>AWS IAM fundamentals (roles, policies, least privilege)<\/li>\n<li>Amazon S3 fundamentals (encryption, bucket policies, lifecycle)<\/li>\n<li>Basic Analytics concepts on AWS:<\/li>\n<li>Athena + Glue Data Catalog<\/li>\n<li>Data lake folder\/prefix design<\/li>\n<li>AWS billing basics (cost allocation tags, cost explorer)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">What to learn after AWS Data Exchange<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Event-driven automation:<\/li>\n<li>EventBridge + Lambda + Step Functions<\/li>\n<li>Data engineering on AWS:<\/li>\n<li>Glue ETL, EMR\/Spark<\/li>\n<li>Data quality frameworks (e.g., Deequ\/Great Expectations)<\/li>\n<li>Governance:<\/li>\n<li>Lake Formation permissions (optional but common in enterprises)<\/li>\n<li>Data classification and access reviews<\/li>\n<li>Warehouse integration:<\/li>\n<li>Redshift loading patterns, Spectrum, performance tuning<\/li>\n<li>FinOps for Analytics:<\/li>\n<li>Athena scan optimization<\/li>\n<li>S3 storage optimization and lifecycle<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Job roles that use it<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Data Engineer \/ Analytics Engineer<\/li>\n<li>Cloud Engineer (data platform)<\/li>\n<li>Solutions Architect (analytics)<\/li>\n<li>Data Platform Engineer<\/li>\n<li>Security Engineer (data governance)<\/li>\n<li>FinOps Analyst (data\/analytics cost governance)<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Certification path (AWS)<\/h3>\n\n\n\n<p>AWS Data Exchange is usually covered as part of broader analytics knowledge rather than a single dedicated certification. Consider:\n&#8211; AWS Certified Data Engineer \u2013 Associate (if available in your track; verify current AWS certification catalog)\n&#8211; AWS Certified Solutions Architect \u2013 Associate\/Professional\n&#8211; AWS Certified Security \u2013 Specialty (for governance-heavy roles)<\/p>\n\n\n\n<p>Always confirm current certification names and availability:\n&#8211; https:\/\/aws.amazon.com\/certification\/<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">Project ideas for practice<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Build a \u201cvendor data landing pipeline\u201d:<\/li>\n<li>EventBridge \u2192 Step Functions \u2192 export revision \u2192 Glue crawl \u2192 Athena views<\/li>\n<li>Implement schema drift detection across revisions and alert on changes.<\/li>\n<li>Create a cost-optimized curated zone:<\/li>\n<li>Convert CSV to Parquet, partition by date, enforce lifecycle\/retention.<\/li>\n<li>Build a metadata inventory:<\/li>\n<li>track product\/dataset\/revision ids and ingestion status in DynamoDB.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\">22. Glossary<\/h2>\n\n\n\n<figure class=\"wp-block-table\"><table>\n<thead>\n<tr>\n<th>Term<\/th>\n<th>Definition<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>AWS Data Exchange<\/td>\n<td>AWS service for subscribing to and consuming third-party data products on AWS<\/td>\n<\/tr>\n<tr>\n<td>Data product<\/td>\n<td>The subscribe-able package containing datasets plus commercial terms<\/td>\n<\/tr>\n<tr>\n<td>Dataset<\/td>\n<td>A logical container of data within a product<\/td>\n<\/tr>\n<tr>\n<td>Revision<\/td>\n<td>A versioned snapshot\/update of a dataset<\/td>\n<\/tr>\n<tr>\n<td>Asset<\/td>\n<td>A concrete deliverable item within a revision, often a file<\/td>\n<\/tr>\n<tr>\n<td>Entitlement<\/td>\n<td>The granted right to access a subscribed product\u2019s datasets<\/td>\n<\/tr>\n<tr>\n<td>Export (to S3)<\/td>\n<td>Copying entitled assets into your S3 bucket for consumption<\/td>\n<\/tr>\n<tr>\n<td>Landing zone<\/td>\n<td>The initial storage location for ingested data (commonly S3 raw\/quarantine)<\/td>\n<\/tr>\n<tr>\n<td>Glue Data Catalog<\/td>\n<td>Central metadata store for tables\/schemas used by Athena and other services<\/td>\n<\/tr>\n<tr>\n<td>Athena<\/td>\n<td>Serverless SQL query service over data in S3<\/td>\n<\/tr>\n<tr>\n<td>SSE-S3<\/td>\n<td>S3-managed server-side encryption using AES-256<\/td>\n<\/tr>\n<tr>\n<td>SSE-KMS<\/td>\n<td>Server-side encryption using AWS KMS keys, enabling key-level access controls<\/td>\n<\/tr>\n<tr>\n<td>EventBridge<\/td>\n<td>Event bus used to route events such as \u201cnew revision available\u201d to automation<\/td>\n<\/tr>\n<tr>\n<td>Schema drift<\/td>\n<td>Changes to columns\/types\/structure between dataset revisions<\/td>\n<\/tr>\n<tr>\n<td>Lifecycle policy<\/td>\n<td>S3 rules to transition or expire objects to control storage cost and retention<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">23. Summary<\/h2>\n\n\n\n<p>AWS Data Exchange is AWS\u2019s managed service for <strong>discovering, subscribing to, and consuming external data products<\/strong> for Analytics. It matters because it standardizes the messy \u201cdata procurement + delivery\u201d problem into an AWS-native workflow using <strong>datasets, revisions, and assets<\/strong>, enabling repeatable ingestion and better governance.<\/p>\n\n\n\n<p>It fits best at the <strong>data acquisition layer<\/strong> of your AWS analytics platform, typically landing data into <strong>Amazon S3<\/strong> and then leveraging <strong>AWS Glue<\/strong> and <strong>Amazon Athena<\/strong> (or Redshift\/EMR\/SageMaker) for downstream processing and insights.<\/p>\n\n\n\n<p>Cost and security success comes from:\n&#8211; understanding that <strong>providers set data product prices<\/strong>, while you pay AWS for storage\/compute\/query\n&#8211; controlling S3 destinations, encryption (often SSE-KMS), IAM permissions, and audit trails\n&#8211; optimizing Athena\/Glue usage to avoid unnecessary scanning and storage growth<\/p>\n\n\n\n<p>Use AWS Data Exchange when you need governed, subscription-based access to third-party datasets inside AWS. Next step: build an automated revision-ingestion pipeline with EventBridge and validate schemas and costs as the dataset grows.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Analytics<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[21,20],"tags":[],"class_list":["post-117","post","type-post","status-publish","format-standard","hentry","category-analytics","category-aws"],"_links":{"self":[{"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/posts\/117","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/comments?post=117"}],"version-history":[{"count":0,"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/posts\/117\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/media?parent=117"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/categories?post=117"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.devopsschool.com\/tutorials\/wp-json\/wp\/v2\/tags?post=117"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}