HomeBig DataHow Fannie Mae constructed an information mesh structure to allow self-service utilizing...

How Fannie Mae constructed an information mesh structure to allow self-service utilizing Amazon Redshift knowledge sharing

This submit is co-written by Kiran Ramineni and Basava Hubli, from Fannie Mae.

Amazon Redshift knowledge sharing allows on the spot, granular, and quick knowledge entry throughout Amazon Redshift clusters with out the necessity to copy or transfer knowledge round. Information sharing supplies dwell entry to knowledge in order that customers all the time see probably the most up-to-date and transactionally constant views of information throughout all customers as knowledge is up to date within the producer. You may share dwell knowledge securely with Amazon Redshift clusters in the identical or completely different AWS accounts, and throughout Areas. Information sharing allows safe and ruled collaboration inside and throughout organizations in addition to exterior events.

On this submit, we see how Fannie Mae carried out an information mesh structure utilizing Amazon Redshift cross-account knowledge sharing to interrupt down the silos in knowledge warehouses throughout enterprise items.

About Fannie Mae

Chartered by U.S. Congress in 1938, Fannie Mae advances equitable and sustainable entry to homeownership and high quality, inexpensive rental housing for tens of millions of individuals throughout America. Fannie Mae allows the 30-year fixed-rate mortgage and drives accountable innovation to make homebuying and renting simpler, fairer, and extra accessible. We’re targeted on rising operational agility and effectivity, accelerating the digital transformation of the corporate to ship extra worth and dependable, trendy platforms in help of the broader housing finance system.


To meet the mission of facilitating equitable and sustainable entry to homeownership and high quality, inexpensive rental housing throughout America, Fannie Mae embraced a contemporary cloud-based structure which leverages knowledge to drive actionable insights and enterprise selections. As a part of the modernization technique, we launched into a journey emigrate our legacy on-premises workloads to AWS cloud together with managed providers equivalent to Amazon Redshift and Amazon S3. The fashionable knowledge platform on AWS cloud serves because the central knowledge retailer for analytics, analysis, and knowledge science. As well as, this platform additionally serves for governance, regulatory and monetary experiences.

To handle capability, scalability and elasticity wants of a giant knowledge footprint of over 4PB, we decentralized and delegated possession of the info shops and related administration capabilities to their respective enterprise items. To allow decentralization, and environment friendly knowledge entry and administration, we adopted an information mesh structure.

Information mesh answer structure

To allow a seamless entry to knowledge throughout accounts and enterprise items, we checked out varied choices to construct an structure that’s sustainable and scalable. The information mesh structure allowed us to maintain knowledge of the respective enterprise items in their very own accounts, however but allow a seamless entry throughout the enterprise unit accounts in a safe method.  We reorganized the AWS account construction to have separate accounts for every of the enterprise items whereby, enterprise knowledge and dependent purposes had been collocated of their respective AWS Accounts.

With this decentralized mannequin, the enterprise items independently handle the duty of hydration, curation and safety of their knowledge.  Nonetheless, there’s a vital have to allow seamless and environment friendly entry to knowledge throughout enterprise items and a capability to control the info utilization. Amazon Redshift cross-account knowledge sharing meets this want and allows us with enterprise continuity.

To facilitate the self-serve functionality on the info mesh, we constructed an internet portal that permits for knowledge discovery and talent to subscribe to knowledge within the Amazon Redshift knowledge warehouse and Amazon Easy Storage Service (Amazon S3) knowledge lake (lake home). As soon as a client initiates a request on the net portal, an approval workflow is triggered with notification to the governance and enterprise knowledge proprietor. Upon profitable completion of the request workflow, an automation course of is triggered to grant entry to the patron, and a notification is shipped to the patron. Subsequently, the patron is ready to entry the requested datasets. The workflow technique of request, approval, and subsequent provisioning of entry was automated utilizing APIs and AWS Command Line Interface (AWS CLI) instructions, and full course of is designed to finish inside a couple of minutes.

With this new structure utilizing Amazon Redshift cross-account knowledge sharing, we had been ready implement and profit from the next key ideas of an information mesh structure that match very nicely for our use case:

  • An information as a product method
  • A federated mannequin of information possession
  • The power for customers to subscribe utilizing self-service knowledge entry
  • Federated knowledge governance with the flexibility to grant and revoke entry

The next structure diagram exhibits the high-level knowledge mesh structure we carried out at Fannie Mae. Information from every of the operational methods is collected and saved in particular person lake homes and subscriptions are managed by way of an information mesh catalog in a centralized management aircraft account.

Fig 1. High level Data Mesh catalog architecture

Fig 1. Excessive degree Information Mesh catalog structure

Management aircraft for knowledge mesh

With a redesigned account construction, knowledge are unfold out throughout separate accounts for every enterprise software space in S3 knowledge lake or in Amazon Redshift cluster. We designed a hub and spoke point-to-point knowledge distribution scheme with a centralized semantic search to reinforce the info relevance. We use a centralized management aircraft account to retailer the catalog data, contract element, approval workflow insurance policies, and entry administration particulars for the info mesh. With a coverage pushed entry paradigm, we allow fine-grained entry administration to the info, the place we automated Information as a Service enablement with an optimized method. It has three modules to retailer and handle catalog, contracts, and entry administration.

Information catalog

The information catalog supplies the info glossary and catalog data, and helps absolutely fulfill governance and safety requirements. With AWS Glue crawlers, we create the catalog for the lake home in a centralized management aircraft account, after which we automate the sharing course of in a safe method. This permits a query-based framework to pinpoint the precise location of the info. The information catalog collects the runtime details about the datasets for indexing functions, and supplies runtime metrics for analytics on dataset utilization and entry patterns. The catalog additionally supplies a mechanism to replace the catalog by way of automation as new datasets turn into obtainable.

Contract registry

The contract registry hosts the coverage engine, and makes use of Amazon DynamoDB to retailer the registry insurance policies. This has the main points on entitlements to bodily mapping of information, and workflows for the entry administration course of. We additionally use this to retailer and preserve the registry of present knowledge contracts and allow audit functionality to find out and monitor the entry patterns. As well as, the contract registry serves as the shop for state administration performance.

Entry administration automation

Controlling and managing entry to the dataset is finished by way of entry administration. This supplies a just-in-time knowledge entry by way of IAM session insurance policies utilizing a persona-driven method. The entry administration module additionally hosts occasion notification for knowledge, equivalent to frequency of entry or variety of reads, and we then harness this data for knowledge entry lifecycle administration. This module performs a vital function within the state administration and supplies intensive logging and monitoring capabilities on the state of the info.

Course of movement of information mesh utilizing Amazon Redshift cross-account knowledge sharing

The method movement begins with making a catalog of all datasets obtainable within the management aircraft account. Shoppers can request entry to the info by way of an internet front-end catalog, and the approval course of is triggered by way of the central management aircraft account. The next structure diagram represents the high-level implementation of Amazon Redshift knowledge sharing by way of the info mesh structure. The steps of the method movement are as follows:

  1. All the info merchandise, Amazon Redshift tables, and S3 buckets are registered in a centralized AWS Glue Information Catalog.
  2. Information scientists and LOB customers can browse the Information Catalog to search out the info merchandise obtainable throughout all lake homes in Fannie Mae.
  3. Enterprise purposes can devour the info in different lake homes by registering a client contract. For instance, LOB1-Lakehouse can register the contract to make the most of knowledge from LOB3-Lakehouse.
  4. The contract is reviewed and permitted by the info producer, which subsequently triggers a technical occasion by way of Amazon Easy Service Notification (Amazon SNS).
  5. The subscribing AWS Lambda perform runs AWS CLI instructions, ACLs, and IAM insurance policies to arrange Amazon Redshift knowledge sharing and make knowledge obtainable for customers.
  6. Shoppers can entry the subscribed Amazon Redshift cluster knowledge utilizing their very own cluster.
Fig 2. Data Mesh architecture using Amazon Redshift data sharing

Fig 2. Information Mesh structure utilizing Amazon Redshift knowledge sharing

The intention of this submit is to not present detailed steps for each facet of making the info mesh, however to offer a high-level overview of the structure carried out, and the way you need to use varied analytics providers and third-party instruments to create a scalable knowledge mesh with Amazon Redshift and Amazon S3. If you wish to check out creating this structure, you need to use these steps and automate the method utilizing your instrument of selection for the front-end consumer interface to allow customers to subscribe to the dataset.

The steps we describe listed below are a simplified model of the particular implementation, so it doesn’t contain all of the instruments and accounts. To arrange this scaled-down knowledge mesh structure, we reveal utilizing cross-account knowledge sharing utilizing one management aircraft account and two client accounts. For this, you need to have the next stipulations:

  • Three AWS accounts, one for the producer <ProducerAWSAccount1>, and two client accounts: <ConsumerAWSAccount1> and <ConsumerAWSAccount2>
  • AWS permissions to provision Amazon Redshift and create an IAM function and coverage
  • The required Amazon Redshift clusters: one for the producer within the producer AWS account, a cluster in ConsumerCluster1, and optionally a cluster in ConsumerCluster2
  • Two customers within the producer account, and two customers in client account 1:
    • ProducerClusterAdmin – The Amazon Redshift consumer with admin entry on the producer cluster
    • ProducerCloudAdmin – The IAM consumer or function with rights to run authorize-data-share and deauthorize-data-share AWS CLI instructions within the producer account
    • Consumer1ClusterAdmin – The Amazon Redshift consumer with admin entry on the patron cluster
    • Consumer1CloudAdmin – The IAM consumer or function with rights to run associate-data-share-consumer and disassociate-data-share-consumer AWS CLI instructions within the client account

Implement the answer

On the Amazon Redshift console, log in to the producer cluster and run the next statements utilizing the question editor:



For sharing knowledge throughout AWS accounts, you need to use the next GRANT USAGE command. For authorizing the info share, sometimes it is going to be accomplished by a supervisor or approver. On this case, we present how one can automate this course of utilizing the AWS CLI command authorize-data-share.


aws redshift authorize-data-share --data-share-arn <DATASHARE ARN> --consumer-identifier <CONSUMER ACCOUNT>

For the patron to entry the shared knowledge from producer, an administrator on the patron account must affiliate the info share with a number of clusters. This may be accomplished utilizing the Amazon Redshift console or AWS CLI instructions. We offer the next AWS CLI command as a result of that is how one can automate the method from the central management aircraft account:

aws redshift associate-data-share-consumer --no-associate-entire-account --data-share-arn <DATASHARE ARN> --consumer-arn <CONSUMER CLUSTER ARN>

/* Create Database in Client Account */

Optionally available:

GRANT USAGE ON DATABASE ds_db TO consumer/group;

/* Optionally available:Grant utilization on database to customers or teams */
GRANT USAGE ON SCHEMA Schema_from_datashare TO GROUP Analyst_group;

To allow Amazon Redshift Spectrum cross-account entry to AWS Glue and Amazon S3, and the IAM roles required, discuss with How can I create Amazon Redshift Spectrum cross-account entry to AWS Glue and Amazon S3.


Amazon Redshift knowledge sharing supplies a easy, seamless, and safe platform for sharing knowledge in a domain-oriented distributed knowledge mesh structure. Fannie Mae deployed the Amazon Redshift knowledge sharing functionality throughout the info lake and knowledge mesh platforms, which presently hosts over 4 petabytes value of enterprise knowledge. The potential has been seamlessly built-in with their Simply-In-Time (JIT) knowledge provisioning framework enabling a single-click, persona-driven entry to knowledge. Additional, Amazon Redshift knowledge sharing coupled with Fannie Mae’s centralized, policy-driven knowledge governance framework enormously simplified entry to knowledge within the lake ecosystem whereas absolutely conforming to the stringent knowledge governance insurance policies and requirements. This demonstrates that Amazon Redshift customers can create knowledge share as product to distribute throughout many knowledge domains.

In abstract, Fannie Mae was capable of efficiently combine the info sharing functionality of their knowledge ecosystem to convey efficiencies in knowledge democratization and introduce a better velocity, close to real-time entry to knowledge throughout varied enterprise items. We encourage you to discover the info sharing characteristic of Amazon Redshift to construct your personal knowledge mesh structure and enhance entry to knowledge for what you are promoting customers.

Concerning the authors

Kiran Ramineni is Fannie Mae’s Vice President Head of Single Household, Cloud, Information, ML/AI & Infrastructure Structure, reporting to the CTO and Chief Architect. Kiran and group spear headed cloud scalable Enterprise Information Mesh (Information Lake) with help for Simply-In-Time (JIT), and Zero Belief because it applies to Citizen Information Scientist and Citizen Information Engineers. Prior to now Kiran constructed/lead a number of web scalable always-on platforms.

Basava Hubli is a Director & Lead Information/ML Architect at Enterprise Structure. He oversees the Technique and Structure of Enterprise Information, Analytics and Information Science platforms at Fannie Mae. His major focus is on Structure Oversight and Supply of Modern technical capabilities that clear up for vital Enterprise enterprise wants. He leads a passionate and motivated group of architects who’re driving the modernization and adoption of the Information, Analytics and ML platforms on Cloud. Underneath his management, Enterprise Structure has efficiently deployed a number of scalable, modern platforms & capabilities that features, a fully-governed Information Mesh which hosts peta-byte scale enterprise knowledge and a persona-driven, zero-trust based mostly knowledge entry administration framework which solves for the group’s knowledge democratization wants.

Rajesh Francis is a Senior Analytics Buyer Expertise Specialist at AWS. He makes a speciality of Amazon Redshift and focuses on serving to to drive AWS market and technical technique for knowledge warehousing and analytics. Rajesh works carefully with massive strategic clients to assist them undertake our new providers and options, develop long-term partnerships, and feed buyer necessities again to our product growth groups to information the course of our product choices.

Kiran Sharma is a Senior Information Architect in AWS Skilled Companies. Kiran helps clients architecting, implementing and optimizing peta-byte scale Large Information Options on AWS.


Most Popular

Recent Comments