Search
00
GBAF Logo
trophy
Top StoriesInterviewsBusinessFinanceBankingTechnologyInvestingTradingVideosAwardsMagazinesHeadlinesTrends

Subscribe to our newsletter

Get the latest news and updates from our team.

Global Banking and Finance Review

Global Banking & Finance Review

Company

    GBAF Logo
    • About Us
    • Profile
    • Privacy & Cookie Policy
    • Terms of Use
    • Contact Us
    • Advertising
    • Submit Post
    • Latest News
    • Research Reports
    • Press Release
    • Awards▾
      • About the Awards
      • Awards TimeTable
      • Submit Nominations
      • Testimonials
      • Media Room
      • Award Winners
      • FAQ
    • Magazines▾
      • Global Banking & Finance Review Magazine Issue 79
      • Global Banking & Finance Review Magazine Issue 78
      • Global Banking & Finance Review Magazine Issue 77
      • Global Banking & Finance Review Magazine Issue 76
      • Global Banking & Finance Review Magazine Issue 75
      • Global Banking & Finance Review Magazine Issue 73
      • Global Banking & Finance Review Magazine Issue 71
      • Global Banking & Finance Review Magazine Issue 70
      • Global Banking & Finance Review Magazine Issue 69
      • Global Banking & Finance Review Magazine Issue 66
    Top StoriesInterviewsBusinessFinanceBankingTechnologyInvestingTradingVideosAwardsMagazinesHeadlinesTrends

    Global Banking & Finance Review® is a leading financial portal and online magazine offering News, Analysis, Opinion, Reviews, Interviews & Videos from the world of Banking, Finance, Business, Trading, Technology, Investing, Brokerage, Foreign Exchange, Tax & Legal, Islamic Finance, Asset & Wealth Management.
    Copyright © 2010-2025 GBAF Publications Ltd - All Rights Reserved.

    Editorial & Advertiser disclosure

    Global Banking and Finance Review is an online platform offering news, analysis, and opinion on the latest trends, developments, and innovations in the banking and finance industry worldwide. The platform covers a diverse range of topics, including banking, insurance, investment, wealth management, fintech, and regulatory issues. The website publishes news, press releases, opinion and advertorials on various financial organizations, products and services which are commissioned from various Companies, Organizations, PR agencies, Bloggers etc. These commissioned articles are commercial in nature. This is not to be considered as financial advice and should be considered only for information purposes. It does not reflect the views or opinion of our website and is not to be considered an endorsement or a recommendation. We cannot guarantee the accuracy or applicability of any information provided with respect to your individual or personal circumstances. Please seek Professional advice from a qualified professional before making any financial decisions. We link to various third-party websites, affiliate sales networks, and to our advertising partners websites. When you view or click on certain links available on our articles, our partners may compensate us for displaying the content to you or make a purchase or fill a form. This will not incur any additional charges to you. To make things simpler for you to identity or distinguish advertised or sponsored articles or links, you may consider all articles or links hosted on our site as a commercial article placement. We will not be responsible for any loss you may suffer as a result of any omission or inaccuracy on the website.

    Home > Banking > CAN THE NEW DATA LAKE PLATFORMS SOLVE BANKING’S BIGGEST ANALYTICS DILEMMA?
    Banking

    CAN THE NEW DATA LAKE PLATFORMS SOLVE BANKING’S BIGGEST ANALYTICS DILEMMA?

    CAN THE NEW DATA LAKE PLATFORMS SOLVE BANKING’S BIGGEST ANALYTICS DILEMMA?

    Published by Gbaf News

    Posted on May 26, 2017

    Featured image for article about Banking

    While the vision for data lakes has always been focused completely on making data more quickly available, few companies have managed to meet the challenge of satisfying the needs and of business end users as a central focus.

    The reality for many banks is that the data contained within most data lakes is not accessible to the average business user, who are relying on central engineering teams to construct queries to extract the data sets.

    There are two stark realities when it comes to serving business users (or not, as is often the case):

    Scenario #1:
    This bank had built its data lake, including many ingested data sources over a number of years.
    However, instead of being focused on building analytics and improving the quality of the data lake, the central engineering team spent their entire time dealing with requests to ingest new data sources, doing little data improvement and only focusing on building the first few layers of the transformation.

    The bank’s analysts and data scientists had become frustrated because the central engineering team had become a bottleneck. To get any new data source in, they had to go through the central data lake engineering team to go through an approvals process. This involved so much manual work in ingesting new data sources that it slowed down the entire analytics journey.

    Scenario #2:
    In dealing with its custom engineered data lakes, the bank’s senior IT executives dedicated extensive budget to building a huge team with excellent capacity. The team then spent months bringing in data sources without really consulting the businesses with what the use cases should be.

    Ultimately, IT didn’t get the buy in because the business didn’t understand how to get value from the data. Because it took so long to ingest a new data sources, productionise the new uses cases and manage the data quality and governance, the business got bored and went away to find another tool.

    The end result was a userless data lake with no subscribers and no stakeholders involved.

    The Verdict
    If data lakes are standing in the way of business users easily accessing data or use cases articulating a clear path to decision making that will lead to ROI, banks know they’ve got a problem. Here, we discuss three changes businesses are making to data lakes to ensure that relevant use cases see the light of day and that business users can put them to the test.

    Rising to the Self-Service Challenge

    From the outset, the main purpose of data lakes has been to give business users immediate access to all data, freeing them from relying on data warehousing teams to model that data, or simply to give them access.

    The point is that nothing was meant to stand in the way between business users and data, but the reality is much more complicated than this. Business users often struggle with self-service using data lakes and ultimately end up relying on engineers to construct complex queries to extract that data, which slows the release cycles. This is simply because open source tools do not feature any sort of self-service capability and so this has to be built.

    In order for use cases to be timely, relevant and useful, business users need to be able to get to and query their data. Data discovery needs to be made simple, allowing user to build queries to access the data to build data products that support analysis.

    One of the important differentiators in the next generation of data lake platforms is that they feature self-service capabilities for non-technical users. Many feature Google-like search functionality against both data and metadata, allowing users to quickly scan the schema catalog for relevant resources.

    Fixing Data Quality Control Issues

    Simply put, business users must be able to rely on the quality of data in a data lake, which is something many companies cannot guarantee.

    Data in data lakes that have been custom engineered tends to devolve over time. Without the proper approvals process over the data quality tools in place, people often continue to engineer in new data sources and integrate these into the data lake as one-offs. This ends up being a quickly forgotten process that does not focus on the quality of data.

    So what can companies do to ensure data quality is monitored and maintained? Automation must be put in place to ensure data is refreshed regularly, and to monitor the quality of data. Without this automation, over time, data quality begins to degrade and becomes useless in analysis.

    Get Governance in Place

    Companies often don’t have the experience, capability or skills or fully enable the governance or security they need to safely and productively maintain a data lake. While the flexibility of data lakes is one of their top selling points, without data lake governance, a data lake can quickly become a data swamp.

    Metadata management, as well as data cataloging and indexing are essential if user are to be able to query and use data in data lakes.

    In order to be able to be able to build the features that excite the business and solve real problems, banks need to put the next generation of data lake platforms to the test. Solving the path to self service, data quality and governance are great steps in the right direction to making data lakes user-centric and straightforward, and to solving banks biggest analytics dilemmas.

    Maurizio Colleluori is Principal Data Engineer for Kylo™, Think Big Analytics

    While the vision for data lakes has always been focused completely on making data more quickly available, few companies have managed to meet the challenge of satisfying the needs and of business end users as a central focus.

    The reality for many banks is that the data contained within most data lakes is not accessible to the average business user, who are relying on central engineering teams to construct queries to extract the data sets.

    There are two stark realities when it comes to serving business users (or not, as is often the case):

    Scenario #1:
    This bank had built its data lake, including many ingested data sources over a number of years.
    However, instead of being focused on building analytics and improving the quality of the data lake, the central engineering team spent their entire time dealing with requests to ingest new data sources, doing little data improvement and only focusing on building the first few layers of the transformation.

    The bank’s analysts and data scientists had become frustrated because the central engineering team had become a bottleneck. To get any new data source in, they had to go through the central data lake engineering team to go through an approvals process. This involved so much manual work in ingesting new data sources that it slowed down the entire analytics journey.

    Scenario #2:
    In dealing with its custom engineered data lakes, the bank’s senior IT executives dedicated extensive budget to building a huge team with excellent capacity. The team then spent months bringing in data sources without really consulting the businesses with what the use cases should be.

    Ultimately, IT didn’t get the buy in because the business didn’t understand how to get value from the data. Because it took so long to ingest a new data sources, productionise the new uses cases and manage the data quality and governance, the business got bored and went away to find another tool.

    The end result was a userless data lake with no subscribers and no stakeholders involved.

    The Verdict
    If data lakes are standing in the way of business users easily accessing data or use cases articulating a clear path to decision making that will lead to ROI, banks know they’ve got a problem. Here, we discuss three changes businesses are making to data lakes to ensure that relevant use cases see the light of day and that business users can put them to the test.

    Rising to the Self-Service Challenge

    From the outset, the main purpose of data lakes has been to give business users immediate access to all data, freeing them from relying on data warehousing teams to model that data, or simply to give them access.

    The point is that nothing was meant to stand in the way between business users and data, but the reality is much more complicated than this. Business users often struggle with self-service using data lakes and ultimately end up relying on engineers to construct complex queries to extract that data, which slows the release cycles. This is simply because open source tools do not feature any sort of self-service capability and so this has to be built.

    In order for use cases to be timely, relevant and useful, business users need to be able to get to and query their data. Data discovery needs to be made simple, allowing user to build queries to access the data to build data products that support analysis.

    One of the important differentiators in the next generation of data lake platforms is that they feature self-service capabilities for non-technical users. Many feature Google-like search functionality against both data and metadata, allowing users to quickly scan the schema catalog for relevant resources.

    Fixing Data Quality Control Issues

    Simply put, business users must be able to rely on the quality of data in a data lake, which is something many companies cannot guarantee.

    Data in data lakes that have been custom engineered tends to devolve over time. Without the proper approvals process over the data quality tools in place, people often continue to engineer in new data sources and integrate these into the data lake as one-offs. This ends up being a quickly forgotten process that does not focus on the quality of data.

    So what can companies do to ensure data quality is monitored and maintained? Automation must be put in place to ensure data is refreshed regularly, and to monitor the quality of data. Without this automation, over time, data quality begins to degrade and becomes useless in analysis.

    Get Governance in Place

    Companies often don’t have the experience, capability or skills or fully enable the governance or security they need to safely and productively maintain a data lake. While the flexibility of data lakes is one of their top selling points, without data lake governance, a data lake can quickly become a data swamp.

    Metadata management, as well as data cataloging and indexing are essential if user are to be able to query and use data in data lakes.

    In order to be able to be able to build the features that excite the business and solve real problems, banks need to put the next generation of data lake platforms to the test. Solving the path to self service, data quality and governance are great steps in the right direction to making data lakes user-centric and straightforward, and to solving banks biggest analytics dilemmas.

    Maurizio Colleluori is Principal Data Engineer for Kylo™, Think Big Analytics

    Related Posts
    DeFi and banking are converging. Here’s what banks can do.
    DeFi and banking are converging. Here’s what banks can do.
    Are Neo Banks Offering Better Metal Debit Cards Than Traditional Banks?
    Are Neo Banks Offering Better Metal Debit Cards Than Traditional Banks?
    Banking at the Intersection: From Nashville to Cannes, A Strategic Call to Action
    Banking at the Intersection: From Nashville to Cannes, A Strategic Call to Action
    Driving Efficiency and Profit Through Customer-Centric Banking
    Driving Efficiency and Profit Through Customer-Centric Banking
    How Ecosystem Partnerships Are Redefining Deposit Products
    How Ecosystem Partnerships Are Redefining Deposit Products
    CIBC Private Banking wins four 2025 Global Banking & Finance Awards
    CIBC Private Banking wins four 2025 Global Banking & Finance Awards
    How Banks Can Put AI to Work Now and Prove ROI in 90 Days
    How Banks Can Put AI to Work Now and Prove ROI in 90 Days
    Top 5 AI quality assurance framework providers for Banks and Financial Services firms.
    Top 5 AI quality assurance framework providers for Banks and Financial Services firms.
    The Unbanked Paradox: How Banking Access Creates Economic Resilience
    The Unbanked Paradox: How Banking Access Creates Economic Resilience
    Hyper-Personalised Banking - Shaping the Future of Finance
    Hyper-Personalised Banking - Shaping the Future of Finance
    The End of Voice Trust: How AI Deepfakes Are Forcing Banks to Rethink Authentication
    The End of Voice Trust: How AI Deepfakes Are Forcing Banks to Rethink Authentication
    Predicting and Preventing Customer Churn in Retail Banking
    Predicting and Preventing Customer Churn in Retail Banking

    Why waste money on news and opinions when you can access them for free?

    Take advantage of our newsletter subscription and stay informed on the go!

    Subscribe

    Previous Banking PostAURIGA URGES BANKS TO MAKE THE MOST OF TECHNOLOGY TO PREPARE FOR BANK HOLIDAY CASH DASH
    Next Banking PostONLINE BANKING REACHING CRITICAL MASS, CX KEY TO UNLOCK BRANCH POTENTIAL

    More from Banking

    Explore more articles in the Banking category

    Growth and Impact: Banreservas Leads Dominican Republic Economic Expansion

    Growth and Impact: Banreservas Leads Dominican Republic Economic Expansion

    Turning Insight into Impact: Making AI and Analytics Work in Retail Banking

    Turning Insight into Impact: Making AI and Analytics Work in Retail Banking

    KeyBank Embraces Next-Generation AI Platform to Transform Fraud and Financial Crime Prevention

    KeyBank Embraces Next-Generation AI Platform to Transform Fraud and Financial Crime Prevention

    Understanding Association Banking: Financial Solutions for Community Success

    Understanding Association Banking: Financial Solutions for Community Success

    Applying Symbiosis for advantage in APAC banking

    Applying Symbiosis for advantage in APAC banking

    AmBank Islamic Berhad Earns Triple Recognition for Excellence in Islamic Banking

    AmBank Islamic Berhad Earns Triple Recognition for Excellence in Islamic Banking

    FinTok Strategy: How Banks Are Reaching Gen Z Through Social Media

    FinTok Strategy: How Banks Are Reaching Gen Z Through Social Media

    Rethinking Retail Banking Sustainability: Why the ATM is an Asset in the Sustainable Transition

    Rethinking Retail Banking Sustainability: Why the ATM is an Asset in the Sustainable Transition

    How private banks can survive the neo-broker revolution

    How private banks can survive the neo-broker revolution

    Next-Gen Bank Branches: The Evolution from Transaction Hubs to Experience Centers

    Next-Gen Bank Branches: The Evolution from Transaction Hubs to Experience Centers

    The Banking Talent Crunch: How Financial Institutions Are Competing for Digital-Native Skills

    The Banking Talent Crunch: How Financial Institutions Are Competing for Digital-Native Skills

    Beyond Interest: How Banks Are Reimagining Revenue in the Digital Age

    Beyond Interest: How Banks Are Reimagining Revenue in the Digital Age

    View All Banking Posts