본문으로 바로가기

Share this webinar

Close your data and AI skills gap

We're the only platform uniquely engineered to advance data and AI skills across your entire organization. Let's explore a tailored program.

Book an Enterprise Demo
Upskilling a small team?Get started today
Artificial Intelligence

Scaling Data Quality in the Age of Generative AI

January 2025

Your Presenter(s)

Barr Moses 헤드샷

Barr Moses

CEO & Co-Founder of Monte Carlo

Barr Moses is CEO & Co-Founder of Monte Carlo, a data reliability company backed by Accel, GGV, Redpoint, and other top Silicon Valley investors. Previously, she was VP Customer Operations at Gainsight, a management consultant at Bain & Company, and served in the Israeli Air Force as a commander of an intelligence data analyst unit. Barr graduated from Stanford with a B.Sc. in Mathematical and Computational Science.

Prukalpa Sankar 헤드샷

Prukalpa Sankar

Co-founder of Atlan

Prukalpa Sankar is the Co-founder of Atlan. Atlan is a modern data collaboration workspace (like Github for engineering or Figma for design). By acting as a virtual hub for data assets ranging from tables and dashboards to models & code, Atlan enables teams to create a single source of truth for all their data assets, and collaborate across the modern data stack through deep integrations with tools like Slack, BI tools, data science tools and more. A pioneer in the space, Atlan was recognized by Gartner as a Cool Vendor in DataOps, as one of the top 3 companies globally. Prukalpa previously co-founded SocialCops, world leading data for good company (New York Times Global Visionary, World Economic Forum Tech Pioneer). SocialCops is behind landmark data projects including India’s National Data Platform and SDGs global monitoring in collaboration with the United Nations. She was awarded Economic Times Emerging Entrepreneur for the Year, Forbes 30u30, Fortune 40u40, Top 10 CNBC Young Business Women 2016. TED Speaker.

George Fraser 헤드샷

George Fraser

CEO at Fivetran

George founded Fivetran to help data engineers simplify the process of working with disparate data sources. He has grown Fivetran to be the de facto standard platform for data movement. In 2023, he was named a Datanami Person to Watch. George has a PhD in neurobiology.

Summary

In the era of generative AI, the need for maintaining top-notch data quality is highly emphasized. With AI and machine learning becoming essential for organizations, the challenge of assuring data reliability is a primary concern. Experts Bar Moses, Prakalpa Sankar, and George Frazier explore the complex nature of data quality, discussing its inherent challenges and the changing environment. As trust in data becomes critical, the conversation points to the shift from traditional data management to more advanced systems that tackle emerging challenges. Despite technological progress, data quality remains a constant issue, with organizations dealing with the pressures of delivering reliable generative AI products. The discussion also highlights the cultural aspects of data quality, the importance of teamwork, and the need for sturdy frameworks to enhance data trust. Ensuring data quality is not solely a technical challenge but also a cultural one, requiring alignment across teams and the adoption of new methodologies to handle evolving data issues.

Key Takeaways:

  • Data quality is vital in the era of generative AI, with organizations under pressure to deliver trustworthy products.
  • Trust in data is as important as the data itself; issues often stem from cultural misunderstandings and technical failures.
  • Generative AI requires high-quality, proprietary data for competitive differentiation.
  • Teamwork between data producers and consumers is essential to close trust gaps.
  • The environment of data quality is changing, with new challenges requiring innovative solutions.

Deep Dives

The Changing Environment of Data Quality

Data quality has always been a challenge in the industry, but the advent of generative AI has raised the stakes. As Bar Moses noted, "The goalpost on data quality is shifting every year." Organizations face increasing pressure from the C-suite and market to produce generative AI products, yet many data leaders feel their data isn't ready. This disconnect highlights the need for a new approach in how data is managed. While technological progress has improved data processing and storage, data management practices have not kept up. As a result, many organizations still rely on manual approaches to data quality. The focus is now on moving beyond detection to understanding and resolving these issues at their root, which often involves complex systems and multi-team collaboration.

Trust as a Key Component of Data Quality

Prakalpa Sankar stressed the importance of trust in data quality, arguing that trust breaks not when something goes wrong, but when stakeholders learn about issues from someone else. A sturdy trust framework is vital for maintaining data quality, especially in fast-paced, real-time ecosystems where things can quickly go wrong. Building this trust involves creating awareness of data issues before they impact users, thus preventing the erosion of stakeholder confidence. The solution is not to prevent errors entirely but to manage them efficiently and transparently, ensuring that data consumers can rely on the information they receive.

Cultural and Teamwork Challenges in Data Quality

The cultural aspect of data quality cannot be overlooked. As organizations become more data-driven, aligning the objectives of data producers and consumers becomes necessary. George Frazier highlighted this by discussing the teamwork needed between business and data teams to tackle data quality issues. A lack of shared understanding and context can lead to mistrust and inefficiencies. Establishing a culture that promotes communication and aligns on data quality standards is essential. This involves setting clear expectations and metrics, such as SLAs, to ensure that data is timely and reliable, ultimately driving better business outcomes.

Generative AI and Data Quality: New Considerations

Generative AI introduces new considerations to data quality. As organizations strive to leverage AI, the need for high-quality proprietary data becomes clear. Bar Moses pointed out that the competitive advantage lies in the proprietary data companies can provide to generative AI models. This requires careful attention to data quality and governance, ensuring that the data fed into AI systems is trustworthy and accurate. The conversation around data quality in AI is just beginning, and as organizations experiment with AI applications, they must prioritize data integrity to ensure meaningful and reliable outcomes.


관련된

white paper

Your Organization's Guide to Data Maturity

Learn how evaluate and scale data maturity throughout your organization

webinar

Laying the Foundations: Data Quality in the Age of AI

Join Susan Walsh and Scott Taylor as they walk us through how data leaders can make meaningful gains on their data quality initiatives, and the nuances of scaling a data quality initiative with AI in mind. 

webinar

How AI is Changing Data Quality

Gorkem Sevinc, CEO and Founder at Qualytics, and Piyush Mehta, CEO at Data Dynamics, will explore how AI is transforming data quality management.

webinar

Scaling Data & AI Literacy with a Persona-Driven Framework

In this session, three experts walk you through the steps of creating a successful data training program.

webinar

Scaling Data & AI Literacy with a Persona-Driven Framework

In this session, three experts walk you through the steps of creating a successful data training program.

webinar

Radar Data & AI Literacy Edition: Laying the Foundations: Data Quality in the Age of AI

Join Susan Walsh and Scott Taylor as they walk us through how data leaders can make meaningful gains on their data quality initiatives, and the nuances of scaling a data quality initiative with AI in mind.