Skip to main content
HomePodcastsPodcast

Unlocking Scalable ROI for Data Teams

Shane Murray, Field CTO at Monte Carlo, joins the show to discuss how data leaders can scale the ROI of their teams. 
Feb 2023

Photo of Shane Murray
Guest
Shane Murray

Shane Murray is the Field CTO at Monte Carlo, a data reliability company that created the industry's first end-to-end Data Observability platform. Shane’s career has taken him through a successful 9-year tenure at The New York Times, where he grew the data analytics team from 12 to 150 people and managed all core data products. Shane is an expert when it comes to data observability, enabling effective ROI for data initiatives, scaling high-impact data teams, and more.


Photo of Adel Nehme
Host
Adel Nehme

Adel is a Data Science educator, speaker, and Evangelist at DataCamp where he has released various courses and live training on data analysis, machine learning, and data engineering. He is passionate about spreading data skills and data literacy throughout organizations and the intersection of technology and society. He has an MSc in Data Science and Business Analytics. In his free time, you can find him hanging out with his cat Louis.

Key quotes

I think you can approach the problems in enabling ROI for data leaders through the lens of how we think about observability. You have detection solutions, you have resolution, and you have prevention. On the detection side you have automated machine learning driven monitors. You have ways to target your alerting to different teams to make sure you're managing that signal-to-noise ratio in terms of alerts. Then on resolution, you have tools where you can actually look upstream as an analyst, see the initial cause of the data incident that you're investigating, be able to resolve it, and talk to the right partner upstream. And then also for those data producers to be able to look downstream and see the full scope of an incident on their side, I think that's just a phenomenal innovation in this space.

We typically think of one of the issues of data quality being downtime: the erroneous, missing, incomplete, or delayed data that often plague data initiatives. The consequence of downtime can range from this almost trivial outcome where engineers or analysts respond, and the result is the hours lost to address the issue, to actually more existential, where you're losing trust, revenue, or even customers. And then, at the far end of the scale, you could actually be putting in danger the reputation of the business.

Key takeaways

1

Before decentralizing a data team, it’s important that the data team is sufficiently mature to be able to handle decentralization efficiently and effectively.

2

Data teams should be focused on building data products that actually drive revenue in line with the organization’s goals.

3

It’s important to get the basics in place that free up your data team to do more expansive data roadmap work, such as self-service access, so stakeholders can get answers to basic questions without taking up team bandwidth.

Related

Top 10 Data Science Tools To Use in 2024

The essential data science tools for beginners and data practitioners to efficiently ingest, process, analyze, visualize, and model the data.

Abid Ali Awan

9 min

Google Cloud for Data Scientists: Harnessing Cloud Resources for Data Analysis

How can using Google Cloud make data analysis easier? We explore examples of companies that have already experienced all the benefits.
Oleh Maksymovych's photo

Oleh Maksymovych

9 min

A Guide to Docker Certification: Exploring The Docker Certified Associate (DCA) Exam

Unlock your potential in Docker and data science with our comprehensive guide. Explore Docker certifications, learning paths, and practical tips.
Matt Crabtree's photo

Matt Crabtree

8 min

Bash & zsh Shell Terminal Basics Cheat Sheet

Improve your Bash & zsh Shell skills with the handy shortcuts featured in this convenient cheat sheet!
Richie Cotton's photo

Richie Cotton

6 min

Functional Programming vs Object-Oriented Programming in Data Analysis

Explore two of the most commonly used programming paradigms in data science: object-oriented programming and functional programming.
Amberle McKee's photo

Amberle McKee

15 min

A Comprehensive Introduction to Anomaly Detection

A tutorial on mastering the fundamentals of anomaly detection - the concepts, terminology, and code.
Bex Tuychiev's photo

Bex Tuychiev

14 min

See MoreSee More