Artwork

Indhold leveret af The Data Flowcast. Alt podcastindhold inklusive episoder, grafik og podcastbeskrivelser uploades og leveres direkte af The Data Flowcast eller deres podcastplatformspartner. Hvis du mener, at nogen bruger dit ophavsretligt beskyttede værk uden din tilladelse, kan du følge processen beskrevet her https://da.player.fm/legal.
Player FM - Podcast-app
Gå offline med appen Player FM !

Scaling Airflow to 11,000 DAGs Across Three Regions at Intercom with András Gombosi and Paul Vickers

34:24
 
Del
 

Manage episode 522613252 series 2948506
Indhold leveret af The Data Flowcast. Alt podcastindhold inklusive episoder, grafik og podcastbeskrivelser uploades og leveres direkte af The Data Flowcast eller deres podcastplatformspartner. Hvis du mener, at nogen bruger dit ophavsretligt beskyttede værk uden din tilladelse, kan du følge processen beskrevet her https://da.player.fm/legal.

The evolution of Intercom’s data infrastructure reveals how a well-built orchestration system can scale to serve global needs. With thousands of DAGs powering analytics, AI and customer operations, the team’s approach combines technical depth with organizational insight.

In this episode, András Gombosi, Senior Engineering Manager of Data Infra and Analytics Engineering, and Paul Vickers, Principal Engineer, both at Intercom, share how they built one of the largest Airflow deployments in production and enabled self-serve data platforms across teams.

Key Takeaways:

00:00 Introduction.

04:24 Community input encourages confident adoption of a common platform.

08:50 Self-serve workflows require consistent guardrails and review.

09:25 Internal infrastructure support accelerates scalable deployments.

13:26 Batch LLM processing benefits from a configuration-driven design.

15:20 Standardized development environments enable effective AI-assisted work.

19:58 Applied AI enhances internal analysis and operational enablement.

27:27 Strong test coverage and staged upgrades protect stability.

30:36 Proactive observability and on-call ownership improve outcomes.

Resources Mentioned:

András Gombosi

https://www.linkedin.com/in/andrasgombosi/

Paul Vickers

https://www.linkedin.com/in/paul-vickers-a22b76a3/

Intercom | LinkedIn

https://www.linkedin.com/company/intercom/

Intercom | Website

https://www.intercom.com

Apache Airflow

https://airflow.apache.org/

dbtLabs

https://www.getdbt.com/

Snowflake Cortex AI

https://www.snowflake.com/en/product/features/cortex/

Datadog

https://www.datadoghq.com/

Thanks for listening to “The Data Flowcast: Mastering Apache Airflow® for Data Engineering and AI.” If you enjoyed this episode, please leave a 5-star review to help get the word out about the show. And be sure to subscribe so you never miss any of the insightful conversations.

#AI #Automation #Airflow

  continue reading

82 episoder

Artwork
iconDel
 
Manage episode 522613252 series 2948506
Indhold leveret af The Data Flowcast. Alt podcastindhold inklusive episoder, grafik og podcastbeskrivelser uploades og leveres direkte af The Data Flowcast eller deres podcastplatformspartner. Hvis du mener, at nogen bruger dit ophavsretligt beskyttede værk uden din tilladelse, kan du følge processen beskrevet her https://da.player.fm/legal.

The evolution of Intercom’s data infrastructure reveals how a well-built orchestration system can scale to serve global needs. With thousands of DAGs powering analytics, AI and customer operations, the team’s approach combines technical depth with organizational insight.

In this episode, András Gombosi, Senior Engineering Manager of Data Infra and Analytics Engineering, and Paul Vickers, Principal Engineer, both at Intercom, share how they built one of the largest Airflow deployments in production and enabled self-serve data platforms across teams.

Key Takeaways:

00:00 Introduction.

04:24 Community input encourages confident adoption of a common platform.

08:50 Self-serve workflows require consistent guardrails and review.

09:25 Internal infrastructure support accelerates scalable deployments.

13:26 Batch LLM processing benefits from a configuration-driven design.

15:20 Standardized development environments enable effective AI-assisted work.

19:58 Applied AI enhances internal analysis and operational enablement.

27:27 Strong test coverage and staged upgrades protect stability.

30:36 Proactive observability and on-call ownership improve outcomes.

Resources Mentioned:

András Gombosi

https://www.linkedin.com/in/andrasgombosi/

Paul Vickers

https://www.linkedin.com/in/paul-vickers-a22b76a3/

Intercom | LinkedIn

https://www.linkedin.com/company/intercom/

Intercom | Website

https://www.intercom.com

Apache Airflow

https://airflow.apache.org/

dbtLabs

https://www.getdbt.com/

Snowflake Cortex AI

https://www.snowflake.com/en/product/features/cortex/

Datadog

https://www.datadoghq.com/

Thanks for listening to “The Data Flowcast: Mastering Apache Airflow® for Data Engineering and AI.” If you enjoyed this episode, please leave a 5-star review to help get the word out about the show. And be sure to subscribe so you never miss any of the insightful conversations.

#AI #Automation #Airflow

  continue reading

82 episoder

所有剧集

×
 
Loading …

Velkommen til Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Hurtig referencevejledning

Lyt til dette show, mens du udforsker
Afspil