In episode 120 of The Gradient Podcast, Daniel Bashir speaks to Sasha Luccioni.

Sasha is the AI and Climate Lead at HuggingFace, where she spearheads research, consulting, and capacity-building to elevate the sustainability of AI systems. A founding member of Climate Change AI (CCAI) and a board member of Women in Machine Learning (WiML), Sasha is passionate about catalyzing impactful change, organizing events and serving as a mentor to under-represented minorities within the AI community.

Have suggestions for future podcast guests (or other feedback)? Let us know here or reach Daniel at editor@thegradient.pub

Subscribe to The Gradient Podcast:  Apple Podcasts  | Spotify | Pocket Casts | RSS
Follow The Gradient on Twitter

Subscribe now

Outline:

(00:00) Intro

(00:43) Sasha’s background

(01:52) How Sasha became interested in sociotechnical work

(03:08) Larger models and theory of change for AI/climate work

(07:18) Quantifying emissions for ML systems

(09:40) Aggregate inference vs training costs

(10:22) Hardware and data center locations

(15:10) More efficient hardware vs. bigger models — Jevons paradox

(17:55) Uninformative experiments, takeaways for individual scientists, knowledge sharing, failure reports

(27:10) Power Hungry Processing: systematic comparisons of ongoing inference costs

(28:22) General vs. task-specific models

(31:20) Architectures and efficiency

(33:45) Sequence-to-sequence architectures vs. decoder-only

(36:35) Hardware efficiency/utilization

(37:52) Estimating the carbon footprint of Bloom and lifecycle assessment

(40:50) Stable Bias

(46:45) Understanding model biases and representations

(52:07) Future work

(53:45) Metaethical perspectives on benchmarking for AI ethics

(54:30) “Moral benchmarks”

(56:50) Reflecting on “ethicality” of systems

(59:00) Transparency and ethics

(1:00:05) Advice for picking research directions

(1:02:58) Outro

Links:

Sasha’s homepage and Twitter

Papers read/discussed

Climate Change / Carbon Emissions of AI Models

Quantifying the Carbon Emissions of Machine Learning

Power Hungry Processing: Watts Driving the Cost of AI Deployment?

Tackling Climate Change with Machine Learning

CodeCarbon

Responsible AI

Stable Bias: Analyzing Societal Representations in Diffusion Models

Metaethical Perspectives on ‘Benchmarking’ AI Ethics

Measuring Data

Mind your Language (Model): Fact-Checking LLMs and their Role in NLP Research and Practice

Read More in  The Gradient