Never miss a new edition of The Variable, our weekly newsletter featuring a top-notch selection of editors’ picks, deep dives, community news, and more.
Many practitioners like to jump headfirst into the nitty-gritty details of implementing AI-powered tools. We get it: tinkering your way into a solution can sometimes save you time, and it’s often a fun way to go about learning.
As the articles we’re highlighting this week show, however, it’s crucial to gain a high-level understanding of how the different pieces in your workflow come together. Sooner or later, when something — say, your data pipeline, or your team’s most-prized metric — goes awry, having this mental model in place will keep you focused and effective as a data or AI leader.
Let’s explore what systemic thinking looks like in practice.
How to Build an Over-Engineered Retrieval System
Ida Silfverskiöld‘s new deep dive, which pieces together a detailed retrieval pipeline as part of a broader RAG solution, assumes that for most AI engineering challenges, “there’s no real blueprint to follow.” Instead, we have to rely on extensive trial and error, optimization, and iteration.
Data Culture Is the Symptom, Not the Solution
Careful planning, prioritizing, and strategizing doesn’t only benefit specific tools or teams. As Jens Linden explains, it’s essential for organizations to thrive and for investments in data to pay off.
Building a Monitoring System That Actually Works
Follow along Mariya Mansurova’s guide to learn about “different monitoring approaches, how to build your first statistical monitoring system, and what challenges you’ll likely encounter when deploying it in production.”
This Week’s Most-Read Stories
Catch up with three of our most popular recent articles, covering code efficiency, LLMs in the service of data analysis, and GraphRAG design.
Run Python Up to 150× Faster with C, by Thomas Reid
LLM-Powered Time-Series Analysis, by Sara Nobrega
Do You Really Need GraphRAG? A Practitioner’s Guide Beyond the Hype, by Partha Sarkar
Other Recommended Reads
From tips on boosting your chances in Kaggle competitions to actionable advice on how to ace your next ML system-design interview, here are a few more articles you shouldn’t miss.
- Understanding Convolutional Neural Networks (CNNs) Through Excel, by Angela Shi
- Javascript Fatigue: HTMX Is All You Need to Build ChatGPT (Part 1, Part 2), by Benjamin Etienne
- How to Evaluate Retrieval Quality in RAG Pipelines (Part 3): DCG@k and NDCG@k, by Maria Mouschoutzi
- Organizing Code, Experiments, and Research for Kaggle Competitions, by Ibrahim Habib
- How to Crack Machine Learning System-Design Interviews, by Aliaksei Mikhailiuk
Meet Our New Authors
We hope you take the time to explore the excellent work from the latest cohort of TDS contributors:
- Mohannad Elhamod challenges the conventional wisdom that more data necessarily leads to better performance, and looks into the interplay of sample size, attribute set, and model complexity.
- Udayan Kanade shared an eye-opening exploration of the ties between contemporary LLMs and old-school randomized algorithms.
- Andrey Chubin leans on his AI leadership experience to unpack the common mistakes companies make when they attempt to integrate ML into their workflows.
We love publishing articles from new authors, so if you’ve recently written an interesting project walkthrough, tutorial, or theoretical reflection on any of our core topics, why not share it with us?
We’d Love Your Feedback, Authors!
Are you an existing TDS author? We invite you to fill out a 5-minute survey so we can improve the publishing process for all contributors.


