Data Science Briefing #314


​(view in browser)​

Apr 15th

Next webinar:
​
Apr 22, 2026 -
Count down to 2026-04-22T17:00:00.000Z​

Dear Reader,

Welcome to the US Tax Day edition of our newsletter!

Announcements

 

Graph Neural Networks are transforming how we model everything from social networks to drug discovery. Whether you're looking for a refresher or a starting point, the introductory guide in our latest blog post over on Substack is a perfect place to begin:

πŸ‘‰ Graph Neural Networks 101

Check it out and Subscribe so you don't miss another post.

The current wave of AI progress is starting to look less like a story of raw model size and more like a story of systems: routines that turn repeated tasks into dependable workflows, tool use that lets models reach beyond the prompt, and a growing realization that real-world performance depends as much on orchestration as on intelligence itself.

This shift runs through this week’s reading, from the rise of practical automation and local coding assistants to sharper debates over how agents should call tools without collapsing into integration chaos. At the same time, the ​latest industry snapshot makes clear that investment, deployment, and capability gains are accelerating, even as compute partnerships and infrastructure deals reveal just how expensive this new era has become. The future of AI will be shaped not just by smarter models but by better interfaces, better engineering, and a much clearer understanding of where the real bottlenecks now lie.

On the academic front, this week's papers sketch a more complicated picture of AI progress than the usual triumphalist narrative: the field is advancing quickly, but not always on foundations sturdy enough to trust.

On one side are genuinely exciting signs of scientific reach, from hybrid deep learning approaches that map protein interactions at scale to emerging visions of β€œneural computers” and agent architectures that expand model capability through memory, tools, and externalized skills rather than sheer parameter count alone.

On the other is a sobering reminder that capability without rigor can mislead just as easily as it can illuminate: disease-prediction systems trained on shaky data, writing assistants that subtly nudge users’ social and political attitudes, and even safety interventions that may introduce new forms of harm all point to the same core lesson. As AI becomes more embedded in science, medicine, and everyday reasoning, the real challenge is no longer just making systems more powerful but ensuring they are grounded, transparent, and resilient in the messy environments where their outputs actually matter.

Our current book recommendation is "Designing Data-Intensive Applications" by M. Kleppmann and C. Riccomini. You can find all the previous book reviews on our website. In this week's video, we have an overview of AI in Videogames.

Data shows that the best way for a newsletter to grow is by word of mouth, so if you think one of your friends or colleagues would enjoy this newsletter, go ahead and forward this email to them. This will help us spread the word!

Semper discentes,

The D4S Team


"Designing Data-Intensive Applications" by M. Kleppmann and C. Riccomini is the kind of book that quietly raises the level of everyone who reads it. In this new edition, the authors do an outstanding job of explaining the core ideas behind modern data systems, like replication, consistency, storage, streaming, fault tolerance, and scalability, without reducing them to buzzwords or vendor-specific recipes. That makes the book especially valuable for data scientists and machine learning engineers as it bridges the gap between building models and understanding the data infrastructure those models depend on in production.

What makes the book so compelling is its focus on first principles. Rather than teaching a single stack or a fleeting set of tools, it gives readers a durable framework for thinking about trade-offs in real systems. That is incredibly useful for ML engineers working on pipelines, model serving, retrieval systems, or any workflow where reliability and performance matter as much as model quality. The downside is that it is more conceptual than hands-on, and readers looking for quick code examples or direct coverage of topics like feature stores, vector databases, or modern LLM infrastructure may wish it connected the dots more explicitly.

Still, that broader systems lens is exactly why the book stands out. It is thoughtful, clear, and deeply practical in the ways that matter over the long run. For anyone in data science or machine learning who wants to understand not just how to build models, but how to build the systems that let those models survive contact with reality, this is an easy book to recommend.


  1. ​Automate work with routines [code.claude.com]
  2. ​Stanford Artificial Intelligence Index Report 2026 [hai.stanford.edu]
  3. ​The Mythical Agent-Month [wesmckinney.com]
  4. ​How We Broke Top AI Agent Benchmarks: And What Comes Next [rdi.berkeley.edu]
  5. ​Anthropic Will Use CoreWeave’s AI Capacity to Power Claude [bloomberg.com]
  6. ​Tool calling, open source, and the MΓ—N problem​ [thetypicalset.com]
  7. ​I ran Gemma 4 as a local model in Codex CLI [medium.com/google-cloud]


AI and Videogames

video preview​

All the videos of the week are now available in our YouTube playlist.

Upcoming Events:

Opportunities to learn from us
​

On-Demand Videos:

Long-form tutorials
​

​Unsubscribe​

Data For Science, Inc

I'm a maker and blogger who loves to talk about technology. Subscribe and join over 3,000+ newsletter readers every week!

Read more from Data For Science, Inc

(view in browser) May 13th Next webinar: May 27, 2026 - Code Development with AI Assistants [Register] Dear Reader, Announcements Ever wonder how we can turn thousands of unstructured news articles into structured, actionable insights? In the latest post from Data4Sci, we dive into the fascinating process of transforming raw text from news articles into interconnected networks of information. If you're interested in Natural Language Processing (NLP), entity extraction, and how to connect the...

(view in browser) May 6th Next webinar: May 27, 2026 - Code Development with AI Assistants [Register] Dear Reader, Announcements Ever wonder how we can turn thousands of unstructured news articles into structured, actionable insights? In the latest post from Data4Sci, we dive into the fascinating process of transforming raw text from news articles into interconnected networks of information. If you're interested in Natural Language Processing (NLP), entity extraction, and how to connect the...

(view in browser) Apr 30th Next webinar: May 6, 2026 - Automate the Boring Developer Stuff with LLMs [Register] Dear Reader, Announcements ✈️ Mapping the skies: How do we visualize airline traffic between states? We often think of air travel in terms of airports, but viewing it as a network of state-to-state connections reveals fascinating patterns in how our country moves. Our latest substack uses data visualization to turn raw statistics into a clear story about infrastructure and mobility....