AI Risk and Threat Taxonomies

Posted on Di 05 August 2025 in security

It seems like every week my LinkedIn feed is filled with new just released AI risk taxonomies, threat models or AI governance handbooks. Usually these taxonomies come from governance consultants or standards authorities and are a great reference for understanding the wide variety of risks AI systems1 bring with …


Continue reading

Algorithmic-based Guardrails: External guardrail models and alignment methods

Posted on Mo 28 Juli 2025 in ml-memorization

You've probably at some point heard the term "guardrails" when talking about security or safety in AI systems like LLMs or multi-modal models (i.e. models that include and produce multiple modalities, like speech and image, videos, image and text).

Are you a visual learner? There's a YouTube video for …


Continue reading

Blocking AI/ML Memorization with Software Guardrails

Posted on Fr 11 Juli 2025 in ml-memorization

One common way to control memorization in today's deep learning systems is to fix the problem by building software around it. This software can also be used to deal with other undesired behavior, like producing hate speech or mentioning criminal activities.

Are you a visual learner? There's a YouTube video …


Continue reading

Defining Privacy Attacks in AI and ML

Posted on Do 12 Juni 2025 in ml-memorization

In this article series, you've been able to investigate memorization in AI/deep learning systems -- often via interesting attack vectors. In security modeling, it's useful to explicitly define the threats you are defending against, so you can both discuss and address them and compare potential interventions.

Prefer to learn by …


Continue reading

Priveedly: your private and personal content reader and recommender

Posted on Do 23 Januar 2025 in personal-ai

I'm excited to open-source a project that I've been using for the past 2 and a half years: a private/personal reader and recommender.

It works with:

and comes with an example Jupyter Notebook for training your own text-based recommendation model once you have …


Continue reading

Adversarial Examples Demonstrate Memorization Properties

Posted on Mi 15 Januar 2025 in ml-memorization

In this article, the last in the problem exploration section of the series, you'll explore adversarial machine learning - or how to trick a deep learning system.

Adversarial examples demonstrate a different way to look at deep learning memorization and generalization. They can show us how important the learned decision space …


Continue reading

Differential Privacy as a Counterexample to AI/ML Memorization

Posted on Do 02 Januar 2025 in ml-memorization

At this point in reading the article series on AI/ML memorization you might be wondering, how did the field get so far without addressing the memorization problem? How did seminal papers like Zhang et al's Understanding Deep Learning Requires Rethinking Generalization not fundamentally change machine learning research? And maybe …


Continue reading

How Memorization Happens: Overparametrized Models

Posted on Mi 18 Dezember 2024 in ml-memorization

You've heard claims that we will "run out of data" to train AI systems. Why is that? In this article in the series on machine learning memorization you'll explore model size as a factor in memorization and the trend for bigger models as a general problem in machine learning.

Prefer …


Continue reading

How memorization happens: Novelty

Posted on Mo 09 Dezember 2024 in ml-memorization

So far in this series on memorization in deep learning, you've learned how massively repeated text and images incentivize training data memorization, but that's not the only training data that machine learning models memorize. Let's take a look at another proven memorization: novel examples.

Prefer to learn by video? This …


Continue reading

How memorization happens: Repetition

Posted on Di 03 Dezember 2024 in ml-memorization

In this article in the deep learning memorization series, you'll learn how one part of memorization happens -- highly repeated data from the "head" of the long-tailed distribution.

Prefer to learn by video? This post is summarized on Probably Private's YouTube.

Recall from the data collection article that some examples are …


Continue reading