Discussion about this post

User's avatar
Somil Aggarwal's avatar

Strongly believe in the idea that value will shift to our agency over data exposure. Another hypothetical is imaging the decreasing friction (read, time it takes to reach ideal potential customer) it takes for a business to reach it's desired audience and thus where the value shifts. Loved reading this!

Expand full comment
Herbie Bradley's avatar

Interesting piece, but two flaws I see:

1. Literal privacy preserving tech is likely unnecessary, relative to the existing guarantees companies give for not training on your data. For example, I don't think opting out of ChatGPT training on your conversations does anything except give you the company's promise that they will exclude that data from training, and most people are probably satisfied with this.

For enterprise use there are guarantees of containerization/sandboxing they give to companies data. There appears to be a model where in some enterprise/corporate use cases, OpenAI try to incentivize customers to give data by giving free usage credits when they do so. Potentially this is how more personalized data might be collected, and potentially it's sufficient to make models good enough to just do this via enterprise deals.

2. We don't yet know the actual amount of data that will be sufficient to train models. Clearly, a diversity of data is good, which is why labs want to gather data from many people, especially for personalization. But it's plausible that the amount needed is in the low millions of data points, which makes pure contracting/enterprise deals possible.

On your final paragraphs, I think most people are not long-term thinking enough to value their data so highly, even if it is used to automate their job. This is basically a "humans are irrational" position, though one could argue that it's rational based on comparative advantage to accept the automation and be on the frontier of AI automation, then hope that you are still employable by being able to move to a higher level of abstraction (staying at the moving frontier up the pyramid). The history of data privacy shows most people care much less about it than researchers concerned about it think, and that in most cases this turned out to be fine.

You also don't mention the value of having one's work be trained on, to make future AIs more sensitive to you. This is part of the reason that I write.

Expand full comment
3 more comments...

No posts