Climbing Further Up the Stack
As Gen AI programming tools continue to develop, I’m wondering what things will look like when we remove the human from the loop.
Bytes that get stuck in your teeth.
Some practical advice on supporting AI agents from Diwank Tomer.
Two things that stood out to me:
As Gen AI programming tools continue to develop, I’m wondering what things will look like when we remove the human from the loop.
Generating a podcast from an academic paper via NotebookLM is a killer feature.
Casey Handmer:
A model I’ve long been interested in is the Corporation as a stand in for AGI. We need some non-human autonomous legal and economic entity. A corporation is just that. The Fortune 500 are already non-human super-intelligence. They operate 24/7/365 according to inscrutable internal logic, routinely execute feats of production unthinkable for any human or other biological organism, often outlive humans, can exist in multiple places at once, etc etc.
Simon Willison:
Using LLMs to write code is difficult and unintuitive. It takes significant effort to figure out the sharp and soft edges of using them in this way, and there’s precious little guidance to help people figure out how best to apply them.
Cal Paterson:
A general pattern seems to be that Artificial Intelligence is used when first doing some new thing. Then, once the value of doing that thing is established, society will find a way to provide the necessary data in a machine readable format, obviating (and improving on) the AI models.
iA Writer dims the text you paste from AI tools. As you edit ChatGPT’s input and make it your own, iA Writer keeps track of what is yours and what isn’t.
From a leaked internal Google document:
While our models still hold a slight edge in terms of quality, the gap is closing astonishingly quickly. Open-source models are faster, more customizable, more private, and pound-for-pound more capable. They are doing things with $100 and 13B params that we struggle with at $10M and 540B. And they are doing so in weeks, not months.
The latest version of Photoshop Beta now includes a feature called generative fill.
I can imagine it’s easy to compromise the authenticity in your photography if you overuse these kinds of tools. With that said, and I know it’s potentially a slippery slide, there are situations where they can be incredibly useful.
The ChatGPT model is huge, but it’s not huge enough to retain every exact fact it’s encountered in its training set.
Eugene Yan:
Rule #1: Don’t be afraid to launch a product without machine learning.
Machine learning is cool, but it requires data. Theoretically, you can take data from a different problem and then tweak the model for a new product, but this will likely underperform basic heuristics. If you think that machine learning will give you a 100% boost, then a heuristic will get you 50% of the way there.
The Case for Learned Index Structures:
Indexes are models: a B-Tree-Index can be seen as a model to map a key to the position of a record within a sorted array, a Hash-Index as a model to map a key to a position of a record within an unsorted array, and a BitMap-Index as a model to indicate if a data record exists or not. In this exploratory research paper, we start from this premise and posit that all existing index structures can be replaced with other types of models, including deep-learning models, which we term learned indexes.
I used to find it odd that these hypothetical AIs were supposed to be smart enough to solve problems that no human could, yet they were incapable of doing something most every adult has done: taking a step back and asking whether their current course of action is really a good idea. Then I realized that we are already surrounded by machines that demonstrate a complete lack of insight, we just call them corporations.