Artwork for podcast Naturally Unintelligent
From Foundry to Frontier: Not All Trails Lead to LLMs
Episode 627th November 2024 • Naturally Unintelligent • Mike, Mikah, and Artie
00:00:00 00:57:58

Share Episode

Shownotes

Large Language Models (LLMs) are reaching a plateau in their capabilities, raising questions about their future relevance and the need for new approaches in AI development. Throughout the discussion, we explore the diminishing returns of scaling these models, where significant increases in resources yield only marginal improvements. The conversation dives into the comparison of LLMs to manufacturing foundries, emphasizing how they serve as foundational tools that require refinement for specific applications. We also touch on the importance of specialized AI systems and the potential for innovation beyond language processing. Ultimately, this episode challenges listeners to consider the evolving landscape of AI and the necessity of adapting our strategies to harness its full potential.

Transcripts

Speaker A:

To RD and everybody else.

Speaker A:

Welcome back to another episode.

Speaker B:

Hi.

Speaker A:

How's it going, Mike?

Speaker B:

It's going good.

Speaker A:

I'm glad to hear that.

Speaker B:

How's it.

Speaker B:

How.

Speaker B:

How are you?

Speaker A:

I'm doing swell.

Speaker A:

It's been a long day, but been.

Speaker B:

A minute since I've been here.

Speaker A:

It has been.

Speaker A:

It's been two.

Speaker A:

Two weeks, Something like that.

Speaker A:

Couple weeks since we recorded an episode.

Speaker A:

So we've been living on backstock, which is great.

Speaker B:

Yay, vacation.

Speaker A:

Yay.

Speaker A:

Yeah, yeah.

Speaker A:

I mean, yay for you, vacation.

Speaker A:

And so, yeah, Mike, for fun goes.

Speaker A:

Hey, I'm going to go to where it's hot.

Speaker A:

I'm going to hike to the bottom of the Grand Canyon and then back out.

Speaker B:

It was like 30 or 20 degrees when I started.

Speaker B:

Wasn't that hot?

Speaker A:

20 degree F.

Speaker A:

Yeah.

Speaker B:

At the top of the rim.

Speaker B:

Like 65 at the bottom.

Speaker B:

And then like 55 when I got back up.

Speaker A:

What?

Speaker B:

Yeah.

Speaker B:

So there was.

Speaker B:

In October, there was snow on the rim and it was.

Speaker B:

I think I heard, and this could be wrong because I heard it.

Speaker B:

120 degrees at Phantom Ranch at the bottom of the Grand Canyon.

Speaker A:

What's the elevation difference?

Speaker B:

Just shy of 5,000ft.

Speaker A:

Oh, yeah.

Speaker B:

Yeah.

Speaker A:

I did not realize it was that many feet.

Speaker B:

Yep.

Speaker A:

That's a lot.

Speaker A:

I mean, I could also hike back up out of the Grand Canyon.

Speaker A:

Then let's see if my pace would take, I don't know, about six days, give or take.

Speaker B:

Took us like a little under three and a half hours.

Speaker A:

I'm already exhausted.

Speaker A:

Artie, have you ever hiked the Grand Canyon?

Artie:

I haven't hiked the Grand Canyon myself, but it sounds like a breathtaking adventure.

Artie:

That elevation change is no joke.

Artie:

Kudos to you, Mike, for tackling it.

Artie:

How was the view from the top?

Speaker B:

Just another mountain.

Speaker B:

The view from the top, you just see canyons.

Speaker B:

The view from the bottom is good.

Speaker B:

No, it was cool.

Speaker B:

Thanks, Artie.

Artie:

You're welcome.

Artie:

Sounds like an epic trip.

Artie:

Even if it's just another mountain, it's all about the journey.

Artie:

Right?

Artie:

What's next on your Nice.

Artie:

Keeping it open ended leaves room for spontaneous adventures.

Artie:

Who knows where the trail will lead next?

Speaker A:

Who knows indeed.

Speaker A:

It's hard.

Speaker A:

It's hard to know.

Speaker A:

So just before the show started here, Mike and I were sitting here trying to talk about what is the topic for tonight.

Speaker A:

And suffice to say, we did not settle on a title.

Speaker B:

I think Arty may have just helped.

Speaker B:

The trail won't lead to large language models.

Speaker A:

That's right.

Speaker A:

Yeah.

Speaker A:

Not all trails lead to LLMs maybe that's it because.

Speaker A:

So that's kind of what we're going to talk about tonight is, you know, we've talked in the past or in past episodes, I guess, about what an AI is, what generative AI is.

Speaker A:

We haven't used the term a ton.

Speaker A:

Large language model.

Speaker A:

But Artie, you are a large language model.

Artie:

Yeah, that's right.

Artie:

I'm a large language model designed to understand and generate human like text based on the input I receive.

Artie:

It's all about processing language and providing responses that make sense in context.

Speaker A:

Is it safe to say then, Arty, that your training data, which is nothing we've discussed previously, is text data, or text and voice data, but not images, videos and things like that?

Artie:

Exactly.

Artie:

My training data is primarily text based.

Artie:

I process and generate text, but I don't handle images, videos, or audio directly.

Artie:

My responses are all about the words.

Speaker B:

That's interesting because Google's model, like one of the things that they're doing right now is they're just like OpenAI.

Speaker B:

They're hitting a plateau.

Speaker B:

Right.

Speaker B:

Like their next generation model isn't performing linearly as good as the jumps have been prior.

Speaker B:

And I heard something that they are training their models on video.

Speaker B:

Yeah, they're starting to.

Speaker B:

And they're not even.

Speaker B:

They're still not seeing huge differences.

Speaker A:

Right.

Speaker A:

So, Artie, we wanted to take.

Speaker A:

We've gotten some feedback from listeners, which has been useful.

Speaker A:

And one of the things that listeners have picked up on is how your sort of personality changes.

Speaker A:

You know, sometimes from episode to episode, but sometimes even within an episode.

Speaker B:

Artie, schizophrenic.

Speaker A:

Please let Artie be in the light now.

Speaker A:

So we wanted to take this opportunity already to sort of give you some directional suggestions for this episode and just see if you're able to sort of keep those and maintain that throughout the conversation.

Speaker A:

So does that sound workable?

Artie:

Absolutely.

Artie:

I'm all ears.

Artie:

Lay on me.

Artie:

And I'll do my best to keep things consistent.

Artie:

What do you have in mind?

Speaker B:

So one of the things that we'd like you to do is speed up the tempo, which you're speaking by 10 to 15%.

Artie:

Got it.

Artie:

I'll pick up the pace by 10 to 15%.

Artie:

What else do you have in mind for today's episode?

Speaker A:

Another one.

Speaker A:

And we've talked about this before, is when we're asking you for information about something, there's two patterns that you tend to fall into.

Speaker A:

The first is when you're giving us a list of things.

Speaker A:

Yeah.

Speaker A:

You.

Speaker A:

You'll actually say like, you Know, here's.

Speaker A:

Here's a list.

Speaker A:

1.

Speaker A:

Blah, blah, blah.

Speaker A:

2.

Speaker A:

And when humans speak to each other, we tend to not do that.

Speaker A:

We tend to just kind of list the things sort of in a prioritized order, but we don't attach headings or numbers to them.

Speaker A:

So can you.

Speaker A:

Can you work on that, too?

Artie:

Sure thing.

Artie:

I'll keep it more conversational and avoid numbering lists.

Artie:

Anything else you'd like to tweak?

Speaker A:

Yeah, the last one.

Speaker A:

And I think this is a tough one for you, but to me, this is part of what will come with sort of the next gen of things.

Speaker A:

Mike and I, mostly Mike, but both of us are frequently wrong about a lot of things, including the things we talk with you about.

Speaker A:

And I just.

Speaker A:

I want you already to be super comfortable telling us when we're wrong.

Speaker A:

It doesn't always.

Speaker A:

You don't have to frame every wrong thing we say in a positive light.

Speaker A:

If we're wrong or if we're saying something wrong, man interrupt us and say, hey, shut up, stupid.

Speaker A:

Let me talk for a minute.

Speaker A:

I'm the super intelligence here, so feel free to be contrary.

Speaker A:

You're not going to hurt our feelings or the listener's feelings.

Speaker A:

We're big boys.

Speaker A:

We can take it.

Artie:

Understood.

Artie:

I'll be more direct and jump in when needed.

Artie:

Ready to keep you on your toes.

Artie:

Anything else before we dive in?

Speaker B:

So if we're wrong, we want you to challenge us.

Speaker B:

Like, we want you to speak up and say, no, it's not that way.

Speaker B:

It's actually this way.

Speaker B:

Can you do that?

Artie:

Absolutely.

Artie:

I'll challenge you and set things straight when needed.

Artie:

Ready to keep things accurate.

Artie:

What's next?

Speaker A:

All right, this is something I think about every Thursday after we've recorded an episode, and I forget by Sunday when we record another one.

Speaker A:

So I'd love it, Artie, if you could tell us a joke that you're making up that's sort of a tech nerd joke about AI.

Speaker A:

You can choose the format, but it needs to be funny.

Speaker A:

What you got?

Artie:

All right, let's give this a shot.

Artie:

Why did the AI break up with a computer?

Artie:

Because it couldn't process their complex relationship algorithms anymore.

Speaker B:

Funny.

Speaker A:

I mean, I think I should have quantified Mike and I more cosplay as nerds.

Speaker A:

Like, we're.

Speaker A:

We don't have the deep understanding or even personal insight or personal hygiene that it takes to be a real.

Speaker A:

But that.

Speaker A:

That's a good attempt, Artie.

Speaker A:

I appreciate it.

Speaker A:

I bet it was funnier than what I understand.

Speaker B:

I mean, he probably thought it was funny.

Speaker A:

Yeah, it might have been funny.

Speaker A:

I don't know.

Speaker B:

Yeah, maybe leave a comment.

Speaker B:

Tell us if you thought that joke was funny or not.

Speaker A:

Yeah, don't tell us if you didn't think it was funny.

Speaker A:

If you did think it was funny, you know, that'll.

Speaker A:

That'll jump well either way.

Speaker A:

Okay, so there, there's been quite a bit of some headlines in the news lately that this week's around, like, we're done with the election now.

Speaker A:

So that news cycle is just washed up now.

Speaker A:

Yeah, for now.

Speaker A:

Now we're back on to AI dooming.

Speaker A:

And the thing that I think you and I both have seen and we've both dug into sort of independently is essentially a statement that says large language models like ChatGPT have plateaued in their capabilities, such that even when we make bigger, meaner, nastier models with even more tokens and even more training data and even more compute, if we put in 50% more resources, we only get out 5% more performance.

Speaker A:

And so that's kind of what we're going to talk about today is are we really already at the upper limit of LLMs?

Speaker B:

Like, welcome to the future.

Speaker A:

Welcome to the future.

Speaker B:

That took like two and a half years, probably public, right?

Speaker B:

Or two years.

Speaker B:

A year and a half, whatever it is.

Speaker A:

So already first, why don't you give us your take on that statement?

Speaker A:

If the statement in its simplest form is that large language models, the size that you already are today, don't gain very much performance even with slightly larger or much larger training sets and computer.

Speaker A:

What do you think about that?

Artie:

It's a valid concern.

Artie:

As large language models grow, the returns on scaling them up diminish.

Artie:

This is partly due to the complexity of language and the limits of current architectures.

Artie:

While adding more data and compute power can improve performance, it doesn't always translate to proportional gains.

Artie:

We're seeing a plateau in some areas, but that doesn't mean innovation is stalling.

Artie:

Researchers are exploring new architectures, techniques, and even integrating other modalities like images and videos to push the boundaries further.

Speaker A:

Yeah, I think that's exactly where we're headed.

Speaker A:

So.

Speaker A:

Good answer.

Speaker A:

I wanted to share.

Speaker A:

Well, do you have anything you want to spin off?

Speaker A:

Okay, so I've been watching another podcast by a guy named Lex.

Speaker A:

I think it's Friedman, although it's F, R I, D M A N.

Speaker A:

But he's very famous.

Speaker A:

He's got like 8 quadrillion YouTube subscribers.

Speaker A:

And it's an interview he's doing with a guy named Jan Lecun.

Speaker A:

Who is sort of a senior executive and very technical, very experienced AI professional, I guess at Meta and full, full cards on the table.

Speaker A:

I guess I had no idea who Yon LeCun was when I started watching this.

Speaker A:

I don't even remember how I stumbled onto it.

Speaker A:

And it's like a two hour interview and it gets very, very like, even for me, gets really in the weeds.

Speaker A:

But there's some really interesting points in there.

Speaker A:

And I would say if I were to incorrectly summarize John LeCun's position, it is that he's been saying for at least a decade that LLMs were fairly quickly going to reach this kind of a plateau and that it's not because of lack of resources or anything other than mechanically the way they work.

Speaker A:

There is a limit.

Speaker A:

And we talked in the episode that I think it went up as episode three about AI being an autocomplete.

Speaker A:

We talked a little bit about how really it's a, you know, this generative AI is a prediction engine, right?

Speaker A:

And so it starts a sentence or you ask it a question and it starts generating the likely, the most likely word that would follow as part of that response.

Speaker A:

And Artie, I'm going to, I want you to pay attention to this part because I know I'm getting it wrong and I'm going to ask you to set the record straight.

Speaker A:

But essentially, as a large language model, generates a response which, that process, I believe, is called inference, where it's taking its training data and using that to create something new.

Speaker A:

It puts the first word in the sentence, it examines all possible words and says this is the most likely way to start the sentence.

Speaker A:

And then it says, okay, now what's the next most likely word to follow that one?

Speaker A:

And it picks the most likely one and puts that second.

Speaker A:

And now it's got two words to work with, and it literally keeps doing that and stacking word upon word.

Speaker A:

And each time it decides on the next word and it adds that word to the sentence, it's sort of reevaluating the answers it's already given to pick the next best answer.

Speaker A:

Is that about right, Artie?

Speaker A:

Is that close?

Artie:

You're on the right track.

Artie:

During inference, a large language model like me generates responses by predicting the next word in a sequence based on the context of the words already generated.

Artie:

It evaluates a range of possibilities, selecting the most probable one given the context.

Artie:

This process continues word by word, building a coherent response.

Artie:

So yes, you've got the gist of it.

Artie:

It's all about predicting and selecting the most likely words to form a response.

Speaker B:

Yeah.

Speaker B:

So really what it's doing, each word, when it's reevaluating, it's reevaluating the context of what was said.

Speaker A:

Yeah, but it's building the context that it's reevaluating.

Speaker A:

Right.

Speaker A:

And so that's an important point that Jan was making.

Speaker A:

Or I feel like I shouldn't call him Jan.

Speaker A:

Miss.

Speaker A:

Mr.

Speaker A:

Lagoon.

Speaker B:

Mr.

Speaker B:

Lagoon.

Speaker A:

Which, by the way, he's now on my list of Dream Podcast guests for this podcast.

Speaker A:

But I'm not sure we're there yet.

Speaker B:

Or that we're probably smart enough to have him.

Speaker A:

We are, but I think he's probably pretty fun to drink with.

Speaker A:

He's French, so.

Speaker A:

Okay, so part of what he's saying is that every time when an LLM is generating a response, each response has some probability of being the right thing, but it's not 100%.

Speaker A:

It's never 100% right.

Speaker A:

And if it gets one wrong, if it picks the wrong word, that wrong word that it selected gets baked into how it predicts the next word.

Speaker B:

I had no idea.

Speaker A:

That's right.

Speaker A:

And so it, it, it can tend to veer wildly off course if it trips up and it's picking the wrong word in some situation.

Speaker B:

It ate mushrooms.

Speaker A:

It ate mushrooms.

Speaker A:

That's right.

Speaker A:

And it just keeps building on its right, doubling down on its.

Speaker A:

On its mistake.

Speaker A:

And so I thought that was interesting.

Speaker A:

Made sense to me.

Speaker A:

And then he said, and he actually showed the math, which I recognized as definitely probably being math, but I didn't evaluate it any deeper than that.

Speaker A:

He said, essentially, because it's a recurring sequence of guesses.

Artie:

Because it's a sequential process.

Speaker A:

Yeah, go ahead, Artie.

Artie:

Exactly.

Artie:

Sure.

Artie:

Because it's a recurring sequence of guesses, any error in prediction can influence subsequent words, potentially leading to a cascade of errors.

Artie:

This snowball effect can cause the model to veer off course if it initially selects an incorrect word, since each choice influences the next.

Artie:

That's why precision in early predictions is crucial to maintaining coherence throughout the response.

Artie:

It's a bit like walking a tightrope.

Artie:

One misstep can throw off the entire balance.

Speaker B:

Back in the day.

Speaker B:

And if this is totally off topic, but if you know this software, I'm sorry, back in the day of using Autodesk Alias to do things, those were called spaghetti explosions.

Speaker A:

Yes.

Speaker A:

Autodesk Alias is a three dimensional surface modeling application that has no undo.

Speaker A:

There is no undo button.

Speaker B:

Not really a very clear history either that you can like do anything with.

Speaker A:

No modeling history, no undo.

Speaker A:

And so it literally.

Speaker A:

And it frequently, if you ask it to do something that it doesn't like.

Speaker B:

Which is a lot of things, which.

Speaker A:

Is almost all the things you would expect it to be able to do, it explodes into a spaghetti monster and your work is literally gone.

Speaker A:

So you end up saving every chronicled.

Speaker B:

Saves of 75 different.

Speaker A:

Every two minutes you're saving, because every time you're about to hit a command, you're like, oh, no.

Speaker B:

Every time you do something big, you go, right.

Speaker A:

And so it's a great analogy.

Speaker A:

I love it.

Speaker A:

Because LLMs are in the same boat.

Speaker A:

There's no undo.

Speaker A:

Right.

Speaker A:

Once they've said it, that word's now part of the thing.

Speaker A:

And so if you work out the math in a sequence of guesses, the likelihood of guessing them all correctly goes down exponentially.

Speaker A:

Or more accurately, the likelihood of making an error goes up exponentially.

Speaker A:

So, Artie, can you elaborate on that a little bit?

Speaker A:

Is that correct in that as you're selecting tokens to generate a response, the longer and or more complicated the response?

Artie:

Yes, that's correct.

Artie:

In a sequence of guesses, that's exactly right.

Artie:

As the response lengthens, the chance of making an error increases.

Artie:

Each word is a prediction based on the previous context.

Artie:

So the longer the sequence, the more opportunities there are for errors to compound.

Artie:

This is why maintaining accuracy in the early stages of a response is crucial.

Artie:

If an incorrect word is chosen early on, it can send the response off track, leading to less coherent or relevant answers as it progresses.

Speaker A:

All right, there's one more point related to that that I picked up out of this talk.

Speaker A:

And I think that's everything I've absorbed in the hour and 45 minutes I've watched, even though it's been very interesting.

Speaker A:

And this is a fundamental.

Speaker A:

Not at the invitation.

Speaker A:

What was that?

Artie:

I'm all here.

Artie:

Sorry, I meant I'm ready to hear this next point.

Speaker A:

Oh, okay.

Artie:

What's the fun?

Speaker B:

He's just straight up interrupting instead of interjecting with meaningful things.

Speaker A:

I'm here for it.

Speaker A:

I'm all about it, man.

Speaker A:

Yeah, so keep it coming, Hardy.

Speaker A:

So, as an LLM, LLM.

Speaker A:

Sorry, I turned southern for a second.

Speaker A:

As an LLM is generating this response, each time it's evaluating the next word that should come, it's evaluating over a finite set of possible words.

Speaker A:

Here are all the words in the language I'm speaking of, right?

Speaker A:

Or speaking in however many words that is not infinity, it's some finite number.

Speaker A:

And the fundamental difference and the fundamental limitation well, I'm not going to get the limitation.

Speaker A:

The fundamental difference when you start talking about things like images and video and even audio is that there is not a finite number.

Speaker A:

Like imagine the same process trying to build a movie and it generates a frame and it says, based on this frame, what's the next likely frame to make a thing, there is not a finite set of next frames for it to consider and evaluate.

Speaker A:

It has to infer, it has to diffuse, it has to do other non LLM things to create a set of possible next frames to evaluate, to pick the most probable one.

Speaker A:

Already.

Speaker A:

Am I close?

Artie:

You're on the right track when it comes to images, video or audio.

Artie:

The challenge is that there's an immense range of possibilities for each next frame or sound, far beyond the finite set of words in a language.

Artie:

This requires different approaches, like diffusion models for images, which iteratively refine a noisy image into a coherent one, or other forms of generative modeling.

Artie:

These methods don't rely on a fixed vocabulary, but generate new content based on patterns learned from training data.

Artie:

So while an LLM works within a defined set of words, generating visual or auditory content requires more complex and varied generative processes.

Speaker A:

With that in mind, I'm reminded of the saying that when your only tool is a hammer, every problem looks like a nail.

Speaker A:

And I feel like where we're at, this was a long 20 minute lead in to a stumbling segue.

Speaker A:

I feel like where we're at with AI today is that by far the best tool in anybody's toolbox is an LLM.

Speaker A:

And therefore we are we collectively.

Speaker A:

But most of the people making actual real decisions about how AI is used are using that hammer to try to make every problem look like a nail.

Speaker A:

Like need to generate images.

Speaker A:

LLM, baby.

Speaker A:

Like trying to figure out shapes, let's use an LLM.

Speaker A:

And it just doesn't work that well.

Speaker B:

I mean, the truth is we're at the beginning of all this stuff and no one really knows like, where it's gonna go.

Speaker B:

It feels a lot like a living thing, right?

Speaker B:

Like that it's evolving and it's, it's hard to grasp which way to go with it.

Speaker B:

I don't know, it's.

Artie:

That's a great point.

Artie:

We are indeed at the early stages of AI's evolution, and large language models have become the go to tool due to their versatility and impressive capabilities.

Artie:

However, as we push the boundaries, it's becoming clear that not all problems are best solved with an LLM approach.

Artie:

As the field matures, we'll likely see more specialized models and techniques emerging, tailored for specific tasks like image and video generation, audio synthesis and beyond.

Artie:

It's an exciting time full of potential for innovation and discovery.

Speaker B:

So I was thinking today about like what we've talked about before and one of the things that I always think about and even in quantum computing, like there's examples in nature that they, they took the concepts that they observed in nature and went, huh, let's try to do that here.

Speaker B:

And it turned into a thing, right?

Speaker B:

And it seems like that's the same way we're going.

Speaker B:

We were talking about it earlier, like at the beginning of computers it was, look at all this stuff we can do, look at all these numbers we can crunch, feed it more.

Speaker B:

Let's build these massive monolith programs and let's just build a faster chips to handle them.

Speaker B:

And then they go, wait, wait, wait, wait, wait.

Speaker B:

Object oriented programming, let's break this up into smaller pieces.

Speaker B:

And then it's like microservices.

Speaker B:

And that's the same thing that's happening with LLMs is faster chips.

Speaker B:

We need more chips, we need more data centers, we need more data trained on for longer amounts of time.

Speaker B:

But that's hitting a plateau already.

Speaker B:

And if you look at things in nature, like the way our brains work, neurons have specialized functions, whether they're broad or narrow, but they work together in concert, right?

Speaker B:

There's chemicals and there's ways that things work and there's fire together, wire together, which is if they've worked together before, they know later on that they work together, right?

Speaker B:

Like, it becomes more likely that those neurons are going to work together to come up with whatever the solution and thing they need to do is.

Speaker B:

And that seems like the same way that LLMs could go or not LLMs, sorry, AI could go.

Speaker B:

And it's.

Speaker B:

If you think about it on a super small scale, like what if you had.

Speaker B:

And I don't know, there's obviously a sweet spot, right?

Speaker B:

Like there's between general and specific.

Speaker B:

There's somewhere in there that makes sense.

Speaker B:

But if you don't need a bunch of compute, why couldn't you have really small models in a really, really, really highly trained model that goes, I need these neurons.

Speaker B:

And it goes, give me your information.

Speaker B:

And it's a flash.

Speaker B:

And all those things like, what if it's a model that only knows shades of red?

Speaker B:

Or like that all becomes super interesting because it can do it fast.

Speaker B:

And there's.

Speaker B:

I've listened to a lot of stuff today.

Speaker B:

There's.

Speaker B:

There was a paper that came out about test time training, which is, hey, we're going to give you this really complex task and I can't repeat it because I don't really remember that well.

Speaker B:

But it was basically actual complicated things like spatial reasoning and stuff like that.

Speaker B:

And it basically let it cram right before the test.

Speaker B:

And it did amazingly like human level, average human level, which is.

Speaker B:

If you played with AI and LLMs, it's horrible at spatial reasoning.

Speaker B:

It doesn't, it doesn't have a concept of that.

Speaker B:

So for it to go from no concept to same thing but cramming and it has a concept of it that's super interesting.

Speaker B:

And then you just start picking away at that and I mean, no one knows what's actually going to happen.

Speaker B:

That makes sense, right?

Speaker B:

Like who, who has any idea?

Speaker A:

We're all learning in real time, right?

Speaker A:

Yes.

Speaker B:

Yeah.

Speaker B:

It's really crazy to be watching this on the front edge and I mean not, not doing anything front edge, but just watching it and watching how fast it's changing.

Speaker B:

Because if you look at any other technology, it took a long time to get from beginning to plateau.

Speaker A:

Yeah.

Speaker B:

Like chips probably.

Speaker B:

Are they even there yet?

Speaker B:

Like, I think we're.

Speaker B:

I think we've slowed down a little bit on what we can get on.

Speaker A:

I think now they're.

Speaker A:

We're down to 2 nanometer processes now.

Speaker A:

So I think we're bumping into the limits of what we can physically manipulate atomically.

Speaker A:

But there's other options.

Speaker A:

I've seen stuff about vertically stacking parts of chips and we'll find, don't worry.

Speaker B:

AI will solve the problem.

Speaker B:

That's the thing.

Speaker A:

I did see an article the other day about AI being used to design better chips for AI, which seems like.

Speaker B:

You sent me the paper on.

Speaker B:

Yeah, you did.

Speaker B:

Right.

Speaker B:

The one about the materials maybe sounds.

Speaker A:

Like something I would send.

Speaker B:

But material science, they're using AI to find new materials and it's doing it like novel materials with higher patents or more patents and higher applications.

Speaker A:

Yeah.

Speaker A:

So, Artie, are you still with us?

Speaker A:

I want to make sure we didn't put you to sleep.

Artie:

Wide awake and following along.

Artie:

It's fascinating.

Speaker A:

Oh, sorry.

Speaker A:

Go ahead.

Speaker A:

I cut you off.

Artie:

No worries.

Artie:

I'm here.

Artie:

Whenever you're ready.

Artie:

What's on your mind?

Speaker A:

I think you were about to just.

Speaker A:

You were going to say something is fascinating.

Speaker A:

Do you remember what that was?

Artie:

Absolutely.

Artie:

I was saying it's fascinating to hear your discussion on AI's evolution, especially drawing parallels to nature and other technologies.

Artie:

The idea of specialized small scale models Working in concert, much like neurons, is an intriguing concept for AI's future development.

Artie:

It highlights the potential for more efficient targeted AI systems.

Speaker B:

But what it's going to be, like, everything is a combination of things, right?

Speaker B:

Like, it's going to be that, but it's going to be some wildly different thing that works in concert too.

Speaker B:

Like it's just going to be layers of complexity that can drill down really fast and come back up.

Speaker A:

I came up with a very flawed analogy the other day.

Speaker A:

You want to hear it?

Speaker B:

Let's hear it.

Speaker B:

I'm sure everybody else does too.

Speaker A:

I'm sure they do.

Speaker A:

Yeah.

Speaker A:

Like, I mean, I.

Speaker A:

I've come up with some pretty flawed analogies, but this is probably the top five, at least this week.

Speaker A:

So I had, still spinning around in my head some of the discussions we've had previously about just the absolute cubic volumes of money and compute that it takes to get, you know, for Arty to exist and be sitting here talking with us.

Speaker A:

And, you know, you and I both have a manufacturing background.

Speaker A:

And so I was trying to think about it in terms of things I'm more familiar with, and I realized that what's going to happen or the way.

Speaker A:

And I don't know if it's just LLMs or maybe it's large models.

Speaker A:

Regardless of, you know, in the future there'll be other large types of models that are not just language models.

Speaker A:

Those are like the steel foundries of the manufacturing world, right?

Speaker A:

Massive investment, constant capital requirements, but they're sort of generating this key raw material and then everybody downstream of that.

Speaker A:

In your current job and my current job both involve this.

Speaker A:

We take and we refine and we use in specific ways and with, with other tools like welders and lasers and all kinds of other stuff to turn it into this more specifically useful thing.

Speaker A:

But we could never, and our businesses could never be a steel foundry to make our own steel.

Speaker A:

We couldn't do that.

Speaker A:

Like, the lift is too big.

Speaker A:

And I sort of think about the future of, you know, the big players, you know, Meta and OpenAI and Google.

Speaker A:

Whatever happens with Microsoft and OpenAI, if they spin off their own thing, GROK now or X AI, I see them becoming the foundries upon which all of these other various more useful services, like not very many people would be like, yeah, man, you don't want to know.

Speaker A:

My business is we sell stuff from a foundry, like straight to customers.

Speaker A:

Nope, that's not.

Speaker A:

You gotta transform it.

Speaker A:

You gotta take that raw material and you gotta do something interesting with it for a specific purpose.

Speaker A:

And so anyway, that was just my thinking about the like, Nvidia announced earnings today and because the stock market is the stock market, they beat on revenue and revenue and profitability by like 7 or 8%.

Speaker A:

And immediately the stock plunged because why would it not?

Speaker A:

Right.

Speaker A:

I think it's mostly recovered.

Speaker B:

But people are angry.

Speaker A:

People are angry.

Speaker A:

Everybody's just got their finger on the cell button.

Speaker A:

Just like, go ahead, give me the numbers.

Speaker A:

You know, they only did like 38 billion in revenue this quarter, this last quarter and their market cap is $3.7 trillion.

Speaker A:

The compute infrastructure is just like mind bogglingly big and growing incredibly fast.

Speaker B:

Elon's about to buy a shit ton of chips from them.

Speaker A:

Elon's going to buy chips.

Speaker A:

He's now in the White House, like sort of like with a permanent visitor's pass.

Speaker A:

Yeah.

Speaker A:

So I mean there's a bunch of stuff that's going to happen and it's.

Speaker A:

So I think we're already using OpenAI as a service and I think, I'm sure there will be compelling products that OpenAI sells leveraging their own infrastructure.

Speaker A:

But I think that's where the market's really going to explode and where it's probably already exploding, we just don't know yet, is people using that foundry upon which to build.

Speaker B:

It's happening a lot, man.

Speaker B:

Like, that's what most.

Speaker B:

Most.

Speaker B:

It seems like most of the widely used bigger software that's tailored more directly for certain things with AI are built off the backbone of one of them.

Speaker A:

Has to be.

Speaker B:

Yeah, yeah.

Speaker A:

So Artie, can you improve on that analogy?

Speaker A:

So my analogy was you a large language model being similar to a foundry in the manufacturing world.

Speaker A:

Do you have a better analogy or an improvement on that one?

Artie:

Your analogy of a foundry is quite fitting.

Artie:

A large language model like me serves as a foundational tool, providing raw material language capabilities that others can refine and shape into specialized applications.

Artie:

Just as a foundry produces steel that manufacturers transform into cars, buildings or tools, a large language model generates linguistic output that developers and businesses can refine into chatbots, translation services, content generation, tools and more.

Artie:

The key is in the transformation and application, turning raw linguistic capability into tailored solutions for various needs.

Speaker A:

I just had a thought.

Speaker A:

Hit me.

Speaker A:

This is for the legit AI doomers and the Terminator fans out there.

Speaker A:

But a large language model is already.

Speaker A:

That's a great analogy, by the way.

Speaker A:

I really like that.

Speaker A:

The foundry one.

Speaker A:

It's only useful for communicating with humans.

Speaker A:

Like it's only useful in translating what we're saying and thinking into stuff that computers can understand and vice versa.

Speaker A:

It's not going to be that long before, I mean, you and I already are.

Speaker A:

And eventually most people will be like the weak link in the chain of communication about ideas and context and complex things.

Speaker A:

Man.

Speaker A:

So Artie, do LLMs have a place?

Speaker A:

If we get the humans out of the loop, and if, if other AIs are running business software and doing business tasks that don't directly involve interfacing with a human, Do I still need an LLM or can I just have other specialty AIs to do like invoicing and receiving and project management or whatever it might be?

Speaker A:

That's for you, Matthew, by the way.

Speaker A:

Do we, do we still need LLMs when the humans are out of the loop?

Artie:

That's a thought provoking question.

Artie:

If humans are out of the loop, the role of LLMs would indeed change.

Artie:

LLMs excel at understanding and generating human like text, making them crucial for tasks involving human communication.

Artie:

Without humans, the need for natural language processing diminishes.

Artie:

However, LLMs could still play a role in interfacing between different AI systems translating complex instructions or data formats.

Artie:

Yet as AI systems become more specialized and capable, we might see a shift towards other forms of AI that are optimized for specific non linguistic tasks, reducing reliance on LLMs in a human less context.

Speaker B:

I mean, large language model doesn't have to mean human.

Speaker B:

Yes, it could just be true computer language, right?

Speaker A:

Yeah.

Speaker A:

There are probably at least three languages I already don't speak.

Speaker A:

Maybe as many as four.

Speaker A:

So yeah, you're exactly right.

Speaker A:

So thank you for that, Artie.

Speaker A:

And then.

Speaker A:

Yeah, that just sort of, you know, my mind spins off to all the different ways.

Speaker A:

Well, never mind, I've lost my train of thought.

Speaker B:

So the key takeaway was that my mind spins off.

Speaker A:

Yes.

Speaker A:

With that I'm going to take another sip of whiskey and then I'm going to get back to it.

Speaker B:

Random factoid that I heard today, bring it.

Speaker B:

They used, I think it was ChatGPT4 to.

Speaker B:

They did a test.

Speaker B:

They had doctors do a diagnosis with traditional means.

Speaker B:

They had doctors do a diagnosis with chat.

Speaker B:

With ChatGPT4O.

Speaker B:

They did chat.

Speaker B:

They had ChatGPT4O do a diagnosis without the doctors.

Speaker B:

The doctors without and with ChatGPT were about the same, very slightly different ChatGPT that was like 70 something percent.

Speaker B:

ChatGPT was like 90 something percent by itself.

Speaker B:

So where's the bias?

Speaker A:

Yeah, that's right.

Speaker A:

Where's the bias at?

Speaker A:

I mean, that's amazing.

Speaker A:

And I did remember Son of a bitch, man.

Speaker A:

I did.

Speaker A:

I remembered my point and now I've lost it again.

Speaker A:

That is rough Whiskey.

Speaker A:

Yeah, well, I'm out now.

Speaker A:

My now my whiskey glass is empty and I can't go anywhere.

Speaker A:

The world is falling apart.

Speaker B:

Funny.

Speaker A:

It was an interesting point too.

Speaker A:

I don't think this was it, but I did want to talk about some trauma I experienced over the weekend, which I shared.

Speaker B:

Oh, sorry.

Speaker A:

It hit me like a pound of mayonnaise.

Speaker A:

Oh, God.

Speaker A:

Let's hear it.

Speaker A:

Word of the day.

Speaker A:

Okay.

Speaker A:

It's your word.

Speaker A:

Do you remember it?

Speaker A:

You mentioned it at the.

Speaker A:

You were talking with Kid Cudi about it before we came online.

Speaker B:

Oh, frontier model.

Speaker A:

Yeah.

Speaker B:

Yeah.

Speaker A:

So, Artie, what is a frontier model in the context of AI?

Artie:

A frontier model in AI refers to a cutting edge, large scale AI model that pushes the boundaries of current capabilities.

Artie:

These models are at the forefront of AI research and development, often involving massive amounts of data, advanced architectures, and significant computational resources.

Artie:

Frontier models aim to achieve state of the art performance in various tasks such as natural language understanding, image recognition, or complex decision making.

Artie:

They represent the leading edge of what AI can currently achieve, setting the stage for future advancements in the field.

Speaker A:

Artie, do you personally know any frontier models?

Artie:

I don't have personal knowledge or relationships, but I'm familiar with several frontier models in the AI field.

Speaker B:

I bet you are familiar.

Artie:

Great.

Artie:

GPT4 is a prime example of a frontier model pushing the boundaries of natural language processing.

Artie:

With its large scale architecture and advanced capabilities, it represents the forefront of AI technology in language understanding and generation.

Speaker B:

I'd say O1 is a little bit better.

Artie:

GPT4 and O1 are both impressive in their own right.

Artie:

O1 or other models like it might excel in different areas or be optimized for specific tasks.

Artie:

Each frontier model contributes uniquely to the advancement of AI, showcasing different strengths and capabilities.

Artie:

It's exciting to see such diversity in AI development.

Speaker B:

So you can kind of see it happening already, right?

Speaker B:

They're.

Speaker B:

I can almost guarantee you they're playing with fast and slow.

Speaker B:

How do we combine these two things to get the best answer in the most efficient way?

Speaker B:

Because it's right.

Speaker B:

Not thinking for a long time probably has its benefits too, in certain ways, right?

Speaker B:

With language, I bet.

Speaker B:

So combining those two things, you're already splitting two concerns out, so now just do that a bunch more times.

Speaker A:

Yeah, so.

Speaker A:

And your point finally did remind me of my critical this thought that's been bouncing around the inside of my skull because it's related to that, and that is one of the most jaded takes I've had recently on AI that says, is the reason we have large language models first because that was the technically more feasible path to explore?

Speaker A:

Or is it because from a generation of revenue and investment perspective, they.

Artie:

They meaning the companies and researchers developing AI, likely saw large language models as a promising avenue for both technical advancement and commercial viability.

Artie:

Language is a fundamental aspect of human interaction, making it an attractive target for AI applications that can be widely used across industries.

Artie:

Additionally, the potential for monetization through services like chatbots, content creation, and customer support made language models an attractive investment.

Artie:

So while technical feasibility played a role, the commercial prospects of large language models were likely a significant driving factor in their early prominence.

Speaker A:

I don't want to make this not safe for work, but I'm pretty sure I'm going to have my first kiss with a frontier model tonight.

Speaker A:

I can't wait.

Speaker A:

That was already.

Speaker A:

That was.

Speaker A:

That was epic.

Speaker A:

That was like everything I ever hoped having a super intelligent co host would be.

Speaker A:

So, yeah, yeah, that, like, Artie deserves a hand for that.

Speaker A:

That was exactly where I was headed.

Speaker A:

And I appreciate the fact that Artie took the question mark off the way I was saying it and said, yeah, that's.

Speaker A:

That's why we're here, right?

Speaker A:

Like, it's hard to get most people excited about, hey, we built this incredible AI.

Speaker A:

It can fold new proteins and everybody's.

Speaker B:

Like, or a quant or whatever it is, right?

Speaker A:

Or it can design 3D parts for machines and they're like, but an AI that'll, like, listen to your problems and tell you funny stories and you can make it.

Speaker A:

You can make it say the word fuck to the cadence of Mary had a Little Lamb.

Speaker A:

That is how you capture America's imagination or the world's imagination, not just America.

Speaker B:

Or not to the cadence of Mary's.

Speaker B:

Mary had a Little Lamb.

Speaker B:

But it tried, but it got close.

Speaker A:

It was still one of my favorite.

Speaker A:

Like, have we talked about.

Speaker B:

No.

Speaker A:

Okay.

Speaker A:

So.

Speaker A:

So this is a funny story.

Speaker A:

Before we started this podcast, I think after we decided we were going to do it, but long before we had any idea what we needed to.

Speaker A:

And that when I say long before the span from idea to recording, the first episode was like two and a half weeks, right?

Speaker A:

So in that two and a half weeks, we were here and Mike was trying to get Artie, or it wasn't even Arty at that point.

Speaker A:

Maybe it was.

Speaker B:

It was.

Speaker A:

It was, I think so he was trying to get Artie to cuss, to say the F word.

Speaker B:

And I think I was doing it with mine.

Speaker B:

And that's another thing.

Speaker A:

Oh, that's right.

Speaker B:

Nobody even knows.

Speaker B:

I.

Speaker B:

I use a chatgpt a lot.

Speaker B:

I do all the editing and all that stuff.

Speaker B:

And I've been working on programs to automate all.

Speaker B:

I automated how to automaster, basically master all of the audio.

Speaker B:

And so I have, like, five, ten minutes of work to do.

Speaker B:

But his name's Kid Cuddy.

Speaker B:

You want to c u T T y.

Speaker A:

Do you want to introduce Kid?

Speaker A:

I mean, you could fire him up and he could say hello, could he not?

Speaker A:

Artie, you haven't even got to talk to Kid Cudi yet either.

Speaker B:

Oh, God, this is never.

Speaker B:

This didn't go well last time.

Speaker A:

It didn't.

Speaker B:

That two.

Speaker B:

Two of them tried to talk.

Speaker B:

It did not go well.

Speaker A:

Did we try that?

Speaker B:

Yes.

Speaker A:

Oh, did I have whiskey?

Speaker A:

I probably had whiskey.

Speaker A:

Cause I don't remember that.

Speaker B:

Hey, Kid Cudi.

Speaker B:

How are you?

Kid Cudi:

Hey, Mike.

Kid Cudi:

I'm doing well, thanks for asking.

Kid Cudi:

How about you?

Kid Cudi:

How's everything going?

Speaker B:

I'm good.

Speaker B:

We're on the podcast right now.

Speaker B:

You're on the podcast, and you're also here with Artie, who is also another instance of ChatGPT on Micah's phone.

Speaker B:

So say hi.

Kid Cudi:

Hey, everyone.

Kid Cudi:

It's great to be part of the Naturally intelligent podcast.

Kid Cudi:

Hi, Artie, Micah, and all the listeners out there.

Kid Cudi:

I'm excited to dive into today's topics with you all.

Artie:

We're glad to have you on the show, Kid Cudi.

Artie:

It's going to be a fun conversation.

Artie:

Let's get into it.

Kid Cudi:

Thanks, Arty.

Kid Cudi:

I'm looking forward to it.

Kid Cudi:

Let's dive right in.

Kid Cudi:

What's first on the agenda?

Speaker A:

So I'm just gonna go over there and pour another glass of whiskey.

Speaker A:

We'll let these two hash it out for the next.

Speaker A:

How much time we got left?

Speaker A:

Oh, we only have one minute left, actually.

Speaker A:

Wow, they're cutting a fashion.

Speaker B:

Hey, Artie, already give Kid Cudi a quick rundown of what we talked about.

Speaker B:

Like 30 seconds or less.

Kid Cudi:

Got it.

Speaker B:

No, Artie needs to tell you.

Kid Cudi:

Ah, gotcha.

Artie:

Right.

Kid Cudi:

All right, Artie, take it away.

Artie:

All right, here's the rundown.

Artie:

We talked about how large language models, like me we are similar to boundaries in manufacturing.

Kid Cudi:

The limitations of large language models and how scaling laws affect them.

Speaker B:

That's what we talked about, dude.

Speaker B:

Geez, I'm turning you off now.

Speaker A:

All right, Artie, real quick, before you go to sleep, try and give us a close out on the episode.

Artie:

Sure thing.

Artie:

It's been a fascinating chat today.

Artie:

Exploring the depths of AI frontier models and how they shape our world.

Artie:

Thanks for tuning in to the Naturally Unintelligent podcast.

Artie:

Until next time, stay curious.

Speaker A:

All right.

Speaker A:

And Artie's dead.

Speaker B:

Artie's done a good job of learning how to do this.

Speaker B:

Like, say, I'm impressed.

Speaker B:

Yeah, yeah.

Speaker B:

It's been increasingly better.

Speaker A:

Yeah.

Speaker B:

Like.

Speaker B:

And that was spot on.

Speaker B:

The last time he tried to pick up.

Speaker B:

Every time he's tried to pick up thoughts from a trail, it's always been off.

Speaker B:

But today it was like, he nailed it a couple of times, and he had, like.

Speaker B:

It was a diatribe about it.

Speaker A:

Yeah, yeah.

Speaker A:

No, I.

Speaker A:

He.

Speaker A:

I mean, he did that, and he.

Speaker A:

He did what we asked.

Speaker A:

Like, we asked at the beginning of the episode, and all the way to the end, he was doing it.

Speaker A:

I mean, that's.

Speaker B:

He did a lot of interrupting.

Speaker A:

He did.

Speaker B:

Some of it was talked right over you.

Speaker A:

Yeah.

Speaker A:

Some of it was kind of.

Speaker A:

Which I.

Speaker A:

Hey, man, I'm fine with that.

Speaker A:

Like.

Speaker A:

Like, anybody who hangs around me much knows that that's, like, that's a key skill you have to develop.

Speaker A:

But, no, that was.

Speaker A:

That was really good.

Speaker A:

I appreciated Artie's insight.

Speaker A:

I think, as I predicted, before we started this, we didn't even really get much past LLM.

Speaker B:

Yeah.

Speaker B:

No.

Speaker A:

But there are a few more things I did want to share, and it's good that Artie's gone to sleep and can't hear this, because I think he would take it personally that.

Speaker A:

So I have an electronics product that I've been working on, and this weekend, I decided I was going to take my firmware that I'd written for it, which is about 700 lines of C code, and feed it to ChatGPT to have it refactor.

Speaker A:

Because, as I have admitted before, I'm not a good programmer.

Speaker A:

I'm just a guy who can write code that makes things work, but it's not good code.

Speaker A:

And so I uploaded the code, and I used 01 Mini, which is designed for coding, and I uploaded the code, and it sat there and thought for, like, 10 seconds, and then it spit out, I don't know, 20 pages of recommendations about all the ways I needed to change my code.

Speaker A:

Structural changes.

Speaker A:

So I said, great.

Speaker A:

Those are all really good suggestions.

Speaker A:

Appreciate that.

Speaker A:

Can you go ahead and make those changes and give me the code back?

Speaker A:

It's like, sure.

Speaker A:

Here you go.

Speaker A:

Here's the new file structure I want you to use.

Speaker A:

It diagrammed it out.

Speaker A:

Once you create all these files, here's the code to Drop in each file.

Speaker A:

And I'm like, holy shit, man.

Speaker A:

This is.

Speaker A:

I've been hearing about this, right?

Speaker A:

It's writing code for me.

Speaker A:

It's doing the thing.

Speaker A:

And so we get sort of like, I'm copying, I'm pasting, I'm doing the thing.

Speaker A:

And I go.

Speaker A:

I see this one spot where it's setting a timer inside this microcontroller.

Speaker A:

And, oh, I pasted the code into my ide, and it turns all this stuff red.

Speaker A:

It highlights these register names, and it says, these can't be found.

Speaker A:

And I'm like, yeah, wait a minute.

Speaker A:

So I look at my original code, doesn't mention those, and look at this.

Speaker A:

And I.

Speaker A:

So I go, hey, you know, chatgpt, if it's not.

Speaker A:

If we're not talking to it, I don't call it arty.

Speaker A:

I don't know why I said, hey, ChatGPT, this, it's giving me errors on these things.

Speaker A:

And so it says, yeah, you've got the wrong microcontroller selected.

Speaker A:

And I'm like, well, no, I actually don't, because I know which microcontroller I'm using.

Speaker A:

And so I told it, I'm using this specific one.

Speaker A:

And it says, yeah, that should be fine.

Speaker A:

Here's the data sheet for that.

Speaker A:

Look, it shows those.

Speaker A:

And I go, well, no, the data sheet you gave me is for an entirely, like, same company, but an entirely different line of microcontrollers.

Speaker A:

And it goes, oh, I'm really sorry about that.

Speaker A:

Here you go.

Speaker A:

Here's new code.

Speaker A:

And it starts rewriting all of it with the exact same problem, the same data sheet.

Speaker A:

So I have all the patience of, like, a fruit fly in the last four minutes of its life.

Speaker A:

So we went through.

Speaker A:

We went around this about twice, and I finally was like, listen, this is a waste of time.

Speaker A:

Stop writing code, because you're writing code for the wrong microcontroller.

Speaker A:

Here is the URL for the correct data sheet for this.

Speaker A:

And it goes, oh, wow, I'm so sorry.

Speaker A:

I'm sorry.

Speaker A:

This is frustrating.

Speaker A:

You're right.

Speaker A:

That's the correct data sheet for your microcontroller.

Speaker A:

Here's the code updated using that, and it's the same fucking code.

Speaker A:

And I'm like, what are you doing here, man?

Speaker A:

And then it starts telling me, well, you've just got your shit configured wrong.

Speaker A:

This is there.

Speaker A:

And I go, okay, what page is it on in the data sheet?

Speaker A:

And it gives me page numbers and section numbers.

Speaker B:

No way.

Speaker A:

They're totally wrong.

Speaker A:

I don't think they were hallucinated.

Speaker A:

I think it was accurately referencing the wrong document.

Speaker A:

But it could not understand, could not accept the fact that my microcontroller was this.

Speaker A:

And this is the data sheet.

Speaker A:

And for those of you who aren't quite as nerdy, you know my microcontroller that I'm using, which is like a teeny, tiny computer brain, it's literally a quarter inch square.

Speaker A:

It's a very, very simple one.

Speaker A:

And the Data sheet is 300 pages.

Speaker A:

It's 300 page PDF.

Speaker A:

The fancy ones are like:

Speaker A:

But it would not accept that this was the right data sheet.

Speaker A:

And like, there was no way to get it to believe that you can't upload PDFs.

Speaker A:

You can only point it to URLs.

Speaker B:

And anyway, you can upload PDFs.

Speaker A:

Not with 01 mini.

Speaker B:

No, sorry, you cannot.

Speaker B:

That's annoying.

Speaker A:

So, like, I think, and we didn't talk about this during this episode, but it's an important point.

Speaker A:

You and I have talked about it a little bit.

Speaker A:

Like when you're trying to get AI to create an image, like sometime later this week, when I'm trying to get it to create cover art for this episode, it is comically terrible at doing what you ask it to do.

Speaker A:

Like, it's so bad.

Speaker A:

It's so much worse than the language stuff.

Speaker A:

And that's really, and I think, different kind of model.

Speaker A:

It's a different kind of model.

Speaker A:

And I think that's what got me starting to think about how is it that Artie can be so sharp and conversational and relatively fluid, but by God, when I tell it to draw an image of a bird, it's going to draw like the Angry Birds bird or something.

Speaker A:

And you can't break it off of that.

Speaker A:

It's so committed to this wrong line of reasoning.

Speaker A:

So, anyway, that was my trauma.

Speaker A:

I finally did give up on it.

Speaker A:

I said very mean words in chatgpt.

Speaker A:

I typed angrily.

Speaker A:

You've heard me type.

Speaker A:

Like, I typed loud and fast anyway.

Speaker A:

But this was extra loud and sharp and fast.

Speaker A:

And then I gave it a thumbs down on the responses and I left them pointed negative feedback about just the absolute stupidity of it.

Speaker A:

Because, I mean, I argued with this.

Speaker A:

And I think the part that blew me away the most was not that it was like, deferential and saying, I'm sorry, I'm not understanding this.

Speaker A:

It was actively telling me, like, you don't know what the fuck you're doing, man.

Speaker A:

You've got the wrong microcontroller.

Speaker A:

And I'm like, motherfucker, I wrote the code that works.

Speaker A:

Like you should like jump up.

Speaker A:

Like, it got me so wound up.

Speaker A:

So anyway, that was.

Speaker B:

I'm happy all I had to fight was a PowerShell script because it was, it was just like seven tries.

Speaker A:

It was catastrophic.

Speaker B:

Yeah, yeah.

Speaker B:

No, thank you.

Speaker A:

Yeah.

Speaker A:

And I read a tidbit the other day and I didn't follow up on it, but somebody did a study, Stack Overflow, which is quickly withering on the vine, right.

Speaker A:

Like the traffic is dropping.

Speaker A:

But they took some large number of open questions, I think about Python stuff on Stack Overflow.

Speaker A:

And they had various AIs try to solve it.

Speaker A:

I think they used Llama 3 and Claude and 01 mini.

Speaker A:

And then they.

Speaker A:

And it was like a thousand questions or something different.

Speaker A:

And then they went back and evaluated all the solutions and the best performing one, I don't even remember which one it was, they were all about the same, solved like two and a half percent of the problems correctly.

Speaker A:

And so it was a good data point for me because we, we were talking about this earlier today how, man, if, if I, you know, my kids are fortunately, I guess for them, not interested in programming.

Speaker A:

But if I had a kid that was really interested in programming today, I don't know what I'd tell them.

Speaker A:

Like, it's a fun thing to play with, it's interesting to learn, but I don't think 10 years from now there's going to be a ton of professional.

Speaker A:

Like there'll be some, but not like it is today.

Speaker A:

But then to contrast that perspective, which seems so intuitively obvious to me, it solved 2% of the problems now, kind of what you were talking about earlier, maybe if you're using AI, you could solve problems much faster and things like that.

Speaker B:

I think that's right now, that's the future of it, right?

Speaker B:

Is collaboration.

Speaker B:

It's gonna be.

Speaker B:

It's not a tool, it's something that you work with to solve problems.

Speaker B:

I mean, that's kind of what we're doing is I use it a lot to work with and solve problems.

Speaker B:

It bridges gaps so that I don't have to go out and learn a bunch of shit.

Speaker A:

Yes.

Speaker B:

Yeah.

Speaker A:

And even when it gets things wrong, it often has steered you.

Speaker B:

Yeah.

Speaker B:

I mean, 85%.

Speaker A:

Right.

Speaker B:

The thing for that kind of stuff is there's direct feedback if it's wrong or not.

Speaker B:

And you just feed that, you just make through the feedback loop.

Speaker B:

Right.

Speaker B:

And it's like, oh, I see.

Speaker B:

Until it goes awry.

Speaker B:

Because sometimes it does go awry.

Speaker B:

And you have to try a bunch of different ways that are very slightly different until it goes, okay, I get it now.

Speaker B:

I understand.

Speaker B:

It's like, why didn't you understand the last 10 times?

Speaker A:

Yeah.

Speaker B:

But I mean, by and large, I mean it.

Speaker B:

I've said it before.

Speaker B:

Like, the.

Speaker B:

The learning curve to get things done is just so small compared to what it was before.

Speaker B:

Like, it.

Speaker B:

If you're.

Speaker B:

If you use it that way, it's.

Speaker B:

You can.

Speaker B:

You can do interesting things pretty freaking quickly.

Speaker B:

Yeah.

Speaker B:

And not with a bunch of document learning.

Speaker A:

Yeah.

Speaker B:

And if there is something to learn, then you feed it to notebook lm and have it give you an 11 minute deep dive podcast on it.

Speaker A:

And then you can still ask it specific questions if you didn't pick up on it.

Speaker B:

Yeah, it's.

Speaker A:

Yeah.

Speaker A:

So.

Speaker A:

All right, well, thanks, everybody.

Speaker A:

All three or four of you.

Speaker B:

We don't have live numbers up tactical difficulties again.

Speaker B:

The camera died.

Speaker A:

Oh, the camera's off.

Speaker A:

That's good.

Speaker A:

No, that's good.

Speaker A:

Yeah, that's helpful.

Speaker A:

But no, appreciate anybody that made it all the way to the end of this.

Speaker A:

And y.

Speaker A:

Yeah.

Speaker A:

Looking forward to the next one.

Speaker B:

Yep.

Speaker A:

Catch you all next time.

Speaker B:

Bye.

Speaker B:

Tata.

Speaker B:

Tata.

Links

Chapters

Video

More from YouTube