Artwork for podcast Data Driven
Navigating the Complexity of Operationalizing ML Models
Episode 224th December 2023 • Data Driven • Data Driven
00:00:00 00:49:07

Share Episode

Shownotes

In this episode of Data Driven, our Andy Leonard and Frank La Vigne are joined by Chris McDermott, VP of Engineering at Wallaroo.AI. Together, they explore the challenges and advancements in the ever-evolving world of machine learning and artificial intelligence.

From the importance of ongoing care for machine learning models to the rise of edge computing and decentralized networks, they touch on the critical need for flexibility and data privacy. Chris shares his insights on the technical challenges of AI and ML adoption, as well as his unique career journey. They also discuss the evolution of technology and the potential future impact of these innovations.

Join us for a deep dive into the world of AI, technology, and the future of machine learning with Chris McDermott on this episode of Data Driven.

Show Notes

00:00 Exploring AI, data science, and data engineering.

06:20 Training and inferring are different stages.

08:12 Legacy AI doesn't require neural networks or GPUs.

12:09 Machine learning models require consistent care and monitoring.

15:10 MLOps merges skills, breaks down silos, collaborates.

16:47 Prefer MLOps to avoid namespace collision. DevOps parallels original Star Wars plot.

20:27 Internet-scale operations require automation and resilience.

24:13 Challenges of integrating AI into business processes.

28:03 New push for edge computing in technology industry.

32:05 Edge technology critical, discussed in government tech symposium.

34:50 Navigating from SendGrid to Twilio simplified processes.

36:15 First foray into data, growing knowledge.

39:33 Technology evolves, builds complexity over time.

44:41 Book recommendation: "Seeing Like a State" by James C. Scott discusses legibility and centralization of power in society.

46:28 Predictable tree farming fails due to ecosystem complexity.

Speaker Bio

Chris McDermott is a software engineer and entrepreneur who is passionate about creating products that make machine learning more accessible and manageable for users. His focus is on developing a platform that allows for easy deployment and management of machine learning models using any framework and on any architecture or hardware. He believes that current solutions in the market force users into a specific platform, and he aims to provide a more flexible and efficient alternative. With a strong belief in the potential of his product, Chris is dedicated to making machine learning more accessible and user-friendly for people across various industries.

Transcripts

Speaker:

Welcome to show 350 of data driven.

Speaker:

In this episode, Frank and Andy interview Chris McDermott,

Speaker:

VP of engineering at Wallaroo. Wallaroo helps

Speaker:

customers operationalize machine learning to ROI in the cloud,

Speaker:

in decentralized networks, and at the edge. It's

Speaker:

a fun conversation on MLOps and the future of intelligence systems

Speaker:

and model management. Now on to the show.

Speaker:

Hello, and welcome to data driven, the podcast where we explore the emergent

Speaker:

fields of artificial intelligence, Data science and of course, data

Speaker:

engineering, the fundamental thing that kind of underpins it all. And with

Speaker:

me on this epic road trip down the information superhighway

Speaker:

Is my favorite data engineer of all time, Andy Leonard. How's it going

Speaker:

Andy? Good Frank, how are you? I'm doing alright. I'm doing

Speaker:

alright. We just, we're chatting

Speaker:

in the, virtual green room about some of the logistical challenges we had,

Speaker:

with Microsoft Bookings and how Kind of like you can only have, you

Speaker:

know, like, remember that the pick any 2 triangle, right? Good, fast, and

Speaker:

cheap? Yep. Yep. Like, we can only have 2 things, 2 features of what we

Speaker:

needed to do. Right. Alright.

Speaker:

Despite logistical challenges, we are excited here to have,

Speaker:

Chris McDermott who is the VP of engineering at Wallaroo,

Speaker:

and, he is a passionate, and intellectually

Speaker:

curious professional With excellent communication skills, he

Speaker:

loves hard problems, then he must have definitely loved the

Speaker:

process to get on the show, And,

Speaker:

have yet to meet 1 he couldn't solve somehow. Maybe we should get you,

Speaker:

Chris, to help us with our scheduling stuff. Really? You visit

Speaker:

later? Yeah. So welcome to the

Speaker:

show, Chris. Thank you. Thank you. It's great to be on. It's nice to meet

Speaker:

you both. Well, likewise. Likewise. So you're coming to us from the,

Speaker:

Mile High City. That's right. Awesome place. It's, I was

Speaker:

there once, for internal Microsoft

Speaker:

conference actually. Oh, nice. And beautiful town, like, it was

Speaker:

just really cool. I think it was the 2nd biggest

Speaker:

event that in Denver history was the Microsoft thing. Wow.

Speaker:

And they they literally ran out of hotel rooms like it was.

Speaker:

Oh, wow. It was pretty wild. Yeah. I think it was, just

Speaker:

before one of the big parties had a convention there. And,

Speaker:

they Oh, yeah. Yeah. Yeah. Yeah. I was so I'm actually

Speaker:

slated to head back there next year for a Red Hat

Speaker:

conference, so we'll see Let's see if the hotel situation has

Speaker:

improved. I think it's improved a little bit. The city's been growing a lot. So,

Speaker:

Yeah. Lots of government. Isn't Denver the place that has, like,

Speaker:

the large bear up against the conference center that

Speaker:

Yeah. Yeah. Yeah. Yeah. Yeah. That's exactly right. A giant blue bear appearing in the

Speaker:

window of the conference center. Yeah. I was there. And,

Speaker:

and and I remembered that. That was the That was the first thing I

Speaker:

remember. It was, I was there in

Speaker:

2007 for a Kind of a Microsoft conference. It was

Speaker:

a, Professional Association for SQL

Speaker:

Server. That's what it was called back then. And, I was

Speaker:

actually the first one I spoke at. I've spoken at a bunch since then,

Speaker:

but 2007 in Denver was the first. And,

Speaker:

yeah. Like, I echo what Frank said, beautiful city

Speaker:

and, just very picture picturesque. Yeah.

Speaker:

Yeah. The weather in the mountains are beautiful. Mhmm. Yeah. And

Speaker:

it's funny, like, you know, on the East Coast, we talk about mountains,

Speaker:

but It's nothing like that. Like Right. Yeah. We quite the

Speaker:

same. You would laugh at what we call mountains. Yeah. Right. But

Speaker:

I remember a Robin Williams bit Where he said something like that

Speaker:

people he admired the people in Denver because they got to

Speaker:

Denver and they looked at the mountains and went, Well, I can't say what

Speaker:

he said, but he had a kind of an Elon

Speaker:

moment.

Speaker:

There's so many of those. There's so many No more. We're stopping right here. We're

Speaker:

not going over those mountains. So,

Speaker:

You're VP of engineering at Wallaroo. So tell us a little bit about Wallaroo.

Speaker:

Mhmm. Plus you're also ex data robot too. That that's interesting.

Speaker:

Yes. Yep. Exadata robot. Yeah. So I've been working in the machine

Speaker:

learning and AI space for, about 7 years now, I guess, or 6

Speaker:

years. And, it's been really fun. You know, it's, it's a

Speaker:

good time to be in the business. There's a lot of development

Speaker:

happening, very fast pace of change, which I appreciate.

Speaker:

And, you know, Wallaroo has been really great. Like,

Speaker:

the team is fantastic, and the people are wonderful. And it it's a lot of

Speaker:

fun working, with people that you enjoy hanging out with and and you respect

Speaker:

and everything, that's that's very important to me. That's awesome. But I also just

Speaker:

I think the product is awesome. It's really, I think,

Speaker:

playing well in the market. Like, we are focusing on making it as easy as

Speaker:

possible to deploy And manage machine learning models.

Speaker:

And the focus is really on any model using any framework and being

Speaker:

able to deploy onto sort of any architecture, any hardware,

Speaker:

and being able to leverage GPUs if you need them or different kinds of CPUs,

Speaker:

different acceleration libraries that people have tailored to the different

Speaker:

architectures. And, honestly, there are not a lot

Speaker:

of other solutions that tackle those 2 problems for people. Right.

Speaker:

A lot of the other companies that we're competing with, they are trying to be

Speaker:

like an end to end solution or, like, really force people into, you know, their

Speaker:

platform. So you train on their platform, you deploy on their platform, you manage on

Speaker:

their platform. But it's very limiting in terms of what you can bring on to

Speaker:

the platform and and being able to, deploy on the different types of

Speaker:

architectures and, platforms and things like that. So it's really

Speaker:

exciting. It's fun. I think that's really important that you bring up

Speaker:

the CPU solutions. As I've been tinkering,

Speaker:

you know, the past couple of years with, you know, with the

Speaker:

different, different platforms that are out there, it's

Speaker:

That's definitely a smaller market, but maybe it's emerging now. I'm

Speaker:

just not sure. Mhmm. Yeah. I wonder yeah. Sorry. Go ahead.

Speaker:

Well, I was gonna say, you know, a lot of the time people conflate training

Speaker:

and inferring, which is, you know, sort of the 2 different stages. Like,

Speaker:

1st, you have to train a model, but then you use the model to make

Speaker:

inferences, which, you know, it's really like asking the model to make a prediction

Speaker:

or you give it some input and it gives you some output. And,

Speaker:

they're they're very, very different tasks. And just because, you know,

Speaker:

like, you may wanna use some hardware GPUs for training Doesn't

Speaker:

necessarily mean mean that you need the GPUs when you are in production and

Speaker:

you're asking it for predictions. A lot of the time, you know,

Speaker:

The model is small enough that you really don't need to, but there's

Speaker:

so much hype. It it's hard sometimes to separate the hype from the, you

Speaker:

know, The real stuff and Yeah. Yeah. The hype the hype

Speaker:

machine is real. I mean, like, it's and and and I I wanna get your

Speaker:

thoughts on, you know, I mean, I love generative AI. I'm not

Speaker:

knocking generative AI, but it feels like it's taken all the oxygen out of the

Speaker:

room for All the other kinds of AI.

Speaker:

Yeah. Yeah. Yeah. Because there are a lot of, you

Speaker:

know, great models. I like XGBoost is a very standard one. It's

Speaker:

been around for, you know, a long time, meaning at least for, you know, 5

Speaker:

or 10 years now or something. But, that really honestly solves so

Speaker:

many problems, and it's such a Small, easy model to deploy.

Speaker:

I I wish people would focus more on on that kind of thing rather than

Speaker:

hype. Right. No. That's a good point. And I think you

Speaker:

bring up an interesting point because not all not

Speaker:

all AI workloads are created equal. Right? Obviously, there's,

Speaker:

I heard this term the other day and I had to spit my coffee out

Speaker:

because it was just so funny. Legacy AI. Yeah.

Speaker:

Yeah. There's generative AI now. There's legacy AI. That's

Speaker:

crazy talk. You know? And I was just like,

Speaker:

wow. But,

Speaker:

you know, because, you know, legacy AI, basically,

Speaker:

you're not using deep learning, you're not using neural networks,

Speaker:

Generally, you don't get a good boost from GPU's.

Speaker:

Correct. Right. And that's something that when when you tell that to

Speaker:

Even tactical decision makers, they they they

Speaker:

kinda look at you like, you know, what sorcery is that? Like, you know, because

Speaker:

they'll they'll They'll say, like, oh, we don't have enough GPUs. There's no budget for

Speaker:

GPUs. Like, what what what types of workloads are you running? And I

Speaker:

tell them, it's like, well, it's not really a concern for you. Like, you don't

Speaker:

need them. Yeah. And you see, you know,

Speaker:

the the the the people who are doing the actual data science, they're like, yeah,

Speaker:

duh, that's what we're trying to tell you. Yeah. But you see, like, the leaders

Speaker:

of these teams are like, like, you know,

Speaker:

it's, now Just for my own education,

Speaker:

there wasn't there something called RAPIDS, and it was an

Speaker:

acronym that let you use GPU's For

Speaker:

certain types of like XGBoost, I think was one of them. Random

Speaker:

forest there. I don't know. Oh. You See, it's funny because

Speaker:

it was an it's an NVIDIA thing and obviously it only optimizes on. But,

Speaker:

like, it was I remember Hearing about

Speaker:

it in 2019, and I'm thinking, wow, this is gonna change

Speaker:

everything, and you haven't heard of it. Only,

Speaker:

like, one ever per other person I met in the wild has ever heard of

Speaker:

it, and he was at the same conference I was at where we heard about

Speaker:

it. So I'm like, That's kind of unusual,

Speaker:

but, you know, we gotta watch so

Speaker:

fast, you know, and it's really hard to tell sometimes What

Speaker:

what, which new developments are gonna end up being the future and which ones

Speaker:

are gonna end up as dead ends? Right. You know, and even all

Speaker:

the transformer stuff that that is powering GPT and and those similar types

Speaker:

of models, I think that was originally written up in a

Speaker:

white paper in, like, 2017 something. Mhmm. And it just kinda sat around for a

Speaker:

while, and nobody paid a whole lot of attention to it until OpenAI really

Speaker:

ran with it. So yeah. Pension is all you

Speaker:

need. I think that's was that the paper? Sounds right.

Speaker:

Yeah. And then we're gonna go. Oh, sorry. Go ahead. Sorry,

Speaker:

Andy. I cut you off your point. No. I I don't wanna go too far

Speaker:

downstream before I say cred boost for using the phrase I don't

Speaker:

know. Oh, nice. Somebody with your

Speaker:

credentials, you know, saying I don't know. That's that's super

Speaker:

cool. So cred Honestly, there's way too much to know. There's no

Speaker:

way anyone person could know that. I I like to joke. I

Speaker:

haven't checked my phone or, like, news Feed in like half an hour

Speaker:

and I'm like woefully behind now. Yeah.

Speaker:

But it feels that way like in the whole Oh, no. It does. Yeah. Especially

Speaker:

it was especially interesting when the whole drama on OpenAI, and I

Speaker:

don't wanna go down that rabbit hole too far. But when all of that soap

Speaker:

opera kinda unfolded Yeah. Yep. It was kind of like,

Speaker:

what's the latest? Like, is he back? Is he gone? Is he working at Microsoft?

Speaker:

Like, he did work at Microsoft for like 10 minutes, and now he doesn't.

Speaker:

Like, Yeah. You know, at

Speaker:

at some some point down the middle of it, it's like, call me when this

Speaker:

is over, and I'll deal with the, things yeah. I'll

Speaker:

check-in again. But that's just the human

Speaker:

side of it, let alone the let alone technology side of it.

Speaker:

So Operationalization. I think that's gonna

Speaker:

be the buzzword. Obviously, chatty b t and JennyIs, taking

Speaker:

all the air out of there. And I think the next buzzword It's gonna be

Speaker:

operationalization. 1, because it's kinda

Speaker:

hard to say, and I'm not gonna lie, I've had to practice.

Speaker:

But, it's something that I think companies and organizations

Speaker:

that adopt AI, whether it's legacy AI

Speaker:

Or generative AI. They're gonna have to realize, like, it's one thing to build

Speaker:

the model, and then it becomes a, okay, now

Speaker:

what? Yeah. Yeah. Well and models

Speaker:

really are just like any other software. It's not something that you just

Speaker:

write once and you, you know, Throw it out there, and it runs forever

Speaker:

without being touched. Right? All of it requires care and feeding, and

Speaker:

and machine learning models are no different. So, I think

Speaker:

part of it is, you know, how do you deploy it? And then, you know,

Speaker:

how do you keep that that deployment up to date, you know, getting critical

Speaker:

patches and vulnerability fixes and things like that. But also

Speaker:

how do you monitor the model and how it's performing and how it's performing

Speaker:

relative to the real world, Because the world doesn't stand

Speaker:

still right. So even if the model was trained on some data and it was

Speaker:

98% accurate when it was trained, as the world shifts and

Speaker:

and the situation around it shifts, that accuracy will

Speaker:

almost certainly start to degrade over time. So You need to monitor that. You need

Speaker:

to know when to retrain the model. And you have to be kind of

Speaker:

keeping track of, new training data. Right? So the

Speaker:

the the new environment that the model is operating in, you need to be recording

Speaker:

all of the the inputs and also paying attention to the ground truth of, You

Speaker:

know, what was the outcome of that prediction that the model made? Was it accurate

Speaker:

or not, after the fact? And and correlating that back into your

Speaker:

training data So you can retrain the models and, you know, keep them going

Speaker:

over time. And that's just, you know, assuming you're gonna be using the same model

Speaker:

forever. But as we just finished talking about new models coming out all the

Speaker:

time, new approaches, new techniques. So, yeah, it really is

Speaker:

is something you've gotta pay attention to. It's an

Speaker:

extremely Yeah. It's an extremely dynamic space.

Speaker:

Mhmm. I've heard this called

Speaker:

MLOps for the longest time. Mhmm. Mhmm. But I've also heard a new term

Speaker:

kinda pop up on the radar called AI ops Mhmm.

Speaker:

For this. What do you call it? I

Speaker:

generally call it MLOps. You know, one, I

Speaker:

I sort of per like, AI and ML, there's an

Speaker:

interesting, you know, difference there in in terms of who uses the different terms and

Speaker:

when they use them. For me, AI is

Speaker:

more of a general term that I use conversationally. And most of the time if

Speaker:

I'm trying to be fairly technical and specific, I'll usually revert to ML,

Speaker:

Because in fact, most of these things are machine learning. AI is a much more

Speaker:

nebulous concept, and I I don't even think everybody agrees on on what AI is

Speaker:

or What the threshold would be, you know, if you're

Speaker:

doing statistical analysis, I think most people probably would not call that

Speaker:

AI. But there are a lot of machine learning models that do work that way.

Speaker:

And and that's definitely, like, part of the gradient. You know? I've

Speaker:

noticed that too. Like, there it is a gradient too. Like, there's not a, like,

Speaker:

a hard, like, You know, typically it depends on the audience. Right? If they're

Speaker:

if they're BDMs, business decision makers, they're gonna use

Speaker:

AI. Yeah. They're technically focused people. They tend to prefer the term

Speaker:

ML. Yeah. That's also been my experience Interesting.

Speaker:

Quite often. So I like MLOps because, one,

Speaker:

it sort of grounds you a little bit more in that technical perspective. Mhmm.

Speaker:

And, and it's sort of a like, To me, I think I came up

Speaker:

through DevOps a lot of my, you know, first half of my career was DevOps

Speaker:

and infrastructure and things like that. And, I

Speaker:

think part of the appeal of the term MLOps is it taps into a

Speaker:

lot of the DevOps, associations. Right? And

Speaker:

Right. The concepts and the themes of DevOps, which is really about,

Speaker:

merging different skill sets and breaking down silos and getting different teams to

Speaker:

communicate with each other and And to collaborate more,

Speaker:

being more dynamic. Not just, you

Speaker:

know, putting software out there and and letting it run forever, but Keeping it up

Speaker:

to date, monitoring it, recording the logs, you know, all of that kind of

Speaker:

stuff, and and getting into a flow of continuous

Speaker:

deployment, you You know, continuous integration, continuous testing, continuous

Speaker:

deployment. And I think on the ML side, that's also where

Speaker:

MLOps really shines and and is bringing those themes

Speaker:

to the party, rather than a data scientist training a

Speaker:

model, deploying it, and, you know, Throwing it over the

Speaker:

wall to to, like, an operations team or something. It's

Speaker:

getting all these different teams and skill sets to work together. It's

Speaker:

building a continuous, you know, pipeline, with

Speaker:

monitoring and and feedback loops and so on. So that's that's why

Speaker:

I like MLOps. No. I like that too. So in order to prevent

Speaker:

any hate mail come in or or but actually comments, AI

Speaker:

ops is also used, I've heard, in,

Speaker:

the telcos and network operators tend to have a term

Speaker:

called AI ops, where they use AI to help operate their network. So that is

Speaker:

Got it. It's it's a it's a namespace collision,

Speaker:

which I've free further which I prefer MLOps for to avoid the namespace

Speaker:

collision, plus all the reasons you said. You

Speaker:

know, what's interesting is and I came from a software engineering

Speaker:

background and, you know, and I'll be honest, I was not

Speaker:

initially a big, believer in in DevOps, but

Speaker:

kind of as time went on, I became a convert. But I think

Speaker:

that, you know, when you look at how AI models, ML

Speaker:

models, whatever, how they get operationalized.

Speaker:

You look at it And I I often I often can tell

Speaker:

who the fans of the new Star Wars movies are by using this analogy,

Speaker:

because I'll say it's The 2015 Star

Speaker:

Wars movie and the 1977 movie, DevOps.

Speaker:

DevOps being kinda like the original, episode 4 And then the

Speaker:

new, the the first new one, right? It's the same

Speaker:

plot. I mean, the characters have changed, some things are

Speaker:

different, But very effectively, it's the same plot. And, you

Speaker:

know, some people will laugh like you did, and some people I can

Speaker:

see will, Their their faces turn red. But,

Speaker:

but I mean it's like it's it's the same plat plot. The names, the places

Speaker:

have some have changed. But you're right. I mean, I think and there's a lot

Speaker:

of lessons we can learn Yeah. In the ML community

Speaker:

from the DevOps world. Right? Because, You know prior

Speaker:

to DevOps, you know, the developers and operations had a

Speaker:

very antagonistic relationship for the most part. I'm sure there's always

Speaker:

exceptions. You know, I was I was joking that they would only meet,

Speaker:

they only have to interact 3 times a year, and one of those was the

Speaker:

holiday Christmas party. You know what I mean? And

Speaker:

Yeah. But if you wanna deploy something in a far more agile

Speaker:

way where they have to, you know, you put it In some extreme cases, every

Speaker:

few hours, some new bit of code gets gets pushed up. That's obviously on

Speaker:

on one fore end of the spectrum. But for the most part, you know, a

Speaker:

couple times a month is not unreasonable. You have to automate that. You have to

Speaker:

have processes in place. Yep. And I I see a day, and if that day

Speaker:

has not already come, I would be surprised, That AI is gonna be the

Speaker:

same thing or ML. Right? You're you're gonna have to get but to your point,

Speaker:

right, this is a continuous process, You know? Yep. Yeah.

Speaker:

We can't get away with, you know, you have this isolated team of data scientists.

Speaker:

They they they kinda go off to their little area 51 type labs

Speaker:

in secret. Right. I then come back with some model,

Speaker:

and and I'm guilty of this too. I've done this. Right? Where I'm like, I

Speaker:

built the model. I'm done. I did the math. I did the hard

Speaker:

part. How do you get the play it? Not my problem. Not my

Speaker:

problem. And it's funny, like Yeah. You know, I caught

Speaker:

myself. Right. I caught myself doing that as I, you know, you

Speaker:

know, doing that. Like recently, I had to I had to do a demo

Speaker:

and I had to work on a kind of a It's basically a predictive

Speaker:

maintenance type thing, and I took all this data, had the model, and I

Speaker:

just said, here's the here's the link to the model, Have at it,

Speaker:

pal. Mhmm. And then as I sent that, I was like, you know, I should

Speaker:

probably be more involved in getting this on a race for it.

Speaker:

Right. Yeah. Yeah. Yeah. Yeah. No. I think that's a big part of it.

Speaker:

Another big part of it though is, scale, you know. And I think scale

Speaker:

scaling of compute and, how How people were using compute and how

Speaker:

much compute was required was a big part of what drove DevOps.

Speaker:

You know, if you were a sysadmin responsible for a 100

Speaker:

servers, That's, you know, challenging, but it's feasible. Like, you can do

Speaker:

that. You can keep them all up to date. You can keep them all in

Speaker:

sync with each other. Make sure they they all have the same patch levels and

Speaker:

and so on. But you scale that up to a 1,000 servers?

Speaker:

That gets a lot trickier. You try to go to a 100,000 or, you know,

Speaker:

if you're doing Internet scale things like Google or Facebook or somebody, We're talking

Speaker:

millions, tens of millions. And Right. That level of scale

Speaker:

requires you know, everything has to be automated. Everything has

Speaker:

to Has to work that way and it has to be resilient and it has

Speaker:

to, you know, have automatic fail over and stuff. You know, there's the,

Speaker:

x k CD where they're, You know, they get to a certain point. They're just

Speaker:

roping off entire data centers and being like, alright. We're throwing that one away and

Speaker:

moving on to the next one. And for AI, I

Speaker:

think a lot of the same stuff is happening. When, you know, 10 years ago

Speaker:

or so when when people were just getting started on this journey And as an

Speaker:

entity, as a business entity, if you're talking about 1 or 2 use

Speaker:

cases, you know, you can have humans curate that stuff and hand

Speaker:

craft it, hand roll it, hand deploy it, and hand manage it. But

Speaker:

if you're a a big enterprise company and you you wanna have hundreds of use

Speaker:

cases in production or thousands or tens of thousands, there's just no way.

Speaker:

You have to automate it. No. That that that's a that's

Speaker:

an excellent point. Like, one way I've heard to describe is that if you're

Speaker:

baking a loaf of bread for your family and friends Or loads of

Speaker:

bread. You can do it in your kitchen. Right? You don't have to do anything

Speaker:

special, but if you're the Wonder Bread Corporation or

Speaker:

Mhmm. And you wanna deliver at that scale, that's no longer an

Speaker:

option. Mhmm. And I think that we're at that point where and

Speaker:

correct me if I'm wrong, where I think AI and ML adoption or AI

Speaker:

adoption is still new enough where there's enough of naivete out

Speaker:

there of, oh, we don't need to scale to that degree. Like, we don't need

Speaker:

the production line. I think I think that's ending. I think we're getting close to

Speaker:

the the end of that era, but that's kind of been my yeah. I think

Speaker:

so too. Yeah. Because they're they're more and more, ML

Speaker:

tools in everybody's toolbox. Right? So you were talking about telcos

Speaker:

routing network traffic using ML models. That's not

Speaker:

gonna be 1 model. Right? Like, with latency and and

Speaker:

everything else, you're gonna need, you know, Very small. Lots

Speaker:

and lots of very small models deployed on every router, every top of rack

Speaker:

switch, every, you know, whatever 5 gs cell phone tower,

Speaker:

whatever you're talking about. There are a lot of cell phone towers. So you're

Speaker:

not managing 1 model. You're managing a fleet of models, right, across

Speaker:

different geos and all kinds of things. No. That's that's an

Speaker:

excellent point. Sorry, Andy. That's okay. It does seem to scale like that,

Speaker:

though. Right? It's almost it's almost tectonic. There's

Speaker:

a whole new layer going down. You know? That's that's the new surface.

Speaker:

I noticed on the website, I I popped over to wallaroo dot,

Speaker:

aiwallar00.ai.

Speaker:

And I noticed a familiar looking, blurb just

Speaker:

below the top of page. And it's familiar to me because, I

Speaker:

started off in business intelligence. I'm still working in BI.

Speaker:

And there's a note, 90% of AI

Speaker:

initiatives Failed to produce ROI. And I saw this

Speaker:

in, you know, it's very similar number, 85% in,

Speaker:

in BI back in the day. It's probably still true. So where

Speaker:

does that number come from? Well, I think it reflects a lot of

Speaker:

things. You know? Some of them we were just talking about and

Speaker:

and where MLOps is coming from is is, a lot of the failure

Speaker:

modes were teams not really working with each other. Right?

Speaker:

Somebody decided we should be doing AI, so they hired the data scientist.

Speaker:

And the data scientist works in the corner for a while and,

Speaker:

You know, 1, they don't have access to all the data. They don't know what

Speaker:

the data is, where to find it, how to access it, how to clean it,

Speaker:

what it means to the business. There there are a whole set of challenges there.

Speaker:

And then, you know, they may train some models and and get something, you

Speaker:

know, to a point where they think that it's gonna solve a problem. But Then

Speaker:

you've got to work with an IT organization to stand up infrastructure. You've got to

Speaker:

work with somebody to package the model and build, you know, an API around

Speaker:

it or a UI of some sort And figure out how to deploy

Speaker:

it, train people on how to use it, and and actually, like,

Speaker:

somehow integrate it into your business process So that it's

Speaker:

it's driving business outcomes. And all of those are really tough

Speaker:

challenges. And all of them require breaking down those

Speaker:

silos and getting a bunch of different People within an organization to

Speaker:

talk to each other and communicate and to work together to solve something.

Speaker:

I don't think ML or AI is is a magic wand that you just

Speaker:

wave and magically provide value to a business. You've got to really

Speaker:

think about What is your business doing? And, you

Speaker:

know, machine learning at at heart, it it's

Speaker:

really just like a a more efficient way of

Speaker:

Making decisions, you know, faster and more accurately,

Speaker:

and with less human input. And so you've got to look for places where your

Speaker:

business can either save a lot of money or make a lot of money by

Speaker:

being able to answer a a simple question repeatedly very,

Speaker:

very efficient. And that sounds easy, but in practice,

Speaker:

defining a business problem is often one of the hardest parts. So now I'm

Speaker:

seeing even more parallels. Uh-huh. Yeah.

Speaker:

You know, that was the problem we were trying to solve, with

Speaker:

business intelligence as well. So didn't mean to cut you off. Sorry about

Speaker:

that. No worries. So I yeah. I think I agree with you. It it there

Speaker:

are tons of parallels there. I think there are a lot of similar lessons to

Speaker:

be learned, and I think we are applying them in this In this space in

Speaker:

ways that we've applied them to other spaces in the past. I also

Speaker:

think there are technical challenges. You know, part of it is the field is moving

Speaker:

so fast. So there's just this constant stream

Speaker:

of of new frameworks, new models, new techniques, and you

Speaker:

have to kinda stay on top of that. You have to be careful with your

Speaker:

tool selection, to make sure you're not, you know,

Speaker:

going whole hog into some tool. That sounds

Speaker:

great today, but it's just not flexible, and it's not gonna be able to support,

Speaker:

like, all these new things that are coming out. Yeah.

Speaker:

Or that company could have internal internal political

Speaker:

strife, which was crazy talk. Right? Cast Absolutely. Right. Cast

Speaker:

doubt on their future. Alright. That would never happen. That would never

Speaker:

happen. Sorry. Yeah. You were talking about privacy, which I think is

Speaker:

another key thing. Yeah. Data residency, data privacy, see data

Speaker:

security. You know, all of those things matter tremendously.

Speaker:

And for for a business trying to, get

Speaker:

value out of AI and ML. You know, a lot of it, depends on

Speaker:

having good data and, Cleaning it and curating it

Speaker:

and getting it ready for things. But then it it forces the

Speaker:

the organization to really kind of do an inventory. What do we have? What's useful?

Speaker:

What's not useful? Well, how much do we store? How much do we not store?

Speaker:

How do we comply with various regulatory

Speaker:

environments? Right? GDPR is is the big one everybody, you know,

Speaker:

loves to throw out there. It's it's big and it's complicated, but, you know,

Speaker:

things like that matter a lot. And And there's 300 +1000000 people behind

Speaker:

that. They're covered or whatever. I think that, you know, that that

Speaker:

is not only do they have a big stick, but they have a big arm

Speaker:

that they can wave that stick wet. Yes. You

Speaker:

know, if if a small country with, like, you know, 50 people in it, and

Speaker:

that could something like GDPR, people would just walk around it. But I think

Speaker:

that, a block with I've heard different numbers, but

Speaker:

it's for, you know, pushing 4 to 500,000,000

Speaker:

people. That's a huge that's a big enough market nobody can really ignore.

Speaker:

Yeah. What's interesting is on the LinkedIn page

Speaker:

for Wallaroo I love the website, by the way. I checked that out too. Thank

Speaker:

you. It talks about decentralized

Speaker:

networks Mhmm. And at the edge. Yes. What how would

Speaker:

you define decentralized network? Yeah. This is a big new push for us that we've

Speaker:

been focused on for, I mean, we've been focused on it kinda for the

Speaker:

last year, but it was a lot of, development on on the back end. And

Speaker:

we just released kind of our 1st edge features and product,

Speaker:

in October, so it's kind of a new thing for us. But,

Speaker:

As you think about ML and edge or ML and AI,

Speaker:

and the the fleets of models that we talked about and all these use cases

Speaker:

And, you know, telcos and and five g cell phone towers and all of those

Speaker:

types of things, intersecting with data and data

Speaker:

residency and privacy and security, It it really seems to

Speaker:

indicate to me and and to us at Wallaroo in general that the

Speaker:

future is lots and lots of models being deployed in lots of

Speaker:

locations. And I think that one

Speaker:

big sort of industry wide theme that I'm seeing is if the

Speaker:

last 20 years, let's say, was the story of Everybody

Speaker:

picking up from their colos and moving to the cloud and centralizing

Speaker:

all of their IT, I think that the next 20 years are gonna be

Speaker:

Not like deconstructing the cloud. I think the clouds are here to stay and they're

Speaker:

gonna continue to grow, right, year over year. But there will be more

Speaker:

of a push out to more edge computing environments. Cell phones

Speaker:

are getting more and more powerful. Cars are getting more and more powerful. Like, there's

Speaker:

more computer stuff happening, all over the place, and the compute

Speaker:

available, the memory and the storage available is all through the roof compared to

Speaker:

what it was 20 years ago. And, I think we're

Speaker:

gonna see more push for smaller, more specific machine learning models, And

Speaker:

they're gonna be pushed out to all these edge locations so that they can run

Speaker:

close to where the data is. So you're not schlepping this sensitive data all over

Speaker:

the Internet and other people's networks. Yeah.

Speaker:

But, you know, you're taking advantage of of compute resources that you

Speaker:

have local to the data and making very fast decisions,

Speaker:

you know, very efficiently. So I I have to jump in

Speaker:

because, you you just made me feel really good.

Speaker:

About a year ago, I built a large server here

Speaker:

at home, which I hadn't done in a decade. Actually, my my

Speaker:

20 year old son built it. But he and he helped me with,

Speaker:

with picking out the new shiny fast parts, on it because I was

Speaker:

so out of practice with this such confessing.

Speaker:

But, and it's really cool to see, you know, all of his All of

Speaker:

his skills. He does edge. We just picked up the

Speaker:

Raspberry Pis are back in stock, finally. Yep. And I just picked up,

Speaker:

like, 3 for $35, You know, the 1 gig force.

Speaker:

Yep. Anyway, super excited about that. One of the things I built

Speaker:

at the time I built a box About a year ago, you

Speaker:

couldn't do a local GPT or anything close

Speaker:

to that. And I said, Eventually, we're

Speaker:

gonna be able to do this. I I made that guess, and it was a

Speaker:

guess. Yeah. But about 6 months later, about 6 months

Speaker:

ago, All of a sudden, I started seeing these 7,000,000,000

Speaker:

token machines showing up and it started clicking.

Speaker:

It was like, holy smokes, you can do this. I did make one stupid mistake

Speaker:

and he didn't catch me on it. I bought a 12 gig GPU

Speaker:

because that's super crazy huge From 10 years

Speaker:

ago. And that wasn't super crazy huge at all. No. No.

Speaker:

No. But it's interesting. Now they're back now. They can run on, You know, on

Speaker:

the 12 gigs. And like you said, you mentioned the CPU models. So I just

Speaker:

learned a ton as I've been going through this. And, That

Speaker:

it's it's very encouraging to hear that. I had not heard anybody

Speaker:

say edge and running small ML models on the edge.

Speaker:

That's, I mean, that's what we've been trying to do here. And I I love

Speaker:

the redundant you know, the idea of a redundant array of whatevers,

Speaker:

you know, MLs. It's almost like a swarm of MLs. I've heard,

Speaker:

yeah. Yeah. Yeah. That's true. Right? And, you know, I think there's a lot

Speaker:

of interesting stuff happening on the battlefields in Ukraine right now drones.

Speaker:

And Right. That Yeah. Was also a fascinating space and

Speaker:

very much, I think, heading in the direction of lots of ML running at the

Speaker:

edge. It's it's funny you mentioned that. So I live in a DC area,

Speaker:

and, I was at a government tech

Speaker:

symposium about 2, 3 weeks ago now. And

Speaker:

they were talking about that that, you know, edge is gonna be much more important

Speaker:

in the future of warfare. And he said presumably

Speaker:

elsewhere too. Right? He was permanent primarily a government in defense. It was definitely a

Speaker:

military industrial complex, type of type of event. But he was

Speaker:

explaining like, you know, in the past, you know, 20 years,

Speaker:

we've not dealt with adversaries. We've

Speaker:

only dealt with adversaries in in battle space conditions

Speaker:

where it was, you know, we controlled the airwaves.

Speaker:

Mhmm. And he, I think he used an interesting term. We

Speaker:

had airspace and electromagnetic electromagnetic

Speaker:

dominance. I was also like, Wow. Yeah. That was yeah. Yeah. I was, like, oh,

Speaker:

that's interesting. So, like, the whole idea of these disconnected

Speaker:

decentralized networks, I mean, I think

Speaker:

you're I think you're spot on. It's the future for

Speaker:

geopolitical reasons, but also just for, you know,

Speaker:

Privacy and just kind of flexibility reasons. Yeah. The

Speaker:

question I have though is, like,

Speaker:

Organizations can barely manage the infrastructure they have now and barely manage

Speaker:

the software they have now. What are they gonna do when the software starts Not

Speaker:

thinking for itself, but, like, this becomes another workload Yeah. On

Speaker:

top of that. Like, what Well, for one thing, that's why Wallaroo It

Speaker:

is focused where we are, and we're trying to build this platform to help people,

Speaker:

you know, with this capability of being able to deploy models and manage a fleet

Speaker:

of them at the edge. Because, yeah, there aren't a lot of good

Speaker:

solutions for that today. Yeah. Interesting. I I think the

Speaker:

general answer to your question is probably some combination of cloud and edge.

Speaker:

You know, like, it does make sense to centralize a lot of things, and it

Speaker:

makes the the maintenance easier and, more efficient. And

Speaker:

You can get some economies of scale and, you know, all that kind of stuff.

Speaker:

But, we are gonna have to get good at managing a bunch of,

Speaker:

disparate types of things in desperate locations. I think all of

Speaker:

us. Interesting.

Speaker:

So this is the part of the show where we'll switch over to

Speaker:

The premade questions, and for your convenience,

Speaker:

I will, paste that in here.

Speaker:

Hopefully, paste it. And there we go.

Speaker:

So You had an interesting career looking at LinkedIn. You were at

Speaker:

SendGrid. You were then you were at DataRobot, and you said you made a switch

Speaker:

into the the data world, which begs the question, How did you

Speaker:

find your way into data? Did data find you or did you

Speaker:

find your way to data? I I

Speaker:

guess that is a good question. I think that, it was probably a

Speaker:

little bit of both.

Speaker:

Finding my way to data, I think that the beginning of the story is probably

Speaker:

at SendGrid. And I joined SendGrid as a DevOps engineer.

Speaker:

And to be honest, I had not really heard of SendGrid at the time. I

Speaker:

knew a little bit about it, but it, you know, I didn't really understand what

Speaker:

it was, too much with the scale. SendGrid, by the way, is now owned by

Speaker:

Twilio. But they have an API for sending email, and

Speaker:

they make it just really easy to integrate with, websites and applications

Speaker:

and and software so you don't have to worry about SMTP and, you know,

Speaker:

DKIM signing and all the other, like, gnarly bits of of

Speaker:

email. Turns out that Sengrid had a

Speaker:

ton of data. They're handling billions of emails a day,

Speaker:

and, you know, there's a lot of metadata there. The the actual data of the

Speaker:

email and and so on, the recipients and who to send it to and all

Speaker:

that stuff. And so working in that space,

Speaker:

I was dealing with tons and tons and tons of data. I mean, we

Speaker:

had, we were using mostly MySQL, and we had these

Speaker:

massive massive clusters. I think we had,

Speaker:

like, 30 or 40, you know, schemas under management. Each

Speaker:

schema was a cluster of anywhere from, Like, 6

Speaker:

to 40 plus servers, Wow.

Speaker:

You know, with lots of compute and everything else. So that was probably my

Speaker:

1st foray into, like, really thinking about data as a first class

Speaker:

citizen. And, and even to the extent of, like,

Speaker:

You know, building an architecture around the data. Right? So

Speaker:

that you can optimize the flow of the data, and being able to store it

Speaker:

and process it and transmit it fast enough to keep up with, with the

Speaker:

flow. And so, yeah, from there,

Speaker:

you know, had a lot of fun, learned a lot of things about, startups

Speaker:

about industry, about, DevOps and and all kinds of

Speaker:

things. Management as well and leadership because that's where I first,

Speaker:

started managing teams And then moved to data robot and,

Speaker:

into the ML space. And then it was a whole another learning journey

Speaker:

about, you know, data,

Speaker:

engineering, feature engineering, transformation tools. How do you

Speaker:

curate your data? And how do you really, like, know what you

Speaker:

have and inventory it and, make it available

Speaker:

to people within the business so that they can get value out of it.

Speaker:

Interesting. Very much. So our next question

Speaker:

is what's your favorite part of your current gig?

Speaker:

I think it's actually, I'm gonna cheat and I'm gonna say I have 2 favorite

Speaker:

things. And I I kind of always have I I

Speaker:

Figured out this formula a while back, in terms of what

Speaker:

motivates me. And it's one part the people that I work with

Speaker:

and another part, the problems that I had yet to solve.

Speaker:

So I wanna work with smart people. I I really don't like being like, feeling

Speaker:

like the smartest person in the room. I much prefer to surround myself

Speaker:

with people that are smarter than me and I respect and I can learn

Speaker:

from. But that also, you know, I enjoy. Right?

Speaker:

We spend a lot of time at work, so it helps to to enjoy the

Speaker:

people that you're working with. True. So that's a big part of it. And

Speaker:

then, finding tough problems, hard challenges. You know, if I

Speaker:

don't have hard challenges to keep me, to keep my mind

Speaker:

engaged and occupied, I start to get bored and, that's no fun. I

Speaker:

prefer to to always have something new to to to, you know, be chewing

Speaker:

at. So, yeah, good people, smart people,

Speaker:

and hard challenges. That is that is really awesome. I feel the

Speaker:

same way about about both of those things. The, for me though, I

Speaker:

I, Trying to find people that are smarter than me is

Speaker:

really easy. So I I enjoy that part a

Speaker:

lot. Like Frank. Frank is smarter than me.

Speaker:

Well, thank you. So

Speaker:

we have a couple of, complete this sentence, questions.

Speaker:

The first one is, when I'm not working, I enjoy

Speaker:

blank. When I'm not working, I enjoy

Speaker:

reading. I Enjoy movies. I go biking sometimes.

Speaker:

That's part enjoyment, part exercise. You know, it's good for me, but,

Speaker:

There's a lot of good, road biking in particular around Denver and a lot of

Speaker:

beautiful scenery. So you can, you know, just ride for a while and find yourself

Speaker:

up in the mountains or something, which is great. Yeah.

Speaker:

Traveling, cooking, all these things are good.

Speaker:

Our next fill in the blank is I think the coolest thing about

Speaker:

technology today is blank.

Speaker:

I I don't think it's necessarily something about today, but I think the coolest thing

Speaker:

about technology is how it builds on itself. I remember

Speaker:

Years years ago, I was studying for the CCNA exam, and

Speaker:

that was such a formative moment for me to

Speaker:

suddenly understand How networks worked all the way

Speaker:

from the physical, you know, sending

Speaker:

electricity down a copper wire, and it can be on or it can be off.

Speaker:

And that's it. Right? And you can do that really, really fast. Switch from on

Speaker:

to off, on to off, on to off, all the way up to,

Speaker:

web 2.0 and and Ajax and, you know, Asynchronous

Speaker:

JavaScript stuff happening in Google Maps. Right? And I can just drag my map

Speaker:

around. It's just mind blowing. And, honestly,

Speaker:

like, that That journey from the zeros and the

Speaker:

ones up to Google Maps, that was, you

Speaker:

know, what, 50, 60 years of,

Speaker:

technology building on itself of people solving very small simple

Speaker:

problems, but you add up all those small simple solutions and you get

Speaker:

something incredibly complex And absolutely mind blowing.

Speaker:

Excellent. Very interesting. The last, the 3rd and

Speaker:

final, Complete the sentence. I look forward to the

Speaker:

day when I can use technology to do blank.

Speaker:

I I would love, a Personal assistant, you know, like

Speaker:

Jarvis from from Marvel Comics or something or, I don't know,

Speaker:

from I I'm big into sci fi and and things like that when I read.

Speaker:

So, there are plenty of examples, but some kind of a smart personal

Speaker:

assistant that, you know, I can chat with and it keeps track of my calendar

Speaker:

and reminds me of appointments and, you know, when to call

Speaker:

my dad and whatever else, stuff like that. I just think that's

Speaker:

so cool. And I don't you know, with Especially with all the new LLMs

Speaker:

and and GPT stuff that's happening, I don't think we're super far from that. So

Speaker:

it's kind of exciting to me. No. You're right. Like, I

Speaker:

you know, if you watched, you know, when I was a kid, Star Trek next

Speaker:

generation was on, And the way that they were able to interact with the

Speaker:

computer just through their voice. Yep. And I mean, the 1st Star

Speaker:

Trek show had that too, but, like, the way the conversations I thought were more

Speaker:

richer and more kinda interactive. Mhmm. Mhmm. We

Speaker:

have a lot of that now. Yeah. I think some of the fundamental pieces are

Speaker:

in place now. Yeah. It'll probably take a little while to put

Speaker:

them all together and make it work right. But yeah. Agreed.

Speaker:

So our next one is, share something different about yourself.

Speaker:

But we, always remind guests that we're trying to keep our clean

Speaker:

rating. Yeah. On Itunes. So

Speaker:

I don't know. I think one of the more interesting things about my

Speaker:

Journey is that I don't have a background, like a a degree

Speaker:

in anything technical. I went to college and I got

Speaker:

my undergrad Studying Greek and Latin and classics. And

Speaker:

so it was mostly history, archaeology, languages, and things like

Speaker:

that. And Computers have always been a hobby of mine and and I

Speaker:

definitely did some computer science stuff in high school. I took 1 or 2 classes

Speaker:

in college, but I didn't really make my way into that

Speaker:

Professionally until a few years after college.

Speaker:

And, you know, honestly, I I don't think it's hurt me at

Speaker:

all. And in many ways, I think it's helped me partly

Speaker:

because, you know, it it helps a lot with management and leadership, just

Speaker:

to To kind of have a broad background and and understand, you know,

Speaker:

different people and perspectives and and where they might be coming from.

Speaker:

And I'm sure that some of the languages, you know, studying languages helped me

Speaker:

picking up computer languages as well. I think there are a lot of similarities in

Speaker:

In, human languages and and computer, you know, programming languages. But

Speaker:

What? Yeah. But, yeah, it is somewhat unique, and I don't run

Speaker:

into too many other classics majors, At, you know, tech startups.

Speaker:

I could definitely see the convergence, especially now when we're talking about

Speaker:

LLMs and the like. Right. You know, the the

Speaker:

nearest neighbor algorithms and all of that that are that are being applied

Speaker:

because my understanding is that's that's, You know, that's how that

Speaker:

works as it picks the next best word Right. You know, in a in a

Speaker:

sentence. And so syntax and grammar and all of the things you

Speaker:

studied in-depth, That should be very helpful.

Speaker:

Yeah. No. That that's awesome. There

Speaker:

is that good value in,

Speaker:

like a classics education. I I went to Jesuit

Speaker:

High School and Jesuit College, you know. Mhmm. I was forced into studying Latin

Speaker:

and things like that, like, didn't do it voluntarily. I'm not gonna

Speaker:

admit that, not do that. But but like as I get older, like, it's

Speaker:

definitely like, Oh, I get this. Like, you

Speaker:

know, especially when dealing with a lot of lawyers, there's a lot of Latin in

Speaker:

that. And so I'll hear them, like, you know, Excuse some

Speaker:

words. I'm like, I think I know what that means. Yeah.

Speaker:

Audible sponsors data driven.

Speaker:

And you mentioned you read a lot. Do you do audiobooks and

Speaker:

sci fi? Do you have any recommendations? Yeah.

Speaker:

There was a really good book that I read recently. Like, this is maybe

Speaker:

a year ago or something, but, best book I've read recently.

Speaker:

It's, The title of the book is called Seeing Like a State,

Speaker:

and it's by, James c Scott. The the longer

Speaker:

subtitle is, something like how Some

Speaker:

schemes to improve the human condition have failed or something like that. But,

Speaker:

it talks about this concept of legibility and how a lot of

Speaker:

The developments over the course of the enlightenment, the industrial revolution,

Speaker:

and, in the last few 100 years in in

Speaker:

Our society have been primarily

Speaker:

driven by the centralization of power in states

Speaker:

And the state needing to administer all of these people,

Speaker:

taxes, lands, land ownership, and all these different things.

Speaker:

And, you know, as part of, like, the the enlightenment, the

Speaker:

scientific revolution, we all got very enamored with, like,

Speaker:

rational thought and Logic and and all of this stuff. And

Speaker:

we thought, we're understanding the principles of the universe. We can predict

Speaker:

the motions of the planets and all these things. Well, we can solve all these

Speaker:

problems about, you know, around human civilization and humans as well.

Speaker:

And in a lot of cases, it failed. Right? And we didn't know as much

Speaker:

as we thought we did. And one of the sort of basic,

Speaker:

like, premises of the book, I guess, or arguments that it's trying to make is

Speaker:

that we routinely Underestimate, the

Speaker:

complexity of the natural world and how necessary it is.

Speaker:

And we think we can Simplify things and strip out all these

Speaker:

variables and go, you know, monocultures in our in our agriculture,

Speaker:

for example, and do industrial scale agriculture. You need

Speaker:

timber for building ships. Great. We'll just plant Norwegian pines in straight

Speaker:

rows. This is gonna be great. It's so predictable. We know exactly what,

Speaker:

You know, an acre of that will yield after 10 years.

Speaker:

But it turns out you can't strip out all the variables because the whole thing

Speaker:

falls apart. You need the complexity of the ecosystem to keep all those trees

Speaker:

healthy. And so all that predictability you thought you had

Speaker:

disappears within a couple of generations because, it can't

Speaker:

sustain itself. Wow. So, anyway, it it's a very, like,

Speaker:

complicated book. I'm not really doing it justice,

Speaker:

but I definitely recommend it. Interesting. It's on Audible.

Speaker:

Yeah, yeah, so definitely check it out.

Speaker:

The show. So if you go to the date is ribbon book.com,

Speaker:

you'll be routed to an Audible page. And if you choose to get a subscription,

Speaker:

to Audible. You will give us

Speaker:

you'll get a free book, and then we'll get like a little bit of a

Speaker:

bump on the head, and pat on the back, and Probably enough to

Speaker:

buy a cup of coffee. It started Which will share. Which will

Speaker:

share. Yes. Yes. And the final question,

Speaker:

Where can people learn more about you and Wallaroo? And they even

Speaker:

made that rhyme. Yeah. Great. I

Speaker:

think the best place to go is the Wallaroo website, which, as

Speaker:

Andy mentioned earlier is wallaroo.ai. So wallar00.ai.

Speaker:

And we've got a ton of great stuff on there. Lots of, you know,

Speaker:

documentation and and white papers and, tutorials and things about the

Speaker:

product and what we're doing there. And for myself,

Speaker:

I'm on LinkedIn. That's probably the easiest place to find me, Chris McDermott.

Speaker:

And, I think I even have that as my, like, LinkedIn

Speaker:

Profile name or whatever sits in the, you know, in the URL.

Speaker:

Cool. It is, actually c s m

Speaker:

c s McDermott. Okay. Well, thank you. Close. I was just looking at

Speaker:

it, and I was also looking at the website. It's a very nice website. Thank

Speaker:

you. Great design. And, although I can't design

Speaker:

great websites, when I look at one, I can tell whether it's great or

Speaker:

not. Me too. Me too. Same boat. I can't do it myself, but I definitely

Speaker:

appreciate it. I I can't cook, but I appreciate a good meal. There

Speaker:

we go. Yeah. That's it. And with that, we'll let

Speaker:

Bailey finish the show. Thanks, Frank and

Speaker:

Andy. And thank you, Chris, for putting up with our broken

Speaker:

calendaring system. Satya should really look into that

Speaker:

now that the drama around open a I is over.

Speaker:

Well, over for now at least. Maybe g p t

Speaker:

five can fix it.

Links

Chapters

Video

More from YouTube