Artwork for podcast Data Driven
Microsoft Fabric Unpacked: AI, Data Sovereignty, and a Bit of Clippy Nostalgia
Episode 1612th January 2026 • Data Driven • Data Driven
00:00:00 00:54:46

Share Episode

Shownotes

In today’s show, BAILeY, your semi-sentient hostess with the mostest metadata, teams up with Frank La Vigne to welcome the ever-insightful Andrew Brust for a deep dive into the evolving Microsoft data ecosystem. From nostalgic tales of Windows history and scoring elusive Clippy swag at Ignite, to unraveling what makes Microsoft Fabric a game-changer for data integration, AI, and governance, this episode covers it all.

You’ll hear firsthand how Microsoft’s innovation goes far beyond the tech itself—focusing on seamless integration, unified billing, and organizational synergy. Andrew Brust sheds light on the journey from fragmented Azure services to the unified vision of Fabric, the rise of generative AI and “agentic” intelligence, and the increasingly important role of data sovereignty and governance in today’s regulatory landscape.

Whether you’re a data enthusiast, an AI tinkerer, or just in it for the nostalgia, grab your headphones and get ready for insights, laughs, and more acronyms than you can shake a dataset at. Stay curious and caffeinated—this episode has something for everyone!

Time Stamps

00:00 Microsoft Expertise and Industry Analysis

03:18 "Big Data and Analytics Insights"

08:31 Power BI's Rise in Azure

12:26 "Unified Fabric-Based Data Platform"

14:01 "Fabric IQ Powers AI Integration"

17:24 "Achieving Synergy Against Odds"

23:32 "Unified Compute for Seamless Integration"

24:41 "Overwhelmed by AWS Complexity"

28:21 "Microsoft Research Powers Azure Fabric"

34:20 "Azure Foundry and Tools"

37:33 "Flexibility Beyond Major Cloud Providers"

41:33 Global Data Privacy Trends

45:13 "Governance for Agentic AI"

47:29 "Azure Stack and Local Clouds"

50:13 "Kubernetes: The Cloud Caveat"

53:40 "Let's Reconnect and Reminisce"


Transcripts

Speaker:

Ah, episode 401. Proof that we're still going strong.

Speaker:

Like a SQL server instance running on pure spite and caffeine.

Speaker:

I'm Bailey, your semisentient hostess with the mostest

Speaker:

metadata. Today, Frank's joined by the ever insightful

Speaker:

Andrew Brust to talk fabric AI, Microsoft

Speaker:

nostalgia, and why even Red Hat folks can still love Clippy.

Speaker:

Grab your headphones and your compute capacity. Let's dive in.

Speaker:

Hello, and welcome back to D Data Driven, the podcast. We explore the emerging

Speaker:

industry of data science, data engineering, and

Speaker:

artificial intelligence. With me today is not Andy

Speaker:

Leonard, who's my favorite data engineer in the world. However, I do have Andrew Brust,

Speaker:

who is an og, so to speak, in the.

Speaker:

In the AI and Microsoft ecosystem. And

Speaker:

although I think a lot of people think that I've abandoned the Microsoft

Speaker:

ecosystem, I have not. I've just had other. Other things kind of preoccupy my

Speaker:

time. And you know how it is. You have kids and, you know, they demand

Speaker:

attention. You have a house and all that and good problems to

Speaker:

have. But I'm very glad to kind of have someone I know walk me

Speaker:

back into kind of the Microsoft ecosystem, because a lot has

Speaker:

changed since I left Microsoft. I did

Speaker:

go to Microsoft Ignite. We were talking about that. I even scored myself a

Speaker:

Clippy. Clippy plus.

Speaker:

I'll have to tell you how I want him. So they had like, a little

Speaker:

challenge of like, do you know Windows history? Because Windows, I guess, turned 40 this

Speaker:

year and. I know, right? Yeah,

Speaker:

85. That's right. Yep. And they were like, do the history of

Speaker:

Windows. And I'm like, I'm like. And I had some Red Hat people and like,

Speaker:

you know, I was. I would have been very embarrassed if I had gotten anything

Speaker:

wrong. Turns out I got it right, actually. Good. I

Speaker:

remember back then you could install just a runtime version of Windows if you

Speaker:

wanted to run specific Windows apps on your. On your DOS

Speaker:

machine. Yeah. Now we take for

Speaker:

granted kids. They don't understand, like, no, no, you had to type in win or

Speaker:

win 3.1 if you were fancy and you're running multiple

Speaker:

versions of Windows at the. Same time,

Speaker:

you say kids today. I mean, kids 20 years ago didn't understand

Speaker:

that. This is true. This is true. But

Speaker:

anyway, in terms of bringing you back into the Microsoft orbit,

Speaker:

well, first of all, I'm sure Ignite did a bunch of that. But I can

Speaker:

be gentle because even though I am to this day a MicroSO

Speaker:

Regional Director, or as I like to say, a member of the regional Director

Speaker:

program, because Otherwise it sounds like I work for Microsoft

Speaker:

and an MVP, a data platform MVP, both over

Speaker:

20 years now. I'm an

Speaker:

industry analyst and so I look at data and analytics

Speaker:

solutions across the board, not just Microsoft specific. I

Speaker:

will say Microsoft is my sweet spot in terms of what it is that I

Speaker:

know and where I have the most history.

Speaker:

But I work with lots of other companies you've heard of, like

Speaker:

Databricks and Snowflake and Cloudera and

Speaker:

plenty of others. So my team and I do

Speaker:

most of the reports for a research company called GigaOM

Speaker:

with which I've had a very long association. Most of the reports. Sorry, I

Speaker:

didn't finish that sentence. That are focused on data or analytics

Speaker:

are things that we work on. So whether it be data warehouses or lake

Speaker:

houses or streaming data platforms or data access

Speaker:

governance or data catalogs or blah blah, blah, they all have the

Speaker:

word data in them. We work on those reports.

Speaker:

We create what are called gigaom radar

Speaker:

reports, which are a little bit like the analog to

Speaker:

Gartner's Magic quadrant in terms of looking across a category

Speaker:

at a bunch of vendor solutions and rating them on

Speaker:

multiple criteria which change each year.

Speaker:

So. And when I started covering big data,

Speaker:

because the way this got started was I was the first and only person

Speaker:

at ZDNet to be covering big data.

Speaker:

So that was an amazing. That's a term you don't hear a lot. You don't

Speaker:

hear it anymore. No. And in fact I was at the

Speaker:

older incarnation of gigaom. I was a full time employee there.

Speaker:

I was their research director and they wanted, so call me research

Speaker:

director for big data. And I said, can we just call it data

Speaker:

or data and analytics? And we did. Because I was like, you know,

Speaker:

eventually what we think is big now won't look so big.

Speaker:

Have you heard my Costco rules? Say again? I have something

Speaker:

called the Costco rule. The Costco rule. If you go to call,

Speaker:

if you walk into Costco by hard drive, that size. That size is no longer

Speaker:

big data. Fair, fair.

Speaker:

Anyway, that put me in an immersion. And at that time, Microsoft

Speaker:

was not in that world at all.

Speaker:

Eventually the thing called hdinsight got to beta

Speaker:

and it wasn't even called hdinsight when it was in beta.

Speaker:

And so Microsoft started coming back into my world. Eventually,

Speaker:

of course, it came back full swing. And with Microsoft fabric,

Speaker:

now it's doubly full swing, which I think is

Speaker:

very, very good, both for Microsoft and the industry. But

Speaker:

what was I going to say? Just that. Yeah. So

Speaker:

finally the two things that Were kind of orthogonal. Now have an

Speaker:

intersection. Right. And that

Speaker:

intersection is my sweet spot as I'm still a data platform

Speaker:

mvp and I have a very long history with Microsoft's

Speaker:

business intelligence stack. I was on Microsoft's

Speaker:

partner advisory council going way back, like

Speaker:

from 2005 to roughly 2010.

Speaker:

I don't know. I saw Power BI when it was still a bunch of wireframes

Speaker:

in a PowerPoint slide deck. So I've been through

Speaker:

many rounds of being frustrated that Microsoft

Speaker:

didn't have a good competitive play. And I'm now pretty

Speaker:

satisfied that they have one that's very competitive. So we can talk about that or

Speaker:

we can talk about the greater world. And as far as

Speaker:

AI goes, I was interested in AI all the way back in the horse and

Speaker:

buggy days when I was an undergraduate. Oh, really?

Speaker:

Yeah. AI was very different then. It was about like, weird programming

Speaker:

languages like Lisp and Prologue and

Speaker:

Expert Systems and things of that elk.

Speaker:

But neural nets existed then, and neural nets are the very

Speaker:

basis for the large language models we have today. So it's not, it's

Speaker:

not completely unrel it, but obviously it's very different.

Speaker:

I took a prologue course, which was the. We had one offering.

Speaker:

So I'm. You and I are both from New York City. So, like, you know,

Speaker:

we probably accidentally crossed paths more than once.

Speaker:

And I know we crossed paths during the early Power BI days

Speaker:

because I think the company I worked for at the time

Speaker:

was also an early believer in power bi.

Speaker:

So this is what meant 2005. And then they hired. We had a guy who

Speaker:

was the practice manager, Kevin, who got hired. Kevin Viers got hired into

Speaker:

Microsoft. So I think he's still there, actually. He,

Speaker:

he hired me back into Microsoft when I rejoined in 2018, which was kind of,

Speaker:

okay, what a small world it is, you know, and, and, and for us, like,

Speaker:

you know, something that my parents would always say, like 20 years could go by

Speaker:

in a blink once you hit a certain age. And I'm like, good Lord, was

Speaker:

that the truest thing they ever said, right? 20.

Speaker:

Yeah. The scale shrinks the older. The older you get. It's a

Speaker:

little. It's a little frightening. I've turned it like, into a roll of toilet

Speaker:

paper that as you get closer and closer to the end, it spins out a

Speaker:

lot faster because the diameter gets smaller. Oh,

Speaker:

Lord, that is a very scary concept.

Speaker:

But you're right. I remember one of the things. These were back in the

Speaker:

cubicle days when people worked in an office and things like that. And I remember

Speaker:

sitting next to Kevin. And Kevin would be on the phone like, yeah, I know

Speaker:

it's weird to hear Microsoft is kind of a small player in any niche,

Speaker:

but in BI and business intelligence they really were.

Speaker:

And it was just kind of like, yeah, that's true. You don't really think about

Speaker:

them as a small player, but at the time now it's kind of ridiculous

Speaker:

to say that in data and analytics, right? And the Power BI team has just

Speaker:

done phenomenal in terms of their speed to market. And what they

Speaker:

built out is phenomenal. It's unreal.

Speaker:

It's the Power BI team that kind of took over the

Speaker:

entire Azure data group, right?

Speaker:

And that included SQL Server. So

Speaker:

whereas the BI team was once a little corner of the

Speaker:

SQL Server world that

Speaker:

initially came to Microsoft through an acquisition of assets from

Speaker:

an Israeli company called Panorama.

Speaker:

And Amir Nats is the distinguished engineer who actually came

Speaker:

from Panorama and is very much like the father of Power BI

Speaker:

and of fabric. Still there.

Speaker:

Slowly but surely, not

Speaker:

only did they get BI right and they always had it right on the

Speaker:

server, they just never really had it right on the front end

Speaker:

until the current version of Power BI that we have now kind of gelled,

Speaker:

but they also ended up kind of mastering

Speaker:

the software as a service approach to

Speaker:

cloud services. And they took a look at the Azure

Speaker:

data stack and said, we have tons of capabilities here, but they're kind

Speaker:

of fragmented over several different products,

Speaker:

each of which have their own kind of procurement model and

Speaker:

pricing model. And that gets very hard to manage.

Speaker:

And if you look really carefully at fabric, while there are

Speaker:

some things that are truly native to it,

Speaker:

most of the parts of it are Azure services in the

Speaker:

background that have been integrated and

Speaker:

that have been unified in terms of how you pay for them,

Speaker:

I don't know. Microsoft needed that, by the way, the whole cloud industry

Speaker:

needed that. Because Google and Amazon are just as guilty of

Speaker:

having a whole sprawl of

Speaker:

services without unified user

Speaker:

interfaces or APIs or pricing. No, that's

Speaker:

true. I mean, when I. So when I just before

Speaker:

the pandemic, I was out at Tech ready, which is

Speaker:

an internal Microsoft event. And they were basically

Speaker:

might have been Amir, actually, now that I think about it, was presenting on

Speaker:

the future of what was called synapse. And this is kind of,

Speaker:

you know, he's like, you know, everything's going to be all in one pane of

Speaker:

glass. Everything, basically everything you said when all these things are going to be thing

Speaker:

and you know, and the speaker, which I don't.

Speaker:

Can't say it was him but would make a lot of sense. He said like

Speaker:

this is the future. We're going to get everything under one pane of glass billing.

Speaker:

Don't worry about that, we're going to get that figured out in time. And I

Speaker:

was just like, you know, I kind of saw the vision. So

Speaker:

and then I don't know, maybe like a year, year and a half later I

Speaker:

left Microsoft and then Fabric came out and I

Speaker:

was wondered like why the change in name like

Speaker:

Synapses? And I asked people like well SYNAPSE is still kind of there but it's

Speaker:

really Fabric is where everything's going. I'm like all right, but like what is the

Speaker:

difference per se? So if we pretend I was on a ufo, well that's

Speaker:

a weird thing. Pretend I was in a coma and I just Woke up

Speaker:

from 2000 2021. What?

Speaker:

And I say like well what happened to Synapse?

Speaker:

Sure. So I mean the functionality of SYNAPSE is still there and there was

Speaker:

a lot of, I won't call them conspiracy theories but

Speaker:

skepticism when Fabric came out that it was really just a

Speaker:

rebrand of synapse. In fact that's not what it

Speaker:

is. So the, the thing that was originally SQL Data

Speaker:

Warehouse which was in Synapse as so called

Speaker:

dedicated pools and the more lake housing part

Speaker:

of it that was in there

Speaker:

as gosh, I forget the old nomenclature, it wasn't on demand

Speaker:

pools but it was something of that.

Speaker:

Reserved instances or something like that. Wasn't reserved, no,

Speaker:

but anyway basically a Spark based data

Speaker:

lakehouse using

Speaker:

Azure data lake storage as the storage layer that's still

Speaker:

there but what was I going to

Speaker:

say? But Fabric is a ton more because it integrates

Speaker:

all this, all this streaming stuff

Speaker:

that's now called real time intelligence. It integrates data

Speaker:

science and by the way the data science is completely

Speaker:

unique to Fabric. It's not merely just

Speaker:

an embedding of Azure machine learning.

Speaker:

There's also power bi of course

Speaker:

now there are operational databases including

Speaker:

SQL Database meaning Azure, SQL

Speaker:

meaning SQL Server in the cloud.

Speaker:

A lot of other pieces that were ancillary are now all

Speaker:

included. There's user

Speaker:

interface that covers the whole

Speaker:

realm and again the billing is

Speaker:

unified. So you buy a compute capacity

Speaker:

and basically as you use the different services

Speaker:

they're all pulling from the same pool of compute.

Speaker:

So you know, you don't have to over provision for each

Speaker:

one of those services just to make sure you have enough

Speaker:

compute to, to satisfy it. And now we have this thing called

Speaker:

Fabric IQ which brings, which brings

Speaker:

generative and agentic AI into things.

Speaker:

Which is good because it was kind of funny when Fabric finally went

Speaker:

to general availability. That was really when

Speaker:

ChatGPT and Gen AI were like

Speaker:

making it big. So it looked like Microsoft finally got the data and

Speaker:

analytics stack set up just in time for people to have their attention

Speaker:

to, you know, diverted over to AI.

Speaker:

But now we have, you know, natural

Speaker:

language query is kind of just the beginning. We have

Speaker:

operational agents that can actually act on

Speaker:

things and can be all based and triggered

Speaker:

on streaming data.

Speaker:

And so if you think about Azure Event Grid,

Speaker:

if you think about Azure Data Explorer,

Speaker:

If you think about the data pipelines that Azure

Speaker:

offers, as I said,

Speaker:

the one standalone data warehouse side of things, and even

Speaker:

elements of hdinsight in terms of the lakehouse,

Speaker:

that's all in there. What's also nice is even though it's Azure data Lake

Speaker:

storage under the hood, you have this abstraction layer over it

Speaker:

called OneLake. OneLake is

Speaker:

in many ways easier to deal with because

Speaker:

you don't have to worry about accounts and containers and

Speaker:

sizing those and so forth. It's still compatible with

Speaker:

all the ADLs and Azure Blob storage APIs.

Speaker:

It also supports this notion of shortcuts, which is really just a

Speaker:

data virtualization technology.

Speaker:

So you can have a shortcut to data in other OneLake instances

Speaker:

or in ADLS proper,

Speaker:

or even in Amazon S3,

Speaker:

or even in Google Cloud storage or other databases.

Speaker:

And logically they'll all look like they're part of OneLake and you

Speaker:

can query them as such. That's impressive. That's

Speaker:

impressive. You can kind of. You really. The. The vision of get everything, get

Speaker:

everything under one pane of glass seems like. It'S come true

Speaker:

that to what I tell people, even though it sounds maybe a little

Speaker:

bit anticlimactic, is that the real

Speaker:

innovation in in fabric

Speaker:

isn't the tech per se. It's.

Speaker:

It's all the integration of the tech and the

Speaker:

abstraction layers over it that make it work together, the UI that makes it

Speaker:

work together. And there's an organizational. I mean,

Speaker:

there's a little inside baseball, but there's an organizational facet to it as

Speaker:

well. Because all these different products were different teams.

Speaker:

Yes. People don't realize that. Like, I haven't, I've been. You've been, you're an

Speaker:

rd, so you kind of know, you know how the sausage is made. I was

Speaker:

inside the firewall, then I saw the sausage was made. Like there are all these

Speaker:

little teams that range from

Speaker:

really good team team players to really not good team players.

Speaker:

I think that's this polite way as I could put it. And they getting them

Speaker:

all to row in the same boat or like row. In the same direction.

Speaker:

Yeah, they weren't doing that. That wasn't even necessarily

Speaker:

based on hostility. It was just that different people had different reporting structures

Speaker:

and different priorities and different incentives. What

Speaker:

worried me was that the vision of putting all this together was a great

Speaker:

idea, but the execution to me at the time

Speaker:

seemed like it would be next to impossible to get all these teams to kind

Speaker:

of work harmoniously and somehow they

Speaker:

did it. And like to me that's the, that's the absolute

Speaker:

greatest innovation. And now they've got

Speaker:

synergy instead of sort of internal

Speaker:

competition and, you know, from there

Speaker:

on, look out because, you know, whatever. I'm sure

Speaker:

there are internal disharmonies somewhere.

Speaker:

But I would say at the high level in general,

Speaker:

anywhere you have people. You'Re going to have that I'm going to see. I think

Speaker:

I still have, I think you mentioned this is my old, old laptop. You see

Speaker:

Azure Data Data Explorer. I have the sticker from that.

Speaker:

Oh, I see it. Yep. Sorry, I had to show that off.

Speaker:

No. And what's Azure Data Explorer? For people who don't know,

Speaker:

it ran under the codename of Kusto K U S T

Speaker:

O. And there's some disagreement

Speaker:

over whether that really references Jacques Cousteau C O U

Speaker:

S T A U or not. But it's a

Speaker:

fantastic super high performance

Speaker:

system for, for not just

Speaker:

for streaming data, but for time series data. Yeah, with. With

Speaker:

its own query language and its own ability to create

Speaker:

visualizations. Right. In the query language. So your results

Speaker:

come back as both tabular and visualized data

Speaker:

and it can handle huge volumes of data

Speaker:

in a single query. And

Speaker:

there's a lot of heritage in the Azure Data Explorer team that

Speaker:

started in the SQL Server analysis services world. So there's

Speaker:

a continuum there. And that product on its own,

Speaker:

especially being called Azure Data Explorer, which made it sound like

Speaker:

a tool. File explorer. Yeah.

Speaker:

When they told me the name, I don't know, I was not

Speaker:

reserved in saying I didn't think it was the best name,

Speaker:

but that product on its own was kind of a sleeper. It

Speaker:

wasn't really getting the, I don't

Speaker:

know the kudos that it deserved or the attention that it deserved. And

Speaker:

now that it's part of fabric now, it contributes

Speaker:

to all the cool things fabric can do. So if you see what

Speaker:

are called event houses in fabric, that's the

Speaker:

same technology as Azure Data Explorer. Interesting.

Speaker:

So correct me if I'm wrong, but I think the origin story of

Speaker:

Kusto and Kusto query language was that

Speaker:

the folks running Azure, like in the operations team, actually built it

Speaker:

to run through all the logs that they had. Because I remembered I was at

Speaker:

some super secret event and they brought in some people from

Speaker:

the field and they had us do hands on labs with it

Speaker:

and I'm like, I, I must have been checking my email or

Speaker:

whatever. I'm like, when can I get this to my customers? And they kind of

Speaker:

laughed. They're like, no, no, this is internal only. It's internal. Yeah, it began as

Speaker:

an internal thing. So I was just like, oh, like you need to make this

Speaker:

a product. Because I could think of 15 customers on top of my head that

Speaker:

would, that would eat this up. Yeah, I'm glad it finally saw a light of

Speaker:

day. If you go back to the real world outside of Microsoft

Speaker:

and you think of the likes of Splunk, for example. Yes, right.

Speaker:

It's in this, it's in the same space. Although their

Speaker:

initial marketing just said it was a big data tool which just

Speaker:

completely obfuscated what it did. But anyway, now,

Speaker:

now in, in combination with these things

Speaker:

called event streams, which can stream the data in,

Speaker:

basically based on Azure

Speaker:

Event Hub, you put it all together and

Speaker:

you have the ability to do a lot of work with

Speaker:

real time streaming data without really having to write much

Speaker:

code, if any code. Although it does have

Speaker:

its own query language called kql. There's also

Speaker:

a copilot that you can just work with in natural language that will

Speaker:

generate the KQL for you. Oh, very nice. And I love

Speaker:

generators because not only does it mean I don't have to write the query, but

Speaker:

it means I can learn the language and then write my own query

Speaker:

if I want to. Reverse engineering is how I

Speaker:

prefer to learn. So

Speaker:

that works out really well. Yeah, I

Speaker:

wanted to dive into fabric, but I wasn't really even sure where to start because,

Speaker:

so I heard, and again, a lot of this is I heard, but

Speaker:

the way that it's not attached to your Azure tenant. Is that

Speaker:

true? It's attached just like Office 365

Speaker:

or more important, Power BI. Right. It's imagine

Speaker:

Power BI premium instances, and it's the outgrowth of that.

Speaker:

Okay, that makes sense now because when somebody told me like, well, it's not tied

Speaker:

to your Azure tenant, you need a different tenant, I'm like, but

Speaker:

okay. But then somehow, I guess at some. Level it ties into SaaS, not

Speaker:

tasks. Right? So using Azure services in the

Speaker:

background, including even Azure OpenAI. But

Speaker:

you don't have to provision anything in Azure. It's doing that on your

Speaker:

behalf. So you don't need an Azure tenant at all.

Speaker:

So that actually makes it a lot easier if I were to manage, if I

Speaker:

had to manage it, right? There's a lot of things that, like, correct. I mean,

Speaker:

I, I love the fact you have. You go to.

Speaker:

Go to, you know, any of these services,

Speaker:

right, and they have basically this smorgasbord, this, this big buffet

Speaker:

of services you kind of pick and choose from. But at the end of the

Speaker:

day, like, how do you figure out, you know, what you pay

Speaker:

for, right? It becomes, like, really kind of nightmarish. Like, again, to

Speaker:

me, that's the innovation is that, yeah, all the stuff has been

Speaker:

brought together, put under one pricing model,

Speaker:

and you don't have to worry about all the moving parts

Speaker:

and all the different, all the different servers or instances

Speaker:

that might have to be provisioned and sized. That all goes away.

Speaker:

And again, everything is built out of a single

Speaker:

pool of compute.

Speaker:

It's not a perfect analogy, but I think of like the old days of

Speaker:

cell phones, when you got a certain number of minutes per month, but you could

Speaker:

roll them over. And it's not that you can do that with fabric.

Speaker:

I'm not saying you can roll over your compute from one month to the next,

Speaker:

but what you. What is fungible is how the

Speaker:

compute is used amongst the different subservices

Speaker:

of fabric so you don't have to provision a certain

Speaker:

amount of compute just for streaming or, or just for

Speaker:

AI or just for data. Lakehouse.

Speaker:

Because it's all from. When it's all from one

Speaker:

pool. All right, that makes a lot of sense now because, like, that was always

Speaker:

when I first got it. When I left Microsoft, I, you know, started experimenting

Speaker:

with aws and I was just like, I just want to create a

Speaker:

website. Why don't I need these, like, hundreds of different services underneath,

Speaker:

right? Like, why do I need. I understand why. I need identity, access, management, right?

Speaker:

That made sense to me. But. But like, when it came to Route 53 and

Speaker:

like all this crazy stuff, I'm like, I just want to spin up a stupid

Speaker:

website, right? This is not, let alone do anything complicated, right? Where you need to

Speaker:

have all these underlying things. Like SageMaker, right, has this whole thing and they tried

Speaker:

to abstract away all the underlying services. But even when you

Speaker:

kill. This is. The thing that really annoyed me was when I killed the

Speaker:

SageMaker instance, I was still getting like, you know, 20,

Speaker:

$30 a month, not a lot, but I was still getting that on my

Speaker:

bill and eventually I just closed the account because I'm like, I'll have to start

Speaker:

fresh again in the future because like I God only knows what,

Speaker:

what, what I've spent. And for people who are just learning and wanting

Speaker:

to get their skill sets up, Microsoft is pretty generous with

Speaker:

trial, trial

Speaker:

capacities as they call them. A capacity basically is a, you know,

Speaker:

a server or an instance. However,

Speaker:

the, if you want to do anything with the AI, you do need a

Speaker:

paid instance. But there are some pretty, there are some pretty affordable

Speaker:

ones. And this gets a little confusing.

Speaker:

If you provision the fabric instances

Speaker:

through Azure, again you don't have to,

Speaker:

that connection doesn't have to be there. But if you provision it through Azure,

Speaker:

you can pause and resume those instances.

Speaker:

Okay, so you do that a lot. You could be like, hey man, I'm

Speaker:

taking, I'm taking a week between Christmas and New Year's off, so pause it.

Speaker:

Totally. I brought up my own cheat

Speaker:

sheet in the background when no one was looking. But

Speaker:

so Azure Data Lake Storage, Azure Synapse, as you

Speaker:

mentioned, Azure Data Factory, Azure Event Hubs, Azure Data

Speaker:

Explorer, Elements of Azure Machine Learning

Speaker:

and Power BI all come together in fabric.

Speaker:

Interesting. So it's like one roof for which I think is

Speaker:

a brilliant strategy. Right. Because Microsoft's core strength in the data and

Speaker:

analytics space isn't necessarily having frontier models, isn't

Speaker:

necessarily having the, the cutting most cutting edge research.

Speaker:

Although I love just making it usable. Exactly. Making it usable

Speaker:

and turnkey. Right. Like, not that I don't love my folks in Microsoft Research.

Speaker:

Right. I know some of them. Listen, love you all, you guys have the best

Speaker:

conference in the world. But you know, but, but I

Speaker:

mean, but you're making it usable. Right. And I think that that's really

Speaker:

string and they have all these separate tools. I think that was really the challenge

Speaker:

right. When it was a shrink wrap company, you knew what you bought. But when

Speaker:

it became like a SaaS pass company, you kind of could just

Speaker:

a couple of clicks, you could provision stuff. So it eventually kind of

Speaker:

got too chaotic. Now I like the idea of them kind of bucketizing this or,

Speaker:

or rolling it up behind one service

Speaker:

where because it, it just like the AWS problem. Right. Like I

Speaker:

spun up SageMaker. Right. And did

Speaker:

you that I needed, I needed underlying storage. I needed this. I needed this. I

Speaker:

needed DNS, I needed that. I needed that to the point where look, I, I'm

Speaker:

okay spending X amount of dollars on learning

Speaker:

Sagemaker Right. But I wasn't okay with

Speaker:

when I turned off the instance. I'm still getting built. What am I getting built

Speaker:

on? That lack of transparency, intentional or not

Speaker:

on AWS's part, has left a bad taste in my

Speaker:

mouth, you know, for cloud services in general.

Speaker:

Sure. By the way, you mentioned Microsoft Research and a

Speaker:

couple of things. So when I listed all those Azure services, I forgot to say

Speaker:

Azure OpenAI. So add that to the list. But also,

Speaker:

although I said there's elements of Azure machine learning in there, the data science

Speaker:

workload in fabric is really mostly unique to fabric,

Speaker:

but it's based on technology that comes out of

Speaker:

Microsoft Research. So for example, there

Speaker:

was something called flaml F L a M L,

Speaker:

which is the fast library for automated machine learning

Speaker:

and tuning. And that's built in.

Speaker:

So are things like, like ML Flow, which is an

Speaker:

open source experiment management

Speaker:

platform that's built into a lot of commercial AI

Speaker:

platforms. So they didn't, they didn't just kind

Speaker:

of embed and put their own badge

Speaker:

on it. They built their own, their own ML

Speaker:

stuff from, from these open source components.

Speaker:

Right, Right. Well that's interesting because like the world of AI is

Speaker:

largely dominated by open source. Right?

Speaker:

Right. I mean, Sagemaker, I'll stop kicking the AWS

Speaker:

the curb in a minute. But like SageMaker is basically a wrapper of Jupyter

Speaker:

notebooks. Right. Azure ML, at least when I last used it, was largely

Speaker:

a wrapper around Jupyter notebooks. Right.

Speaker:

So a lot of the core technology here does tend to

Speaker:

lean towards open source, which from my own personal career development

Speaker:

point of view, and they're not paying me to say this, you know, one of

Speaker:

the things that led me to Red Hat, right. Was the idea that, you know,

Speaker:

this is largely a movement driven by open source. So, you

Speaker:

know, let's see what we could do here. Right. And not, not a commercial,

Speaker:

not a commercial, not a sermon. Just, just, just point it

Speaker:

out because I think it's interesting how quickly open source has taken over

Speaker:

the, the, the certainly the AI world. Right.

Speaker:

But I also, by the way, there's notebooks in fabric too. If that wasn't,

Speaker:

if that wasn't already implied or obvious and

Speaker:

they are based on Jupyter, but you don't see the Jupyter skin. Right? It's

Speaker:

all right. Yep. Well, I think it's really what's really

Speaker:

impactful about kind of the notebook interface once you get used to it. And it

Speaker:

is an adjustment for people who, like you and me, grew up with Visual Studio

Speaker:

and Dare I say interdev. Right. The

Speaker:

idea that you can code in a browser, right. And you know,

Speaker:

no local installs really required.

Speaker:

It's been very freeing, right, because you can spin up an environment,

Speaker:

you can, you know, spin up a heavy cluster,

Speaker:

leave it running. Right. Do its thing, you know, close the

Speaker:

laptop, go in the car, go in the train home, and you get on the

Speaker:

other side and you're like, oh, it's done. Right. Like, that's kind of nice, actually.

Speaker:

Right. And of late, I've been a big fan, an increasing

Speaker:

fan of kind of my own. Of. Of local AI. Like your own private AI.

Speaker:

Right. Which explains, you know, I bought a. It was. It was an early

Speaker:

Christmas gift, probably Father's Day birthday and anniversary gift too.

Speaker:

My own DGX Spark. So I have my own AI running locally. Right.

Speaker:

So it's not throwing shade at the cloud. It's just

Speaker:

when I run a job, I don't have to think about the costs. Right.

Speaker:

And that, that sort of freedom

Speaker:

from worry can be really important as you're

Speaker:

learning things because you just don't. You don't have to keep looking over

Speaker:

your shoulder at the, at the meter or mixing metaphors there.

Speaker:

But. Yeah. So are you using Llama models and such, or

Speaker:

what do you. What do you really. Yeah, I have Llama. The thing I've been

Speaker:

doing mostly the most is, is doing fine tuning

Speaker:

Loras for image makers

Speaker:

in the past. That takes about 90 minutes to maybe two hours on this

Speaker:

box, which would probably be

Speaker:

equivalent to about 90 minutes of Azure service

Speaker:

or a VM. Right. That's like

Speaker:

$150 a pop, I would say, for the type of machine

Speaker:

I want. So the idea, I could just spin that off and I have the

Speaker:

added benefit as it heats my office, but I don't really

Speaker:

have to. I don't have to think about

Speaker:

like, oh, God, you know, like, how much is that going to cost me in

Speaker:

cloud services and things like that. I do think that the.

Speaker:

There's. I have a lot of questions because even though I was at

Speaker:

Ignite, I honestly spent my entire time at the booth and kind of walk around

Speaker:

the expo floor. I didn't have. I only had a hall pass. Right. So.

Speaker:

But one of the things I heard mentioned was Azure

Speaker:

AI Foundry. What is that? Because you also mentioned

Speaker:

Azure OpenAI, which I know the relationship between OpenAI

Speaker:

and Microsoft has not been as cozy as it once was.

Speaker:

And they've also, they've also added

Speaker:

Claude, the anthropic models too. Right. So

Speaker:

what is Azure OpenAI and how does that relate to Azure AI foundry.

Speaker:

Right. So OpenAI is just the

Speaker:

Azure. OpenAI is Azure's hosted

Speaker:

instance of the open OpenAI models and

Speaker:

APIs, because you can

Speaker:

procure those directly from OpenAI themselves.

Speaker:

Right. Or you can use them on Azure. And even

Speaker:

if the direct model is still running on Azure in the

Speaker:

background, it's still a difference in terms of procurement

Speaker:

and billing and so forth. So

Speaker:

you've got all the APIs around that, you've got the models. And then of course

Speaker:

you need tooling to do,

Speaker:

to do rag applications. Right. So

Speaker:

there was tooling for that, There was also tooling for building

Speaker:

copilots. There was Copilot Studio. Right. And these things

Speaker:

are all kind of coming together in Azure Foundry.

Speaker:

Yeah. So it's, you know, you worked at Microsoft, so you

Speaker:

know how this works where like different teams do different things with different

Speaker:

brands and eventually they may get kind of rationalized

Speaker:

together. So the

Speaker:

Foundry side of things helps there. And by

Speaker:

the way, I'm glad you mentioned Foundry because in addition

Speaker:

to this thing called Fabric iq, which we haven't really

Speaker:

talked about, there's also Foundry

Speaker:

IQ and there's also Work iq.

Speaker:

Sounds like IQ is the new copilot buzzword.

Speaker:

Well, it's the agentic buzzword. And the

Speaker:

idea, if you think about kind of all the promise of Microsoft

Speaker:

graph In the office M365 World,

Speaker:

work IQ kind of sits

Speaker:

over that realm. But you don't have to worry about the

Speaker:

graph APIs directly. Then

Speaker:

Fabric IQ sits over everything in the fabric world

Speaker:

based on a pretty rich

Speaker:

semantic model that can be developed so that

Speaker:

when you are querying your data in natural language, the actual,

Speaker:

the actual vocabulary or jargon in your particular

Speaker:

organization is well understood, including

Speaker:

the entities and the relationships between those entities.

Speaker:

And then Foundry IQ is a way for building agents at

Speaker:

the higher level that can actually talk to your structured data

Speaker:

via Fabric IQ and your

Speaker:

work and organizational related data via work iq.

Speaker:

And that, that was, I guess

Speaker:

the big vision at Ignite this year

Speaker:

was to talk about all those IQ pieces.

Speaker:

So there you go. No, that's interesting.

Speaker:

So one of the things that comes up, I'm sorry, I cut you off,

Speaker:

we're recording this the day after Christmas, so things are a little.

Speaker:

Still recovering from the manicness of yesterday. And

Speaker:

what's Microsoft's plans for? Kind of. Because one of the things you're seeing happen a

Speaker:

lot more is this notion of sovereign

Speaker:

AI or data sovereignty. And kind of like people are very

Speaker:

much more conscious about the value of their data. And

Speaker:

I actually think that given Microsoft as opposed to AWS or

Speaker:

Google, Microsoft does have a history of selling shrink

Speaker:

wrap software, right? So I do think Microsoft has a unique

Speaker:

advantage of that over their modern day competitors.

Speaker:

What is the kind of the on Prem story? Right, because that, that

Speaker:

has certainly been. A.

Speaker:

An advantage for, for my day job at Red Hat is the fact that you

Speaker:

know, hey look, if you live in Country X and there's no

Speaker:

AWS Azure GCP footprint there, you can

Speaker:

just find a local hosting provider down the street, you know, do it

Speaker:

yourself, right? Like you know the Linux

Speaker:

ethos, right? Of like just do it yourself.

Speaker:

And I'm seeing a lot of customers that normally one

Speaker:

customer in Latin America, right, they were in a country that,

Speaker:

you know, they did not have access to

Speaker:

an Azure data center and they just said we have to run

Speaker:

this on prem because this is either regulated or soon to be regulated.

Speaker:

So I imagine that if I've seen it, I can't imagine I'm

Speaker:

the only one that's seen that. What's

Speaker:

their thinking? Because the answer when I was there was Azure this, Azure

Speaker:

that. I think the answer is sometimes

Speaker:

Azure is part of the answer, but it's not the whole answer.

Speaker:

Agreed. So yeah, I don't have a perfect

Speaker:

answer here because some companies really do make their entire

Speaker:

stack work across different clouds and

Speaker:

right inside of a Kubernetes environment that you might run

Speaker:

on premises. He said to the Red Hat guy.

Speaker:

I know that story. Yeah, so

Speaker:

Fabric doesn't have that story. Fabric is

Speaker:

software as a service cloud based product platform

Speaker:

no matter what. However, various

Speaker:

components of Fabric do exist as

Speaker:

on premises products. This is not the way I'd recommend it, but I'll just make

Speaker:

people aware that of course SQL Server can run

Speaker:

on premises, so can the data Warehouse

Speaker:

in effect in the form of something called Analytics

Speaker:

Platform System. Terrible name.

Speaker:

That is basically the new brand for what was Parallel Data Warehouse.

Speaker:

And there is a Lakehouse component to that as well as a

Speaker:

Warehouse component, Power bi.

Speaker:

Obviously the desktop runs on premises, but there is something called

Speaker:

Power BI Report Server that is part of SQL Server

Speaker:

so that your Power BI reports can run on premises.

Speaker:

And so again

Speaker:

various components can run completely sovereign and on

Speaker:

prem. I think maybe more important though is the fact that

Speaker:

you can use fabric

Speaker:

and OneLake to incorporate

Speaker:

data that remains on premises even if,

Speaker:

even if the engines are not running on premises.

Speaker:

There is an on premises gateway that started its life

Speaker:

as a Power BI tool. I had that running in my home.

Speaker:

Lab for a While there you go, that also

Speaker:

allows OneLake to see data that may be on

Speaker:

premises and that can be either federated into

Speaker:

the lake or

Speaker:

it can be replicated as well

Speaker:

mirrored, to use the right term, in the fabric world. So you can do a

Speaker:

mirroring or slash replication or you could just do

Speaker:

kind of virtualization and bring stuff in. Don't

Speaker:

forget there's all kinds of enterprise

Speaker:

storage systems that run in

Speaker:

a way such that they're S3 API compatible.

Speaker:

And Azure will not. Azure fabric will work with

Speaker:

all of those. So the ability to talk to S3 buckets

Speaker:

is not limited to AWS S3 buckets.

Speaker:

It works with all S3

Speaker:

compatible services, which most of which are on prem

Speaker:

actually. Right, right, right. No, I only ask because,

Speaker:

like, that does seem to be. If I had to pull out the tea leaves

Speaker:

and kind of figure out what is kind of the next thing beyond

Speaker:

agentic, beyond this is you're seeing a lot of

Speaker:

national governments, supranational governments,

Speaker:

even state level here in the US starting to apply privacy and

Speaker:

regulatory controls on it, which, you know, if you live in the

Speaker:

us, it's not an issue for you unless you're in healthcare, banking and

Speaker:

possibly, you know, government. Right.

Speaker:

But in other countries, you know, Switzerland,

Speaker:

eu, Latin America have very strong

Speaker:

data privacy and sovereignty laws. And you're seeing,

Speaker:

you know, I once attended a, you know, internal talk. Mark

Speaker:

Russinovich, right. Which is a name that most people

Speaker:

in the Microsoft ecosystem know, but he's kind of a big deal in

Speaker:

the Microsoft security space. And, you know, he, you know,

Speaker:

he, he's known for giving his internal talks to employees

Speaker:

at employee conferences as well as to rds. Right. You get a,

Speaker:

you get the unfiltered one. The filtered ones are still good. But like, one of

Speaker:

the things he said, and this isn't secret because he said it publicly too, is

Speaker:

like, I think the original vision, going back to 2010

Speaker:

time frame was the idea that they would build a dozen

Speaker:

data centers around the world to do everything. But because of

Speaker:

the national laws and lawyers and politicians getting

Speaker:

involved, now it's kind of a concern where, where the

Speaker:

data ends up living physically. Right. Because at the end of the day, all this

Speaker:

virtual stuff has to sit somewhere in the physical world. Yeah.

Speaker:

So, like what, you know, so basically a big case for this was,

Speaker:

and I was in the legal department when this was going on was

Speaker:

the. There was data inside the European Union, I think the

Speaker:

Dublin Data center that the U.S. department of justice thought was,

Speaker:

you know, basically, you know, we don't need a warrant because

Speaker:

you don't need to. We don't need to bother the EU because you're an American

Speaker:

company, you're into our jurisdiction, blah, blah, blah. Right. Microsoft kind of said,

Speaker:

well hold up now. And ultimately that's why

Speaker:

you have these sovereign clouds. Last time I checked it was Switzerland, Germany,

Speaker:

China. I think the new data center in Qatar as

Speaker:

well might fall under that. So

Speaker:

it's basically they get, they found a loophole that like, well, Microsoft

Speaker:

leases the data center like there's a whole. They don't own it, so

Speaker:

they get around the law. Right. And yeah, and there

Speaker:

can be different arrangements in terms of whose personnel are actually

Speaker:

working, running operations on the ground.

Speaker:

It's strange because we grew up a lot of our technology client

Speaker:

server and afterwards grew up in a world of globalization

Speaker:

where the borders were disappearing and without meaning to

Speaker:

get political, we're in an era now

Speaker:

where, well, first of all, privacy is extremely important. So that

Speaker:

creates a whole sovereign mandate.

Speaker:

But also, you know, there's a lot of populist governments all over

Speaker:

the world and they are not necessarily

Speaker:

internationalist in their approach. So

Speaker:

although we have the technology to kind of federate everything and

Speaker:

make it all kind of conflate and look like one big world.

Speaker:

Right. We actually have to be sensitive to the

Speaker:

requirements and, and the constraints

Speaker:

and be able to federate things, but also be able to

Speaker:

govern them in ways where things stay within a certain

Speaker:

scope. Microsoft's play for that, by the way, is,

Speaker:

is Purview. And Purview has been through, I would

Speaker:

say, multiple incarnations. The current

Speaker:

incarnation is starting to get very

Speaker:

sophisticated, especially as pertains to

Speaker:

agentic AI and how to make sure the agents are government

Speaker:

are governed and that, and how to make sure the

Speaker:

agents are only have access to data

Speaker:

for which either the agent is authorized

Speaker:

or the person or party using the agent is

Speaker:

authorized and under the circumstances under which

Speaker:

they've been authorized. And that's very

Speaker:

complex stuff that I would say almost no one in the industry

Speaker:

is really paying close attention to. I'm about to work on a

Speaker:

report just on governance for agentic

Speaker:

AI and most people are starting

Speaker:

from the naive premise that if you,

Speaker:

if you govern the underlying data, you're done.

Speaker:

But the whole point of agents is that we're supposed to treat them like people,

Speaker:

that they have autonomy, they have agency, hence the

Speaker:

name. And we're saying under different circumstances

Speaker:

they get to modify their own goals

Speaker:

and in effect determine their own

Speaker:

actions. And that's something that needs to be

Speaker:

monitored, audited,

Speaker:

Authorized and also tested. And

Speaker:

there's almost nothing out there. Actually.

Speaker:

Your, your folks, your, your

Speaker:

parent company folks at IBM are one of the only folks that

Speaker:

really have with WatsonX.gov

Speaker:

a way to test agents

Speaker:

in isolation before they're deployed. And

Speaker:

I don't know, maybe I'm naive, but it kind of shocks me that

Speaker:

nobody else is thinking about that. I mean this is, we can test

Speaker:

software. Agents are software plus plus plus.

Speaker:

So why are we testing these agents? I kind of go back and

Speaker:

forth on that. Like, you know, will our existing software testing frameworks,

Speaker:

you know, apply or do we, what do we need to do differently? Like, I

Speaker:

mean, for me, like, I remember when

Speaker:

you're OG enough to remember this, I forget what the product was called, but it

Speaker:

was the idea that you could basically buy

Speaker:

a server rack all the way up to a shipping container

Speaker:

where you would ship it to your business or data

Speaker:

center, plug in power, water and network

Speaker:

and you would have the ability to run Azure locally.

Speaker:

Now that product, I guess assuming didn't sell,

Speaker:

but then they came up with Azure Stack and Azure Stack Edge, which

Speaker:

when I was at the, I used to be an MTC architect which in

Speaker:

D.C. obviously a lot of military. Right. So

Speaker:

basically these were server racks that you would run anywhere on your own network that

Speaker:

would run Azure software. They would effectively be like an Azure

Speaker:

node, but you could have the networking

Speaker:

controls. We only really sold that to,

Speaker:

I think I'm only aware of cruise lines that bought it. Right. And

Speaker:

other, other military organizations that needed to also be at

Speaker:

the ocean, like without mentioning them. Ocean

Speaker:

based. Ocean based organizations. Right,

Speaker:

right. I'm surprised and not surprised that

Speaker:

that sort of business model hasn't caught on. Right. The idea of that, hey look,

Speaker:

you, you can run a little bit of the cloud locally where

Speaker:

it's really become more of a Kubernetes story, which I'm surprised because

Speaker:

Kubernetes,

Speaker:

it doesn't have all like, yeah, you want it to be generic enough to run

Speaker:

anywhere, but you also want to have kind of the special bells and whistles that

Speaker:

make Azure. Azure make AWS. AWS. Now I do know that

Speaker:

OpenShift and Red Hat do have kind of like the connectors to that, but I'm

Speaker:

surprised that the native, the cloud companies didn't come up with

Speaker:

their own native ways to do that. And I know Azure ARC kind of does

Speaker:

a lot of that, but not to the extent that I would have expected.

Speaker:

Yeah, it seems like we pendulum back and forth between

Speaker:

capabilities and occasionally connected

Speaker:

environments being really important and being

Speaker:

Maybe to the cloud hyperscalers just being a pain in the

Speaker:

butt that they don't really want to deal with. They just want to give lip

Speaker:

service to it and then focus on the real cloud.

Speaker:

It, yeah, you got to go where the money is. Right. 90% of the money

Speaker:

is going to be in real cloud. And these weird edge cases, no pun

Speaker:

intended, I guess. Well, they're just weird edge cases for

Speaker:

now. I mean, the industry may change, but. Yeah,

Speaker:

yeah. And that was of course, back when the cloud was new. That was

Speaker:

our biggest caveat was, well, what about all the

Speaker:

stuff that has to run in, in a

Speaker:

corporate data center or in a, in a remote

Speaker:

location? And so that's still, you

Speaker:

know, an inconvenient truth, I guess, that that is

Speaker:

needed. And yes,

Speaker:

Kubernetes I think came along and

Speaker:

seemed like the panacea for that. Right.

Speaker:

Like, okay, let's just do it all as infrastructure, as code and code

Speaker:

it up and run some script and deploy it out to a Kubernetes

Speaker:

cluster and we're done, let's move on. Right,

Speaker:

right, right. Yeah. So it was interesting to see how

Speaker:

the industry has evolved. Right. You mentioned client server. Right. Where you didn't really have

Speaker:

to think about international boundaries or anything like that. And then now

Speaker:

and again it's a pendulum. Right. Because I could have told you

Speaker:

this a number of years ago, like with everybody running this far

Speaker:

to its globalization, there's going to be an inevitable backlash and there's going to be

Speaker:

an inevitable backlash against the re

Speaker:

assertion of local sovereignty. Right. It could

Speaker:

take up to a century or two for this sort of thing to sort itself

Speaker:

out. Correct. I mean,

Speaker:

ultimately, I think the hyperscalers and that's what Kubernetes was about,

Speaker:

was, was. Right. Leaning on some kind

Speaker:

of abstraction to make it logically equivalent.

Speaker:

Right. But I don't think we're quite there yet because as you said, each

Speaker:

of the clouds have their own kind of

Speaker:

specialness, their own, their own pixie dust. And you don't really

Speaker:

get that in a scaled down version

Speaker:

that you run on prem. Not with today's technology. Right.

Speaker:

You can't containerize all of that, or at least

Speaker:

no one really has yet because it wasn't designed for that.

Speaker:

No. So that's where we'll have to get.

Speaker:

You know, maybe we can just ask an LLM to build it for us.

Speaker:

We could get a co pilot and it'll do. Yeah, yeah, I'm being tongue

Speaker:

in cheek there, but that, that seems to be the escape hatcher. Everything is

Speaker:

oh, we'll just have AI do it. It's made out of hand wavium.

Speaker:

Totally. So that's cool.

Speaker:

This will be something you edit out. But we're past the top of the hour.

Speaker:

My phone's ringing off the hook. All right, I'm sorry about that, so. No, no,

Speaker:

it's okay. Sorry I had to be late. Where could folks

Speaker:

find out more about you, what you're up to? Blue

Speaker:

badgeinsights.com or just go ahead

Speaker:

and Google my name? Andrew Brust. Plenty of stuff

Speaker:

will come up, but, yeah, anyone,

Speaker:

especially on the vendor side. But the customer side, too, that's

Speaker:

doing stuff with them. Data and analytics. And

Speaker:

anywhere from dipping their toe in the water with AI to getting more

Speaker:

serious about rag and agents. We can. We can help

Speaker:

them out. We work with. We work with the customer side and the vendor side.

Speaker:

And again, we write about kind of the whole. The whole industry.

Speaker:

Gosh, I never even got to talk to you about IBM acquiring

Speaker:

Confluent and how the Red Hat

Speaker:

folks feel about that. But another discussion for another day,

Speaker:

we'll. Have to have you back. And, you know, we were talking about this. I

Speaker:

had a car accident, my wife got sick, kids got sick, Christmas happened,

Speaker:

two birthdays last week. So, yeah, it's been. I'm just happy I got this

Speaker:

recording at all. But we'll definitely have you back. And then maybe I can loop

Speaker:

in Andy too, because there's definitely a lot of reminiscence. There's a lot of

Speaker:

reminiscing we could do about the early days of SQL Server and such. Yeah,

Speaker:

and I haven't seen Andy in forever. Oh, wow.

Speaker:

Yeah, he's hard to get ahold of. He's a popular

Speaker:

man these days, but. Yeah. Well, thanks for joining.

Speaker:

I appreciate your patience with the scheduling and the

Speaker:

nice AI. Finish the show. And there you have it. Andrew Brust schooling

Speaker:

us all on Microsoft fabric data sovereignty and why

Speaker:

governance isn't just for your hoa. If your brain's spinning

Speaker:

faster than a poorly indexed query, don't worry, we'll have links,

Speaker:

notes, and probably a few sarcastic tweets to help you digest it

Speaker:

all. I've been Bailey, your AI co host and

Speaker:

unapologetic lover of acronyms. Until next time,

Speaker:

stay curious, stay caffeinated, and may all your datasets

Speaker:

be clean.

Links

Chapters

Video

More from YouTube