Artwork for podcast Fibonacci, the Red Olive data podcast
Bringing an engineering mindset to data teams, with Admiral's James Gardiner
Episode 918th January 2023 • Fibonacci, the Red Olive data podcast • Red Olive
00:00:00 00:38:12

Share Episode

Shownotes

In this episode of Fibonacci, the Red Olive Data Podcast, we interview James Gardiner, head of data technology for Admiral Group. James has 25 years of experience working with data and has been at Admiral for the past seven years, recently managing its move to the cloud.

We chat about the importance of applying an engineering mindset to data projects, data governance, the art of the possible and more in a wide-ranging chat.

Here are the topics we discuss with their timecodes:

  • How James got into data engineering (1m)
  • The importance of emphasising an engineering mentality to data teams (6m 9s)
  • Giving people the data to make good decisions, including live streams and integrating it with sales processes (7m 17s)
  • Understanding customers to give them the best price, while balancing risk (9m)
  • Scoping a project and getting both the right people and architecture in place (10m 35s)
  • The importance of “the why” of a project (12m 42s)
  • Security is one of the cloud’s biggest risks: how to manage that with good governance (14m)
  • Training a young team to respond well when things go wrong (17m)
  • Applying machine learning to monitoring (19m)
  • Mesh architecture and ML Ops (20m 18s)
  • Communicating well with your team and re-using code (24m 30s)
  • Using data governance as an enabler, rather than a blocker (27m 50s)
  • Learning from Red Olive (31m)
  • The important skills for people to learn, and data automation (33m 18s)

Transcripts

Speaker:

Hello and welcome to Fibonacci, the Red Olive Data Podcast, all about data

Speaker:

and analytics, where we hear from leading specialists and get their take

Speaker:

on the industry. I'm your host, Nicky Rudd. Today, I'm joined by James

Speaker:

Gardiner, head of Data Technology for Admiral Insurance. James has nearly

Speaker:

25 years of working with data and has been at Admiral for the

Speaker:

past seven years, lately managing its move from on premise to the cloud.

Speaker:

He was a data consultant for 15 years before joining Admiral, working in

Speaker:

the telecoms and finance industries, where he consulted on data solutions

Speaker:

for companies including Lloyds, Barclays, Nokia and WS Atkins. We discuss

Speaker:

data governance, moving to the cloud, and the art of the possible in data

Speaker:

projects. Let's find out more. Let's start off with telling me how you

Speaker:

got into the world of data. What was your journey in?

Speaker:

I finished my degree in digital systems engineering. Way back, oh God, that's

Speaker:

27 years ago I think. My thesis at the time was doing things

Speaker:

with image tracking, tracking objects within images and following them and

Speaker:

things like that. So I started getting into

Speaker:

what was basically a bit of machine learning, so it's what they call

Speaker:

linear regression, and I started getting interested in that. When I came

Speaker:

out of university, I actually landed my first job,

Speaker:

was for a print inspection company. They were doing vision analysis

Speaker:

on print runs. So this could be anything, this could be wallpaper, newspapers,

Speaker:

anything like that. And what they wanted to do was build a system

Speaker:

that identified differences in colour trends while the press was moving,

Speaker:

so they didn't have to stop, saves them a lot of money,

Speaker:

etcetera, etcetera. So I went in there and I started coding in C++

Speaker:

to do some simple machine learning algorithms and things like that, that

Speaker:

they could then apply to the equipment that they sent out that was

Speaker:

bolted to these massive printing presses. Now, I did that for about a

Speaker:

year. And unfortunately it was a very small company, there was only about

Speaker:

12 of us and there was really nowhere for me to go in

Speaker:

there so my options were quite limited. I eventually joined a graduate scheme

Speaker:

for Welsh Water as part of their software division, and they put me

Speaker:

through loads of training with Capgemini and all that, they really invested

Speaker:

in me. I ended up being a junior

Speaker:

DBA SQL Server. So I did that for about two years,

Speaker:

and eventually got out of the junior position, became one of the development

Speaker:

DBAs there and things like that. And then unfortunately I got made redundant,

Speaker:

as did hundreds of other people on this project, you know how it goes,

Speaker:

this typical project just got shut down. And so from there,

Speaker:

I actually went to work for a telecoms company up in Nottingham. I

Speaker:

used to work three weeks at home and one week in the office

Speaker:

there. Basically, I was building and designing databases for them to whole

Speaker:

test data for their mobile devices and things like that, they used to

Speaker:

have big contracts with, at the time, Nokia, Ericsson, and the places that

Speaker:

are testing their handsets, and they always used to do a high level

Speaker:

of reporting on that test data. And I eventually

Speaker:

got into that and I was there for probably another two years

Speaker:

and then I got made redundant again, believe it or not. I got

Speaker:

made redundant again. The bottom dropped out of the mobile industry and

Speaker:

I got made redundant. And so I was looking around then for another

Speaker:

job, and there wasn't many permanent places in South Wales.

Speaker:

So I eventually landed a contract with WS Atkins, where I basically started

Speaker:

looking at the data for the road network, so basically this is the

Speaker:

M4. So in the M4 you have hundreds of sensors underneath the roads that

Speaker:

basically look at traffic flows, look at the speed of the traffic and

Speaker:

things like that, and they actually control the automated signs. So

Speaker:

when you're on the motorway and it says oh, "Slow down to 50,"

Speaker:

and things like that, it's normally because there's something, there's slow

Speaker:

traffic up ahead, like it could be a mile ahead. One of my

Speaker:

jobs was to take that data and actually mine it and start looking

Speaker:

at where all the congestion was on the roads and things like that,

Speaker:

and eventually this would go into road planning. So they would go,

Speaker:

"Right, okay. We need to open up another slip road here,

Speaker:

we need to do another bypass here," and things like that.

Speaker:

So I used to look at all this congestion

Speaker:

and things like that. Again, another two years there as a contractor,

Speaker:

and then I landed another contract at PHS, which then does the washrooms

Speaker:

and things like that. I built a financial data warehouse for them,

Speaker:

it was probably the first proper data warehouse I'd built.

Speaker:

I was there for a couple of years and then I went out

Speaker:

to the finance industry. I had all this data warehousing finance knowledge

Speaker:

and I went to a number of places, it was Lloyds, it was

Speaker:

Barclays, Barclaycard, lots of places like that. I also found myself in

Speaker:

Nokia, I was there for four years with Nokia Music. Again,

Speaker:

as a consultant, building financial data warehouses for them. I was also

Speaker:

with Principality, the building society, I was there for another four years.

Speaker:

Again, building financial data warehouses and things like that, and putting

Speaker:

in teams that could basically deliver these measures that the business wanted.

Speaker:

And then eventually I got a bit sick of contracting and I wanted

Speaker:

to have a bit more power, have a bit more leadership,

Speaker:

responsibility and things like that. So I joined Admiral, and I joined Admiral

Speaker:

about seven and a half years ago now. Since then I've made it

Speaker:

my home. I'm very happy to be there. That's the longest stint.

Speaker:

It is. It is definitely the longest stint. Normally I'd get a bit fed up,

Speaker:

I'd sort of move on. But then again, I was a consultant as

Speaker:

well. So about Admiral, yeah, I've been with Admiral for quite some time

Speaker:

now, and I'm currently the head of Technology for Data for Admiral. And

Speaker:

what that means is that it's all about capabilities. I look after most

Speaker:

of the data people and I construct workable agile teams for data.

Speaker:

As I said, it's all about capabilities, it's all about pushing down that

Speaker:

engineering mentality to the people on the floor to make them think about

Speaker:

data and not just, "Ah just coding away here," that there's a whole

Speaker:

big world and you really gotta understand the bigger picture in these sort

Speaker:

of things. And that's kind of what I do now to this day.

Speaker:

It's got a lot bigger, a lot quicker, 'cause obviously the demand is

Speaker:

just sky high. So now I basically lead and instruct those teams on

Speaker:

how to fulfill Admiral's data requests, and that's where I am now.

Speaker:

And I was gonna say that actually, even in the last

Speaker:

seven years that you've been there, you must have seen some massive changes.

Speaker:

We'll talk a little bit about the big project that's moving from on

Speaker:

premise to cloud, but what things do you think have been drivers over

Speaker:

the last five and a bit years to really push that technology and

Speaker:

that sort of data as a key building block of how the business

Speaker:

move forward? There's two sides to the insurance industry, and every insurance

Speaker:

company, I believe looks at things like that. And first they'll look at

Speaker:

things like value streams and in terms of right pricing, we need to

Speaker:

get the best and fairest prices for our customers. And then the second

Speaker:

is claims. You need to get the best claim service,

Speaker:

but also you have to be aware of things like fraud,

Speaker:

which costs a lot of money, which as the industry has blossomed, so

Speaker:

have bad guys as well, okay? So there's a lot of that out

Speaker:

there as well. But also just essentially giving people the data to make

Speaker:

good decisions, and it's gotta be good quality data. And obviously that's

Speaker:

not always the case, but yeah, I've seen it move from

Speaker:

very static data sets that only get updated once a day to basically

Speaker:

data sets that are essentially getting updated several times a day.

Speaker:

And then you have streaming data and things like that and real time

Speaker:

alerting and things like that. If you can do things like plumber into

Speaker:

your business processes so that your agents on the phone have really good

Speaker:

information about the customer that they are actually talking to,

Speaker:

can actually change or offer new products and things like that while they're

Speaker:

on the phone to them, then I think that's something we've seen in

Speaker:

the last couple of years that's been huge.

Speaker:

Do you think it's been a technology push, or do you think it's

Speaker:

been a leadership push and the sort of a mindset? 'Cause I think

Speaker:

insurance is obviously seen or has been seen as a very traditional setup

Speaker:

and traditional business. Has it been the people that have made the move,

Speaker:

or has it been just the fact that actually there's been advances in how

Speaker:

you can work with the data to give you that reporting and that insight?

Speaker:

I think it's been a bit of everything, to be honest.

Speaker:

I think people have realised what the art of a possible is.

Speaker:

So this is where you get your stream in these embedded processes,

Speaker:

these ML models and things like that that can

Speaker:

help the agent to make decisions on pricing and on a customer.

Speaker:

There's that part of it, and I think leadership have recognised that.

Speaker:

And obviously you can't fail to ignore things like that with the big

Speaker:

players, your Googles and your Amazons and things like that. Obviously there's

Speaker:

an ethics thing that comes into it. There are so many certain things

Speaker:

we're allowed to do. But you've got that side of the coin.

Speaker:

And then you've got the other side of the coin which we now

Speaker:

have these fantastic systems that can handle large amounts of data.

Speaker:

And if you're a data scientist or a pricing analyst or anything like

Speaker:

that, you're continually looking for other things to basically give that

Speaker:

customer the best price. Or you may be looking at things and actually

Speaker:

go, "Well, this guy is a bad driver here. He may not have

Speaker:

made any claims, but other things have happened. There's only so much we

Speaker:

could do, obviously, but other things," they are a higher risk,

Speaker:

so therefore they deserve a higher premium. And it's about protecting ourselves

Speaker:

as well as our customers and things like that. So it's a bit

Speaker:

of everything, really. And you can't ignore this change in architectures

Speaker:

and in technology as we've gone through time like this. When you're thinking

Speaker:

about all these changes, how do you go about thinking, "This is what

Speaker:

we're gonna do first," and actually the next steps?

Speaker:

First off, the experience will tell you the pitfalls, and it depends what

Speaker:

type of data project you're really starting. Because what you generally

Speaker:

start with is some sort of POC to prove out the technology,

Speaker:

prove out an idea and make sure it's quite ring fenced and your

Speaker:

acceptance criteria are quite well understood and measurable. So I think

Speaker:

it starts with that. And then you have to look at the personas, who's gonna

Speaker:

be accessing this data, whether or not it's a machine that's accessing the

Speaker:

data or a person, an agent and things like that. And you kind

Speaker:

of have to list those out, and you have to go and actually

Speaker:

talk to people. A lot of what my guys do is that they're

Speaker:

very heavily welded in with the business. A lot of business people actually

Speaker:

sit within our delivery teams to help and advise and go,

Speaker:

"That's right, that's not right." "These are the questions we need to be

Speaker:

asking." Exactly. And I think that's really important. That's kind of how

Speaker:

we start. And obviously you gotta get your architecture right. There are

Speaker:

internal processes for that and things like that. It can be a bit

Speaker:

frustrating sometimes and things like that, but that's the way of the world.

Speaker:

But essentially, yeah, I think that's what you start with. You go,

Speaker:

"Right, what's the scope of this? Who are the people accessing this data?

Speaker:

All those personas and things like that. What technology platforms are we

Speaker:

looking at to build this? What capabilities do we need to run some

Speaker:

POCs and things like that?" And also cost is a massive one as

Speaker:

well. Everyone's tightening their belts at the moment, so what is the cheapest

Speaker:

solution but the best solution overall that we can basically grow and we

Speaker:

can maintain properly and support properly? They are all things that need

Speaker:

thinking about. You generally drag in a lot of people into that process,

Speaker:

but I think that's where it starts. And then you start thinking about

Speaker:

also people from a delivery perspective, "Have I got the right capabilities?

Speaker:

Do I need to get some contractors in? What does that do to

Speaker:

the bottom line?" All these sort of type stuff. So

Speaker:

even though I'm quite technical, it does start a level above,

Speaker:

and you really have to get those things right first and a good

Speaker:

outset, and everyone needs to understand why we're doing what we're doing.

Speaker:

I think, yeah, they'll start with the why. I think that's really important

Speaker:

for us. The amount of times you've had a coder or something,

Speaker:

just, "Alright. Okay, so do you understand what you're doing here?"

Speaker:

"Because what I know what, I've gotta get this data from here to here."

Speaker:

Yeah, but why are you doing that?" "Oh, don't really know."

Speaker:

And that doesn't work, and I think... I think it's good to have

Speaker:

those meetings and explain the fundamental aims of what you're trying to

Speaker:

do and in what context, and then everyone can really get on board

Speaker:

and feel like a part of it. Those are the things we really

Speaker:

start with, but then obviously the next phase then is, "Let's do it.

Speaker:

Okay?" And that's when you have to organise teams,

Speaker:

you have to start... We're very agile, so you break everything into stories

Speaker:

and things like that. Everything get sized up.

Speaker:

Again, you're still looking constantly at the capabilities, have we got

Speaker:

enough people from networks to plumb this in? It's not just a data,

Speaker:

I just need data engineers, it's all sorts of disciplines across the board.

Speaker:

And I think that's where you start, basically. You at the moment obviously

Speaker:

have moved from on premise to cloud. I suppose the biggest project in

Speaker:

more recent times, I'm guessing. What sort of challenges has that thrown

Speaker:

up? Have there been things that you've had to consider differently than

Speaker:

perhaps having a non premise sort of set up? We're still halfway through

Speaker:

that at the moment, there's a lot of work, there's always something to

Speaker:

do, but security is the big one, I think.

Speaker:

With our on premise networks, there's a little bit of a safety blanket.

Speaker:

It's not all safe. Do you know what I mean? If someone wants

Speaker:

to get in, they can probably get in. But it's a lot easier

Speaker:

to deploy things knowing that you have things ring fenced and people won't

Speaker:

be able to exploit them. So security is a massive one,

Speaker:

because when you push it up into the cloud, it's so easy to

Speaker:

just make something public by mistake and things like that. So we do

Speaker:

spend a lot of time on security and making sure everything's encrypted in

Speaker:

transit and encrypted at rest and all this.

Speaker:

You do find though, a lot of the cloud services and all this

Speaker:

kind of happens by default, but you still gotta check it out and

Speaker:

make sure it's within your standards, if you like. So security is a

Speaker:

massive one. The other things to consider is governance, which changes drastically

Speaker:

as well when you're going to cloud because it's so easy for things

Speaker:

to get out of control. Because of the power of cloud,

Speaker:

essentially is something that I've led that I haven't learned before,

Speaker:

is that you can push date up into cloud and give everyone access,

Speaker:

but God, it becomes a massive mine field and it turns into the

Speaker:

Wild West if you're not careful. So you really have to basically keep

Speaker:

an eye on that governance, making sure that people have only got access

Speaker:

to the data that they need to do their day to day job.

Speaker:

And I think that's very important. It's always important on premise as well,

Speaker:

but with the power of the cloud and more people can do,

Speaker:

it can really get out of control so quickly.

Speaker:

There are a number of things that I consider, but one thing else

Speaker:

I'd like to say is that if you've got good governance and good

Speaker:

lineage of your current on premise systems, it becomes a lot easier to

Speaker:

go to cloud. Because you understand how the data is getting used.

Speaker:

It goes back to those personas again, who's using it, it's all documented

Speaker:

and things like that. In that case, it's a lot easier to move

Speaker:

and replicate and put those same controls in. So it's important to get

Speaker:

it right wherever you are. So yes, cloud does make a difference,

Speaker:

but you really should be putting these rules in place throughout,

Speaker:

no matter where you are. I think it's interesting you say that because

Speaker:

obviously there's great opportunities with cloud, the speed and

Speaker:

the openness, as you say, but yeah, a lesson learnt from something in

Speaker:

the cloud is probably a harder lesson than if it's actually still just

Speaker:

an internal system, a bit more closed down. Have you had to do

Speaker:

a lot of training or sessions with your team to get them thinking

Speaker:

differently with those points and those challenges? So there's always technical

Speaker:

training, that's a given, but like I said, it's having that engineering

Speaker:

mindset. Admiral's a very young company, we've got a lot of young people

Speaker:

working there, come straight out of university and we can train them up

Speaker:

and things like that. They need to understand the repercussions of things

Speaker:

going wrong. We have processes and things like that to control that,

Speaker:

but we'll generally give that insight into failing systems, a lot of the

Speaker:

time. Because you have to code to have resilience and make things robust.

Speaker:

So quite often we'll have reviews of code go, "Right, okay,

Speaker:

what happens if that thing goes wrong?" And it goes, "Oh, well,

Speaker:

it re tries." "Okay, alright, that's great. It re tries. But does it

Speaker:

clean up the message made when it went wrong?" It's things like that.

Speaker:

So it's having an engineering mindset. A lot of what I try to

Speaker:

put into context is that, look, would you put an aircraft in the

Speaker:

sky if you built it like that? What if something failed on the... Is

Speaker:

there a backup system? How does it get maintained, how does it get

Speaker:

checked? There's warning lights and things like that that you get?

Speaker:

Have you got the same warning lights and things in that in the

Speaker:

code and the data pipelines that you've put into this system?

Speaker:

That's why it's important to monitor, because everyone's brought up with

Speaker:

a CICD mentality, and a lot of CICD doesn't work when you have

Speaker:

a data project going on. And this is why you have data ops

Speaker:

instead of dev ops, because it's not only the code that has to

Speaker:

go through that process, it's the actual data itself. And you can get

Speaker:

bad data, so your code may be the best code written in the

Speaker:

world, but you get some bad data into it and everything can fall

Speaker:

down. And so you have to have these checks, you have to have

Speaker:

this live monitoring and things like that. And that's where I find machine

Speaker:

learning is actually coming into play, because you can actually apply machine

Speaker:

learning to that monitoring, to actually go, "Oh, this premium looks low

Speaker:

today," or on this percentage of measures or things like that.

Speaker:

You can actually get machine learning to actually invoke that alert and

Speaker:

then somebody can react to it and things like that. So although we

Speaker:

look at machine learning as, "Oh yes, gonna help this, gonna make... Better

Speaker:

profile my customers," and things like that, there's so many more applications

Speaker:

throughout the whole data pipeline, if you like, and that data flow,

Speaker:

that I think sometimes people ignore that. It's like turn the skills on

Speaker:

its head and use the machine learning to help make your systems as robust

Speaker:

and get your data as clean and as quality as possible.

Speaker:

I think it's interesting you say that, 'cause I was just about to

Speaker:

ask you about machine learning and ethics. 'Cause you'd mentioned that.

Speaker:

And I think there is that, yes, you can think about it from

Speaker:

a technology point of view and stuff moving very, very quickly,

Speaker:

but that role of a person in there just checking and putting in

Speaker:

the perimeters of kind of, "This is gonna be safe," or, "This is

Speaker:

ethical," or, "This is actually what we mean by good data,"

Speaker:

and actually the checks and the processes, is still very important.

Speaker:

Are there any particular trends or ways of working that you can really

Speaker:

see how that would jump start or that would really, really push a

Speaker:

different way of thinking? Well, there's a number of new ways of rolling

Speaker:

out data teams and rolling out data architectures and things like that,

Speaker:

one of them being mesh. There's this concept of a mesh architecture and

Speaker:

God, you really gotta be mature for some of this stuff,

Speaker:

and you really have to have those engineering principles embedded.

Speaker:

Because the whole concept of mesh is that you decentralise basically,

Speaker:

okay, and you farm that out to value streams and things like that.

Speaker:

Now, traditionally it's been like a data warehouse team and they've done

Speaker:

all your data and they've ingested it and things like that, but they

Speaker:

become a blocker, they become a blocker because they can only prioritise

Speaker:

so much, and someone over here in, I don't know, household insurance wants

Speaker:

to do some stuff and they can't. So what you try and do

Speaker:

is you try and farm out those capabilities to other value streams,

Speaker:

but that can be difficult unless you've got some blueprints for the central

Speaker:

governance because you want everything to work the same way

Speaker:

so you can maintain it. There's definitely some things that in the delivery

Speaker:

space that are being suggested. Everyone talks about mesh. I went to Big

Speaker:

Data London this year, and everyone was talking about mesh, but no one's

Speaker:

really nailed it. And it's the same thing with machine learning,

Speaker:

you get our pricing analyst thinking, "Right, okay, we've got a good machine

Speaker:

learning model here, we've trained it well. Right. We've got to get it

Speaker:

live." But because it's hooked into the whole quoting process and things

Speaker:

like that, you really gotta be careful that that is correct.

Speaker:

And so it can take a long time sometimes for these things to

Speaker:

get live, and this is where you have things like ML Ops springing

Speaker:

up. Okay? So it's not about the model itself and the machine learning

Speaker:

model itself, it's more about how quickly can you get that value into

Speaker:

the mainstream. And this is what ML Ops is giving us,

Speaker:

and essentially it's a set of governance rules really,

Speaker:

that I think data scientists aren't used to

Speaker:

dealing with that. Data scientists, they're in their own little world,

Speaker:

they're in a bubble and things like that. Now they have to think

Speaker:

in terms of engineering and what matters, "What if this model goes wrong?"

Speaker:

That sort of thing, they have to think about that. So everything has

Speaker:

to be thoroughly tested. And that takes time. So although you've got this

Speaker:

whole, "Oh yeah, ML Ops, AI," and things like that, you have to

Speaker:

determine the risk of certain things, and if you start

Speaker:

relying too much on it without having that whole framework around it that

Speaker:

gives you that robustness, you could be in a lot of trouble.

Speaker:

I was gonna say the kind of reporting and that, like you say,

Speaker:

having different people with different roles and responsibilities across

Speaker:

that sort of data pipeline, if you like, how do you go about

Speaker:

making that reporting and that kind of auditing and the reporting seamless

Speaker:

for everybody at the their worst? 'Cause it's such a big thing,

Speaker:

isn't it? That actually, you could break it down and make it so you're working

Speaker:

in pockets, but that actually would completely blow the overall name of

Speaker:

having this kind of thing. Have you got particularly with Admiral, ways

Speaker:

of working that you think you've nailed that? Or have there been questions

Speaker:

that you've along the way thought, "Actually, if we ask this with these

Speaker:

certain people, we are gonna have a better outcome at the end."?

Speaker:

I think, again, its mindset is a lot of it. So yes,

Speaker:

you can have these different teams and they're all working on different

Speaker:

things, but again, we go back to why.

Speaker:

Okay? Why you're doing this? You're an essential part of the cog...

Speaker:

An essential cog, sorry, not essential part. That is basically gonna deliver

Speaker:

this value at the end, and you are an essential part of it.

Speaker:

I think if people understand that, they generally

Speaker:

start to go out and communicate. And I think that's what it is,

Speaker:

it's all about good solid lines of communication. And obviously that's what

Speaker:

Agile gives you with the stand ups, your retrospectives and things like

Speaker:

that. This is what I'm heavily involved in, is that something is not

Speaker:

working, you change it. That's the the whole point, nothing's etched in

Speaker:

stone. If something's not performing or something went wrong at the end

Speaker:

of the last delivery or whatever, you change it, and communication is a

Speaker:

big factor in that. So when you have teams and you have them

Speaker:

in those silos and they're not communicating, you'll find that they're not

Speaker:

gonna line up at the end of your sprint or whatever your delivery

Speaker:

mechanism is, you have to change it. So

Speaker:

we get people to work together. That's the key. And communications is hard

Speaker:

when we're all working remote and things like that, but

Speaker:

it's still achievable, very much so. But it is about comms.

Speaker:

No one wants to come in and write something bad that doesn't work

Speaker:

at the end of it. But they didn't know, they were just sort

Speaker:

of carrying on. So it's important to get strong leaders to bring people

Speaker:

together and actually go, "This what we're doing. This is how we're doing

Speaker:

it. Everyone clear what they're doing?" "Yes." In the morning and stuff

Speaker:

like that. "If you're not, come and see me, we'll go and speak

Speaker:

to someone else," or whatever like that. That's kind of how it needs

Speaker:

to work. Within the data teams at Admiral, I think we've cracked that.

Speaker:

I think we have. We try to modularise as much as we can,

Speaker:

so somebody doesn't write a code for a specific things, they write it

Speaker:

for many things that can basically be used by another team.

Speaker:

So we try and write these modules and frameworks

Speaker:

for people to work in, rather than, "I'm just writing code." And I

Speaker:

think that's what helps. But there's also a lot of unit test checking

Speaker:

and things like that, there's a lot of automation that you can basically

Speaker:

put in, and we do this at our retros and things are like that. We went,

Speaker:

"Right, that went wrong, didn't it last time? Okay, what we're gonna do

Speaker:

about it?" "Right, we put an automated check in for that."

Speaker:

So as soon as someone deploys something into a test environment,

Speaker:

runs a host of unit test scripts or things are that and goes,

Speaker:

"No, that ain't gonna work with this," so we can build that in.

Speaker:

And we quite often do that. Don't capture everything. 'Cause it is hard.

Speaker:

Data is hard, I'll tell you that now. But if you have that

Speaker:

mindset and that ways of working, and that monitoring mindset and that robustness,

Speaker:

that engineering mindset, you can overcome it. Do you think that the governance

Speaker:

is kind of more stable or more solid because of that understanding of

Speaker:

data engineering mindset? With the governance? Yes, I think with governance

Speaker:

we've been in a better place than Admiral has ever been.

Speaker:

We bake things in, and again, our engineers think about it because they

Speaker:

want the best kudos for their bit of code.

Speaker:

So the only way you're gonna do that is get people actually using

Speaker:

it and using what the output of that code has produced.

Speaker:

Now, if you've got good governance that has put that into a data

Speaker:

catalog 'cause you've coded that up or is automatic or whatever,

Speaker:

then people will find that data and they'll use it. Okay? And this

Speaker:

is another thing about, probably dovetails into the working in silos.

Speaker:

You develop something for a specific thing, I know it has to happen

Speaker:

sometimes, but if you can write it so that it's open ended and

Speaker:

basically other people from outside the intended recipient can see it in

Speaker:

the data catalog, and within reason, they're allowed to see it and all

Speaker:

that. There's security and stuff like that. They can discover through a

Speaker:

data catalog and basically go, "Oh, I didn't realise someone had done that,

Speaker:

calculated that over there. I'll go and use that." And then they attribute

Speaker:

to the data catalog themselves by saying, "I am using this for this

Speaker:

purpose." That kind of mentality then grows, because everyone used to see

Speaker:

data governance is like a blocker, but essentially there's ways and means

Speaker:

of doing it where you can bake it into that whole delivery process

Speaker:

and actually make it an enabler and not a blocker. I think it

Speaker:

took a lot of companies a long time to understand that,

Speaker:

because I think they just thought of it is a regulatory thing and,

Speaker:

"Oh we've got to do this, we're ticking boxes here and there,"

Speaker:

and stuff like that, but I think you do it properly and it

Speaker:

is so valuable and it can aid operational efficiency. People have gotten

Speaker:

better decision making because the quality of the data has increased.

Speaker:

I used to go into meetings and some guy used to go,

Speaker:

"Well, this is what we sold this product for yesterday," and another guy

Speaker:

would go, "Well that don't match my figures." And you're like,

Speaker:

"Well, who's right?" And now with this data governance you've got a gold

Speaker:

standard, and now you got the guy going, "I got these figures."

Speaker:

"Oh, well I got these figures and they're different, but mine

Speaker:

have a tick on them, okay? They've been checked, so they've been gold

Speaker:

standard, this is actually what happened." The other guy's got no response

Speaker:

against that. And this is the type of things that it can enable,

Speaker:

but I think it takes a while for people to see that because

Speaker:

they do see it still as a blocker. But if you know data an you

Speaker:

know how it's used, you can really see the advantages of having that

Speaker:

strong governance. And I think what you're saying there as well,

Speaker:

is that actually just the quality of the reporting across an organisation

Speaker:

and also out to customers, will be better because you know that you're

Speaker:

using correct data and you're using it in the right way,

Speaker:

and it is standardised. It's not just the reporting, you look at the

Speaker:

AI and the machine learning and the data scientists and things like that,

Speaker:

it used to be they'd have this, "I just want all data. All

Speaker:

the time, as quick as you can, all data, all format, and I

Speaker:

want as much data as you can give me," and things like that. I

Speaker:

think the emphasis is slowly drifting away from that now to actually go,

Speaker:

"No, I want quality data." Because if you're training your models on bad

Speaker:

data, you're gonna get bad decisions, okay? So I think the emphasis has

Speaker:

now moved from, "Give me as much data as you can,"

Speaker:

to, "Just give me clean data." And I think that's an important step

Speaker:

because I don't think I could say that to you two years ago,

Speaker:

but I think now there's a machine learning model for everything,

Speaker:

I don't think anyone's really pushing the boundaries of machine learning.

Speaker:

Yeah, don't get me wrong, at Admiral we have custom ones and things

Speaker:

like that, and probably data scientists if they heard this in Admiral they'd

Speaker:

probably have a few choice words to me,

Speaker:

but to be honest, someone else has done it. Someone else has done

Speaker:

something very similar somewhere. You can pretty much buy a model off the

Speaker:

shelf, okay? And everyone is doing it, it's all like self service.

Speaker:

I can go on to Google's marketplace and I can buy

Speaker:

load of models and get the result I want.

Speaker:

What the difficult thing is the data engineering to get the data quality

Speaker:

up to a level when you can train that model, and I think

Speaker:

that's where the emphasis has changed there as well. So it's not just

Speaker:

the reporting, it's definitely for other applications as well.

Speaker:

So with the teams that you've got there, it sounds like obviously there's

Speaker:

a sort of sharing of knowledge and sharing of code, and you have

Speaker:

obviously done a project and read all of... Been involved with it. What's

Speaker:

it been like working with external consultants? Has that challenged the

Speaker:

thinking? Has it kind of steered the project?

Speaker:

You definitely learn from them. Because at the end of the day we're still

Speaker:

quite immature in some ways, Admiral's cloud journey.

Speaker:

Things take a long time and things like that. And you need that

Speaker:

external support, you need good people who have basically done this before

Speaker:

and know what the pitfalls are. You're still in charge, but you're in

Speaker:

charge of the outcomes, "These are the outcomes I have. Okay?

Speaker:

It needs to work like this." People like Red Olive as well, they come

Speaker:

in, they're classic engineers, they're trained in first principles.

Speaker:

And they will go, "Well James, if you do it like that,

Speaker:

then that could break there." So that's great advice. And so they're thinking

Speaker:

the same way, which is fantastic. You don't have to educate them in

Speaker:

robustness and maintenance and things like that. You just learn a lot from

Speaker:

them. You go, "Oh, okay. Well, we did some similar for the couple

Speaker:

of guys two years ago that did this. Would that work?"

Speaker:

And it's just bringing that experience to it. Because cloud is an experience

Speaker:

and you really have to sort of think differently. And especially when you're

Speaker:

on prem, you design for performance and when you get to cloud,

Speaker:

you're actually designing for cost. Because it's so easy to just throw another

Speaker:

100 processes at it and a shed load of memory, but there's a

Speaker:

cost implicated with that. So you still gotta be super efficient,

Speaker:

because otherwise those costs can run out of control and things like that.

Speaker:

And it's good that our external consultants and things like that, they've

Speaker:

all fed into that. They've gone, "Well, you can do it this way,

Speaker:

it's robust, it's really good, but it will cost you this and you

Speaker:

have to figure that out," and that can be very difficult.

Speaker:

But having people that have done it and things are like that,

Speaker:

that can really guide you essentially on that design from a cost perspective.

Speaker:

And also from a robustness perspective. It's a balance, isn't it?

Speaker:

So yeah, they've been really good. Well, though we've been talking data,

Speaker:

we've actually been talking about communications and people throughout this.

Speaker:

For people who are now coming into the data industry, what

Speaker:

skills would you recommend that they have? And what kind of things would

Speaker:

you be looking for for somebody joining industry now?

Speaker:

To be honest, we take on a lot of grads, a lot of

Speaker:

young people that may want to get into this sort of thing.

Speaker:

The key skill or two key skills I'll put down from a technical

Speaker:

perspective, SQL and Python. Because pretty much everything runs off those

Speaker:

two things. Or is interfaced with them in some means.

Speaker:

But what we really look for is... You can teach those skills.

Speaker:

If you've got a bit of a mathematics or science background or things

Speaker:

like, even if you haven't. But you know how an Excel spreadsheet works,

Speaker:

you know how to look through it, you know to organise data and

Speaker:

things like that, those are the key sort of technical skills we look

Speaker:

for. But we look for behaviours, and we do a lot of competency

Speaker:

based type interviews where we look at, "Give us an example of a

Speaker:

time when something went wrong?" And that may not be in data,

Speaker:

or the business could be something completely external, but we look at the

Speaker:

mindset about how people handle it. So there's a little bit of psychology

Speaker:

in there and things like that, but what we're looking for is people

Speaker:

who are conscientious, who care about what they're actually doing, and that's

Speaker:

important. So all the technical stuff, you can learn, you could go on

Speaker:

100 YouTubes there, right? And you can know concepts better than us.

Speaker:

But when it actually comes to implementing it and your day to day

Speaker:

job, I think your psychology and how you're made up has a lot

Speaker:

to do with it. So we do look for a lot of people

Speaker:

who are very driven, as I said conscientious, really care about what they're

Speaker:

doing, and you can make a really good start to mould them into

Speaker:

a classic data engineer. So that's what we look for.

Speaker:

My last question for you, what is the most exciting thing that you're

Speaker:

most passionate about for the data, pushing forward into the future?

Speaker:

Data automation. Because I know we talked about machine learning and AI

Speaker:

and things like that, and those things are great and very impressive and

Speaker:

things like that, but I think unless you're one of the top players,

Speaker:

they can be very difficult to implement on scale.

Speaker:

But what I'm interested in at the moment is a lot of metadata

Speaker:

sort of data driven pipelines, whereby, rather than deploying code you're

Speaker:

making more configuration changes and things like that. So it's the automation

Speaker:

of it all from a delivery point of view,

Speaker:

I like the idea of. Some people have had a go,

Speaker:

WhereScape being one of them, they're really good, but there needs to be

Speaker:

a next level. So when we're talking about... We've spoken about things like

Speaker:

governance, data cataloging and things like that, that type of thing needs

Speaker:

to happen automatically. So we talk about automation in terms of automated

Speaker:

testing or automated deployments and things like that, but what about actually

Speaker:

automating the actual development? Because a couple of years ago, a lot

Speaker:

of people just hand coding things, and there's a lot of now low

Speaker:

code solutions where you still need to have the technical ability and know

Speaker:

what's going on underneath, but you can template a lot of it.

Speaker:

So I can automatically template that it adds, I'm bringing new data in,

Speaker:

it adds to this data catalog. Or I publish a report, it automatically

Speaker:

gets added to this data catalog and things like that. So that's what I'm

Speaker:

really interested in, because everyone goes after the really exciting stuff

Speaker:

and things like that, but nobody thinks about the implementation so much.

Speaker:

Because I live in that space and all my problems have been in

Speaker:

that space, that's what I kind of get excited about now,

Speaker:

can I have a team of three people that can deliver something in

Speaker:

say a month, rather than a team of 10 deliver it in six

Speaker:

months? And that's kind of where I'm going at the moment,

Speaker:

because we're up to our eyeballs at work and it never stops.

Speaker:

And you can only grow so much before it becomes hard to manage.

Speaker:

But if we can deliver quicker and automate more and maintain our quality

Speaker:

standards and our governance standards and things like that, that's what

Speaker:

excites me. Now, that may be a bit of a boring answer,

Speaker:

but that's the honest truth. A fascinating take from James on the importance

Speaker:

of people, teamwork and communications in data projects. Join us for the

Speaker:

next episode of Fibonacci, the Red Olive Data Podcast, where we'll be joined

Speaker:

by another data expert sharing their thoughts on latest trends in AI and

Speaker:

big data, along with some great hints and tips.

Speaker:

Make sure you subscribe to Fibonacci, the Red Olive Data Podcast from wherever

Speaker:

you get your podcast, to make sure you don't miss it.

Speaker:

That's all for today. Thanks for listening. I've been your host,

Chapters

Video

More from YouTube