In this episode of Fibonacci, the Red Olive Data Podcast, we interview James Gardiner, head of data technology for Admiral Group. James has 25 years of experience working with data and has been at Admiral for the past seven years, recently managing its move to the cloud.
We chat about the importance of applying an engineering mindset to data projects, data governance, the art of the possible and more in a wide-ranging chat.
Here are the topics we discuss with their timecodes:
Hello and welcome to Fibonacci, the Red Olive Data Podcast, all about data
Speaker:and analytics, where we hear from leading specialists and get their take
Speaker:on the industry. I'm your host, Nicky Rudd. Today, I'm joined by James
Speaker:Gardiner, head of Data Technology for Admiral Insurance. James has nearly
Speaker:25 years of working with data and has been at Admiral for the
Speaker:past seven years, lately managing its move from on premise to the cloud.
Speaker:He was a data consultant for 15 years before joining Admiral, working in
Speaker:the telecoms and finance industries, where he consulted on data solutions
Speaker:for companies including Lloyds, Barclays, Nokia and WS Atkins. We discuss
Speaker:data governance, moving to the cloud, and the art of the possible in data
Speaker:projects. Let's find out more. Let's start off with telling me how you
Speaker:got into the world of data. What was your journey in?
Speaker:I finished my degree in digital systems engineering. Way back, oh God, that's
Speaker:27 years ago I think. My thesis at the time was doing things
Speaker:with image tracking, tracking objects within images and following them and
Speaker:things like that. So I started getting into
Speaker:what was basically a bit of machine learning, so it's what they call
Speaker:linear regression, and I started getting interested in that. When I came
Speaker:out of university, I actually landed my first job,
Speaker:was for a print inspection company. They were doing vision analysis
Speaker:on print runs. So this could be anything, this could be wallpaper, newspapers,
Speaker:anything like that. And what they wanted to do was build a system
Speaker:that identified differences in colour trends while the press was moving,
Speaker:so they didn't have to stop, saves them a lot of money,
Speaker:etcetera, etcetera. So I went in there and I started coding in C++
Speaker:to do some simple machine learning algorithms and things like that, that
Speaker:they could then apply to the equipment that they sent out that was
Speaker:bolted to these massive printing presses. Now, I did that for about a
Speaker:year. And unfortunately it was a very small company, there was only about
Speaker:12 of us and there was really nowhere for me to go in
Speaker:there so my options were quite limited. I eventually joined a graduate scheme
Speaker:for Welsh Water as part of their software division, and they put me
Speaker:through loads of training with Capgemini and all that, they really invested
Speaker:in me. I ended up being a junior
Speaker:DBA SQL Server. So I did that for about two years,
Speaker:and eventually got out of the junior position, became one of the development
Speaker:DBAs there and things like that. And then unfortunately I got made redundant,
Speaker:as did hundreds of other people on this project, you know how it goes,
Speaker:this typical project just got shut down. And so from there,
Speaker:I actually went to work for a telecoms company up in Nottingham. I
Speaker:used to work three weeks at home and one week in the office
Speaker:there. Basically, I was building and designing databases for them to whole
Speaker:test data for their mobile devices and things like that, they used to
Speaker:have big contracts with, at the time, Nokia, Ericsson, and the places that
Speaker:are testing their handsets, and they always used to do a high level
Speaker:of reporting on that test data. And I eventually
Speaker:got into that and I was there for probably another two years
Speaker:and then I got made redundant again, believe it or not. I got
Speaker:made redundant again. The bottom dropped out of the mobile industry and
Speaker:I got made redundant. And so I was looking around then for another
Speaker:job, and there wasn't many permanent places in South Wales.
Speaker:So I eventually landed a contract with WS Atkins, where I basically started
Speaker:looking at the data for the road network, so basically this is the
Speaker:M4. So in the M4 you have hundreds of sensors underneath the roads that
Speaker:basically look at traffic flows, look at the speed of the traffic and
Speaker:things like that, and they actually control the automated signs. So
Speaker:when you're on the motorway and it says oh, "Slow down to 50,"
Speaker:and things like that, it's normally because there's something, there's slow
Speaker:traffic up ahead, like it could be a mile ahead. One of my
Speaker:jobs was to take that data and actually mine it and start looking
Speaker:at where all the congestion was on the roads and things like that,
Speaker:and eventually this would go into road planning. So they would go,
Speaker:"Right, okay. We need to open up another slip road here,
Speaker:we need to do another bypass here," and things like that.
Speaker:So I used to look at all this congestion
Speaker:and things like that. Again, another two years there as a contractor,
Speaker:and then I landed another contract at PHS, which then does the washrooms
Speaker:and things like that. I built a financial data warehouse for them,
Speaker:it was probably the first proper data warehouse I'd built.
Speaker:I was there for a couple of years and then I went out
Speaker:to the finance industry. I had all this data warehousing finance knowledge
Speaker:and I went to a number of places, it was Lloyds, it was
Speaker:Barclays, Barclaycard, lots of places like that. I also found myself in
Speaker:Nokia, I was there for four years with Nokia Music. Again,
Speaker:as a consultant, building financial data warehouses for them. I was also
Speaker:with Principality, the building society, I was there for another four years.
Speaker:Again, building financial data warehouses and things like that, and putting
Speaker:in teams that could basically deliver these measures that the business wanted.
Speaker:And then eventually I got a bit sick of contracting and I wanted
Speaker:to have a bit more power, have a bit more leadership,
Speaker:responsibility and things like that. So I joined Admiral, and I joined Admiral
Speaker:about seven and a half years ago now. Since then I've made it
Speaker:my home. I'm very happy to be there. That's the longest stint.
Speaker:It is. It is definitely the longest stint. Normally I'd get a bit fed up,
Speaker:I'd sort of move on. But then again, I was a consultant as
Speaker:well. So about Admiral, yeah, I've been with Admiral for quite some time
Speaker:now, and I'm currently the head of Technology for Data for Admiral. And
Speaker:what that means is that it's all about capabilities. I look after most
Speaker:of the data people and I construct workable agile teams for data.
Speaker:As I said, it's all about capabilities, it's all about pushing down that
Speaker:engineering mentality to the people on the floor to make them think about
Speaker:data and not just, "Ah just coding away here," that there's a whole
Speaker:big world and you really gotta understand the bigger picture in these sort
Speaker:of things. And that's kind of what I do now to this day.
Speaker:It's got a lot bigger, a lot quicker, 'cause obviously the demand is
Speaker:just sky high. So now I basically lead and instruct those teams on
Speaker:how to fulfill Admiral's data requests, and that's where I am now.
Speaker:And I was gonna say that actually, even in the last
Speaker:seven years that you've been there, you must have seen some massive changes.
Speaker:We'll talk a little bit about the big project that's moving from on
Speaker:premise to cloud, but what things do you think have been drivers over
Speaker:the last five and a bit years to really push that technology and
Speaker:that sort of data as a key building block of how the business
Speaker:move forward? There's two sides to the insurance industry, and every insurance
Speaker:company, I believe looks at things like that. And first they'll look at
Speaker:things like value streams and in terms of right pricing, we need to
Speaker:get the best and fairest prices for our customers. And then the second
Speaker:is claims. You need to get the best claim service,
Speaker:but also you have to be aware of things like fraud,
Speaker:which costs a lot of money, which as the industry has blossomed, so
Speaker:have bad guys as well, okay? So there's a lot of that out
Speaker:there as well. But also just essentially giving people the data to make
Speaker:good decisions, and it's gotta be good quality data. And obviously that's
Speaker:not always the case, but yeah, I've seen it move from
Speaker:very static data sets that only get updated once a day to basically
Speaker:data sets that are essentially getting updated several times a day.
Speaker:And then you have streaming data and things like that and real time
Speaker:alerting and things like that. If you can do things like plumber into
Speaker:your business processes so that your agents on the phone have really good
Speaker:information about the customer that they are actually talking to,
Speaker:can actually change or offer new products and things like that while they're
Speaker:on the phone to them, then I think that's something we've seen in
Speaker:the last couple of years that's been huge.
Speaker:Do you think it's been a technology push, or do you think it's
Speaker:been a leadership push and the sort of a mindset? 'Cause I think
Speaker:insurance is obviously seen or has been seen as a very traditional setup
Speaker:and traditional business. Has it been the people that have made the move,
Speaker:or has it been just the fact that actually there's been advances in how
Speaker:you can work with the data to give you that reporting and that insight?
Speaker:I think it's been a bit of everything, to be honest.
Speaker:I think people have realised what the art of a possible is.
Speaker:So this is where you get your stream in these embedded processes,
Speaker:these ML models and things like that that can
Speaker:help the agent to make decisions on pricing and on a customer.
Speaker:There's that part of it, and I think leadership have recognised that.
Speaker:And obviously you can't fail to ignore things like that with the big
Speaker:players, your Googles and your Amazons and things like that. Obviously there's
Speaker:an ethics thing that comes into it. There are so many certain things
Speaker:we're allowed to do. But you've got that side of the coin.
Speaker:And then you've got the other side of the coin which we now
Speaker:have these fantastic systems that can handle large amounts of data.
Speaker:And if you're a data scientist or a pricing analyst or anything like
Speaker:that, you're continually looking for other things to basically give that
Speaker:customer the best price. Or you may be looking at things and actually
Speaker:go, "Well, this guy is a bad driver here. He may not have
Speaker:made any claims, but other things have happened. There's only so much we
Speaker:could do, obviously, but other things," they are a higher risk,
Speaker:so therefore they deserve a higher premium. And it's about protecting ourselves
Speaker:as well as our customers and things like that. So it's a bit
Speaker:of everything, really. And you can't ignore this change in architectures
Speaker:and in technology as we've gone through time like this. When you're thinking
Speaker:about all these changes, how do you go about thinking, "This is what
Speaker:we're gonna do first," and actually the next steps?
Speaker:First off, the experience will tell you the pitfalls, and it depends what
Speaker:type of data project you're really starting. Because what you generally
Speaker:start with is some sort of POC to prove out the technology,
Speaker:prove out an idea and make sure it's quite ring fenced and your
Speaker:acceptance criteria are quite well understood and measurable. So I think
Speaker:it starts with that. And then you have to look at the personas, who's gonna
Speaker:be accessing this data, whether or not it's a machine that's accessing the
Speaker:data or a person, an agent and things like that. And you kind
Speaker:of have to list those out, and you have to go and actually
Speaker:talk to people. A lot of what my guys do is that they're
Speaker:very heavily welded in with the business. A lot of business people actually
Speaker:sit within our delivery teams to help and advise and go,
Speaker:"That's right, that's not right." "These are the questions we need to be
Speaker:asking." Exactly. And I think that's really important. That's kind of how
Speaker:we start. And obviously you gotta get your architecture right. There are
Speaker:internal processes for that and things like that. It can be a bit
Speaker:frustrating sometimes and things like that, but that's the way of the world.
Speaker:But essentially, yeah, I think that's what you start with. You go,
Speaker:"Right, what's the scope of this? Who are the people accessing this data?
Speaker:All those personas and things like that. What technology platforms are we
Speaker:looking at to build this? What capabilities do we need to run some
Speaker:POCs and things like that?" And also cost is a massive one as
Speaker:well. Everyone's tightening their belts at the moment, so what is the cheapest
Speaker:solution but the best solution overall that we can basically grow and we
Speaker:can maintain properly and support properly? They are all things that need
Speaker:thinking about. You generally drag in a lot of people into that process,
Speaker:but I think that's where it starts. And then you start thinking about
Speaker:also people from a delivery perspective, "Have I got the right capabilities?
Speaker:Do I need to get some contractors in? What does that do to
Speaker:the bottom line?" All these sort of type stuff. So
Speaker:even though I'm quite technical, it does start a level above,
Speaker:and you really have to get those things right first and a good
Speaker:outset, and everyone needs to understand why we're doing what we're doing.
Speaker:I think, yeah, they'll start with the why. I think that's really important
Speaker:for us. The amount of times you've had a coder or something,
Speaker:just, "Alright. Okay, so do you understand what you're doing here?"
Speaker:"Because what I know what, I've gotta get this data from here to here."
Speaker:Yeah, but why are you doing that?" "Oh, don't really know."
Speaker:And that doesn't work, and I think... I think it's good to have
Speaker:those meetings and explain the fundamental aims of what you're trying to
Speaker:do and in what context, and then everyone can really get on board
Speaker:and feel like a part of it. Those are the things we really
Speaker:start with, but then obviously the next phase then is, "Let's do it.
Speaker:Okay?" And that's when you have to organise teams,
Speaker:you have to start... We're very agile, so you break everything into stories
Speaker:and things like that. Everything get sized up.
Speaker:Again, you're still looking constantly at the capabilities, have we got
Speaker:enough people from networks to plumb this in? It's not just a data,
Speaker:I just need data engineers, it's all sorts of disciplines across the board.
Speaker:And I think that's where you start, basically. You at the moment obviously
Speaker:have moved from on premise to cloud. I suppose the biggest project in
Speaker:more recent times, I'm guessing. What sort of challenges has that thrown
Speaker:up? Have there been things that you've had to consider differently than
Speaker:perhaps having a non premise sort of set up? We're still halfway through
Speaker:that at the moment, there's a lot of work, there's always something to
Speaker:do, but security is the big one, I think.
Speaker:With our on premise networks, there's a little bit of a safety blanket.
Speaker:It's not all safe. Do you know what I mean? If someone wants
Speaker:to get in, they can probably get in. But it's a lot easier
Speaker:to deploy things knowing that you have things ring fenced and people won't
Speaker:be able to exploit them. So security is a massive one,
Speaker:because when you push it up into the cloud, it's so easy to
Speaker:just make something public by mistake and things like that. So we do
Speaker:spend a lot of time on security and making sure everything's encrypted in
Speaker:transit and encrypted at rest and all this.
Speaker:You do find though, a lot of the cloud services and all this
Speaker:kind of happens by default, but you still gotta check it out and
Speaker:make sure it's within your standards, if you like. So security is a
Speaker:massive one. The other things to consider is governance, which changes drastically
Speaker:as well when you're going to cloud because it's so easy for things
Speaker:to get out of control. Because of the power of cloud,
Speaker:essentially is something that I've led that I haven't learned before,
Speaker:is that you can push date up into cloud and give everyone access,
Speaker:but God, it becomes a massive mine field and it turns into the
Speaker:Wild West if you're not careful. So you really have to basically keep
Speaker:an eye on that governance, making sure that people have only got access
Speaker:to the data that they need to do their day to day job.
Speaker:And I think that's very important. It's always important on premise as well,
Speaker:but with the power of the cloud and more people can do,
Speaker:it can really get out of control so quickly.
Speaker:There are a number of things that I consider, but one thing else
Speaker:I'd like to say is that if you've got good governance and good
Speaker:lineage of your current on premise systems, it becomes a lot easier to
Speaker:go to cloud. Because you understand how the data is getting used.
Speaker:It goes back to those personas again, who's using it, it's all documented
Speaker:and things like that. In that case, it's a lot easier to move
Speaker:and replicate and put those same controls in. So it's important to get
Speaker:it right wherever you are. So yes, cloud does make a difference,
Speaker:but you really should be putting these rules in place throughout,
Speaker:no matter where you are. I think it's interesting you say that because
Speaker:obviously there's great opportunities with cloud, the speed and
Speaker:the openness, as you say, but yeah, a lesson learnt from something in
Speaker:the cloud is probably a harder lesson than if it's actually still just
Speaker:an internal system, a bit more closed down. Have you had to do
Speaker:a lot of training or sessions with your team to get them thinking
Speaker:differently with those points and those challenges? So there's always technical
Speaker:training, that's a given, but like I said, it's having that engineering
Speaker:mindset. Admiral's a very young company, we've got a lot of young people
Speaker:working there, come straight out of university and we can train them up
Speaker:and things like that. They need to understand the repercussions of things
Speaker:going wrong. We have processes and things like that to control that,
Speaker:but we'll generally give that insight into failing systems, a lot of the
Speaker:time. Because you have to code to have resilience and make things robust.
Speaker:So quite often we'll have reviews of code go, "Right, okay,
Speaker:what happens if that thing goes wrong?" And it goes, "Oh, well,
Speaker:it re tries." "Okay, alright, that's great. It re tries. But does it
Speaker:clean up the message made when it went wrong?" It's things like that.
Speaker:So it's having an engineering mindset. A lot of what I try to
Speaker:put into context is that, look, would you put an aircraft in the
Speaker:sky if you built it like that? What if something failed on the... Is
Speaker:there a backup system? How does it get maintained, how does it get
Speaker:checked? There's warning lights and things like that that you get?
Speaker:Have you got the same warning lights and things in that in the
Speaker:code and the data pipelines that you've put into this system?
Speaker:That's why it's important to monitor, because everyone's brought up with
Speaker:a CICD mentality, and a lot of CICD doesn't work when you have
Speaker:a data project going on. And this is why you have data ops
Speaker:instead of dev ops, because it's not only the code that has to
Speaker:go through that process, it's the actual data itself. And you can get
Speaker:bad data, so your code may be the best code written in the
Speaker:world, but you get some bad data into it and everything can fall
Speaker:down. And so you have to have these checks, you have to have
Speaker:this live monitoring and things like that. And that's where I find machine
Speaker:learning is actually coming into play, because you can actually apply machine
Speaker:learning to that monitoring, to actually go, "Oh, this premium looks low
Speaker:today," or on this percentage of measures or things like that.
Speaker:You can actually get machine learning to actually invoke that alert and
Speaker:then somebody can react to it and things like that. So although we
Speaker:look at machine learning as, "Oh yes, gonna help this, gonna make... Better
Speaker:profile my customers," and things like that, there's so many more applications
Speaker:throughout the whole data pipeline, if you like, and that data flow,
Speaker:that I think sometimes people ignore that. It's like turn the skills on
Speaker:its head and use the machine learning to help make your systems as robust
Speaker:and get your data as clean and as quality as possible.
Speaker:I think it's interesting you say that, 'cause I was just about to
Speaker:ask you about machine learning and ethics. 'Cause you'd mentioned that.
Speaker:And I think there is that, yes, you can think about it from
Speaker:a technology point of view and stuff moving very, very quickly,
Speaker:but that role of a person in there just checking and putting in
Speaker:the perimeters of kind of, "This is gonna be safe," or, "This is
Speaker:ethical," or, "This is actually what we mean by good data,"
Speaker:and actually the checks and the processes, is still very important.
Speaker:Are there any particular trends or ways of working that you can really
Speaker:see how that would jump start or that would really, really push a
Speaker:different way of thinking? Well, there's a number of new ways of rolling
Speaker:out data teams and rolling out data architectures and things like that,
Speaker:one of them being mesh. There's this concept of a mesh architecture and
Speaker:God, you really gotta be mature for some of this stuff,
Speaker:and you really have to have those engineering principles embedded.
Speaker:Because the whole concept of mesh is that you decentralise basically,
Speaker:okay, and you farm that out to value streams and things like that.
Speaker:Now, traditionally it's been like a data warehouse team and they've done
Speaker:all your data and they've ingested it and things like that, but they
Speaker:become a blocker, they become a blocker because they can only prioritise
Speaker:so much, and someone over here in, I don't know, household insurance wants
Speaker:to do some stuff and they can't. So what you try and do
Speaker:is you try and farm out those capabilities to other value streams,
Speaker:but that can be difficult unless you've got some blueprints for the central
Speaker:governance because you want everything to work the same way
Speaker:so you can maintain it. There's definitely some things that in the delivery
Speaker:space that are being suggested. Everyone talks about mesh. I went to Big
Speaker:Data London this year, and everyone was talking about mesh, but no one's
Speaker:really nailed it. And it's the same thing with machine learning,
Speaker:you get our pricing analyst thinking, "Right, okay, we've got a good machine
Speaker:learning model here, we've trained it well. Right. We've got to get it
Speaker:live." But because it's hooked into the whole quoting process and things
Speaker:like that, you really gotta be careful that that is correct.
Speaker:And so it can take a long time sometimes for these things to
Speaker:get live, and this is where you have things like ML Ops springing
Speaker:up. Okay? So it's not about the model itself and the machine learning
Speaker:model itself, it's more about how quickly can you get that value into
Speaker:the mainstream. And this is what ML Ops is giving us,
Speaker:and essentially it's a set of governance rules really,
Speaker:that I think data scientists aren't used to
Speaker:dealing with that. Data scientists, they're in their own little world,
Speaker:they're in a bubble and things like that. Now they have to think
Speaker:in terms of engineering and what matters, "What if this model goes wrong?"
Speaker:That sort of thing, they have to think about that. So everything has
Speaker:to be thoroughly tested. And that takes time. So although you've got this
Speaker:whole, "Oh yeah, ML Ops, AI," and things like that, you have to
Speaker:determine the risk of certain things, and if you start
Speaker:relying too much on it without having that whole framework around it that
Speaker:gives you that robustness, you could be in a lot of trouble.
Speaker:I was gonna say the kind of reporting and that, like you say,
Speaker:having different people with different roles and responsibilities across
Speaker:that sort of data pipeline, if you like, how do you go about
Speaker:making that reporting and that kind of auditing and the reporting seamless
Speaker:for everybody at the their worst? 'Cause it's such a big thing,
Speaker:isn't it? That actually, you could break it down and make it so you're working
Speaker:in pockets, but that actually would completely blow the overall name of
Speaker:having this kind of thing. Have you got particularly with Admiral, ways
Speaker:of working that you think you've nailed that? Or have there been questions
Speaker:that you've along the way thought, "Actually, if we ask this with these
Speaker:certain people, we are gonna have a better outcome at the end."?
Speaker:I think, again, its mindset is a lot of it. So yes,
Speaker:you can have these different teams and they're all working on different
Speaker:things, but again, we go back to why.
Speaker:Okay? Why you're doing this? You're an essential part of the cog...
Speaker:An essential cog, sorry, not essential part. That is basically gonna deliver
Speaker:this value at the end, and you are an essential part of it.
Speaker:I think if people understand that, they generally
Speaker:start to go out and communicate. And I think that's what it is,
Speaker:it's all about good solid lines of communication. And obviously that's what
Speaker:Agile gives you with the stand ups, your retrospectives and things like
Speaker:that. This is what I'm heavily involved in, is that something is not
Speaker:working, you change it. That's the the whole point, nothing's etched in
Speaker:stone. If something's not performing or something went wrong at the end
Speaker:of the last delivery or whatever, you change it, and communication is a
Speaker:big factor in that. So when you have teams and you have them
Speaker:in those silos and they're not communicating, you'll find that they're not
Speaker:gonna line up at the end of your sprint or whatever your delivery
Speaker:mechanism is, you have to change it. So
Speaker:we get people to work together. That's the key. And communications is hard
Speaker:when we're all working remote and things like that, but
Speaker:it's still achievable, very much so. But it is about comms.
Speaker:No one wants to come in and write something bad that doesn't work
Speaker:at the end of it. But they didn't know, they were just sort
Speaker:of carrying on. So it's important to get strong leaders to bring people
Speaker:together and actually go, "This what we're doing. This is how we're doing
Speaker:it. Everyone clear what they're doing?" "Yes." In the morning and stuff
Speaker:like that. "If you're not, come and see me, we'll go and speak
Speaker:to someone else," or whatever like that. That's kind of how it needs
Speaker:to work. Within the data teams at Admiral, I think we've cracked that.
Speaker:I think we have. We try to modularise as much as we can,
Speaker:so somebody doesn't write a code for a specific things, they write it
Speaker:for many things that can basically be used by another team.
Speaker:So we try and write these modules and frameworks
Speaker:for people to work in, rather than, "I'm just writing code." And I
Speaker:think that's what helps. But there's also a lot of unit test checking
Speaker:and things like that, there's a lot of automation that you can basically
Speaker:put in, and we do this at our retros and things are like that. We went,
Speaker:"Right, that went wrong, didn't it last time? Okay, what we're gonna do
Speaker:about it?" "Right, we put an automated check in for that."
Speaker:So as soon as someone deploys something into a test environment,
Speaker:runs a host of unit test scripts or things are that and goes,
Speaker:"No, that ain't gonna work with this," so we can build that in.
Speaker:And we quite often do that. Don't capture everything. 'Cause it is hard.
Speaker:Data is hard, I'll tell you that now. But if you have that
Speaker:mindset and that ways of working, and that monitoring mindset and that robustness,
Speaker:that engineering mindset, you can overcome it. Do you think that the governance
Speaker:is kind of more stable or more solid because of that understanding of
Speaker:data engineering mindset? With the governance? Yes, I think with governance
Speaker:we've been in a better place than Admiral has ever been.
Speaker:We bake things in, and again, our engineers think about it because they
Speaker:want the best kudos for their bit of code.
Speaker:So the only way you're gonna do that is get people actually using
Speaker:it and using what the output of that code has produced.
Speaker:Now, if you've got good governance that has put that into a data
Speaker:catalog 'cause you've coded that up or is automatic or whatever,
Speaker:then people will find that data and they'll use it. Okay? And this
Speaker:is another thing about, probably dovetails into the working in silos.
Speaker:You develop something for a specific thing, I know it has to happen
Speaker:sometimes, but if you can write it so that it's open ended and
Speaker:basically other people from outside the intended recipient can see it in
Speaker:the data catalog, and within reason, they're allowed to see it and all
Speaker:that. There's security and stuff like that. They can discover through a
Speaker:data catalog and basically go, "Oh, I didn't realise someone had done that,
Speaker:calculated that over there. I'll go and use that." And then they attribute
Speaker:to the data catalog themselves by saying, "I am using this for this
Speaker:purpose." That kind of mentality then grows, because everyone used to see
Speaker:data governance is like a blocker, but essentially there's ways and means
Speaker:of doing it where you can bake it into that whole delivery process
Speaker:and actually make it an enabler and not a blocker. I think it
Speaker:took a lot of companies a long time to understand that,
Speaker:because I think they just thought of it is a regulatory thing and,
Speaker:"Oh we've got to do this, we're ticking boxes here and there,"
Speaker:and stuff like that, but I think you do it properly and it
Speaker:is so valuable and it can aid operational efficiency. People have gotten
Speaker:better decision making because the quality of the data has increased.
Speaker:I used to go into meetings and some guy used to go,
Speaker:"Well, this is what we sold this product for yesterday," and another guy
Speaker:would go, "Well that don't match my figures." And you're like,
Speaker:"Well, who's right?" And now with this data governance you've got a gold
Speaker:standard, and now you got the guy going, "I got these figures."
Speaker:"Oh, well I got these figures and they're different, but mine
Speaker:have a tick on them, okay? They've been checked, so they've been gold
Speaker:standard, this is actually what happened." The other guy's got no response
Speaker:against that. And this is the type of things that it can enable,
Speaker:but I think it takes a while for people to see that because
Speaker:they do see it still as a blocker. But if you know data an you
Speaker:know how it's used, you can really see the advantages of having that
Speaker:strong governance. And I think what you're saying there as well,
Speaker:is that actually just the quality of the reporting across an organisation
Speaker:and also out to customers, will be better because you know that you're
Speaker:using correct data and you're using it in the right way,
Speaker:and it is standardised. It's not just the reporting, you look at the
Speaker:AI and the machine learning and the data scientists and things like that,
Speaker:it used to be they'd have this, "I just want all data. All
Speaker:the time, as quick as you can, all data, all format, and I
Speaker:want as much data as you can give me," and things like that. I
Speaker:think the emphasis is slowly drifting away from that now to actually go,
Speaker:"No, I want quality data." Because if you're training your models on bad
Speaker:data, you're gonna get bad decisions, okay? So I think the emphasis has
Speaker:now moved from, "Give me as much data as you can,"
Speaker:to, "Just give me clean data." And I think that's an important step
Speaker:because I don't think I could say that to you two years ago,
Speaker:but I think now there's a machine learning model for everything,
Speaker:I don't think anyone's really pushing the boundaries of machine learning.
Speaker:Yeah, don't get me wrong, at Admiral we have custom ones and things
Speaker:like that, and probably data scientists if they heard this in Admiral they'd
Speaker:probably have a few choice words to me,
Speaker:but to be honest, someone else has done it. Someone else has done
Speaker:something very similar somewhere. You can pretty much buy a model off the
Speaker:shelf, okay? And everyone is doing it, it's all like self service.
Speaker:I can go on to Google's marketplace and I can buy
Speaker:load of models and get the result I want.
Speaker:What the difficult thing is the data engineering to get the data quality
Speaker:up to a level when you can train that model, and I think
Speaker:that's where the emphasis has changed there as well. So it's not just
Speaker:the reporting, it's definitely for other applications as well.
Speaker:So with the teams that you've got there, it sounds like obviously there's
Speaker:a sort of sharing of knowledge and sharing of code, and you have
Speaker:obviously done a project and read all of... Been involved with it. What's
Speaker:it been like working with external consultants? Has that challenged the
Speaker:thinking? Has it kind of steered the project?
Speaker:You definitely learn from them. Because at the end of the day we're still
Speaker:quite immature in some ways, Admiral's cloud journey.
Speaker:Things take a long time and things like that. And you need that
Speaker:external support, you need good people who have basically done this before
Speaker:and know what the pitfalls are. You're still in charge, but you're in
Speaker:charge of the outcomes, "These are the outcomes I have. Okay?
Speaker:It needs to work like this." People like Red Olive as well, they come
Speaker:in, they're classic engineers, they're trained in first principles.
Speaker:And they will go, "Well James, if you do it like that,
Speaker:then that could break there." So that's great advice. And so they're thinking
Speaker:the same way, which is fantastic. You don't have to educate them in
Speaker:robustness and maintenance and things like that. You just learn a lot from
Speaker:them. You go, "Oh, okay. Well, we did some similar for the couple
Speaker:of guys two years ago that did this. Would that work?"
Speaker:And it's just bringing that experience to it. Because cloud is an experience
Speaker:and you really have to sort of think differently. And especially when you're
Speaker:on prem, you design for performance and when you get to cloud,
Speaker:you're actually designing for cost. Because it's so easy to just throw another
Speaker:100 processes at it and a shed load of memory, but there's a
Speaker:cost implicated with that. So you still gotta be super efficient,
Speaker:because otherwise those costs can run out of control and things like that.
Speaker:And it's good that our external consultants and things like that, they've
Speaker:all fed into that. They've gone, "Well, you can do it this way,
Speaker:it's robust, it's really good, but it will cost you this and you
Speaker:have to figure that out," and that can be very difficult.
Speaker:But having people that have done it and things are like that,
Speaker:that can really guide you essentially on that design from a cost perspective.
Speaker:And also from a robustness perspective. It's a balance, isn't it?
Speaker:So yeah, they've been really good. Well, though we've been talking data,
Speaker:we've actually been talking about communications and people throughout this.
Speaker:For people who are now coming into the data industry, what
Speaker:skills would you recommend that they have? And what kind of things would
Speaker:you be looking for for somebody joining industry now?
Speaker:To be honest, we take on a lot of grads, a lot of
Speaker:young people that may want to get into this sort of thing.
Speaker:The key skill or two key skills I'll put down from a technical
Speaker:perspective, SQL and Python. Because pretty much everything runs off those
Speaker:two things. Or is interfaced with them in some means.
Speaker:But what we really look for is... You can teach those skills.
Speaker:If you've got a bit of a mathematics or science background or things
Speaker:like, even if you haven't. But you know how an Excel spreadsheet works,
Speaker:you know how to look through it, you know to organise data and
Speaker:things like that, those are the key sort of technical skills we look
Speaker:for. But we look for behaviours, and we do a lot of competency
Speaker:based type interviews where we look at, "Give us an example of a
Speaker:time when something went wrong?" And that may not be in data,
Speaker:or the business could be something completely external, but we look at the
Speaker:mindset about how people handle it. So there's a little bit of psychology
Speaker:in there and things like that, but what we're looking for is people
Speaker:who are conscientious, who care about what they're actually doing, and that's
Speaker:important. So all the technical stuff, you can learn, you could go on
Speaker:100 YouTubes there, right? And you can know concepts better than us.
Speaker:But when it actually comes to implementing it and your day to day
Speaker:job, I think your psychology and how you're made up has a lot
Speaker:to do with it. So we do look for a lot of people
Speaker:who are very driven, as I said conscientious, really care about what they're
Speaker:doing, and you can make a really good start to mould them into
Speaker:a classic data engineer. So that's what we look for.
Speaker:My last question for you, what is the most exciting thing that you're
Speaker:most passionate about for the data, pushing forward into the future?
Speaker:Data automation. Because I know we talked about machine learning and AI
Speaker:and things like that, and those things are great and very impressive and
Speaker:things like that, but I think unless you're one of the top players,
Speaker:they can be very difficult to implement on scale.
Speaker:But what I'm interested in at the moment is a lot of metadata
Speaker:sort of data driven pipelines, whereby, rather than deploying code you're
Speaker:making more configuration changes and things like that. So it's the automation
Speaker:of it all from a delivery point of view,
Speaker:I like the idea of. Some people have had a go,
Speaker:WhereScape being one of them, they're really good, but there needs to be
Speaker:a next level. So when we're talking about... We've spoken about things like
Speaker:governance, data cataloging and things like that, that type of thing needs
Speaker:to happen automatically. So we talk about automation in terms of automated
Speaker:testing or automated deployments and things like that, but what about actually
Speaker:automating the actual development? Because a couple of years ago, a lot
Speaker:of people just hand coding things, and there's a lot of now low
Speaker:code solutions where you still need to have the technical ability and know
Speaker:what's going on underneath, but you can template a lot of it.
Speaker:So I can automatically template that it adds, I'm bringing new data in,
Speaker:it adds to this data catalog. Or I publish a report, it automatically
Speaker:gets added to this data catalog and things like that. So that's what I'm
Speaker:really interested in, because everyone goes after the really exciting stuff
Speaker:and things like that, but nobody thinks about the implementation so much.
Speaker:Because I live in that space and all my problems have been in
Speaker:that space, that's what I kind of get excited about now,
Speaker:can I have a team of three people that can deliver something in
Speaker:say a month, rather than a team of 10 deliver it in six
Speaker:months? And that's kind of where I'm going at the moment,
Speaker:because we're up to our eyeballs at work and it never stops.
Speaker:And you can only grow so much before it becomes hard to manage.
Speaker:But if we can deliver quicker and automate more and maintain our quality
Speaker:standards and our governance standards and things like that, that's what
Speaker:excites me. Now, that may be a bit of a boring answer,
Speaker:but that's the honest truth. A fascinating take from James on the importance
Speaker:of people, teamwork and communications in data projects. Join us for the
Speaker:next episode of Fibonacci, the Red Olive Data Podcast, where we'll be joined
Speaker:by another data expert sharing their thoughts on latest trends in AI and
Speaker:big data, along with some great hints and tips.
Speaker:Make sure you subscribe to Fibonacci, the Red Olive Data Podcast from wherever
Speaker:you get your podcast, to make sure you don't miss it.
Speaker:That's all for today. Thanks for listening. I've been your host,