diff --git a/content/sessions/2023/mini-summits/Apr/Governance/Behaviour-Change-and-Awareness.md b/content/sessions/2023/mini-summits/Apr/Governance/Behaviour-Change-and-Awareness.md index 9f7609e8e70d..f602096801f4 100644 --- a/content/sessions/2023/mini-summits/Apr/Governance/Behaviour-Change-and-Awareness.md +++ b/content/sessions/2023/mini-summits/Apr/Governance/Behaviour-Change-and-Awareness.md @@ -26,6 +26,934 @@ zoom_link : https://us06web.zoom.us/meeting/register/tZIud-mprD0vE9xWCJLSzFi8 --- -## About this session +## Session Transcript: + +Dinis Cruz - 00:00 +Hi, welcome to this open Security seminar session in April 2023 and we're +going to have a very cool panel around behaviour change and awareness. +It's definitely a topic that I'm quite passionate about. I think there's a lot +of things that are related to it and a lot of activities that we can do that +make a big difference. So I'm Dennis Cruz. I'm the chief scientist of glass. +All I'm also the CEO of Holland Barrett, and I think it's about creating a +system and then everything tends to fall into place with it, but also the +reaction and the response element of it. I'm looking forward to explore +maybe tim, over to you. + +Tim Ward - 00:40 +Yeah. My name is Danusia Rolewicz and I specialize in human factors, risk +analysis and management, and particularly security awareness and culture +change. I suppose our whole business is all about trying to take a more +behavioral approach to security awareness and move us away from those +traditional approaches where really it's just a bit of a tick box exercise, +but we can come back to talk about that in a second. + +Danusia Rolewicz - 01:02 +My name is Donna Shara Levic and I specialize in human factors, risk +analysis and management, and particularly security awareness and culture +change. I am contracting at the moment around various organizations to +improve security posture and security culture within the organization, +most recently Apple. I've found that there is a shift in the industry at the +moment towards more empowering and pathetic approaches towards +staff, more employees, staff people rather than users. I think that's having +a really positive effect. I think the more that we focus behaviour change +and empowerment of staff rather than using that tired, weakest link +expression, the further we can get really. + +Dinis Cruz - 02:03 +Cool. All right, so who wants to define what do you mean by behaviour +change and awareness? + +Tim Ward - 02:11 +Pikey, do you want to go first and easier? + +Danusia Rolewicz - 02:17 +I think we all know what behaviors we want to change, right? There's a +lot of talk about click rates, but really when you're looking at behavior +change, you are also looking at the behavior and the dynamic within the +company or the organization, like trust dynamics. Rather than focusing on +things like click rates, you would try to empower your staff, foster a +culture of trust within the company, where rather than focusing on the +click rate, you focus on the reporting rate. You give the credit and +celebrate the wins for those positive outcomes and for those achievements +that often fall by the wayside. When you look at a traditional security +audit and the metrics that go with that. + +Tim Ward - 03:08 +I agree with that completely. I think also just going back to your point +about there being a shift, I think we've traditionally done security +awareness training. It's just been a thing people did partly because they +got that there's a human element and we have to help people know how to +deal with the threats. Partly just because it became a compliance +checkbox. If you step back and think about why you're actually bothering +doing it, well, you go back to that kind of famous stat of 80, 90% of tax +start with a human. If you actually want to reduce that risk, then +awareness doesn't reduce the risk. It's just people are aware, great, well +done. You've got to be trying to change behavior. You've got to be trying +to build secure habits. It might be just reinforcing habits that are already +taking place. I can see Janet just successfully come onto the call. + +Tim Ward - 04:03 +Because if you really want to reduce risk, then it does come down to +different behaviors. + +Danusia Rolewicz - 04:10 +Yeah, that's right. I can see Jeanette's just successfully come onto the call. + +Janet Bonar Law - 04:17 +My sincere apologies. I am technologically incompetent and couldn't find +the right setting in zoom. I lived my life in teams, so zoom is not familiar +to me. + +Dinis Cruz - 04:34 +Just give me a quick intro, Jeanette, and then give us your views on +behavioral change and awareness. + +Janet Bonar Law - 04:41 +Okay, so I work as a lead specialist in people's cyber risk management at +the Coventry Building Society. We take a very broad view of what +awareness is in the Coventry and we concentrate more on the end, which +is that our function is to drive down risk through a secure behavior +change. In order to do that, we'd like to think of ourselves more as human +hackers. We try to get people to do the things that we want them to do by +wanting to do them. + +Danusia Rolewicz - 05:25 +I think that's a really good point, Janet. I think if you make things easy +and accessible for people, then you're more likely to get them doing what +you want rather than taking a top down approach to controls and +behaviors. + +Dinis Cruz - 05:45 +We can use science for this, right? Because there's already data that shows +that, like you said, the more the users engage, the more you make sense to +them, the more their behavior is safe, the better the outcomes, the better +the return on investment that we have. + +Tim Ward - 06:03 +I think there's a kind of a self efficacy thing, isn't there? If you kind of help +people then get better at things, then you get that kind of positive cycle. I +mean, I think one of the interesting because once you're talking about +behavior, it does make you go to the academic stuff and think about +behavioral theories. One of the interesting one is protection motivation +theory that has you thinking about how do you make a decision when it +comes to dealing with risk. Part of that is around understanding the +threat. That's the only bit we've tended to focus on as awareness +practitioners in the past. The other bit is the coping, like, do we know +what to do? Can we cope? I think that's the self efficacy bit. If we give +people more information of how to cope with the problem. We don't +necessarily have to scare them quite so much a bit, maybe, to help them +understand the risk, but we should focus more on the coping, like giving +them an easy action to do. + +Janet Bonar Law - 06:58 +Yeah. + +Danusia Rolewicz - 06:58 +Also on the communication side, I think I'm muted. Am I muted? No. Also +on the communication side of things, to really communicate what's in it +for them, rather than going in and saying, oh, well, we have this objective, +we need to improve our posture by this or that statistic, how will this make +your life easier? Why should you be doing this? I think a lot of security +communication, forgets the why and also just the connection with people. + +Dinis Cruz - 07:40 +What I like about the word behavior is because ultimately what we want +to create is a certain of behaviors in organizations. I'm also a big fan of +taking incidents like P three S and P four S and elevating them to P ones. +In fact, again, I was mentioning that the presentation before this one was +really cool because the guy that presented it was showing really good +examples of real world incidents who costed like from half a million to +multimillion pounds to the companies. He was showing all the things that +the attackers did. In a way, when you look at list, you see this is every one +of them was a failure in behavior in the organization. Something should +have detected that there should have been behaviors in the organization +that are looking for this. They're paying attention that look for this. I like +the word behavior because we can measure that and we can basically say, +look, we want the organization to behave in a certain way and we want +the organization to behave in ways that then allow us to do X-Y-Z. + +Dinis Cruz - 08:38 +When you look at the real world data that either backs up or not the +behavior that you want. + +Danusia Rolewicz - 08:45 +I think it's interesting that you're saying a failure in behavior, Dennis, +because like Tim mentioned earlier, the Gartners of the world talk about +90% of incidents could be attributed to human failure. That I mean, +humans are responsible for every point in the chain. It human failure to +correctly configure a system? It human failure to perhaps misunderstand +the culture of the organization and impose a policy or a procedure that's +probably unrealistic? Or is it the end user inverted comments that failed? + +Tim Ward - 09:37 +Yeah, I think it's an interesting one, isn't it, of that kind of risk of getting +into that blame, starting to talk about repeat offenders and things like +that when it comes to clicking links. I suppose we should recognize that +we are giving people computers with email that has links in and half the +time we expect them to click them and sometimes we don't. So it needs a +bit more thought. I suppose I like doing this point around measurability +because I think that is quite important and that's if you are thinking about +behavior so old traditional awareness, all you can really measure is have +people done it or not? And you get a completion rate. It's like, well, what +does that mean? That doesn't tell you anything. Obviously, measuring +incidents is far too lagging an indicator. With behaviors, you can do things +like that. You can get really observable ones. + +Tim Ward - 10:32 +One of the things we're trying to do with our piece of software red flags is +to try and actually get right there and measure, well, how often are people +doing things like plugging in a USB or uploading files? When you +intervene, are you seeing that change? There are lots of ways to make this +more measurable. As Deniser said earlier, start with reporting rates, not +necessarily quick rates, because reporting is a really positive behavior. If +you've got a culture where that's a high number, then you're really onto a +winner. If not, start building up that as a focus. + +Danusia Rolewicz - 11:04 +If you have a Champions program or thinking of starting one, then there's +all sorts of things that you can use to join up the dots to the reporting rate. +How has this changed? How many people have become involved in +Champions programs? How many meetings have there been? How many +times have they discussed security with their team? Join up the dots of all +of those metrics and you can get an idea of why there's been a change in +reporting rates, for example, or other measurable behaviors. + +Janet Bonar Law - 11:44 +That's a really good point. This idea that we have incidents because of +some kind of single point of failure and this hideous metaphor, which I +loathe about being the weakest link. Humans are the weakest link. Well, +that's not true. Humans exist in a whole ecosystem of vulnerabilities, some +of which are behavioral, some of which are technical, some of which are +cultural. It just so happens that in a particular incident, one step, because +it is never a single point of failure, one set of events happened which +highlighted weaknesses in the ecosystem. And we need to recognize that. + +Danusia Rolewicz - 12:32 +I think that's a really good point, Jeanette. I think just saying the word +weakest link or failure or whatever equivalent thereof, it doesn't send the +right message to the rest of the organization. Why would anyone want to +think of themselves negatively? You would never engage anyone by telling +them a negative thing about themselves. If we have that behavior +ourselves, of using language like that within our team, then how likely is it +that's going to be perpetuated? We're going to spread that message +consciously or unconsciously through the organization and create apathy, +really like, do we want to motivate or demotivate? Really discourage +language is so powerful in that, isn't it? + +Dinis Cruz - 13:27 +I think I could be a little controversial. I would say that 99%, if not 95% of +what we call human error are system errors, right? We put the human in a +place that it can create a huge amount of problems in a way that you +don't understand the side effects. So that's a system error, right. + +Tim Ward - 13:52 +I was going to say one of the quite important things when you start to take +a behavioral approach is to think about what you're going to do with that. +Because you could take a behavioral I mean, phishing simulations +arguably are a behavioral approach, but they have a tendency to then +result in a bit of a stick that says you failed, we're going to tell you off, +thing. What you need to be doing is thinking about what that data can tell +you. To your point, Dennis, it can tell you that your business process is +putting people in a position where the only way they can meet the deadline +the boss is set is I know, Gmail, and send something to their Gmail account +at home. The behavioral data can give you really interesting insights. It +does take a bit of a mental shift, I suppose, to think about what are you +going to do with that information? + +Tim Ward - 14:36 +How are you going to take a more behavioral approach so that it just +doesn't become a consequences culture where you're just telling people off +when they do things wrong. + +Dinis Cruz - 14:44 +I think it's about what question you're asking, right? For me, I always +think that the users should have common sense, right? They should really +understand how to report things that they observe that are a bit weird and +they should be rewarded by positive behavior, right? It means that in a +way, you want over reporting, right? Because in an interesting way, over +reporting is a side effect of your users actually thinking, oh, this is a bit +weird. Oh, this is not very good. Right in the middle of it, there should be +the real scenarios, but also it gives you good data. It allows you to +understand, are you an organization that has a huge amount of problem +with that kind of activities, or which division of your organization has lots +of problems with phishing or abuse and what other divisions don't have, +right? You can look it again at the incidents or the practices and see what +do we have in place to prevent that? + +Dinis Cruz - 15:42 +I feel that again, I like the behavior because behavior allows us to look at +incidents and going, why didn't we detect this here instead of there, right? +Who saw this in all the activities and didn't raise an alarm bell that +suddenly you only got raise an alarm bell because XYZ had happened, +right? You can think about what technology you put in place and what +solutions you put in place so that in a way, when something happens, you +have immediate understanding and reaction about it because the user +clicks on a link, the user doing an action, a stressed engineer doing +something that makes a mistake. It's not at fault. We put the engineers in +that location, right? We put users in that environment. They not at fault, +right. They're just behaving if anything, we should be celebrating the other +99% of the times that they saved, but they didn't do it. + +Dinis Cruz - 16:38 +Right. It's we have this thing where we force the users to drive a car down +the mountain, right, with no rail guards at a crazy speed, and then we +complain about the ones that fell over, right? Actually, actually you should +be celebrating the ones that made it true. Right. Because we gave them a +really dangerous environment to operate in. + +Tim Ward - 16:59 +It does raise some interesting questions about how you think about +security. You can have that traditional approach. Well, everything's going +to be really secure if you lock everything down and control it. That's +perhaps quite an old fashioned approach to security, but that was the way +things work. We can lock everything down, it'll be fine. Turn off the +Internet, we'll all be secure. Actually, if you're starting to understand +behaviors at the level you're talking about Dennis, then you can start to +go, well, actually, these people over here, the way that they work, they +could actually have more human based controls. They don't need to be +locked down. We could just help them behaviorally and therefore your +whole business could work. It could access these tools that will add to +creativity rather than being that we're a bank, therefore we turn off the +internet. We turn off everything. + +Tim Ward - 17:45 +They can't do their home stuff, they can't plug in USBs. Obviously that's +going to have an impact on an organization getting talent to some degree. +I think you can create quite an interesting kind of trade off and say, well, +look, can we take a more behavioral approach? I think that's going to be +hard for security. Some security professionals will find that kind of, well, +hold on with this control. This is really black and white. It's really binary. +They can either do it or they can't. + +Dinis Cruz - 18:08 +I think you need to work first. I think you need to add behavior and risk to +it because they are very connected, right. It also you need to take into +account the threat agents and in a way the regulatory compliance that you +might exist. Right. But the threats are important. The threat agent is very +important because the threat agent determines the level of risk that you +have. If you're an environment that has highly sophisticated, super +efficient, like they can act crazy fast, then yes, you probably need to have +more security controls in place. Majority organizations, that's not where +we are, right? Majority organizations is criminals, is vigilantes, people +who disgruntled employees, even the users. Right. The biggest attacker +when you track this properly that the company has is the company itself. +We just don't call them cybersecurity incidents. Right, but the company +itself will cause a lot of damage to itself. + +Tim Ward - 19:04 +Yes. + +Dinis Cruz - 19:06 +And that's all behavior. Right. It's all behavior because we put users in +places that they can make a lot of those mistakes. And it's not even +mistakes. It's almost a margin of error. Right. If you force a team to +operate at a certain level of conditions, you will have, say, they operate +well 90% of the time, okay, there's going to be 1% mistakes. That's it. You +have to assume that is just almost the cost of that environment that you +set up in place. Right. That your behavior there determines and your risk +appetite determines what that percentage is. + +Tim Ward - 19:38 +Right, absolutely. I mean, I'd be interested in kind of Janet in your kind of +world of commentary, kind of how you go about taking a behavioral +approach. It about just walking through the risk register and saying, +what's the behavioral perspective of all of these things? Where do you +start? + +Janet Bonar Law - 19:58 +It's interesting, isn't it, that Dennis mentioned how risk and response to +threats varies across various pieces of an organization. For that very +reason, in the Coventry, we're moving to a more functionally aligned +pattern of delivering security awareness for wanting a better word so that +particularly high risk functions such as finance and the people team will +have a dedicated awareness specialist who becomes almost part of the +tribe. So it's like being a cultural anthropologist. You go in and you find +out how people are working and what it is they need to be doing and what +support we can put in to take away as much of the pain points as possible +in process and technology and then whatever remains the residual risk we +address through dialogue and behavior change. + +Tim Ward - 21:06 +Yeah, it's interesting because I think it does make you have to think and +take a slightly different approach because for example, if you want good +behaviors around passwords, then you could obviously be and obviously +we focus quite a lot on nudging, but you could be intervening nudging, +saying stuff around passwords. What are you going to tell them to do +instead? If you haven't got a good thing like an action you want them to +do, like you haven't got a password tool or a way of storing them, then +what are you nudging towards? It's much easier to nudge towards a good +behavior than just say, don't do that, without giving any guidance, well, +what do you want me to do? It changes the way you have to think about +things. + +Janet Bonar Law - 21:48 +I think it's interesting you talk about passwords because we have a +particularly horrible password policy, which is super complex and no +password manager to assist with that at all. + +Dinis Cruz - 22:03 +That's a good example of bad security. Right? + +Janet Bonar Law - 22:06 +Makes no sense. We have nothing to help. We have instead of a helpful +password manager, we have a super strict password policy that if your +password on audit quarterly is found not to be secure, you go on a very +serious naughty list and miss disciplinary action. That is shocking and that +needs to change. + +Dinis Cruz - 22:34 +I was in a very interesting meeting where I was saying that we should +weaken our password policy if they had MFA. I was saying if you have +MFA, then in a current setting environment we should have long password +expiration dates and we should allow for weaker passwords compared to +freaking crazy stuff, right? I'm not saying that you allow password one, +but I'm saying that it's about saying the single most important thing to do +in passwords today is multifactorification MFA. + +Danusia Rolewicz - 23:07 +That's like, how are you introducing MFA? How are you facilitating them +to get on board with MFA? Because I was in a bank environment in +another country and the procedure for customers to download the app +and go through all of the motions to get that meant that I heard at least +two different conversations where a customer had come in. It was quite +open concept layout as well, on top of everything in the bank. I could hear +two different people getting help from a clerk on how to set up their MFA, +but in such a way that they were exposing data to anyone that could hear +even a temporary password that the clerk suggested to them was +everything was audible. You can have this really amazing secure app, but +if people are going to find these workarounds and not be able to get on +board and to download it and to get onto it. + +Dinis Cruz - 24:19 +Then it's worth not usable. Absolutely right. That scales more, right, in a +way which actually takes to a point where you could only have behavior if +it's usable. That's what I like about, again, the word behavior because +you've all seen that picture of what's it called the gate, right? All the cars +go around it, right? Because in a way to go around it, again, I like +behavior because behavior you can answer why are we trying to do this? +Right? Why we're trying to do this to protect the users. Actually, again, +there's data that shows that sometimes if you make it more secure, it +actually becomes less safe because people will you actually push users +away. That's why I always talk about safety versus security. When I +presented the boards, I said my job here is not to make to not to have +incidents and not to be secure. + +Dinis Cruz - 25:18 +Our job is to be safe. And that's very different. Right. Going back, Tim, to +your point, you can make a great secure system that's connected to +anything, right? It might be very secure, but it's not very usable and it's +not very safe. It doesn't meet the risk profile of the organization. + +Tim Ward - 25:35 +Your definition of safe is kind of secure and usable. + +Dinis Cruz - 25:39 +Well, I would say it's aligned to the business risk appetite. Right. There's a +balance, right. + +Tim Ward - 25:44 +Appropriate security. + +Dinis Cruz - 25:47 +Here's the irony, right? If you live in a place where you have to have a +house with 25 locks and a machine gun and a guard outside, I can argue +that house has a lot of security. Right? There's a lot of security features on +it. Right. But which one is safer? That one or a house that you can leave +your door unlocked? Now remember that you can only leave your door +unlocked if you're in an environment, in a neighborhood that is safe. And +why is safe? Because there's a lot of other things that will do that from the +geographical location, from the neighbors. There's a lot of other things +that come into play, right, to allow you to leave your door unlocked and +also the kind of attackers you have there. I think you have to balance who +is attacking you, what's your requirements, why are you doing stuff, what +assets you have that should determine the security and the implications of +a compromise. + +Dinis Cruz - 26:39 +Clearly, if I have a pile of medical data and the pile of data is not really +relevant, the security requirements should be very different. Even the +definition of safety is very different from one to the other. Again, context +and behavior matters, right? + +Tim Ward - 26:57 +Yeah. I think arguably some of that is taking that risk based approach, +isn't it? Because even in ignoring behavior, I've been in organizations +where they've just had these well, if it's got data, it's got to be encrypted. +You're like, well, no, what's the threat? Like, what different data has +different kind of requirements? The saying comes down to where people +are involved, I suppose, is just thinking about what's the risk profile and +what behaviors are the ones that we want to see. You get into the whole +world of, okay, and then how do you get those behaviors? How do you +help change and steer people to work in those ways? + +Dinis Cruz - 27:39 +Can we now then move to the second word, which I really like, or third +actually, which is awareness. And I would just throw this. I think +awareness by the users is also crazy important. Sometimes we try to get +awareness by scaring, which is not always the right way to do it. I think +once the users understand what they're dealing with, they actually tend to +be very protective and they actually tend to do the right thing in how they +operate. + +Tim Ward - 28:09 +Yeah, I suppose we quite often go back to BJ Fogg's model in this because +from a behavior perspective and that has in it that element of motivation. +BJ Fox says that any behavior has got a bit of motivation, a bit of ability, +but also some prompt or trigger to act and obviously the tools around you +give you the ability bit and your knowledge. Also you've got to have that +motivation and some of that I think comes from the awareness you're +talking about, didn't it? You've got to help people understand that yes, this +is a threat and that it's relevant because if they don't have that motivation +they're not going to necessarily act. Slightly counter that you've also got +to expect that quite often people aren't going to act securely because +they're busy, they've got other things to do. Therefore when it comes +behaviour change and want to work as hard as you can to make the +easiest behavior the one you want and that's a great example when it +comes to passwords, how hard are you going to make? + +Danusia Rolewicz - 29:15 +Yeah, and equally I think going back to what you were saying Dennis, +about scaring them, I don't think it's ever really the right way. Fear, +uncertainty, doubt, FUD or EPPM, these are strategies that we can see +being exploited historically in marketing or by social engineers. Really, +why not focus more on the positive empowerment? We're pre programmed +to react to positive prompts more. Again, creating security messaging, +creating programs where we can talk and involve people and give them a +feeling of shared responsibility because it's their choice to get involved. +Once they have the buy in of being involved in for example, a Champions +program, then that buy in, that investment that they've made, they're not +just going to throw that away, that's going to be used much more +powerfully. I think something that Tim often talks about is social force. +Once you recruit a few people and they start talking about something and +what they're getting out of it, like how they're benefiting and growing, +then other people will want to join immediately. + +Tim Ward - 30:43 +Yeah, social proof can really be a great helpful thing to bring any behavior +and you can kind of do things like say well, so and so has reported this, +have you reported anything when it comes to fishing? You're just even in a +simple behavior you can kind of frame it like that. We're all social animals +so we want to be helping socially, we want to be seen to be doing what +other people do. There's other cognitive bus like reciprocity which is kind +of similar where you can well, the security team have done all of this work +blocking this stuff, we've given you this, we block this much, could you do +your bit? People kind of feel like they ought to give back a bit? Probably +depends how you word that slightly but there's some useful ways that you +can kind of encourage those behaviors. + +Dinis Cruz - 31:34 +As a Google strategy, which I just want to prefix that I've also seen a very +interesting. Paradox that when you have a supply chain that goes from +very small individuals to very large companies, you actually tend to have a +weird situation where the ones that are most secure are, I would say, the +large companies that have already regulation, compliance, bigger teams +who actually are able to take security a lot more seriously. Right, not more +seriously. They actually have very color resources and they understand +better. Also the smaller companies tend to have that because they +understand the problems and the side effects of cybersecurity in their +world. Sometimes it's the one in the middle that tend to be the worst ones +who tend to actually fall into you want to give us your question? Yeah. + +Speaker 5 - 32:24 +Basically, from what Dennis was saying about the appetite for security, +where a company might have all they might have a situation where they +feel that they're secure enough, but they might think that they have this +false sense of security mindset where they believe that they're too small to +be targeted, which in essence is like for a lot of people that might go after +cybercriminals, they prey on those, but they just don't realize that. It's just +hard to really get them in that mindset that, hey, actually you should be +taking things more seriously. And here's why. + +Tim Ward - 33:27 +It's a good point, Billy, and it goes back to that awareness point Dennis +was making, that there is a bit of need to help people understand the +threat and possibly of highlighting the risk. Not too much fear, but +because just to help people understand that they might not even being +targeted, but they're going to be collateral damage, potentially. Quite +often the way that attacks work is that they're not targeting anyone, +they're targeting just general weaknesses. They'll be firing this out across +thousands of kind of endpoints or people. If your small business hasn't got +some of the basics right, you're not the target, but you've become the +target because you had a week. + +Janet Bonar Law - 34:18 +Yeah. + +Danusia Rolewicz - 34:18 +Especially with this at ransomware as a service model and the outsourcing +involved in that anyone could be targeted just because why not? It doesn't +take very much, really effort on the part of the attacker, but. + +Dinis Cruz - 34:37 +I would follow the assets. If you go back again to behavior change, one of +the places where you can see huge amount of policy behavior is on the +third party supply chain where you start to see pressures from the more +mature companies downwards to do the right thing. To be honest, that is +correct because ultimately it's about understanding that you might be a +company that has quite large, but you depend on a chain of suppliers that +if they are compromised down here, that is going to affect you. Also that's +an awareness for the company because the company might think, well, we +are a small provider who would attack us. If they are aware, going back +to the second thought of the word, right, if they're aware that they can +actually cause a huge amount of damage and the asset or whatever +they're providing is actually a mission critical element of a bigger +company, then they need to take it where? + +Dinis Cruz - 35:27 +Again, I think it's that behavior that you can also drive by creating +economic models or social models. Because again, you talk about security +champions, but it's also to do with the companies who hold key elements +of the supply chain. You can push security downstream and also allow +good investment in security. Going to Billy's point or why should we do it? +What's the value for us? Well, guess what? If companies also can get +rewarded in the marketplace for having good security practices, for +having good security investment, which there's a direct correlation +actually between security investment and good engineering, right, and +good practices. Because most of the stuff that we want to do is common +sense, right? + +Tim Ward - 36:09 +Good governance is quite often means yeah. + +Dinis Cruz - 36:13 +There's a direct correlation even on potential due diligence. I always tell +the companies I work for get us involved in due diligence because we'll +lower the price, right? Because of acquisitions, because security is a great +way to do due diligence. Because you ask really good questions about how +things work, not that they work, it's about what actually is happening +under the hood. A good company, when you ask how things operate, +doesn't go to the whiteboard and start drawing it, right? That's crazy. +Immaturity right. They should know how things behave. Right. I think that +behavior is super critical across organization. I know who mentioned that +we hacked the organization, but that's how I think about this. We're trying +to find the ways to nudge the organization, to behave in the ways that is +on their own interest, to be safe and more efficient. + +Tim Ward - 37:04 +Yeah, I mean, that's an interesting kind of much higher macro level +version of the behavior, isn't it? It's the regulatory environment that the +kind of the market can provide to steer people. Obviously people like +NCSC have been trying to do that with things like cyber essentials, haven't +they? They've been trying to create a framework where, well, okay. Now +you do get, I mean, as a supplier, you get asked, have you got these +different frameworks? And that has a positive impact. I suppose there's +then an interesting debate about where regardless of all these +frameworks, where behavior becomes that entry point where people can +still do something that breaks the whole bottle. How are you going to help +those people? How are you going to help the individuals within those +organizations as well to behave? + +Dinis Cruz - 37:56 +Well, that's why I think we should be measuring behavior, right? Because +behavior, again, driven by incidents, is real. It's not theoretical. Basically, +when something occurs let's take Phishing as an example, right? If a +company is targeting with phishing, what happens? Do people click on +things and then hide it? Do people click on things and then report it? Do +people don't click on things? Right? What's the behavior that you want +from a security team? From my point of view, yes, it's better if they don't +click, but they're going to click on things, right? I much prefer somebody +who's comfortable enough to go to the security team and give us advanced +warning that something just happened than somebody clicking, going, oh, +I hope nothing went wrong, and then not give us advanced warning. We +actually had an incident that was like that. We actually went back to users +and says, look, the problem is not you did XYZ, the problem is you didn't +tell us that you just did that. + +Dinis Cruz - 38:55 +The fact that you did this, the fact that you went to a website and entered +your credentials, for example, we would have done that. I cannot look a +user in the eye and says you had fault when members of my team probably +could have fault for the same thing, right? It looked just like the real stuff +was very authentic. You cannot expect users not to fall for that. Once they +notice that something was weird, what they should have done is tell us +straight away because then we can make sure that the users of that +password is not problematic. Again, what's the behavior you want the +users do you want the users not to click on things and they want the users +to tell you and to be the your eyes and ears on the ground that stuff is +happening. I prefer the second one. + +Tim Ward - 39:42 +Yeah, we talked about it earlier, didn't we? This idea of understanding +what are the behaviors you want to see in your organization, then thinking +about how you can measure them. I think there's of just thinking about +what are you going to do with that? That you don't become so we try our +hardest to try and intervene with behaviors before or as they're +happening, or we use the behavioral psychology of ideas of like priming +and things to steer people before the event. I think there are obviously +tools out there that are looking at kind of post instant and post event and +behavior. There's just a worry in my head there that then you get into this +cycle of punishment with training where you're trying to change the +behavior by going, we spotted that you did this thing, now we're going to +come. However hard you try, I think it's quite hard for that not to be seen +as telling someone off and trying to correct their behavior. + +Tim Ward - 40:39 +Whereas I think if you can try and do it before or as it can be a bit more +looks like this might be about to happen. Are you sure you don't want to +do it this way and you can try and be a bit more back to January, bit more +gentle, a bit more supportive, empathic and help people. I suppose. + +Danusia Rolewicz - 40:58 +Even in the post click scenario I think it doesn't have to be about training. +It could just be a discussion starting off a dialogue that actually could go +into a security champion session or some situation where you could not +expose that person but use their experience to understand the triggers and +the context that are floating around. + +Janet Bonar Law - 41:37 +I would argue that in the post click situation that is possibly the worst time +you could choose to try and deliver any kind of training because they've +just clicked on the email, they're annoyed that they're being caught out +and they're quite upset about it and that is not a good receptive base for +new information. I would argue it's next to pointless. + +Dinis Cruz - 42:06 +I completely agree. We should be thanking them. The other thing is +training should be in context to what their day job is. Then it makes sense, +right? Or else it doesn't again doesn't secure. Billy, I had a follow up +question. Do you want to yeah. + +Speaker 5 - 42:22 +Basically, I spent of time as a truck driver. One of the things one of the +companies did was when went to pick up our load, we would have to do +our pre trip inspection. Sometimes they would hide gift cards in places +where we need to look for our inspection. Like $10 at the local coffee shop +kind of a thing. Just as a reward for doing your job. Of course if you don't +do it properly then well, you lose out. I'm thinking you could do the same +thing in the office space for some things like for instance setting it up in +the office where a policy where if you have a suspicious email send it to +suspicious@thecompanyname.com and then it will look at it, analyze it, +see if it is and then maybe give that person reward. Like hey, you protected +the company by reporting this suspicious email. + +Speaker 5 - 43:53 +Here's a gift card for Amazon or something along those lines. Maybe if +someone at the end of the year, if someone gets whoever gets the most +maybe gets some kind of a reward or something like that, just like a +recognition for doing that. I think that could put people in a proactive +disposition where they're more looking for these so they can report them. +Otherwise it's just like they just don't really care about them. Even if they +do see them as suspicious they might just ignore them or whatever. This +way they're being proactive and I think that would inspire others to do the +same when they see people actually getting rewarded for that. + +Dinis Cruz - 44:42 +Absolutely. + +Danusia Rolewicz - 44:43 +I love the suspicious atcompanyname.com address. That's amazing. + +Janet Bonar Law - 44:49 +I am going to steal that immediately. + +Danusia Rolewicz - 44:53 +There's a lot of things that I've seen around in different organizations of +gamifying security. One thing is like a Pokemon Go kind of approach to +reporting or flagging issues, which it depends on the type of +organizational culture that you're looking at because it might not be +appropriate for everybody, but I think what you're saying could fit in quite +nicely on that gamification shelf. + +Tim Ward - 45:24 +Yeah, I've seen it done quite well with fishing where it's less about tricking +and catching out and it's more about saying, right, we're going to run a +phishing competition. It's going to start really easy and then it's going to +get really hard and the people who get through to the kind of spot, the last +most difficult email will win a prize or something. You've changed it, +you've turned it on its head, you've made the whole thing quite fun and +interesting and people step. The only slight issue is the self selection of +people who are already probably the ones who don't need to know the +trading. You have to think about what you're trying to achieve, but it gets +people liking and talking about security. + +Dinis Cruz - 46:03 +Tim, you tell my question, because the question I was going to ask you is +that when you do a phishing campaign exercise, do you tell the users and +the It team or not? + +Tim Ward - 46:14 +So we don't tend to do them. We think that there are other ways to tackle +this, but I think it all comes down to what you're trying to achieve with it. +Because the reason we don't use I don't think the click rate is so variable +depending on the bait, that you have to think like, why are you doing this? +What are you trying to achieve? I don't think the training afterwards is +that effective, as Janet says. It's kind of, why are you even doing these +things? Maybe it's useful to understand the difference between different +sectors of your organization, like who's most prone to authority based +type attacks, in which case, no, you wouldn't tell them because you are +trying to get some useful insightful data which you can then use. I don't +know, janet, do you tell people? + +Janet Bonar Law - 47:05 +I don't tell anybody. I just set the campaign up, let it know that the box is +going to be busy and hit the Go button. Really? That's what I do. + +Tim Ward - 47:17 +Stand back. Like the blue touch paper. + +Danusia Rolewicz - 47:22 +There was one organization that I was at where people were actively +requesting phishing campaigns that was an indication of and it was +specific teams as well. I think that ties into what Tim was saying, how you +might get a group that really goes for that thing and that really wants to +prove themselves and then maybe the other groups or teams within the +organization, something else could work. Maybe developing a tailored +approach where not everybody gets hit with the same thing or the same +strategy could be useful. + +Dinis Cruz - 48:05 +Yeah. Let me share a couple of things here. Because I think this is on a +topic, maybe some of you guys are already aware of this, but I think it's +quite cool. Can you see my screen? Do about the knife and framework? +Have you heard of this framework? + +Janet Bonar Law - 48:20 +No. + +Dinis Cruz - 48:22 +If you haven't heard about this is pretty awesome. The other one is Worthy +Maps, which we've done a lot of sessions, but I think for this one, it's a +really good behavioral framework created by Dave Snowden that breaks +the world into this different worlds. It's very interesting to understand +where you are and what behaviour change and to see in each of these. It's +a really good way of thinking about how people will react, how systems +will react, how almost like trying to get an understanding of the behavior +that you see is almost a consequence of the environment that exists. But +the environment evolves, right? The evolvement goes from chaotic to +complex to complicated and simple. Also each of these has certain +resilience. A simple, for example, you might think is what you want, but +actually a simple could be a monoculture. There's one thing that happens, +everybody does it, which then is very successful to massive disruption +because if that thing stops working, everything so you actually sometimes +fall from simple to chaotic because things can occur. + +Dinis Cruz - 49:24 +This is quite interesting again how you respond, like you act sense, +respond here, you probe sense and respond. So it's quite cool. Again, I'll +put the links in there and there's some really interesting stuff about the +framework, how you behave, et cetera. I definitely recommend you guys to +see this in the presentation. And then the other one is. + +Tim Ward - 49:46 +Yeah, looks really interesting. + +Dinis Cruz - 49:48 +Yeah. The other one, I see this recently, but have you heard about this one +by this guy Nicholas Means, who talk about the My Line incident. It's a +great story about how this was a nuclear incident occurred in US. How +you can look at this from two different angles completely. One is kind of +what happened in hindsight, almost looking for what went wrong, but +then you can have another view. We does it the second part, which is what +was the behavior that you would expect the individuals and that moment +in time with that background, with that experience, with that stuff, to +respond. You see they actually did what you would expect them to do, +right? Of course, in hindsight you go why do you turn that valve along? +Why do you do this, why do you do that? You would make in how those +individuals were thinking, how they got brought up to their experience, +what the signals they see, even the design of the place where you have all +the alarms that was badly designed, had really bad usability, right? + +Dinis Cruz - 50:49 +It's a great example of you say why did you miss the alarm? Okay, but if +the alarm is on a table with 30 other freaking LEDs and things, and the +mission critical flashing light is next to the splashing light that the elevator +doesn't work. Well, clearly that's a problem. Right. From a Usability point +of view, and there are tons of others. Right. I think these are really cool +examples on behavior, that it's about thinking about the environment that +you create that then leads to the behavior of the individuals, which is, +again, why I think if we can measure the behavior, you can sometimes +even predict the behavior because you know that the ecosystem will occur +on this in this good kind of environment. + +Janet Bonar Law - 51:37 +I think you touched on a really powerful thing there about the power of +stories and connecting with an audience instead of just giving the facts. If +you can weave the facts into a compelling story, the narrative kind of +carries it into the brain. Remember stories better than remember list. + +Dinis Cruz - 52:01 +Absolutely. The story is critical. Yeah. Simon Worley on the worldly maps. +If you haven't seen the sessions on Worldly Maps on the Summit, just +search the Summit website. There's some amazing sessions there, right? +Yeah. Simon Worley talks about how you want maps and how maps give +you context, allowed to understand things, allowed to create a story about +what happens, a narrative that then you then remember. Cool. Well, we +want to talk about the hour. This is a really good session. We need to do +more of this. I love this session. Definitely what the semi is all about. Any +final words? Let me just go around the table. Any final words? + +Danusia Rolewicz - 52:42 +Yeah, I just think definitely I've got a few recommendations that I'm +putting in the chat as you were showing us those great resources. I'm just +trying to it's not really working very well, but yeah, just try to understand +your main audience, try to key into their needs, understand their +behaviour change and empathetic route, and try to marry that bottom up +approach with your top down compliance needs. If you take that route, +then you'll see more positive impact and behavior change than the +compliance focused approach. + +Dinis Cruz - 53:31 +Tim? + +Tim Ward - 53:33 +Yeah, I suppose I've talked about this idea of kind of thinking a bit +differently about security awareness and thinking about the behaviors that +you want to see and then work out how you create a choice architecture, I +suppose, where you make it easy for people to do the secure behavior. Our +approach would tend to be to try and as you said, Dennis, measure things +baseline, see what behaviors are taking place. Once you've understood +that, start thinking about how you're going to steer people in the right +direction. We would talk about trying to do that as in real time as you +possibly can. So you're steering the right behaviors. I'll put some links. +We've got lots of behavioral science articles and awareness articles on our +blog, so I can share that as well. + +Dinis Cruz - 54:19 +Yeah. Also send it to Alana because we can add it to the page. Right. Once +we put the video there, we put it there. That would be really cool. Jeanette, +any final words? + +Janet Bonar Law - 54:30 +Yeah, I think in the past, awareness has sometimes been seen as a +broadcast activity, that we have the information and we broadcast it to +everybody. Because everybody needs to know the same thing, we deliver it +in the same way. Clearly we now understand that's not the case. It's not a +broadcast activity at all. It's a very interactive fiber anthropology exercise +in trying to understand your audience and their needs and fit in your +communications to shape their behaviors. + +Dinis Cruz - 55:08 +Yeah, couldn't agree more. Good stuff. Oh, sorry. Finish any final? You +look like you're going to chip in with a final comment. + +Danusia Rolewicz - 55:17 +No, I just managed to send my message, that's all. + +Dinis Cruz - 55:22 +Yeah. The other one that's really powerful is the checklist, if you see in the +book. What I really liked about that one was the idea that you don't create +a checklist for everything you want to do. You create a checklist to +actually nudge the behaviors and the things that people miss. Right. It's +like, think about a pilot has a checklist, doesn't have a checklist. +Everything they want to do is almost a checklist of either the things they +forget to do or to miss or the things that would trigger other behaviors +that will then get you to do the right in. Right. Again, the way you design +that system makes a big difference. So again, we can end on that. So +thank you very much. This was a really cool session, and I'll see you +around in the other session of the summit. Let's do come up with another +session for the next summit. + +Dinis Cruz - 56:04 +This is really good. + +Tim Ward - 56:06 +Thank you very much. + +Dinis Cruz - 56:07 +Good. + +Tim Ward - 56:07 +See. Bye bye. +