When catastrophes strike, you have no time. You’re under pressure to quickly understand the financial impact of an event and provide estimates to management. At the same time, you (and your team) are constantly tracking the event, processing hazard data, making sure exposure data is accurate, pulling reports, and (hopefully) beginning outreach to insureds. The last item—proactive customer outreach—may suffer, however, when the other to-dos consume your time and resources.
Speed and quality of response following catastrophes can be an asset to your organization, and a key reason why your customers choose you over your competitors. But only if you can make your event response operations run like clockwork. This entails moving away from the status quo and integrating elements of automation into your event response processes. Let’s look at some of the challenges you may face and how to implement a more proactive approach for minimal cost and disruption.
Hurricanes, in particular, illustrate the problem of quickly deriving insight from data. For example, does the following scenario sound familiar?
Imagine a hurricane strikes…
…and it’s impacting Texas, Florida, or the Carolinas (probably not too hard to imagine, actually). Management is asking for the estimated financial impact of this event, and your stress levels are rising. It’s all hands on deck.
1) Get event data – 45 minute clock
You go to the NOAA website, pull down wind datasets from the latest update, and then work to get them into a usable format.
2) Intersect with your portfolio – 60 minute clock
Now, it’s time to intersect the footprint with your portfolio data which may take another hour or so to complete.
3) Update portfolio – 45 minute clock
After you get everything set up, you realize your portfolio is six months old, which may over- or underestimate your actual exposure. Do you pull an updated snapshot of your exposures? Probably not because there isn’t enough time!
4) Run financial model SQL scripts – 45 minute clock
With a manual intersection process, you are likely unable to easily access the impact of policy terms and conditions, so you’ll need to run some financial model scripts to determine the actual exposure for this event.
5) Create and share reports – 45 minute clock
You finally get some financial numbers ready and format them into a nice report for management.
Then, you think about what you had on your to-do list for the day before the hurricane was in the picture. Or wait, maybe not…because just then, you see that NOAA has published the next snapshot of the hurricane.
Rinse and repeat. It’s going to be a long night.
Let’s face it, if you can’t extract insight from data fast enough to mitigate damage or provide a timely course of action, your operational efficiency and downstream customer satisfaction will suffer. And just think, this was for a single data source. Realistically, you must perform these same steps across multiple sources to gain a complete understanding of this event (e.g. KatRisk, Impact Forecasting, JBA flood, NOAA probability surge, etc.).
What makes the above process so time-consuming and inefficient?
- You had to source the data yourself and operationalize it (i.e. get it into a usable format)
- You had to navigate the complexity of the data, which can be exceptionally time-consuming depending on the source, resolution, and other variables
- You realized your portfolio data was out of date, which means you can’t determine the actual financial impact of an event
- You had to manually run a financial model after determining the exposures that could be impacted by the event
- And, of course, you had to manually pull this information together into a report for stakeholders
So, what can you do? API integrations help solve these challenges by ensuring you always have the latest hazard data and portfolio snapshot available. If you invest just a few hours to get your data configured with a data import API like Insurity offers, you’ll always have the latest view of your exposures ready to analyze—without ever lifting a finger. You’ll save countless hours in the long run by investing just a few hours up front. This also enables quicker and more accurate analyses downstream since you won’t be over- or under-stating your exposures (not to mention creating errors by scrambling at the last minute to get a refreshed snapshot).
Imagine another hurricane strikes…but this time you’re set with automation
Those couple of hours that it took to get your portfolio data integrated and automation in place with a solution like SpatialKey Event Response are paying off (no deep breaths required).
Within moments of NOAA publishing an update, you receive an email notifying you of the financial and insured impact. With the click of a button, you’re in a live dashboard, investigating the event, your impacted exposures, and more.
You still need to get those numbers to management, but this time you can breathe easy knowing that your numbers are accurate and the whole process took a fraction of the time. Now when NOAA (or any other public or private data provider) pushes the next update, you’ll be set with a highly scalable infrastructure that enriches your data, calculates financial impact, and produces a report within minutes.
Why was this process much more efficient?
- Since you invested a few hours in implementing API technology, your exposure data was up to date
- You had access to pre-processed, ready-to-use hazard footprints as they became available
- The event was monitored 24/7 so you didn’t need to constantly track it and pull reports to understand what changed
- Custom filters and thresholds ensured you were never inundated with notifications and only received metrics that you care about
- You saved a bundle of time because a financial report was auto-generated for you to pass along to upper management
- You were able to quickly share reports across your organization, enabling your claims team to get a head start on their customer outreach
Now, you’ll never be a bottleneck in the process of understanding and communicating the impact of an event to your stakeholders. And, with all the time you’ve saved, you can use SpatialKey Event Response to contextualize the event and dive deeper into investigating it some more.
Tick tock: It’s time to make your event response run like clockwork
It’s clear, there’s a better way to tackle the growing challenge of deriving insight from data and quickly understanding the impact of an event. If you lack the ability to operationalize and extract insight from time-critical data, you’re operating in status quo when your management team and customers expect to know more about an event, and sooner.
Fortunately, automation doesn’t have to be a time-consuming or costly endeavor. There are simple ways to automate your manual processes, such as API integrations, that save time and steps along the way. “Automation” can carry with it preconceptions of disruption and heavy investment, but this is not true of a solution like SpatialKey Event Response. Automating your event response operations can positively impact your customer retention and drive efficiencies now—not years from now.
Next week, in Part 5 of this series, we’ll discuss the seven questions an automated event response solution should answer for you.
For a complete overview of how to make your event response operations run like clockwork, get the guide.