AI will transform the insurance sector. Two techniques that will have the biggest impact are convolutional neural networks (CNNs) and recurrent neural networks (RNNs). The other key to business success can be found in a mature approach to putting AI in practice — disciplined productionization.
I just got back from the Insurance Data Science Conference in Zurich, where I gave a talk on how insurance companies can streamline their operations with AI, using the example of passenger car insurance claims. In this article I’d like to share my thoughts from that talk.
In 2017 there were over 15.4 million reported vehicle accidents in the United States. In 2016, there were 16.1 million. These numbers include passenger cars, light trucks, large trucks, and motorcycles.
So insurance companies currently process massive amounts of claim data, and will continue to do so for some time. Insurance claims processing requires sorting through heterogeneous inputs — handwriting, photos, video, audio, signs, maps, typed documents, all coming from multiple sources. Insurance companies can streamline claims processing and other business processes with AI. Two modern techniques that will have the biggest impact on insurance are CNNs for Computer Vision and RNNs for Text Processing (handwriting and speech). The key to business success however is not in model accuracy, but in applying the best practices of productionization (more about that later).
Claim processing often begins with a photo of the vehicle’s damage. Instead of having a human verify these images, we can have an AI model that will recognize and classify the type of damage. Moreover, it will verify that the photo actually shows the insured car.
Results of a visual AI model classifying car damage.
Signing insurance policies and processing claims requires gathering information like license number, vehicle identification number (VIN), mileage and other dashboard indicators. Insurers receive scans of vehicle documents, and photos of specific parts of the vehicle, like those containing VIN number plates. All of these can be processed by AI instead of a human to streamline policy signing and processing claims.
As I mentioned before, the two AI techniques that will have the biggest impact on the insurance sector are convolutional neural networks, which are great at recognizing objects in images, and recurrent neural networks, which can “read” handwriting and “understand” speech. Neural networks are state-of-the-art for visual tasks, topping even human accuracy. Thanks to the rise of graphical processors and a number of recently discovered techniques, like smarter optimizers, one-cycle fitting, and transfer learning, we now can train much deeper models than ever before. The outcome is that they are now unbeatable in most visual tasks and they are becoming widely used in business.
The core concept that made neural networks so awesome in visual tasks is convolutions. The concept is modelled after the human brain — this is how we see things. Our brain doesn’t have just one part that is responsible for all object recognition. Instead, we have parts of our eyes, neural system and primary visual cortex, from initial ones responsible for recognizing simple shapes like straight line or curve, to more and more complex shapes and patterns and finally to faces, and entire objects. The amazing thing is that when we train a neural network, it learns what to do on each of these levels, and we don’t have to specify any of them.
So how does it work? What happens inside such a network? I think it is best understood thanks to visualizations like In the images below. For each layer we see the features that the network learned and examples of images best matching these features. The first layer is only able to recognize very simple structures like straight lines and some gradients of colors. Then as we progress through the layers, the objects and features that are recognized become increasingly complex. In the second layer we can see there are rounded shapes. In the 3rd layer we can already see repeating shapes and objects such as a wheel and a part of a car. Later we have dog faces, even more complex features, and finally full objects, like flowers, humans and bikes in the last layer.
Source: Visualizing and Understanding Convolutional Networks, Matthew D. Zeiler and Rob Fergus
When it comes to recognizing text, such as VIN numbers, license plates and vehicle registration documents, the most important modern technique is recurrent neural networks (RNNs).
In particular, architectures based on LSTM (Long Short-Term Memory) allow us to recognize text more precisely. An LSTM are able to remember both information that it has seen recently, as well as information it has seen long ago. The result is that it improves accuracy to unseen levels. Here are a few examples:
First example: an algorithm learns a recurrent network policy that steers its attention around an image; In particular, it learns to read out house numbers from left to right.
Another LSTM example: automatic transcription without a prior segmentation into lines, which was critical in previous approaches:
For more about CNNs and RNNs, here is a nice description.
Accuracy of a model is important, but the key to business success is a mature approach to putting the AI in production. In our experience, key points for successful productionization of an AI model are the following:
In the coming weeks we will release another article describing the best practices for productionization which can be applied to AI as well as a host of other data science projects.
Deep learning architectures based on CNNs and RNNs have gotten to a point where we can successfully apply them to a range of problems in the insurance sector. With the right approach to productionization, these solutions will bring insurance processes to new levels of performance.
The time for AI in insurance is now, and we’re excited to be a part of this change.
You can find me on Twitter @marekrog.