Parsing a large computational process into smaller independent tasks that run in parallel to each other can help actuaries benefit from the time-saving efficiencies of cloud computing. Machine learning has parallel compute capabilities to assist with these tasks. In this article, Milliman’s Joe Long and Dan McCurley discuss how they were able to cut a three-month machine learning project down to four days using open source tools and the Microsoft Azure cloud.
Milliman will debut its proprietary predictive modeling platform at the Insider Tech Conference held in New York City on December 6. Milliman’s recently created analytics software, Solys, uses advanced computer languages, models, and machine learning so that consultants can serve their clients with increased speed, reach, and cost-efficiency.
An internal tool that can be used to benefit Milliman’s current and future clients, Solys simplifies processes, improves data management, and performs advanced predictive analytics using the latest software environments and programming languages. The leading technology increases efficiencies and consultant capabilities in the growing InsurTech field. Milliman consultants will be discussing the tool and the firm’s work in InsurTech at a panel discussion at the Insider Tech event in New York on December 6.
As insurers face disruption around the “Internet of Things,” the shared economy, and autonomous vehicles, it’s vital that their consultants provide the best answers in the fastest and most cost-efficient manner possible. Milliman’s advanced predictive modeling tool enables consultants to address their clients’ InsurTech questions and remain leaders in this rapidly changing industry.
To read Milliman’s InsurTech research, click here. Also, to subscribe to Milliman’s InsurTech updates, contact us here.
How can a company leverage customer data and turn it into actionable information? This was the challenge one transportation provider faced when its modeling system began underperforming after the company implemented it to predict revenue and passenger traffic. In this article, Milliman consultant Antoine Ly discusses how the firm created a machine-learning model that helps the company analyze various aspects of its ridership, leading to more informed financial decisions.
Here is an excerpt:
Working from a mock-up drafted by the client, the [Milliman] team reproduced the dashboard to the client’s specifications, but it is now supported by newly developed software as well as the client’s data warehouse. The dashboard allows the client’s management team to quire different aspects of passenger usage to gain insight into traffic flows and revenue. Colour-coded symbols, which when clicked on, give managers a concise picture of a train’s revenue and traffic. Managers can also quire the system based on selected features for both past usage and anticipated ridership, and are now able to make more informed decisions about pricing, the need for discounts or adjustments to marketing campaigns.
Because the model can adapt to new situations, deviations from the average error are confined to a much more narrow range. This means managers can have more confidence in the model’s predictive value and increases their ability to manage revenue.
This report by Milliman’s Neil Cantle uses advanced machine learning algorithms, such as deep neural networks, to analyse social media conversations about Brexit. The purpose of the study was to examine whether useful information could be extracted from social media in what is effectively real time on a key topic in a political economy.
While machine-learning techniques can improve business processes, predict future outcomes, and save money, they also increase modeling risk because of their complex and opaque features. In this article, Milliman’s Jonathan Glowacki and Martin Reichhoff discuss how model validation techniques can mitigate the potential pitfalls of machine-learning algorithms.
Here is an excerpt:
An independent model validation carried out by knowledgeable professionals can mitigate the risks associated with new modeling techniques. In spite of the novelty of machine-learning techniques, there are several methods to safeguard against overfitting and other modeling flaws. The most important requirement for model validation is for the team performing the model validation to understand the algorithm. If the validator does not understand the theory and assumptions behind the model, then they are likely to not perform an effective model validation on the process. After demonstrating an understanding on the model theory, the following procedures are helpful in performing the validation.
Outcomes analysis refers to comparing modeled results to actual data. For advanced modeling techniques, outcomes analysis becomes a very simple yet useful approach to understanding model interactions and pitfalls. One way to understand model results is to simply plot the range of the independent variable against both the actual and predicted outcome along with the number of observations. This allows the user to visualize the univariate relationship within the model and understand if the model is overfitting to sparse data. To evaluate possible interactions, cross plots can also be created looking at results in two dimensions as opposed to a single dimension. Dimensionality beyond two dimensions becomes difficult to evaluate, but looking at simple interactions does provide an initial useful understanding of how the model behaves with independent variables….
…Cross-validation is a common strategy to help ensure that a model isn’t overfitting the sample data it’s being developed with. Cross-validation has been used to help ensure the integrity of other statistical methods in the past, and with the rising popularity of machine-learning techniques, it has become even more important. In cross-validation, a model is fitted using only a portion of the sample data. The model is then applied to the other portion of the data to test performance. Ideally, a model will perform equally well on both portions of the data. If it doesn’t, it’s likely that the model has been over fit.
Registration for the 2017 Data Science Game is officially open. The Data Science Game is a two-phase competition showcasing teams of data science students from universities around the world. An online qualifier will take place on April 15 with the final stage happening in September.
Students from the Moscow Institute of Physics and Technology (MIPT) won last year’s competition. Will your university win this year? To register your team, visit www.datasciencegame.com. The deadline to register is April 9.
Milliman is a sponsor of the 2017 Data Science Game.