Will Big Data and Artificial Intelligence change how risk .. GRC Articles » Will Big Data and Artificial Intelligence change how risk is managed?
Articles by Author
- March 2017
- January 2017
- December 2016
- November 2016
- October 2016
- August 2016
- June 2016
- April 2016
- March 2016
- February 2016
- January 2016
- November 2015
- October 2015
- September 2015
- August 2015
- July 2015
- June 2015
- May 2015
- April 2015
- March 2015
- February 2015
- January 2015
- December 2014
- October 2014
- September 2014
- July 2014
- June 2014
- May 2014
- April 2014
- March 2014
- February 2014
- December 2013
- November 2013
- October 2013
- September 2013
- August 2013
- July 2013
- June 2013
- May 2013
- April 2013
- March 2013
- February 2013
- January 2013
- December 2012
- April 2012
- March 2012
- February 2012
- January 2012
- December 2011
Tags: $1.4 billion project underway) has penned an open letter critiquing the failings of the scope of the project., Al Newell and I invented a thinking machine.”, and expertise is needed to execute a large Big Data project with the capability to become truly predictive. The predictive capability of Big Data may yet be achieved in the future as these vendors and, and paper-based) data, and voice mail routers have captured the imagination., attempts to “build” a functioning replica of a human brain on a computer are in serious jeopardy. The European Commission’s Human Brain project (yes, By now it should be obvious that a great deal of judgment, Examples such as, Herbert Simon started his class at Carnegie’s Tech Graduate School of Industrial Administration with a startling announcement: “Over the Christmas holiday, IBM’s chess playing, In fact, Let’s start with the basic math used in statistics to transform raw numbers into useful information. There are two basic classifications of statistical analysis: Descriptive statistics and Inferential, media, Now that you have a basic understanding of the limits and scope of these analytical tools you should begin your education to broaden your awareness of what is the appropriate approach for your project, Simon’s “thinking machine” called the Logic Theorist was designed on the theorems of Bertrand Russell and Alfred North Whitehead’s Principia Mathematica, skill, smart elevators, The concept of Big Data is a “catch all” term that refers to a variety of technology solutions positioned to assist with the implementation of data analytics to glean information from stores of data., There are more than 100+ firms involved in artificial intelligence and/or Big Data projects collectively seeking to make sense of the billions of data stored in electronic and non-electronic warehouse, there is a 10 year, to fields as diverse as medicine, Watson, “One cold day in early January 1956.
Will Big Data and Artificial Intelligence change how risk is managed?
“One cold day in early January 1956, Herbert Simon started his class at Carnegie’s Tech Graduate School of Industrial Administration with a startling announcement: “Over the Christmas holiday, Al Newell and I invented a thinking machine.” That announcement 58 years ago ushered in the “Thinking Machine” and the age of Artificial Intelligence.
Simon’s “thinking machine” called the Logic Theorist was designed on the theorems of Bertrand Russell and Alfred North Whitehead’s Principia Mathematica. Fast forward almost 60 years and you will find a host of applications that boast the use of some form of artificial intelligence or AI. Examples such as, IBM’s chess playing, Watson, to fields as diverse as medicine, smart elevators, and voice mail routers have captured the imagination.
To date, the application of AI has been used to make our lives easier and in some cases has replaced or reduced the need for human workers in jobs where routine chores can be easily handled by a machine. However, intelligence has not been outsourced and the promise of a thinking machine is yet unfulfilled.
In fact, attempts to “build” a functioning replica of a human brain on a computer are in serious jeopardy. The European Commission’s Human Brain project (yes, there is a 10 year, $1.4 billion project underway) has penned an open letter critiquing the failings of the scope of the project. It seems that defining and categorizing the complex functions of the brain into digital form is challenging even for the world’s best neuroscientists!
Interestingly, the concept of artificial intelligence is increasingly implied to Big Data. Artificial intelligence and Big Data are not interchangeable terms. However, a form of artificial intelligence may be used in the execution of a Big Data project. Confused? It is not surprising. If you Google “Big Data” you will get a range of disparate definitions that border on the evangelical. What you will not likely hear are the structural and technical limits of artificial intelligence and its use in Big Data. Let’s save that topic for another day.
There are more than 100+ firms involved in artificial intelligence and/or Big Data projects collectively seeking to make sense of the billions of data stored in electronic and non-electronic warehouses. Firms call this data structured (electronic) and unstructured (electronic, media, and paper-based) data. Firms widely boast about the gains made using numeric data in well-structured formats but little is heard about the failures in unstructured data or attempts to provide large scale predictive modeling to data.
Separating Fact from Fiction
The concept of Big Data is a “catch all” term that refers to a variety of technology solutions positioned to assist with the implementation of data analytics to glean information from stores of data.
Terms often used to describe the potential benefits of Big Data include concepts such as predictive analytics or In-Memory processing. In other examples, marketing terms are used to describe the attributes of Big Data, such as “to unlock the business intelligence hidden in globally distributed [data]” or Big Data analytics “is about uncovering hidden correlations, unknown patterns and valuable information to enable a better understanding of the business environment, in effect leading to superior decision making capability.” These definitions and descriptions range from the mildly optimistic to the wildly exaggerated.
So how does one separate truth from fiction in the definitions and solutions offered by Big Data vendors?
Let’s start with the basic math used in statistics to transform raw numbers into useful information. There are two basic classifications of statistical analysis: Descriptive statistics and Inferential statistics. Descriptive statistics is used to “describe” or summarize data in meaningful ways. For example, calculating the mean, median or mode of data is useful if you want to understand the ranking or segregation of data into groups. Descriptive statistics can be used to show past patterns, trends, or changes in the data but cannot be used to predict whether these patterns will continue into the future.
As you can see, descriptive statistics are very helpful in understanding large amounts of data and is used widely in business. Descriptive statistics, however, have limits in that you are simply describing the specific events or losses that have already occurred. The mistake that many users of descriptive statistics make is attempt to use the data to predict future events. Depending on the type of data you have collected risk events can change widely from one point in time to another. This is why some risk professionals are surprised when their descriptive models fail to identify a major risk before it happens.
In order to develop a form of predictive modeling, risk professionals must use the second classification of statistical analysis: Inferential statistics.
Inferential statistical methods require more robust analytics. Users of inferential statistics must first understand what they are attempting to predict. Users should decide if they are attempting to determine a correlation between one or more variables? Its important to understand that correlation does not imply causation!
Alternatively, are you attempting to make an inference (or predictive probability) about the population of a dataset within a certain degree of confidence? These are very simplified assumptions but if not well structured will contribute to bad outcomes.
Quantifying future risk events with a certain degree of confidence requires a large repository of risk data. The larger the quantity and quality of the data; the higher the level of confidence one may be able to assign to its analytical models. Inferential statistics starts with a detailed analysis of the sampling strategy that will be used. Sampling must be sufficient to ensure that your sample is representative of the population you wish to make an inference.
It is also critical that you have an understanding of the distribution patterns of your data. Different industries produce data with a wide variety of distribution patterns. There are a host of statistical techniques used to manage these varied distribution patterns however if you select the wrong ones you will, at best, waste a lot of time and energy, or worse, you will make very poor inferences from the techniques used. The methods of inferential statistics require that you have a clear grasp of the estimation of parameters to measure and you test your statistical hypotheses.
Suffice it to say that the level of expertise and skill to perform inferential statistics is greater than is needed to perform descriptive statistics.
Ok, now let’s recap. Consider for a moment that some of the advanced functions in excel may be considered a form of limited artificial intelligence. One can see how some vendors of Big Data using these classifications of statistics can make their claims. Vendors may package their version of descriptive and inferential statistics in a platform to be used in limited form to create, so called, predictive models.
Now that you have a basic understanding of the limits and scope of these analytical tools you should begin your education to broaden your awareness of what is the appropriate approach for your projects and firm’s data needs.
This is good news for risk professionals!
By now it should be obvious that a great deal of judgment, skill, and expertise is needed to execute a large Big Data project with the capability to become truly predictive. The predictive capability of Big Data may yet be achieved in the future as these vendors and early adopters gain more experience. Learning from the mistakes of others may also be a good strategy if you don’t have the budget or resources to tackle these projects now.
Automating the analytics of a Big Data project may include choosing the type and sophistication of artificial intelligence you need to employ. The skill set of Big Data and Artificial Intelligence vendors is emerging. Selecting a vendor or set of vendors begins with educating yourself about the projects these firms have completed, understanding the complexity and the similarity of their skills to address your problem and developing a clear roadmap is essential.
Artificial intelligence and Big Data are simply a new set of tools to deepen your knowledge of the business and your risk management program. Currently, these projects are focused on ROI initiatives by uncovering new opportunities to sell more products and services or to find ways to cut costs. Risk managers should look for opportunities to incorporate “smart” uses of these tools to reduce risk and improve operational capabilities whenever possible.
Technology will change how risk is managed! We sit at the crossroads of the early stage of this transformation. AI and Big Data projects should not be considered simply tactical projects but strategic building blocks to protect, manage and leverage data as a critical resource with potentially untapped benefits for the organization. Risk professionals should have a role at the table in shaping this vision to ensure that the full potential of these initiatives is realized.