Confidence levels can be used anytime one is estimating or predicting something. Examples include: business, engineering, medicine, technology... or just day-to-day life.
As humans we use confidence levels regularly. Whether you decide to dodge an aisle at the grocery store because you thought you saw your chatty neighbor, or using evidence and intuition to convict a suspected criminal during jury duty, your mind is in a constant state of perceiving its surroundings. It makes decisions based on those perceptions via an inherent estimate of confidence.
And so how does your mind generate confidence levels? Well... this all happens in a split second. The brain using an innate multimodal approach to better perceive external scenarios and draw conclusions. In the first phase of the brain’s decision-making process, most people rely on their 5 senses.
This is called a cognitive shortcut, or heuristic technique, which is immediate and instinctive, but not guaranteed as optimal (examples: an educated guess, sensory instinct etc).
A second layer would then be cross-checking this analysis with previous experience. This indicates a more advanced thought process which either validates or invalidates the brain’s initial confidence level in making a decision.
This logic is based off of basic evolutionary traits that allowed our ancestors to not just survive, but live longer. Making correct decisions meant outliving the bad decision-makers.
When looking at modern day science, confidence levels become more statistics-based and less intuitive.
Do you remember your first Statistics class? The heavy textbook you never used, the lecture hall full of students... and (if you were paying attention) a board that looked something like this?
The image above is an example of estimation statistics. You can see a parabolic representation of a confidence interval (CI) of 0.44 and 0.64. For reference, the CI is the range of experimental results expected to contain a parameter of the population of interest. The confidence level is 95% (probability that you would get similar results if you repeatedly run a test). With this information, we create a data set that is “trusted” with 95% confidence.
Just as your statistics professor painstakingly calculated these data sets on a blackboard, neural network models now provide results much quicker. They also deliver a smaller margin of error that could be caused by humans.
As applied to machine learning, a minimum confidence level is pre-defined by an AI model with existing thresholds to validate results (remaining empirical).
This solution is still based on estimation statistics - except instead of taking 30 minutes, data sets are populated instantaneously.
That’s kind of heavy, so let’s take a lighter example.
Say you live in New York and take a summer vacation to Tokyo. You just finished watching an amazing sumo match, and now you are walking back to your hotel, enjoying some evening sight-seeing.
Suddenly, someone catches your eye as you’re walking through the crowded streets. It’s a familiar face, but you can’t quite pinpoint his identity. As you pass him directly and get a closer view, you realize this person looks a lot like your neighbor in Manhattan, Paul Johnson.
But it can’t be, so you keep walking unphased. The likelihood is low. Paul being in the exact same place as you at the same time, so far away from home- it’s nearly impossible in your mind. And well... nearly impossible statistically-speaking, too.
... Except that it was Paul.
As a recap, in this case, you:
This example showcases the immense power of the human mind.
The missing element was context, as you saw his face, but did not have the existing knowledge that he was indeed on vacation in the same city as you.
In this instance, perhaps your innate confidence level was a 2 out of 10... also showing how much the human brain relies on the contextual approach to making correct decisions.
Just as chemistry and neural architecture of the brain helps humans draw conclusions and recognize faces, AI is able to use statistics and neural networks to detect various components.
Such elements include faces, objects, language and context, just to name a few.
Similar to how humans use basic senses and contextual information to make decisions and estimate confidence levels, AI algorithms can now replicate this process.
Now machines are able to detect what’s in a video, for example, via cross-analysing facial recognition, speech-to-text, metadata... aligning sources of data to better perceive results, in record-breaking time.
That said, just as the human brain has margin for error, the same can be said for deep learning. After-all, the same algorithms were created by humans.
In no way does AI take the place of a human brain- but it does emulate the brain’s perception process via multimodal analysis factors.
Understanding and perceiving the world around us originates from a mixture of fascination, necessity and survival. Nowadays in a digital-focused society, many entities understand the importance of intaking information quickly and on multiple levels.
Here are some examples:
In its third quarter 2019 update, Tesla stated that autopilot technology is “9x safer than the average human driver [in the U.S.].” The report supports this statement with the fact that autopilot cars were only involved in one accident for every 4.34 million miles (7 million km) driven during the quarter- (in the U.S) whereas the average human driver related accident is one for every half million miles (800 thousand km). In terms of machine decision making- the confidence level of Tesla autopilot technology is not just high-functioning it is life-saving.
Perhaps you heard about it in the news. Now China is making facial recognition mandatory for all mobile phone users to protect cyberspace citizens against fraud. Telecom companies now use artificial intelligence among other technical methods to make extremely important decisions when it comes to online identity- so it could be fair enough to say that confidence levels must be extremely high.
Moments Lab created a system of multimodal analysis that uses machine learning to train AI and adapt the algorithm. More specifically, journalists can upload all images and videos to a secure cloud-based platform. From there, digital content is automatically tagged and indexed via multimodal facial recognition (visual), object detection (tactile) and speech-to text technologies (audio), among others. This technology also has the ability to read any “open text” sources to improve results across various scenarios.
Perhaps one of the main takeaways is that as machine learning and AI continue to advance, human logic and decision-making remains at the heart of machine perception.
In turn, many other technological advancements and revolutions to come. Just as humans use confidence levels for making important decisions, today multimodal AI machines use a similar rationale to make judgements and take action in a short period of time.
This is leading to more profitable, secure and sometimes life-saving results.