Five questions to ask before starting your first (next) AI initiative.

D.M. Nichols
Candidit
Published in
5 min readAug 2, 2019

--

What is AI?

AI is often widely understood as a system that comes to conclusions in a way that resembles a human’s approach. Still, AI is defined differently by different communities and its definition will continue to change with future advances in technology. For instance, some define AI loosely as a computerized system that exhibits behavior that is commonly thought of as requiring intelligence. Others define AI as a system capable of rationally solving complex problems or taking appropriate actions to achieve its goals in whatever real-world circumstances it encounters. Still others define AI as a great marketing term to connect to an otherwise non-intelligent system as a means of capitalizing on the recent resurgence of the term from its varying iterations through the years.

To help cut through the clutter, we offer a few questions to ask anytime AI is brought into the conversation:

Is it based on data?

Without data, there is no AI. In fact, rather than starting with assumptions, AI techniques start with data and “learn” certain assumptions based on that data. Many emerging AI technologies and platforms are bringing their own data set … which begs the question, on whose data are the assumptions based?

Additionally, while it’s tempting to think that we could conceive a perfectly rational system, the data used to train AI is often inherently flawed, due to human or other bias, which makes “perfect rationality” nearly impossible to evaluate.

While smaller organizations are not likely to have enough quality data to develop their own AI, larger organizations would be better off developing and training their own AI or machine learning systems with their own data to better reflect their own unique needs and circumstances. This is especially true regarding utilization within the human or organizational elements of the company. Unlike uniform systems, chemical or biological processes, human behavior varies widely based on environment, conditions, situations, etc. If you are interested in solving human problems unique to your organization, and you have a large enough set of reference data, you are often best served to develop your own models to achieve optimal results.

Is it unconcerned with the outcome of its calculations?

Unlike the inherent messiness of human decision-making, an AI’s capacity to make decisions isn’t influenced by ulterior motives or how much sleep it got last night, but is solely focused on the task at hand. Since it doesn’t know good from bad, however, any biases that exist in the data are perpetuated. This is where the need for solid design is paramount. You may not be first at launching your AI-based services, but if your design and your data are superior, you will have the superior product in the end.

Are its abilities learned rather than programmed?

AI can improve iteratively on its own — without being programmed every step of the way, it can learn from its experiences and improve at making future predictions and decisions, resulting in increasingly sophisticated abilities. Managers that want fine-tuned control over their models are best to avoid AI and ML altogether as the internal calculations (generally derived through back propagation).

A deep-learning system doesn’t have any explanatory power. A black box cannot investigate cause. Indeed, the more powerful the deep-learning system becomes, the more opaque it can become.

Recent advances in machine learning have largely been due to the growth of deep learning — a subfield of machine learning. Deep learning borrows from the structure of the brain, by linking lots of simple “neuron” like structures together to do interesting things in a neural network. By stacking many layers of these artificial neurons together (hence “deep”), the network as a whole can learn to do complex tasks. Interestingly, “neurons” in these layers often end up performing specific roles, such as recognizing edges, or the outline of a specific object. The unique strength of deep learning is that these sub-tasks — often known as “features” — are learned directly from the data, rather than being specified by programmers.

Is it constantly evolving?

AI models continually gain in value over time. For this reason, even initiating a clean data collection initiative sooner than later is preferable to waiting or utilizing an external data set.

Let’s take a real-life example for HR: recognizing ideal fit in applicant selection. A classic AI approach would rely upon a human expert trying to distill their own decision-making process and then codify it in the algorithm. For instance, in our applicant tracking system, we might flag specific words that are relevant to the job, or connected to a specific certification. With deep learning, however, we can directly feed profiles of existing employees labeled to indicate whether they’re high-performing or not, and our neural network will learn to pick out the most useful features of the profile for this particular task. This is a classic example of “supervised learning”: we provide some inputs and some desired outputs, and the algorithm learns to map from one to the other.

From this perspective, the more data, the more learning the model can handle, though it is also important to understand that current iterations should be narrow in scope: ie. one job type at a time.

We can also do away with the labels entirely, and ask the algorithm to group the traits or “competencies” that have something in common.This process is known as clustering, and it’s a type of unsupervised learning. Here we’re not providing supervision in the form of labels, we’re simply using deep learning to find structure in the data. In our example, perhaps our profile includes many different competency types — academic, hands-on experiences, and software — and it’d be useful to cluster these before trying to figure out which competencies in each cluster are most important for selection.

In either case, business needs change dynamically, but not swiftly, and as data is continually fed into the model, the model gains capabilities, potentially being able to predict future occupational skill groupings, or ideal skill groupings.

--

--