By
Sudhir Pucha
| Published: November 19, 2019
| Comments
When a customer calls into a contact center, they want a good experience,
and they want to receive accurate information in a timely manner so that
they don’t have to call back. Fulfilling these wants will ensure that they
remain satisfied and loyal.
When it comes to contact centers, time is often equated with money. Metrics
like Average Handle Time (AHT) and First Call Resolution (FCR) are given a
lot of emphasis. These metrics are important for the contact center and
customers.
The longer a customer service representative (CSR) stays on the line with a
customer, the more cost the center incurs. A longer AHT also drives down
customer satisfaction. The cost to the center is tangible: it costs more
per contact to have a live representative on the phone than it does to set
up a system in which customers self-serve and find answers to their own
queries.
The higher cost is also intangible. Customers lose faith in CSRs and the
companies they represent when they are kept on the line waiting or on hold
for answers to queries, which results in a negative perception and loss of
trust by the customer.
More and more, contact centers are employing voice technology to speed up
the time to resolution. Customers have been comfortable with using
interactive voice response (IVR) technology for some time now. On the other
side of the interaction, contact centers are relying on real-time and
post-call voice transcriptions to gain insight into customer behaviors and
needs, using custom algorithms driven by machine learning. Some of this
information used to be gleaned from the IVR in the form of Intent or Call
Reason or after the call as Call Disposition.
Problems occur when managers try to stretch voice transcription technology
beyond its limits. Voice technologies, including machine-based
transcription and voice biometrics, have a place in the contact center, but
they cannot and should not take the place of live agents and human
touchpoints. Technology should equip agents with precise information to
best serve customers. Managers need a balance of technology and physical
agents to best serve customers.
Transcription Software: Limitations
Presently, contact centers use voice transcription services to gain insight
into the basic data of a call but not the actual content of the customer
call. This data, such as the identities of the agent and caller, the
caller’s contact information and the AHT, is just not enough. Organizations
should use voice transcription to drive analytics, which lend insight into
big trends that drive call volume. These analytics help agencies perform
root-cause analysis, which then drives operations into a Continuous
Improvement mode. Analytics also help agencies weed out things that
shouldn’t happen, find use cases for robotics and process automation, find
workflow efficiencies, and gain an understanding of which products or
materials the agent used to drive calls to resolution.
But, however accurate voice transcription services purport to be, managers
are lucky if transcriptions attain more than 50 percent accuracy in
capturing what was said during a call. So while using transcription for
higher-level trend identification makes a lot of sense, the reality is the
technology is not quite there. This quality issue exists because calls lack
context.
If we look at a different voice technology, voice recognition, we see how
critical this context can be for supporting analysis. When a customer
encounters an IVR menu and gives an ID number or responds to a prompt, the
system knows what kind of data to expect. It can confirm the information at
the time of interaction (“You said you want to change your address—is that
correct?”).
Transcriptions lack that context. Add in the imperfections of
recording—background noise, a lack of clarity on the speaker’s part, or bad
cell phone reception—and automated voice transcription software cannot
provide a clear window into the data managers are trying to capture.
While agencies are waiting for transcription accuracy to improve beyond the
50 to 60 percent range, there is still a place for voice transcription
software. Combining the (imperfect) results of transcription with agents’
intelligence and recall, can provide insights that will give customers the
best experience possible when they call in.
Transcription Software: A Good Use Case
Voice transcription currently works best when it helps CSRs work smarter.
In the recent past, CSRs relied on keywords to find the answers to customer
queries. Rather than having to parse thousands of different scripts to find
answers, CSRs could search for a keyword, or a combination of keywords, to
find the information they needed.
This approach meant that CSRs had to memorize thousands of keywords, which
is a lot to ask. Fortunately, as contact center knowledge repositories have
improved, agencies have implemented natural language searches. CSRs can
enter words just as people enter search terms on Google and refine their
results as they go.
Voice transcription can further help CSRs in their quest to find answers.
By “listening” to calls, transcription software works in tandem with an
agency’s natural language query system to suggest information that an agent
might be looking for. While experienced agents may not need suggestions,
seasonal agents hired for surges will rely on them to find the answers they
are looking for.
The more machine-based voice transcription systems are used, the more
intelligent they become. When coupled with natural language search engines,
agencies effectively create a natural language cognitive search engine.
This will be a powerful tool for contact center agents to use when
assisting callers with queries.
Voice Biometrics: Using Voiceprints
While machine-based transcription programs are still in their infancy and
natural language cognitive searching is still a few years away, the use of
voice biometrics in a contact center environment is growing. Using a
customer’s voice to confirm that caller’s identity is much faster than
asking and answering three or four different verification questions (last
name, phone number, the name of your first pet, etc.).
A simple voiceprint, captured as a WAV file when a customer calls for the
first time, is easy to match and nearly impossible for another person to
replicate. Once captured, voiceprints are typically added to a library of
known customers, which helps eliminate the need to authenticate identity
via other means. CSRs can quickly move into the reason for the call, saving
a caller’s time and an agency’s capital. The future is moving away from
knowledge-based authentication toward a hybrid approach of voice biometrics
and a hybrid approach of voice biometrics and real-time technology often
used in fraud detection and forensic analysis.
As the technology catches up to the applications, voice biometrics, a
natural-language search engine, and voice transcription software will work
alongside CSRs will create a swift and powerful response to whatever query
a customer might have. Biometrics will verify a customer’s identity so that
agents can move to the service part of the call. Once CSRs understand the
problem, a natural language search engine, made more intelligent by input
from voice transcription, will speed time to resolution. Both caller and
organization will be happy with the outcome.