Request a quote!

Blog, News &

Case Studies

A Collision of Trust, Cobots, and AI Communications: Themes of the 2021 Collision Conference
By E2E Research | April 23, 2021

Collision 2021 was a four-day, North American tech conference that drew more than 38 000 attendees. I was fortunate to be one of those attendees this year thanks to a ticket kindly donated by ESOMAR. This year, the Collision Conference hosted more than 600 speakers from all walks of life. Just a few of those people included:

 

  • Celebrities: Cindy Crawford, Meaningful Beauty; Maria Sharapova, Therbody; Ashton Kutcher, Sound Ventures; Ryan Reynolds, Mint Mobile
  • CEOs and CMOs from global companies: Geoffrey Hinton, University of Toronto; Ukonwa Ojo, Amazon; Fiona Carter, Goldman Sachs; Martin Wildberger, Royal Bank of Canada
  • Local and global community leaders: Jagmeet Singh, Leader of Canadian New Democrat Party; John Tory, Mayor of Toronto; Katie Porter, Representative at US House of Representatives; Lori Lightfoot, Mayor of Chicago
  • And 13-year-old whiz kids whose expertise and speaking skills rivaled the most experienced speakers in attendance!

 

With hundreds of sessions running simultaneously (and literally colliding with each other!), it was easy to create a personalized stream of content, particularly since no matter the time, a great talk was always just beginning. The stream I created for myself focused on artificial intelligence, robotics, and innovation. Here are the key themes I took away.

 

 

Technology Leaders Must Prove Their Trust

People love their devices. We trust them to help us discover and buy products, make and take phone calls and text messages from our loved ones, and remind us about confidential meetings and doctor’s appointments. We trust our devices will work as expected when we need them to work. However, there is a trust problem and it doesn’t lie with the technology itself. It lies in the fact that we don’t trust the people behind our devices, neither the people building the devices nor our government leaders, to create and hold appropriate boundaries around privacy and security.

 

Companies build trust by having clear values and a clear mission grounded in being authentic, empathetic, transparent, and relatable. We learn to trust companies that shape our experiences in ways that are personalized but at the same time not creepy. We also learn who trust by witnessing which companies hold themselves fully and immediately accountable when they make mistakes. Companies that abuse these expectations will quickly find themselves speaking to a declining audience. A great way to think about trust is that every interaction a company has with a consumer is either a deposit or a withdrawal. You do good or you do bad. There is no neutral.

 

 

Robot, Cobots, and the Inevitable

Did you realize you already have robots in your home? If we follow the strict definition that any automatically operated machine that replaces people is a robot, then your electric toothbrush, your toaster, and your vacuum cleaner (even if it’s NOT a Roomba) are robots. We’re slowly getting used to the idea that robots don’t have to take a human shape to be called robots.

 

A newer take on robots is the idea of cobots. Unlike a lot of robots that run behind the scenes, collaborative robots are designed to interact directly with or next to people. While you may be nervous that robots or cobots will take your job, there are many good reasons to be excited about working with them. Not only do they easily take on jobs that are dull, dirty, and dangerous, they augment our skills and abilities and help us do our work better and with more agility. Robots make us physically stronger and mentally more agile. If we let them, they help us make truly better decisions.

 

As in the case of robots and cobots, if something is inevitable, get enthusiastic about it.

 

 

The Language of AI

One of the main complaints about artificial intelligence comes when it’s used as a substitute for people. For instance, researchers are actively working on building AI tools intended to serve as personal companions for people who are elderly or disabled, and counsellors for people who’ve experienced trauma. Isn’t that impersonal? Isn’t that disrespectful? Well, let’s consider it from a different angle.

 

Think about people who’ve experienced a life of trauma, a life wrecked by abuse, trafficking, trauma, or addiction. A life where people have repeatedly let them down and shown that they can’t be trusted. Those who’ve experienced trauma may find it particularly hard to trust new people and may be far more comfortable beginning their healing process by working with AI.

 

Think about people who have experienced a brain injury or deal with communication disabilities. Or people who aren’t using their native language. Or people who feel more comfortable communicating via email or text. We constantly hear that people should be treated in the way they want and prefer to be treated. That we need to increase accessibility. This could easily be AI.

 

Regardless of the initial need, we need to ensure that these AI communication tools demonstrate empathy and show respect. AI can’t replace human judgement but it can and should reflect good judgement.

 

 

What Does It Mean For Researchers

The research industry talks about trust all the time. We need research participants to trust us enough to share their most personal opinions, their most private click-paths, and their most unusual purchase behaviours. We need research tools that can effectively automate dull and error-prone research tasks leaving us with more time to do our jobs even better and make better decisions.

 

And we really need to focus on language. So much of our work revolves around language – writing questionnaires with respectful wording that everyone can understand, moderating focus groups that accommodate every participant, making the research space accessible to all.

 

I may not have attended a single market research talk but I did indeed come away with new perspectives that will make me rethink how I have conducted research in the past, and what I will do in the future.

Transformation and Trust: An ESOMAR webinar recap
By E2E Research | March 31, 2021

On March 31, 2021, ESOMAR hosted a flagship series webinar on trust which was moderated by Fiona Blades from MESH Experience and Maya Kantak from Disney Parks.

 

The speakers defined and described trust in a number of ways. Paul Neto described it as “consistency over time with positive experiences.” Jeff Marshall and Rogier Verhulst shared that people who trust a brand are more likely to defend it, advocate for it, buy from it, and stay loyal to it. On the other hand, they explained that people can trust a company because they do what they say they will do but, at the same time, they might not trust other aspects of the business. Rob Key clarified that having trust with the brand or the company doesn’t equate to them being right or wrong, or good or bad. For instance, we trust that Amazon will deliver our package tomorrow but we might not trust how they treat our data.

 

Ann Constantine shared that in their company, they build trust from the bottom up through a consultancy process of creating ownership with the people who implement those processes every day.

 

Paul Neto discussed trust as it applies to the research industry. Right now, only about a third of people trust market research. The consequence of insufficient trust is that a third of people will refuse to participate in research and a third will limit the information they share. By building trust, researchers and marketers can unlock more consumer data. We need to be more transparent about the research experience by telling people what’s happening with their data, why they were screened out, and why the survey and incentive weren’t the length they expected. We need to step away from privacy by compliance and move to privacy by design. When we don’t meet these basic expectations, we lose trust and we lose valuable data. An interesting idea Paul raised is whether our attempts to improve the research experience are simply patchwork fixes. Perhaps if people trusted us more, we wouldn’t have to enforce shorter surveys or worry about mobile-first designs. People would trust our words and follow our advice.

 

Nicolas Pochart shared a fascinating example of GSK’s  “Consumer Closeness Program” wherein their research & development scientists spoke directly to consumers using virtual qualitative tools. Though they didn’t see any cost savings and it was actually extra work, talking directly to consumers was a huge success. The scientists were nervous ahead of time because they’d never talked to consumers before. However, they enjoyed the experience and confessed they will remember those consumers’ words forever. It changed how they think about their work. In 2021, GSK will have 500 R&D scientists talk to consumers. It won’t replace their regular qualitative research, but it definitely helped to put consumers at the center of the conversation, it created empathy with the scientists, it helped them understand consumer language, and it was vastly more impactful than receiving a 200-page PowerPoint file.

 

Jeff Marshall and Rogier Verhulst shared how they measured trust at LinkedIn. In 2020, LinkedIn registered the most digital trust among social networks. This trust is at least partly because of the individual authentication that is built into the network. People have to behave well to ensure their career and company are successful. Further, LinkedIn doesn’t allow open APIs and they don’t share their data with third party publishers. To measure trust, they conducted a major research study using narrative analytics and machine learning models to discover the best trust metrics. They followed that up by building a trust tracking program with interviews, focus groups, and surveys. The model is very important because it allows them to identify up and coming problems and be prepared to resolve those problems before it becomes too difficult. This is particularly important because of halo effects – when another social network experiences a trust issue, it can easily be reflected onto other similar brands.

 

Duncan Southgate and Gonca Bubani shared their thoughts on trust in the media. They’ve identified that trust is correlated with purchase and the growth of a brand’s value growth over time. Right now, traditional media, like TV ads, receives the highest trust scores but it’s still very low. When it comes to media channels, Google is the most trust media channel followed by YouTube and Instagram. As we’ve heard so often, people feel the media is biased. A single opinion can have extremely far reaching implications quickly leading to a loss of trust. Media can improve their trust scores by being more relevant, useful, authentic, innovative, fun and entertaining, and ensuring advertisers using their platforms are trustworthy.

 

Speakers and topics from this webinar included:

  • Ann Constantine, Direct Line Group: From Top Down to Bottom up: Transforming Direct Line Group’s Measurement Approach
  • Duncan Southgate and Gonca Bubani, Kantar: Trust in Media – The New Publishing Battleground
  • Jeff Marshall, Protagonist and Rogier Verhulst, LinkedIn: Building and Measuring Customer Trust at LinkedIn
  • Nicolas Pochart, GSK Consumer Healthcare: Transforming our Organisation Through Consumer Closeness
  • Paul Neto, Measure Protocol: Exploring the Use of Trust Principles to Unlock the Next Generation of Consumer Data Collection