Meeting Owl is a smart videoconferencing master

Like Tweet Pin it Share Share Email

Almost every gadget you use in the near future will benefit from AI. We’ll keep using Alexa on speakers like the Echo, but even a fan in your living room will use AI to control the temperature. The lights will notice when we pull a book off a shelf and brighten in one corner of the room. A coffee-maker will adjust using machine learning based on our taste preferences.

It will happen, and it already is with several new products. One that stands out to me is called Meeting Owl, a $799 device that sits on a conference room table and can automatically determine who is talking. It uses machine learning to understand not just who is speaking, but if there’s a conversation between two people and whether the videoconferencing chat should show two or more views for the video. It’s all on the fly, as well. And, the device doesn’t need to be pre-programmed (setting how many people are in a room, for example).

Mark Schnittman, the Owl Labs cofounder and CTO (who also created the Meeting Owl) explained to VentureBeat that there’s a lot going on behind the scenes.

“The Meeting Owl’s artificial intelligence allows it to act as a real-time director and editor,” he says. “When it’s placed in an arbitrary room in an arbitrary location, it can figure out where people are, which people are most relevant to the conversation at any given moment, and how best to display those people to the viewer.”

Inside the device — which looks a bit like an owl — there’s a 360-degree camera and a 360-degree microphone array. Schnittman says a technique called static beam analysis helps identify the speaker — basically, it’s a method that can determine the highest concentration of energy in an audio sample. It means all of the various microphone signals from the device help identify the strongest signal. “These and other sensing systems are fused together to give a frame-by-frame analysis of where people are and how wide of a field-of-view the camera needs to reliably show each person,” he says.

Then, machine learning takes over. The Meeting Owl gathers the sensor data and uses AI to determine who should appear on screen or if there should be multiple people at the same time. “To do this, the Meeting Owl uses a statistical model to optimize how it decides to display the various people,” he explained. “Of course people move, so all this is done dynamically on the fly, all with the goal of optimizing the experience for the remote user.”

What’s interesting about this technique is that the users in the room barely notice. With the best AI, that’s always the goal. The device thinks like a human, showing the right people (or group) at the right time to match the flow of conversation. From their standpoint, it almost seems like a human is moving a camera toward the correct human speaker or multiple people.

It’s a good use case — no one has to tweak the software before the meeting, no one has to push a remote button. The meeting flows. From all indications, this is how many gadgets will operate in the near future. It will seem like someone is behind the curtain, making our gadgets function more efficiently. Now if someone can just create a device that one guy in marketing keeps checking his phone constantly.

Comments (0)

Leave a Reply

Your email address will not be published. Required fields are marked *

اخبار حلويات الاسرة طب عام طعام وشراب