Robot Knows the Right Question to Ask When It's Confused

[] http://spectrum.ieee.org/automaton/robotics/artificial-intelligence/robot-knows-the-right-question-to-ask-when-its-confused


Brown University researcher works with a Baxter robot.Photo: Nick Dentamaro/Brown University

Last week, we (and most of the rest of the internet) covered some research from MIT that uses a brain interface to help robots correct themselves when they’re about to make a mistake. This is very cool, very futuristic stuff, but it only works if you wear a very, very silly hat that can classify your brain waves in 10 milliseconds flat.

At Brown University, researchers in Stefanie Tellex’s lab are working on a more social approach to helping robots more accurately interact with humans. By enabling a robot to model its own confusion in an interactive object-fetching task, the robot can ask relevant clarifying questions when necessary to help understand exactly what humans want. No hats required. 

Whether you ask a human or a robot to fetch you an object, it’s a simple task to perform if the object is unique in some way, and a more complicated task to perform if it involves several similar objects. Say you’re a mechanic, and you want an assistant to bring you a tool. You can point at a shelf of tools and say, “Bring me that tool.” Your assistant, if they’re human, will look where you point, and if there are only a handful tools on the shelf, they’ll probably be able to infer what tool you mean. But if the shelf is mostly full, especially if it’s full of objects that are similar, your assistant might not be able to determine exactly what tool you’re talking about, so they’ll ask you to clarify somehow, perhaps by pointing at a tool and saying, “Is this the one you mean?”

To be useful in situations like these, your assistant has to have an understanding of ambiguity and uncertainty: They have to be able to tell when there is (or is not) enough information to complete a task, and then take the right action to help get more information when necessary, whether that uncertainty comes from the assistant not really getting something, or just from you not being specific enough about what you want. For robot assistants, it’s a much more difficult problem than it is for human assistants, because of the social components involved. Pointing, gestures, gaze, and language cues are all tricks that humans use to communicate information that robots are generally quite terrible at interpreting.

Brown researchers created a system called “FEedback To Collaborative Handoff Partially Observable Markov Decision Process,” or FETCH-POMDP, which is able to understand the implicit meaning of common gestures, and merge those meanings with what the person making the gestures is saying to improve its understanding of what the person wants. 

At Brown, they’ve created a painfully acronym’d system called “FEedback To Collaborative Handoff Partially Observable Markov Decision Process,” or FETCH-POMDP. It’s able to understand the implicit meaning of common gestures, and merge those meanings with what the person making the gestures is saying to improve its understanding of what the person wants. Assuming that the person is being cooperative (not lying about what they want), the system is able to model its own confusion and ask questions only when necessary for accuracy so as not to be unduly bothersome. 

To test out the FETCH-POMDP system, the Brown researchers asked people who had no idea what was going on to ask a Baxter robot to fetch things for them. A third of the time, the robot asked no questions; a third of the time, the robot always asked clarifying questions; and the final third of the time, the robot only asked questions when it decided that a question was necessary. The researchers expected that the robot would be fastest when asking no questions, and most accurate when always asking questions, but it turned out that the intelligent questioning approach managed to be both the fastest and most accurate. This is because human robot interaction is messy: People asking questions led to transcription errors (confusing “yes” with “hand,” for example), so more questions meant more misunderstandings.

Interestingly, the participants in the trials also ascribed all kinds of capabilities to the robot which it didn’t actually have:

During trials, many users used prepositional phrases in order to describe items, such as “Hand me the spoon to the left of the bowl.” Although the language model in this work did not account for referential phrases, the agent was able to use intelligent social feedback to figure out what the human desired. This may explain why many users reported that they thought the robot did understand prepositional phrases. Methods exist to interpret referential language, but problems in understanding will still occur. Our model will help correct those mistakes, regardless of the exact method of state estimation and language understanding.

The researchers do plan to update their model to include referential phrasing like this, and they also want to add things like eye tracking to improve accuracy even more. Adding the ability to place (along with pick) has the potential to make this system much more workplace useful, especially if you have a spaceship that needs fixing



Brown University researcher works with a Baxter robot.Photo: Nick Dentamaro/Brown University

Last week, we (and most of the rest of the internet) covered some research from MIT that uses a brain interface to help robots correct themselves when they’re about to make a mistake. This is very cool, very futuristic stuff, but it only works if you wear a very, very silly hat that can classify your brain waves in 10 milliseconds flat.

At Brown University, researchers in Stefanie Tellex’s lab are working on a more social approach to helping robots more accurately interact with humans. By enabling a robot to model its own confusion in an interactive object-fetching task, the robot can ask relevant clarifying questions when necessary to help understand exactly what humans want. No hats required. 

Whether you ask a human or a robot to fetch you an object, it’s a simple task to perform if the object is unique in some way, and a more complicated task to perform if it involves several similar objects. Say you’re a mechanic, and you want an assistant to bring you a tool. You can point at a shelf of tools and say, “Bring me that tool.” Your assistant, if they’re human, will look where you point, and if there are only a handful tools on the shelf, they’ll probably be able to infer what tool you mean. But if the shelf is mostly full, especially if it’s full of objects that are similar, your assistant might not be able to determine exactly what tool you’re talking about, so they’ll ask you to clarify somehow, perhaps by pointing at a tool and saying, “Is this the one you mean?”

To be useful in situations like these, your assistant has to have an understanding of ambiguity and uncertainty: They have to be able to tell when there is (or is not) enough information to complete a task, and then take the right action to help get more information when necessary, whether that uncertainty comes from the assistant not really getting something, or just from you not being specific enough about what you want. For robot assistants, it’s a much more difficult problem than it is for human assistants, because of the social components involved. Pointing, gestures, gaze, and language cues are all tricks that humans use to communicate information that robots are generally quite terrible at interpreting.

Brown researchers created a system called “FEedback To Collaborative Handoff Partially Observable Markov Decision Process,” or FETCH-POMDP, which is able to understand the implicit meaning of common gestures, and merge those meanings with what the person making the gestures is saying to improve its understanding of what the person wants. 

At Brown, they’ve created a painfully acronym’d system called “FEedback To Collaborative Handoff Partially Observable Markov Decision Process,” or FETCH-POMDP. It’s able to understand the implicit meaning of common gestures, and merge those meanings with what the person making the gestures is saying to improve its understanding of what the person wants. Assuming that the person is being cooperative (not lying about what they want), the system is able to model its own confusion and ask questions only when necessary for accuracy so as not to be unduly bothersome. 

To test out the FETCH-POMDP system, the Brown researchers asked people who had no idea what was going on to ask a Baxter robot to fetch things for them. A third of the time, the robot asked no questions; a third of the time, the robot always asked clarifying questions; and the final third of the time, the robot only asked questions when it decided that a question was necessary. The researchers expected that the robot would be fastest when asking no questions, and most accurate when always asking questions, but it turned out that the intelligent questioning approach managed to be both the fastest and most accurate. This is because human robot interaction is messy: People asking questions led to transcription errors (confusing “yes” with “hand,” for example), so more questions meant more misunderstandings.

Interestingly, the participants in the trials also ascribed all kinds of capabilities to the robot which it didn’t actually have:

During trials, many users used prepositional phrases in order to describe items, such as “Hand me the spoon to the left of the bowl.” Although the language model in this work did not account for referential phrases, the agent was able to use intelligent social feedback to figure out what the human desired. This may explain why many users reported that they thought the robot did understand prepositional phrases. Methods exist to interpret referential language, but problems in understanding will still occur. Our model will help correct those mistakes, regardless of the exact method of state estimation and language understanding.

The researchers do plan to update their model to include referential phrasing like this, and they also want to add things like eye tracking to improve accuracy even more. Adding the ability to place (along with pick) has the potential to make this system much more workplace useful, especially if you have a spaceship that needs fixing

Reducing Errors in Object-Fetching Interactions through Social Feedback,” by David Whitney, Eric Rosen, James MacGlashan, Lawson L.S. Wong, and Stefanie Tellex from Brown University will be presented at ICRA 2017 in Singapore.

Brown ]

Reducing Errors in Object-Fetching Interactions through Social Feedback,” by David Whitney, Eric Rosen, James MacGlashan, Lawson L.S. Wong, and Stefanie Tellex from Brown University will be presented at ICRA 2017 in Singapore.

资源下载链接为: https://pan.quark.cn/s/1f197bf22c2c 随着疫情防控形势的复杂多变,防疫物资的高效管理成为保障社会正常运转的关键环节。传统管理方式存在效率低、信息不透明等问题,难以满足现代疫情防控的动态需求。因此,开发一套基于现代化信息技术的防疫物资管理系统具有重要的现实意义。 本研究旨在通过构建一套高效的防疫物资管理系统,实现物资信息的实时更新与精准管理,提高物资调配效率,降低管理成本,为疫情防控工作提供有力支持,同时为相关领域的信息化建设提供参考借鉴。 本研究将围绕防疫物资管理系统的设计与实现展开,包括系统需求分析、系统设计、系统实现以及系统测试等环节,重点解决系统功能模块的开发与优化问题。 Eclipse是一款功能强大的集成开发环境,支持多种编程语言,为本系统开发提供了便捷的代码编写与调试工具。 JSP(Java Server Pages)是一种基于Java的动态网页技术,能够实现页面内容的动态生成,为系统的用户界面展示提供了技术支持。 B/S(浏览器/服务器)结构是一种网络应用架构,用户通过浏览器访问服务器上的应用程序,具有易于部署、维护方便等优点,适用于本系统的网络环境。 MySQL是一种开源关系型数据库管理系统,具有高性能、高可靠性和易用性等特点,为系统的数据存储与管理提供了可靠的解决方案。 Tomcat是一个开源的Java Servlet容器,支持JSP和Servlet技术,能够高效地运行Java Web应用程序,是本系统运行的重要支撑平台。 当前信息技术的快速发展为本系统的开发提供了成熟的技术基础,相关技术的稳定性和兼容性能够满足系统开发的需求。 系统开发所需的技术资源和硬件设备成本相对较低,且能够有效降低防疫物资管理的运营成本,具有良好的经济可行性。 系统能够提高防疫物资管理的效率和透明度,符合社会对疫情防控工作的需求,具
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值