Targeting a Robotic Brain Capable of Thoughtful Communication

The Hagiwara Lab in the Department of Information and Computer Science of Keio University's Faculty of Science and Technology is trying to realize a robotic brain that can carry on a conversation, or in other words, a robotic brain that can understand images and words and can carry on thoughtful communication with humans.

"Even now, significant progress is being made with robots, and tremendous advancements are being made with the control parts. However, we feel like R&D with regards to the brain has been significantly delayed. When we think about what types of functions are necessary for the brain, the first thing that we as humans do is visual information processing. In other words, the brain needs to be able to process what is seen. The next thing is the language information processing that we as humans implement. By using language capabilities, humans are able to perform extremely advanced intellectual processing. However, even if a robotic brain can process what it sees and use words, it is still lacking one thing, specifically, feelings and emotions. Therefore, as a third pillar, we're conducting research on what is called Kansei Engineering, or affective information processing."

The Hagiwara Lab has adopted an approach of learning from the information processing of the human brain. The team is trying to construct a robotic brain while focusing on three elements: visual information processing, language information processing, and affective information processing, and an even more important point is integrating these three elements.

"With regards to visual information processing, by using a neural network as well, we're trying to recognize items through mechanisms based on experience and intuition in the same manner that is implemented directly by humans without having to use three-dimensional structures or perform complicated mathematical processing. In the conventional object recognition field, patterns from the recognized results are merely converted to symbols. However, by adding language processing to those recognized results, we can comprehensively utilize knowledge to get a better visual image. For example, even if an object is recognized as being a robot, knowledge such as the robot has a human form, or it has arms and legs can also be used. Next will be language information processing because processing of language functions is becoming extremely important. For example, even as a robot, the next step would be for it to recognize something as being cute, not cute, mechanical, or some other type of characteristic. Humans naturally have this type of emotional capability, but in current robotic research, that type of direction is not being researched much. Therefore, at our lab, we're conducting research in a direction that enables robots to understand what they see, to use language information processing to understand what they saw as knowledge, and to then comprehensively use the perspective of feelings and emotions like those of humans as well."

The robotic brain targeted by the Hagiwara Lab is one that is not merely just smart. Instead, the lab is targeting a robotic brain with emotions, feelings, and spirit that will enable it to interact skillfully with humans and other environments. To achieve this, the lab is conducting a broad range of research from the fundamentals of Kansei Engineering to applications thereof in fields such as entertainment, design, and healing.

"Most of the robots thus far move exactly as they are programmed to do. However, within the next 10 years, and perhaps even sooner, I believe that robots will be steadily introduced into the home. And when that happens, the interface with humans, which are the users, will be extremely important. For example, if you have a robot that can undergo a variety of movements rather than being a robot like this that doesn't move, and if amongst those various movements, there is movement that looks like fluctuation, then communication is occurring amongst that movement, or if the contact time with the robot becomes longer, then of course the robot will be able to understand even the user's feelings and personality, and it can then respond and act accordingly. We're trying to build a robot that is capable of that type of attentiveness."

 

 

Source: 

Related Articles

Reno J. Tibke - July 04, 2014

The Best Transformers Movie of 2014... (VIDEO)

LISTEN INSTEAD OF READING
AkihabaraNews MICROCAST

A train festival will be held from August 9 through August 17 at LaLaport Yokohama in Yokohama, Kanagawa prefecture. The main display at the event will be a Shinkansen (bullet train) transforming robot called "Project E5".

Project E5 is made based on JR East's E5 series Shinkansen. It was exhibited to the public at the Tokyo Toy Show 2014, which took place in July, and became very popular.

Akihabaranews.com - Diginfo - Kosmek robotic hand changer

KOSMEK has developed a robotic hand changer that can switch between robot tools automatically - and with high precision. 

Associate Professor Toshiaki Tsuji's Laboratory at Saitama University has developed R-cloud, a rehabilitation support robot that enables users to view how their own muscles move during rehabilitation and training.

Baxter Coming to Japan - AKihabaraNews.com

Attention Kawada, Yaskawa, Nachi, Fanuc - Your American Nemesis is Coming!

Skeletonics at Maker Faire: Tokyo 2013 - AKihabaraNews.com

In addition to our exclusive close-up footage of Team Skeletonics' human-powered exoskeleton, luckily we also collected a few wide angle and close-up stills. Yeah, we've kinda been fawning over Skeletonics of late. But for good reason: you won't believe what they're working on.

Microtech Laboratory exhibited the ultra-small rotary encoder MES-6-125PST16C, a device that detects at high accuracy the rotational angle and speed of mechanical devices and motors, at International Robot Exhibition 2013.

"Typical rotary encoders are large industrial models mostly about palm-size, but this product is very small, able to fit on a fingertip."

Taihei Environmental Science Center has developed a system that automatically performs water quality testing, and can detect bacteria such as E. coli. In the past, this work had to be done manually. The system uses three robots to completely automate all stages of the process, from collecting a predetermined amount of the sample, injecting agar, agitating, coagulating and inverting, to the placement and storage of the culture in incubators. This level of automation not only eliminates human error, substantially improving test accuracy, it also increases processing capacity.

ASIMO Goess Hanami - AkihabaraNews.com
The Japanese and anyone else lucky enough to be here are celebrating an intermittently occurring secular holiday that spans an entire month. Various gatherings, parties, festivals, and no-real-reason-necessary events with groups from a just a few people to several thousand are currently being held across the entire nation. One famous robot, too.
Japanese Technology from the Future SATURDAY but was Supposed to Be Friday!

This week it’s Japan thinking about investing a bunch of cash and tech into cancer treatment in India, which is nice, NTT Data Canada has significantly upped some jobs predictions in Nova Scotia, and even we're saying "NTT Data Canada?", and SCHAFT Robotics has left the DARPA building, which is kind of a bummer, but intriguing. 

• • •

Okay, Trask Industries, We're Infected - What's a Sentinel Robot?
Taking a lesson from the very successful campaigns of films like District 9 and Prometheus, Twentieth Century Fox’s X-Men: Days of Future Past has hit the internets with a slick, sincere Trask Industries website. In the X-Men universe, Trask Industries makes 30ft/9m-tall robotic mutant hunters known as "Sentinels."

 

The Hairlytop Interface is an interactive surface display made of soft solid-state actuators and optical sensors which react to light. Jointly developed by the University of Electro-Communications and Symphodia Phil, when placed on an iPad that is playing a video, it moves organically, like a living thing, in response to changes in the brightness of the screen.

"One feature of this system is that the motion is very cute, like that of an animal. Another feature is that it can be used extremely freely in terms of design."

Auto Mee S Cleaning Robots - AkihabaraNews.com

【Published February 5, 2013 - Top 13 of '13

Takara TOMY Company, Ltd., the Japanese toy company, just released a tiny cleaning robot, called the “Auto Mee S”, designed for cleaning smartphone and tablet LCD displays.

This is a fun gadget to have and would be a great gift that I would love to get…

Pages