Subscribe free to our newsletters via your
. 24/7 Space News .




ROBO SPACE
When Will My Computer Understand Me
by Aaron Dubrow, Science and Technology Writer
Austin TX (SPX) Jun 12, 2013


Katrin Erk, professor of linguistics at The University of Texas at Austin.

It's not hard to tell the difference between the "charge" of a battery and criminal "charges." But for computers, distinguishing between the various meanings of a word is difficult.

For more than 50 years, linguists and computer scientists have tried to get computers to understand human language by programming semantics as software. Driven initially by efforts to translate Russian scientific texts during the Cold War (and more recently by the value of information retrieval and data analysis tools), these efforts have met with mixed success.

IBM's Jeopardy-winning Watson system and Google Translate are high profile, successful applications of language technologies, but the humorous answers and mistranslations they sometimes produce are evidence of the continuing difficulty of the problem.

Our ability to easily distinguish between multiple word meanings is rooted in a lifetime of experience. Using the context in which a word is used, an intrinsic understanding of syntax and logic, and a sense of the speaker's intention, we intuit what another person is telling us.

"In the past, people have tried to hand-code all of this knowledge," explained Katrin Erk, a professor of linguistics at The University of Texas at Austin focusing on lexical semantics. "I think it's fair to say that this hasn't been successful. There are just too many little things that humans know."

Other efforts have tried to use dictionary meanings to train computers to better understand language, but these attempts have also faced obstacles. Dictionaries have their own sense distinctions, which are crystal clear to the dictionary-maker but murky to the dictionary reader. Moreover, no two dictionaries provide the same set of meanings - frustrating, right?

Watching annotators struggle to make sense of conflicting definitions led Erk to try a different tactic. Instead of hard-coding human logic or deciphering dictionaries, why not mine a vast body of texts (which are a reflection of human knowledge) and use the implicit connections between the words to create a weighted map of relationships - a dictionary without a dictionary?

"An intuition for me was that you could visualize the different meanings of a word as points in space," she said.

"You could think of them as sometimes far apart, like a battery charge and criminal charges, and sometimes close together, like criminal charges and accusations ("the newspaper published charges..."). The meaning of a word in a particular context is a point in this space. Then we don't have to say how many senses a word has. Instead we say: 'This use of the word is close to this usage in another sentence, but far away from the third use.'"

To create a model that can accurately recreate the intuitive ability to distinguish word meaning requires a lot of text and a lot of analytical horsepower.

"The lower end for this kind of a research is a text collection of 100 million words," she explained. "If you can give me a few billion words, I'd be much happier. But how can we process all of that information? That's where supercomputers and Hadoop come in."

Applying Computational Horsepower
Erk initially conducted her research on desktop computers, but around 2009, she began using the parallel computing systems at the Texas Advanced Computing Center (TACC). Access to a special Hadoop-optimized subsystem on TACC's Longhorn supercomputer allowed Erk and her collaborators to expand the scope of their research.

Hadoop is a software architecture well suited to text analysis and the data mining of unstructured data that can also take advantage of large computer clusters. Computational models that take weeks to run on a desktop computer can run in hours on Longhorn. This opened up new possibilities.

"In a simple case we count how often a word occurs in close proximity to other words. If you're doing this with one billion words, do you have a couple of days to wait to do the computation? It's no fun," Erk said. "With Hadoop on Longhorn, we could get the kind of data that we need to do language processing much faster. That enabled us to use larger amounts of data and develop better models."

Treating words in a relational, non-fixed way corresponds to emerging psychological notions of how the mind deals with language and concepts in general, according to Erk. Instead of rigid definitions, concepts have "fuzzy boundaries" where the meaning, value and limits of the idea can vary considerably according to the context or conditions. Erk takes this idea of language and recreates a model of it from hundreds of thousands of documents.

Say That Another Way
So how can we describe word meanings without a dictionary? One way is to use paraphrases. A good paraphrase is one that is "close to" the word meaning in that high-dimensional space that Erk described.

"We use a gigantic 10,000-dimentional space with all these different points for each word to predict paraphrases," Erk explained. "If I give you a sentence such as, 'This is a bright child,' the model can tell you automatically what are good paraphrases ('an intelligent child') and what are bad paraphrases ('a glaring child'). This is quite useful in language technology."

Language technology already helps millions of people perform practical and valuable tasks every day via web searches and question-answer systems, but it is poised for even more widespread applications.

Automatic information extraction is an application where Erk's paraphrasing research may be critical. Say, for instance, you want to extract a list of diseases, their causes, symptoms and cures from millions of pages of medical information on the web.

"Researchers use slightly different formulations when they talk about diseases, so knowing good paraphrases would help," Erk said.

In a paper to appear in ACM Transactions on Intelligent Systems and Technology, Erk and her collaborators illustrated they could achieve state-of-the-art results with their automatic paraphrasing approach.

Recently, Erk and Ray Mooney, a computer science professor also at The University of Texas at Austin, were awarded a grant from the Defense Advanced Research Projects Agency to combine Erk's distributional, high dimensional space representation of word meanings with a method of determining the structure of sentences based on Markov logic networks.

"Language is messy," said Mooney. "There is almost nothing that is true all the time. "When we ask, 'How similar is this sentence to another sentence?' our system turns that question into a probabilistic theorem-proving task and that task can be very computationally complex."

In their paper, "Montague Meets Markov: Deep Semantics with Probabilistic Logical Form," presented at the Second Joint Conference on Lexical and Computational Semantics (STARSEM2013) in June, Erk, Mooney and colleagues announced their results on a number of challenge problems from the field of artificial intelligence.

In one problem, Longhorn was given a sentence and had to infer whether another sentence was true based on the first. Using an ensemble of different sentence parsers, word meaning models and Markov logic implementations, Mooney and Erk's system predicted the correct answer with 85% accuracy. This is near the top results in this challenge. They continue to work to improve the system.

There is a common saying in the machine-learning world that goes: "There's no data like more data." While more data helps, taking advantage of that data is key.

"We want to get to a point where we don't have to learn a computer language to communicate with a computer. We'll just tell it what to do in natural language," Mooney said. "We're still a long way from having a computer that can understand language as well as a human being does, but we've made definite progress toward that goal."

.


Related Links
Texas Advanced Computing Center
All about the robots on Earth and beyond!






Comment on this article via your Facebook, Yahoo, AOL, Hotmail login.

Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle








ROBO SPACE
Firefighting Robot Paints 3D Thermal Imaging Picture for Rescuers
San Diego CA (SPX) Jun 11, 2013
Engineers in the Coordinated Robotics Lab at the University of California, San Diego, have developed new image processing techniques for rapid exploration and characterization of structural fires by small Segway-like robotic vehicles. A sophisticated on-board software system takes the thermal data recorded by the robot's small infrared camera and maps it onto a 3D scene constructed from th ... read more


ROBO SPACE
LADEE Arrives at Wallops for Moon Mission

NASA's GRAIL Mission Solves Mystery of Moon's Surface Gravity

Moon dust samples missing for 40 years found in Calif. warehouse

Unusual minerals in moon craters may have been delivered from space

ROBO SPACE
Marks on Martian Dunes May Reveal Tracks of Dry-Ice Sleds

UH Astrobiologists Find Martian Clay Contains Chemical Implicated in the Origin of Life

Mars Rover Opportunity Trekking Toward More Layers

SciTechTalk: Mars rover readies for 'road trip' on the Red Planet

ROBO SPACE
The Body Electric: Researchers Move Closer to Low-Cost, Implantable Electronics

TED conference sets stage for a week of bright ideas

NASA's Orion Spacecraft Proves Sound Under Pressure

Expert slams Congress over ban on U.S.-China space cooperation

ROBO SPACE
China astronauts enter space module

China to send second woman into space: officials

Tiangong-1 ready for docking and entry

Shenzhou-10 mission to teach students in orbit

ROBO SPACE
Russian cargo supply craft separates from International Space Station

Russian Space Freighter to Depart From Orbital Station

Star Canadian spaceman Chris Hadfield retiring

Experiments, Spacewalk Preps and Maintenance for Crew

ROBO SPACE
Mitsubishi Heavy and Arianespace conclude MOU on commercial launches

Sea Launch IS-27 FROB Report Complete

Europe launches record cargo for space station

New chief urges Ariane 5 modification for big satellites

ROBO SPACE
Kepler Stars and Planets are Bigger than Previously Thought

Astronomers gear up to discover Earth-like planets

Stars Don't Obliterate Their Planets (Very Often)

'Dust trap' around distant star may solve planet formation mystery

ROBO SPACE
Chilean, U.S. firms join effort to expand e-waste recycling

Space Debris - One Solution

Moon Radiation Findings May Reduce Health Risks to Astronauts

Sony eyes long game despite console launch triumph




The content herein, unless otherwise known to be public domain, are Copyright 1995-2014 - Space Media Network. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA Portal Reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. Advertising does not imply endorsement,agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement