![](https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEigFfqlTUzKOUDO8NVzlp6B2LIr4cmE0T8TEcg2JtNykgOR3oG6Zjb-HJmkfZgXRqnMpCJWfEkCiq6o3kx-V75AGRaBY8G31-sU2JvFKfZ7InxdtEiz0iiPoBowGuqojU5ZxrcbZofrarsl/s400/oxford-english-dictionary-300x280.jpg)
Lexy Cal Am Big You Itty Rezo Loo Shin.
A computer/science illiterate attempts to understand the intelligence of artificial intelligence.
The most frequently used words in English are highly ambiguous: Webster’s Ninth New Collegiate Dictionary lists 94 meanings for the word “run” as a verb alone.
The word "bat" can describe an animal, a sports apparatus, the use of a sports apparatus, the blink of an eye, and more.
And yet computers are learning all the “runs”.
And all the “bats”.
And all the other words.
And all the other combinations of words.
And all the other combinations of words—in all their combinations.
Or are they?
How can a computer program be made to understand so many words in an almost infinite number of contexts --especially when those contexts are compounded by all the other mitigating conundrums of language like idiom and syntax?
If I knew the answer I’d be working at a LAR Lab.
LAR is Lexical Ambiguity Resolution ( appearing phonetically in the headline of this post) and it’s the central problem in natural language and computational semantics research. And it’s what’s behind all that voice recognition stuff that you get when you call customer service lines, and what Apple built into the new iPhone that let’s you find the nearest Spicy Salmon Roll with a simple query like “I’m in the mood for Sushi.”
Sidebar: When I looked up LAR I found a site that listed LAR as the acronym for 52 things, none of which was Lexical Ambiguity Resolution. Talk about context?
![](https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEiGWZku3Zx_R-pypFKJUcFSDpmFWxeQcBREtHNtQr0WE0cQQYkUK0NzWMbktPfQxbzcLhJKcickwrfowVO0Qfyj9SE7yTBhKkXIdWDOSRN84KbZcUzcQslGiZI924y5u2thQlXXhki0UG6L/s400/8437685-random-collection-of-different-words-and-word-forms-on-magnetic-tiles.jpg)
But programming a computer to “understand” context in language blows my mind. What would they make of:
“They passed the port as dawn rose”
Was that port by the river, or from the wine rack?
Did they pass it by or pass it around?
And did Dawn get a sip?
Or
“Visiting distant relatives can be inconvenient.”
Unless, of course, it's the other way around.
Or
“The rabbi married my sister who was looking for a match ”
Oy Vey.
So I did some reading and here’s what I learned.
“The problem is specifying the nature of context and how it interacts with the rest of an “understanding” system. “
“Advanced computer speech does not need all words to be programmed, it only needs all transitions between any pair of sounds (phonemes) in a language.”
Aha, so it’s sounds not words. It’s not really the language that has to be interpreted, it’s the sound of the components of the language.
Which to my mind begs the question:
Who’s making the sounds? A drawling pig farmer from Arkansas? A nasal soccer mom from Long Gi-land? A Cockney lad from Liverpool? George Bush? Me?
![](https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEj2u1JHFbYKtVuzaPi4LVv6Yk1DmOMgv4M9F6PGJA3oPrpTmwbYNCVjNqH7wivjw4gHzR_AUqcpXELtZhqICOLCp7ERdYWhyXrJQC5kT542-yNEux9mJ4sX-yKn2HntvQ31nOmgJxSzP-HF/s400/images.jpeg)
“The most complicated part of the programming is turning "human readable" text into phonemes. Sounds. Mistakes in this process are what account for the often hilarious results in GPS pronunciation.”
![](https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEgtMT1w4E55ImQ49ogZpnf8feumVQorYAATnuot5wQ3fsAdGAFY_gi6tEs7cSfBRGsrbWcrLcfZEXxPTJEQ4MCRc4Vdl4h278qk0hEGCOy9IfWtovEN7O5XLA4QSHzlHPuiIe_2zj8flYfW/s400/nascar7.gif)
(Most phonetics in the GPS-device market are produced by a company called the Acapela Group…gotta love that)
Good Luck LAR folk, but don’t mess with Ethel . She’s our portable dependable and entertaining Mrs. Malaprop and we like her just the way she are.
No comments:
Post a Comment