Apr. 3, 2013 ? Humans favor speech as the primary means of linguistic communication. Spoken languages are so common many think language and speech are one and the same. But the prevalence of sign languages suggests otherwise. Not only can Deaf communities generate language using manual gestures, but their languages share some of their design and neural mechanisms with spoken languages.
New research by Northeastern University's Prof. Iris Berent further underscores the flexibility of human language and its robustness across both spoken and signed channels of communication.
In a paper published in PLOS ONE, Prof. Berent and her team show that English speakers can learn to rapidly recognize key structures of American Sign Language (ASL), despite no previous familiarity with this language.
Like spoken languages, signed languages construct words from meaningless syllables (akin to can-dy in English) and distinguish them from morphemes (meaningful units, similar to the English can-s). The research group examined whether non-signers might be able to discover this structure.
In a series of experiments, Prof. Berent and her team (Amanda Dupuis, a graduate student at Northeastern University, and Dr. Diane Brentari of the University of Chicago) asked English speakers to identify syllables in novel ASL signs. Results showed that these non-signing adults quickly learned to identify the number of signed syllables (one vs. two), and they could even distinguish syllables from morphemes.
Remarkably, however, people did not act as indiscriminate general-purpose learners. While they could easily learn to discern the structure of ASL signs, they were unable to do so when presented with signs that were equally complex, but violated the structure of ASL (as well as any known human language).
The results suggest that participants extended their linguistic knowledge from spoken language to sign language. This finding is significant because it shows that linguistic principles are abstract, and they can apply to both speech and sign. Nonetheless, Dr. Berent explains, language is also constrained, as not all linguistic principles are equally learnable. "Our present results do not establish the origin of these limitations -- whether they only result from people's past experience with English, or from more general design properties of the language system. But regardless of source, language transcends speech, as people can extend their linguistic knowledge to a new modality."
Share this story on Facebook, Twitter, and Google:
Other social bookmarking and sharing tools:
Story Source:
The above story is reprinted from materials provided by Northeastern University College of Science, via EurekAlert!, a service of AAAS.
Note: Materials may be edited for content and length. For further information, please contact the source cited above.
Journal Reference:
- Iris Berent, Amanda Dupuis, Diane Brentari. Amodal Aspects of Linguistic Design. PLOS ONE, 2013; 8 (4): e60617 DOI: 10.1371/journal.pone.0060617
Note: If no author is given, the source is cited instead.
Disclaimer: This article is not intended to provide medical advice, diagnosis or treatment. Views expressed here do not necessarily reflect those of ScienceDaily or its staff.
cujo karen handel hangout todd haley kareem abdul jabbar miramonte elementary school mark jenkins
কোন মন্তব্য নেই:
একটি মন্তব্য পোস্ট করুন