Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
letter
. 2019 Oct 15;116(44):21972–21973. doi: 10.1073/pnas.1912302116

Sighted people’s language is not helpful for blind individuals’ acquisition of typical animal colors

Markus Ostarek a,1, Jeroen van Paridon a, Guillermo Montero-Melis a,b
PMCID: PMC6825296  PMID: 31615880

How do you learn what things look like if you cannot see? Kim et al. (1) tackle this intriguing question by assessing knowledge about animal appearance in blind and sighted individuals. The authors evaluated 2 plausible hypotheses: The learn-from-description hypothesis that blind individuals learn directly from sighted people’s descriptions (e.g., “elephants are gray”) and the learn-from-kind hypothesis that blind people infer visual animal properties from knowledge they have about the animal’s taxonomic class (e.g., a crow is a bird and birds have feathers).

While group differences were observed for all visual properties, blindness had the largest effect on color knowledge: Only sighted participants consistently grouped animals with the same canonical color together. This striking difference between blind and sighted participants arose despite the finding that colors were the easiest to verbalize of all properties tested. From this the authors concluded that blind people do not use verbal descriptions (e.g., “elephants are gray”) as a primary source of information. This conclusion rests on the assumption that information about highly verbalizable properties (such as color) is conveyed in speech. Hence, if blind individuals learn from verbal descriptions produced by sighted people they should have no problem acquiring the canonical colors of animals.

We performed an analysis of cooccurrence statistics in a large corpus of spoken language (2) which revealed that this assumption is not met (Fig. 1): For 23 out of the 30 animals used in Kim et al. (1), the color mentioned most often was noncanonical (“white elephant”) rather than canonical (“gray elephant”), and in only 25% of all of the instances where animals were described as having a color was that color canonical. Thus, contrary to the authors’ claim, their results for color are compatible with the learn-from-description hypothesis; inconsistent descriptions are associated with inconsistent responses in blind individuals. The authors overlooked the fact that language use is geared toward efficiency (3) such that it avoids redundant information (people rarely talk about a “round ball”). Since indices of verbalizability as used by Kim et al. (1) do not appear to be a good proxy for the language input that people get, it will be imperative that future research systematically evaluates to what extent language input predicts what blind individuals know.

Fig. 1.

Fig. 1.

Probability of a color’s being used to describe an animal, conditioned on the occurrence of the animal itself, in a 750-million-word corpus of pseudoconversational speech (2). Our tally includes both explicit descriptive statements of the type “crows are black” and more implicit descriptions of the type “black crow.”

On a theoretical level, the alternative account proposed by Kim et al. (1), whereby blind people primarily learn via inferences from ontological kind, runs into a circular reasoning problem: Knowledge about ontological kind itself has to be learned and this is likely achieved via verbal input. Moreover, inferences from kind are limited in the specificity of knowledge that can be derived. Hence, the most plausible scenario is that learning from verbal descriptions and learning via inferences are deeply intertwined, so that one cannot happen without the other and one cannot be said to take precedence over the other when it comes to the acquisition of knowledge in the blind.

Acknowledgments

We thank the OpenSubtitles.org team for making their data available for research purposes. Funding for this work was provided by the Swedish Research Council Grant 2018-00245 (G.M.-M.).

Footnotes

The authors declare no competing interest.

Data deposition: The code for the corpus analyses and additional information have been deposited in Zenodo (https://doi.org/10.5281/zenodo.3406143).

References

  • 1.Kim J. S., Elli G. V., Bedny M., Knowledge of animal appearance among sighted and blind adults. Proc. Natl. Acad. Sci. U.S.A. 116, 11213–11222 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Lison P., Tiedemann J., “OpenSubtitles2016: Extracting large parallel corpora from movie and TV subtitles” in Proceedings of the 10th Annual Conference on Language Resources and Evaluation, Calzolari N., et al., Eds. (European Language Resources Association, Paris, 2016), pp. 923–929. [Google Scholar]
  • 3.Grice H. P., “Logic and conversation” in Syntax and Semantics, Volume 3: Speech Acts, Cole P., Morgan J. L., Eds. (Academic Press, 1975), pp. 41–58. [Google Scholar]

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES