Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2021 Nov 8;16(11):e0258376. doi: 10.1371/journal.pone.0258376

Where to draw the line?

Heping Sheng 1, John Wilder 2, Dirk B Walther 2,*
Editor: Markus Lappe3
PMCID: PMC8575256  PMID: 34748556

Abstract

We often take people’s ability to understand and produce line drawings for granted. But where should we draw lines, and why? We address psychological principles that underlie efficient representations of complex information in line drawings. First, 58 participants with varying degree of artistic experience produced multiple drawings of a small set of scenes by tracing contours on a digital tablet. Second, 37 independent observers ranked the drawings by how representative they are of the original photograph. Matching contours between drawings of the same scene revealed that the most consistently drawn contours tend to be drawn earlier. We generated half-images with the most- versus least-consistently drawn contours and asked 25 observers categorize the quickly presented scenes. Observers performed significantly better for the most compared to the least consistent half-images. The most consistently drawn contours were more likely to depict occlusion boundaries, whereas the least consistently drawn contours frequently depicted surface normals.

Introduction

Humans have used line drawings to depict scenes from their lives for at least 45,500 years [1]. Line drawings can capture the essence of shapes and spatial relationships in complex scenes to an extent that makes them not only reasonable stand-ins for photorealistic images, but can emphasize particular aspects of the scene unencumbered by other, extraneous features that might otherwise obscure the clarity arising from a purely contour-driven depiction. For this reason, line drawings are used for technical drawings, such as architectural plans, designs for complex machinery, or charts for complex production processes. Even though line drawings lack texture and color, they nevertheless represent essential information, making them suitable for efficiently conveying visual information non-verbally, for instance in signs that request patrons to wear masks inside stores, instructions for the Heimlich maneuver, or step-by-step instructions for assembling a MALM chest of drawers from IKEA.

Line drawings of real-world scenes are recognized as quickly and accurately as photographs [2, 3]. Perception of line drawings is innate to human children [46] and does not rely on an acquired cultural framework [7, 8]. Even chimpanzees were found to understand line drawings without prior training [9]. In their first attempts to artistically reflect their world, children resort to line drawings [10]. Attempts at photorealistic painting tend to come at a later age. Neural representations of scenes, objects and faces elicited by line drawings were found to be equivalent to those elicited by the corresponding color photographs in visual cortex [1113], although the equivalence for scenes appears to be restricted to the level of scene categories, not scene identity [14]. These similarities of line drawings to photo-realistic depiction provide interesting insights into visual brain functions [15].

However, not all line drawings convey visual information equally well. The quality of a drawing, and thereby its usefulness to convey visual content, depends on the experience and on the particular choices of the artist. Lines are placed for different reasons, e.g., to indicate the boundaries of objects, structure within objects, occlusion of one object by another, texture elements, or the outlines of shadows cast by objects [1619]. We here demonstrate that some lines are more important than others for conveying visual information, and we establish a tight relationship between the importance of contour lines, artistic expertise, and the ability of line drawings to convey scene content. In a series of three experiments, participants with varying degrees of artistic experience were asked to trace the important contours in a set of photographs of complex real-world scenes (Experiment 1). Then, an independent group of participants rank-ordered the drawings belonging to the same photograph, allowing us to establish a measure of contour importance (Experiment 2). Finally, we test the effect of contour importance for scene categorization (Experiment 3).

Experiment 1: Tracing scene photographs

In this experiment we seek to determine how consistent individuals are at drawing lines as well as effects of artistic expertise on tracing photographs. To this end, we performed a controlled drawing experiment, in which we asked participants with varying degrees of artistic experience to trace the important contours in a set of 18 photographs of real-world scenes. We chose a tracing task rather than free drawing or copying the scenes from the photographs so that we could match the contours in the drawings by different participants based on their spatial location. The task still gives participants freedom to decide which lines they deem important enough to include in their drawings.

Methods

Participants

We recruited two groups of participants: 14 students majoring in visual arts or architecture from the University of Toronto and the Ontario College of Art and Design University (artists) participated in a two-hour drawing experiment for monetary compensation. In addition, 44 undergraduate students majoring in psychology at the University of Toronto (non-artists) participated in the study for one hour for course credit. The sample size of the artist group was determined by the number of artists who responded to our call for participation. We chose the size of the control sample to approximately equate the total number of drawings produced by the two groups.

We asked all participants to indicate how many years of experience in producing visual arts they had, how many full-year arts courses they had taken, and how much time they typically spend drawing per week. All participants reported normal or corrected-to-normal vision and provided written informed consent. The experiment was approved by the Research Ethics Board of the University of Toronto (Protocol #30999) and followed the guidelines set out in the Declaration of Helsinki.

Images

The 18 images used in this experiment were chosen from a set of photographs of real-world scenes [13]. The images have been rated by an independent group of participants on how well they represent one of six scene categories: beaches, city streets, forests, highways, mountains, and offices. Since knowledge of category membership may aid people in constructing a clearer representation of the scene [20], the top three exemplar photographs from each scene category were chosen for a total of 18 stimulus images. Four images distinct from the experimental categories were used for practice trials: a regular concave decagon (five-pointed star), a line drawing of a banana, a photograph of a leaf in isolation, and a photograph of penguins with background (Fig 1, left).

Fig 1. Setup for Experiment 1.

Fig 1

Left: Images similar to these were use as practice stimuli to familiarize participants with the tracing task and program interface. Images under Creative Commons Licence by Supachoteamorn, Aryadna Pons, Ksanyi, and Christopher Michel. Right Experiment setup with a pen display graphics tablet. The computer screen displayed instructions and the original 800x600 pixels photograph as reference during the trial. The graphics tablet displayed a 1440 x 1080 pixels copy of the image with a semi-transparent overlay for tracing. The Next, Finish, Show, and Undo functions corresponding to each button on the side of the graphics table were displayed during the entire experiment. Images by the authors. The set of 18 photographs are available at: https://osf.io/x9uj5/.

Procedure

Participants were seated at a computer station, equipped with a cathode ray tube monitor (Dell) as well as a Wacom Cintiq 13HD Interactive Pen Display graphics tablet. The dual-screen experiment was programmed using the Psychophysics Toolbox in Matlab [21], running under the Linux operating system. The CRT computer screen displayed instructions and the original 800x600 pixels photograph as reference during each trial, while the graphics tablet displayed a 1440x1080 pixels copy of the original reference image overlaid with a blank white sheet at 50% transparency for tracing.

Participants received both verbal and written instructions as well as a demonstration on how to use the graphics tablet at the beginning of the experiment. They started each contour line by pressing a button on the side of the graphics pen. They could draw smooth curves as well as set anchor points for quickly making straight edges while holding down the button. The latter function is similar to the Polygonal Lasso Tool in Adobe Photoshop. Traced contours were rendered in black on the semi-transparent overlay. There are 4 control buttons on the non-dominant-hand side of the tablet: NEXT and FINISH for navigating between trials, UNDO for deleting the most recent contour, and SHOW for displaying the drawing without the underlying tracing image (Fig 1). Participants were asked to draw clean outlines of the original image and to avoid extraneous features including repeated lines, shading or labels.

Participants performed four practice trials with stimuli of increasing complexity, intended to familiarize them with the tracing task and the software interface (Fig 1, right). Then, before the experimental trials began, they were given the following instructions:

For every image, please annotate all important and salient lines, including closed loops (e.g., boundary of a monitor) and open lines (e.g., boundaries of a road). Our requirement is that, by looking only at the annotated line drawings, a human observer can recognize the scene and salient objects within the image.

During each trial, participants had up to 10 minutes to trace a photograph from the stimulus set. They could proceed at their own pace by finishing a drawing early, or taking a break before the start of the next trial. Non-artists completed 5 trials. Artists were asked to trace as many pictures as they could comfortably do in 2 hours. The 18 stimuli were presented in random order to minimize the influence of practice and fatigue. We recorded spatial coordinates and timing of the strokes for all lines in the final drawings.

Results

One artist was excluded from the analysis, because she obviously did not make an effort to produce recognizable drawings. The remaining 13 artists (12 women; age 18–48, mean age 25; 3 left-handed) reported to have 4 to 32 years of arts experience (median 10), had taken 0.5 to 20 full-year arts courses (median 4.5), and draw for 1 to 52 hours per week (median 5). We excluded two non-artists from the analysis–one for adding handwritten words to the drawings, and one for exceedingly sloppy work. The remaining 42 non-artists (26 female, age 17–24 years, mean age 19 years; 2 left-handed) reported up to 12 years of arts experience (median 1) but did not report having taken any arts courses or regularly practicing drawing.

As expected, artists had more years of self-reported art experience, with a median of 10 years compared to 1 year for controls (t(12) = 5.75, p < 0.001, two-sample t test). The median number of full year-equivalent drawing-related courses taken by artists was 4.5 (range 0.5–18) and they spent a median 5 hours/week practicing drawing (range 1–52). The artist group was older (25.23 versus 19.24 years), more female predominant (92% vs 62%), and less right hand dominant (77% vs 95%).

We collected a total of 194 line drawings from artists and 159 from non-artists. We added the 18 drawings commissioned from trained artists for [13] for a grand total of 371 line drawings, with 15 to 24 (median 21) drawings for each photograph.

Experiment 2: Ranking line drawings

The quality of the line drawings collected in Experiment 1 varied considerably. For an impartial assessment, we performed a ranking experiment with an independent group of observers. Furthermore, we establish the match of contours between drawings of the same scene. The match allows us to count how frequently a particular contour was drawn across participants in Experiment 1. We use this drawing frequency as a proxy for the importance of contour lines.

Methods

Participants

We recruited a separate group of 37 undergraduate students of psychology from the University of Toronto. Twenty-five participants spent one hour on the experiment in lieu of partial course credit. Twelve participants volunteered their time and spent less than one hour, performing fewer ranking trials. This sample size was estimated to produce approximately 10 rankings for each of the 18 original images. Participants were unfamiliar with the photographs and line drawings in the study. All participants reported normal or corrected-to-normal vision and provided written informed consent. The experiment was approved by the Research Ethics Board of the University of Toronto (Protocol #30999) and followed the guidelines set out in the Declaration of Helsinki.

Ranking procedure

We printed all 371 line drawings from Experiment 1 on cardstock (15 x 10 cm) and grouped them according to the identity of the original photographs. Each printed drawing was labeled on the back with a unique identifier. We also printed the 18 original color photographs for reference. Participants were asked to sort each set of drawings from most to least representative of the photo by physically arranging the printed drawings on a table and writing down the identifiers in order. To reduce the difficulty of comparing up to 24 drawings at once, raters were instructed to first rank 5 randomly selected drawings, and then insert each additional one into its appropriate location in the list (Fig 2A). Participants proceeded through randomly selected sets of images at their own pace, finishing between 1 and 7 sets (median 6).

Fig 2. Procedures for Experiment 2.

Fig 2

(A) Ranking of line drawings was performed by re-arranging images printed on cardstock on a large table. Participants were instructed to first rank five randomly chosen drawings, and to then insert subsequent drawings at the appropriate position. (B) Constructing a super-reference. The highest-ranked drawing served as the initial reference. The other drawings were matched to the reference in ranking orders. Unmatched contours were added to the reference, eventually yielding a super-reference that contains the super set of contours from all drawings. Original photograph under Creative Commons License by Flickr user Cyndy Sims Parr.

Data analysis

After collecting all rankings, each drawing was assigned a normalized rank score, with 100% indicating high rank and 0% low rank with respect to all drawings produced for a particular photo. To examine the consistency among raters, we computed Kendall’s coefficient of concordance for all rankings of drawings for each image, and performed significance testing using chi-squared statistics. Trials that did not result in a complete ranking of a set of drawings were excluded from the analysis.

For further analysis, we averaged the normalized rank scores for each drawing across raters. The most representative drawing for each image was chosen as the reference for subsequent contour matching between drawings of the same original photograph.

We examined the factors that may have contributed to the ranking with a multiple linear regression analysis. We sought to predict ranking of a particular line drawing based on the artistic experience of its author and the total number of pixels in the drawing, as it is plausible that the amount of detail in a drawing also affects perceived representativeness.

Contour matching

To identify which contour lines participants in Experiment 1 most agree on when depicting complex natural scenes, we developed an algorithm to match contour lines. Given two drawings of the same photograph, a human observer can easily identify lines in each drawing depicting the same element, even though the strokes may vary somewhat in length, number, spatial location and shape. Here we strive to replicate this ability algorithmically.

Let us assume that we have two line drawings, a reference (R) and a query (Q). For each contour in Q we wish to determine the contour in R that is the closest match. In our drawings, contours are composed of a contiguous set of straight line segments. To make the analysis of long line segments easier, we split segments longer than 10 pixels into equal-length shorter segments such that all line segments are at most 10 pixels long.

Next, we calculated the Euclidean distance of a given query line segment to each reference line segment. We use the center point of each line segment for this computation. The reference line segment with the shortest distance is considered a match as long the distance is at most 30 pixels. The identifier of the corresponding reference contour is stored with the query contour segment. Any contour segments more than 30 pixels away from the closest reference contour are considered to be without a match in the reference.

This computation potentially leads to conflicting assignments along the length of the query contour. That is, different segments of the same query contour could be assigned to different reference contours. We resolve such conflicts with a two-stage voting process. First, we implemented a majority vote with a sliding window of 25 segments across the query contour. That is, the reference contour identifier most frequently assigned to the individual line segments within the window along the query contour was assigned to the contour segment at the center of the window. This voting process is analogous to median filtering for noise reduction in image processing [22]. Remember that contour segments are at most 10 pixels long, so the sliding window in this procedure is at most 250 pixels wide.

Second, another global majority vote was then applied to all line segments within the query contour to determine the best matching reference contour. This process usually led to a unique assignment of the reference contour and only rarely led to a tie. When ties occurred, they were broken by determining which of the contending reference contours was closest to the query contour over its entire length. Specifically, we computed the sum of the distances of the individual query line segments to the corresponding reference line segment, weighted by the length of the query line segments. The identifier of the reference contour with the smallest sum (the closest overall distance) was assigned to the query contour.

This matching procedure takes into consideration differences in stroke length, number and spatial location as well as minor variations in shape. This method also takes into account the possibility that one contour in the reference is depicted as more than one contour in a query, but does not account for the reverse.

Constructing the super-reference

We used the contour matching algorithm in order to construct a super-reference that contained the superset of all contours drawn by any of the individuals in Experiment 1 for a given photograph. In the first step, we used the highest-ranked drawing as the reference and the second highest drawing as the query and proceeded with the matching algorithm as described above. Any query contours that could not be matched to the reference were added to the reference. Next, the third-highest ranked drawing was used as the query. Any query contours that did not match any of the contours in the updated reference were added to the reference, and so on, until all contours in all drawings were accounted for. As a result, this super-reference captures the superset of all contours, i.e., all non-repeating contours for each scene (Fig 2B).

In a second round of matching, each drawing was matched to the super-reference. In this round, we counted the number of times each contour in the super-reference was matched with a contour in a query drawing. Dividing the count by the number of drawings for this particular image resulted in a normalized measure of the frequency with which a particular contour was depicted. We call this number the consistency score.

Results

Each set of line drawings was ranked by 5 to 12 participants (median 9). Participants strongly agreed on how much each line drawing represents the original photograph, as evidenced by the significant coefficients of concordance that ranged between 0.57 and 0.92 (Table 1). All 18 super-references are shown in Fig 3, with color of the contours coding for their consistency scores.

Table 1. Inter-rater agreement of line drawing rankings.

Image ID Kendall’s W df p-value
Beach 1 0.57 18 7.62·10−12
Beach 2 0.67 20 6.52·10−14
Beach 3 0.60 17 1.60·10−10
City 1 0.82 21 0
City 2 0.88 22 0
City 3 0.85 21 0
Forest 1 0.79 14 0
Forest 2 0.72 20 0
Forest 3 0.64 23 0
Highway 1 0.72 20 0
Highway 2 0.68 16 4.96·10−6
Highway 2 0.69 22 0
Mountains 1 0.92 22 0
Mountains 2 0.87 21 0
Mountains 3 0.67 17 2.22·10−16
Office 1 0.85 20 0
Office 2 0.83 19 0
Office 3 0.81 20 0

Fig 3. Super-references for all 18 scene images.

Fig 3

We performed a multiple linear regression analysis to predict average normalized rank scores (SR) of individual line drawings based on artistic experience (X) of the author of the drawing (measured in number of years of artistic experience) and based on the total number of pixels (NP) in the drawing as a covariate. We excluded four drawings by one outlier artist with 32 years of experience from the analysis. The range of artistic experience without this artist was 0–17 years. We also did not include the drawings originally commissioned for [13], since we had no information about the authors of these drawings.

We found a significant regression equation (F(3, 345) = 104.8, p < 0.001) with an adjusted R2 of 0.472 (Fig 4). The resulting regression equation is:

S^R=3.214+2.092X+2.633NP0.045XNP (1)

where artistic experience X is expressed in years and NP in 1000s of pixels. Artistic experience and total number of pixels were significant predictors of rank with positive coefficients (Table 2). That is, more detailed drawings (with more pixels) and drawings by artists with more experience were ranked more highly. The two factors contributed to the ranking of the drawings independently, as there was no significant interaction.

Fig 4. Average rank of drawings versus the ranks predicted by the multiple linear regression model.

Fig 4

The regression line is shown in red with the 95% confidence interval in gray.

Table 2. Regression table for the regression in Eq 1.

coefficient SE t-stat p-value
Intercept -3.214 3.902 -0.824 0.411
Artistic Experience 2.092 0.626 3.344 9.17·10−4
Number of Pixels 2.633 0.246 10.719 < 2·10−16
Experience*Pixels -0.045 0.037 -1.228 0.220

If individuals show such high agreement on a certain set of contours, do they also draw these lines earlier in the trial? To answer this question, we divided the time when a line was drawn within a trial by the total duration of the trial, resulting in relative timing in the range between zero and one for each of the 353 drawings produced in Experiment 1. Since we did not have timing information for the 18 drawings from [13], we excluded these drawings from the analysis. Consistency scores of contours in the super-reference drawings were mapped back to the matching contours in the individual drawings.

We fitted a linear mixed-effects model to explain the relationship of consistency scores of individual contours (58480 contours from 353 drawings) with relative timing (fixed effect), with drawing identity as a random-effect covariate. We found a highly significant negative effect of timing (β = -0.206; 95% confidence interval = [-0.213,-0.199]; t(58478) = -55.88; p < 0.001), with a smaller random effect of image identity (β = 0.101; CI = [0.093, 0.109]). This result shows that the most consistent lines tend to be drawn earlier, suggesting that both the consistency and the timing of drawings reflect their importance for perception. We test this hypothesis explicitly in Experiment 3.

Experiment 3: Effect of line consistency for scene perception

In Experiment 2, we computed the consistency score by matching the drawings of the same photograph by several participants to a super-reference drawing. Does a high consistency score for a particular contour mean that the contour is more important for representing scene content? We address this question with a fast scene categorization experiment with stimuli designed to maximize the difference in consistency scores.

Since contours represent what participants deemed to be important and salient edges in the scene, we hypothesized that the frequency at which a contour is drawn across participants is a good measure of its relative importance for perceiving the scene. To test this hypothesis, we used the consistency scores computed for the super-references in Experiment 2 to construct half-drawings that contained either the most or the least consistent half of the contours. We used these half-drawings as stimuli for a six-alternative forced-choice (6AFC) scene categorization experiment. We hypothesized that the most consistent half-drawings would enable more accurate recognition compared to the least consistent half drawings.

Furthermore, we sought to determine what distinguishes more important contour lines from less important ones. Contours in a drawing can have different physical causes in the real scene. For instance, they can represent depth discontinuities, such as occlusion boundaries, relate to changes in surface curvature, or be part of textures. We here investigate whether any of these physical roles are more important for conveying scene information than others in drawings of complex real-world scenes.

Methods

Participants

Twenty-five undergraduate students of Psychology (ages 17–28, mean 18.9; 16 female) at the University of Toronto participated in the study for partial course credit. This experiment was performed as an add-on for another, similarly structured main experiment. The sample size was determined based on the requirements of the main experiment.

All participants reported normal or corrected-to-normal vision and provided written informed consent. The experiment was approved by the Research Ethics Board of the University of Toronto (Protocol #30999) and followed the guidelines set out in the Declaration of Helsinki.

Stimuli

In order to examine the perceptual characteristics of the most-consistently and least-consistently drawn contours, we divided each of the 18 super-references into a high- and a low-consistency half-drawing according to their consistency scores as computed in Experiment 2. The two drawings did not share any contours, and each contained approximately 50% of the total pixels in the original super-reference (Fig 5A). As a result, we obtained 36 half-line drawings, 18 with the most consistent contours (Fig 6A) and 18 with the least consistent contours (Fig 6B).

Fig 5. Design of Experiment 3.

Fig 5

(A) Construction of half drawings according to consistency scores. The two drawings did not share any contours, and each contained approximately 50% of total pixels. (B) Experimental procedure for the six-alternative forced-choice scene categorization experiment.

Fig 6. Stimuli for Experiment 3.

Fig 6

(A) Most consistent half-drawings. (B) Least consistent half-drawings.

Procedure

The experiment was performed on a personal computer with Matlab and the Psychophysics Toolbox [21] running on Windows 10. A cathode ray tube monitor (Dell) was used to display the experiment at a resolution of 800x600 pixels with a refresh rate of 150 Hz. Participants were seated approximately 57 cm from the screen.

Participants performed the experiment as an add-on to another, similarly structured main experiment, which will be reported elsewhere. For the main experiment, participants were asked to categorize briefly presented line drawings of real-world scenes into six categories (beaches, city streets, forests, highways, mountains, offices) by pressing one of six buttons on a keyboard (s, d, f for the left hand and j, k, l for the right hand). The assignment of keys to categories was randomized for each participant.

Each trial of the experiment started with a 500 ms fixation period, followed by a brief presentation of the target image, followed by a perceptual mask for 500 ms. The mask consisted of randomly distributed black lines on a white background. After the mask, a blank screen was displayed until a response key was detected (Fig 4B).

In an initial training phase, images were presented for 233 ms (stimulus onset asynchrony, SOA). Participants heard a low tone when they made an error. When participants responded correctly in 17 of the last 18 trials or in 72 trials in total, they moved on to the ramping phase. In the ramping phase, the SOA was linearly decreased from 200 ms to 33 ms over the course of 54 trials. During the subsequent testing phase (360 trials), the SOA was fixed to 53 ms, and participants no longer received feedback.

Following the testing phase for the main experiment, participants performed an additional block of 36 test trials with the half-drawings created for this experiment. Key assignment, SOA, and temporal structure of the experiment remained the same as in the main experiment (see Fig 5). Accuracy of categorizing the drawings was recorded separately for high- and low-consistency drawings and compared using a paired t test.

Contour types

To analyze contour types, the first author manually labelled each contour in the super-reference drawings by sequentially overlaying them over the original photograph. She classified contours into four different edge types according to their physical cause [1618]: texture/albedo edges (change in reflectance across smooth surface), occlusion/depth boundaries (boundaries of objects) [23], surface normal discontinuities (intersecting surfaces, ridges and valleys), and shadow edges (boundary of cast shadows) [19]. For cases where a contour included more than one type of origin, the type that corresponds to the longer portion was chosen. Any contours that could not be assigned clearly to one of these types were labeled as “other”. The “other” categories includes “suggestive contours,” which occur in locations where surfaces bend away from the observer but do not form a true depth discontinuity [24]. Identifying suggestive contours with certainty occurred too rarely in the real-world complex scenes in this study to justify its own contour category. Once the contours were categorized, we totalled the number of pixels in contours belonging to each type within the most and least consistent half line drawings, and performed significance testing with a fixed-effects two-way ANOVA followed by Bonferroni-corrected paired t tests.

Results

Participants categorized the half-drawings containing the contours with the lowest consistency scores with 43% accuracy (chance: 16.7%) and the half-drawings with the high-consistency contours with 63% accuracy (Fig 7A). The difference was highly significant (t(14) = 4.21; p < 0.001).

Fig 7. Results of Experiment 3.

Fig 7

(A) Categorization accuracy for least and most consistent half drawings. (B) Number of pixels in particular types of contours in least and most consistent half drawings. ** p < 0.01, *** p < 0.001.

What makes the most consistently drawn contours so much better at conveying scene information? We annotated all contours in the super references according to their physical cause and analyzed the types separately for the least and the most consistent half drawings (Fig 7B). A 2x5 analysis of variance (ANOVA) with most/least consistent and contour type as factors showed a significant main effect for contour type (F(4,170) = 79.59, p < 0.001) but not for most/least consistent half drawing (F(1,170) = 0.014, p = 0.906), as expected since the half drawings are by design equated in the total number of pixels. Importantly, we found a significant interaction (F(4,170) = 10.98; p < 0.001). The most consistent half drawings contained significantly more pixels belonging to occlusion boundaries, typically object boundaries, than the least consistent half drawings (t(17) = 8.346, p < 0.001; Bonferroni-corrected for multiple comparisons). This was balanced by more pixels for the least than the most consistent half drawings belonging to contours associated with surface normals (t(17) = -4.28, p = 0.00254) and contours that could not be clearly assigned (t(17) = -4.78, p < 0.001). There was no difference in the number of pixels associated with texture boundaries (t(17) = -1.94, p = 0.348) or cast shadows (t(17) = -1.65, p = 0.588).

Discussion

In the three experiments presented here we have established that (1) line drawings are rated as more representative of the depicted scene when they are drawn by experienced artists; (2) the most consistently drawn contours are drawn earlier; (3) drawings containing the most consistently drawn contours are more recognizable than drawings with the least consistently drawn contours; and (4) the most consistent contours are more likely than inconsistent contours to represent occlusion boundaries, that is, the boundaries of objects.

While these results may seem unsurprising in hindsight, this is the first time, to our knowledge, that the ability of humans of various levels of artistic expertise to convey essential information in line drawings has been quantitatively measured for complex real-world scenes. Specifically, we used artistic expertise in the participant populations as a control variable to confirm that artistic experience does indeed result in drawings that are objectively more recognizable. To this end, we established an algorithmic measure of the importance of contours by determining how consistently contours were drawn across all participants. More consistently drawn contours tended to be drawn earlier in the drawing process. Importantly, the fast scene categorization task in Experiment 3 established the perceptual advantage in an unbiased, objective way, since participants in that study were not aware of the authorship of the stimulus drawings. Finally, we found that this perceptual advantage was tied to occlusion boundaries, which represent the shapes of objects in the scenes (Fig 7B).

Recent work on human sketches of individual objects showed similar results for the most agreed-upon contours as well the temporal order of strokes [25]. Interestingly, these experiments included a freehand sketch condition, which resulted in similar profiles of contour agreement as well as temporal order of contours as the tracing condition.

Human proficiency at perceiving objects in complex scenes has been previously linked to edges created from surface normals and depth boundaries in studies with synthetic scenes [16] as well as isolated objects [17, 18, 25] and portraits [26]. Specifically, contours drawn by artists were found to be in good agreement with one another [17], contours were found to be placed at locations optimal for depicting 3D shape and aesthetic quality [18], and artists’ drawings were found to define the convex hull area of the depicted objects early on in the drawing process [25]. Our analysis of the roles of contours in Experiment 3 confirms these findings in a real-world scene setting. These findings are consistent with the important role of contour junctions for the perception of objects and scenes, as contour junctions serve as a low-level cue to spatial relations such as occlusion [2729]. Moreover, we show how artists prioritize contours that lead to drawings which are more representative of the depicted scene (Eq 1) and that lead to better perception of scene gist (Fig 7A). This finding is likely related to the technique of “blocking-in”–a coarse, block-like outline of the proportions of figures and objects in the initial phase of drawing [30]. This technique may lead trained artists to initially prioritize contours that convey global shape over contours that convey finer details (see also [25]).

While our work involves tracing an image, previous work has shown that when asked to produce a line drawing of a recently viewed scene, non-artists draw many of the object boundaries, and in roughly the correct location [31]. This suggests that humans intuitively know which lines are important to convey the meaning of a scene, but that artistic training improves upon this ability, including improvements in the efficiency in object encoding [18, 32]. Further supporting this notion, aphantasia has been linked specifically to deficits in correctly recollecting objects in a drawing task [33].

Recent computer vision work has demonstrated that non-artists can be trained to prioritize the most important lines and draw them earlier than unimportant lines [34]. Furthermore, even artificial sketch generators, when trained to create a sketch to convey the essence of an image with as few strokes as possible, learn to first draw lines that convey global structure and shape, prior to any details [34]. In fact, there have recently been a number of artificial neural networks trained to generate sketches that are as easily recognizable as those generated by a human [35, 36]. We here show that drawings that take advantage of the visual system’s mechanisms for understanding scenes will be more easily interpreted [37].

It is important to point out that we are here not addressing any issues of artistic style, artistic expression or their relationship to perceived aesthetic value. In particular, we make no claims of labeling the line drawings created in this highly controlled study as “artwork.” Visual artistic expression, although sometimes concerned with faithful representation of the real world, involves many more aspects, such as composition, emotional content, and frequently metaphorical allusions that transcend the figurative content of the physical artwork. Nevertheless, our findings may help to illuminate what attributes make artwork recognizable, often despite extreme distortions or extreme simplification–for example, gesture drawings or Cubist paintings.

To conclude, we have presented a set of controlled experiments on the production and perception of line drawings for conveying the content of complex real-world scenes. We found that contours drawn most consistently across individuals are most effective at conveying scene content, that contours drawn earlier in the drawing process show higher consistency, and that trained artists are more likely to draw consistent contours. More consistent contours are more likely than less consistent contours to convey occlusion boundaries, which signal the shape of objects in a scene as well as their spatial relationships.

Acknowledgments

We thank Profs. Natalie Waldburger and Amy Swartz for insightful discussions and for help with recruiting participants from OCAD University.

Data Availability

All data and data analysis scripts as well as all individual line drawings and the super-references are available on the Open Science Framework website: https://osf.io/x9uj5/.

Funding Statement

This work was supported by grants by the Natural Sciences and Engineering Research Council of Canada (RGPIN-2020-04097) and the Social Sciences and Humanities Research Council of Canada (430-2017-01189) to DBW. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Brumm A., Oktaviana A.A., Burhan B., Hakim B., Lebe R., Zhao J., et al., Oldest cave art found in Sulawesi, Science Advances. 7 (2021) eabd4648. doi: 10.1126/sciadv.abd4648 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Biederman I., Ju G., Surface versus edge-based determinants of visual recognition., Cognitive Psychology. 20 (1988) 38–64. doi: 10.1016/0010-0285(88)90024-2 [DOI] [PubMed] [Google Scholar]
  • 3.Potter M.C., Levy E.I., Recognition memory for rapid sequences of pictures, Journal of Experimental Psychology. 81 (1969) 10–15. doi: 10.1037/h0027470 [DOI] [PubMed] [Google Scholar]
  • 4.Hochberg J., Brooks V., Pictorial recognition as an unlearned ability: A study of one child’s performance, The American Journal of Psychology. 75 (1962) 624–628. [PubMed] [Google Scholar]
  • 5.Jahoda G., Derȩgowski J., Ampene E., Williams N., Pictorial recognition as an unlearned ability, in: The Child’s Representation of the World, Springer, 1977: pp. 203–217. [Google Scholar]
  • 6.Yonas A., Arterberry M.E., Infants perceive spatial structure specified by line junctions, Perception. 23 (1994) 1427–1435. doi: 10.1068/p231427 [DOI] [PubMed] [Google Scholar]
  • 7.Kennedy J.M., Ross A.S., Outline picture perception by the Songe of Papua, Perception. 4 (1975) 391–406. [Google Scholar]
  • 8.Kennedy, John M, A Psychology of Picture Perception: Images and Information, Jossey-Bass Publishers, 1974.
  • 9.Itakura S., Differentiated responses to different human conditions by chimpanzees, Perceptual and Motor Skills. 79 (1994) 1288–1290. doi: 10.2466/pms.1994.79.3.1288 [DOI] [PubMed] [Google Scholar]
  • 10.Goodnow J., Children’s Drawing (The Developing Child), HarperCollins Publishers, 1980. [Google Scholar]
  • 11.Ishai A., Ungerleider L.G., Haxby J.V., Distributed neural systems for the generation of visual images, Neuron. 28 (2000) 979–990. doi: 10.1016/s0896-6273(00)00168-9 [DOI] [PubMed] [Google Scholar]
  • 12.Spiridon M., Kanwisher N., How distributed is visual category information in human occipito-temporal cortex? An fMRI study, Neuron. 35 (2002) 1157–1165. doi: 10.1016/s0896-6273(02)00877-2 [DOI] [PubMed] [Google Scholar]
  • 13.D.B. Walther, B. Chai, E. Caddigan, D.M. Beck, L. Fei-Fei, Simple line drawings suffice for functional MRI decoding of natural scene categories, Proceedings of the National Academy of Sciences. 108 (2011) 9661–9666. 10.1073/pnas.1015666108. [DOI] [PMC free article] [PubMed]
  • 14.O’Connell T.P., Sederberg P.B., Walther D.B., Representational differences between line drawings and photographs of natural scenes: A dissociation between multi-voxel pattern analysis and repetition suppression, Neuropsychologia. 117 (2018) 513–519. doi: 10.1016/j.neuropsychologia.2018.06.013 [DOI] [PubMed] [Google Scholar]
  • 15.Sayim B., Cavanagh P., What line drawings reveal about the visual brain, Frontiers in Human Neuroscience. 5 (2011) 118. doi: 10.3389/fnhum.2011.00118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.A. Bansal, A. Kowdle, D. Parikh, A. Gallagher, L. Zitnick, Which edges matter?, in: Proceedings of the IEEE International Conference on Computer Vision Workshops, 2013: pp. 578–585.
  • 17.F. Cole, A. Golovinskiy, A. Limpaecher, H.S. Barros, A. Finkelstein, T. Funkhouser, et al., Where do people draw lines?, in: ACM SIGGRAPH 2008 Papers, 2008: pp. 1–11.
  • 18.F. Cole, K. Sanik, D. DeCarlo, A. Finkelstein, T. Funkhouser, S. Rusinkiewicz, et al., How well do line drawings depict shape?, in: ACM SIGGRAPH 2009 Papers, 2009: pp. 1–9.
  • 19.Metzger W., Spillmann L.T., Lehar S.T., Stromeyer M.T., Wertheimer M.T., Laws of seeing., MIT Press, 2006. [Google Scholar]
  • 20.Torralbo A., Walther D.B., Chai B., Caddigan E., Fei-Fei L., Beck D.M., Good exemplars of natural scene categories elicit clearer patterns than bad exemplars but not greater BOLD activity, PLoS One. 8 (2013) e58594. doi: 10.1371/journal.pone.0058594 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Brainard D.H., The Psychophysics Toolbox, Spatial Vision. 10 (1997) 433–436. 10.1163/156856897x00357. [DOI] [PubMed] [Google Scholar]
  • 22.T. Huang, G. Yang, G. Tang, A fast two-dimensional median filtering algorithm, IEEE Transactions on Acoustics, Speech, and Signal Processing. 27 (1979) 13–18.
  • 23.Koenderink J.J., What does the occluding contour tell us about solid shape?, Perception. 13 (1984) 321–330. doi: 10.1068/p130321 [DOI] [PubMed] [Google Scholar]
  • 24.D. DeCarlo, A. Finkelstein, S. Rusinkiewicz, A. Santella, Suggestive contours for conveying shape, in: ACM SIGGRAPH 2003 Papers, 2003: pp. 848–855.
  • 25.Wang Z., Qiu S., Feng N., Rushmeier H., McMillan L., Dorsey J., Tracing Versus Freehand for Evaluating Computer-Generated Drawings, ACM Transactions on Graphics (TOG). 40 (2021) Article 52. [Google Scholar]
  • 26.Berger I., Shamir A., Mahler M., Carter E., Hodgins J., Style and abstraction in portrait sketching, ACM Transactions on Graphics (TOG). 32 (2013) 1–12. [Google Scholar]
  • 27.Biederman I., Recognition-by-components: A theory of human image understanding, Psychological Review. 94 (1987) 115–147. doi: 10.1037/0033-295X.94.2.115 [DOI] [PubMed] [Google Scholar]
  • 28.Choo H., Walther D.B., Contour junctions underlie neural representations of scene categories in high-level human visual cortex, NeuroImage. 135 (2016) 32–44. doi: 10.1016/j.neuroimage.2016.04.021 [DOI] [PubMed] [Google Scholar]
  • 29.Walther D.B., Shen D., Nonaccidental Properties Underlie Human Categorization of Complex Natural Scenes, Psychological Science. (2014). doi: 10.1177/0956797613512662 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Sanmiguel D., Art of Drawing: The Complete Course, 2nd edition, Sterling, New York, NY, 2003. [Google Scholar]
  • 31.Bainbridge W.A., Hall E.H., Baker C.I., Drawings of real-world scenes during free recall reveal detailed object and spatial information in memory, Nature Communications. 10 (2019) 1–13. doi: 10.1038/s41467-018-07882-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Perdreau F., Cavanagh P., Drawing skill is related to the efficiency of encoding object structure, I-Perception. 5 (2014) 101–119. doi: 10.1068/i0635 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Bainbridge W.A., Pounder Z., Eardley A.F., Baker C.I., Quantifying Aphantasia through drawing: Those without visual imagery show deficits in object but not spatial memory, Cortex. 135 (2021) 159–172. doi: 10.1016/j.cortex.2020.11.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Bhunia A.K., Das A., Riaz Muhammad U., Yang Y., Hospedales T.M., Xiang T., et al., Pixelor: A Competitive Sketching AI Agent. So you think you can beat me?, in: SIGGRAPH Asia, 2020. [Google Scholar]
  • 35.Das A., Yang Y., Hospedales T., Xiang T., BézierSketch: A generative model for scalable vector sketches, in: Springer, 2020: pp. 632–647. [Google Scholar]
  • 36.C. Gao, Q. Liu, Q. Xu, L. Wang, J. Liu, C. Zou, SketchyCOCO: image generation from freehand scene sketches, in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020: pp. 5174–5183.
  • 37.Hertzmann A., Why do line drawings work? A realism hypothesis, Perception. 49 (2020) 439–451. doi: 10.1177/0301006620908207 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Markus Lappe

6 Sep 2021

PONE-D-21-25287Where to draw the line?PLOS ONE

Dear Dirk,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please ensure that your decision is justified on PLOS ONE’s publication criteria and not, for example, on novelty or perceived impact.

Please submit your revised manuscript by Oct 21 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

I look forward to receiving your revised manuscript.

Best wishes,

Markus 

---

Markus Lappe

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

3. Please modify the title to ensure that it is meeting PLOS’ guidelines (https://journals.plos.org/plosone/s/submission-guidelines#loc-title). In particular, the title should be "specific, descriptive, concise, and comprehensible to readers outside the field" and in this case it is not informative and specific about your study's scope and methodology.

4. Thank you for stating the following in the Acknowledgments Section of your manuscript: 

"We thank Profs. Natalie Waldburger and Amy Swartz for insightful discussions and for help with recruiting participants from OCAD University. This work was supported by the Natural Sciences and Engineering Research Council of Canada (RGPIN-2020-04097) and the Social Sciences and Humanities Research Council of Canada (430-2017-01189)."

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. 

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: 

"This work was supported by grants by the Natural Sciences and Engineering Research Council of Canada (RGPIN-2020-04097) and the Social Sciences and Humanities Research Council of Canada (430-2017-01189) to DBW. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

5. We note that Figures 1 & 2 in your submission contain copyrighted images. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

We require you to either (1) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (2) remove the figures from your submission:

a. You may seek permission from the original copyright holder of Figures 1 & 2 to publish the content specifically under the CC BY 4.0 license. 

We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission. 

In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

b. If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Overall, this paper presents a welcome addition to the literature on line drawing, showing that the most-commonly drawn lines on real scenes correspond to occlusion boundaries, and connecting these lines to artist experience and drawing order. As I am not experienced in publishing this kind of psychological study, I cannot evaluate the details of the experimental setup according to the standards of this field, but, to my eye, it all appears sensible and worthwhile.

My only concern is a lack of citations to highly-relevant studies performed with human artists to address several of these same questions. The submission provides complementary information to these studies, since it operates on a dataset of photographs of real-world scenes, rather than computer-generated imagery of individual objects. There are various pros and cons to the different methodologies. An advantage of the submission is that these are real photographs; a disadvantage of the submission is that the line coding is ad hoc and omits some categories of lines (that may or may not be relevant to these scene categories).

The most immediately-relevant paper was published very recently. This paper analyzes drawing order and uses precise definitions of line styles, and uses photorealistic imagery:

Zeyu Wang, Sherry Qiu, Nicole Feng, Holly Rushmeier, Leonard McMillan, Julie Dorsey

Tracing Versus Freehand for Evaluating Computer-Generated Drawings

ACM Transactions on Graphics (SIGGRAPH), 2021

These older papers are also highly relevant, including analysis of which lines people draw and the order in different cases:

Forrester Cole, Aleksey Golovinskiy, Alex Limpaecher, Heather Stoddart Barros, Adam Finkelstein, Thomas Funkhouser, and Szymon Rusinkiewicz.

"Where Do People Draw Lines?"

ACM Transactions on Graphics 27(3), August 2008.

"How Well Do Line Drawings Depict Shape?," Forrester Cole, Kevin Sanik, Doug DeCarlo, Adam Finkelstein, Thomas Funkhouser, Szymon Rusinkiewicz, and Manish Singh, ACM Transactions on Graphics 28(3)

Berger, I., Shamir, A., Mahler, M., Carter, E., & Hodgins, J. (2013). Style and abstraction in portrait sketching. ACM Transactions on Graphics (TOG), 32(4), 1-12.

Reviewer #2: The manuscript, entitled "Where to draw the line?", by H. Sheng, J. Wilder, and D. B. Walther reports a study testing how people draw line-drawings representing 3D scenes. This is an important and fundamental question in Vision science and the authors addressed this question using a sophisticated method. The study is composed of 3 experiments and they are closely tied to one another. Their line-drawings were generated by people (artists and non-artists) under a controlled condition from photos of 3D scenes in Experiment 1. These drawings were evaluated by different groups of people in Experiments 2 and 3. The drawings were also analyzed by the authors and the results of their analysis were compared with the results of their experiments. The authors found that there were contours that were drawn consistently across the participants (artists and non-artists) in Experiment 1. These consistent contours, which were drawn earlier than the other contours, represented the scenes drawn in the drawings well. These consistent contours often represented the occluding boundaries of objects in the scenes.

The study is very interesting, as well as done very well, and I have only a few minor suggestions.

1) The authors categorize contours the drawings into 4 types (L. 385-388): texture/albedo edges, occlusion/depth boundaries, surface normal discontinuities, and shadow edges. There are, however, some other types of contours that represent 3D information in scenes: e.g. ridges and suggestive-contours. The 3D perception from these two types of contours was tested in Cole et al. (Cole, Sanik, DeCarlo, Finkelstein, Funkhouser, Rusinkiewicz & Singh, 2009, ACM SIGGRAPH).

2) The perception of contours representing shadow edges is discussed in Metzger (1936/2006, Figure 132).

3) L.229-242. This paragraph is unclear and should be revised for clarity.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Nov 8;16(11):e0258376. doi: 10.1371/journal.pone.0258376.r002

Author response to Decision Letter 0


22 Sep 2021

Toronto, Sept. 21st, 2021

Dear Prof. Lappe, dear Reviewers

Thank you for your overall positive assessment of our manuscript “Where to draw the line?” We appreciate the input from the two reviewers regarding additional references as well as a more in-depth discussion of contour types. We have revised our manuscript accordingly. Please find our point-by-point response below.

On behalf of all authors

Dirk B. Walther

Reviewer #1: Overall, this paper presents a welcome addition to the literature on line drawing, showing that the most-commonly drawn lines on real scenes correspond to occlusion boundaries, and connecting these lines to artist experience and drawing order. As I am not experienced in publishing this kind of psychological study, I cannot evaluate the details of the experimental setup according to the standards of this field, but, to my eye, it all appears sensible and worthwhile.

We thank the reviewer for the positive feedback and appreciation of our work.

My only concern is a lack of citations to highly-relevant studies performed with human artists to address several of these same questions. The submission provides complementary information to these studies, since it operates on a dataset of photographs of real-world scenes, rather than computer-generated imagery of individual objects. There are various pros and cons to the different methodologies. An advantage of the submission is that these are real photographs; a disadvantage of the submission is that the line coding is ad hoc and omits some categories of lines (that may or may not be relevant to these scene categories).

The most immediately-relevant paper was published very recently. This paper analyzes drawing order and uses precise definitions of line styles, and uses photorealistic imagery:

Zeyu Wang, Sherry Qiu, Nicole Feng, Holly Rushmeier, Leonard McMillan, Julie Dorsey,Tracing Versus Freehand for Evaluating Computer-Generated Drawings, ACM Transactions on Graphics (SIGGRAPH), 2021

These older papers are also highly relevant, including analysis of which lines people draw and the order in different cases:

Forrester Cole, Aleksey Golovinskiy, Alex Limpaecher, Heather Stoddart Barros, Adam Finkelstein, Thomas Funkhouser, and Szymon Rusinkiewicz.

"Where Do People Draw Lines?", ACM Transactions on Graphics 27(3), August 2008.

"How Well Do Line Drawings Depict Shape?," Forrester Cole, Kevin Sanik, Doug DeCarlo, Adam Finkelstein, Thomas Funkhouser, Szymon Rusinkiewicz, and Manish Singh, ACM Transactions on Graphics 28(3)

Berger, I., Shamir, A., Mahler, M., Carter, E., & Hodgins, J. (2013). Style and abstraction in portrait sketching. ACM Transactions on Graphics (TOG), 32(4), 1-12.

We thank the reviewer for making us aware of these publications, especially the recent SIGGRAPH paper. We had neglected to consider the field of computer graphics in our literature search. But as the reviewer points out, these papers are highly relevant! We now include them in the discussion section and specifically position our work with respect to them.

Recent work on human sketches of individual objects showed similar results for the most agreed-upon contours as well the temporal order of strokes (25). Interestingly, these experiments included a freehand sketch condition, which resulted in similar profiles of contour agreement as well as temporal order of contours as the tracing condition.

Human proficiency at perceiving objects in complex scenes has been previously linked to edges created from surface normals and depth boundaries in studies with synthetic scenes (16) as well as isolated objects (17,18,25) and portraits (26). Specifically, contours drawn by artists were found to be in good agreement with one another (17), contours were found to be placed at locations optimal for depicting 3D shape and aesthetic quality (18), and artists’ drawings were found to define the convex hull area of the depicted objects early on in the drawing process (25). Our analysis of the roles of contours in Experiment 3 confirms these findings in a real-world scene setting. These findings are consistent with the important role of contour junctions for the perception of objects and scenes, as contour junctions serve as a low-level cue to spatial relations such as occlusion (27–29). Moreover, we show how artists prioritize contours that lead to drawings which are more representative of the depicted scene (Equation 1) and that lead to better perception of scene gist (Figure 7A). This finding is likely related to the technique of “blocking-in” – a coarse, block-like outline of the proportions of figures and objects in the initial phase of drawing (30). This technique may lead trained artists to initially prioritize contours that convey global shape over contours that convey finer details (see also (25)).

Reviewer #2: The manuscript, entitled "Where to draw the line?", by H. Sheng, J. Wilder, and D. B. Walther reports a study testing how people draw line-drawings representing 3D scenes. This is an important and fundamental question in Vision science and the authors addressed this question using a sophisticated method. The study is composed of 3 experiments and they are closely tied to one another. Their line-drawings were generated by people (artists and non-artists) under a controlled condition from photos of 3D scenes in Experiment 1. These drawings were evaluated by different groups of people in Experiments 2 and 3. The drawings were also analyzed by the authors and the results of their analysis were compared with the results of their experiments. The authors found that there were contours that were drawn consistently across the participants (artists and non-artists) in Experiment 1. These consistent contours, which were drawn earlier than the other contours, represented the scenes drawn in the drawings well. These consistent contours often represented the occluding boundaries of objects in the scenes.

The study is very interesting, as well as done very well, and I have only a few minor suggestions.

We thank the reviewer for the overall positive assessment.

1) The authors categorize contours the drawings into 4 types (L. 385-388): texture/albedo edges, occlusion/depth boundaries, surface normal discontinuities, and shadow edges. There are, however, some other types of contours that represent 3D information in scenes: e.g. ridges and suggestive-contours. The 3D perception from these two types of contours was tested in Cole et al. (Cole, Sanik, DeCarlo, Finkelstein, Funkhouser, Rusinkiewicz & Singh, 2009, ACM SIGGRAPH).

We thank the reviewer for this comment. And thank you for pointing out that we neglected to reference the Cole et al. paper. Regarding the contour types mentioned in that paper, we classified "ridges and valleys" as "surface normals" in our manuscript. Suggestive contours were grouped into "other", since we could not determine a physical reason for the particular contours. We now make this correspondence explicitly clear in the Methods section for Experiment 3:

To analyze contour types, the first author manually labelled each contour in the super-reference drawings by sequentially overlaying them over the original photograph. She classified contours into four different edge types according to their physical cause (16–18): texture/albedo edges (change in reflectance across smooth surface), occlusion/depth boundaries (boundaries of objects) (23), surface normal discontinuities (intersecting surfaces, ridges and valleys), and shadow edges (boundary of cast shadows) (19). For cases where a contour included more than one type of origin, the type that corresponds to the longer portion was chosen. Any contours that could not be assigned clearly to one of these types were labeled as “other”. The “other” categories includes “suggestive contours,” which occur in locations where surfaces bend away from the observer but do not form a true depth discontinuity (24). Identifying suggestive contours with certainty occurred too rarely in the real-world complex scenes in this study to justify its own contour category. Once the contours were categorized, we totalled the number of pixels in contours belonging to each type within the most and least consistent half line drawings, and performed significance testing with a fixed-effects two-way ANOVA followed by Bonferroni-corrected

2) The perception of contours representing shadow edges is discussed in Metzger (1936/2006, Figure 132).

Thank you for pointing out this important reference. We have added it to the introduction and the methods section for Experiment 3.

3) L.229-242. This paragraph is unclear and should be revised for clarity.

We have split the paragraph into two parts and added additional explanations to improve clarity. Here is the revised text:

This computation potentially leads to conflicting assignments along the length of the query contour. That is, different segments of the same query contour could be assigned to different reference contours. We resolve such conflicts with a two-stage voting process. First, we implemented a majority vote with a sliding window of 25 segments across the query contour. That is, the reference contour identifier most frequently assigned to the individual line segments within the window along the query contour was assigned to the contour segment at the center of the window. This voting process is analogous to median filtering for noise reduction in image processing (22). Remember that contour segments are at most 10 pixels long, so the sliding window in this procedure is at most 250 pixels wide.

Second, another global majority vote was then applied to all line segments within the query contour to determine the best matching reference contour. This process usually led to a unique assignment of the reference contour and only rarely led to a tie. When ties occurred, they were broken by determining which of the contending reference contours was closest to the query contour over its entire length. Specifically, we computed the sum of the distances of the individual query line segments to the corresponding reference line segment, weighted by the length of the query line segments. The identifier of the reference contour with the smallest sum (the closest overall distance) was assigned to the query contour.

Attachment

Submitted filename: ResponseToReviewers.docx

Decision Letter 1

Markus Lappe

27 Sep 2021

Where to draw the line?

PONE-D-21-25287R1

Dear Dirk,

I am pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Markus Lappe

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Markus Lappe

27 Oct 2021

PONE-D-21-25287R1

Where to draw the line?

Dear Dr. Walther:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Markus Lappe

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: ResponseToReviewers.docx

    Data Availability Statement

    All data and data analysis scripts as well as all individual line drawings and the super-references are available on the Open Science Framework website: https://osf.io/x9uj5/.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES