4serendipity logo
   Home  | Résumé  |  4serendipity Web Search Utility  |  Some of my pages  |  Links

   Résumé Home   |  Projects   |  Functional Résumé   |  Chronological Résumé  

        arrowUsability Evaluation of WebBrain.com

Usability Test Report:  WebBrain Online Search Directory

Andrew Stevens and Caren Spencer-Smith

August 10, 2000

 

Table of Contents:

I.     Overview

        A.    Introduction

        B.    Background

        C.    Test Objectives

II.    Procedure

        A.    Methods

III.  Results

        A.    Summary

        B.    Efficiency

        C.    Effectiveness

        D.    Satisfaction

IV.  Discussion

        A.    Interface Findings

        B.    Search Directory Findings

        C.    Recommendations

V.    Appendix

        Table 1:  Efficiency Results

        Figure 1:  Efficiency Results

        Figure 2:  Efficiency Mean Times

        Figure 3:  Efficiency Standard Deviation

        Table 2:  Effectiveness Results

        Figure 4:  Effectiveness Results

        Table 3:  Successful Task Completion

        Table 4:  Post-evaluation Survey

        Table 5:  Post-evaluation Comments

        Data Records


I.     Overview

A.    Introduction

        WebBrain is a dynamic graphical representation of the Open Directory Project (ODP) web-based search directory. The interface is a horizontally split screen where the upper pane displays the search directory that dynamically centers around a user's selection of a directory topic; related information is displayed in a thesaurus-like structure. When a user selects a topic containing website links (as opposed to exclusively subdirectories), those links are displayed in the lower part of the display. A standard search engine box, capable of driving both the visual display and list of resulting page links, separates the two halves.

B.    Background

        The problems of the system relate to the navigation and effort associated with the WebBrain interface, as well as the ability of the ODP directory structure to adequately classify information. Thus, a test of this interface both reflects a user's ability to manipulate and navigate the interface, as well as a user's ability to correctly interpret a directory's potential content. We speculate that WebBrain’s unconventional interface and the use of a directory whose hierarchy is often quite ambiguous could produce significant usability problems and inefficiencies. More specifically, we believe that some of this site’s characteristics (dual pane interface, its visual hierarchy representation, and dependence on scrolling) are inherently flawed.

C.    Test Objectives

  1. How easy are very simple tasks (finding a website link only one level deep in the directory hierarchy) to complete using the WebBrain interface?
  2. Can users complete ordinary, common tasks (finding a site whose topic is 3 to 4 levels deep in ODP’s hierarchy) within the WebBrain interface?
  3. Do users easily grasp WebBrain’s graphical representation (the location and meaning of parent, child, sibling, and jump categories) of ODP’s hierarchy, as indicated by their use or verbal expressions?
  4. Does the dual pane interface of Webrain lead to users focusing excessively on the upper frame, as indicated by their ability to find the correct directory link and not the actual website link in the lower frame?
  5. Are users subjectively dissatisfied with WebBrain’s unconventional interface?
  6. What obstacles, if any, impede task completion?

II.    Procedure

A.    Methods

After completion of a pre-evaluation survey, each of five subjects were asked to enter the testing room of the SLIS Usability Lab, listen to the introduction script, and asked to sequentially complete six tasks. Subjects were asked to limit their search interaction to the graphical portion of the directory during this evaluation, but told they could access other features, including online help, at any time. Results were evaluated on the basis of time taken to complete each task, the degree to which the task was completed, and the total number of "clicks" made by each test subject for each task. Participants were asked to use the "think aloud" protocol during testing. Qualitative expressions were recorded, but not quantified. A post-evaluation survey was used to measure satisfaction and solicit written comments.

B.    Test Environment

        User testing was conducted in the SLIS Usability Lab, using the PC computer and the MS Internet Explorer web browser. The lab design follows the classic usability testing lab design with separate testing and observation rooms. A wall-mounted camera was used to record the user's facial expressions, while a microphone was used to record comments they made during the testing procedure. Synchronized data capture of the screen (using a scan converter) was used to record mouse movement and target selection.

        Tasks were presented on 3x5 cards, and progressed from “easy” to “difficult.” Test subjects were asked to complete a post-evaluation survey immediately following the last task

III.  Results

A.    Summary

        Results were based on the ability of participants to complete all or part of the tasks, the time taken to complete each task as related to previously established benchmarks, satisfaction measures obtained using a post-evaluation survey, and qualitative expressions recorded during testing and as part of the post-evaluation survey.

        Tasks given to each subject were to find links to the following websites:

1.      CNN (Cable News Network (1 level down, scrolling in website window)

2.      A History of Traditional Games (1 level down, no scrolling required)

3.      Ultramagic Balloons (3 levels down, no scrolling reqired)

4.      Quotation Ring (3 levels, no scrolling required)

5.      Nutrition Remedies for Allergies (5 levels, scrolling in website window required)

6.      Inner City Handball Association (2 levels, lateral scrolling through category display window required)

        The results for task 4 indicate a possible problem with task wording rather than user interaction with the interface. User 2 was successful in selecting the appropriate third-level link and completing the task; all other users stopped at an earlier page and selected a website link that was very similar to the target. Although this task was included in result summaries, its validity is questionable. Where applicable, data including and excluding this measure were given.

        Test subjects were able to select the first category link correctly from the home page 86% of the time, indicating a good organization of the directory’s top level. Subjects, however, experienced difficulty in making selections from subsequent levels, expressed by either wrong choice of category topic or extended time spent on a page reading through category links.

        Test subjects both demonstrated and expressed difficulty and confusion interpreting the data presented through the interface. Specific problems included the ability to shift focus from one frame to another, subject recognition the significance of the category positioning in the category display, and recognition of scroll bars in the category display window.

B.    Efficiency

        Efficiency was based on the time taken to complete each task. Benchmarks for the easy tasks (1 and 2) were:  excellent =   15 sec, OK = 16-60 sec, poor = 61-179 sec, and bad =  180 sec. Benchmarks for the difficult tasks (3 through 6) were:  excellent =  60 sec, OK = 61-179 sec, poor = 180-539 sec, and bad =  540 sec. (see Table 1, and Figures 1 through 3 in Appendix).

        Learning was not directly demonstrated by a consistent efficiency improvement for any test subject. Overall, 33% of the recorded times were excellent, 27% of the times were OK, 27% of the times were poor, and 13% of the times were bad. Again, although efficiency for task 4 was excellent or OK for all users, the validity is questionable due to task wording.

C.    Effectiveness

        Effectiveness was based on the number of clicks made by a test subject to complete a task, as well as the ability of a test subject to complete all or part of the task.

        Benchmarks for effectiveness were rated from excellent (the fewest number of clicks required for task completion) to poor (just over twice that number) (see Table 2, and Figures 4 and 5 in Appendix). Overall, 43% of the effectiveness scores were excellent, 17% were OK, and 40% were poor. Excluding data from Task 4, where 80% of the test subjects achieved excellent scores but failed to complete the task, the results are:  36% excellent, 20% OK, and 44% poor.

        Task completion was also documented (see Table 3 in Appendix). Overall, 60% of the tasks were completed and 40% were not. Of those tasks not completed, the following results were noted:

  • Users 4 and 5 both selected the correct category link from the home page (and each selected the correct page several times) but failed to find the target resource in the website window. Completion of the task required scrolling down in the website window.

  • User 3 completed 33% of task 3.

  • Users 1, 3, 4, and 5 completed 67% of task 4; task wording was a problem.

  • Users 1 and 3 completed 20% of task 5, users 2 and 4 completed 60%, and user 5 completed 80%. All users correctly selected the top-level category and then incorrectly selected the second level category (although users 2, 4, and 5 were able to recover from this to varying degrees).

    D.    Satisfaction

            Satisfaction was measured using a 20-question survey based on WAMMI, a previously validated instrument (see Table 4 in the Appendix). The 5-point scale of each question ranged from Strongly Agree to Strongly Disagree. Although test subjects were neutral in over half of their responses to the post-evaluation survey, several points are evident from the remaining responses:

    Users strongly agreed (1) that:

    • Remembering where I am on the WebBrain web site is difficult. (mean=1.2, S.D.=0.45)
    • It is difficult to tell if the WebBrain web site has what I want. (1.6, 0.55)
    • The WebBrain web site needs more introductory explanations. (1.8, 1.30)
    • The WebBrain web site has some annoying features. (2.2, 1.30)

    User's strongly disagreed (5) that:

    • Using the WebBrain web site for the first time is easy. (mean=4, S.D.=1.00)
    • I feel in control when I’m using the WebBrain web site. (3.8, 1.10)
    • The WebBrain web site is too slow. (3.8, 0.84)
    • Everything on the WebBrain web site is easy to understand. (3.8, 0.84)

            As part of the post-evaluation survey, each user was asked for the most negative and positive aspects of the WebBrain search directory (see Table 5 in the Appendix).

    IV.  Discussion

    A.    Interface Findings

    1.         Test subjects did not recognize the significance of the category positioning in the category display.

  • Supported by comments made by the users during testing and observed use of browser navigation for navigation within the directory structure. Some users demonstrated and commented on their understanding of the Parent-Sibling relationship, and used it for navigation. Comments related to other category positions, when made, were incorrect.

  •       "Oh, here it is...but that is totally by serendipity."

  •       "Oh, so these are the top links and these are bottom links. I don't know, I think still it's confusing. The thing that I really liked in the beginning is now confusing."

  •       "OK. I'm getting confused. This thing is moving me around in a way that is kind of disorienting. It's moving me from concepts, I realize."

  •       "This is a nightmare."

  •       "Actually, I find the navigation kind of difficult on here, as far as the sequential type of navigation."

    2.         Subjects often failed to notice target resource’s existence in the web site window.

  • Supported by user comments and observed use of the interface. User 4 failed to notice the links in the website window, as stated in a comment, until 2 min 18 sec into the task.  

  •       "I wonder where you actually get to a link to something.... Oh, oh wait; it's pulling them up down here."

  •       "I think it's confusing because I'm looking up, and then I realize something is down there."

  •       "See, this is why I'm kind of confused because when I clicked on games, on the last task, it pulled up things down here, and this doesn't have things down there, so I really wasn't paying much attention."

    3.         Subjects did not access any assistance resources and were dependent upon browser functions for navigation.

  • Supported by observation.

    4.         Subjects sometimes failed to notice scrolls available for the different category positions within the category link window

  • Supported by user comments and observations. Most users verbally identified the category scrolls during the third task or later. User 4 did not notice category scroll bars until last task.

  •       “Boy, I hope I didn’t need these before now. I probably did.”

  •       “Red navigation arrows are too subtle.”

    5.         Subjects were delayed by number of topics in category display window

  • Supported by user comments and observations, as well as efficiency ratings.

  •       "I can't believe it took me that long."

  •       "There are too many options, I was 'overwhelmed' with links, categories and silly animated graphics. Give me Yahoo instead any day."

    B.    Search Directory Findings

    1.      The partial/incomplete display of categories' hierarchical position in the category display window resulted in users getting “lost” in the directory structure.

  • Supported by verbal comments, effectiveness ratings, and suggestions included with post-evaluation survey

  •       "Too many options one-level down."

  •       "Show a navigation path-I got lost"

  •       "Actually, you know, the search engine looks pretty good, I mean, as far as the specificity, if that's the word, with which it's pulling things up; I can't find...I don't know where I'm going."

    2.      The ODP directory structure is sometimes ambiguous and redundant.

  • Supported by verbal and post-evaluation comments, as well as effectiveness ratings. Users were often confused to find the same category under different directory listings (News as a main heading and News under Television).

  •       "Too many options one-level down."

  •       "As with all pre-coordinated systems, it is difficult to determine where fringe topics will be located (handball, balloons). It seems as if these should be connected via sports & entertainment-but if this was the case, why couldn't I find the allergy site? It should have been cross-listed as well."

  •       "I thought it would be here.... It seems like it would be there."

    C.    Recommendations

    1.      Use simple text labels to highlight the significance of each category type within the category display window and category scrolls (addresses Interface Findings 1 and 4).

    2.      Purpose of each window should be highlighted in some way, preferably on the home page (addresses Interface Findings 1 and 2)

    3.      Dynamic resizing of website window to assist focusing user attention on either the category link or website window at any given time (addresses Interface Finding 2).

    4.      Link text labels to a “Help” section; provide a specifically labeled start screen or blatantly labeled “Help” link (addresses Interface Finding 3, but also others).

    5.      Restructure directory to reduce number of category links in any given category position, or consider using a directory structure better suited for the interface (addresses Search Directory Findings 1 and 2).

    V.    Appendices

    Table 1:  Efficiency Results

    Time Taken

    Benchmarks

    User #1

    User #2

    User #3

    User #4

    User #5

    Task 1

    Excellent:  15 sec

    OK:   16-60 sec

    Poor:  61-179 sec

    Bad:   180 sec

    49 sec

    27 sec

    61 sec

    175 sec

    450 sec

    Task 2

    Excellent:   15 sec

    OK:  16-60 sec

    Poor: 61-179 sec

    Bad:   180 sec

    19 sec

    17 sec

    9 sec

    420 sec

    105 sec

    Task 3

    Excellent:   60 sec

    OK:  61-179 sec

    Poor:  180-539 sec

    Bad:   540 sec

    192 sec

    59 sec

    467 sec

    58 sec

    30 sec

    Task 4

    Excellent:   60 sec

    OK:  61-179 sec

    Poor:  180-539 sec

    Bad:   540 sec

    76 sec

    53 sec

    58 sec

    176 sec

    30 sec

    Task 5

    Excellent:   60 sec

    OK:  61-179 sec

    Poor: 180-539 sec

    Bad:   540 sec

    336 sec

    562 sec

    396 sec

    562 sec

    446 sec

    Task 6

    Excellent:   60 sec

    OK:  61-179 sec

    Poor:  180-539 sec

    Bad:   540 sec

    21 sec

    22 sec

    118 sec

    94 sec

    28 sec


    Figure 1:  Efficiency Results

    Figure 2:  Efficiency Mean Times

    Mean Time to Task Completion


     

    Figure 3:  Efficiency Standard Deviation


     

    Table 2:  Effectiveness Results

    Accuracy

    Benchmarks

    User #1

    User #2

    User #3

    User #4

    User #5

    Task 1

    Excellent:  1

    OK:  2-3

    Poor:   4

    1

    1

    3

    12

    40

    Task 2

    Excellent:  1

    OK:  2-3

    Poor:   4

    2

    1

    1

    25

    3

    Task 3

    Excellent:  3

    OK:   4-6

    Poor:   7

    7

    3

    27

    3

    3

    Task 4

    Excellent:  3

    OK:  4-6

    Poor:   7

    2

    3

    2

    8

    2

    Task 5

    Excellent:  5

    OK:  6-11

    Poor:   12

    11

    32

    30

    40

    40

    Task 6

    Excellent:  2

    OK:  3-4

    Poor:   5

    2

    2

    5

    5

    4

    Figure 4:  Effectiveness Results


    Figure 5:  Effectiveness Mean Clicks


     

    Table 3:  Successful Task Completion

    Task #

    User 1

    User 2

    User 3

    User 4

    User 5

    1

    Complete

    Complete

    Complete

    Fail

    Fail

    2

    Complete

    Complete

    Complete

    Complete

    Complete

    3

    Complete

    Complete

    Fail

    Complete

    Complete

    4

    Fail

    Complete

    Fail

    Fail

    Fail

    5

    Fail

    Fail

    Fail

    Fail

    Fail

    6

    Complete

    Complete

    Complete

    Complete

    Complete

    Table 4:  Post-Evaluation Survey

    (1=Strongly Agree, 5=Strongly Disagree)

    Average

    Std. Dev.

    The WebBrain web site has much that of interest to me.  

    2.6

    0.89

    It is difficult to move around the WebBrain web site. 

    2.8

    0.45

    I can quickly find what I want on the WebBrain web site.

    3

    0.71

    The WebBrain web site seems logical to me.  

    2.8

    0.84

    The WebBrain web site needs more introductory explanations.  

    1.8

    1.30

    The pages on the WebBrain web site are very attractive.  

    2.8

    1.30

    I feel in control when I’m using the WebBrain web site.  

    3.8

    1.10

    The WebBrain web site is too slow.  

    3.8

    0.84

    The WebBrain site helps me find what I am looking for.  

    3

    0.00

    Learning to find my way around the WebBrain web site is a problem.

    2.8

    0.84

    I don’t like using the WebBrain web site.  

    3.4

    1.82

    I can easily contact the people I want to on the WebBrain web site.  

    3.4

    0.89

    I feel efficient when I'm using the WebBrain web site.  

    3.6

    0.55

    It is difficult to tell if the WebBrain web site has what I want.  

    1.6

    0.55

    Using the WebBrain web site for the first time is easy.  

    4

    1.00

    The WebBrain web site has some annoying features.  

    2.2

    1.30

    Remembering where I am on the WebBrain web site is difficult.  

    1.2

    0.45

    Using the WebBrain web site is a waste of time.  

    3.6

    0.89

    I get what I expect when I click on things on the WebBrain web site.

    3.2

    0.45

    Everything on the WebBrain web site is easy to understand.  

    3.8

    0.84

     

    Table 5:  Post-evaluation Comments

    Negative aspects of system:

    • "-too much java script - diff browser will not read"
    • "-links are hard to tell"
    • "-too much animation"
    • "Too many options one-level down"
    • "Ridiculous interface/not needed"
    • "The bottom view screen (scroll) is too small"
    • "There are too many options, I was 'overwhelmed' with links, categories and silly animated graphics. Give me Yahoo instead any day."
    • "It is difficult to tell where I am in the structure after moving from the first level"
    • "Red navigational arrows are too subtle"
    • "All the reorganization in the upper blue part of the screen freaks one out. It's startling."
    • "As with all pre-coordinated systems, it is difficult to determine where fringe topics will be located (handball, balloons). It seems as if these should be connected via sports & entertainment-but if this was the case, why couldn't I find the allergy site? It should have been cross-listed as well."
    • "Difficult first-time user experience"
    • "Navigation can be disorienting"
    • "No alternative, traditional hierarchical scheme to be browsed."
    • "Show a navigation path-I got lost"
    • "Indicate how information is related hierarchically."
    • "Provide a tutorial for new users"
    • "The scrolling"
    • "(Not) Knowing where you have been"
    • "Unpredictable categories"

    Positive aspects of system:

    • "-cool design/idea"
    • "-has lots of info"
    • "-good categorization"
    • "none"
    • "Site is uncluttered, and makes one think more logically about which link to choose"
    • "The lines connecting concepts are interesting & may be helpful when I am more familiar with the site"
    • "Fun to use"
    • "Browsable in a more flexible way than other engines"
    • "The learning curve isn't too bad. After a few clicks you get the hang."
    • "Navigation may be a little less efficient and somewhat confusing, but the site affords a style of browsing not supported by other search engines."
    • "The way it flips :)"
    • "The feel"
    • "The colors (Go U of M!!)
    • "With some improvement with the categories, this could be useful"