1 / 33

Toward a Reliable Evaluation of Mixed-Initiative Systems

Toward a Reliable Evaluation of Mixed-Initiative Systems. Gabriella Cortellessa and Amedeo Cesta National Research Council of Italy Institute for Cognitive Science and Technology Rome, Italy. Outline. Motivations Aims of the study

adia
Download Presentation

Toward a Reliable Evaluation of Mixed-Initiative Systems

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Toward a Reliable Evaluation of Mixed-Initiative Systems Gabriella Cortellessa and Amedeo Cesta National Research Council of Italy Institute for Cognitive Science and Technology Rome, Italy

  2. Outline • Motivations • Aims of the study • Users’ attitude towards the mixed-initiative paradigm • Role of explanation during problem solving • Evaluation Method • Results • Conclusions and future work

  3. Motivations • Lack of studies that investigate users attitude towards this solving paradigm • Lack of methodologies for evaluating different aspects of mixed-initiative problem solving This work applies an experimental approach (from HCI and Psychology) to the problem of understanding users’ attitude towards the mixed-initiative approach and investigating the importance of explanation as a means to foster users’ involvement in the problem solving

  4. Artificial problem solver Interaction Module User User Two alternative Problem Solving approaches Automated approach Mixed-Initiative approach

  5. considering users’ requirements and judgment Evaluating Mixed-Initiative Systems • Measuring the overall problem solving performance • The pair human-artificial system is supposed to exhibit better performances (metrics). • Evaluating aspects related to users’ requirements and judgment on the system. • Usability, level of trust, clarity of presentation, user satisfaction etc.

  6. Aims of the study • Users’ attitude towards the solving strategy selection. • Automated vs mixed-initiative • The recourse to explanation during problem solving • Explanations for solvers’ choices and failures Differences between experts and non experts

  7. Solving strategy selection No empirical studies in the mixed-initiative area explore the context of strategy selection (who and why choose a solving strategy) However: Decision Support Systems • Empirical evidence of low trust toward automated advices during decision making processes (Jones & Brown, 2002). Human-Computer Interaction • Artificial solver as a competitor rather than a collaborator (Langer, 1992; Nass & Moon, 2000).

  8. Solving strategy selection: Hypotheses Two variables are supposed to influence the selection of the solving strategy (automated vs. mixed-initiative): user’s expertise, and problem difficulty Hypothesis 1: It is expected that expert users exploit the automated procedure more than non-experts; and, conversely, non-expert users exploit the mixed-initiative approach more than experts. Hypothesis 1a: It is expected that inexperienced users prefer the mixed-initiative approach when solving easy problems, and the automated strategy when solving difficult problems, while expert users are expected to show the opposite behavior.

  9. Explanation Recourse No empirical studies in the mixed-initiative research field investigate the role of explanations in cooperative problem solving However: Knowledge-Based Systems • explanation recourse is more frequent in case of systems failures (Gilbert, 1989; Schank, 1986; Chandrasekaran & Mittal, 1999). • explanation recourse is more frequent in case of collaborative problem solving (Gregor, 2001) • individual differences in the motivations for explanations recourse (Mao & Benbasat, 1996; Ye, 1995).

  10. Explanation Recourse: Hypotheses The following variables are supposed to influence the recourse to explanation: user’s expertise, problem difficulty,strategy selection,failure. Hypothesis 2: The access to explanation is more frequent in case of failure than in case of success. Hypothesis 3 : Access to explanation is related to the solving strategy selection. • In particular participants who choose the automated solving strategy access more frequently to explanation than those who use the mixed-initiative approach.

  11. Explanation Recourse: Hypotheses Hypothesis 4: During problem solving non experts access explanations more frequently than experts. Hypothesis 5: Access to explanation is more frequent in case of difficult problems.

  12. Evaluation Method • Participants: • 96 participants balanced with respect to gender, education, age and profession, subdivided in two groups based on the level of expertise (40 experts and 56 non experts). • Experimental apparatus: • COMIREM problem solver • Planning and scheduling problems • Procedure: • Web-based apparatus • Stimuli: Problems solution • Questionnaires

  13. A mixed-initiative problem solver: COMIREM COMIREM: Continuous Mixed-Initiative Resource Management Developed at Carnegie Mellon University Automated Solver Interaction Module User (Smith et al, 2003)

  14. DataBase http://pst2.istc.cnr.it/experiment Procedure • Training session • Two experimental sessions presented randomly : • Session 1: easy problems • Questionnaire 1 • Session 2: difficult problems • Questionnaire 2 • For each session participants were asked to choose between mixed and automated strategy Web-based

  15. Tasks Stimuli • 4 scheduling problems defined in the field of a broadcast TV station resources management: • 2 solvable • 2 unsolvable Questionnaires aiming to • Assessing the difficulty of the task: 5-steps Likert scale (Manipulation check of variable difficulty) • Evaluating the clarity of textual and graphical representations: (5-steps Likert scale) • Investigating the reasons for choosing the selected strategy (multiple choice) • Studying the reasons for accessing the explanation (only 2nd questionnaire)

  16. Solving Strategy Selection Results

  17. Dependent Variables Std. N Mean Deviation .6786 n_auto Non Expert 56 .6786 .7653 Expert 40 1.3750 .7048 1.3750 Total 96 .9688 .8137 n_mista Non Expert 1.3214 56 1.3214 .7653 Expert 40 .6250 .7048 .6250 Total 96 1.0313 .8137 expertise F(1,94) = 20.62, p < .001 Influence of expertise on strategy Choice_auto Choice_mixed Influence of expertise on solving strategy selection (statistics)

  18. Influence of expertise on strategy Hypothesis 1: Solving strategy selection (automated vs mixed-initiative) depends upon users’ expertise VERIFIED: p < .001 Experts automated Non experts  mixed-initiative

  19. Influence of difficulty on strategy strategy Easy Problems expertise Automated Mixed 24 32 Chi-square = 9.80, df=1, p< .01 Non expert 30 10 Expert 42 54 Total strategy expertise Difficult Problems Automated Mixed 24 32 Chi-square = 3.6 , df=1, n. s. Non expert 25 15 Expert 47 49 Total

  20. Influence of difficulty on strategy • Hypothesis 1a: Solving strategy selection (automated vs mixed-initiative) is related to problem difficulty PARTIALLY VERIFIED: Easy problems  experts: automated, non experts: mixed(p< .01) Difficult problems(n. s.)

  21. Reasons for strategy selection Automated -- Easy Automated -- Difficult Chi-square = .92 , df=2, n. s. Chi-square = 3.9 , df=2, p< .05 Mixed -- Easy Mixed -- Difficult Chi-square = 1.32 , df=2, n. s. Chi-square = 1.15 , df=2, n. s.

  22. Explanation Recourse Results

  23. Dependent Variables Std. Mean Deviation N .8111 Access_failure .8111 .3716 90 .3702 Access_correct .3702 .3354 90 F(1,89) = 85.37, p< .001 Correlation Analysis r= .035, n.s. r = .86 p < .001. in case of failure in case of success Influence of failures on explanation

  24. Influence of failures on explanation Hypothesis 2: The access to explanation is more frequent in case of failure than in case of success. VERIFIED p< .001

  25. Dependent Variables I_AC_DIF Indice di accesso alla spiegazione in caso I_AC_FAC Indice di accesso alla spiegazione in caso di compiti DIFFICILI di compiti FACILI Access easy Access difficult Std. Std. N Mean Deviation N Mean Deviation .6297 .8769 Automated 49 .6297 .2959 Automated 54 .8769 .3373 Mixed .2790 47 .2790 .2709 Mixed 42 .2802 .3202 .2802 Total 96 .4580 .3329 Total 96 .6158 .4430 F(1,94) = 77.26, p< .001 F(1,94) = 36.60, p< .05 Influence of strategy on explanation Easy problems Difficult problems

  26. Influence of strategy on explanation Hypothesis 3: Access to explanation is related to the solving strategy selection. Access to explanation is more frequent in case of automated strategy choice VERIFIED Easy problems p< .001 Difficult problems p< .05

  27. Dependent Variables Std. Expertise Mean Deviation N .5423 .7187 Access_easy Non Expert .5423 .4760 56 Expert .7187 .3740 40 .6158 Total .6158 .4430 96 Access_difficult Non Expert .3829 .5632 .3829 .3177 56 Expert .5632 .3289 40 .4580 Total .4580 .3329 96 expertise F(1,94) = 7.34, p< .01 difficulty F(1,94) = 12.54, p< .01 interaction F(1,94) = .002, n. s. Influence of expertise and difficulty on explanation

  28. difficulty p< .01 Influence of expertise and difficulty on explanation • Hypotheses 4 e 5: • During problem solving non experts rely on explanation more frequently than experts • Access to explanation is more frequent in case of difficult problems. FALSIFIED expertise p< .01

  29. Reasons for accessing explanation Non Experts Experts Understand the problem Understand automated solvers choices Chi-square = 2,28 , df=1, n. s.

  30. Conclusions • Solving strategy selection depends upon users’ expertise • Experts  automated • Non experts  mixed-initiative • The mixed initiative approach is chosen to maintain the control over the problem solving • Explanation during problem solving is frequently accessed (73 out of 96 respondents), the access being more frequent in case of: • Failures during problem solving • When using the automated strategy • Explanation is accessed to understand solvers choices

  31. Contributions • Empirical proof that the mixed-initiative approach responds to a specific need of end users to keep the control over automated systems. • The study confirms the need for developing problem solving systems in which humans play an active role • Need for designing different interaction styles to support the existing individual differences (e.g., expert vs non experts) • Empirical proof of the usefulness of explanation during problem solving. Failures have been identified as a main prompt to increase the frequency of access to explanation

  32. Remarks • Need for designing evaluation studies which takes into consideration the human component of the mixed-initiative system (importing methodologies from other fields) • At present we have inherited the experience from disciplines like HCI and Psychology and adapted them to our specific case. • The same approach can be followed to broaden the testing of different mixed-initiative features.

  33. Future work • Investigating the impact of strategy (automated vs mixed-initiative) and explanation recourse on problem solving performance. • Application of the evaluation methodology to measure different features of the mixed-initiative systems. • Synthesis of “user-oriented” explanations

More Related