260 likes | 399 Views
Finding Solutions in Goal Models: An Interactive Backward Reasoning Approach. Jennifer Horkoff 1 Eric Yu 2 Department of Computer Science 1 Faculty of Information 2 jenhork@cs.utoronto.ca yu@ischool.utoronto.ca University of Toronto November 1, 2010 ER’10. Outline. Early RE
E N D
Finding Solutions in Goal Models: An Interactive Backward Reasoning Approach Jennifer Horkoff1 Eric Yu2 Department of Computer Science1 Faculty of Information2 jenhork@cs.utoronto.cayu@ischool.utoronto.ca University of Toronto November 1, 2010 ER’10
Outline • Early RE • A Framework for Iterative, Interactive Analysis of Agent-Goal Models in Early RE • Background: Goal Models • Background: Interactive Forward Satisfaction Analysis • Need for Backward Analysis • Interactive, Iterative Backward Analysis • Example • Implementation • Qualitative Studies • Conclusions and Future Work Finding Solutions in Goal Models - Horkoff, Yu
Early Requirements Engineering • Domain understanding in the early stages of a project is crucial for the success of a system • Characterized by incomplete and imprecise information • Many non-functional, hard to quantify success criteria: customer happiness, privacy, job satisfaction, security,… • (Ideally) a high degree of stakeholder participation • Goal models can be a useful conceptual tool for Early RE • Represent imprecise concepts (softgoals) and relationships (contribution links) • Relatively simple syntax • Amenable for high-level analysis and decision making Finding Solutions in Goal Models - Horkoff, Yu
Example: inflo Case Study Finding Solutions in Goal Models - Horkoff, Yu
Framework for Iterative, Interactive Analysis of Agent-Goal Models in Early RE • We develop a framework to support iteration and stakeholder participation in early goal model analysis • Survey and analysis of existing analysis procedures (SAC’11) • Interactive forward evaluation (CAiSE’09 Forum, PoEM’09, IJISMD) • Interactive backward evaluation (this work!) • Visualizations for backward evaluation (REV’10) • User testing (PoEM’10) • Suggested methodology (CAiSE’09, PoEM’09, IJISMD,…) • More… Finding Solutions in Goal Models - Horkoff, Yu
Background: Goal Models We use i* as an example goal modeling framework Finding Solutions in Goal Models - Horkoff, Yu
Interactive Forward Satisfaction Analysis • A question/scenario/alternative is placed on the model and its affects are propagated “forward” through model links • Dependency links propagate values as is • AND/OR (Decomposition/Means-Ends) links take the min/max • Contribution links propagated as follows: • Interactive: user input (human judgment) is used to decide on partial or conflicting evidence “What is the resulting value?” Finding Solutions in Goal Models - Horkoff, Yu
Example: Interactive Forward Satisfaction Analysis Human Judgment Human Judgment What if… the application asked for secret question and did not restrict the structure of password? Finding Solutions in Goal Models - Horkoff, Yu
Need for Backward Analysis Is it possible for Attract Users to be Satisfied? Partially Satisfied? If so how? If not, why not? We can ask “what if…?” questions, but what about “Is this possible?” “How?” “If not, why?” Finding Solutions in Goal Models - Horkoff, Yu
Background: SAT and UNSAT Core • SAT Solver: Algorithms which solve Boolean Satisfiability Problem • Accept a Boolean formula in conjunctive normal form (CNF), composed of a conjunction of clauses. • Searches for an assignment of the formula’s clauses in which it makes the formula true. • Although the SAT problem is NP-Complete, algorithms and tools which can solve many SAT problems in a reasonable amount of time have been developed. • When a SAT problem does not have a solution, we can find the UNSAT core • UNSAT core: an unsatisfiable subset of clauses in a CNF • We use the zChaff and zMinimal implementations in our tool Finding Solutions in Goal Models - Horkoff, Yu
Interactive, Iterative Backward Analysis • Procedure overview: • i* model and target values are encoded into CNF • Iteratively call a SAT solver on the CNF representation: • After each iteration, prompt for human judgment for intentions with conflicting or multiple sources of partial evidence • Re-encode CNF formula • If SAT solver finds an answer and no human judgment is needed: • Success: answer provided • If SAT solver cannot find an answer • Display UNSAT core • Backtrack over last round of human judgment • If no more human judgment to backtrack over: • Failure: no answer Finding Solutions in Goal Models - Horkoff, Yu
Formally Expressing i* • i* model: a tupleM = <I, R, A> • I is a set of intentions • R is a set of relations between intentions • A is a set of actors • Intention type: each intention maps to one type in IntentionType = {Softgoal, Goal, Task, Resource} • Relation type: Each relation maps to one type in RelationType = {Rme, Rdec, Rdep, Rc} • Rc can be broken down into {Rm, Rhlp, Ru, Rhrt, Rb} • Relation behavior: Rdep and Rc are binary (one intention relates to one intention), Rme and Rdec are (n+1)-ary (one to many intentions relate to one intention) Finding Solutions in Goal Models - Horkoff, Yu
Analysis Predicates • Formal expression of analysis predicates • Similar to predicates used in Tropos Framework (Sebastiani et al. CAiSE’04) • For i∈ I: • Example: PS(Attract Users), D(Ask for Secret Question) • Conflict label: C(i), • Conflict situation: for i∈I a predicate from more than one of the following sets hold: • {S(i), PS(i)}, {U(i)}, {C(i)}, {PD(i), D(i)} Finding Solutions in Goal Models - Horkoff, Yu
Propagation in CNF • We use the SAT formula from the Tropos Framework: Φ = ΦTarget∧ (the target values representing the question) ΦForward ∧ (axioms describing forward propagation) ΦBackward∧ (axioms describing backward propagation) ΦInvariants ∧ (axioms describing invariant properties) ΦConstraints (constraints on propagation) ΦInvariants: S(i) → PS(i) D(i) → PD(i) ΦConstraints: ∀i∈I s.t. i is a leaf: PS(i) ⋁C(i) ⋁ U(i) ⋁ PD(i) ∀i∈I s.t. i is a non-softgoal leaf: i must not have conflicting predicates (?) Finding Solutions in Goal Models - Horkoff, Yu
Forward Propagation Axioms • We develop axioms to express forward and backward propagation • Forward Propagation, for i∈ I, R:i1 x … x in→i: • (Some combination of v(i1)…v(in)) → v(i) • Samples (complete list in paper): • Rhlp: S(i) → PS(i) PS(i) → PS(i) U(i) → U(i) C(i) → C(i) PD(i) → PD(i) D(i) → PD(i) Finding Solutions in Goal Models - Horkoff, Yu
Forward Propagation Axioms • Forward Propagation, for i∈ I, R:i1 x … x in→i: • (Some combination of v(i1)…v(in)) → v(i) • Rdec: Finding Solutions in Goal Models - Horkoff, Yu
Backward Propagation Axioms • Decomposition, Means-Ends and Dependency backward axioms are the inverse of forward axioms • Forward Propagation, for i∈ I, R:i1 x … x in→i: • (Some combination of v(i1)…v(in)) → v(i) • Backward Propagation, for i∈ I, R:i1 x … x in→i: • v(i)→(Some combination of v(i1)…v(in)) • Contribution links are different – information is lost during value combination in forward direction • In the backward direction, we can make limited assumptions Finding Solutions in Goal Models - Horkoff, Yu
Human Judgment • An intention, i ∈ I, needs human judgment if: • i is the recipient of more than one contribution link, AND: • There is a conflict situation • OR, PS(i) or PD(i) holds and i has not received a human judgment in the previous iteration (allows promotion of partial evidence) • When a judgment is provided, the CNF encoding is adjusted as follows: • Forward and backward axioms propagating to or from i are removed • (Any combination of v(i1)…v(in)) → v(i) • v(i)→ (Any combination of v(i1)…v(in)) • New axioms representing the judgment are added • Forward: (v(i1) ⋀ … ⋀ v(in)) → judgment(i) • Backward: judgment(i) → (v(i1) ⋀ … ⋀ v(in)) Finding Solutions in Goal Models - Horkoff, Yu
Example: Interactive Backward Satisfaction Analysis Target(s) unsatisfiable The following intentions are involved in the conflict: Security PS Attract Users PS Usability PS The following intentions are the sources of the conflict: Restrict Structure of Password PS, D, not PD, PD Human Judgment Backtrack Human Judgment Backtrack Conflict Is it possible for Attract Users to be Partially Satisfied? If so how? If not, why not? Finding Solutions in Goal Models - Horkoff, Yu
Implementation • Implemented in OpenOME • Worst case run time: O(6q(l * n2 + n * runtime(SAT))) • There are 6 evaluation labels • n = number of intentions • q = maximum number of sources for intentions • l = number of links in the model • Procedure has been applied to several medium to large sized models • If the user does not repeat judgments, procedure will terminate Finding Solutions in Goal Models - Horkoff, Yu
Qualitative Studies of Interactive Goal Model Analysis • Tested the procedure with a group implementing the inflo “back-of-the-envelope” calculation modeling tool • Ten individual case studies (PoEM’10) • Qualitative analysis of results (no statistical significance) • Observations showed benefits and revealed usability issues • Individuals and groups able to understand and apply backward analysis • In some cases interesting discoveries were made about the model and domain (model incompleteness, meaning of intentions) • Analysis may be less useful on incomplete, smaller models, or models without tradeoffs • i* and evaluation training, or presence of an experienced facilitator is needed to get the full intended benefits Finding Solutions in Goal Models - Horkoff, Yu
Inflo Example Finding Solutions in Goal Models - Horkoff, Yu
Conclusions and Future Work • Expanded the analysis ability of i* models as part of an interactive, iterative framework aimed for Early RE • Future work includes: • Procedure optimizations • Optionally reusing human judgment • Deriving judgments from existing judgments • Improving run time • Further visualizations • Highlighting intentions involved in human judgment • Links coloring for conflicts • Further applications • More realistic action research setting Finding Solutions in Goal Models - Horkoff, Yu
Thank you • jenhork@cs.utoronto.ca • www.cs.utoronto.ca/~jenhork • yu@ischool.utoronto.ca • www.cs.utoronto.ca/~eric • OpenOME: • https://se.cs.toronto.edu/trac/ome Finding Solutions in Goal Models - Horkoff, Yu
Related Work S, PD PS, PD PS, D PS • Giorgini et al. have introduced a formal framework for (forward and) backward reasoning with goal models (CAiSE’04) • We use the CNF formula and some of the axioms from this work; however, we make the following expansions and modifications: • Incorporating interaction through human judgment – making the procedure more amenable to Early RE • Accounting for additional agent-goal syntax (dependencies, unknown) and analysis values (conflict, unknown) • Producing results with only one value per intention • Providing information on model conflicts when a solution cannot be found Finding Solutions in Goal Models - Horkoff, Yu