<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://vrl.cs.brown.edu/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Eric+Sodomka</id>
	<title>VrlWiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="http://vrl.cs.brown.edu/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Eric+Sodomka"/>
	<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php/Special:Contributions/Eric_Sodomka"/>
	<updated>2026-04-19T21:31:51Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3379</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3379"/>
		<updated>2009-05-01T16:53:22Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Parallel Framework for Evaluation Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Contributions =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[I have removed almost everything that does not describe itself as a contribution.  If your part is missing, a revision is in order and you can find what the former version on the page linked below (David)]&lt;br /&gt;
&lt;br /&gt;
[[CS295J/Thursday 2pm version of final contribution|Thursday 2pm version of final contribution]]&lt;br /&gt;
&lt;br /&gt;
== Workflow, Multi-tasking and Interruptions ==&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop a Theory of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
* Build a Meta-work Assistance Tool&lt;br /&gt;
:# We will perform a series of ecologically-valid studies to compare user performance between a state of the art/popular task management system (control group) and our meta-work assistance tool (experimental group) &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Didn&#039;t your study show that there is nothing close to a killer app in this department? - Steven]&amp;lt;/span&amp;gt; [By &amp;quot;state of the art&amp;quot; I mean a popular, well-recognized system (or systems), such as Outlook, that are used by large numbers of people.]&lt;br /&gt;
* Evaluate the Module-based Model of HCI by using it to predict the outcome of the above-mentioned evaluation; then compare with empirical results&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Theory of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:* Internal dependencies: none&lt;br /&gt;
* Empirical Evaluation of Meta-work Assistance Tool in an Ecologically-valid Context&lt;br /&gt;
:* Internal dependencies: Completion of a testable meta-work tool&lt;br /&gt;
* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:* Internal dependencies: Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration (we might consider calling these deliverables)&#039;&#039;&#039;&lt;br /&gt;
* A predictive, quantiative model of controlled task performance and interruption resumption time for varying sized working sets&lt;br /&gt;
* An ecologically valid evaluation which shows that the completed meta-work assistance tool outperforms a state of the art system such as Outlook&lt;br /&gt;
* A comparison between the result predicted by the module-based model of HCI and the evaluation mentioned above; accompanied by a detailed analysis of factors contributing to the result&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;External Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Dependent on most of the rest of the proposal being completed to the point of being usable to predict the outcome of the meta-work tool evaluation&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Salvucci et al&#039;s integrated theory of multi-tasking continuum integrates concurrent and sequential task performance&lt;br /&gt;
* Gloria Mark&#039;s research group has conducted a number of empirical studies into multi-tasking and interruption&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Preliminary Work&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* [[Study of Meta-work Tools and Strategies of 7 Information Workers]]&lt;br /&gt;
:* Results suggest there is not a single &amp;quot;ideal&amp;quot; meta-work tool today, motivating our proposal to build a new meta-work support tool based in part on a scientific theory of multi-tasking and interruptions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reviews&#039;&#039;&#039;&lt;br /&gt;
* Not in the expected format -- split into appropriate parts and label them as in the other sections (David)&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[How do you propose to get around the problem Andrew brought up of accurately linking this low-level data to the interface/program&#039;s state at the moment it occurred? - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Low-level interactions will need to be logged along with a markup-language-style state description of the application itself.  This would either require access to the source code, or a detailed interface description from the interface designer that would allow for complete simulation of the application.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Perhaps something should be said about sectioning this data off into modules - i.e. compiling all e-mail writing recordings to analyze separately from e-mail sorting recordings in order to facilitate analysis. - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I believe this is addressed in the &amp;quot;semantic-level chunking&amp;quot; section, given below.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Things like pupil-tracking can be quantitatively evaluated through users tests.  For instance, we could instruct users to shift the focus of their eyes from one on-screen object to another as quickly as possible some fixed number of times. We could then compare the results from our system to determine accuracy and speed, and by varying the size and distance of the target objects, we could obtain richer data for analysis.  With respect to clustering low-level interactions, I believe that issue is addressed in the following section.]&amp;lt;/span&amp;gt;&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Should this take into account the need to match recordings to interface state? - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Possibly, if we go down the simulation route.  If we&#039;re actually logging all of the events from within the source code itself, I believe the demonstration of syncing low-level interaction with application-state is self-evident.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;Can you briefly summarize the utility with one sentence? Or is that what the next bullet does? (Jon)&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[In the instance of this work, utility was demonstrated with respect to automatically generating presentations from data explorations, and providing quantitative interface-usage analysis that aided future decisions for interface designers.  i.e., this operations was performed a lot, and this one was not; therefore we should focus on making the first more efficient, perhaps at the expense of the second.  The results of the work are fairly basic, but they do demonstrate that logging histories is a useful practice for a number of reasons.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
* Two thumbs up. - Steven&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I don&#039;t intend for these techniques to be performed manually.  Application-specific rules for chunking will be created manually, most likely by the interface designer, but those rules will be applied automatically to filter low-level primitives and application-states into semantic-level interactions.  The success of such methods could be evaluated by having automatically-generated semantic-level histories compared to those manually created by a user.  Of course, there&#039;s no guarantee that what a user manually creates is any good, so I think it would be more effective to have a quantitative study of user performance in which timed-tasks are completed with and without the ability to review and revisit semantic-level histories.  The test without any history at all would serve as a baseline, and then performance with histories created using different chunking rules would speak to the effectiveness of different strategies for chunking.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I&#039;d like to see this fleshed out in a theoretical manner. - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I believe the previous section speaks to this question, but perhaps not with enough detail.  It seems appropriate to have this discussion in class.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)  &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[This is a great question.  It really depends on what&#039;s meant by accuracy.  If the only level of semantics we want is that defined by the interface-designer, then it&#039;s not really an interesting question.  As we begin to talk about semantics with respect to what the user &#039;&#039;perceives&#039;&#039; or what the user&#039;s &#039;&#039;intent&#039;&#039; might be, it becomes more interesting, and the term accuracy becomes more ambiguous.  I suppose accuracy could be tested by having users verbally speak their intent as they use an interface, and compare their descriptions with the output of the system, but I think that might be a messy study.  Interesting to think about.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;Can you include something about the broader impact of this contribution? (Jon)&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I think the broader impact here is offering interface designers tools to analyze interaction at a high enough level that they can begin to observe user&#039;s intent with an interface, and potentially infer overarching user-goals.  If these tools prove effective, and could be implemented in real-time, there&#039;s a wealth of possibility in adaptive interfaces.  In addition, the semantic-level interaction history construct has the potential to provide meaningful training data for machine-learning algorithms that could probabilistically model user-action and be used to offer suggested interactions.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A framework for the unification of established heuristic usability guidelines and accepted cognitive principles.  We propose a matrix pairing cognitive principles and heuristic guidelines into specific interface improvement recommendations.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Language cleaned up.  While I still think it is important to give more &amp;quot;weight&amp;quot; to cognitive/heuristic principle pairs that agree more strongly than others and output more useful recommendations, I believe I&#039;ve pitched it wrong and made it too prominent an attribute of this module.  I&#039;ve adjusted. [[User:E J Kalafarski|E J Kalafarski]] 15:34, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A human &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A human &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several recommendations from a matrix of cogntive/heuristic principle pairs.&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the principle pairs versus separate application of cognitive and heuristic guidelines.&lt;br /&gt;
* Matrix can be improved incrementally by applying &amp;quot;weight&amp;quot; to each cell of the matrix, increasing its influence on final output recommendations, based on the measurable success of this study.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Added more explanatory description of methodology as a bullet under Demo.  [[User:E J Kalafarski|E J Kalafarski]] 15:49, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon) &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Yes. [[User:E J Kalafarski|E J Kalafarski]] 15:49, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Gideon&#039;s comment brings up an important point - in what form will the weights be presented? - Steven &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Values which contribute to the &amp;quot;influence&amp;quot; of each recommendation. [[User:E J Kalafarski|E J Kalafarski]] 15:49, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
* Could be more explicit about whether the experts are simulated agents or real people (Jon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a set of interface evaluation modules for analysis, we demonstrate the proposed efficiency and accuracy of a method for aggregating these interface suggestions.  In this preliminary demonstration, we limit the field to a strictly-defined interface and 2–3 CPM principles (e.g. Fitts&#039; Law and Affordance).  Unified recommendations are generated for application to the interface for incremental improvement in usability.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A limited but strictly-controlled study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Framework for the input of a given interface to individual modules, and standardization of the output of said modules.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon) &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I limited the scope of this initial demonstration. [[User:E J Kalafarski|E J Kalafarski]] 15:46, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I think this is the same as my contribution, and that we should merge them together. You&#039;ve covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation. You also gave a demonstration of comparing recommendations, while I only covered evaluations. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* I think this contribution and mine (the CPM-GOMS improvement) will need to be reconciled into a single unit. - Steven&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Is it feasible to combine all three of these contributions? [[User:E J Kalafarski|E J Kalafarski]] 15:47, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
* Interesting to note that this could describe the overarching architecture but it depends upon components of the architecture to work.  Would it depend on any other modules? (Jon) &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Yes, several modules are listed as dependencies. [[User:E J Kalafarski|E J Kalafarski]] 15:46, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# In user studies, this module predicted time-to-completion more accurately than the standard CPM-GOMS model.&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with proposed cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# User studies were conducted during which participants brain&#039;s were scanned with EEG to measure activation levels.&lt;br /&gt;
:::# Self-report measures were obtained describing the amount of concentration/effort required on the part of the user.&lt;br /&gt;
:::# The literature base provided an accurate assessment of cognitive load in the tasks involved.&lt;br /&gt;
:::# Facial recognition allows us to approximate spans of high-load (e.g., furled brow)&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Self-Report Measure&lt;br /&gt;
::::# Database of load per cognitive tasks&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Frustration is measured by asking participants to respond on the self-report scale.&lt;br /&gt;
:::# Facial recognition allows us to approximate spans of frustration (e.g., frowning, sighing, furled brow)&lt;br /&gt;
:::# Galvanic Skin Response allows detection of emotional arousal&lt;br /&gt;
:::# Heart rate monitoring allows detection of emotional arousal&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response input&lt;br /&gt;
::::# Heart rate response input&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I agree that this pretty close to E.J.&#039;s contribution, and that maybe they should be merged. E.J. covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation, and his gave a demonstration of comparing recommendations, while I only covered evaluations.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* I feel like this structure&#039;s clarity may come at expense of specificity, I&#039;d like to know more than that it&#039;s parallel. - Steven&lt;br /&gt;
* Very thorough demonstration(s)! (Jon)&lt;br /&gt;
* I think that the contribution is much more clearly stated here -- I would argue for integrating them together. (Andrew)&lt;br /&gt;
&lt;br /&gt;
== Anti-Pattern Recognition &amp;amp; Resolution ==&lt;br /&gt;
Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims:&#039;&#039;&#039;&lt;br /&gt;
* Develop a framework for analyzing an interface and recognizing violations of  established patterns of possible interaction&lt;br /&gt;
* Develop an algorithm to quantitatively rate the amount of anti-pattern violations&lt;br /&gt;
* Integrate cognitive theory in providing suitable alternatives&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions:&#039;&#039;&#039;&lt;br /&gt;
* Creation of established patterns of interaction as determined by user studies&lt;br /&gt;
* Creation of algorithm to detect match and library elements to interface elements&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039;&lt;br /&gt;
* The viability of this module will be tested through providing it with a poorly designed interface. &lt;br /&gt;
** The design would be analyzed and tested by: a usability expert, normal users, the module whereby each would have to eventually provide a value as to how much the interface exhibits anti-patterns.&lt;br /&gt;
** Based on the results provided, we should expect to see the expert and module reports to be similar&lt;br /&gt;
	&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039;&lt;br /&gt;
* Library of standard design patterns&lt;br /&gt;
* Affordances module&lt;br /&gt;
* Custom design pattern libraries (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs:&#039;&#039;&#039;&lt;br /&gt;
* Formal interface description&lt;br /&gt;
* Tasks which can be performed within the interface&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs:&#039;&#039;&#039;&lt;br /&gt;
* Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
* Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[In theory could the non-semantic inputs be automatically read from the interface? - Steven]&amp;lt;/span&amp;gt;&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Could this module benefit from eye-tracking? - Steven]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3378</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3378"/>
		<updated>2009-05-01T16:53:06Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Parallel Framework for Evaluation Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Contributions =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[I have removed almost everything that does not describe itself as a contribution.  If your part is missing, a revision is in order and you can find what the former version on the page linked below (David)]&lt;br /&gt;
&lt;br /&gt;
[[CS295J/Thursday 2pm version of final contribution|Thursday 2pm version of final contribution]]&lt;br /&gt;
&lt;br /&gt;
== Workflow, Multi-tasking and Interruptions ==&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop a Theory of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
* Build a Meta-work Assistance Tool&lt;br /&gt;
:# We will perform a series of ecologically-valid studies to compare user performance between a state of the art/popular task management system (control group) and our meta-work assistance tool (experimental group) &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Didn&#039;t your study show that there is nothing close to a killer app in this department? - Steven]&amp;lt;/span&amp;gt; [By &amp;quot;state of the art&amp;quot; I mean a popular, well-recognized system (or systems), such as Outlook, that are used by large numbers of people.]&lt;br /&gt;
* Evaluate the Module-based Model of HCI by using it to predict the outcome of the above-mentioned evaluation; then compare with empirical results&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Theory of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:* Internal dependencies: none&lt;br /&gt;
* Empirical Evaluation of Meta-work Assistance Tool in an Ecologically-valid Context&lt;br /&gt;
:* Internal dependencies: Completion of a testable meta-work tool&lt;br /&gt;
* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:* Internal dependencies: Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration (we might consider calling these deliverables)&#039;&#039;&#039;&lt;br /&gt;
* A predictive, quantiative model of controlled task performance and interruption resumption time for varying sized working sets&lt;br /&gt;
* An ecologically valid evaluation which shows that the completed meta-work assistance tool outperforms a state of the art system such as Outlook&lt;br /&gt;
* A comparison between the result predicted by the module-based model of HCI and the evaluation mentioned above; accompanied by a detailed analysis of factors contributing to the result&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;External Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Dependent on most of the rest of the proposal being completed to the point of being usable to predict the outcome of the meta-work tool evaluation&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Salvucci et al&#039;s integrated theory of multi-tasking continuum integrates concurrent and sequential task performance&lt;br /&gt;
* Gloria Mark&#039;s research group has conducted a number of empirical studies into multi-tasking and interruption&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Preliminary Work&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* [[Study of Meta-work Tools and Strategies of 7 Information Workers]]&lt;br /&gt;
:* Results suggest there is not a single &amp;quot;ideal&amp;quot; meta-work tool today, motivating our proposal to build a new meta-work support tool based in part on a scientific theory of multi-tasking and interruptions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reviews&#039;&#039;&#039;&lt;br /&gt;
* Not in the expected format -- split into appropriate parts and label them as in the other sections (David)&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[How do you propose to get around the problem Andrew brought up of accurately linking this low-level data to the interface/program&#039;s state at the moment it occurred? - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Low-level interactions will need to be logged along with a markup-language-style state description of the application itself.  This would either require access to the source code, or a detailed interface description from the interface designer that would allow for complete simulation of the application.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Perhaps something should be said about sectioning this data off into modules - i.e. compiling all e-mail writing recordings to analyze separately from e-mail sorting recordings in order to facilitate analysis. - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I believe this is addressed in the &amp;quot;semantic-level chunking&amp;quot; section, given below.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Things like pupil-tracking can be quantitatively evaluated through users tests.  For instance, we could instruct users to shift the focus of their eyes from one on-screen object to another as quickly as possible some fixed number of times. We could then compare the results from our system to determine accuracy and speed, and by varying the size and distance of the target objects, we could obtain richer data for analysis.  With respect to clustering low-level interactions, I believe that issue is addressed in the following section.]&amp;lt;/span&amp;gt;&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Should this take into account the need to match recordings to interface state? - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Possibly, if we go down the simulation route.  If we&#039;re actually logging all of the events from within the source code itself, I believe the demonstration of syncing low-level interaction with application-state is self-evident.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;Can you briefly summarize the utility with one sentence? Or is that what the next bullet does? (Jon)&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[In the instance of this work, utility was demonstrated with respect to automatically generating presentations from data explorations, and providing quantitative interface-usage analysis that aided future decisions for interface designers.  i.e., this operations was performed a lot, and this one was not; therefore we should focus on making the first more efficient, perhaps at the expense of the second.  The results of the work are fairly basic, but they do demonstrate that logging histories is a useful practice for a number of reasons.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
* Two thumbs up. - Steven&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I don&#039;t intend for these techniques to be performed manually.  Application-specific rules for chunking will be created manually, most likely by the interface designer, but those rules will be applied automatically to filter low-level primitives and application-states into semantic-level interactions.  The success of such methods could be evaluated by having automatically-generated semantic-level histories compared to those manually created by a user.  Of course, there&#039;s no guarantee that what a user manually creates is any good, so I think it would be more effective to have a quantitative study of user performance in which timed-tasks are completed with and without the ability to review and revisit semantic-level histories.  The test without any history at all would serve as a baseline, and then performance with histories created using different chunking rules would speak to the effectiveness of different strategies for chunking.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I&#039;d like to see this fleshed out in a theoretical manner. - Steven]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I believe the previous section speaks to this question, but perhaps not with enough detail.  It seems appropriate to have this discussion in class.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)  &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[This is a great question.  It really depends on what&#039;s meant by accuracy.  If the only level of semantics we want is that defined by the interface-designer, then it&#039;s not really an interesting question.  As we begin to talk about semantics with respect to what the user &#039;&#039;perceives&#039;&#039; or what the user&#039;s &#039;&#039;intent&#039;&#039; might be, it becomes more interesting, and the term accuracy becomes more ambiguous.  I suppose accuracy could be tested by having users verbally speak their intent as they use an interface, and compare their descriptions with the output of the system, but I think that might be a messy study.  Interesting to think about.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;Can you include something about the broader impact of this contribution? (Jon)&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I think the broader impact here is offering interface designers tools to analyze interaction at a high enough level that they can begin to observe user&#039;s intent with an interface, and potentially infer overarching user-goals.  If these tools prove effective, and could be implemented in real-time, there&#039;s a wealth of possibility in adaptive interfaces.  In addition, the semantic-level interaction history construct has the potential to provide meaningful training data for machine-learning algorithms that could probabilistically model user-action and be used to offer suggested interactions.]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A framework for the unification of established heuristic usability guidelines and accepted cognitive principles.  We propose a matrix pairing cognitive principles and heuristic guidelines into specific interface improvement recommendations.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Language cleaned up.  While I still think it is important to give more &amp;quot;weight&amp;quot; to cognitive/heuristic principle pairs that agree more strongly than others and output more useful recommendations, I believe I&#039;ve pitched it wrong and made it too prominent an attribute of this module.  I&#039;ve adjusted. [[User:E J Kalafarski|E J Kalafarski]] 15:34, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A human &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A human &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several recommendations from a matrix of cogntive/heuristic principle pairs.&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the principle pairs versus separate application of cognitive and heuristic guidelines.&lt;br /&gt;
* Matrix can be improved incrementally by applying &amp;quot;weight&amp;quot; to each cell of the matrix, increasing its influence on final output recommendations, based on the measurable success of this study.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Added more explanatory description of methodology as a bullet under Demo.  [[User:E J Kalafarski|E J Kalafarski]] 15:49, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon) &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Yes. [[User:E J Kalafarski|E J Kalafarski]] 15:49, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Gideon&#039;s comment brings up an important point - in what form will the weights be presented? - Steven &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Values which contribute to the &amp;quot;influence&amp;quot; of each recommendation. [[User:E J Kalafarski|E J Kalafarski]] 15:49, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
* Could be more explicit about whether the experts are simulated agents or real people (Jon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a set of interface evaluation modules for analysis, we demonstrate the proposed efficiency and accuracy of a method for aggregating these interface suggestions.  In this preliminary demonstration, we limit the field to a strictly-defined interface and 2–3 CPM principles (e.g. Fitts&#039; Law and Affordance).  Unified recommendations are generated for application to the interface for incremental improvement in usability.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A limited but strictly-controlled study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Framework for the input of a given interface to individual modules, and standardization of the output of said modules.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon) &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I limited the scope of this initial demonstration. [[User:E J Kalafarski|E J Kalafarski]] 15:46, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I think this is the same as my contribution, and that we should merge them together. You&#039;ve covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation. You also gave a demonstration of comparing recommendations, while I only covered evaluations. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* I think this contribution and mine (the CPM-GOMS improvement) will need to be reconciled into a single unit. - Steven&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Is it feasible to combine all three of these contributions? [[User:E J Kalafarski|E J Kalafarski]] 15:47, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
* Interesting to note that this could describe the overarching architecture but it depends upon components of the architecture to work.  Would it depend on any other modules? (Jon) &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[Yes, several modules are listed as dependencies. [[User:E J Kalafarski|E J Kalafarski]] 15:46, 1 May 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# In user studies, this module predicted time-to-completion more accurately than the standard CPM-GOMS model.&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with proposed cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# User studies were conducted during which participants brain&#039;s were scanned with EEG to measure activation levels.&lt;br /&gt;
:::# Self-report measures were obtained describing the amount of concentration/effort required on the part of the user.&lt;br /&gt;
:::# The literature base provided an accurate assessment of cognitive load in the tasks involved.&lt;br /&gt;
:::# Facial recognition allows us to approximate spans of high-load (e.g., furled brow)&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Self-Report Measure&lt;br /&gt;
::::# Database of load per cognitive tasks&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Frustration is measured by asking participants to respond on the self-report scale.&lt;br /&gt;
:::# Facial recognition allows us to approximate spans of frustration (e.g., frowning, sighing, furled brow)&lt;br /&gt;
:::# Galvanic Skin Response allows detection of emotional arousal&lt;br /&gt;
:::# Heart rate monitoring allows detection of emotional arousal&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response input&lt;br /&gt;
::::# Heart rate response input&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt; &amp;lt;span style=&amp;quot;color: blue;&amp;quot;&amp;gt;[I think this pretty close to E.J.&#039;s contribution, and that maybe they should be merged. E.J. covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation, and his gave a demonstration of comparing recommendations, while I only covered evaluations.]&amp;lt;/span&amp;gt;&lt;br /&gt;
* I feel like this structure&#039;s clarity may come at expense of specificity, I&#039;d like to know more than that it&#039;s parallel. - Steven&lt;br /&gt;
* Very thorough demonstration(s)! (Jon)&lt;br /&gt;
* I think that the contribution is much more clearly stated here -- I would argue for integrating them together. (Andrew)&lt;br /&gt;
&lt;br /&gt;
== Anti-Pattern Recognition &amp;amp; Resolution ==&lt;br /&gt;
Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims:&#039;&#039;&#039;&lt;br /&gt;
* Develop a framework for analyzing an interface and recognizing violations of  established patterns of possible interaction&lt;br /&gt;
* Develop an algorithm to quantitatively rate the amount of anti-pattern violations&lt;br /&gt;
* Integrate cognitive theory in providing suitable alternatives&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions:&#039;&#039;&#039;&lt;br /&gt;
* Creation of established patterns of interaction as determined by user studies&lt;br /&gt;
* Creation of algorithm to detect match and library elements to interface elements&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039;&lt;br /&gt;
* The viability of this module will be tested through providing it with a poorly designed interface. &lt;br /&gt;
** The design would be analyzed and tested by: a usability expert, normal users, the module whereby each would have to eventually provide a value as to how much the interface exhibits anti-patterns.&lt;br /&gt;
** Based on the results provided, we should expect to see the expert and module reports to be similar&lt;br /&gt;
	&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039;&lt;br /&gt;
* Library of standard design patterns&lt;br /&gt;
* Affordances module&lt;br /&gt;
* Custom design pattern libraries (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs:&#039;&#039;&#039;&lt;br /&gt;
* Formal interface description&lt;br /&gt;
* Tasks which can be performed within the interface&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs:&#039;&#039;&#039;&lt;br /&gt;
* Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
* Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[In theory could the non-semantic inputs be automatically read from the interface? - Steven]&amp;lt;/span&amp;gt;&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Could this module benefit from eye-tracking? - Steven]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3276</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3276"/>
		<updated>2009-04-29T17:06:21Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Logic */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[...HCI design principles? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I think this is the same as my contribution, and that we should merge them together. You&#039;ve covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation. You also gave a demonstration of comparing recommendations, while I only covered evaluations. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[More detail on each demonstration would be helpful. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I really like the xml, too. I&#039;m guessing you just have placeholder text for some of these values (e.g. &amp;quot;pulling hair out&amp;quot;), but maybe we should think more about what exactly is returned.  I was thinking the evaluation values would be some sort of utility value between 0-1, but this makes the &amp;quot;time&amp;quot; dimension tricky (what does a &amp;quot;time&amp;quot; score of 0.75 mean?). -Eric] &amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3275</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3275"/>
		<updated>2009-04-29T17:06:08Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Logic */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[...HCI design principles? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I think this is the same as my contribution, and that we should merge them together. You&#039;ve covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation. You also gave a demonstration of comparing recommendations, while I only covered evaluations. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[More detail on each demonstration would be helpful. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I really like the xml, too. I&#039;m guessing you just have placeholder text for some of these values (e.g. &amp;quot;pulling hair out&amp;quot;), but maybe we should think more about what exactly is returned.  I was thinking the evaluation values would be some sort of utility value between 0-1, but this makes the &amp;quot;time&amp;quot; dimension tricky (what does a &amp;quot;time&amp;quot; score of 0.75 mean?).] &amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3274</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3274"/>
		<updated>2009-04-29T16:58:18Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Evaluation Metrics */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[...HCI design principles? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I think this is the same as my contribution, and that we should merge them together. You&#039;ve covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation. You also gave a demonstration of comparing recommendations, while I only covered evaluations. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[More detail on each demonstration would be helpful. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3273</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3273"/>
		<updated>2009-04-29T16:52:48Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Evaluation for Recommendation and Incremental Improvement */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[...HCI design principles? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
* &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I think this is the same as my contribution, and that we should merge them together. You&#039;ve covered some things that I didn&#039;t, such as a more detailed demonstration of how our framework is better than a human interface designer&#039;s evaluation. You also gave a demonstration of comparing recommendations, while I only covered evaluations. -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3272</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3272"/>
		<updated>2009-04-29T16:47:42Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Reconciling Usability Heuristics with Cognitive Theory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[...HCI design principles? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[There is still the question of how these weights are determined. Is this the job of the third expert? Or is the third expert given these weights, and he just determines how to apply them? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3271</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3271"/>
		<updated>2009-04-29T16:45:42Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Reconciling Usability Heuristics with Cognitive Theory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[...HCI design principles? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3270</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3270"/>
		<updated>2009-04-29T16:43:23Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Reconciling Usability Heuristics with Cognitive Theory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is there an explicit, quantifiable demonstration you can propose?  Such some kind of conclusive comparison between automated capture techniques and manual techniques?  Or maybe a demonstration of some kind of methodical, but still low-level, procedure for clustering related interaction primitives? [[User:E J Kalafarski|E J Kalafarski]] 15:02, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;review&#039;&#039;&#039;&#039;&lt;br /&gt;
* This is recording everything possible during an interaction. It is necessary to do for our project. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[As above, how will this be validated quantitatively?  Comparing two methodologies for manual chunking?  What would these differing methodologies be? [[User:E J Kalafarski|E J Kalafarski]] 15:06, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
* With what level of accuracy can this be performed. I like the idea, its worthy of a nice project in and of itself. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that the wording is is a little confusing. The framework is not weighted, but rather it weights the different guidelines/principles. It might also be worth explaining how this is a useful contribution, e.g. does it allow for more accurate interface evaluation? -Eric]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* This seems like the 2d matrix? Is this implemented as a module? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* I see this as being the &amp;quot;black box&amp;quot; for our architecture? If so, good. Wouldn&#039;t the dependencies be any/all modules? (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I agree that a metric (or several) is crucial to our framework, but what is your demonstration of the feasibility or usability of these particular values you&#039;ve chosen.  On an unrelated note, we talked a lot about how some of these might be &amp;quot;convertible&amp;quot; into others…do you see these metrics as different sides of the same coin?  Can the units of one be &amp;quot;converted&amp;quot; into the units of another, like gallons to liters? [[User:E J Kalafarski|E J Kalafarski]] 15:10, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Words cannot express how amazing this contribution is. It should be given as much money as requested without question.&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* What exactly are the differences between this and EJ&#039;s earlier contribution? I think that if they are the same, this one is a bit more clear, IMHO. (Gideon)&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I have a similar question to Gideon&#039;s.  My hunch is that this defines a sort of &amp;quot;namespace&amp;quot; for the modules themselves, while the other contribution (the one you guys assigned to me) is more explicitly the aggregator of these outputs.  Correct me if I&#039;m wrong.  But if that&#039;s the case, the two contributions might need to be validated similarly/together. [[User:E J Kalafarski|E J Kalafarski]] 15:17, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;review&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Cool- I like the xml. (Gideon)&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
Evaluation method:&lt;br /&gt;
* Evaluate an interface via the proposed framework, as well as traditional CPM-GOMS and ACT-R methods.  Also have a small team of interface design experts evaluate the interface.  Solicit comments on the current interface and suggestions for improvement from each team/method, and compare the accuracy and validity of the results.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*Provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&lt;br /&gt;
*Integrate this module into our novel cognitive framework for interface evaluation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* [Not sure here.  Is this really novel?]  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[I like it.  I think it&#039;s necessary.  I think we demonstrated in class that this has not been formalized and standardized already. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module by generating a formal description of an interface for scientific visualization, a formal description of a task to perform with the interface, and we will correlate the estimated time to complete the task based on Fitt&#039;s law with the actual time required based on several user traces.  &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Check. [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a formal description of the interface with graph nodes representing clickable interface elements and graph edges representing the physical (on-screen) distance between adjacent nodes. &amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Is this graph format you&#039;re suggesting part of the proposal, or is the &amp;quot;language&amp;quot; itself a dependency from literature? [[User:E J Kalafarski|E J Kalafarski]] 15:22, 29 April 2009 (UTC)]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Module Description&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Inputs&lt;br /&gt;
**A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
**A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
**The physical distances between interface elements along those paths.&lt;br /&gt;
**The width of those elements along the most likely axes of motion.&lt;br /&gt;
**Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
*To provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing specified tasks.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
*A quantitative measure of the extent to which an interface suggests to the user the actions that it is capable of performing.&lt;br /&gt;
*A quantitative, indirect measure of the extent to which an interface facilitates (or hinders) the use of fast perceptual mechanisms.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Again, I&#039;m a fan.  I don&#039;t think this has been formalized already. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstrations&#039;&#039;&#039;&lt;br /&gt;
*We will demonstrate the feasibility of this module through the following experiment:&lt;br /&gt;
**Specify a task for a user to perform with scientific visualization software.&lt;br /&gt;
**There should be several different ways to complete the task (paths through the space of possible interface actions).&lt;br /&gt;
**Some of these paths will be more direct than others.&lt;br /&gt;
**We will then measure the number of task-relevant affordances that were perceived and acted upon by analyzing the user trace, and the time required to complete the task.&lt;br /&gt;
**Use the formula: (affordances perceived) / [(relevant affordances present) * (time to complete task)].&lt;br /&gt;
**Correlate the resulting scores with verbal reports on naturalness and ease-of-use for the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
*Requires a method for capturing user traces that include formalized records of interface elements used for a particular, pre-specified task or set of tasks.&lt;br /&gt;
*Providing suggestions/recommendations will require interaction with other modules that analyze the perceptual salience of interface elements.&lt;br /&gt;
&amp;lt;span style=&amp;quot;color: gray;&amp;quot;&amp;gt;[Some kind of formalized list or arbitrary classification of affordances might be necessary to limit the scope of this contribution. [[User:E J Kalafarski|E J Kalafarski]] 15:24, 29 April 2009 (UTC)]&amp;lt;/span&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Description of the Module&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
*Module Inputs&#039;&#039;&lt;br /&gt;
**Formalized descriptions of...&lt;br /&gt;
***Interface elements&lt;br /&gt;
***Their associated actions&lt;br /&gt;
***The functions of those actions&lt;br /&gt;
***A particular task&lt;br /&gt;
***User traces for that task&lt;br /&gt;
&lt;br /&gt;
*Processing&lt;br /&gt;
**Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
*Output&lt;br /&gt;
**The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3241</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3241"/>
		<updated>2009-04-28T20:44:26Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Parallel Framework for Evaluation Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3240</id>
		<title>CS295J/Final contributions</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Final_contributions&amp;diff=3240"/>
		<updated>2009-04-28T20:44:03Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
[David: I have removed some unattributed stuff that wasn&#039;t in the format described in assignment 13.  If that was in error, go back to an earlier version to recover the text, attribute it, and put in the correct format.]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multi-modal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* Techniques for pupil-tracking, auditory-recognition, and muscle-activity monitoring will be evaluated with respect to accuracy, sampling rate, and computational cost in a series of quantitative studies.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Hardware for pupil-tracking and muscle activity monitoring.  Commercial software packages for such software may be required.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic-chunking techniques for interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;&lt;br /&gt;
* These techniques will be validated quantitatively with respect to efficiency and user error rate on a set of timed-tasks over a series of case studies.  In addition, qualitative feedback on user satisfaction will be incorporated in the evaluation of our methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies&#039;&#039;&#039;&lt;br /&gt;
* Software framework for collecting user interactions.&lt;br /&gt;
* Formal description of interface functionality.&lt;br /&gt;
* Description of data objects that can be manipulated through interaction.&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution&#039;&#039;&#039;: A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m having trouble parsing this contribution.  How do you weight a framework?  Also, I&#039;d like to have a little more sense of how this might fit into the bigger picture.  The assignment didn&#039;t ask for that, but is there some way to provide some of that context?]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;: Using a limited set of interface evaluation modules for analysis, we demonstrate, in a narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
[David: I&#039;m a little lost.  Can you make this long sentence into a couple shorter simpler ones?  I think I can imagine what you are getting at, but I&#039;m not sure.]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
[David: nice demonstration!]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
[David: I think this also has, as a dependency, some kind of framework for the modules.  &amp;quot;narrow but comprehensive&amp;quot; sounds challenging.  ]&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contribution:&#039;&#039;&#039; &#039;Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration:&#039;&#039;&#039; Run a series of user studies and compare users&#039; performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
# Traditional, manual interface evaluation &lt;br /&gt;
#* As a baseline.&lt;br /&gt;
# Using our system with a single module&lt;br /&gt;
#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
#* For validating the use of a dynamic weighting system.&lt;br /&gt;
# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependencies:&#039;&#039;&#039; Requires a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
[David: I think that the work proposed here is interesting, but I&#039;m not sure that it is organized by &amp;quot;contribution&amp;quot;.  Try labeling the contributions and the demonstrations explicitly.  A &amp;quot;tool&amp;quot; is usually not a contribution -- an evaluation of the tool could be.  If this distinction isn&#039;t clear, let&#039;s talk.  For these particular contributions, phrasing and integrating them is a challenge because some of the contributions are demonstrations of other contributions...]&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
=== Anti-Pattern Conflict Resolution ===&lt;br /&gt;
* Owner: [[User: Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
Interface design patterns are defined as reusable elements which provide solutions to common problems. For instance, we expect that an arrow in the top left corner of a window which points to the left, when clicked upon, will take the user to a previous screen. Furthermore, we expect that buttons in the top left corner of a window will in some way relate to navigation. &lt;br /&gt;
&lt;br /&gt;
An anti-pattern is a design pattern which breaks standard design convention, creating more problems than it solves. An example of an anti-pattern would be if in place of the &#039;back&#039; button on your web browser, there would be a &#039;view history&#039; button.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module would be to analyze an interface to see if any anti-patterns exist, identify where they are in the interface, and then suggest alternatives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
*Formal interface description&lt;br /&gt;
*Tasks which can be performed within the interface&lt;br /&gt;
*A library of standard design patterns&lt;br /&gt;
*Outputs from the &#039;Affordances&#039; module&lt;br /&gt;
*Uncommon / Custom additional pattern library (optional)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
*Identification of interface elements whose placement or function are contrary to the pattern library&lt;br /&gt;
*Recommendations for alternative functionality or placement of such elements.&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The interface design process is critical to the creation of a quality end product. The process of creating an interface can also be used as a model for analyzing a finished one.&lt;br /&gt;
&lt;br /&gt;
There are a number of different philosophies on how to best design software and in turn, interface. Currently, agile development using an incremental process such as [http://en.wikipedia.org/wiki/SCRUM Scrum] has become a well known and generally practiced procedure.&lt;br /&gt;
&lt;br /&gt;
The steps to create interfaces varies significantly from text to text, although the [http://http://cfg.cit.cornell.edu/design/process.html Common Front Group at Cornell] has succinctly been able to reduce this variety into six simple steps:&lt;br /&gt;
	&lt;br /&gt;
*Requirement Sketching&lt;br /&gt;
*Conceptual Design&lt;br /&gt;
*Logical Design&lt;br /&gt;
*Physical Design&lt;br /&gt;
*Construction&lt;br /&gt;
*Usability Testing&lt;br /&gt;
&lt;br /&gt;
This can be broken down further into just information architecture design followed by physical design and testing.&lt;br /&gt;
&lt;br /&gt;
As far as this proposal is concerned in the context of interface design, the goal of our proposal is to improve what is involved with later end of the middle two portions: logical and physical design. Prior to feeding an interface to the system we are proposing, designers should have already created a baseline model for review that should exhibit the majority, if not all, of the functionality listed in the interface requirements. Once this initial interface has been created, our system will aid in rapidly iterating through the physical design process. The ultimate end products are then subject to human usability testing.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3216</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3216"/>
		<updated>2009-04-28T16:06:49Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependence&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
Using a limited set of interface evaluation modules for analysis, we demonstrate, in narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its modules taken in isolation.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
*Dependencies: Having a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3215</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3215"/>
		<updated>2009-04-28T16:06:29Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
A weighted framework for the unification of established heuristic usability guidelines and accepted cognitive principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: Three groups of experts anecdotally apply cognitive principles, heuristic usability principles, and a combination of the two.&lt;br /&gt;
* A &amp;quot;cognition expert,&amp;quot; given a constrained, limited-functionality interface, develops an independent evaluative value for each interface element based on accepted cognitive principles.&lt;br /&gt;
* A &amp;quot;usability expert&amp;quot; develops an independent evaluative value for each interface element based on accepted heuristic guidelines.&lt;br /&gt;
* A third expert applies several unified cognitive analogues from a matrix of weighted cognitive and &lt;br /&gt;
* User testing demonstrates the assumed efficacy and applicability of the matricized analogues versus independent application of analogued principles.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: A set of established cognitive principles, selected with an eye toward heuristic analogues.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependence&#039;&#039;&#039;: A set of established heuristic design guidelines, selected with an eye toward cognitive analogues.&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
Using a limited set of interface evaluation modules for analysis, we demonstrate, in narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Create a framework that provides better interface evaluations than currently existing techniques, and a module weighting system that provides better evaluations than any of its individual modules.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
*Dependencies: Having a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3213</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3213"/>
		<updated>2009-04-28T16:04:35Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Reconciling Usability Heuristics with Cognitive Theory ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
== Evaluation for Recommendation and Incremental Improvement ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski|E J Kalafarski]] 14:56, 28 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
Using a limited set of interface evaluation modules for analysis, we demonstrate, in narrow and controlled manner, the proposed efficiency and accuracy of a method of aggregating individual interface suggestions based on accepted CPM principles (e.g. Fitts&#039; Law and Affordance) and applying them to the incremental improvement of the interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Demonstration&#039;&#039;&#039;: A narrow but comprehensive study to demonstrate the efficacy and efficiency of automated aggregation of interface evaluation versus independent, analysis.  Shows not total replacement, but a gain in speed of evaluation comparable to the loss of control and feature.&lt;br /&gt;
* Given a carefully-constrained interface, perhaps with as few as two buttons and a minimalist feature set, expert interface designer given the individual results of several basic evaluation modules makes recommendations and suggestions to the design of the interface.&lt;br /&gt;
* Aggregation meta-module conducts similar survey of module outputs, outputting recommendations and suggestions for improvement of given interface.&lt;br /&gt;
* Separate independent body of experts then implements the two sets of suggestions and performs user study on the resultant interfaces, analyzing usability change and comparing it to the time and resources committed by the evaluation expert and the aggregation module, respectively.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of Fitts&#039; Law as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dependency&#039;&#039;&#039;: Module for the analysis of a Affordance as it applies to the individual elements of a given interface.&lt;br /&gt;
&lt;br /&gt;
== Evaluation Metrics ==&lt;br /&gt;
Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
:*Architecture Outputs&lt;br /&gt;
::* Time (time to complete task)&lt;br /&gt;
:::# Performs as well or better than CPM-GOMS, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
:::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::* Cognitive Load&lt;br /&gt;
:::# Predicts cognitive load during tasks, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# CPM-GOMS Module with cognitive load extension&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::* Frustration&lt;br /&gt;
:::# Accurately predicts users&#039; frustration levels, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Facial Gesture Recognition Module&lt;br /&gt;
::::# Galvanic Skin Response Module&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
::* Aesthetic Appeal&lt;br /&gt;
:::# Analyzes if the interface is aesthetically unpleasing, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Aesthetics Module &lt;br /&gt;
::* Simplicity&lt;br /&gt;
:::# Analyzes how simple the interface is, demonstrated with user tasks&lt;br /&gt;
:::* Dependencies&lt;br /&gt;
::::# Interface Efficiency Module&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Create a framework that provides better interface evaluations than any of its individual modules.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
*Dependencies: Having a good set of modules to plug into the framework.&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3199</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3199"/>
		<updated>2009-04-28T13:35:25Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Create a framework that provides better interface evaluations than any of its individual modules.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3198</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3198"/>
		<updated>2009-04-28T13:35:04Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Created a framework that provides better interface evaluations than any of its individual modules.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3197</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3197"/>
		<updated>2009-04-28T13:12:51Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Contributions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Created a framework that provides better interface evaluations than any of its individual components.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3196</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3196"/>
		<updated>2009-04-28T13:11:05Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Parallel Framework for Evaluation Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# specification (language?) of how to define an interface evaluation module and how to integrate it into a larger system.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
==Design guidelines==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Recording User-Interaction Primitives ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop system for logging low-level user-interactions within existing applications.  By low-level interactions, we refer to those interactions for which a quantitative, predictive, GOMS-like model may be generated.&lt;br /&gt;
* At the same scale of granularity, integrate explicit interface-interactions with multimodal-sensing data.  i.e. pupil-tracking, muscle-activity monitoring, auditory recognition, EEG data, and facial and posture-focused video. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Extensible, multimodal, HCI framework for recording rich interaction-history data in existing applications.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Provide rich history data to serve as the basis for novel quantitative interface evaluation.&lt;br /&gt;
* Aid expert HCI and visual designers in traditional design processes.&lt;br /&gt;
* Provide data for automated machine-learning strategies applied to interaction.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* The utility of interaction histories with respect to assessing interface design has been demonstrated in &amp;lt;ref name = bob&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008) &amp;lt;/ref&amp;gt;.&lt;br /&gt;
* In addition, data management histories have been shown effective in the visualization community in &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-MED.pdf Callahan-2006-MED] &amp;lt;/ref&amp;gt;  &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Callahan-2006-VVM.pdf Callahan-2006-VVM] &amp;lt;/ref&amp;gt; &amp;lt;ref&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Bavoil-2005-VEI.pdf Bavoil-2005-VEI]&amp;lt;/ref&amp;gt;, providing visualizations by analogy &amp;lt;ref&amp;gt; [http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&amp;amp;arnumber=4376187&amp;amp;isnumber=4376125 Querying and Creating Visualizations by Analogy] &amp;lt;/ref&amp;gt; and offering automated suggestions &amp;lt;ref&amp;gt; [http://www.cs.utah.edu/~juliana/pub/tvcg-recommendation2008.pdf VisComplete: Automating Suggestions from Visualization Pipelines] &amp;lt;/ref&amp;gt;, which we expect to generalize to user interaction history.&lt;br /&gt;
&lt;br /&gt;
== Semantic-level Interaction Chunking ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Aims&#039;&#039;&#039;&lt;br /&gt;
* Develop techniques for &#039;&#039;chunking&#039;&#039; low-level interaction primitives into &#039;&#039;semantic-level&#039;&#039; interactions, given an application&#039;s functionality and data-context.  (And de-chunking? Invertible mapping needed?)&lt;br /&gt;
* Perform user-study evaluation to validate chunking methods.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
* Design and user-study evaluation of semantic chunking techniques.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Significance&#039;&#039;&#039;&lt;br /&gt;
* Low-level &amp;lt;--&amp;gt; semantic-level mapping allows for cognitive-modeling to be applied at a functionality level, where low-level interaction techniques can be swapped out.  This will allow our interface assessment system to make feasible suggestions for more optimal interface design.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Background and Related Work&#039;&#039;&#039;&lt;br /&gt;
* Chunking interactions has been studied in the HCI community as in &amp;lt;ref name = bob/&amp;gt; [http://www.cs.brown.edu/people/trevor/Papers/Heer-2008-GraphicalHistories.pdf Graphical Histories for Visualization: Supporting Analysis, Communication, and Evaluation] (InfoVis 2008).&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==== Contributions ====&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Created a framework that provides better interface evaluations than any of its individual components.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Demonstrate by running user studies and comparing this performance to expected performance, as given by the following interface evaluation methods:&lt;br /&gt;
*# Traditional, manual interface evaluation &lt;br /&gt;
*#* As a baseline.&lt;br /&gt;
*# Using our system with a single module&lt;br /&gt;
*#* &amp;quot;Are any of our individual modules better than currently existing methods of interface evaluation?&amp;quot;.&lt;br /&gt;
*# Using our system with multiple modules, but have aggregator give a fixed, equal weighting to each module&lt;br /&gt;
*#* As a baseline for our aggregator: want to show that the value of adding the dynamic weighting.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, but have each module get a single weighting for all dimensions (time, fatigue, etc.)&lt;br /&gt;
*#* For validating the use of a dynamic weighting system.&lt;br /&gt;
*# Using our system with multiple modules, and allow the aggregator to adjust weightings for each module, and allow the module to give different weightings to every dimension of the module (time, fatigue, etc.) &lt;br /&gt;
*#* For validating the use of weighting across multiple utility dimensions.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* A set of interaction functions. These specify all of the information that the application wants to give users or get from them. It is not tied to a specific interface. For example, the fact that an applications shows videos would be included here. Whether it displays them embedded, in a pop-up window or fullscreen would not.&lt;br /&gt;
* A mapping of interaction functions to interface elements (e.g., buttons, windows, dials,...). Lots of optional information describing visual properties, associated text, physical interactions (e.g., turning the dial clockwise increases the input value) and timing.&lt;br /&gt;
* Functional user traces - sequences of interaction functions that represent typical user interactions with the application. Could include a goal hierarchy, in which case the function sequence is at the bottom of the hierarchy.&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
A &amp;quot;meta-module&amp;quot; called the aggregator will be responsible for assembling and formatting the output of all other modules into a structure that is both extensible and immediately usable, by both an automated designer or a human designer.&lt;br /&gt;
&lt;br /&gt;
===Requirements===&lt;br /&gt;
The aggregator&#039;s functionality, then, is defined by its &#039;&#039;&#039;inputs&#039;&#039;&#039;, the outputs of the other modules, and the desired &#039;&#039;&#039;output&#039;&#039;&#039; of the system as a whole, per its position in the architecture.  Its purpose is largely formatting and reconciliation of the products of the multitudinous (and extensible) modules.  The output of the aggregator must meet several requirements: first, to generate a set of human-readable suggestions for the improvement of the given interface; second, to generate a machine-readable, but also analyzable, evaluation of the various characteristics of the interface and accompanying user traces.&lt;br /&gt;
&lt;br /&gt;
From these specifications, it is logical to assume that a common language or format will be required for the output of individual modules.  We propose an XML-based file format, allowing: (1) a section for the standardized identification of problem areas, applicable rules, and proposed improvements, generalized by the individual module and mapped to a single element, or group of elements, in the original interface specification; (2) a section for specification of generalizable &amp;quot;utility&amp;quot; functions, allowing a module to specify how much a measurable quantity of utility is positively or negatively affected by properties of the input interface; (3) new, user-definable sections for evaluations of the given interface not covered by the first two sections.  The first two sections are capable of conveying the vast majority of module outputs predicted at this time, but the XML can extensibly allow modules to pass on whatever information may become prominent in the future.&lt;br /&gt;
&lt;br /&gt;
===Specification===&lt;br /&gt;
 &amp;lt;module id=&amp;quot;Fitts-Law&amp;quot;&amp;gt;&lt;br /&gt;
 	&amp;lt;interface-elements&amp;gt;&lt;br /&gt;
 		&amp;lt;element&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;submit button&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;problem&amp;gt;&lt;br /&gt;
 				&amp;lt;desc&amp;gt;size&amp;lt;/desc&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;width *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;suggestion&amp;gt;height *= 2&amp;lt;/suggestion&amp;gt;&lt;br /&gt;
 				&amp;lt;human-suggestion&amp;gt;Increase size relative to other elements&amp;lt;/human-suggestion&amp;gt;&lt;br /&gt;
 			&amp;lt;/problem&amp;gt;&lt;br /&gt;
 		&amp;lt;/element&amp;gt;&lt;br /&gt;
 	&amp;lt;/interface-elements&amp;gt;&lt;br /&gt;
 	&amp;lt;utility&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;time&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0:15:35&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;frustration&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;pulling hair out&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 		&amp;lt;dimension&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;efficiency&amp;lt;/desc&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;13.2s/KPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 			&amp;lt;value&amp;gt;0.56m/CPM task&amp;lt;/value&amp;gt;&lt;br /&gt;
 		&amp;lt;/dimension&amp;gt;&lt;br /&gt;
 	&amp;lt;/utility&amp;gt;&lt;br /&gt;
 	&amp;lt;tasks&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;complete form&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;lookup SSN&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 		&amp;lt;task&amp;gt;&lt;br /&gt;
 			&amp;lt;desc&amp;gt;format phone number&amp;lt;/desc&amp;gt;&lt;br /&gt;
 		&amp;lt;/task&amp;gt;&lt;br /&gt;
 	&amp;lt;/tasks&amp;gt;&lt;br /&gt;
 &amp;lt;/module&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Logic===&lt;br /&gt;
This file provided by each module is then the input for the aggregator.  The aggregator&#039;s most straightforward function is the compilation of the &amp;quot;problem areas,&amp;quot; assembling them and noting problem areas and suggestions that are recommended by more than one module, and weighting them accordingly in its final report.  These weightings can begin in an equal state, but the aggregator should be capable of learning iteratively which modules&#039; results are most relevant to the user and update weightings accordingly.  This may need to be accomplished with manual tuning, or a machine-learning algorithm capable of determining which modules most often agree with others.&lt;br /&gt;
&lt;br /&gt;
Secondly, the aggregator compiles the utility functions provided by the module specs.  This, again, is a summation of similarly-described values from the various modules.&lt;br /&gt;
&lt;br /&gt;
When confronted with user-defined sections of the XML spec, the aggregator is primarily responsible for compiling them and sending them along to the output of the machine.  Even if the aggregator does not recognize a section or property of the evaluative spec, if it sees the property reported by more than one module it should be capable of aggregating these intelligently.  In future versions of the spec, it should be possible for a module to provide instructions for the aggregator on how to handle unrecognized sections of the XML.&lt;br /&gt;
&lt;br /&gt;
From these compilations, then, the aggregator should be capable of outputting both aggregated human-readable suggestions on interface improvements for a human designer, as well as a comprehensive evaluation of the interface&#039;s effectiveness at the given task traces.  Again, this is dependent on the specification of the system as a whole, but is likely to include measures and comparisons, graphings of task versus utility, and quantitative measures of an element&#039;s effectiveness.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
Schneiderman&#039;s Eight Golden Rules and Jakob Nielsen&#039;s Ten Heuristics are perhaps the most famous and well-regarded heuristic design guidelines to emerge over the last twenty years.  Although the explicit theoretical basis for such heuristics is controversial and not well-explored, the empirical success of these guidelines is established and accepted.  This module will parse out up to three or four common (that is, intersecting) principles from these accepted guidelines and apply them to the input interface.&lt;br /&gt;
&lt;br /&gt;
As an example, we identify an analogous principle that appears in Schneiderman (&amp;quot;Reduce short-term memory load&amp;quot;)&amp;lt;ref&amp;gt;http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html&amp;lt;/ref&amp;gt; and Nielsen (&amp;quot;Recognition rather than recall/Minimize the user&#039;s memory load&amp;quot;)&amp;lt;ref&amp;gt;http://www.useit.com/papers/heuristic/heuristic_list.html&amp;lt;/ref&amp;gt;.  The input interface is then evaluated for the consideration of the principle, based on an explicit formal description of the interface, such as XAML or XUL.  The module attempts to determine how effectively the interface demonstrates the principle.  When analyzing an interface for several principles that may be conflicting or opposing in a given context, the module makes use of a hard-coded but iterative (and evolving) weighting of these principles, based on (1) how often they appear in the training set of accepted sets of guidelines, (2) how analogues a heuristic principle is to a cognitive principle in a parallel training set, and (3) how effective the principle&#039;s associated suggestion is found to be using a feedback mechanism.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
# A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
# A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&amp;lt;br/&amp;gt;&lt;br /&gt;
Standard XML-formatted file containing problem areas of the input interface, suggestions for each problem area based on principles that were found to have a strong application to a problem element and the problem itself, and a human-readable generated analysis of the element&#039;s affinity for the principle.  Quantitative outputs will not be possible based on heuristic guidelines, and the &amp;quot;utility&amp;quot; section of this module&#039;s output is likely to be blank.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Assignment 13&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
::* Scientific Study of Multi-tasking Workflow and the Impact of Interruptions&lt;br /&gt;
:::# We will undertake detailed studies to help understand the following questions:&lt;br /&gt;
::::# How does the size of a user&#039;s working set impact interruption resumption time?&lt;br /&gt;
::::# How does the size of a user&#039;s working set, when used for rapid multi-tasking, impact performance metrics?&lt;br /&gt;
::::# How does a user interface which supports multiple simultaneous working sets benefit interruption resumption time?&lt;br /&gt;
:::* No Dependencies&lt;br /&gt;
::* Meta-work Assistance Tool&lt;br /&gt;
:::# We will perform a series of ecologically-valid studies to compare user performance between a state of the art task management system (control group) and our meta-work assistance tool (experimental group)&lt;br /&gt;
:::* Dependent on core study completion, as some of the specific design decisions will be driven by the results of this study.  However, it is worth pointing out that this separate contribution can be researched in parallel to the core study.&lt;br /&gt;
::* Baseline Comparison Between Module-based Model of HCI and Core Multi-tasking Study&lt;br /&gt;
:::# We will compare the results of the above-mentioned study in multi-tasking against results predicted by the module-based model of HCI in this proposal; this will give us an important baseline comparison, particularly given that multi-tasking and interruption involve higher brain functioning and are therefore likely difficult to predict&lt;br /&gt;
:::* Dependent on core study completion, as well as most of the rest of the proposal being completed to the point of being testable&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Text for Assignment 12:&lt;br /&gt;
&lt;br /&gt;
Add text here about how this can be used to evaluate automatic framework&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Models&lt;br /&gt;
** Beddeley&#039;s Model of Working Memory&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Baddeley%27s_model_of_working_memory&amp;lt;/ref&amp;gt;&lt;br /&gt;
*** Episodic Buffer&lt;br /&gt;
** George Miller&#039;s &amp;quot;The magic number 7 plus or minus 2&amp;quot;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/The_Magical_Number_Seven,_Plus_or_Minus_Two&amp;lt;/ref&amp;gt;&lt;br /&gt;
** The magic number 4&amp;lt;ref&amp;gt;Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24, 87-185.&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Chunking&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Chunking_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Priming&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Priming_(psychology)&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Subitizing and Counting&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Subitizing&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Visual Stimuli&lt;br /&gt;
* Audio Stimuli&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Remembered percepts&lt;br /&gt;
* Half-Life of percepts&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Learning&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Learning#Mathematical_models_of_learning&amp;lt;/ref&amp;gt;&lt;br /&gt;
** Logan&#039;s instance theory of automatization&amp;lt;ref&amp;gt;http://74.125.95.132/search?q=cache:IZ-Zccsu3SEJ:psych.wisc.edu/ugstudies/psych733/logan_1988.pdf+logan+isntance+teory&amp;amp;cd=1&amp;amp;hl=en&amp;amp;ct=clnk&amp;amp;gl=us&amp;amp;client=firefox-a&amp;lt;/ref&amp;gt;&lt;br /&gt;
* Fluency&lt;br /&gt;
** As meta-cognitive information &amp;lt;ref&amp;gt;http://www.sciencedirect.com/science?_ob=ArticleURL&amp;amp;_udi=B6VH9-4SM7PFK-4&amp;amp;_user=10&amp;amp;_rdoc=1&amp;amp;_fmt=&amp;amp;_orig=search&amp;amp;_sort=d&amp;amp;view=c&amp;amp;_acct=C000050221&amp;amp;_version=1&amp;amp;_urlVersion=0&amp;amp;_userid=10&amp;amp;md5=10cd279fa80958981fcc3c06684c09af&amp;lt;/ref&amp;gt;&lt;br /&gt;
** As a cognitive &#039;&#039;heuristic&#039;&#039;&amp;lt;ref&amp;gt;http://en.wikipedia.org/wiki/Fluency_heuristic&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
* Interface&lt;br /&gt;
* User goals&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
* Learning curve&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3142</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3142"/>
		<updated>2009-04-24T13:11:09Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Module Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3141</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3141"/>
		<updated>2009-04-24T13:10:55Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Module Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
***Either a single recommendation or a set of different recommendations can be output&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3140</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3140"/>
		<updated>2009-04-24T13:09:42Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Module Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendation(s) for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3139</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3139"/>
		<updated>2009-04-24T13:09:09Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Aggregator Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action (in progress) ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Formalism eases translation of theory into scripting language&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Assumes symbolic representation&lt;br /&gt;
&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
**Advantages&lt;br /&gt;
***Emphasis on bodily and environmental constraints&lt;br /&gt;
**Disadvantages&lt;br /&gt;
***Lack of formalism hinders translation of theory into scripting language&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendations for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
Outputs for the aggregator are the same as the outputs for each module. The difference is that the aggregator will consider all the module outputs, and arrive at a merged output based on the past performance of each of the modules.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3135</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3135"/>
		<updated>2009-04-24T12:59:28Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Module Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
== Perception and Action ==&lt;br /&gt;
&lt;br /&gt;
*Information Processing Approach&lt;br /&gt;
*Ecological (Gibsonian) Approach&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Inputs (Incomplete) ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendations for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Inputs ====&lt;br /&gt;
The aggregator receives as input the outputs of all the modules.&lt;br /&gt;
&lt;br /&gt;
==== Aggregator Outputs ====&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3133</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3133"/>
		<updated>2009-04-24T12:57:30Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Parallel Framework for Evaluation Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==== Module Outputs ====&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendations for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3132</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3132"/>
		<updated>2009-04-24T12:56:51Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Parallel Framework for Evaluation Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;[http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf Borchers, Jan O.  &amp;quot;A Pattern Approach to Interaction Design.&amp;quot;  2000.]&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;[http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf Ivory, M and Hearst, M.  &amp;quot;The State of the Art in Automated Usability Evaluation of User Interfaces.&amp;quot; ACM Computing Surveys (CSUR), 2001.]&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;[http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774 Czerwinski, Horvitz, and White. &amp;quot;A Diary Study of Task Switching and Interruptions.&amp;quot;  Proceedings of the SIGCHI conference on Human factors in computing systems, 2004.]&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* A set of utility dimensions {d1, d2, ...} are defined in the framework. These could be {d1=&amp;quot;time&amp;quot;, d2=&amp;quot;fatigue&amp;quot;, ...}&lt;br /&gt;
* Every module ouputs at least one of the following:&lt;br /&gt;
** An evaluation of the interface&lt;br /&gt;
*** This can be on any or all of the utility dimensions, e.g. evaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
*** This can alternately be an overall evaluation, ignoring dimensions, e.g. evaluation={score}&lt;br /&gt;
**** In this case, the aggregator will treat this as the module giving the same score to all dimensions. Which dimension this evaluator actually predicts well on can be learned by the aggregator over time.&lt;br /&gt;
** Recommendations for improving the interface&lt;br /&gt;
***This can be a textual description of what changes the designer should make&lt;br /&gt;
***This can alternately be a transformation that can automatically be applied to the interface language (without designer intervention)&lt;br /&gt;
***In addition to the textual or transformational description of the recommendation, a &amp;quot;change in evaluation&amp;quot; is output to describe how specifically the value will improve the interface&lt;br /&gt;
****Recommendation = {description=&amp;quot;make this change&amp;quot;, Δevaluation={d1=score1, d2=score2, ...}&lt;br /&gt;
****Like before, this Δevaluation can cover any number of dimensions, or it can be generic.&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section is necessarily defined by the output of the individual modules (which I already expect to be of varied and arbitrary structure) and the desired output of the machine as a whole.  It will likely need to be revised heavily after other modules and the &amp;quot;Parellel Framework&amp;quot; section are defined.&#039;&#039; [[User:E J Kalafarski|E J Kalafarski]] 12:34, 24 April 2009 (UTC)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
&amp;lt;references/&amp;gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3126</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3126"/>
		<updated>2009-04-24T10:00:46Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Introduction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]], [[User: Trevor O&#039;Brien | Trevor]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Propose:&#039;&#039;&#039; The design, application and evaluation of a novel, cognition-based, computational framework for assessing interface design and providing automated suggestions to optimize usability.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Evaluation Methodology:&#039;&#039;&#039;  Our techniques will be evaluated quantitatively through a series of user-study trials, as well as qualitatively by a team of expert interface designers. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions and Significance:&#039;&#039;&#039;  We expect this work to make the following contributions:  &lt;br /&gt;
# design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces.  &lt;br /&gt;
# design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
# an extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring.  &lt;br /&gt;
# (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability)&lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
== Cognitive Models ==&lt;br /&gt;
&lt;br /&gt;
I plan to port over most of the background on cognitive models of HCI from the old proposal&lt;br /&gt;
&lt;br /&gt;
Additions will comprise of:&lt;br /&gt;
*CPM-GOMS as a bridge from GOMS architecture to the promising procedural optimization of the Model Human Processor&lt;br /&gt;
**Context of CPM development, discuss its relation to original GOMS and KLM&lt;br /&gt;
***Establish the tasks which were relevant for optimization when CPM was developed and note that its obsolescence may have been unavoidable&lt;br /&gt;
**Focus on CPM as the first step in transitioning from descriptive data, provided by mounting efforts in the cognitive sciences realm to discover the nature of task processing and accomplishment, to prescriptive algorithms which can predict an interface’s efficiency and suggest improvements&lt;br /&gt;
**CPM’s purpose as an abstraction of cognitive processing – a symbolic representation not designed for accuracy but precision&lt;br /&gt;
**CPM’s successful trials, e.g. Ernestine&lt;br /&gt;
***Implications of this project include CPM’s ability to accurately estimate processing at a psychomotor level&lt;br /&gt;
***Project does suggest limitations, however, when one attempts to examine more complex tasks which involve deeper and more numerous cognitive processes&lt;br /&gt;
*ACT-R as an example of a progressive cognitive modeling tool&lt;br /&gt;
**A tool clearly built by and for cognitive scientists, and as a result presents a much more accurate view of human processing – helpful for our research&lt;br /&gt;
**Built-in automation, which now seems to be a standard feature of cognitive modeling tools&lt;br /&gt;
**Still an abstraction of cognitive processing, but makes adaptation to cutting-edge cognitive research findings an integral aspect of its modular structure&lt;br /&gt;
**Expand on its focus on multi-tasking, taking what was a huge advance between GOMS and its CPM variation and bringing the simulation several steps closer to approximating the nature of cognition in regards to HCI&lt;br /&gt;
**Far more accessible both for researchers and the lay user/designer in its portability to LISP, pre-construction of modules representing cognitive capacities and underlying algorithms modeling paths of cognitive processing&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Design guidelines (in progress…)==&lt;br /&gt;
A multitude of rule sets exist for the design of not only interface, but architecture, city planning, and software development.  They can range in scale from one primary rule to as many Christopher Alexander&#039;s 253 rules for urban environments,&amp;lt;ref&amp;gt;http://hci.rwth-aachen.de/materials/publications/borchers2000a.pdf&amp;lt;/ref&amp;gt; which he introduced with the concept design patterns in the 1970s.  Study has likewise been conducted on the use of these rules:&amp;lt;ref&amp;gt;http://stl.cs.queensu.ca/~graham/cisc836/lectures/readings/tetzlaff-guidelines.pdf&amp;lt;/ref&amp;gt; guidelines are often only partially understood, indistinct to the developer, and &amp;quot;fraught&amp;quot; with potential usability problems given a real-world situation.&lt;br /&gt;
&lt;br /&gt;
===Application to AUE===&lt;br /&gt;
And yet, the vast majority of guideline sets, including the most popular rulesets, have been arrived at heuristically.  The most successful, such as Raskin&#039;s and Schneiderman&#039;s, have been forged from years of observation instead of empirical study and experimentation.  The problem is similar to the problem of circular logic faced by automated usability evaluations: an automated system is limited in the suggestions it can offer to a set of preprogrammed guidelines which have often not been subjected to rigorous experimentation.&amp;lt;ref&amp;gt;http://www.eecs.berkeley.edu/Pubs/TechRpts/2000/CSD-00-1105.pdf&amp;lt;/ref&amp;gt;  In the vast majority of existing studies, emphasis has traditionally been placed on either the development of guidelines or the application of existing guidelines to automated evaluation.  A mutually-reinforcing development of both simultaneously has not been attempted.&lt;br /&gt;
&lt;br /&gt;
Overlap between rulesets is inevitable and unavoidable.  For our purposes of evaluating existing rulesets efficiently, without extracting and analyzing each rule individually, it may be desirable to identify the the overarching &#039;&#039;principles&#039;&#039; or &#039;&#039;philosophy&#039;&#039; (max. 2 or 3) for a given ruleset and determining their quantitative relevance to problems of cognition.&lt;br /&gt;
&lt;br /&gt;
===Popular and seminal examples===&lt;br /&gt;
Schneiderman&#039;s [http://faculty.washington.edu/jtenenbg/courses/360/f04/sessions/schneidermanGoldenRules.html Eight Golden Rules] date to 1987 and are arguably the most-cited.  They are heuristic, but can be somewhat classified by cognitive objective: at least two rules apply primarily to &#039;&#039;repeated use&#039;&#039;, versus &#039;&#039;discoverability&#039;&#039;.  Up to five of Schneiderman&#039;s rules emphasize &#039;&#039;predictability&#039;&#039; in the outcomes of operations and &#039;&#039;increased feedback and control&#039;&#039; in the agency of the user.  His final rule, paradoxically, removes control from the user by suggesting a reduced short-term memory load, which we can arguably classify as &#039;&#039;simplicity&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Raskin&#039;s [http://www.mprove.de/script/02/raskin/designrules.html Design Rules] are classified into five principles by the author, augmented by definitions and supporting rules.  While one principle is primarily aesthetic (a design problem arguably out of the bounds of this proposal) and one is a basic endorsement of testing, the remaining three begin to reflect philosophies similar to Schneiderman&#039;s: reliability or &#039;&#039;predictability&#039;&#039;, &#039;&#039;simplicity&#039;&#039; or &#039;&#039;efficiency&#039;&#039; (which we can construe as two sides of the same coin), and finally he introduces a concept of &#039;&#039;uninterruptibility&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
Maeda&#039;s [http://lawsofsimplicity.com/?cat=5&amp;amp;order=ASC Laws of Simplicity] are fewer, and ostensibly emphasize &#039;&#039;simplicity&#039;&#039; exclusively, although elements of &#039;&#039;use&#039;&#039; as related by Schneiderman&#039;s rules and &#039;&#039;efficiency&#039;&#039; as defined by Raskin may be facets of this simplicity.  Google&#039;s corporate mission statement presents [http://www.google.com/corporate/ux.html Ten Principles], only half of which can be considered true interface guidelines.  &#039;&#039;Efficiency&#039;&#039; and &#039;&#039;simplicity&#039;&#039; are cited explicitly, aesthetics are once again noted as crucial, and working within a user&#039;s trust is another application of &#039;&#039;predictability&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
===Elements and goals of a guideline set===&lt;br /&gt;
Myriad rulesets exist, but variation becomes scarce—it indeed seems possible to parse these common rulesets into overarching principles that can be converted to or associated with quantifiable cognitive properties.  For example, it is likely &#039;&#039;simplicity&#039;&#039; has an analogue in the short-term memory retention or visual retention of the user, vis a vis the rule of [http://books.google.com/books?hl=en&amp;amp;lr=&amp;amp;id=j5q0VvOGExYC&amp;amp;oi=fnd&amp;amp;pg=PA357&amp;amp;dq=seven+plus+or+minus+two&amp;amp;ots=prI3PKJBar&amp;amp;sig=vOZnqpnkXKGYWxK6_XlA4I_CRyI Seven, Plus or Minus Two].  &#039;&#039;Predictability&#039;&#039; likewise may have an analogue in Activity Theory, in regards to a user&#039;s perceptual expectations for a given action; &#039;&#039;uninterruptibility&#039;&#039; has implications in cognitive task-switching;&amp;lt;ref&amp;gt;http://portal.acm.org/citation.cfm?id=985692.985715&amp;amp;coll=Portal&amp;amp;dl=ACM&amp;amp;CFID=21136843&amp;amp;CFTOKEN=23841774&amp;lt;/ref&amp;gt; and so forth.&lt;br /&gt;
&lt;br /&gt;
Within the scope of this proposal, we aim to reduce and refine these philosophies found in seminal rulesets and identify their logical cognitive analogues.  By assigning a quantifiable taxonomy to these principles, we will be able to rank and weight them with regard to their real-world applicability, developing a set of &amp;quot;meta-guidelines&amp;quot; and rules for applying them to a given interface in an automated manner.  Combined with cognitive models and multi-modal HCI analysis, we seek to develop, in parallel with these guidelines, the interface evaluation system responsible for their application.&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Specific Aims&#039;&#039;&#039;&lt;br /&gt;
* Incorporate interaction history mechanisms into a set of existing applications.&lt;br /&gt;
* Perform user-study evaluation of history-collection techniques.&lt;br /&gt;
* Distill a set of cognitive principles/models, and evaluate empirically?&lt;br /&gt;
* Build/buy sensing system to include pupil-tracking, muscle-activity monitoring, auditory recognition.&lt;br /&gt;
* Design techniques for manual/semi-automated/automated construction of &amp;lt;insert favorite cognitive model here&amp;gt; from interaction histories and sensing data.&lt;br /&gt;
* Design system for posterior analysis of interaction history w.r.t. &amp;lt;insert favorite cognitive model here&amp;gt;, evaluating critical path &amp;lt;or equivalent&amp;gt; trajectories.&lt;br /&gt;
* Design cognition-based techniques for detecting bottlenecks in critical paths, and offering optimized alternatives. &lt;br /&gt;
* Perform quantitative user-study evaluations, collect qualitative feedback from expert interface designers.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Contributions&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design-space analysis and quantitative evaluation of cognition-based techniques for assessing user interfaces. &lt;br /&gt;
* Design and quantitative evaluation of techniques for suggesting optimized interface-design changes. &lt;br /&gt;
* An extensible, multimodal software architecture for capturing user traces integrated with pupil-tracking data, auditory recognition, and muscle-activity monitoring. &lt;br /&gt;
* (there may be more here, like testing different cognitive models, generating a markup language to represent interfaces, maybe even a unique metric space for interface usability) &lt;br /&gt;
&lt;br /&gt;
--&lt;br /&gt;
&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
I’m hoping to have some input on this section, because it seems to be the crux of the “black box” into which we take the inputs of interface description, user traces, etc. and get our outputs (time, recommendations, etc.).  I know at least a few people have pretty strong thoughts on the matter and we ought to discuss the final structure.&lt;br /&gt;
&lt;br /&gt;
That said – my proposal for the module:&lt;br /&gt;
*In my opinion the concept of the Model Human Processor (at least as applied in CPM) is outdated – it’s too economic/overly parsimonious in its conception of human activity.  I think we need to create a structure which accounts for more realistic conditions of HCI including multitasking, aspects of distributed cognition (and other relevant uses of tools – as far as I can tell CPM doesn’t take into account any sort of productivity aids), executive control processes of attention, etc.  ACT-R appears to take steps towards this but we would probably need to look at their algorithms to know for sure.&lt;br /&gt;
*Critical paths will continue to play an important role – we should in fact emphasize that part of this tool’s purpose will be a description not only of ways in which the interface should be modified to best fit a critical path, but also ways in which the user’s ought to be instructed in their use of the path.  This feedback mechanism could be bidirectional – if the model’s predictions of the user’s goals are incorrect, the critical path determined will also be incorrect and the interface inherently suboptimal.  The user could be prompted with a tooltip explaining in brief why and how the interface has changed, along with options to revert, select other configurations (euphemized by goals), and to view a short video detailing how to properly use the interface.&lt;br /&gt;
*Call me crazy but, if we assume designers will be willing to code a model of their interfaces into our ACT-R-esque language, could we allow that model to be fairly transparent to the user, who could use a gui to input their goals to find an analogue in the program which would subsequently rearrange its interface to fit the user’s needs?  Even if not useful to the users, such dynamic modeling could really help designers (IMO)&lt;br /&gt;
*I think the model should do its best to accept models written for ACT-R and whatever other cognitive models there are out there – gives us the best chance of early adoption&lt;br /&gt;
*I would particularly appreciate input on the number/complexity/type of inputs we’ll be using, as well as the same qualities for the output.&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module provides an estimate of the required time to complete various tasks that have been decomposed into formalized sequences of interactions with interface elements, and will provide evaluations and recommendations for optimizing the time required to complete those tasks using the interface.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1. A formal description of the interface and its elements (e.g. buttons).&lt;br /&gt;
&lt;br /&gt;
2. A formal description of a particular task and the possible paths through a subset of interface elements that permit the user to accomplish that task.&lt;br /&gt;
&lt;br /&gt;
3. The physical distances between interface elements along those paths.&lt;br /&gt;
&lt;br /&gt;
4. The width of those elements along the most likely axes of motion.&lt;br /&gt;
&lt;br /&gt;
5. Device (e.g. mouse) characteristics including start/stop time and the inherent speed limitations of the device.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The module will then use the Shannon formulation of Fitt&#039;s Law to compute the average time needed to complete the task along those paths.&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on a measure of the extent to which the user perceives the relevant affordances of the interface when performing a number of specified tasks.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Formalized descriptions of...&lt;br /&gt;
&lt;br /&gt;
1. Interface elements&lt;br /&gt;
&lt;br /&gt;
2. Their associated actions&lt;br /&gt;
&lt;br /&gt;
3. The functions of those actions&lt;br /&gt;
&lt;br /&gt;
4. A particular task&lt;br /&gt;
&lt;br /&gt;
5. User traces for that task.  &lt;br /&gt;
&lt;br /&gt;
Inputs (1-4) are then used to generate a &amp;quot;user-independent&amp;quot; space of possible functions that the interface is capable of performing with respect to a given task -- what the interface &amp;quot;affords&amp;quot; the user.  From this set of possible interactions, our model will then determine the subset of optimal paths for performing a particular task.  The user trace (5) is then used to determine what functions actually were performed in the course of a given task of interest and this information is then compared to the optimal path data to determine the extent to which affordances of the interface are present but not perceived.  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Output&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The output of this module is a simple ratio of (affordances perceived) / [(relevant affordances present) * (time to complete task)] which provides a quantitative measure of the extent to which the interface is &amp;quot;natural&amp;quot; to use for a particular task.&lt;br /&gt;
&lt;br /&gt;
=== Workflow, Multi-tasking and Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
There are, at least, two levels at which users work ([http://portal.acm.org/citation.cfm?id=985692.985707&amp;amp;coll=portal&amp;amp;dl=ACM&amp;amp;CFID=20781736&amp;amp;CFTOKEN=83176621 Gonzales, et al., 2004]).  Users accomplish individual low-level tasks which are part of larger &#039;&#039;working spheres&#039;&#039;; for example, an office worker might send several emails, create several Post-It (TM) note reminders, and then edit a word document, each of these smaller tasks being part of a single larger working sphere of &amp;quot;adding a new section to the website.&amp;quot;  Thus, it is important to understand this larger workflow context - which often involves extensive levels of multi-tasking, as well as switching between a variety of computing devices and traditional tools, such as notebooks.  In this study it was found that the information workers surveyed typically switch individual tasks every 2 minutes and have many simultaneous working spheres which they switch between, on average every 12 minutes.  This frenzied pace of switching tasks and switching working spheres suggests that users will not be using a single application or device for a long period of time, and that affordances to support this characteristic pattern of information work are important.&lt;br /&gt;
&lt;br /&gt;
The purpose of this module is to integrate existing work on multi-tasking, interruption and higher-level workflow into a framework which can predict user recovery times from interruptions.  Specifically, the goals of this framework will be to:&lt;br /&gt;
&lt;br /&gt;
* Understand the role of the larger workflow context in user interfaces&lt;br /&gt;
* Understand the impact of interruptions on user workflow&lt;br /&gt;
* Understand how to design software which fits into the larger working spheres in which information work takes place&lt;br /&gt;
&lt;br /&gt;
It is important to point out that because workflow and multi-tasking rely heavily on higher-level brain functioning, it is unrealistic within the scope of this grant to propose a system which can predict user performance given a description of a set of arbitrary software programs.  Therefore, we believe this module will function much more in a qualitative role to provide context to the rest of the model.  Specifically, our findings related to interruption and multi-tasking will advance the basic research question of &amp;quot;how do you users react to interruptions when using working sets of varying sizes?&amp;quot;.  This core HCI contribution will help to inform the rest of the outputs of the model in a qualitative manner.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Inputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outputs&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
N/A&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
===Workflow, Multi-tasking, and Interruption===&lt;br /&gt;
&lt;br /&gt;
====I.  &#039;&#039;&#039;Goals&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
The goals of the preliminary work are to gain qualitative insight into how information workers practice metawork, and to determine whether people might be better-supported with software which facillitates metawork and interruptions.  Thus, the preliminary work should investigate, and demonstrate, the need and impact of the core goals of the project.&lt;br /&gt;
&lt;br /&gt;
====II.  &#039;&#039;&#039;Methodology&#039;&#039;&#039;====&lt;br /&gt;
&lt;br /&gt;
Seven information workers, ages 20-38 (5 male, 2 female), were interviewed to determine which methods they use to &amp;quot;stay organized&amp;quot;.  An initial list of metawork strategies was established from two pilot interviews, and then a final list was compiled.  Participants then responded to a series of 17 questions designed to gain insight into their metawork strategies and process.  In addition, verbal interviews were conducted to get additional open-ended feedback.&lt;br /&gt;
&lt;br /&gt;
====III.  &#039;&#039;&#039;Final Results&#039;&#039;&#039;====&lt;br /&gt;
A histogram of methods people use to &amp;quot;stay organized&amp;quot; in terms of tracking things they need to do (TODOs), appointments and meetings, etc. is shown in the figure below.&lt;br /&gt;
&lt;br /&gt;
[[Image:AcbGraph.jpg]]&lt;br /&gt;
&lt;br /&gt;
In addition to these methods, participants also used a number of other methods, including:&lt;br /&gt;
&lt;br /&gt;
* iCal&lt;br /&gt;
* Notes written in xterms&lt;br /&gt;
* &amp;quot;Inbox zero&amp;quot; method of email organization&lt;br /&gt;
* iGoogle Notepad (for tasks)&lt;br /&gt;
* Tag emails as &amp;quot;TODO&amp;quot;, &amp;quot;Important&amp;quot;, etc.&lt;br /&gt;
* Things (Organizer Software)&lt;br /&gt;
* Physical items placed to &amp;quot;remind me of things&amp;quot;&lt;br /&gt;
* Sometimes arranging windows on desk&lt;br /&gt;
* Keeping browser tabs open&lt;br /&gt;
* Bookmarking web pages&lt;br /&gt;
* Keep programs/files open scrolled to certain locations sometimes with things selected&lt;br /&gt;
&lt;br /&gt;
In addition, three participants said that when interrupted they &amp;quot;rarely&amp;quot; or &amp;quot;very rarely&amp;quot; were able to resume the task they were working on prior to the interruption.  Three of the participants said that they would not actively recommend their metawork strategies for other people, and two said that staying organized was &amp;quot;difficult&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
Four participants were neutral to the idea of new tools to help them stay organized and three said that they would like to have such a tool/tools.&lt;br /&gt;
&lt;br /&gt;
====IV.  &#039;&#039;&#039;Discussion&#039;&#039;&#039;====&lt;br /&gt;
These results qunatiatively support our hypothesis that there is no clearly dominant set of metawork strategies employed by information workers.  This highly fragemented landscape is surprising, even though most information workers work in a similar environment - at a desk, on the phone, in meetings - and with the same types of tools - computers, pens, paper, etc.  We believe that this suggests that there are complex tradeoffs between these methods and that no single method is sufficient.  We therefore believe that users will be better supported with a new set of software-based metawork tools.&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J&amp;diff=3087</id>
		<title>CS295J</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J&amp;diff=3087"/>
		<updated>2009-04-21T21:49:22Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Cognition, Human-Computer Interaction, and Visual Analysis&#039;&#039;&#039;: in this graduate seminar we will learn about models of human cognition and perception and explore potential implications of the models on how computers and humans can interact effectively when performing scientific analyses.  Participants will be responsible for reading assigned materials, taking turns guiding discussions of the readings, and preparing a final paper and presentation.  It is recommended that participants have some background in at least one of the areas of study. &lt;br /&gt;
&lt;br /&gt;
*Notes from Class&lt;br /&gt;
**[[/Class 1|Class 1]] 23 Jan 2009, [[/Theory of Visualization outline|Theory of Visualization outline]]&lt;br /&gt;
**[[/Class 2|Class 2]] 30 Jan 2009&lt;br /&gt;
**[[/Class 3|Class 3]] 6 Feb 2009&lt;br /&gt;
**[[/Class 4|Class 4]] 13 Feb 2009 (time changed: 10:30-12 and 1-2)&lt;br /&gt;
***[[/Posters from class 4|Posters from class 4]]&lt;br /&gt;
**[[/Class 5|Class 5]] 20 Feb 2009&lt;br /&gt;
**[[/Class 6|Class 6]] 27 Feb 2009, [[/Literature to read for class 6|Reading for class 6]], [[/Application Critiques|Application Critiques]], [[/Experiment results from class 6|Experiment results from class 6]]&lt;br /&gt;
**[[/Class 7|Class 7]] 6 Mar 2009, [[/Literature to read for class 7|Reading for class 7]], [[/Boards from class 7|Boards from class 7]] (preliminary results first sentence(s))&lt;br /&gt;
**[[/Class 8|Class 8]] 13 Mar 2009, [[/Literature to read for class 8|Reading for class 8]], [[/Proposal reviews from class 8|Proposal reviews from class 8]], [[/Boards from class 8|Boards from class 8]] (big picture intro stuff)&lt;br /&gt;
**[[/Class 9|Class 9]] 20 Mar 2009, [[/Literature to read for class 9|Reading for class 9]]&lt;br /&gt;
**Spring break&lt;br /&gt;
**[[/Class 10|Class 10]] 3 Apr 2009, [[/Literature to read for class 10|Reading for class 10]]&lt;br /&gt;
**[[/Class 11|Class 11]] 10 Apr 2009, [[/Literature to read for class 11|Reading for class 11]] &lt;br /&gt;
**[[/Class 12|Class 12]] 17 Apr 2009, [[/Literature to read for class 12|Reading for class 12]] &lt;br /&gt;
*[[/Assignments|Assignments]]&lt;br /&gt;
*&#039;&#039;&#039;[[/Research proposal|Research proposal]]&#039;&#039;&#039;&lt;br /&gt;
*&#039;&#039;&#039;[[CS295J/Research proposal (draft 2) | Research proposal (draft 2)]]&#039;&#039;&#039;&lt;br /&gt;
*[[/Relevant Journals and Conferences|Relevant Journals and Conferences]]&lt;br /&gt;
*[[/Literature|Literature]]&lt;br /&gt;
**[[/Literature week 2|Literature to read for class 3]] (2/6/09)&lt;br /&gt;
**[[/Literature to read for class 4|Reading for class 4]] (2/13/09)&lt;br /&gt;
**[[/Literature to read for class 5|Reading for class 5]] (2/20/09)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
*[[/Rule Lists|Rule Lists]]&lt;br /&gt;
*[[/Model elements|Model elements]]&lt;br /&gt;
*[[/Project Concepts|Project Concepts]]&lt;br /&gt;
&lt;br /&gt;
[[/How Tos|How Tos]] -- [[/Class Members&#039; Pages|Class Members&#039; Pages]] -- [http://groups.google.com/group/cs295j Mailing list archive]&lt;br /&gt;
&lt;br /&gt;
[[Category:Courses]]&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Assignments&amp;diff=3085</id>
		<title>CS295J/Assignments</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Assignments&amp;diff=3085"/>
		<updated>2009-04-21T16:59:36Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Assignment 12 (out April 21, due April 24, 2009) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Assignment 12 (out April 21, due April 24, 2009) ==&lt;br /&gt;
&lt;br /&gt;
Complete all of your owned subsections for the [[CS295J/Research proposal (draft 2) | Research proposal (draft 2)]].&lt;br /&gt;
&lt;br /&gt;
== Assignment 11 (out April 10, due April 17, 2009) ==&lt;br /&gt;
&lt;br /&gt;
=== Part A (due Tuesday at noon) ===&lt;br /&gt;
First, revise any contribution that you added to the proposal to reflect our converging view of what we are proposing. Given the evolution of the proposal since the contributions were originally created, if you wish to start from a fresh set of contributions, you can outline them [[CS295J/Contributions for class 12 |here]].&lt;br /&gt;
&lt;br /&gt;
Own one or more of the possible tasks below (enough to spend 8-10 hours on).  Edit this page to indicate your ownership and to describe enough details of what you&#039;ll do for the group to avoid duplication.  It&#039;s ok, for exmaple, for two folks to try using or extending CPM-GOMS with different extensions or a different application; indicating what your extension/application is, though would be helpful.  It&#039;s also fine to add tasks to the list.&lt;br /&gt;
&lt;br /&gt;
=== Part B (due class time) ===&lt;br /&gt;
&lt;br /&gt;
Revise any parts of the preliminary results that you have added.  That may include adding things that you would like to do, taking out pieces that don&#039;t fit into our worldview any longer, or revising things.  The &amp;quot;CPM-GOMS for gmail&amp;quot; project presented today, for example, should go into the preliminary work section to establish viability of CPM-GOMS in the context of a more modern application.&lt;br /&gt;
&lt;br /&gt;
Complete owned task(s).&lt;br /&gt;
&lt;br /&gt;
=== Possible Tasks (also consider those from last week -- add here if you pick one) ===&lt;br /&gt;
&lt;br /&gt;
* Possible additions to the proposed research&lt;br /&gt;
** Experiments with EEG to determine if they can inform modeling (&#039;&#039;&#039;Andrew Bragdon&#039;&#039;&#039; -- I will look into the EEG literature from the CHI community and determine what could be put in the proposal on this topic)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;CONCLUSION:&#039;&#039; Buying the BioSemi system used by Tan et. al. would cost about 50,000+ Euros.  I think that we should add this to the budget, and incorporate EEG measurements into our methodology.  Also, my conclusion is that making one giant equation which outputs only time may not be realistic.  We should refine what our theory can predict, and we should consider working memory load as a model of cognitive load.&lt;br /&gt;
&lt;br /&gt;
** Maybe similar experiments with muscle sensing or body motion&lt;br /&gt;
* Possible preliminary experiments&lt;br /&gt;
** Try some EEG&lt;br /&gt;
** Try some low-cost muscle tracking&lt;br /&gt;
** Try to decode some captured events&lt;br /&gt;
** Try to get some pupil-tracking thing going and sync&#039;ed with other events (I found an open-source eye-tracking program called [http://www.inference.phy.cam.ac.uk/opengazer/ OpenGazer], but it&#039;s written for linux and might require considerable work to port. Would someone with a linux machine like to try it out? Adam)&lt;br /&gt;
** Try some light-weight JavaScript-based mouse tracking (will see how far I can get in 10 hours) ([[User:E J Kalafarski|E J Kalafarski]] 15:16, 14 April 2009 (UTC))&lt;br /&gt;
* Other&lt;br /&gt;
** Get &amp;quot;Human information processor model&amp;quot; concept into related work; ernestine is canonical example, but other more recent work too (Steven,  Eric (Steven wrote core of the article, but I&#039;ll add to it as I come across other work))&lt;br /&gt;
** Revise proposal introduction again, emphasis on our new converging view ([[User:E J Kalafarski|E J Kalafarski]] 15:13, 14 April 2009 (UTC), [[User:Trevor|Trevor O&#039;Brien]], Eric)&lt;br /&gt;
** Will revise and expand &amp;quot;big world&amp;quot; significance and contributions, based on the output/architecture we discussed last week ([[User:E J Kalafarski|E J Kalafarski]] 15:15, 14 April 2009 (UTC), [[User:Trevor|Trevor O&#039;Brien]], Eric)&lt;br /&gt;
** Move 30-hour projects out of the contributions sections (Eric)&lt;br /&gt;
* CPM-GOMS&lt;br /&gt;
** Extend the cognition element in the CPM-GOMS model to account for dual-process theories of reasoning. (OWNER - Gideon; Jon)&lt;br /&gt;
** Extend the model by incorporating an &#039;Affect&#039; output, instead of just relying on time as the only dependent variable. (SUGGESTED BY - Gideon; Ian (I like this idea))&lt;br /&gt;
&lt;br /&gt;
== Assignment 10 (out April 3, due April 10, 2009) ==&lt;br /&gt;
&lt;br /&gt;
=== Part A (due Tuesday at noon) ===&lt;br /&gt;
Own one or more of the possible tasks below (enough to spend 8-10 hours on).  Edit this page to indicate your ownership and to describe enough details of what you&#039;ll do for the group to avoid duplication.  It&#039;s ok, for exmaple, for two folks to try using or extending CPM-GOMS with different extensions or a different application; indicating what your extension/application is, though would be helpful.  It&#039;s also fine to add tasks to the list.&lt;br /&gt;
&lt;br /&gt;
=== Part B (due class time) ===&lt;br /&gt;
&lt;br /&gt;
Complete owned task(s).&lt;br /&gt;
&lt;br /&gt;
=== Possible Tasks ===&lt;br /&gt;
&lt;br /&gt;
* read CPM-GOMS and consider as theme (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;, &#039;&#039;&#039;Eric&#039;&#039;&#039;, &#039;&#039;&#039;Jon&#039;&#039;&#039;, &#039;&#039;&#039;Gideon&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Steven Ellis|Steven]]&#039;&#039;&#039;, &#039;&#039;&#039;Ian&#039;&#039;&#039;)&lt;br /&gt;
* converge proposal (ID weaknesses &amp;amp; fix or propose fixes)&lt;br /&gt;
** make contributions and aims agree (&#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
** make contributions and aims compelling and significant&lt;br /&gt;
*** intro acceptable (merge the two best) (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will rewrite intro, merging the two best &amp;quot;in progress&amp;quot; intros, with an emphasis on the project&#039;s interdisciplinary aspect)&lt;br /&gt;
*** improves world, lots of people, in big ways, long time (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will brainstorm and add to &amp;quot;big picture&amp;quot; contributions, but someone else should brainstorm this as well, &#039;&#039;&#039;Eric&#039;&#039;&#039;)&lt;br /&gt;
*** extends knowledge, increased productivity (designers and users) (&#039;&#039;&#039;Eric&#039;&#039;&#039;)&lt;br /&gt;
*** bg &amp;amp; significance section consistent with contribs/aims (&#039;&#039;&#039;Ian&#039;&#039;&#039;)&lt;br /&gt;
* try CPM-GOMS, maybe w/ 1 extension (&#039;&#039;&#039;Gideon&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Steven Ellis|Steven]]&#039;&#039;&#039;)&lt;br /&gt;
* what&#039;s input and output (architecture)? (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will attempt to come up with an architecture that accommodates all seven proposed modules and presents a simple, useful final product, &#039;&#039;Jon: Will brainstorm b/c I&#039;ve been wondering about this for a long time&#039;&#039;, &#039;&#039;&#039;Adam&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
* Gajos paper (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will read, consider architecture method, &#039;&#039;&#039;Adam&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
&lt;br /&gt;
== Assignment 9 (out March 20, due April 3, 2009) ==&lt;br /&gt;
&lt;br /&gt;
Part A is what we talked about in class -- a revision of the introduction to re-evaluate the big picture.  Part B is the original assignment.  While I&#039;ve asked for both below, it may not be realistic given the time available.  Please make sure to finish A by class time and to do as much of B as you can.&lt;br /&gt;
&lt;br /&gt;
=== Part A, due noon Tuesday March 31 ===&lt;br /&gt;
&lt;br /&gt;
Refine your introduction to be consistent with the theme of bringing knowledge of cognitive and perceptual modeling to bear in a principled way on human-computer interface design.  Evaluate each of the suggested contributions for their impact, their relevance to the theme, and their cost.  Triage as appropriate to achieve the best overall proposal given the 5 years with 5 people scope.&lt;br /&gt;
&lt;br /&gt;
* [[CS295J/Proposal intros from class 9|Link to Intros]]&lt;br /&gt;
* [[CS295J/Triage for class 10|Link to Triages]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due at class time ===&lt;br /&gt;
&lt;br /&gt;
Complete your 30 hour preliminary work project and writeup in the proposal.  It should explicitly support the contributions of the proposal.  Consider this a hard, externally imposed deadline, i.e., you have to have something as finished as possible that could be reviewed by an outside reviewer.  Trim scope, if necessary, but finish!&lt;br /&gt;
&lt;br /&gt;
Pick another gap in the proposal and plug it.  Any section that you have &amp;quot;owned&amp;quot; should also be &amp;quot;final&amp;quot; and reviewable.&lt;br /&gt;
&lt;br /&gt;
====Gaps====&lt;br /&gt;
* &amp;lt;s&amp;gt;Significance (expand further with unified description of relatable components)&amp;lt;/s&amp;gt;&lt;br /&gt;
* Specific contributions &amp;gt; Multi-modal HCI (question marks, needs content)&lt;br /&gt;
* Background &amp;gt; Workflow analysis &amp;gt; Interface improvements (incomplete)&lt;br /&gt;
* Research plan (we may have to cut this section, it is woefully underdeveloped)&lt;br /&gt;
&lt;br /&gt;
Read [[CS295J/Literature to read for class 10|&amp;quot;must reads&amp;quot;]] -- add one if you want.&lt;br /&gt;
&lt;br /&gt;
== Assignment 8 (out March 13, due March 20, 2009) ==&lt;br /&gt;
=== Part A, due Tuesday noon ===&lt;br /&gt;
* Outline a coherent 250 word summary to a coherent proposal [[CS295J/Proposal intros from class 9|here]]&lt;br /&gt;
* Select a gap to own&lt;br /&gt;
** &#039;&#039;&#039;Andrew Bragdon&#039;&#039;&#039;:  Metawork Support Tool proposal; will examine integrating this into the main proposal vs. making a separate proposal.&lt;br /&gt;
** &#039;&#039;&#039;EJ&#039;&#039;&#039;: &amp;quot;Significance/intellectual merit&amp;quot; section is currently bare, I can take a stab at that.  I believe this can/needs to incorporate a gap Trevor identified, &amp;quot;mapping between individual contributions and centralized theme of the proposal;&amp;quot; I&#039;ll try to start to illustrate the relationships between our individual projects. [[User:E J Kalafarski|E J Kalafarski]] 12:20, 17 March 2009 (UTC)&lt;br /&gt;
**&#039;&#039;&#039;Jon&#039;&#039;&#039;: The &amp;quot;models of perception&amp;quot; section needs to be revised and expanded.  I&#039;ve changed &amp;quot;Gibsonianism&amp;quot; to &amp;quot;The Ecological Approach to Perception&amp;quot; and I&#039;ve added a section on &amp;quot;The Computational Approach to Perception&amp;quot;; I will update these for Friday.&lt;br /&gt;
**&#039;&#039;&#039;Gideon&#039;&#039;&#039;: More up-to-date background section on distributed cognition. I know Jon is doing this for the other areas, but I feel that this is very important.&lt;br /&gt;
** &#039;&#039;&#039;Trevor&#039;&#039;&#039;:  I&#039;ll take a pass through the &#039;&#039;Specific Aims&#039;&#039; section, which currently lacks specificity.   --- [[User:Trevor O&amp;amp;#39;Brien|Trevor O&amp;amp;#39;Brien]] 15:29, 17 March 2009 (UTC) &lt;br /&gt;
**&#039;&#039;&#039;Eric&#039;&#039;&#039;: I&#039;ll flesh out the &#039;&#039;Workflow Analysis&#039;&#039; section. What&#039;s there could be cleaned up, and more could be added, particularly on learning from interaction histories.&lt;br /&gt;
**&#039;&#039;&#039;Steven&#039;&#039;&#039;: There&#039;s currently no background (slash anything) on the topic of embodied models of cognition, I&#039;ll fill in some background on that.&lt;br /&gt;
* suggest 0.5 &amp;quot;must read&amp;quot; [[CS295J/Literature to read for class 9|papers]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due in class ===&lt;br /&gt;
* Finish writing the coherent 250 word summary to a coherent proposal [[CS295J/Proposal intros from class 9|still here]]&lt;br /&gt;
* Fill your gap&lt;br /&gt;
* read &amp;quot;must read&amp;quot;s for discussion w.r.t. proposal relevance&lt;br /&gt;
* get to another week&#039;s worth of results in your preliminary results -- make sure to be consistent with your summary!&lt;br /&gt;
* be prepared to tell us about those new results in class, tied to the intellectual claims&lt;br /&gt;
&lt;br /&gt;
== Assignment 7 (out March 6, due March 13, 2009) ==&lt;br /&gt;
=== Part A, due Tuesday noon ===&lt;br /&gt;
* review [[CS295J/Research proposal|proposal]] for&lt;br /&gt;
** intellectual contribution (1-2 paragraphs)&lt;br /&gt;
** major gaps (bullet list) (don&#039;t duplicate gaps already listed)&lt;br /&gt;
** review goes [[CS295J/Proposal reviews from class 8|here]]&lt;br /&gt;
** feel free to improve any part of the proposal rather than criticizing it, if you wish :-)&lt;br /&gt;
* suggest 0.5 &amp;quot;must read&amp;quot; [[CS295J/Literature to read for class 8|papers]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due in class ===&lt;br /&gt;
* read &amp;quot;must read&amp;quot;s for discussion w.r.t. proposal relevance&lt;br /&gt;
* get to another week&#039;s worth of results in your preliminary results&lt;br /&gt;
* be prepared to tell us about those new results in class, tied to the intellectual claims&lt;br /&gt;
&lt;br /&gt;
== Assignment 6 (out February 27, due March 6, 2009) ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* If you are not familiar with NIH proposals, skim through [http://vis.cs.brown.edu/docs/pdf/bib/Laidlaw-2005-DA2.pdf this active proposal] to get a sense of the sections.  The [http://vis.cs.brown.edu/docs/pdf/bib/NIH-2001-PHS.pdf guide to nih proposals] may also be helpful.&lt;br /&gt;
* Create some preliminary results&lt;br /&gt;
** Start the work you proposed in your poster for assignment 4&lt;br /&gt;
** Start the work someone else proposed :-)&lt;br /&gt;
** Create a more detailed critique-enabled workflow of a scientific user&lt;br /&gt;
** Working in pairs is fine; preferably, pair members should have different backgrounds&lt;br /&gt;
* Write them up as a subsection of the preliminary results section of the proposal.&lt;br /&gt;
** By Wednesday 9am have an outline of the section you will produce&lt;br /&gt;
*** This should be in past tense, as it will be when the work is done.&lt;br /&gt;
*** At the top level, it should state how the preliminary results enable the overall multi-year proposal or how they demonstrate feasibility of some questionable or risky part.&lt;br /&gt;
*** Check out the NIH proposal for examples, albeit in a different domain.&lt;br /&gt;
** By class have as much filled in from that outline as possible&lt;br /&gt;
* Add any &amp;quot;must reads&amp;quot; to the [[CS295J/Literature to read for class 6]] page.  You are &amp;quot;expected&amp;quot; to add 1/2 of a reference.&lt;br /&gt;
* Read and be prepared to discuss the &amp;quot;must reads&amp;quot;.&lt;br /&gt;
** Put in fictional placeholders for parts that are not done by Friday.&lt;br /&gt;
* By Wednesday 5pm e-mail constructive comments about at least 1/2 of the outlines to the entire class&lt;br /&gt;
* Success criteria for this assignment&lt;br /&gt;
*# Proposal section will demonstrate a feasibility or prerequisite for interesting research&lt;br /&gt;
*# Results by class are complete and concrete enough that they are interesting, even though some parts may be missing&lt;br /&gt;
&lt;br /&gt;
== Assignment 5 (out February 20, due February 27, 2009) ==&lt;br /&gt;
# Videotape a 15-20 minute interactive session with a user doing a visually challenging task for which they are at least an advanced beginner (ie, they don&#039;t have to look up what to do, but they have not yet internalized the operations and made them subconscious).  Analyze the session and report your observations and conclusions.&lt;br /&gt;
&lt;br /&gt;
== Assignment 4 (out February 13, due February 20, 2009) ==&lt;br /&gt;
&lt;br /&gt;
# Be prepared to decide on a subset of applications to critique.  Flesh out at least one potential application in [[CS295J/Application Critiques]], including a specific workflow.  Modify any others you want, particularly in terms of arguments for or against proceeding with them.  These should be completed by Thursday noon.&lt;br /&gt;
#* Some possibilities: Google scholar, Mathematica, Tableau, Google notebook, Matlab, paper, media wiki, Ensight Gold, AVS, VisTrails.  Note that a given piece of software could be represented more than once with a different workflow.&lt;br /&gt;
#* Possible criteria: main purpose is analysis, perhaps scientific; interative; scientific; amenable to cognition-driven improvemennts; interesting; fun&lt;br /&gt;
# Bring your ranking of the subset of (application+workflow)s we should critique&lt;br /&gt;
# Be prepared to decide on the [[CS295J/Model elements]] of cognition we will emulate to critique each application.  Revise some part of this list so that it can be used as a concrete basis for evaluation.&lt;br /&gt;
# Add any &amp;quot;must reads&amp;quot; to the [[CS295J/Literature to read for class 5]] page.  You are &amp;quot;expected&amp;quot; to add 1/3 of a reference.&lt;br /&gt;
# Read the &amp;quot;must reads&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
== Assignment 3 (out February 6, due February 13) ==&lt;br /&gt;
* Flesh out and support a contribution within the proposal&lt;br /&gt;
** Add any new references to the literature section.&lt;br /&gt;
** Many of our readings date from 8+ years ago; check to make sure your contribution hasn&#039;t already been done.&lt;br /&gt;
** If any of the new references are &amp;quot;must reads&amp;quot;, add to the [[CS295J/Literature to read for class 4]] (2/13/09) page.  You are &amp;quot;expected&amp;quot; to add 1/2 of a reference.&lt;br /&gt;
** Estimate the impact of the contribution.&lt;br /&gt;
** Estimate the risk and costs of the contribution.&lt;br /&gt;
** Propose a 3-week (30 hour) result that you could create to demonstrate feasibility.&lt;br /&gt;
** Bring a printout of your contribution concept to class.  It should be legible from 2 meters away, so use big text.  You can hand-write it on posterboard or paste together printouts.&lt;br /&gt;
* Bring a list of holes in the proposal -- e.g., &amp;quot;background section needs something on workflow capture.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Assignment 2 (out January 30, 2009) ==&lt;br /&gt;
* Refine literature short summaries to include the relationship to our project.&lt;br /&gt;
* Draft by tuesday noon a subsection in the background section of the [[CS295J/Research proposal]], making sure to include citations to the relevant materials in the literature page.  Add any new references to the literature page.&lt;br /&gt;
* Read and comment/edit by Thursday noon all background sections&lt;br /&gt;
* Revise your background section by Friday noon (so David can print before class)&lt;br /&gt;
&lt;br /&gt;
* Add to or refine one or more specific contribution in the [[CS295J/Research proposal]]; each contribution must have a list of ways it can be demonstrated.  Some will become part of the preliminary results, others will be parts of the future work that will be proposed.&lt;br /&gt;
* Add to or refine one or more specific aim in the [[CS295J/Research proposal]]; make it consistent with the contribution you added.&lt;br /&gt;
&lt;br /&gt;
* Identify the one additional most important paper for us to read this week also by Tuesday noon; be prepared to summarize relevance in 2 minutes in class&lt;br /&gt;
* Read those &amp;quot;most important&amp;quot; papers for class discussion.&lt;br /&gt;
&lt;br /&gt;
* Be prepared to summarize to the class your contributions to the background, contributions, and aims sections.&lt;br /&gt;
* If there is preliminary work that will help to make decisions about contributions and aims, please get started on it (and be ready to report on what you&#039;d like to do or what you have done).&lt;br /&gt;
&lt;br /&gt;
== Assignment 1 (out January 23, 2009) ==&lt;br /&gt;
* spend 10 hours adding to any part of the wiki you think is relevant&lt;br /&gt;
* by Monday noon add any potential readings.  If you&#039;ve got a tentative summary evaluation, go ahead and add it.  It&#039;s ok to edit folks summary evaluations, but try to make the result more accurate or precise without losing information.&lt;br /&gt;
* by Wednesday noon finish with any summary evaluation and also identify at least one as-relevant-as-possible reading as yours.  Put your name on that entry in the reading list as the &amp;quot;owner&amp;quot; so that there are no duplicates.&lt;br /&gt;
* by Wednesday 5pm -- select 2 additional relevant readings that are owned and that you will read by Friday and be prepared to discuss.  Put your name as a &amp;quot;discussant&amp;quot; in the reading list; there should be a max of two discussants per reading.&lt;br /&gt;
* by Friday class -- author a summary description, less than 250 words, in the wiki of how the reading you own relates to our project.  Be prepared to describe, in two minutes, how your reading relates to the project. Also be prepared for everyone in class to discuss after your description.  You may bring notes for yourself, but no slides.  The wiki page for your reading will be displayed while you talk.&lt;br /&gt;
* by Friday class -- read and be prepared to discuss the other two readings you choose.&lt;br /&gt;
* by Friday class -- make one more wiki page titled &amp;quot;&#039;&#039;&#039;&amp;lt;Last-Name&amp;gt; week 1&#039;&#039;&#039;&amp;quot; with a list of the keys for the citations you added, the readings  you summarized, the reading you presented, the two readings you were a discussant on, and any other readings you did in detail.&lt;br /&gt;
* Let me know if you have any kind of problems.  You should be spending right around 10 hours -- if that&#039;s a problem, let&#039;s talk.&lt;br /&gt;
* The [[../How Tos|How Tos]] page has some tips.  Edit or add as you find others.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3084</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3084"/>
		<updated>2009-04-21T16:55:45Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Specific Aims and Contributions (to be separated later) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
In order to use this framework, a designer will have to provide:&lt;br /&gt;
* Functional specification - what are the possible interactions between the user and the application. This can be thought of as method signatures, with a name (e.g., setVolume), direction (to user or from user) and a list of value types (boolean, number, text, video, ...) for each interaction.&lt;br /&gt;
* GUI specification - a mapping of interactions to interface elements (e.g., setVolume is mapped to the grey knob in the bottom left corner with clockwise turning increasing the input number).&lt;br /&gt;
* Functional user traces - sequences of representative ways in which the application is used. Instead of writing them, the designer could have users use the application with a trial interface and then use our methods to generalize the user traces beyond the specific interface (The second method is depicted in the diagram). As a form of pre-processing, the system also generates an interaction transition matrix which lists the probability of each type of interaction given the previous interaction.&lt;br /&gt;
* Utility function - this is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
Each of the modules can use all of this information or a subset of it. Our approach stresses flexibility and the ability to give more meaningful feedback the more information is provided. After processing the information sent by the system of experts, the aggregator will output:&lt;br /&gt;
* An evaluation of the interface. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
* Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Each person should come up with a single paragraph describing fictional (or not) preliminary results pertaining to their owned specific aims and contributions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3082</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3082"/>
		<updated>2009-04-21T16:33:53Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
== Inputs ==&lt;br /&gt;
&lt;br /&gt;
== Outputs ==&lt;br /&gt;
&lt;br /&gt;
Also passed as input is the utility function to optimize over. This utility function is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. For example, a user at NASA probably cares more about interface accuracy than speed. By passing this information to our committee of experts, we can create interfaces that are tuned to maximize the utility of a particular user type.&lt;br /&gt;
&lt;br /&gt;
As output, the aggregator will provide an evaluation of the interface and a set of recommended improvements. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
&lt;br /&gt;
Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Each person should come up with a single paragraph describing fictional (or not) preliminary results pertaining to their owned specific aims and contributions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3081</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3081"/>
		<updated>2009-04-21T16:29:49Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Outputs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
== Inputs ==&lt;br /&gt;
&lt;br /&gt;
== Outputs ==&lt;br /&gt;
&lt;br /&gt;
Also passed as input is the utility function to optimize over. This utility function is a weighting of various performance metrics (time, cognitive load, fatigue, etc.), where the weighting expresses the importance of a particular dimension to the user. The utility function is used by the aggregator to provide evaluations and recommendations appropriate to user preferences.&lt;br /&gt;
&lt;br /&gt;
As output, the aggregator will provide an evaluation of the interface and a set of recommended improvements. Evaluations are expressed both in terms of the utility function components (i.e. time, fatigue, cognitive load, etc.), and in terms of the overall utility for this interface (as defined by the utility function). These evaluations are given in the form of an efficiency curve, where the utility received on each dimension can change as the user becomes more accustomed to the interface. &lt;br /&gt;
&lt;br /&gt;
Suggested improvements for the GUI are also output. These suggestions are meant to optimize the utility function that was input to the system. If a user values accuracy over time, interface suggestions will be made accordingly.&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Each person should come up with a single paragraph describing fictional (or not) preliminary results pertaining to their owned specific aims and contributions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3079</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3079"/>
		<updated>2009-04-21T16:01:46Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Specific Aims and Contributions (to be separated later) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
== Inputs ==&lt;br /&gt;
&lt;br /&gt;
== Outputs ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Each person should come up with a single paragraph describing fictional (or not) preliminary results pertaining to their owned specific aims and contributions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3078</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3078"/>
		<updated>2009-04-21T16:01:14Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Specific Aims and Contributions (to be separated later) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
Each person should add the background related to their specific aims.&lt;br /&gt;
&lt;br /&gt;
* Steven Ellis - Cognitive models of HCI, including GOMS variations and ACT-R&lt;br /&gt;
* EJ - Design Guidelines&lt;br /&gt;
* Jon - Perception and Action&lt;br /&gt;
* Andrew - Multiple task environments&lt;br /&gt;
* Gideon - Cognition and dual systems&lt;br /&gt;
* Ian - Interface design process&lt;br /&gt;
* Trevor - User trace collection methods (especially any eye-tracking, EEG, ... you want to suggest using)&lt;br /&gt;
&lt;br /&gt;
= Specific Aims and Contributions (to be separated later) =&lt;br /&gt;
See the [http://vrl.cs.brown.edu/wiki/images/b/b1/Flowchart2.pdf flowchart] for a visual overview of our aims.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;The user traces that are collected are tied to a specific interface. In order to use them with different interfaces to the same application, they should be generalized to be based only on the functional description of the application and the user&#039;s goal hierarchy. This would abstract away from actions like accessing a menu.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;In addition to specific user traces, many modules could use a transition probability matrix based on interaction predictions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Parallel Framework for Evaluation Modules ==&lt;br /&gt;
* Owner: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section will describe in more detail the inputs, outputs and architecture that were presented in the introduction.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Integration into the Design Process ==&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section outlines the process of designing an HCI interface and at what stages our proposal fits in and how.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Each person should come up with a single paragraph describing fictional (or not) preliminary results pertaining to their owned specific aims and contributions.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=File:Flowchart2.pdf&amp;diff=3077</id>
		<title>File:Flowchart2.pdf</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=File:Flowchart2.pdf&amp;diff=3077"/>
		<updated>2009-04-21T15:58:39Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3067</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3067"/>
		<updated>2009-04-21T14:44:02Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Evaluation and Recommendation via Modules */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3064</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3064"/>
		<updated>2009-04-21T14:41:57Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Interruptions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3063</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3063"/>
		<updated>2009-04-21T14:41:18Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Interaction Prediction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section could include an example or two of established design guidelines that could easily be implemented as modules.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Affordances ===&lt;br /&gt;
* Owner: [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will provide interface evaluations and recommendations based on perceived affordances and if possible a comparison to actual affordances.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Working Memory Load ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module measures how much information the user needs to retain in memory while interacting with the interface and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Automaticity of Interaction ===&lt;br /&gt;
* Owner: [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Measures how easily the interaction with the interface becomes automatic with experience and makes suggestions for improvements.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Assignments&amp;diff=3057</id>
		<title>CS295J/Assignments</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Assignments&amp;diff=3057"/>
		<updated>2009-04-21T14:04:06Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Assignment 12 (out April 21, due April 24, 2009) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Assignment 12 (out April 21, due April 24, 2009) ==&lt;br /&gt;
&lt;br /&gt;
Coming soon.&lt;br /&gt;
&lt;br /&gt;
== Assignment 11 (out April 10, due April 17, 2009) ==&lt;br /&gt;
&lt;br /&gt;
=== Part A (due Tuesday at noon) ===&lt;br /&gt;
First, revise any contribution that you added to the proposal to reflect our converging view of what we are proposing. Given the evolution of the proposal since the contributions were originally created, if you wish to start from a fresh set of contributions, you can outline them [[CS295J/Contributions for class 12 |here]].&lt;br /&gt;
&lt;br /&gt;
Own one or more of the possible tasks below (enough to spend 8-10 hours on).  Edit this page to indicate your ownership and to describe enough details of what you&#039;ll do for the group to avoid duplication.  It&#039;s ok, for exmaple, for two folks to try using or extending CPM-GOMS with different extensions or a different application; indicating what your extension/application is, though would be helpful.  It&#039;s also fine to add tasks to the list.&lt;br /&gt;
&lt;br /&gt;
=== Part B (due class time) ===&lt;br /&gt;
&lt;br /&gt;
Revise any parts of the preliminary results that you have added.  That may include adding things that you would like to do, taking out pieces that don&#039;t fit into our worldview any longer, or revising things.  The &amp;quot;CPM-GOMS for gmail&amp;quot; project presented today, for example, should go into the preliminary work section to establish viability of CPM-GOMS in the context of a more modern application.&lt;br /&gt;
&lt;br /&gt;
Complete owned task(s).&lt;br /&gt;
&lt;br /&gt;
=== Possible Tasks (also consider those from last week -- add here if you pick one) ===&lt;br /&gt;
&lt;br /&gt;
* Possible additions to the proposed research&lt;br /&gt;
** Experiments with EEG to determine if they can inform modeling (&#039;&#039;&#039;Andrew Bragdon&#039;&#039;&#039; -- I will look into the EEG literature from the CHI community and determine what could be put in the proposal on this topic)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;CONCLUSION:&#039;&#039; Buying the BioSemi system used by Tan et. al. would cost about 50,000+ Euros.  I think that we should add this to the budget, and incorporate EEG measurements into our methodology.  Also, my conclusion is that making one giant equation which outputs only time may not be realistic.  We should refine what our theory can predict, and we should consider working memory load as a model of cognitive load.&lt;br /&gt;
&lt;br /&gt;
** Maybe similar experiments with muscle sensing or body motion&lt;br /&gt;
* Possible preliminary experiments&lt;br /&gt;
** Try some EEG&lt;br /&gt;
** Try some low-cost muscle tracking&lt;br /&gt;
** Try to decode some captured events&lt;br /&gt;
** Try to get some pupil-tracking thing going and sync&#039;ed with other events (I found an open-source eye-tracking program called [http://www.inference.phy.cam.ac.uk/opengazer/ OpenGazer], but it&#039;s written for linux and might require considerable work to port. Would someone with a linux machine like to try it out? Adam)&lt;br /&gt;
** Try some light-weight JavaScript-based mouse tracking (will see how far I can get in 10 hours) ([[User:E J Kalafarski|E J Kalafarski]] 15:16, 14 April 2009 (UTC))&lt;br /&gt;
* Other&lt;br /&gt;
** Get &amp;quot;Human information processor model&amp;quot; concept into related work; ernestine is canonical example, but other more recent work too (Steven,  Eric (Steven wrote core of the article, but I&#039;ll add to it as I come across other work))&lt;br /&gt;
** Revise proposal introduction again, emphasis on our new converging view ([[User:E J Kalafarski|E J Kalafarski]] 15:13, 14 April 2009 (UTC), [[User:Trevor|Trevor O&#039;Brien]], Eric)&lt;br /&gt;
** Will revise and expand &amp;quot;big world&amp;quot; significance and contributions, based on the output/architecture we discussed last week ([[User:E J Kalafarski|E J Kalafarski]] 15:15, 14 April 2009 (UTC), [[User:Trevor|Trevor O&#039;Brien]], Eric)&lt;br /&gt;
** Move 30-hour projects out of the contributions sections (Eric)&lt;br /&gt;
* CPM-GOMS&lt;br /&gt;
** Extend the cognition element in the CPM-GOMS model to account for dual-process theories of reasoning. (OWNER - Gideon; Jon)&lt;br /&gt;
** Extend the model by incorporating an &#039;Affect&#039; output, instead of just relying on time as the only dependent variable. (SUGGESTED BY - Gideon; Ian (I like this idea))&lt;br /&gt;
&lt;br /&gt;
== Assignment 10 (out April 3, due April 10, 2009) ==&lt;br /&gt;
&lt;br /&gt;
=== Part A (due Tuesday at noon) ===&lt;br /&gt;
Own one or more of the possible tasks below (enough to spend 8-10 hours on).  Edit this page to indicate your ownership and to describe enough details of what you&#039;ll do for the group to avoid duplication.  It&#039;s ok, for exmaple, for two folks to try using or extending CPM-GOMS with different extensions or a different application; indicating what your extension/application is, though would be helpful.  It&#039;s also fine to add tasks to the list.&lt;br /&gt;
&lt;br /&gt;
=== Part B (due class time) ===&lt;br /&gt;
&lt;br /&gt;
Complete owned task(s).&lt;br /&gt;
&lt;br /&gt;
=== Possible Tasks ===&lt;br /&gt;
&lt;br /&gt;
* read CPM-GOMS and consider as theme (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;, &#039;&#039;&#039;Eric&#039;&#039;&#039;, &#039;&#039;&#039;Jon&#039;&#039;&#039;, &#039;&#039;&#039;Gideon&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Steven Ellis|Steven]]&#039;&#039;&#039;, &#039;&#039;&#039;Ian&#039;&#039;&#039;)&lt;br /&gt;
* converge proposal (ID weaknesses &amp;amp; fix or propose fixes)&lt;br /&gt;
** make contributions and aims agree (&#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
** make contributions and aims compelling and significant&lt;br /&gt;
*** intro acceptable (merge the two best) (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will rewrite intro, merging the two best &amp;quot;in progress&amp;quot; intros, with an emphasis on the project&#039;s interdisciplinary aspect)&lt;br /&gt;
*** improves world, lots of people, in big ways, long time (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will brainstorm and add to &amp;quot;big picture&amp;quot; contributions, but someone else should brainstorm this as well, &#039;&#039;&#039;Eric&#039;&#039;&#039;)&lt;br /&gt;
*** extends knowledge, increased productivity (designers and users) (&#039;&#039;&#039;Eric&#039;&#039;&#039;)&lt;br /&gt;
*** bg &amp;amp; significance section consistent with contribs/aims (&#039;&#039;&#039;Ian&#039;&#039;&#039;)&lt;br /&gt;
* try CPM-GOMS, maybe w/ 1 extension (&#039;&#039;&#039;Gideon&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Steven Ellis|Steven]]&#039;&#039;&#039;)&lt;br /&gt;
* what&#039;s input and output (architecture)? (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will attempt to come up with an architecture that accommodates all seven proposed modules and presents a simple, useful final product, &#039;&#039;Jon: Will brainstorm b/c I&#039;ve been wondering about this for a long time&#039;&#039;, &#039;&#039;&#039;Adam&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
* Gajos paper (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will read, consider architecture method, &#039;&#039;&#039;Adam&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
&lt;br /&gt;
== Assignment 9 (out March 20, due April 3, 2009) ==&lt;br /&gt;
&lt;br /&gt;
Part A is what we talked about in class -- a revision of the introduction to re-evaluate the big picture.  Part B is the original assignment.  While I&#039;ve asked for both below, it may not be realistic given the time available.  Please make sure to finish A by class time and to do as much of B as you can.&lt;br /&gt;
&lt;br /&gt;
=== Part A, due noon Tuesday March 31 ===&lt;br /&gt;
&lt;br /&gt;
Refine your introduction to be consistent with the theme of bringing knowledge of cognitive and perceptual modeling to bear in a principled way on human-computer interface design.  Evaluate each of the suggested contributions for their impact, their relevance to the theme, and their cost.  Triage as appropriate to achieve the best overall proposal given the 5 years with 5 people scope.&lt;br /&gt;
&lt;br /&gt;
* [[CS295J/Proposal intros from class 9|Link to Intros]]&lt;br /&gt;
* [[CS295J/Triage for class 10|Link to Triages]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due at class time ===&lt;br /&gt;
&lt;br /&gt;
Complete your 30 hour preliminary work project and writeup in the proposal.  It should explicitly support the contributions of the proposal.  Consider this a hard, externally imposed deadline, i.e., you have to have something as finished as possible that could be reviewed by an outside reviewer.  Trim scope, if necessary, but finish!&lt;br /&gt;
&lt;br /&gt;
Pick another gap in the proposal and plug it.  Any section that you have &amp;quot;owned&amp;quot; should also be &amp;quot;final&amp;quot; and reviewable.&lt;br /&gt;
&lt;br /&gt;
====Gaps====&lt;br /&gt;
* &amp;lt;s&amp;gt;Significance (expand further with unified description of relatable components)&amp;lt;/s&amp;gt;&lt;br /&gt;
* Specific contributions &amp;gt; Multi-modal HCI (question marks, needs content)&lt;br /&gt;
* Background &amp;gt; Workflow analysis &amp;gt; Interface improvements (incomplete)&lt;br /&gt;
* Research plan (we may have to cut this section, it is woefully underdeveloped)&lt;br /&gt;
&lt;br /&gt;
Read [[CS295J/Literature to read for class 10|&amp;quot;must reads&amp;quot;]] -- add one if you want.&lt;br /&gt;
&lt;br /&gt;
== Assignment 8 (out March 13, due March 20, 2009) ==&lt;br /&gt;
=== Part A, due Tuesday noon ===&lt;br /&gt;
* Outline a coherent 250 word summary to a coherent proposal [[CS295J/Proposal intros from class 9|here]]&lt;br /&gt;
* Select a gap to own&lt;br /&gt;
** &#039;&#039;&#039;Andrew Bragdon&#039;&#039;&#039;:  Metawork Support Tool proposal; will examine integrating this into the main proposal vs. making a separate proposal.&lt;br /&gt;
** &#039;&#039;&#039;EJ&#039;&#039;&#039;: &amp;quot;Significance/intellectual merit&amp;quot; section is currently bare, I can take a stab at that.  I believe this can/needs to incorporate a gap Trevor identified, &amp;quot;mapping between individual contributions and centralized theme of the proposal;&amp;quot; I&#039;ll try to start to illustrate the relationships between our individual projects. [[User:E J Kalafarski|E J Kalafarski]] 12:20, 17 March 2009 (UTC)&lt;br /&gt;
**&#039;&#039;&#039;Jon&#039;&#039;&#039;: The &amp;quot;models of perception&amp;quot; section needs to be revised and expanded.  I&#039;ve changed &amp;quot;Gibsonianism&amp;quot; to &amp;quot;The Ecological Approach to Perception&amp;quot; and I&#039;ve added a section on &amp;quot;The Computational Approach to Perception&amp;quot;; I will update these for Friday.&lt;br /&gt;
**&#039;&#039;&#039;Gideon&#039;&#039;&#039;: More up-to-date background section on distributed cognition. I know Jon is doing this for the other areas, but I feel that this is very important.&lt;br /&gt;
** &#039;&#039;&#039;Trevor&#039;&#039;&#039;:  I&#039;ll take a pass through the &#039;&#039;Specific Aims&#039;&#039; section, which currently lacks specificity.   --- [[User:Trevor O&amp;amp;#39;Brien|Trevor O&amp;amp;#39;Brien]] 15:29, 17 March 2009 (UTC) &lt;br /&gt;
**&#039;&#039;&#039;Eric&#039;&#039;&#039;: I&#039;ll flesh out the &#039;&#039;Workflow Analysis&#039;&#039; section. What&#039;s there could be cleaned up, and more could be added, particularly on learning from interaction histories.&lt;br /&gt;
**&#039;&#039;&#039;Steven&#039;&#039;&#039;: There&#039;s currently no background (slash anything) on the topic of embodied models of cognition, I&#039;ll fill in some background on that.&lt;br /&gt;
* suggest 0.5 &amp;quot;must read&amp;quot; [[CS295J/Literature to read for class 9|papers]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due in class ===&lt;br /&gt;
* Finish writing the coherent 250 word summary to a coherent proposal [[CS295J/Proposal intros from class 9|still here]]&lt;br /&gt;
* Fill your gap&lt;br /&gt;
* read &amp;quot;must read&amp;quot;s for discussion w.r.t. proposal relevance&lt;br /&gt;
* get to another week&#039;s worth of results in your preliminary results -- make sure to be consistent with your summary!&lt;br /&gt;
* be prepared to tell us about those new results in class, tied to the intellectual claims&lt;br /&gt;
&lt;br /&gt;
== Assignment 7 (out March 6, due March 13, 2009) ==&lt;br /&gt;
=== Part A, due Tuesday noon ===&lt;br /&gt;
* review [[CS295J/Research proposal|proposal]] for&lt;br /&gt;
** intellectual contribution (1-2 paragraphs)&lt;br /&gt;
** major gaps (bullet list) (don&#039;t duplicate gaps already listed)&lt;br /&gt;
** review goes [[CS295J/Proposal reviews from class 8|here]]&lt;br /&gt;
** feel free to improve any part of the proposal rather than criticizing it, if you wish :-)&lt;br /&gt;
* suggest 0.5 &amp;quot;must read&amp;quot; [[CS295J/Literature to read for class 8|papers]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due in class ===&lt;br /&gt;
* read &amp;quot;must read&amp;quot;s for discussion w.r.t. proposal relevance&lt;br /&gt;
* get to another week&#039;s worth of results in your preliminary results&lt;br /&gt;
* be prepared to tell us about those new results in class, tied to the intellectual claims&lt;br /&gt;
&lt;br /&gt;
== Assignment 6 (out February 27, due March 6, 2009) ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* If you are not familiar with NIH proposals, skim through [http://vis.cs.brown.edu/docs/pdf/bib/Laidlaw-2005-DA2.pdf this active proposal] to get a sense of the sections.  The [http://vis.cs.brown.edu/docs/pdf/bib/NIH-2001-PHS.pdf guide to nih proposals] may also be helpful.&lt;br /&gt;
* Create some preliminary results&lt;br /&gt;
** Start the work you proposed in your poster for assignment 4&lt;br /&gt;
** Start the work someone else proposed :-)&lt;br /&gt;
** Create a more detailed critique-enabled workflow of a scientific user&lt;br /&gt;
** Working in pairs is fine; preferably, pair members should have different backgrounds&lt;br /&gt;
* Write them up as a subsection of the preliminary results section of the proposal.&lt;br /&gt;
** By Wednesday 9am have an outline of the section you will produce&lt;br /&gt;
*** This should be in past tense, as it will be when the work is done.&lt;br /&gt;
*** At the top level, it should state how the preliminary results enable the overall multi-year proposal or how they demonstrate feasibility of some questionable or risky part.&lt;br /&gt;
*** Check out the NIH proposal for examples, albeit in a different domain.&lt;br /&gt;
** By class have as much filled in from that outline as possible&lt;br /&gt;
* Add any &amp;quot;must reads&amp;quot; to the [[CS295J/Literature to read for class 6]] page.  You are &amp;quot;expected&amp;quot; to add 1/2 of a reference.&lt;br /&gt;
* Read and be prepared to discuss the &amp;quot;must reads&amp;quot;.&lt;br /&gt;
** Put in fictional placeholders for parts that are not done by Friday.&lt;br /&gt;
* By Wednesday 5pm e-mail constructive comments about at least 1/2 of the outlines to the entire class&lt;br /&gt;
* Success criteria for this assignment&lt;br /&gt;
*# Proposal section will demonstrate a feasibility or prerequisite for interesting research&lt;br /&gt;
*# Results by class are complete and concrete enough that they are interesting, even though some parts may be missing&lt;br /&gt;
&lt;br /&gt;
== Assignment 5 (out February 20, due February 27, 2009) ==&lt;br /&gt;
# Videotape a 15-20 minute interactive session with a user doing a visually challenging task for which they are at least an advanced beginner (ie, they don&#039;t have to look up what to do, but they have not yet internalized the operations and made them subconscious).  Analyze the session and report your observations and conclusions.&lt;br /&gt;
&lt;br /&gt;
== Assignment 4 (out February 13, due February 20, 2009) ==&lt;br /&gt;
&lt;br /&gt;
# Be prepared to decide on a subset of applications to critique.  Flesh out at least one potential application in [[CS295J/Application Critiques]], including a specific workflow.  Modify any others you want, particularly in terms of arguments for or against proceeding with them.  These should be completed by Thursday noon.&lt;br /&gt;
#* Some possibilities: Google scholar, Mathematica, Tableau, Google notebook, Matlab, paper, media wiki, Ensight Gold, AVS, VisTrails.  Note that a given piece of software could be represented more than once with a different workflow.&lt;br /&gt;
#* Possible criteria: main purpose is analysis, perhaps scientific; interative; scientific; amenable to cognition-driven improvemennts; interesting; fun&lt;br /&gt;
# Bring your ranking of the subset of (application+workflow)s we should critique&lt;br /&gt;
# Be prepared to decide on the [[CS295J/Model elements]] of cognition we will emulate to critique each application.  Revise some part of this list so that it can be used as a concrete basis for evaluation.&lt;br /&gt;
# Add any &amp;quot;must reads&amp;quot; to the [[CS295J/Literature to read for class 5]] page.  You are &amp;quot;expected&amp;quot; to add 1/3 of a reference.&lt;br /&gt;
# Read the &amp;quot;must reads&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
== Assignment 3 (out February 6, due February 13) ==&lt;br /&gt;
* Flesh out and support a contribution within the proposal&lt;br /&gt;
** Add any new references to the literature section.&lt;br /&gt;
** Many of our readings date from 8+ years ago; check to make sure your contribution hasn&#039;t already been done.&lt;br /&gt;
** If any of the new references are &amp;quot;must reads&amp;quot;, add to the [[CS295J/Literature to read for class 4]] (2/13/09) page.  You are &amp;quot;expected&amp;quot; to add 1/2 of a reference.&lt;br /&gt;
** Estimate the impact of the contribution.&lt;br /&gt;
** Estimate the risk and costs of the contribution.&lt;br /&gt;
** Propose a 3-week (30 hour) result that you could create to demonstrate feasibility.&lt;br /&gt;
** Bring a printout of your contribution concept to class.  It should be legible from 2 meters away, so use big text.  You can hand-write it on posterboard or paste together printouts.&lt;br /&gt;
* Bring a list of holes in the proposal -- e.g., &amp;quot;background section needs something on workflow capture.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Assignment 2 (out January 30, 2009) ==&lt;br /&gt;
* Refine literature short summaries to include the relationship to our project.&lt;br /&gt;
* Draft by tuesday noon a subsection in the background section of the [[CS295J/Research proposal]], making sure to include citations to the relevant materials in the literature page.  Add any new references to the literature page.&lt;br /&gt;
* Read and comment/edit by Thursday noon all background sections&lt;br /&gt;
* Revise your background section by Friday noon (so David can print before class)&lt;br /&gt;
&lt;br /&gt;
* Add to or refine one or more specific contribution in the [[CS295J/Research proposal]]; each contribution must have a list of ways it can be demonstrated.  Some will become part of the preliminary results, others will be parts of the future work that will be proposed.&lt;br /&gt;
* Add to or refine one or more specific aim in the [[CS295J/Research proposal]]; make it consistent with the contribution you added.&lt;br /&gt;
&lt;br /&gt;
* Identify the one additional most important paper for us to read this week also by Tuesday noon; be prepared to summarize relevance in 2 minutes in class&lt;br /&gt;
* Read those &amp;quot;most important&amp;quot; papers for class discussion.&lt;br /&gt;
&lt;br /&gt;
* Be prepared to summarize to the class your contributions to the background, contributions, and aims sections.&lt;br /&gt;
* If there is preliminary work that will help to make decisions about contributions and aims, please get started on it (and be ready to report on what you&#039;d like to do or what you have done).&lt;br /&gt;
&lt;br /&gt;
== Assignment 1 (out January 23, 2009) ==&lt;br /&gt;
* spend 10 hours adding to any part of the wiki you think is relevant&lt;br /&gt;
* by Monday noon add any potential readings.  If you&#039;ve got a tentative summary evaluation, go ahead and add it.  It&#039;s ok to edit folks summary evaluations, but try to make the result more accurate or precise without losing information.&lt;br /&gt;
* by Wednesday noon finish with any summary evaluation and also identify at least one as-relevant-as-possible reading as yours.  Put your name on that entry in the reading list as the &amp;quot;owner&amp;quot; so that there are no duplicates.&lt;br /&gt;
* by Wednesday 5pm -- select 2 additional relevant readings that are owned and that you will read by Friday and be prepared to discuss.  Put your name as a &amp;quot;discussant&amp;quot; in the reading list; there should be a max of two discussants per reading.&lt;br /&gt;
* by Friday class -- author a summary description, less than 250 words, in the wiki of how the reading you own relates to our project.  Be prepared to describe, in two minutes, how your reading relates to the project. Also be prepared for everyone in class to discuss after your description.  You may bring notes for yourself, but no slides.  The wiki page for your reading will be displayed while you talk.&lt;br /&gt;
* by Friday class -- read and be prepared to discuss the other two readings you choose.&lt;br /&gt;
* by Friday class -- make one more wiki page titled &amp;quot;&#039;&#039;&#039;&amp;lt;Last-Name&amp;gt; week 1&#039;&#039;&#039;&amp;quot; with a list of the keys for the citations you added, the readings  you summarized, the reading you presented, the two readings you were a discussant on, and any other readings you did in detail.&lt;br /&gt;
* Let me know if you have any kind of problems.  You should be spending right around 10 hours -- if that&#039;s a problem, let&#039;s talk.&lt;br /&gt;
* The [[../How Tos|How Tos]] page has some tips.  Edit or add as you find others.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Assignments&amp;diff=3056</id>
		<title>CS295J/Assignments</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Assignments&amp;diff=3056"/>
		<updated>2009-04-21T14:03:55Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Assignment 12 (out April 21, due April 24, 2009) ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Assignment 11 (out April 10, due April 17, 2009) ==&lt;br /&gt;
&lt;br /&gt;
=== Part A (due Tuesday at noon) ===&lt;br /&gt;
First, revise any contribution that you added to the proposal to reflect our converging view of what we are proposing. Given the evolution of the proposal since the contributions were originally created, if you wish to start from a fresh set of contributions, you can outline them [[CS295J/Contributions for class 12 |here]].&lt;br /&gt;
&lt;br /&gt;
Own one or more of the possible tasks below (enough to spend 8-10 hours on).  Edit this page to indicate your ownership and to describe enough details of what you&#039;ll do for the group to avoid duplication.  It&#039;s ok, for exmaple, for two folks to try using or extending CPM-GOMS with different extensions or a different application; indicating what your extension/application is, though would be helpful.  It&#039;s also fine to add tasks to the list.&lt;br /&gt;
&lt;br /&gt;
=== Part B (due class time) ===&lt;br /&gt;
&lt;br /&gt;
Revise any parts of the preliminary results that you have added.  That may include adding things that you would like to do, taking out pieces that don&#039;t fit into our worldview any longer, or revising things.  The &amp;quot;CPM-GOMS for gmail&amp;quot; project presented today, for example, should go into the preliminary work section to establish viability of CPM-GOMS in the context of a more modern application.&lt;br /&gt;
&lt;br /&gt;
Complete owned task(s).&lt;br /&gt;
&lt;br /&gt;
=== Possible Tasks (also consider those from last week -- add here if you pick one) ===&lt;br /&gt;
&lt;br /&gt;
* Possible additions to the proposed research&lt;br /&gt;
** Experiments with EEG to determine if they can inform modeling (&#039;&#039;&#039;Andrew Bragdon&#039;&#039;&#039; -- I will look into the EEG literature from the CHI community and determine what could be put in the proposal on this topic)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;CONCLUSION:&#039;&#039; Buying the BioSemi system used by Tan et. al. would cost about 50,000+ Euros.  I think that we should add this to the budget, and incorporate EEG measurements into our methodology.  Also, my conclusion is that making one giant equation which outputs only time may not be realistic.  We should refine what our theory can predict, and we should consider working memory load as a model of cognitive load.&lt;br /&gt;
&lt;br /&gt;
** Maybe similar experiments with muscle sensing or body motion&lt;br /&gt;
* Possible preliminary experiments&lt;br /&gt;
** Try some EEG&lt;br /&gt;
** Try some low-cost muscle tracking&lt;br /&gt;
** Try to decode some captured events&lt;br /&gt;
** Try to get some pupil-tracking thing going and sync&#039;ed with other events (I found an open-source eye-tracking program called [http://www.inference.phy.cam.ac.uk/opengazer/ OpenGazer], but it&#039;s written for linux and might require considerable work to port. Would someone with a linux machine like to try it out? Adam)&lt;br /&gt;
** Try some light-weight JavaScript-based mouse tracking (will see how far I can get in 10 hours) ([[User:E J Kalafarski|E J Kalafarski]] 15:16, 14 April 2009 (UTC))&lt;br /&gt;
* Other&lt;br /&gt;
** Get &amp;quot;Human information processor model&amp;quot; concept into related work; ernestine is canonical example, but other more recent work too (Steven,  Eric (Steven wrote core of the article, but I&#039;ll add to it as I come across other work))&lt;br /&gt;
** Revise proposal introduction again, emphasis on our new converging view ([[User:E J Kalafarski|E J Kalafarski]] 15:13, 14 April 2009 (UTC), [[User:Trevor|Trevor O&#039;Brien]], Eric)&lt;br /&gt;
** Will revise and expand &amp;quot;big world&amp;quot; significance and contributions, based on the output/architecture we discussed last week ([[User:E J Kalafarski|E J Kalafarski]] 15:15, 14 April 2009 (UTC), [[User:Trevor|Trevor O&#039;Brien]], Eric)&lt;br /&gt;
** Move 30-hour projects out of the contributions sections (Eric)&lt;br /&gt;
* CPM-GOMS&lt;br /&gt;
** Extend the cognition element in the CPM-GOMS model to account for dual-process theories of reasoning. (OWNER - Gideon; Jon)&lt;br /&gt;
** Extend the model by incorporating an &#039;Affect&#039; output, instead of just relying on time as the only dependent variable. (SUGGESTED BY - Gideon; Ian (I like this idea))&lt;br /&gt;
&lt;br /&gt;
== Assignment 10 (out April 3, due April 10, 2009) ==&lt;br /&gt;
&lt;br /&gt;
=== Part A (due Tuesday at noon) ===&lt;br /&gt;
Own one or more of the possible tasks below (enough to spend 8-10 hours on).  Edit this page to indicate your ownership and to describe enough details of what you&#039;ll do for the group to avoid duplication.  It&#039;s ok, for exmaple, for two folks to try using or extending CPM-GOMS with different extensions or a different application; indicating what your extension/application is, though would be helpful.  It&#039;s also fine to add tasks to the list.&lt;br /&gt;
&lt;br /&gt;
=== Part B (due class time) ===&lt;br /&gt;
&lt;br /&gt;
Complete owned task(s).&lt;br /&gt;
&lt;br /&gt;
=== Possible Tasks ===&lt;br /&gt;
&lt;br /&gt;
* read CPM-GOMS and consider as theme (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;, &#039;&#039;&#039;Eric&#039;&#039;&#039;, &#039;&#039;&#039;Jon&#039;&#039;&#039;, &#039;&#039;&#039;Gideon&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Steven Ellis|Steven]]&#039;&#039;&#039;, &#039;&#039;&#039;Ian&#039;&#039;&#039;)&lt;br /&gt;
* converge proposal (ID weaknesses &amp;amp; fix or propose fixes)&lt;br /&gt;
** make contributions and aims agree (&#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
** make contributions and aims compelling and significant&lt;br /&gt;
*** intro acceptable (merge the two best) (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will rewrite intro, merging the two best &amp;quot;in progress&amp;quot; intros, with an emphasis on the project&#039;s interdisciplinary aspect)&lt;br /&gt;
*** improves world, lots of people, in big ways, long time (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will brainstorm and add to &amp;quot;big picture&amp;quot; contributions, but someone else should brainstorm this as well, &#039;&#039;&#039;Eric&#039;&#039;&#039;)&lt;br /&gt;
*** extends knowledge, increased productivity (designers and users) (&#039;&#039;&#039;Eric&#039;&#039;&#039;)&lt;br /&gt;
*** bg &amp;amp; significance section consistent with contribs/aims (&#039;&#039;&#039;Ian&#039;&#039;&#039;)&lt;br /&gt;
* try CPM-GOMS, maybe w/ 1 extension (&#039;&#039;&#039;Gideon&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Steven Ellis|Steven]]&#039;&#039;&#039;)&lt;br /&gt;
* what&#039;s input and output (architecture)? (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will attempt to come up with an architecture that accommodates all seven proposed modules and presents a simple, useful final product, &#039;&#039;Jon: Will brainstorm b/c I&#039;ve been wondering about this for a long time&#039;&#039;, &#039;&#039;&#039;Adam&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
* Gajos paper (&#039;&#039;&#039;[[User:E J Kalafarski|E J Kalafarski]]&#039;&#039;&#039;: will read, consider architecture method, &#039;&#039;&#039;Adam&#039;&#039;&#039;, &#039;&#039;&#039;[[User:Trevor O&#039;Brien|Trevor]]&#039;&#039;&#039;)&lt;br /&gt;
&lt;br /&gt;
== Assignment 9 (out March 20, due April 3, 2009) ==&lt;br /&gt;
&lt;br /&gt;
Part A is what we talked about in class -- a revision of the introduction to re-evaluate the big picture.  Part B is the original assignment.  While I&#039;ve asked for both below, it may not be realistic given the time available.  Please make sure to finish A by class time and to do as much of B as you can.&lt;br /&gt;
&lt;br /&gt;
=== Part A, due noon Tuesday March 31 ===&lt;br /&gt;
&lt;br /&gt;
Refine your introduction to be consistent with the theme of bringing knowledge of cognitive and perceptual modeling to bear in a principled way on human-computer interface design.  Evaluate each of the suggested contributions for their impact, their relevance to the theme, and their cost.  Triage as appropriate to achieve the best overall proposal given the 5 years with 5 people scope.&lt;br /&gt;
&lt;br /&gt;
* [[CS295J/Proposal intros from class 9|Link to Intros]]&lt;br /&gt;
* [[CS295J/Triage for class 10|Link to Triages]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due at class time ===&lt;br /&gt;
&lt;br /&gt;
Complete your 30 hour preliminary work project and writeup in the proposal.  It should explicitly support the contributions of the proposal.  Consider this a hard, externally imposed deadline, i.e., you have to have something as finished as possible that could be reviewed by an outside reviewer.  Trim scope, if necessary, but finish!&lt;br /&gt;
&lt;br /&gt;
Pick another gap in the proposal and plug it.  Any section that you have &amp;quot;owned&amp;quot; should also be &amp;quot;final&amp;quot; and reviewable.&lt;br /&gt;
&lt;br /&gt;
====Gaps====&lt;br /&gt;
* &amp;lt;s&amp;gt;Significance (expand further with unified description of relatable components)&amp;lt;/s&amp;gt;&lt;br /&gt;
* Specific contributions &amp;gt; Multi-modal HCI (question marks, needs content)&lt;br /&gt;
* Background &amp;gt; Workflow analysis &amp;gt; Interface improvements (incomplete)&lt;br /&gt;
* Research plan (we may have to cut this section, it is woefully underdeveloped)&lt;br /&gt;
&lt;br /&gt;
Read [[CS295J/Literature to read for class 10|&amp;quot;must reads&amp;quot;]] -- add one if you want.&lt;br /&gt;
&lt;br /&gt;
== Assignment 8 (out March 13, due March 20, 2009) ==&lt;br /&gt;
=== Part A, due Tuesday noon ===&lt;br /&gt;
* Outline a coherent 250 word summary to a coherent proposal [[CS295J/Proposal intros from class 9|here]]&lt;br /&gt;
* Select a gap to own&lt;br /&gt;
** &#039;&#039;&#039;Andrew Bragdon&#039;&#039;&#039;:  Metawork Support Tool proposal; will examine integrating this into the main proposal vs. making a separate proposal.&lt;br /&gt;
** &#039;&#039;&#039;EJ&#039;&#039;&#039;: &amp;quot;Significance/intellectual merit&amp;quot; section is currently bare, I can take a stab at that.  I believe this can/needs to incorporate a gap Trevor identified, &amp;quot;mapping between individual contributions and centralized theme of the proposal;&amp;quot; I&#039;ll try to start to illustrate the relationships between our individual projects. [[User:E J Kalafarski|E J Kalafarski]] 12:20, 17 March 2009 (UTC)&lt;br /&gt;
**&#039;&#039;&#039;Jon&#039;&#039;&#039;: The &amp;quot;models of perception&amp;quot; section needs to be revised and expanded.  I&#039;ve changed &amp;quot;Gibsonianism&amp;quot; to &amp;quot;The Ecological Approach to Perception&amp;quot; and I&#039;ve added a section on &amp;quot;The Computational Approach to Perception&amp;quot;; I will update these for Friday.&lt;br /&gt;
**&#039;&#039;&#039;Gideon&#039;&#039;&#039;: More up-to-date background section on distributed cognition. I know Jon is doing this for the other areas, but I feel that this is very important.&lt;br /&gt;
** &#039;&#039;&#039;Trevor&#039;&#039;&#039;:  I&#039;ll take a pass through the &#039;&#039;Specific Aims&#039;&#039; section, which currently lacks specificity.   --- [[User:Trevor O&amp;amp;#39;Brien|Trevor O&amp;amp;#39;Brien]] 15:29, 17 March 2009 (UTC) &lt;br /&gt;
**&#039;&#039;&#039;Eric&#039;&#039;&#039;: I&#039;ll flesh out the &#039;&#039;Workflow Analysis&#039;&#039; section. What&#039;s there could be cleaned up, and more could be added, particularly on learning from interaction histories.&lt;br /&gt;
**&#039;&#039;&#039;Steven&#039;&#039;&#039;: There&#039;s currently no background (slash anything) on the topic of embodied models of cognition, I&#039;ll fill in some background on that.&lt;br /&gt;
* suggest 0.5 &amp;quot;must read&amp;quot; [[CS295J/Literature to read for class 9|papers]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due in class ===&lt;br /&gt;
* Finish writing the coherent 250 word summary to a coherent proposal [[CS295J/Proposal intros from class 9|still here]]&lt;br /&gt;
* Fill your gap&lt;br /&gt;
* read &amp;quot;must read&amp;quot;s for discussion w.r.t. proposal relevance&lt;br /&gt;
* get to another week&#039;s worth of results in your preliminary results -- make sure to be consistent with your summary!&lt;br /&gt;
* be prepared to tell us about those new results in class, tied to the intellectual claims&lt;br /&gt;
&lt;br /&gt;
== Assignment 7 (out March 6, due March 13, 2009) ==&lt;br /&gt;
=== Part A, due Tuesday noon ===&lt;br /&gt;
* review [[CS295J/Research proposal|proposal]] for&lt;br /&gt;
** intellectual contribution (1-2 paragraphs)&lt;br /&gt;
** major gaps (bullet list) (don&#039;t duplicate gaps already listed)&lt;br /&gt;
** review goes [[CS295J/Proposal reviews from class 8|here]]&lt;br /&gt;
** feel free to improve any part of the proposal rather than criticizing it, if you wish :-)&lt;br /&gt;
* suggest 0.5 &amp;quot;must read&amp;quot; [[CS295J/Literature to read for class 8|papers]]&lt;br /&gt;
&lt;br /&gt;
=== Part B, due in class ===&lt;br /&gt;
* read &amp;quot;must read&amp;quot;s for discussion w.r.t. proposal relevance&lt;br /&gt;
* get to another week&#039;s worth of results in your preliminary results&lt;br /&gt;
* be prepared to tell us about those new results in class, tied to the intellectual claims&lt;br /&gt;
&lt;br /&gt;
== Assignment 6 (out February 27, due March 6, 2009) ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* If you are not familiar with NIH proposals, skim through [http://vis.cs.brown.edu/docs/pdf/bib/Laidlaw-2005-DA2.pdf this active proposal] to get a sense of the sections.  The [http://vis.cs.brown.edu/docs/pdf/bib/NIH-2001-PHS.pdf guide to nih proposals] may also be helpful.&lt;br /&gt;
* Create some preliminary results&lt;br /&gt;
** Start the work you proposed in your poster for assignment 4&lt;br /&gt;
** Start the work someone else proposed :-)&lt;br /&gt;
** Create a more detailed critique-enabled workflow of a scientific user&lt;br /&gt;
** Working in pairs is fine; preferably, pair members should have different backgrounds&lt;br /&gt;
* Write them up as a subsection of the preliminary results section of the proposal.&lt;br /&gt;
** By Wednesday 9am have an outline of the section you will produce&lt;br /&gt;
*** This should be in past tense, as it will be when the work is done.&lt;br /&gt;
*** At the top level, it should state how the preliminary results enable the overall multi-year proposal or how they demonstrate feasibility of some questionable or risky part.&lt;br /&gt;
*** Check out the NIH proposal for examples, albeit in a different domain.&lt;br /&gt;
** By class have as much filled in from that outline as possible&lt;br /&gt;
* Add any &amp;quot;must reads&amp;quot; to the [[CS295J/Literature to read for class 6]] page.  You are &amp;quot;expected&amp;quot; to add 1/2 of a reference.&lt;br /&gt;
* Read and be prepared to discuss the &amp;quot;must reads&amp;quot;.&lt;br /&gt;
** Put in fictional placeholders for parts that are not done by Friday.&lt;br /&gt;
* By Wednesday 5pm e-mail constructive comments about at least 1/2 of the outlines to the entire class&lt;br /&gt;
* Success criteria for this assignment&lt;br /&gt;
*# Proposal section will demonstrate a feasibility or prerequisite for interesting research&lt;br /&gt;
*# Results by class are complete and concrete enough that they are interesting, even though some parts may be missing&lt;br /&gt;
&lt;br /&gt;
== Assignment 5 (out February 20, due February 27, 2009) ==&lt;br /&gt;
# Videotape a 15-20 minute interactive session with a user doing a visually challenging task for which they are at least an advanced beginner (ie, they don&#039;t have to look up what to do, but they have not yet internalized the operations and made them subconscious).  Analyze the session and report your observations and conclusions.&lt;br /&gt;
&lt;br /&gt;
== Assignment 4 (out February 13, due February 20, 2009) ==&lt;br /&gt;
&lt;br /&gt;
# Be prepared to decide on a subset of applications to critique.  Flesh out at least one potential application in [[CS295J/Application Critiques]], including a specific workflow.  Modify any others you want, particularly in terms of arguments for or against proceeding with them.  These should be completed by Thursday noon.&lt;br /&gt;
#* Some possibilities: Google scholar, Mathematica, Tableau, Google notebook, Matlab, paper, media wiki, Ensight Gold, AVS, VisTrails.  Note that a given piece of software could be represented more than once with a different workflow.&lt;br /&gt;
#* Possible criteria: main purpose is analysis, perhaps scientific; interative; scientific; amenable to cognition-driven improvemennts; interesting; fun&lt;br /&gt;
# Bring your ranking of the subset of (application+workflow)s we should critique&lt;br /&gt;
# Be prepared to decide on the [[CS295J/Model elements]] of cognition we will emulate to critique each application.  Revise some part of this list so that it can be used as a concrete basis for evaluation.&lt;br /&gt;
# Add any &amp;quot;must reads&amp;quot; to the [[CS295J/Literature to read for class 5]] page.  You are &amp;quot;expected&amp;quot; to add 1/3 of a reference.&lt;br /&gt;
# Read the &amp;quot;must reads&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
== Assignment 3 (out February 6, due February 13) ==&lt;br /&gt;
* Flesh out and support a contribution within the proposal&lt;br /&gt;
** Add any new references to the literature section.&lt;br /&gt;
** Many of our readings date from 8+ years ago; check to make sure your contribution hasn&#039;t already been done.&lt;br /&gt;
** If any of the new references are &amp;quot;must reads&amp;quot;, add to the [[CS295J/Literature to read for class 4]] (2/13/09) page.  You are &amp;quot;expected&amp;quot; to add 1/2 of a reference.&lt;br /&gt;
** Estimate the impact of the contribution.&lt;br /&gt;
** Estimate the risk and costs of the contribution.&lt;br /&gt;
** Propose a 3-week (30 hour) result that you could create to demonstrate feasibility.&lt;br /&gt;
** Bring a printout of your contribution concept to class.  It should be legible from 2 meters away, so use big text.  You can hand-write it on posterboard or paste together printouts.&lt;br /&gt;
* Bring a list of holes in the proposal -- e.g., &amp;quot;background section needs something on workflow capture.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Assignment 2 (out January 30, 2009) ==&lt;br /&gt;
* Refine literature short summaries to include the relationship to our project.&lt;br /&gt;
* Draft by tuesday noon a subsection in the background section of the [[CS295J/Research proposal]], making sure to include citations to the relevant materials in the literature page.  Add any new references to the literature page.&lt;br /&gt;
* Read and comment/edit by Thursday noon all background sections&lt;br /&gt;
* Revise your background section by Friday noon (so David can print before class)&lt;br /&gt;
&lt;br /&gt;
* Add to or refine one or more specific contribution in the [[CS295J/Research proposal]]; each contribution must have a list of ways it can be demonstrated.  Some will become part of the preliminary results, others will be parts of the future work that will be proposed.&lt;br /&gt;
* Add to or refine one or more specific aim in the [[CS295J/Research proposal]]; make it consistent with the contribution you added.&lt;br /&gt;
&lt;br /&gt;
* Identify the one additional most important paper for us to read this week also by Tuesday noon; be prepared to summarize relevance in 2 minutes in class&lt;br /&gt;
* Read those &amp;quot;most important&amp;quot; papers for class discussion.&lt;br /&gt;
&lt;br /&gt;
* Be prepared to summarize to the class your contributions to the background, contributions, and aims sections.&lt;br /&gt;
* If there is preliminary work that will help to make decisions about contributions and aims, please get started on it (and be ready to report on what you&#039;d like to do or what you have done).&lt;br /&gt;
&lt;br /&gt;
== Assignment 1 (out January 23, 2009) ==&lt;br /&gt;
* spend 10 hours adding to any part of the wiki you think is relevant&lt;br /&gt;
* by Monday noon add any potential readings.  If you&#039;ve got a tentative summary evaluation, go ahead and add it.  It&#039;s ok to edit folks summary evaluations, but try to make the result more accurate or precise without losing information.&lt;br /&gt;
* by Wednesday noon finish with any summary evaluation and also identify at least one as-relevant-as-possible reading as yours.  Put your name on that entry in the reading list as the &amp;quot;owner&amp;quot; so that there are no duplicates.&lt;br /&gt;
* by Wednesday 5pm -- select 2 additional relevant readings that are owned and that you will read by Friday and be prepared to discuss.  Put your name as a &amp;quot;discussant&amp;quot; in the reading list; there should be a max of two discussants per reading.&lt;br /&gt;
* by Friday class -- author a summary description, less than 250 words, in the wiki of how the reading you own relates to our project.  Be prepared to describe, in two minutes, how your reading relates to the project. Also be prepared for everyone in class to discuss after your description.  You may bring notes for yourself, but no slides.  The wiki page for your reading will be displayed while you talk.&lt;br /&gt;
* by Friday class -- read and be prepared to discuss the other two readings you choose.&lt;br /&gt;
* by Friday class -- make one more wiki page titled &amp;quot;&#039;&#039;&#039;&amp;lt;Last-Name&amp;gt; week 1&#039;&#039;&#039;&amp;quot; with a list of the keys for the citations you added, the readings  you summarized, the reading you presented, the two readings you were a discussant on, and any other readings you did in detail.&lt;br /&gt;
* Let me know if you have any kind of problems.  You should be spending right around 10 hours -- if that&#039;s a problem, let&#039;s talk.&lt;br /&gt;
* The [[../How Tos|How Tos]] page has some tips.  Edit or add as you find others.&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=Talk:CS295J/Research_proposal_(draft_2)&amp;diff=3055</id>
		<title>Talk:CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=Talk:CS295J/Research_proposal_(draft_2)&amp;diff=3055"/>
		<updated>2009-04-21T11:22:13Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;People in the class to assign tasks:&lt;br /&gt;
* [[User:Adam Darlow | Adam Darlow]]&lt;br /&gt;
* [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
* [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
* [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
* [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
* [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
* [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
* [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
* [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
TODO:&lt;br /&gt;
* More evenly distribute the work amongst people in the class&lt;br /&gt;
* Clean up the intro&lt;br /&gt;
* Add an intro to the methodology section&lt;br /&gt;
* Breakdown the related work section into subsections?&lt;br /&gt;
* Add description to &#039;Generalizing User Traces&#039; section&lt;br /&gt;
* Add a figure?&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=Talk:CS295J/Research_proposal_(draft_2)&amp;diff=3054</id>
		<title>Talk:CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=Talk:CS295J/Research_proposal_(draft_2)&amp;diff=3054"/>
		<updated>2009-04-21T11:21:42Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;People in the class to assign tasks:&lt;br /&gt;
* [[User:Adam Darlow | Adam Darlow]]&lt;br /&gt;
* [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
* [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
* [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
* [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
* [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
* [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
* [[User:Gideon Goldin | Gideon Goldin]]&lt;br /&gt;
* [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
TODO:&lt;br /&gt;
* More evenly distribute the work amongst people in the class&lt;br /&gt;
* Clean up the intro&lt;br /&gt;
* Add an intro to the methodology section&lt;br /&gt;
* Breakdown the related work section into subsections?&lt;br /&gt;
* Add description to &#039;Generalizing User Traces&#039; section&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3053</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3053"/>
		<updated>2009-04-21T11:15:09Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Cognitive/HCI Guidelines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module weights various interface guidelines from the HCI and cognitive science literature, and uses these weights to provide suggested improvements for the given interface. It identifies interface elements that are detrimental to user performance and suggests effective alternatives.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given the history of user interactions, we may be able to improve user performance by modifying the interface so that frequently performed interactions from a given state are more readily accessible to the user. This module recommendations such modifications based on user traces.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3052</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3052"/>
		<updated>2009-04-21T11:14:39Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Cognitive/HCI Guidelines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module weights various interface guidelines from HCI and cognitive science, and uses these weights to provide suggested improvements for the given interface. It identifies interface elements that are detrimental to user performance and suggests effective alternatives.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given the history of user interactions, we may be able to improve user performance by modifying the interface so that frequently performed interactions from a given state are more readily accessible to the user. This module recommendations such modifications based on user traces.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3051</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3051"/>
		<updated>2009-04-21T11:07:01Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Interaction Prediction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given the history of user interactions, we may be able to improve user performance by modifying the interface so that frequently performed interactions from a given state are more readily accessible to the user. This module recommendations such modifications based on user traces.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3050</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3050"/>
		<updated>2009-04-21T11:04:57Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Interaction Prediction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given the history of user interactions, we may be able to improve user performance by modifying the interface so that frequently performed interactions from a given state are more readily accessible to the user.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3049</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3049"/>
		<updated>2009-04-21T10:47:17Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Preliminary Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
* Owner: [[User:Ian Spector | Ian Spector]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3048</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3048"/>
		<updated>2009-04-21T10:46:41Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Preliminary Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Initially, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3047</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3047"/>
		<updated>2009-04-21T10:46:28Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Preliminary Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;To start, let&#039;s make up some fictional (but reasonable) preliminary results that we&#039;d like to see and think we can accomplish before submitting the proposal.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3046</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3046"/>
		<updated>2009-04-21T10:44:31Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* [Criticisms] */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Any criticisms or questions we have regarding the proposal can go here.&#039;&#039;&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
	<entry>
		<id>http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3045</id>
		<title>CS295J/Research proposal (draft 2)</title>
		<link rel="alternate" type="text/html" href="http://vrl.cs.brown.edu/wiki/index.php?title=CS295J/Research_proposal_(draft_2)&amp;diff=3045"/>
		<updated>2009-04-21T10:36:56Z</updated>

		<summary type="html">&lt;p&gt;Eric Sodomka: /* Interruptions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Introduction =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
We propose a framework for interface evaluation and recommendation that integrates behavioral models and design guidelines from both cognitive science and HCI. Our framework behaves like a committee of specialized experts, where each expert provides its own assessment of the interface, given its particular knowledge of HCI or cognitive science. For example, an expert may provide an evaluation based on the GOMS method, Fitts&#039;s law, Maeda&#039;s design principles, or cognitive models of learning and memory. An aggregator collects all of these assessments and weights the opinions of each expert, and outputs to the developer a merged evaluation score and a weighted set of recommendations.&lt;br /&gt;
&lt;br /&gt;
Systematic methods of estimating human performance with computer interfaces are used only sparsely despite their obvious benefits, the reason being the overhead involved in implementing them. In order to test an interface, both manual coding systems like the GOMS variations and user simulations like those based on ACT-R/PM and EPIC require detailed pseudo-code descriptions of the user workflow with the application interface. Any change to the interface then requires extensive changes to the pseudo-code, a major problem because of the trial-and-error nature of interface design. Updating the models themselves is even more complicated. Even an expert in CPM-GOMS, for example, can&#039;t necessarily adapt it to take into account results from new cognitive research.&lt;br /&gt;
&lt;br /&gt;
Our proposal makes automatic interface evaluation easier to use in several ways. First of all, we propose to divide the input to the system into three separate parts, functionality, user traces and interface. By separating the functionality from the interface, even radical interface changes will require updating only that part of the input. The user traces are also defined over the functionality so that they too translate across different interfaces. Second, the parallel modular architecture allows for a lower &amp;quot;entry cost&amp;quot; for using the tool. The system includes a broad array of evaluation modules some of which are very simple and some more complex. The simpler modules use only a subset of the input that a system like GOMS or ACT-R would require. This means that while more input will still lead to better output, interface designers can get minimal evaluations with only minimal information. For example, a visual search module may not require any functionality or user traces in order to determine whether all interface elements are distinct enough to be easy to find. Finally, a parallel modular architecture is much easier to augment with relevant cognitive and design evaluations.&lt;br /&gt;
&lt;br /&gt;
= Overview of Contributions =&lt;br /&gt;
* Owners: [[User:Adam Darlow | Adam Darlow]], [[User:Eric Sodomka | Eric Sodomka]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Note: For reference, this is the aggregate set contributions from last week. Maybe we can edit/add/remove from this as needed.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Design and user-study evaluation of novel techniques for collecting and filtering user traces with respect to user goals.&lt;br /&gt;
* Extensible, low-cost architecture for integrating pupil-tracking, muscle-activity monitoring, and auditory recognition with user traces in existing applications.&lt;br /&gt;
* System for isolating cognitive, perceptual, and motor tasks from an interface design to generate CPM_GOMS models for analysis.&lt;br /&gt;
* Design and quantitative evaluation of semi-automated techniques for extracting critical paths from an existing CPM_GOMS model.&lt;br /&gt;
* Novel algorithm for analyzing and optimizing critical paths based on established research in cognitive science.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements, based on a &#039;&#039;unified matrix&#039;&#039; of cognitive principles and heuristic design guidelines.&lt;br /&gt;
* The creation of a language for &#039;&#039;abstractly representing user interfaces&#039;&#039; in terms of the layout of graphical components and the functional relationships between these components.&lt;br /&gt;
* A system for &#039;&#039;generating interaction histories&#039;&#039; within user interfaces to facilitate individual and collaborative scientific discovery, and to enable researchers to more easily document and analyze user behavior.&lt;br /&gt;
* A system that takes user traces and creates a GOMS model that decomposes user actions  into various cognitive, perceptual, and motor control tasks. &lt;br /&gt;
* The development of other evaluation methods using various cognitive/HCI models and guidelines.&lt;br /&gt;
* A design tool that can provide a designer with recommendations for interface improvements. These recommendations can be made for a specific type of user or for the average user, as expressed by a utility function.&lt;br /&gt;
&lt;br /&gt;
= Background / Related Work =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Methodology =&lt;br /&gt;
&#039;&#039;&#039;TODO: Add some intro paragraph here.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Collecting User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Given an interface, our first step is to run users on the interface and log these user interactions. We want to log actions at a sufficiently low level so that a GOMS model can be generated from the data. When possible, we&#039;d also like to log data using additional sensing technologies, such as pupil-tracking, muscle-activity monitoring and auditory recognition; this information will help to analyze the explicit contributions of perception, cognition and motor skills with respect to user performance.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Generalizing User Traces ==&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
== Evaluation and Recommendation via Modules ==&lt;br /&gt;
* Owners: [[User:E J Kalafarski | E J Kalafarski]], [[User:Jon Ericson | Jon Ericson]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This section describes the aggregator, which takes the output of multiple independent modules and aggregates the results to provide (1) an evaluation and (2) recommendations for the user interface. We should explain how the aggregator weights the output of different modules (this could be based on historical performance of each module, or perhaps based on E.J.&#039;s cognitive/HCI guidelines).&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Sample Modules ==&lt;br /&gt;
&lt;br /&gt;
=== CPM-GOMS ===&lt;br /&gt;
* Owners: [[User:Gideon Goldin | Gideon Goldin]], [[User:Steven Ellis | Steven Ellis]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This module will provide interface evaluations and suggestions based on a CPM-GOMS model of cognition for the given interface. It will provide a quantitative, predictive, cognition-based parameterization of usability. From empirically collected data, user trajectories through the model (critical paths) will be examined, highlighting bottlenecks within the interface, and offering suggested alterations to the interface to induce more optimal user trajectories.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Cognitive/HCI Guidelines ===&lt;br /&gt;
* Owner: [[User:E J Kalafarski | E J Kalafarski]]&lt;br /&gt;
&lt;br /&gt;
=== Fitts&#039;s Law ===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;This simple module will use Fitts&#039;s Law to provide interface evaluations and recommendations.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interruptions ===&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;While most usability testing focuses on low-level task performance, there is also previous work suggesting that users also work at a higher, working sphere level. This module attempts to evaluate a given interface with respect to these higher-level considerations, such as task switching.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=== Interaction Prediction ===&lt;br /&gt;
* Owner: [[User:Trevor O&#039;Brien | Trevor O&#039;Brien]]&lt;br /&gt;
&lt;br /&gt;
= Preliminary Results =&lt;br /&gt;
&lt;br /&gt;
= [Criticisms] =&lt;br /&gt;
* Owner: [[User:Andrew Bragdon | Andrew Bragdon]]&lt;/div&gt;</summary>
		<author><name>Eric Sodomka</name></author>
	</entry>
</feed>