24
University of British Columbia Software Analysis and Testing Lab Technical Report Series Understanding JavaScript Event-Based Interactions Saba Alimadadi, Sheldon Sequeira, Ali Mesbah, and Karthik Pattabiraman Report UBC-SALT-2014-001

Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

  • Upload
    others

  • View
    45

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

University of British ColumbiaSoftware Analysis and Testing LabTechnical Report Series

Understanding JavaScript Event-Based Interactions

Saba Alimadadi, Sheldon Sequeira, Ali Mesbah, and Karthik Pattabiraman

Report UBC-SALT-2014-001

Page 2: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

UBC-SALT-2014-001

Published, produced and distributed by:

Software Analysis and Testing (SALT) LabDepartment of Electrical and Computer EngineeringFaculty of Applied ScienceUniversity of British Columbia5500–2332 Main MallVancouver BC V6T 1Z4Canada

SALT Technical Reports:http://salt.ece.ubc.ca

Note: This paper is a pre-print of:

Saba Alimadadi, Sheldon Sequeira, Ali Mesbah, and Karthik Pattabiraman. Understanding JavaScriptEvent-based Interactions. In Proceedings of the ACM/IEEE International Conference on Software Engi-neering (ICSE). 2014.

c© copyright 2014, by the authors of this report. Software Analysis and Testing Lab, Department of Elec-trical and Computer Engineering, Faculty of Applied Science, University of British Columbia. All rightsreserved. No part of this series may be reproduced in any form or by any means without prior writtenpermission of the authors.

Page 3: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Understanding JavaScript Event-Based Interactions

Saba Alimadadi Sheldon Sequeira Ali Mesbah Karthik PattabiramanElectrical and Computer Engineering

University of British ColumbiaVancouver, BC, Canada

{saba, sheldon, amesbah, karthikp}@ece.ubc.ca

ABSTRACTWeb applications have become one of the fastest growingtypes of software systems today. Despite their popularity,understanding the behaviour of modern web applications isstill a challenging endeavour for developers during develop-ment and maintenance tasks. The challenges mainly stemfrom the dynamic, event-driven, and asynchronous natureof the JavaScript language. We propose a generic techniquefor capturing low-level event-based interactions in a web ap-plication and mapping those to a higher-level behaviouralmodel. This model is then transformed into an interactivevisualization, representing episodes of triggered causal andtemporal events, related JavaScript code executions, andtheir impact on the dynamic DOM state. Our approach,implemented in a tool called Clematis, allows developers toeasily understand the complex dynamic behaviour of theirapplication at three different semantic levels of granularity.The results of our industrial controlled experiment show thatClematis is capable of improving the task accuracy by 61%,while reducing the task completion time by 47%.

Categories and Subject DescriptorsD.2.5 [Software Engineering]: Testing and Debugging-Tracing; D.2.7 [Software Engineering]: Distribution, Main-tenance, and Enhancement

General TermsDesign, Algorithms, Experimentation

KeywordsProgram comprehension, event-based interactions, JavaScript

1. INTRODUCTIONJavaScript is widely used today to create interactive mod-

ern web applications that replace many traditional desktopapplications. However, understanding the behaviour of webapplications is a challenging endeavour for developers [18,23]. Program comprehension is known to be an essentialstep in software engineering, consuming up to 50% [7] of theeffort in software maintenance and analysis activities.

First, the weakly-typed and highly-dynamic nature of Java-Script makes it a particularly difficult language to analyze.Second, JavaScript code is extensively used to seamlesslymutate the Document Object Model (DOM) at runtime.This dynamic interplay between two separate entities, name-ly JavaScript and the DOM, can become quite complex to

follow [17]. Third, JavaScript is an event-driven language al-lowing developers to register various event listeners on DOMnodes. While most events are triggered by user actions, tim-ing events and asynchronous callbacks can be fired with nodirect input from the user. To make things even more com-plex, a single event can propagate on the DOM tree andtrigger multiple listeners according to the event capturingand bubbling properties of the event model [21].

Unfortunately, despite its importance and challenges, the-re is currently not much research dedicated to supportingprogram comprehension for web applications [9]. Populartools, such as Firebug and Chrome DevTools, are limited intheir capabilities to support web developers effectively.

In this paper, we present a generic, non-intrusive tech-nique, called Clematis, for supporting web application com-prehension. Through a combination of automated Java-Script code instrumentation and transformation, we capturea detailed trace of a web application’s behaviour during aparticular user session. Our technique transforms the traceinto an abstract behavioural model, preserving temporal andcausal relations within and between involved components.The model is then presented to the developers as an interac-tive visualization that depicts the creation and flow of trig-gered events, the corresponding executed JavaScript func-tions, and the mutated DOM nodes, within each episode.

To the best of our knowledge, we are the first to provide ageneric technique for capturing low-level event-based inter-actions in a JavaScript application , and mapping and visu-alizing those interactions as higher-level behavioural models.Our work makes the following key contributions:

• We propose a generic technique for capturing and pre-senting the complex dynamic behaviour of web appli-cations. In particular, our technique:

– Captures the consequences of JavaScript and DOMevents in terms of the executed JavaScript code,including the functions that are called indirectlythrough event propagation on the DOM tree.

– Extracts the source-and-target relations for asyn-chronous events, i.e., timing events and XML-HttpRequest requests/callbacks.

– Identifies and tracks mutations to the DOM re-sulting from each event.

• We build a novel model for capturing the event-driveninteractions as well as an interactive, visual interfacesupporting the comprehension of the program throughthree different semantic levels of zooming granularity.

• We implement our technique in a generic open sourcetool called Clematis [1], which (1) does not modify the

1

Page 4: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

web browser, (2) is independent of the server technol-ogy, and (3) requires no extra effort from the developerto use.

• We empirically evaluate Clematis through two con-trolled experiments comprising 34 users in total. Oneof the studies is carried out in a lab environment, whilethe other is carried out in an industrial setting. Theresults of the industrial experiment show that Clema-tis can reduce the task completion time by 47%, whileimproving the accuracy by 61%.

2. RELATED WORKUI Feature Location. Li and Wohlstadter [12] present atool called Script Insight to locate the implementation of aDOM element in JavaScript code. Similarly, Maras et al.[13, 14] propose a technique for deriving the implementa-tion of a UI feature on the client side. While similar to ourwork at a high level, in these approaches the user needs toselect a visible DOM element and its relevant behaviour inorder to investigate its functionality. This manual effort caneasily frustrate the user in large applications. Further, thesetechniques are not concerned with capturing event-based in-teractions. Finally, the model they derive and present to theuser contains low-level information and noise, which can ad-versely influence program comprehension.

Capture and Replay. Related to our work are ‘captureand replay’ techniques for web applications [4, 5, 15, 16, 20].The goal in most of these techniques is to find a deterministicway of replaying the same set of user events for debuggingpurposes. Instead of simply replaying recorded events, ourapproach aims at detecting causal and temporal event-basedinteractions and linking them to their impact on JavaScriptcode execution and DOM mutations.

Visualization. There are many tools that use visualiza-tion to improve the process of understanding the behaviourof software applications. For instance, Extraviz [8] visual-izes dynamic traces of Java applications to assist with pro-gram comprehension tasks. However, their approach doesnot concern itself with building a model of the web applica-tion, while ours does.

Zaidman et al. [23] propose a Firefox add-on called FireDe-tective, which captures and visualizes a trace of executionon both the client and the server side. Their goal is to makeit easier for developers to understand the link between clientand server components, which is different from our approachwhich aims to make it easier for developers to understandthe client-side behaviour of the web application.

FireCrystal [18] is another Firefox extension that storesthe trace of a web application in the browser. It then vi-sualizes the events and changes to the DOM in a timeline.FireCrystal records the execution trace selectively similar toour work. But unlike Clematis, FireCrystal does not cap-ture the details about the execution of JavaScript code orasynchronous events. Another limitation of FireCrystal isthat it does not link the triggering of events with the dy-namic behaviour of the application, as Clematis does. Dy-naRIA [3] focuses on investigating the structural and qualityaspect of the code. While DynaRIA captures a trace of theweb application, Clematis facilitates the process of com-prehending the dynamic behaviour using a high-level modeland visualization based on a semantically partitioned trace.

1 <BODY>2 <FIELDSET class="registration">3 Email: <INPUT type="text" id="email"/>4 <BUTTON id="submitBtn">Submit </BUTTON >5 <DIV id="regMsg"></DIV>6 </FIELDSET >7 </BODY>

Figure 1: Initial DOM state of the running example.

3. CHALLENGES AND MOTIVATIONModern web applications are largely event-driven. Their

client-side execution is normally initiated in response to auser-action triggered event, a timing event, or the receiptof an asynchronous callback message from the server. As aresult, web developers encounter many program comprehen-sion challenges in their daily development and maintenanceactivities. We use an example, presented in Figures 1–2, toillustrate these challenges. Note that this is a simple exam-ple and these challenges are much more potent in large andcomplex web applications.

Challenge 1: Event Propagation. The DOM eventmodel [21] makes it possible for a single event, fired on aparticular node, to propagate through the DOM tree hierar-chy and indirectly trigger a series of other event-handlers at-tached to other nodes. There are typically two types of thisevent propagation in web applications; (1) with bubbling en-abled, an event first triggers the handler of the deepest childelement on which the event was fired, and then it bubbles upand triggers the parents’ handlers. (2) when capturing is en-abled, the event is first captured by the parent element andthen passed to the event handlers of children, with the deep-est child element being the last. Hence, a series of lower-levelevent-handlers, executed during the capturing and bubblingphases, may be triggered by a single user action. The ex-istence or the ordering of these handlers is often inferredmanually by the developer, which becomes more challeng-ing as the size of the code/DOM tree increases.Example. Consider the sample code shown in Figures 1–2.

Figure 1 represents the initial DOM structure of the appli-cation. It mainly consists of a fieldset containing a set ofelements for the user to enter their email address to be reg-istered for a service. The JavaScript code in Figure 2 partlyhandles this submission. When the user clicks the submit

button, a message appears indicating that the submissionwas successful. This message is displayed from within theevent-handler submissionHandler() (line 7), which is at-tached to the button on line 2 of Figure 2. However, after afew seconds, the developer observes that the message unex-pectedly starts to fade out. In the case of this simple exam-ple, she can read the whole code and find out that the clickon the submit button has bubbled up to its parent element,namely fieldset. Closer inspection reveals that fieldset’sanonymous handler function is responsible for changing thevalue of the same DOM element through a setTimeout func-tion (lines 3–5 in Figure 2). In a more complex application,the developer may be unaware of the existence of the par-ent element, its registered handlers, or the complex eventpropagation mechanisms such as bubbling and capturing.

Challenge 2: Asynchronous Events. Web browsers pro-vide a single thread for web application execution. To cir-cumvent this limitation and build rich responsive web ap-plications, developers take advantage of the asynchronous

2 UBC-SALT-2014-001

Page 5: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

1 $(document).ready(function () {2 $( '#submitBtn ').click(submissionHandler);3 $( ' fieldset.registration ').click(function () {4 setTimeout(clearMsg , 3000);5 }); });6 ...7 function submissionHandler(e) {8 $( '#regMsg ').html("Submitted!");9 var email = $( '#email ').val();

10 if (isEmailValid(email)) {11 informServer(email);12 $( '#submitBtn ').attr("disabled", true);13 }14 }15 ...16 function informServer(email) {17 $.get( '/register/ ' , { email }, function(data) {18 $( '#regMsg ').append(data);19 });20 return;21 }22 ...23 function clearMsg () {$( '#regMsg ').fadeOut (2000) ;}

Figure 2: JavaScript code of the running example.

capabilities offered by modern browsers, such as timeoutsand XMLHttpRequest (XHR) calls. Asynchronous program-ming, however, introduces an extra layer of complexity inthe control flow of the application and adversely influencesprogram comprehension.

Timeouts: Events can be registered to fire after a cer-tain amount of time or at certain intervals in JavaScript.These timeouts often have asynchronous callbacks that areexecuted when triggered. In general, there is no easy wayto link the callback of a timeout to its source, which is im-portant to understand the program’s flow of execution.

XHR Callbacks: XHR objects are used to exchange dataasynchronously with the server, without requiring a pagereload. Each XHR goes through three main phases: open,send, and response. These three phases can be scatteredthroughout the code. Further, there is no guarantee on thetiming and the order of XHR responses from the server. Asin the case of timeouts, mapping the functionality triggeredby a server response back to its source request is a challeng-ing comprehension task for developers.

Example. Following the running example, the developermay wish to further investigate the unexpected behaviour:the message has faded out without a direct action from thedeveloper. The questions that a developer might ask at thispoint include: “What exactly happened here?” and “Whatwas the source of this behaviour?”. By reviewing the code,she can find out that the source of this behaviour was theexpiration of a timeout that was set in line 4 of Figure 2 bythe anonymous handler defined in lines 3–5. However thecallback function, defined on line 23 of Figure 2, executesasynchronously and with a delay, long after the execution ofthe anonymous handler function has terminated. While inthis case, the timing behaviour can be traced by reading thecode, this approach is not practical for large applications. Asimilar problem exists for asynchronous XHR calls. For in-stance, the anonymous callback function of the request sentin the informServer function (line 17, Figure 2) updates theDOM (line 18).

Challenge 3: Implications of Events. Another chal-lenge in understanding the flow of web applications lies inunderstanding the consequences of (in)directly triggered ev-

ents. Handlers for a (propagated) DOM event, and callbackfunctions of timeouts and XHR requests, are all JavaScriptfunctions. Any of these functions may change the observablestate of the application by modifying the DOM. Currently,developers need to read the code and make the connectionsmentally to see how an event affects the DOM state, whichis quite challenging. In addition, there is no easy way of pin-pointing the dynamic changes made to the DOM state as aresult of event-based interactions. Inferring the implicationsof events is, therefore, a significant challenge for developers.Example. After the submitBtn button is clicked in the

running example, a confirmation message will appear on-screen and disappear shortly thereafter (lines 8&23, Figure2). Additionally, the attributes of the button are altered todisable it (line 12). It can be difficult to follow such DOM-altering features in an application’s code.

4. APPROACHIn this section, we describe our approach for address-

ing the challenges mentioned in the previous section. Anoverview of the overall process is depicted in Figure 3, whichconsists of the following main steps:

• First, our technique captures a fine-grained trace of allsemantically related event-based interactions within aweb application’s execution, in a particular user ses-sion. The collection of this detailed trace is enabledthrough a series of automated JavaScript transforma-tions (Section 4.1).

• Next, a behavioural model is extracted from the infor-mation contained within the trace. The model struc-tures the captured trace and identifies the implicitcausal and temporal relationships between various eve-nt-based interactions (Section 4.2).

• Finally, based on the inferred behavioural model, ourapproach generates an interactive (web-based) user in-terface, visualizing and connecting all the pieces to-gether. This interactive visualization assists develop-ers during their web application comprehension andmaintenance tasks (Section 4.3).

We describe each step further below.

4.1 JavaScript Transformation and TracingTo automatically trace semantically related event-based

interactions and their impact, we transform the JavaScriptcode on-the-fly. Our approach generates a trace comprisingmultiple trace units. A trace unit contains information ac-quired through the interception of a particular event-basedinteraction type, namely, DOM events, timing events, XHRcalls and callbacks, function calls, and DOM mutations. Theobtained trace is used to build a behavioural model (as de-scribed in subsection 4.2).

Interposing on DOM Events. There are two ways eventlisteners can be bound to a DOM element in JavaScript.The first method is programatically using the DOM Level1 e.click=handler or DOM Level 2 e.addEventListener

methods [21] in JavaScript code. To record the occurrence ofsuch events, our technique replaces the default registrationof these JavaScript methods such that all event listeners arewrapped within a tracing function that logs the occurringevent’s time, type, and target.

3

Page 6: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

ProxyServer Browser

JavaScript Transformation

Transformaionto Java DS

Episodes & Story Creation

Linking DOMMutations

BehaviouralModel

GenerationVisualization

DOM events

Timeouts

XHR callbacks

Function calls

DOM mutations

JSON

Trace Collection

1

2

3

4

567

Figure 3: A processing overview of our approach.

The second and more traditional way to register an eventlistener is inline in the HTML code, e.g., <DIV onclick=‘-

handler();’>. The effect of this inline assignment is seman-tically the same as the first method. Our technique inter-poses on inline-registered listeners by removing them fromtheir associated HTML elements, annotating the HTML el-ements, and re-registering them using the substituted ad-

dEventListener function. This way we can handle themsimilarly to the programmatically registered event handlers.

Capturing Timeouts and XHRs. For tracing timeouts,we replace the browser’s setTimeout() method and the call-back function of each timeout with wrapper functions, whichallow us to track the instantiation and resolution of eachtimeout. A timeout callback usually happens later and trig-gers new behaviour, and thus we consider it as a sepa-rate component than a setTimeout(). We link these to-gether through a timeout_id and represent them as a causalconnection later. In our model, we distinguish betweenthree different components for the open, send, and response

phases of each XHR object. We intercept each componentby replacing the XMLHttpRequest object of the browser. Thenew object captures the information about each componentwhile preserving its functionality.

Recording Function Traces. To track the flow of execu-tion within a JavaScript-based application, we instrumentthree code constructs, namely function declarations, returnstatements, and function calls. Each of these code constructsare instrumented differently, as explained below.

Function Declarations: Tracing code is automatically add-ed to each function declaration allowing us to track the flowof control between developer-defined functions by loggingthe subroutine’s name, arguments, and line number. In caseof anonymous functions, the line number and source file ofthe subroutine are used as supplementary information toidentify the executed code.

As this communication is done each time a function is ex-ecuted, argument values are recorded dynamically at thecost of a slight overhead. Figure 4 contains the simpleclearMsg() JavaScript function from the running exampleshown in Figure 2 (line 22), which has been instrumentedto record both the beginning and end of its execution (lines2 and 4). Figure 4 contains a simple JavaScript function

1 function clearMsg () {2 send(JSON.stringify ({ messageType: "←↩

FUNCTION_ENTER", fnName: "clearMsg", args: ←↩null , ...}));

4 $( '#submissionMsg ').fadeOut (2000);

6 send(JSON.stringify ({ messageType: "←↩FUNCTION_EXIT", fnName: "clearMsg", ...}));

7 }

Figure 4: Instrumented JavaScript function decla-ration.

which has been instrumented to record both the beginningand end of its execution (lines 3 and 12).Return Statements: Apart from reaching the end of a

subroutine, control can be returned back to a calling func-tion through a return statement. There are two reasonsfor instrumenting return statements: (1) to accurately tracknested function calls, and (2) to provide users with the linenumbers of the executed return statements. Without record-ing the execution of return statements, it would be difficultto accurately track nested function calls. Furthermore, byrecording return values and the line number of each returnstatement, Clematis is able to provide users with contex-tual information that can be useful during the debuggingprocess.

Figure 5 illustrates the instrumentation for the returnstatement of informServer(), a function originally shownin the running example (Figure 2, lines 16-21). The wrap-per function RSW receives the return value of the functionand the line number of the return statement and is respon-sible for recording this information before execution of theapplication’s JavaScript is resumed.Function Calls: In order to report the source of a function

invocation, our approach also instruments function calls.When instrumenting function calls, it is important to pre-serve both the order and context of each dynamic call. Toaccurately capture the function call hierarchy, we modifyfunction calls with an inline wrapper function. This allowsus to elegantly deal with two challenging scenarios. First,when multiple function calls are executed from within a sin-

4 UBC-SALT-2014-001

Page 7: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

1 function informServer(email) {2 $.get( '/register/ ' , { email }, function(data) {3 $( '#regMsg ').append(data);4 });5 return RSW(null , 5);6 }

Figure 5: Instrumented JavaScript return state-ment.

1 function submissionHandler(e) {2 $( '#regMsg ')[FCW("html", 2)]("Submitted!");3 var email = $( '#email ')[FCW("value", 3)]();4 if (FCW(isEmailValid , 4)(email)) {5 FCW(informServer , 5)(email);6 $( '#submitBtn ')[FCW("attr",

6)]("disabled", true);7 }8 }9 function clearMsg () {

10 $( '#regMsg ')[FCW("fadeOut", 10)](2000);11 }12 function FCW(fnName) { // Function Call Wrapper13 send(JSON.stringify ({ messageType: "←↩

FUNCTION_CALL" ,..., targetFunction: fnName←↩}));

14 return fnName;15 }

Figure 6: Instrumented JavaScript function calls forrunning example.

gle line of JavaScript code, it allows us to infer the orderof these calls without the need for complex static analysis.Second, inline instrumentation enables us to capture nestedfunction calls. Figure 6 depicts the instrumentation of func-tion calls for two methods from Figure 1, submissionHan-dler() and clearMsg().

Once instrumented using our technique, the function callsto isEmailValid() and informServer() are wrapped byfunction FCW (lines 4 and 5). The interposing function FCW()

executes immediately before each of the original functioncalls and interlaces our function logging with the applica-tion’s original behaviour. Class methods html(), value(),attr(), and fadeOut() are also instrumented in a similarway (lines 2, 3, 6, and 11 respectively).

For comparison, an alternative instrumentation techniqueis shown on line 8 of figure 7. While such a technique mightbe sufficient for measuring function coverage, it does notcapture the order of execution accurately for nested functioncalls or when multiple function calls are made from a singleline. Doing so would require more complex static analysis.

DOM Mutations. Information about DOM mutations canhelp developers relate the observable changes of an applica-tion to the corresponding events and JavaScript code. Tocapture this important information, we introduce an ob-server module into the system. This information is inter-leaved with the logged information about events and func-tions, enabling us to link DOM changes with the JavaScriptcode that is responsible for these mutations.

4.2 Capturing a Behavioural ModelWe use a graph-based model to capture and represent

a web application’s event-based interactions. The graphis multi-edge and directed. It contains an ordered set ofnodes, called episodes, linked through edges that preserve

1 Before Instrumentation:2 getRegistrationDate(getStudentNumber(document.←↩

getElementById( ' username ').value));

4 After Metis Instrumentation:5 FCW(getRegistrationDate)(FCW(getStudentNumber)(←↩

documentFCW("getElementById")( ' username ').←↩value));

7 Alternative Instrumentation:8 FCW(getRegistrationDate);9 FCW(getStudentNumber);

10 FCW(getElementById);11 getRegistrationDate(getStudentNumber(document.←↩

getElementById( ' username ').value));

Figure 7: Comparison of instrumention techniquesfor JavaScript function calls.

the chronological order of event executions.1 In addition,causal edges between the nodes represent asynchronous even-ts. We describe the components of the graph below.

Episode Nodes. An episode is a semantically meaningfulpart of the application behaviour, initiated by a synchronousor an asynchronous event. An event may lead to the exe-cution of JavaScript code, and may change the DOM stateof the application. An episode node contains informationabout the static and dynamic characteristics of the applica-tion, and consists of three main parts:

1. Source: This is the event that started the episode andits contextual information. This source event is eithera DOM event, a timeout callback, or a response to anXHR request, and often causes a part of the JavaScriptcode to be executed.

2. Trace: This includes all the functions that are exe-cuted either directly or indirectly after the source eventoccurs. A direct execution corresponds to functionsthat are called from within an event handler on a DOMelement. An indirect execution corresponds to func-tions that get called due to the bubbling and capturingpropagation of DOM events. The trace also includesall (a)synchronous events that were created within theepisode. All the invoked functions and initiated eventsare captured in the trace part, and their original orderof execution and dependency relations are preserved.

3. Result: This is a section in each episode summariz-ing the changes to the DOM state of the application.These changes are caused by the execution of the episo-de trace and are usually observable by the end-user.

Edges. In our model, edges represent a progression of timeand are used to connect episode nodes. Two types of edgesare present in the model:

• Temporal: The temporal edges connect one episodenode to another, indicating that an episode succeededthe previous one in time.

• Causal: These edges are used to connect different com-ponents of an asynchronous event, e.g., timeouts andXHRs. A causal edge from episode s to d indicatesepisode d was caused by episode s in the past.

Story. The term story refers to an arrangement of episodenodes encapsulating a sequence of interactions with a web1Because JavaScript is single-threaded on all browsers, the events are

totally ordered in time.

5

Page 8: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Figure 8: Top: menu of Clematis. Bottom: overview of a captured story.

Algorithm 1: Story Creation

input : traceoutput: story

Procedure CreateModel() begin1 G < V,E > story ← ∅2 ecurr, eprev ← ∅3 Σtu← ExtractAndSortTraceUnits(trace)4 foreach tu ∈ Σtu do5 if eprev == ∅||eprev.ended()&&

tu.type == episodeSource then6 ecurr ← CreateEpisode()7 ecurr.source← SetEpisodeSource(tu)8 V ← V ∪ ecurr

9 else if tu.type == FunctionTrace||EventHandler ||(tu.type == XHRCallback||TimeoutCallback&& ¬episodeEndCriteria) then

10 ecurr.trace← ecurr.trace ∪ tu11 else if tu.type == DOMMutation then12 ecurr.results← ecurr.results ∪ tu13 if episodeEndCriteriaSatisfied then14 E ← E ∪ createTemporalLink(eprev, ecurr)15 eprev ← ecurr

16 timeoutMap<TimeoutSet, TimeoutCallback>← MapTimeoutTraceUnits(Σtu)

17 XHRMap<XHROpen, XHRSend, XHRCallback>← MapXHRTraceUnits(Σtu)

18 E ← E ∪ ExtractCausalLinks(timeoutMap, XHRMap)19 story ← BuildStory(G < V,E >)20 return story

application. Different stories can be captured according todifferent features, goals, or use-cases that need investigation.

Algorithm 1 takes the trace collected from a web appli-cation as input and outputs a story with episodes and theedges between them. First, the trace units are extractedand sorted based on the timestamp of their occurrence (line3). Next, the algorithm iteratively forms new episodes andassigns trace units to the source, trace, and the result fieldsof individual episodes. If it encounters a trace unit thatcould be an episode source (i.e., an event handler, a time-out, or an XHR callback), a new episode is created (lines5–6) and added to the list of nodes in the story graph (line8). The encountered trace unit is added to the episode as itssource (line 7). Line 9 shows different types of trace unitsthat could be added to the trace field of the episode. Thistrace is later processed to form the complete function callhierarchy as well as each function’s relation with the eventsinside that episode. Next, DOM mutation units that wereinterleaved with other trace units are organized and linked totheir respective episode (lines 11–12). An episode terminatessemantically when the execution of the JavaScript code re-lated to that episode is finished. The algorithm also waitsfor a time interval τ to ensure that the execution of immedi-

ate asynchronous callbacks is completed (line 13). When allof the trace units associated with the source, trace, and re-sult of the episode are assigned and the episode terminationcriteria are met, a temporal edge is added to connect therecently created episode node to the previous one (line 14).The same process is repeated for all episodes by proceedingto the next episode captured in the trace (line 15). Afterall episodes have been formed, the linkages between distantasynchronous callbacks – those that did not complete im-mediately – are extracted and added to the graph as causaledges (lines 16–18). Finally, the story is created based onthe whole graph and returned (lines 19–20).

4.3 Visualizing the Captured ModelIn the final step, our technique produces an interactive

visualization of the generated model, which can be used bydevelopers to understand the behaviour of the application.The main challenge in the visualization is to provide a wayto display the model without overwhelming the developerwith the details. To this end, our visualization follows afocus+context [6] technique that provides the details basedon a user’s demand. The idea is to start with an overviewof the captured story, let the users determine which episodethey are interested in, and provide an easy means to drilldown to the episode of interest. With integration of focuswithin the context, developers can semantically zoom intoeach episode to gain more details regarding that episode,while preserving the contextual information about the story.

Multiple Sessions, Multiple Stories. The user can cap-ture multiple sessions that leads to creation of multiple sto-ries. After each story is recorded, it will be added to the listof captured stories. The name of each story is the date andtime at which it was captured. Figure 9 shows a screenshotof sample captured stories in the visualization of Clematis.Once the users select their desired story, the browser opensa new page dedicated to that story. The initial view of astory contains a menu bar that helps the user navigate thevisualization (Figure 8, top). It also displays an overview ofall captured episodes inside the story (Figure 8, bottom).

Figure 9: Overview of all captured stories.

6 UBC-SALT-2014-001

Page 9: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Figure 10: A DOM mutation as displayed in zoomlevel 2

Story Map, Queries, and Bookmarking. A menu bar isdesigned for the visualization that contains two main parts:the story map and the query mechanism (Figure 8, top).The story map represents a general overview of the wholestory as a roadmap. Panning and (semantic) zooming areavailable for all episode and may cause users to lose the gen-eral overview of the story. Hence, based on the user’s inter-action with the story (e.g., episode selection), the episodesof interest are highlighted on the roadmap to guide the user.The query section enables users to search and filter the infor-mation visualized on the screen. Users can filter the episodesdisplayed on the screen by the episode types (i.e., Event,Timeout, or XHR). They can also search the textual con-tent of the events as well as the actual code. Moreover, theyhave the option to bookmark one or more episodes whileinteracting with the target web application. Those episodesare marked with a star in the visualization to help users tonarrow the scope and spot related episodes (e.g., episode #6in Figure 8 is bookmarked). The episodes’ timing informa-tion is also shown.

Semantic Zoom Levels. The visualization provides 3 se-mantic zoom levels.

Zoom Level 0. The first level displays all of the episodesin an abstracted manner, showing only the type and thetimestamp of each episode (Figure 8, bottom). The type ofeach episode is displayed by the text of the episode as wellas its background color. The horizontal axis is dedicated totime and episodes are sorted from left to right according tothe time of their occurrence (temporal relations). The causaledges between different sections of each timeout or XHRobject are shown by additional edges under the episodes.

Zoom Level 1. When an episode is selected, the view tran-sitions into the second zoom level, which presents an outlineof the selected episode, providing more information aboutthe source event as well as a high-level trace (Figure 11,middle). The trace at this level contains only the names ofthe (1) invoked functions, (2) triggered events, and (3) DOMmutations, caused directly or indirectly by the source event.The user can view multiple episodes to have a side-by-sidecomparison.

Zoom Level 2. The final zoom level exhibits all the infor-mation embedded in each episode. Clicking on the “Event”tab will display the the type of the event that started theepisode (DOM, timeout or XHR event). The contextualinformation of the event are displayed based on its type.Choosing the “DOM mutations” tab will list all the changesthat were made to the DOM after the execution of thisepisode. For each DOM element that was added, removedor modified, an item is added to the list of mutation thatidentifies the modified element, the type of the change andadditional information about the change (Figure 10). The

Figure 11: Three semantic zoom levels in Clematis.Top: overview. Middle: zoomed one level into anepisode, while preserving the context of the story.Bottom: drilled down into the selected episode.

third and final tab depicts a detailed trace of the episode.The trace of at this level includes a customized sequencediagram of the dynamic flow of all the invoked JavaScriptfunctions and events within that episode. When the userclicks on any of the functions or events in the diagram, theJavaScript code of each executed function is displayed andhighlighted (Figure 11, bottom).

RESTful API.We deployed a RESTful API that providesaccess to details about captured stories and allows the ap-proach to remain portable and scalable. This architecturaldecision enables all users, independent of their environments,to take advantage of the behavioural model. By invoking au-thorized calls to the API, one can represent the model as acustom visualization, or use it as a service in the logic of aseparate application. The interface of this API can be foundin Section A.

7

Page 10: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Table 1: Adapted comprehension activities.Activity DescriptionA1 Investigating the functionality of (a part of) the sys-

temA2 Adding to / changing the system’s functionalityA3 Investigating the internal structure of an artifactA4 Investigating the dependencies between two artifactsA5 Investigating the run-time interaction in the systemA6 Investigating how much an artifact is usedA7 Investigating the asynchronous aspects of JavaScriptA8 Investigate the hidden control flow of event handling

4.4 Tool Implementation: ClematisWe implemented our approach in a tool called Clematis,

which is freely available [1]. We use a proxy server to au-tomatically intercept and inspect HTTP responses destinedfor the client’s browser. When a response contains Java-Script code, it is transformed by Clematis. We also usethe proxy to inject a JavaScript-based toolbar into the webapplication, which allows the user to start/stop capturingtheir interactions with the application. The trace data col-lected is periodically transmitted from the browser to theproxy server in JSON format. To observe low-level DOMmutations, we build on and extend the JavaScript MutationSummary library [11]. The model is automatically visual-ized as a web-based interactive interface. Our current im-plementation does not capture the execution of JavaScriptcode that is evaluated using eval. Clematis provides ac-cess to details of captured stories through a RESTFul API(see Section A).

5. CONTROLLED EXPERIMENTSTo assess the efficacy of our approach, we conducted two

controlled experiments [22], one in a research lab settingand the other in an industrial environment. Common designelements of both experiments are described in this section.The next two sections (6–7) are each dedicated to describingthe specific characteristics and results of each experimentseparately. Details of the tasks used for both experimentsare attached in Sections B and C. The consent form, and thepre and post questionnaire forms can be found in SectionsE–G.

Our evaluation aims at addressing the following researchquestions:RQ1 Does Clematis decrease the task completion duration

for common tasks in web application comprehension?RQ2 Does Clematis increase the task completion accuracy

for common tasks in web application comprehension?RQ3 For what types of tasks is Clematis most effective?RQ4 What is the performance overhead of using Clema-

tis? Is the overall performance acceptable?

5.1 Experimental DesignThe experiments had a “between-subject” design; i.e., the

subjects were divided into two groups: experimental groupusing Clematis and control group using other tools. The as-signment of participants to groups was done manually, basedon the level of their expertise in web development. We useda 5-point Likert scale in a pre-questionnaire to collect thisinformation, and distributed the level of expertise in a bal-anced manner between the two groups. None of the partic-ipants had any previous experience with Clematis and allof them volunteered for the study.

Table 2: Comprehension tasks used in study 1.Task Description Activity

T1 Locating the implementation of afeature modifying the DOM

A1, A4

T2 Finding the functions called after aDOM event (nested calls)

A1, A4, A5

T3.a Locating the place to add a new func-tionality to a function

A2, A3

T3.b Finding the caller of a function A4, A5T4.a Finding the functions called after a

DOM event (nested calls + bubbling)A1, A4, A5

T4.b Locating the implementation of a UIbehavior

A1, A3, A4

T5.a Finding the functions called after aDOM event (bubbling + capturing)

A1, A5, A8

T5.b Finding the changes to DOM result-ing from a user action

A4, A5

T6.a Finding the total number of sentXHRs

A6, A7

T6.b Finding if there exists an un-responded XHR

A4, A5, A7

Task Design. The subjects were required to perform aset of tasks during the experiment, representing tasks nor-mally used in software comprehension and maintenance ef-forts. We adapted the activities proposed by Pacione et al.[19], which cover categories of common tasks in programcomprehension, to web applications by replacing two items.The revised activities are shown in Table 1. We designeda set of tasks for each experiment to cover these activities.Tables 2 and 3 show the tasks for studies 1 and 2 accord-ingly. Because study 2 was conducted in an industrial set-ting, participants had limited time. Therefore, we designedfewer tasks for this study compared to study 1.

Independent Variable (IV). This is the tool used for per-forming the tasks, and has two levels: Clematis representsone level, and other tools used in the experiment representthe other level (e.g., Chrome developer tools, Firefox devel-oper tools, Firebug).

Dependent Variables (DV). These are (1) task comple-tion duration, which is a continuous variable, and (2) accu-racy of task completion, which is a discrete variable.

Data Analysis. For analyzing the results of each study, weuse two types of statistical tests to compare dependent vari-ables across the control and experimental groups. Indepen-dent-samples t-tests with unequal variances are used for du-ration and accuracy in study 1, and for duration in study2. However, the accuracy data in study 2 was not normallydistributed, and hence we use a Mann-Whitney U test forthe analysis of accuracy in this study. We use the statisticalanalysis package R [10] for the analysis.

5.2 Experimental ProcedureAll experiments consisted of four main phases. First, the

subjects were asked to fill a pre-questionnaire regarding theirexpertise in the fields related to this study.

In the next phase, the participants in the experimentalgroup were given a tutorial on Clematis. They were thengiven a few minutes to familiarize themselves with the toolafter the tutorial.

In the third phase, each subject performed a set of tasks,as outlined in Tables 2 and 3. Each task was given to a par-ticipant on a separate sheet of paper, which included instruc-tions for the task and had room for the participant’s answer.Once completed, the form was to be returned immediatelyand the subject was given the next task sheet. This allowed

8 UBC-SALT-2014-001

Page 11: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Table 3: Comprehension tasks used in study 2.Task Description Activity

T7 Extracting the control flow of anevent with delayed effects

A1, A4, A5,A7

T8 Finding the mutations in DOM afteran event

A1, A5

T9 Locating the implementation of amalfunctioning feature

A1, A2, A3

T10 Extracting the control flow of anevent with event propagation

A1, A5, A8

us to measure each task’s completion time accurately, toanswer RQ1 and RQ3. To address RQ2 and RQ3, the ac-curacy of each task was later evaluated and marked from 0to 100 according to a rubric that we had created prior toconducting the experiment. The design of the tasks allowedthe accuracy of the results to be quantified numerically. Thetasks and sample rubrics are available in Sections B to D.

In the final phase, participants filled out a post-questionn-aire form providing feedback on their experience with thetool used (e.g., limitations, strength, usability). We catego-rized the qualitative data and present it in Section 9.4.

6. EXPERIMENT 1: LAB ENVIRONMENTThe first controlled experiment was conducted in a lab set-

ting with students and postdocs at the University of BritishColumbia (UBC).

6.1 ApproachExperimental Design. For this experiment, both groupsused Mozilla Firefox 19.0. The control group used Firebug1.11.2. We chose Firebug in the control group since it is thede facto tool used for understanding, editing, and debuggingmodern web applications.2 Firebug has been used in othersimilar studies [23].

Experimental Subjects. We recruited 16 participants forthe study, 3 females and 13 males. The participants weredrawn from different educational levels: 2 undergraduatestudents, 5 Master’s students, 8 Ph.D. students, and 1 Post-doctoral fellow, at UBC. The participants represented dif-ferent areas of software and web engineering and had skillsin web development ranging from beginner to professional.The tasks used in this study are enumerated in Table 2.

Experimental Object. We decided to use a web-basedsurvey application that was developed in our lab. The ap-plication had modest size and complexity, so that it could bemanaged within the time frame anticipated for the experi-ment. Yet it covered the common comprehension activitiesdescribed in Table 1.

Experimental Procedure. We followed the general proce-dure described in section 5.2. After filling the pre-questionnaireform, the participants in the control group were given a tuto-rial on Firebug and had time to familiarize themselves withit, though most of them were already familiar with Firebug.

6.2 ResultsDuration. To address RQ1, we measured the amount oftime (minutes:seconds) spent on each task by the partici-pants, and compared the task durations between Clema-tis and Firebug using a t-test. According to the results

2Firebug has over 3 million active daily users: https://addons.

mozilla.org/en-US/firefox/addon/firebug/statistics/usage/

16:0020:0024:0028:0032:0036:0040:0044:0048:0052:00

To

tal T

ime

(m

:s)

Clematis FirebugFigure 12: t-test analysis with unequal variances oftask completion duration by tool type. Lower valuesare better.

Tim

e (

m:s

)0:00

2:00

4:00

6:00

8:00

10:00

12:00

14:00

T1 Clem

atis

T1 Fire

bug

T2 Clem

atis

T2 Fire

bug

T3 Clem

atis

T3 Fire

bug

T4 Clem

atis

T4 Fire

bug

T5 Clem

atis

T5 Fire

bug

T6 Clem

atis

T6 Fire

bug

Figure 13: Box plots of task completion durationdata per task for each tool. Lower values are better

of the test, there was a statistically significant difference (p-value=0.002) in the durations between Clematis (M=23:22,SD=4:24) and Firebug (M=36:35, SD=8:35). Figure 12shows the results of the comparisons.

To investigate whether certain categories of tasks (Table2) benefit more from using Clematis (RQ3), we tested eachtask separately. The results showed improvements in timefor all tasks. The improvements were statistically significantfor tasks 2 and 5, and showed a 60% and 46% average timereduction with Clematis, respectively. The mean times ofall tasks for Clematis and Firebug are presented in Figure13. The results show that on average, participants usingClematis require 36% less time than than the control groupusing Firebug, for performing the same tasks.

Accuracy. The accuracy of answers was calculated in per-centages. We compared the accuracy of participants’ an-swers using a t-test. The results were again in favour ofClematis and were statistically significant (p=0.02): Clema-tis (M=83%, SD=18%) and Firebug (M=63%, SD=16%).This comparison of accuracy between tools is depicted inFigure 14. As in the duration case, individual t-tests werethen performed for comparing accuracy per task (related toRQ3). Clematis showed an increased average accuracy forall tasks. Further, the difference was statistically significantin favour of Clematis for task 5, and subtasks 4.a and 5.a.The results show that participants using Clematis achieved22% higher accuracy than participants in the control group.We plot the average accuracies of all tasks for Clematis andFirebug in Figure 15. We discuss the implications of theseresults in Section 9.

9

Page 12: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

30

40

50

60

70

80

90

100T

ota

l A

ccu

racy (

%)

Clematis FirebugFigure 14: t-test analysis with unequal variances oftask completion accuracy by tool type. Higher val-ues are better.

Accu

racy (

%)

0

20

40

60

80

100

T1 Clem

atis

T1 Fire

bug

T2 Clem

atis

T2 Fire

bug

T3 Clem

atis

T3 Fire

bug

T4 Clem

atis

T4 Fire

bug

T5 Clem

atis

T5 Fire

bug

T6 Clem

atis

T6 Fire

bug

Figure 15: Box plots of task completion accuracydata per task for each tool. Higher values are better.

7. EXPERIMENT 2: INDUSTRIALTo investigate Clematis’s effectiveness in more realistic

settings, we conducted a second controlled experiment at alarge software company in Vancouver, where we recruitedprofessional developers as participants and used an open-source web application as the experimental object.

7.1 ApproachExperimental Design. Similar to the first experiment,the participants were divided into experimental and controlgroups. The experimental group used Clematis throughoutthe experiment. Unlike the previous experiment, membersof the control group were free to use the tool of their choicefor performing the tasks. The intention was for the par-ticipants to use whichever tool they were most comfortablewith. 5 participants used Google Chrome’s developer tools,2 used Firefox’s developer tools, and 3 used Firebug.

Experimental Subjects. We recruited 20 developers froma large software company in Vancouver, 4 females and 16males. They were 23 to 42 years old and had medium tohigh expertise in web development.

Task Design. For this experiment, we used fewer but morecomplex tasks compared to the first experiment. We de-signed 4 tasks (Table 3) spanning the categories: followingthe control flow, understanding event propagation, detect-ing DOM mutations, locating feature implementation, anddetermining delayed code execution using timeouts.

Experimental Object. Phormer [2] is an online photogallery in PHP, JavaScript, CSS and XHTML. It provides

T7−

Ctr

l

T7−

Exp

T8−

Ctr

l

T8−

Exp

T9−

Ctr

l

T9−

Exp

T10

−C

trl

T10

−E

xp

Tota

l−C

trl

Tota

l−E

xp

Duration (mm:ss)

0:00

8:20

16:40

25:00

33:20

41:40

50:00

Figure 16: Notched box plots of task completionduration data per task and in total for the controland experimental groups (lower values are desired).

features such as uploading, commenting, rating, and display-ing slideshows for users’ photos. It contains typical mecha-nisms such as dynamic DOM mutation, asynchronous calls(XHR and timeouts), and event propagation. Phormer hasaround 6,000 lines of JavaScript, PHP and CSS code in to-tal. It was rated 5.0 star on SourceForge and had over 38,000downloads at the time of conducting the experiment.

Experimental Procedure. We followed the same proce-dure described in 5.2, with one difference: the participantsin the control group were not given any tutorial regardingthe tool they used throughout the experiment, as they wereall proficient users of the tool of their choice.

7.2 ResultsBox plots of task completion duration and accuracy, per

task and in total, for the control (Ctrl) and experimental(Exp) groups, are depicted in Figures 16 and 17, respec-tively.

Duration. Similar to the previous experiment, we ran a setof t-tests for the total task duration as well as for the timespent on individual tasks. The results of the tests showeda statistically significant difference (p-value = 0.0009) be-tween the experimental group using Clematis (M=15:37,SD=1:43) and the control group (M=29:12, SD=5:59), interms of total task completion duration. The results showedimprovements in duration when using Clematis for all fourtasks. We found significant differences in favour of Clema-tis for tasks T7, T8 and T9. The results show that developersusing Clematis took 47% less time on all tasks compared todevelopers in the control group.

Accuracy. We used Mann-Whitney U tests for compar-ing the results of task accuracy between the control and theexperimental group, since the data was not normally dis-tributed. For the overall accuracy of the answers, the tests

10 UBC-SALT-2014-001

Page 13: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

●●

● ●

T7−

Ctr

l

T7−

Exp

T8−

Ctr

l

T8−

Exp

T9−

Ctr

l

T9−

Exp

T10

−C

trl

T10

−E

xp

Tota

l−C

trl

Tota

l−E

xp

0

20

40

60

80

100

Accuracy (%)

Figure 17: Notched box plots of task completionaccuracy data per task and in total for the controland experimental groups (higher values are desired).

revealed a statistically significant difference with high confi-dence (p-value = 0.0005) between Clematis (M=90%, SD-=25%) and other tools (M=35%, SD=20%). We then per-formed the comparison between individual tasks. Again, forall tasks the experimental group using Clematis performedbetter on average. We observed statistical significant im-provements in the accuracy of developers using Clematisfor tasks T7, T8 and T10. The results show that developersusing Clematis performed more accurately across all tasksby 61% on average, compared to developers in the controlgroup.

8. PERFORMANCE OVERHEADWith respect to RQ4, there are two sources of performance

overhead: (1) instrumentation overhead, and (2) executionoverhead. The former pertains to the overhead incurred dueto the instrumentation code added by Clematis, while thelatter pertains to the overhead of processing the trace andconstructing the model. We do not measure the overhead ofvisualization as this is dependent on the user task performed.

Phormer, the experimental object in study 2, was used tocollect performance measurements over 10 one-minute trials.The results were as follows:

Instrumentation overhead. Average delays of 15.04 and1.80 seconds were experienced for pre and post processingphases with Clematis respectively. And a 219.30 ms ad-ditional delay was noticed for each page. On average, eachcaptured episode occupies 11.88 KB within our trace.

Execution overhead. For processing one minute of ac-tivity with Phormer, Clematis experienced an increase of250.8 ms, 6.1 ms and 11.6 ms for DOM events, timeouts andXHRs, respectively.

Based on our experiments, there was no noticeable delayfor end-users when interacting with a given web applicationthrough Clematis.

9. DISCUSSION

9.1 Task Completion DurationTask completion duration is a measure of task perfor-

mance. Therefore, Clematis improves web developers’ per-formance by significantly decreasing the overall time requiredto perform a set of code comprehension tasks (RQ1).

Dynamic Control Flow. Capturing and bubbling mech-anisms are pervasive in JavaScript-based web applicationsand can severely impede a developer in understanding thedynamic behaviour of an application. These mechanismsalso complicate the control flow of an application, as de-scribed in Section 3. Our results show that Clematis signif-icantly reduces the time required for completing tasks thatinvolve a combination of nested function calls, event prop-agation, and delayed function calls due to timeouts withina web application (T2, T5.a, and T7). Hence, Clematismakes it more intuitive to comprehend and navigate the dy-namic flow of the application (RQ3).

One case that needs further investigation is T10. Thistask mainly involves following the control flow when mostof the executed functions are invoked through event propa-gation. The results of this task indicate that although us-ing Clematis caused an average of 32% reduction in taskcompletion duration, the difference was not statistically sig-nificant. However, closer inspection of the results revealsthat the answers given using Clematis for T10 are 68%more accurate in average. This huge difference shows thatmany of the developers in the control group were unawareof occurrences of event propagation in the application, andterminated the task early. Hence, they scored significantlylower than the experimental group in task accuracy and stillspent more time to find the (inaccurate) answers.

Feature Location. Locating features, finding the appro-priate place to add a new functionality, and altering existingbehaviour are a part of comprehension, maintenance anddebugging activities in all software tools, not only in webapplications. The results of study 1 suggested that Clema-tis did reduce the average time spent on the tasks involvingthese activities (T1, T3, T4.b), but these reductions werenot statistically significant. These tasks mostly dealt withstatic characteristics of the code and did not involve any ofthe features specific to JavaScript-based web applications.Study 2, however, involved more complicated tasks in morerealistic settings. T9 represented the feature location activ-ity in this study, and the results showed that using Clema-tis improved the average time spent on this task by 68%.Thus, we see that Clematis speeds up the process of locat-ing a feature or a malfunctioning part of the web application(RQ3).

State of the DOM. The final category of comprehensionactivities investigated in this work is the implications ofevents on the state of the DOM. Results of Study 1 displayeda significant difference in duration of the task involving find-ing DOM mutations in favour of Clematis (T5). The re-sults of Study 2 further confirmed the findings of Study 1 byreducing the duration in almost half (T8). Thus, Clema-tis aids understanding the behaviour of web applications byextracting the mutated elements of the DOM, visualizingcontextual information about the mutations, and linking themutations back to the corresponding JavaScript code (RQ3).

11

Page 14: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

9.2 Task Completion AccuracyTask completion accuracy is another metric for measuring

developers’ performance. According to the results of bothexperiments, Clematis increases the accuracy of develop-ers’ actions significantly (RQ2). The effect is most visiblewhen the task involves event propagation (RQ3). The out-come of Study 1 shows that Clematis addresses Challenge 1(described in Section 3) in terms of both time and accuracy(T5.a). Study 2 further indicates that Clematis helps de-velopers to be more accurate when faced with tasks involvingevent propagation and control flow detection in JavaScriptapplications (67% and 68% improvement for T7 and T10respectively).

For the remaining tasks of Study 1, the accuracy wassomewhat, though not significantly, improved. We believethis is because of the simplistic design of the experimentalobject used in Study 1, as well as the relative simplicity ofthe tasks. This led us towards the design of Study 2 withprofessional developers as participants and a third-party webapplication as the experiment object in the evaluation ofClematis. According to the results of Study 2, Clematissignificantly improves the accuracy of completion of tasks(T8) that require finding the implications of executed codein terms of DOM state changes (RQ3). This is related toChallenge 3 as described in Section 3.

For the feature location task (T9), the accuracy resultswere on average slightly better with Clematis. However,the experimental group spent 68% less time on the task com-pared to the control group. This is surprising as this task iscommon across all applications and programming languagesand we anticipated that the results for the control groupwould be comparable with those of the experimental group.

9.3 Consistent PerformanceLooking at Figures 16 and 17, it can be observed that

using Clematis not only improves both duration and accu-racy of individual and total tasks, but it also helps devel-opers to perform in a much more consistent manner. Thehigh variance in the results of the control group shows thatindividual differences of developers (or tools in Study 2) in-fluence their performance. However, the low variance in allthe tasks for the experimental group shows that Clematishelped all developers in the study to perform consistentlybetter by making it easier to understand the internal flowand dependency of event-based interactions.

9.4 Participants’ FeedbackWe analyzed the qualitative data obtained through the

post-questionnaire forms. Overall the feedback was verypositive. The main features that the participants found mostuseful were the (1) semantic zooming: presenting the over-view first and providing more details on demand, (2) visual-izing the hierarchy of functions and events in the customizedsequence diagram, (3) linking the visualization back to Java-Script code, and (4) extracting DOM mutations per event.The participants also requested for a number of features tobe included in future versions of the tool. These features in-cluded (1) filtering and query options for DOM mutations,(2) ability to attach notes to bookmarked episodes, and (3)integrating Clematis with debugging techniques such asbreakpoints. Overall, according to two of our industrial par-ticipants, Clematis is “Helpful and easy to use” and “Veryuseful. A lot of potential for this tool!”.

9.5 Threats to ValidityInternal Threats. The first threat is that different levelsof expertise in each subject group could affect the results.We mitigated this threat by manually assigning the subjectsto experimental and control groups such that the level of ex-pertise was balanced between the two groups. The secondthreat is that the tasks in the experiment were biased to-wards Clematis. We eliminated this threat by adoptingthe tasks from a well-known framework of common codecomprehension tasks [19]. A third threat arises from theinvestigators’ bias towards Clematis when rating the ac-curacy of subjects’ answers. We addressed this concern bydeveloping an answer key for all the tasks before conduct-ing the experiments. A similar concern arises regarding thetask completion duration measurements. We mitigated thisthreat by presenting each task to subjects on a separatesheet of paper and asking them to return it upon completion.The duration of each task was calculated from the point asubject received the task until they returned the paper tothe investigators, thus eliminating our bias in measuring thetime (and the subjects’ bias in reporting the time). Finally,we avoided an inconsequential benchmark by choosing a toolfor the control group in Study 1 that was stable and widely-deployed, namely Firebug. In Study 2, the developers inthe control group were given the freedom to choose any toolthey preferred (and had experience with).

External Threats. An external threat to validity is thatthe tasks used in the experiment may not be representa-tive of general code comprehension activities. As mentionedabove, we used the Pacione’s framework and thus these tasksare generalizable. A similar threat arises with the represen-tativeness of the participants. To address this threat, weused both professional web developers and students/post-docs with previous web development experience.

Reproducibility. As for replicating our experiments, Clema-tis [1], the experimental object Phormer [2], and the detailsof our experimental design (e.g., tasks and questionnaires)are all available making our results reproducible.

10. CONCLUDING REMARKSModern web applications are highly dynamic and inter-

active, and offer a rich experience for end-users. This in-teractivity is made possible by the intricate interactions be-tween user-events, JavaScript code, and the DOM. How-ever, web developers face numerous challenges when tryingto understand these interactions. In this paper, we pro-posed a portable and fully-automated technique for relatinglow-level interactions in JavaScript-based web applicationsto high level behaviour. We proposed a behavioural modelto capture these event interactions, and their temporal andcausal relations. We presented a novel interactive visual-ization mechanism based on focus+context techniques, forpresenting these complex event interactions in a more com-prehensible format to web developers. Our approach is im-plemented in a code comprehension tool, called Clematis.The evaluation of Clematis points to the efficacy of the ap-proach in reducing the overall time and increasing the accu-racy of developer actions, compared to state-of-the-art webdevelopment tools. The greatest improvement was seen fortasks involving control flow detection, and especially eventpropagation, showing the power of our approach.

12 UBC-SALT-2014-001

Page 15: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

As part of future work, we plan to improve the interac-tive visualization and extend the details captured in eachstory to allow the programmer to gain a better insight intothe application. Another direction we will pursue is in de-bugging, where Clematis can potentially help developersto better detect and localize faulty behaviour of JavaScriptapplications.

11. ACKNOWLEDGMENTSThis work was supported in part by an NSERC Strategic

Project Grant and a research gift from Intel Corporation.We thank Mohammed Ali for his contributions in improv-

ing the visualization of Clematis. We are grateful to allparticipants of the controlled experiments for their time andcommitment. We especially thank SAP Labs Vancouver forall their help and support.

12. REFERENCES[1] Clematis.

http://salt.ece.ubc.ca/software/clematis/.

[2] Phormer PHP photo gallery. http://p.horm.org/er/.

[3] D. Amalfitano, A. Fasolino, A. Polcaro, andP. Tramontana. The DynaRIA tool for thecomprehension of Ajax web applications by dynamicanalysis. Innovations in Systems and SoftwareEngineering, pages 41–57, 2014.

[4] S. Andrica and G. Candea. WaRR: A tool forhigh-fidelity web application record and replay. InProceedings of the International Conference onDependable Systems & Networks (DSN), pages403–410. IEEE Computer Society, 2011.

[5] B. Burg, R. Bailey, A. J. Ko, and M. D. Ernst.Interactive record/replay for web applicationdebugging. In Proceedings of the Symposium on UserInterface Software and Technology (UIST), pages473–484. ACM, 2013.

[6] A. Cockburn, A. Karlson, and B. B. Bederson. Areview of overview+detail, zooming, andfocus+context interfaces. ACM Computing Surveys,41(1):2:1–2:31, 2009.

[7] T. A. Corbi. Program understanding: Challenge forthe 1990s. IBM Systems Journal, 28(2):294–306, 1989.

[8] B. Cornelissen, A. Zaidman, and A. van Deursen. Acontrolled experiment for program comprehensionthrough trace visualization. IEEE Transactions onSoftware Engineering, 37(3):341–355, 2011.

[9] B. Cornelissen, A. Zaidman, A. van Deursen,L. Moonen, and R. Koschke. A systematic survey ofprogram comprehension through dynamic analysis.IEEE Transactions on Software Engineering,35(5):684–702, 2009.

[10] R. Gentleman and R. Ihaka. The R project forstatistical computing. http://www.r-project.org.

[11] Google. Mutation Summary Library.http://code.google.com/p/mutation-summary/.

[12] P. Li and E. Wohlstadter. Script InSight: Usingmodels to explore JavaScript code from the browserview. In Proceedings of the 9th InternationalConference on Web Engineering (ICWE), pages260–274. Springer-Verlag, 2009.

[13] J. Maras, J. Carlson, and I. Crnkovi. Extractingclient-side web application code. In Proceedings of theInternational Conference on World Wide Web(WWW), pages 819–828. ACM, 2012.

[14] J. Maras, M. Stula, and J. Carlson. Generatingfeature usage scenarios in client-side web applications.In Proceeding of the International Conference on WebEngineering (ICWE), pages 186–200. Springer, 2013.

[15] J. Mickens, J. Elson, and J. Howell. Mugshot:Deterministic capture and replay for Javascriptapplications. In Proceedings of the 7th USENIXConference on Networked Systems Design andImplementation, NSDI’10, pages 159–174. USENIXAssociation, 2010.

[16] P. Montoto, A. Pan, J. Raposo, F. Bellas, andJ. Lopez. Automating navigation sequences in Ajaxwebsites. In Proceedings of the InternationalConference on Web Engineering (ICWE), pages166–180. Springer, 2009.

[17] F. Ocariza, K. Bajaj, K. Pattabiraman, andA. Mesbah. An empirical study of client-sideJavaScript bugs. In Proceedings of the ACM/IEEEInternational Symposium on Empirical SoftwareEngineering and Measurement (ESEM), pages 55–64.IEEE Computer Society, 2013.

[18] S. Oney and B. Myers. FireCrystal: Understandinginteractive behaviors in dynamic web pages. InProceedings of the Symposium on Visual Languagesand Human-Centric Computing, pages 105–108. IEEEComputer Society, 2009.

[19] M. J. Pacione, M. Roper, and M. Wood. A novelsoftware visualisation model to support softwarecomprehension. In Proceedings of the WorkingConference on Reverse Engineering, pages 70–79.IEEE Computer Society, 2004.

[20] K. Sen, S. Kalasapur, T. Brutch, and S. Gibbs.Jalangi: a selective record-replay and dynamic analysisframework for JavaScript. In Proceedings of the JointMeeting on Foundations of Software Engineering,ESEC/FSE 2013, pages 488–498. ACM, 2013.

[21] W3C. Document Object Model (DOM) level 2 eventsspecification.http://www.w3.org/TR/DOM-Level-2-Events/, 13November 2000.

[22] C. Wohlin, P. Runeson, M. Host, M. C. Ohlsson,B. Regnell, and A. Wesslen. Experimentation insoftware engineering: an introduction. Kluwer, 2000.

[23] A. Zaidman, N. Matthijssen, M.-A. Storey, and A. vanDeursen. Understanding Ajax applications byconnecting client and server-side execution traces.Empirical Software Engineering, 18(2):181–218, 2013.

13

Page 16: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

APPENDIXA. RESTFUL API

We provide a RESTful API for transferring the informa-tion regarding the stories and their details to the users.Clematis’s visualization uses this API to acquire the in-formation that need to be displayed. This section providesa list of valid calls to the REST API of Clematis’ visualiza-tion.

• Retrieve the number of episodes in a story.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/howmany

• Retrieve All the episodes of a story.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/

• Display All the episodes in a more readable format.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/pretty

• Retrieve a certain Episode by providing its id.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/{id}

• Retrieve the Dom of a certain episode by providing itsid.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/{id}/DOM

• Retrieve the trace of a certain episode by providing itsid.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/{id}/trace

• Retrieve the source of a certain episode by providingits id.

http://localhost:8080/rest/clematis-api/

{storyName}/episodes/{id}/source

• Retrieve a list of causal links.

http://localhost:8080/rest/clematis-api/

{storyName}/story/causalLinks

B. TASKS: EXPERIMENT 11. In the Contest Agreement section, when the user

agrees with the terms and conditions and checks thebox, the fields for entering and submitting user’s nameand email get activated. What JavaScript function isresponsible for this functionality?

2. In the Contest Agreement section, please specify thenames of developer-defined functions that get called di-rectly or indirectly after clicking the Enter button (pre-serve the orders).

3. In the Contest Agreement section, we want to addsome code for checking the name and email right beforethey are sent to the server using a XMLHttpRequest ob-ject.

(a) Find the right place (JavaScript function and theline number in that function) in the code to addthis check.

(b) Let’s call the function that you found foo(); whatfunction calls foo()?

4. In the Education section, selecting graduate or un-

dergraduate degree affects the allowed number of yearsin the program shown in the dropdown menu.

(a) Name all developer-defined functions that get calleddirectly or indirectly after a choice is made betweengraduate and undergraduate.

(b) Name the responsible function for handling the changein the dropdown menu for years in program.

5. In the Submit section, the end user can submit thesurvey form by pressing the Submit button.

(a) How many developer-defined functions are calledafter the end user presses the Submit button? Pleasespecify the names of these functions (preserve theorders).

(b) What element(s) in the Document Object Model(DOM) change after submission? What is the change?

6. How many XMLHttpRequest objects were sent to serverin total? Is there any XMLHttpRequest sent to theserver that does not receive a response? If yes, pleasespecify its ID.

C. TASKS: EXPERIMENT 21. Open the slideshow page and pause the slideshow. The

slideshow will stop and you can see a picture on thescreen. After you press play, in order to display thenext image, what developer-defined functions get calleddirectly or indirectly? Preserve the order of calls.

2. When the picture changes and a new picture is dis-played in the slideshow, what elements of the DOMchange? And what is the change?

3. After a picture is loaded and displayed, there is a codethat gathers some statistical information regarding theaction. This functionality does not work properly andwe want to find the place where it is implemented. Findthe responsible function for this functionality.

4. List the developer-defined functions that get called di-rectly or indirectly after you click the “smaller size”option in the slideshow.

D. RUBRICSThis section contains samples of grading rubrics for both

lab and industrial experiments.

D.1 Experiment 1: Lab EnvironmentThe first two sample tasks are chosen from the lab exper-

iment tasks. The questions and the answers are both shownin this section.

Sample Question (Q4). In the Education section, se-lecting graduate or undergraduate degree affects the al-lowed number of years in the program shown in the drop-down menu.

1. Name all developer-defined functions that get calleddirectly or indirectly after a choice is made betweengraduate and undergraduate.

Answer:

• educationDivClicked()

• addOption()

14 UBC-SALT-2014-001

Page 17: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

• handleFormClick()

2. Name the responsible function for handling the changein the dropdown menu for years in program.

Answer:

• educationDivClicked()

Sample Question (Q5). In the Submit section, the enduser can submit the survey form by pressing the Submit

button.

1. How many developer-defined functions are called afterthe end user presses the Submit button? Please specifythe names of these functions (preserve the orders).

Answer:

• handleClickOfSubmission()

• submitForm()

• sendSubmissionToServer()

• onreadystatechange

• handleSubmissionDivClick()

• handlerFormClick()

• After timeout expiration: warnInactiveSubmis-

sion()

2. What element(s) in the Document Object Model (DOM)change after submission? What is the change?

Answer:

• DIV element with ID: submissionFinalizedMsg

• DIV’s text changes

D.2 Experiment 2: Industrial EnvironmentThis section contains two sample tasks from the indus-

trial experiment, as well as the pre-defined rubrics for theanswers.

Sample Question (Q1). Open the slideshow page andpause the slideshow. The slideshow will stop and you cansee a picture on the screen. After you press play, in orderto display the next image, what developer-defined functionsget called directly or indirectly? Preserve the order of calls.Answer:

• onclick()

• ss_playpause()

• dg()

• ss_run()

• mp_table()

• mp_granny()

• mp_center()

• ss_slideshow()

• ss_update()

• hideElem()

• dg()

• inlineElem()

• ss_run()

• sts_data_collection()

• updateNumOfLoads()

• storeUserInfo()

• sendStsToServer()

• ss_loaddone()

Sample Question (Q3). After a picture is loaded anddisplayed, there is a code that gathers some statistical in-formation regarding the action. This functionality does notwork properly and we want to find the place where it isimplemented. Find the responsible function for this func-tionality.Answer:

• sts_data_collection()

15

Page 18: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Information Consent Form

Effects of Using Clematis on Comprehension,

Debugging and Maintenance of Web Applications

Human Ethics Application ID: H13-00632

Principal Investigator:Dr. Ali MesbahAssistant ProfessorElectrical and Computer Engineering Department, University of British Columbia(604) [email protected]

Co-Investigators:Dr. Karthik PattabiramanAssistant ProfessorElectrical and Computer Engineering Department, University of British Columbia(604) [email protected]

Saba AlimadadiGraduate StudentElectrical and Computer Engineering Department, University of British Columbia(604) [email protected]

Sheldon SequeiraGraduate StudentElectrical and Computer Engineering Department, University of British Columbia(778) [email protected]

Sponsor: N/A

Page 1 of 3

E. INFORMATION CONSENT FORM

16 UBC-SALT-2014-001

Page 19: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Purpose:The purpose of this study is to investigate whether our tool, Clematis, improves compre-hension, maintenance and debugging of web applications. In this study, you will be askedto perform some tasks with Clematis, or using other tools such as the well-known Firefoxadd-on Firebug. The results will be compared later in terms of task completion durationand accuracy.

Study Procedure:During the period of this experiment, first you will be informed of the consent for the studyand sign the consent form. Then you will fill a pre-questionnaire survey, indicating yourlevel of expertise in the fields related to the current experiment. Afterwards, you will beasked to perform a list of tasks and write down the results. After finishing the study, youwill fill a post-questionnaire form about your opinion on the tool and finally you will beshortly interviewed about your experience with the tools.

Consent forms will be administered in person and thus signatures will be collected in person.You may also ask questions about the consent form and the procedure of the experiment.You may leave the experiment at any time that you wish. You can withdraw your volun-teered participation without consequences.

Potential Risks:The risks related to this project are minimal and are no more than what you would en-counter in your regular day.

Potential Benefits:The results of this study will help assess and improve our under-development tool for mod-elling web applications. Web developers can use the tool, Clematis, in order to help themcomprehend, debug and maintain their code.

Results and Future Contact: You can obtain the results of this study from Saba Al-imadadi via email ([email protected]) after the results have been processed, analyzed anddocumented. After Completion of this experiment, there will be no contact with you fromthe investigator team. However, feel free to contact a co-investigator should you have anycomments, questions or concerns.

Complaints:Complaints or concerns about your treatment or rights as a research participant should bedirected to:The Research Subject Information Line in the UBC Office of Research Services

Page 2 of 3

17

Page 20: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

(604) [email protected]

Confidentiality:The data collected during this experiment is completely anonymous and total confidential-ity is promised.

Data will be recorded in the form of text documents and will not include any personalinformation about the participant. The gathered data will be put in a password protectedfolder in the investigators flash memory stick, and will be locked in a secure cabinet for 5years after the date of the study.

Consent:Your participation in this study is entirely voluntary and you may refuse to participate orwithdraw at any time during the experiment.

You consent to participate in the study on Effects of Modelling Web Applications withClematis on Code Comprehension, Debugging and Maintenance. You have understood thenature of this project and wish to participate.

Your signature below indicates that you have received a copy of this consent form for yourown records.

———————————————————————————————————————Participant’s Name

———————————————————————————————————————Participant’s Signature

———————————————————————————————————————Date

Page 3 of 3

18 UBC-SALT-2014-001

Page 21: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Pre-Questionnaire

Name:Job Title:Degree:Sex:Age:

1. From 1 to 5, what best describes your familiarity with web development(1: no/very little experienced, 5: professional)1 2 3 4 5

2. Please name the tools that you usually use for understanding, maintainingand debugging the web-application code.From 1 to 5, what best describes your familiarity with each tool?(1: no/very little experienced, 5: professional)

Tool Familiarity

1) ........................................... 1 2 3 4 52) ........................................... 1 2 3 4 53) ........................................... 1 2 3 4 5

3. From 1 to 5, what best describes your familiarity with Firebug(1: no/very little experienced, 5: professional)1 2 3 4 5

4. What is the browser that you normally use for web development?

5. How many years of experience have you had with web development?

6. How many years of experience have you had with software development?

1

F. PRE-QUESTIONNAIRE FORM

19

Page 22: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

Post-Questionnaire

ID Number:

1) Please name the tool(s) you used in the experiment.

2) Did you find the above tool(s) helpful for understanding the web application? Whatcapabilities of the tool(s) did you find more useful?

3) What were the limitations of the tool(s) used in the experiment? What improvementswould you suggest?

1

G. POST-QUESTIONNAIRE FORM

20 UBC-SALT-2014-001

Page 23: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon
Page 24: Understanding JavaScript Event-Based Interactionssalt.ece.ubc.ca/publications/docs/UBC-SALT-2014-001.pdf · Understanding JavaScript Event-Based Interactions Saba Alimadadi Sheldon

UBC-SALT-2014-001