Attention Please! Learning Analytics for Visualization and

8
Attention Please! Learning Analytics for Visualization and Recommendation Erik Duval Dept. Computer Science Katholieke Universiteit Leuven Celestijnenlaan 200A B3000 Leuven, Belgium [email protected] ABSTRACT This paper will present the general goal of and inspiration for our work on learning analytics, that relies on attention metadata for visualization and recommendation. Through information visualization techniques, we can provide a dash- board for learners and teachers, so that they no longer need to ”drive blind”. Moreover, recommendation can help to deal with the ”paradox of choice”and turn abundance from a problem into an asset for learning. 1. INTRODUCTION Attention is a core concern in learning: as learning resources become available in more and more abundant ways, atten- tion becomes the scarce factor, both on the side of learners as well as on the side of teachers. (This is a wider concern, as we evolve towards an ’attention economy’ [10].) Learners and teachers leave many traces of their attention: some are immediately obvious to others, for instance in the form of posts and comments on blogs, or as twitter mes- sages. These explicit traces are human readable, but can be difficult to cope with in a world of abundance [29]. Although some refer to ’information overload’, we prefer Shirky’s ”fil- ter failure” as a way to think about the problem of dealing with this abundance [30]. In any case, human attention traces are extremely valuable, but do not scale very well. In this paper, we will explain how machine readable traces of attention can be used to filter and suggest, provide aware- ness and support social links. This paper is structured as follows: section 2 provides a brief background on the field of analytics in general. The section thereafter focuses on applications from the world of jogging, as these provide a particularly rich source of inspiration for our work. The general concept of goal oriented visualiza- tions is at the core of learning dashboard applications: that is why it is the topic of section 4. In order to scale up the cur- This text is a PREPRINT. Please cite as: E. Duval Attention Please! Learning Analytics for Visualization and Recommendation. To appear in: Proceedings of LAK11: 1st International Conference on Learning Analytics and Knowledge 2011. rent work on learning analytics and achieve broad adoption, it is imperative to establish a global open infrastructure, as we briefly explain in section 5. The two next sections briefly present the two approaches we’ve explored so far to lever- age learning analytics: learning dashboard (section 6) and learning recommenders (section 7). Before concluding the paper, we briefly mention exciting opportunities that learn- ing analytics provides for data based research on learning in section 8. 2. BACKGROUND ON ANALYTICS The new field of learning analytics is quite related to similar evolutions in other domains, such as Big Data [1], e-science [14, 9, 26], web analytics [7], educational data mining [25]. All of these have in common that they rely on large col- lections of quite detailed data in order to detect patterns. This detection of patterns can be based on data mining tech- niques, so that for instance recommendations can be made for resources, activities, people, etc. that are likely to be relevant. Alternatively, the data can be processed so that they can be visualized in a way that enables the teacher or learner rather than the software to make sense of them. In fact, the research in my team gets much of its inspira- tion from tools like wakoopa (http://social.wakoopa.com) or rescuetime (http://www.rescuetime.com), that install tracker tools on the machine of a user and then automati- cally record all activities (applications launched, documents accessed, web sites visited, music played, etc.) by that user. A typical illustration of the visualizations that such tools provide is figure 1, where a simple overview is presented of the software applications I used last week and last month and how their usage is distributed over time. (Tuesday- Thursday were travel days...) In this way, such an applica- tion can help a user to be more aware of her activities. Moreover, based on these tracking data, the wakoopa tool can compare the activity of a user with that of other users and recommend software or contacts - see figure 2. It doesn’t require much imagination to see how similar visualizations could be useful in the world of learning, for instance to chart learning activities, tools used or recommended, and peer learners or suitable teachers in one’s social network. It is important to note that the tracking occurs without any manual effort by the user - although it is of course important

Transcript of Attention Please! Learning Analytics for Visualization and

Page 1: Attention Please! Learning Analytics for Visualization and

Attention Please!Learning Analytics for Visualization and Recommendation

Erik DuvalDept. Computer Science

Katholieke Universiteit LeuvenCelestijnenlaan 200A

B3000 Leuven, [email protected]

ABSTRACTThis paper will present the general goal of and inspirationfor our work on learning analytics, that relies on attentionmetadata for visualization and recommendation. Throughinformation visualization techniques, we can provide a dash-board for learners and teachers, so that they no longer needto ”drive blind”. Moreover, recommendation can help todeal with the ”paradox of choice” and turn abundance froma problem into an asset for learning.

1. INTRODUCTIONAttention is a core concern in learning: as learning resourcesbecome available in more and more abundant ways, atten-tion becomes the scarce factor, both on the side of learnersas well as on the side of teachers. (This is a wider concern,as we evolve towards an ’attention economy’ [10].)

Learners and teachers leave many traces of their attention:some are immediately obvious to others, for instance in theform of posts and comments on blogs, or as twitter mes-sages. These explicit traces are human readable, but can bedifficult to cope with in a world of abundance [29]. Althoughsome refer to ’information overload’, we prefer Shirky’s ”fil-ter failure” as a way to think about the problem of dealingwith this abundance [30]. In any case, human attentiontraces are extremely valuable, but do not scale very well.

In this paper, we will explain how machine readable tracesof attention can be used to filter and suggest, provide aware-ness and support social links.

This paper is structured as follows: section 2 provides a briefbackground on the field of analytics in general. The sectionthereafter focuses on applications from the world of jogging,as these provide a particularly rich source of inspiration forour work. The general concept of goal oriented visualiza-tions is at the core of learning dashboard applications: thatis why it is the topic of section 4. In order to scale up the cur-

This text is a PREPRINT. Please cite as: E. Duval Attention Please!Learning Analytics for Visualization and Recommendation. To appearin: Proceedings of LAK11: 1st International Conference on LearningAnalytics and Knowledge 2011.

rent work on learning analytics and achieve broad adoption,it is imperative to establish a global open infrastructure, aswe briefly explain in section 5. The two next sections brieflypresent the two approaches we’ve explored so far to lever-age learning analytics: learning dashboard (section 6) andlearning recommenders (section 7). Before concluding thepaper, we briefly mention exciting opportunities that learn-ing analytics provides for data based research on learning insection 8.

2. BACKGROUND ON ANALYTICSThe new field of learning analytics is quite related to similarevolutions in other domains, such as Big Data [1], e-science[14, 9, 26], web analytics [7], educational data mining [25].All of these have in common that they rely on large col-lections of quite detailed data in order to detect patterns.This detection of patterns can be based on data mining tech-niques, so that for instance recommendations can be madefor resources, activities, people, etc. that are likely to berelevant. Alternatively, the data can be processed so thatthey can be visualized in a way that enables the teacher orlearner rather than the software to make sense of them.

In fact, the research in my team gets much of its inspira-tion from tools like wakoopa (http://social.wakoopa.com)or rescuetime (http://www.rescuetime.com), that installtracker tools on the machine of a user and then automati-cally record all activities (applications launched, documentsaccessed, web sites visited, music played, etc.) by that user.

A typical illustration of the visualizations that such toolsprovide is figure 1, where a simple overview is presented ofthe software applications I used last week and last monthand how their usage is distributed over time. (Tuesday-Thursday were travel days...) In this way, such an applica-tion can help a user to be more aware of her activities.

Moreover, based on these tracking data, the wakoopa toolcan compare the activity of a user with that of other usersand recommend software or contacts - see figure 2. It doesn’trequire much imagination to see how similar visualizationscould be useful in the world of learning, for instance tochart learning activities, tools used or recommended, andpeer learners or suitable teachers in one’s social network.

It is important to note that the tracking occurs without anymanual effort by the user - although it is of course important

Page 2: Attention Please! Learning Analytics for Visualization and

Figure 1: My wakoopa dashboard

Figure 2: My Wakoopa recommendations

Page 3: Attention Please! Learning Analytics for Visualization and

that the user is aware that her activities are being tracked.Actually, such tools typically also make it possible to pausetracking. Some applications allow users to set goals (”spendless than 1 hour per day on email” or ”play computer gamesfor less than 1,5 hour per day” or ”write more than 3 hoursper day”) and will notify them when they are in dangerof not meeting their goal, when they get close to the self-imposed limit - or signal them that they did reach their goal.Moreover, they provide quite detailed visualizations of allthe activities of a user, so that she can analyze where mostof her on-line activity takes place and make better informeddecisions on how to manage these activities.

A similar tool is tripit (http://www.tripit.com): notewor-thy about this tool is that when a user forwards flight orhotel reservations to a tripit email address, all the struc-tured data is extracted and a calendar is created with allthe relevant information. This is an excellent example ofautomatic metadata generation [6] or information extraction[5], an essential technology if we want to collect metadataof resources, activities and people at scale. Note also that,if other people from the user’s network are near, tripit willmention that - see figure 3.

Of course, more mainstream tools like google offer similarfunctionality, such as for instance ”google history” that pro-vides an overview of every search that a user ever submitted(when logged in) or that indicates who from a user’s socialcircle tweeted about an item included in a search result, etc.

3. INSPIRATION FOR OUR WORKA particularly inspiring set of applications comes from thedomain of jogging, and sports in general, where applica-tions like nikeplus (http://nikerunning.nike.com/, see fig-ure 4) or runkeeper (http://runkeeper.com/) provide de-tailed statistics on how fast, far, often, etc. one runs.

What is particularly relevant in a learning context is thatmany running applications also help runners to set goals(”run a marathon in november”), develop a plan to achievethat goal, find running routes in a foreign town, locate otherrunners with a similar profiles, challenge them so as to main-tain motivation, etc. Sometimes, such tools even take a morepro-active role and send messages to users to enquire whythey have stopped uploading activities, whether they needto re-define goals and plans, or want to be connected to otherusers that can help, etc.

Although there are few studies that show whether these spe-cial purpose social networks actually change user behavior,[16] did find that ”users’ weight changes correlated positivelywith the number of their friends and their friends’ weight-change performance” and ”users’ weight changes have rip-pling effects in the Online Social Networks due to the so-cial influence. The strength of such online influence and itspropagation distance appear to be greater than those in thereal-world social network.”. An early overview of how thecombination of tracking and social network services can leadto a more patient-driven approach to medicine is providedin [32].

One assumption underlying our work is that similar appli-cations can be built to track learner progress, to assist in

developing and maintaining motivation, to help define re-alistic goals and develop plans to achieve them, as well asconnect learners or teachers with other learning actors, etc.In that way, they can help to realize a more learner-drivenapproach to education, training and learning in general.

4. GOAL ORIENTED VISUALIZATIONSMany of these inspiring applications take a visual approach.Yet, if visualizations are to have any effect beyond the ini-tial ”wow” factor, it would be useful to have more clarityon what the intended goal is and how to assess whetherthat goal is achieved. Many visualizations look good - andsome are actually beautiful. But how we can connect visu-alization not only with meaning or truth, but with takingactions? This is very much a ”quantified self” approach (seehttp://quantifiedself.com/) [31], where for instance a vi-sualization of eating habits can help to lead a healthier life,or where a visualization of mobility patterns can help toexplore alternative modes of transport, etc. Such visualiza-tions are successful if they trigger the intended behaviour(change). That can be measured, as in ”people smoke lesswhen they use this visualization” or ”people discover newpublications based on this visualization” (we are actuallyevaluating such an application) or ”people run more usingthis visualization” etc.

It would be really useful if we could draw up some guide-lines to design effective goal oriented visualizations. As anexample, it is probably kind of useful to be able to visualiseprogress towards a goal - or lack thereof. If you want to runfurther, a visualization can help you to assess whether you’remaking progress. Or if you want to spend less time doingemail, a simple visualization can help. Another guidelinecould relate to social support, that enables you to compareyour progress with that of others.

5. TECHNICAL INFRASTRUCTURE FORLEARNING ANALYTICS

If we want to apply learning analytics at a broader scale,then it is imperative that we realize an infrastructure thatcan support the development of tools and services. Such aninfrastructure will need basic technical agreement on com-mon standards and protocols [8].

A first question is how to model the relevant data. Ourearly work on Contextualized Attention Metadata (CAM)[19] [36] defines a simple model to structure attention meta-data, i.e. the interactions that people have with objects.The ontology-based user interaction context model (UICO)[24] focuses more on the tasks that people carry out whileinteracting with resources. Either we need to better un-derstand how to map and translate automatically betweendifferent such models, or we need to find a way to achievebroad consensus on and adoption of a common schema ora small set of schemas, as in the case of learning resourceswhere nearly everyone has now adopted Learning ObjectMetadata (LOM) or Dublin Core (DC) [8].

Similar to the way we manage learning objects and theirmetadata [33], we will need a service architecture that canpower a plethora of tools and applications. One interest-ing approach is to rely on technologies like widgets that

Page 4: Attention Please! Learning Analytics for Visualization and

Figure 3: My tripit dahsboard

Figure 4: My Nike Plus dashboard

Page 5: Attention Please! Learning Analytics for Visualization and

enable the dynamic embedding of small application com-ponents - an approach at the core of Personal LearningEnvironments (PLE’s), researched in the ROLE project onResponsive Open Learning Environments (see http://www.

role-project.eu/) [15] [12]. Another approach is the Learn-ing Registry architecture that makes ”user data trails” avail-able through a network of nodes that provide services topublish, access, distribute, broker or administer paradata(see http://www.learningregistry.org/).

6. LEARNING DASHBOARDSFor learners and teachers alike, it can be extremely usefulto have a visual overview of their activities and how theyrelate to those of their peers or other actors in the learningexperience.

In fact, such visualizations can also be quite useful for otherstakeholders, like for instance system administrators. Figure5 provides an early example of such a visualization that dis-plays the number of events in different widgets deployed inthe ROLE context [27]. From the visualization, it is ratherobvious that users were most active in the May-July period(towards the left of the diagram), that they enter chat rooms(top of area 4 on Figure 5) much more often than they postmessages (third row of area 4 on Figure 5), etc. Such in-formation can help a teacher to re-organize the activities oreven to retract or add widgets that learners can deploy intheir PLE.

Similarly, [28] describes a tool that includes a ”zeitgeist” ofaction types (opening a document, sending a message, etc.)and specific user actions. By selecting a time period and therelevant action types, the user can control the visualizationof relevant data (see also http://www.role-showcase.eu/

role-tool/cam-zeitgeist).

Following a similar visual approach, the Student ActivityMonitor (SAM) supports self-monitoring for learners andawareness for teachers [13]: In area A on figure 6, every linerepresents a student in a course. The horizontal axis repre-sents calendar time and the vertical axis total time spent.If the line ascends fast, then the student worked intenselyduring that period. If the line stays flat, the student did notwork much on the course. For example, student s1 startedlate and worked very hard for a very short time. Students2 started early and then worked harder in about the sameperiod as student s1. At the bottom, a smaller version of thevisualization is shown with a slider on top to select a partof the period for analysis of data dense areas. Area 2 dis-plays global course statistics on time spent and documentuse. The colored dots represent minimum, maximum andaverage time spent per student and the time spent for thecurrently logged in user and for a user selected in one of thevisualizations. The recommendation area in Box 3 enablesexploration of document recommendations (see also section7). The parallel coordinates in area B display

1. the total time spent on the course,

2. the average time spent on a document,

3. the number of documents used and

4. the average time of the day that the students work.

For example, the green line (the logged in user) works onaverage in the early evening and is spending an average timein line with the majority. He does not use so many differentdocuments and on average looks at these for a short time.He scores the worst here. The average student of the class(in yellow) is also presented. This is a somewhat complexvisualization, but our evaluation studies show that studentsconsidered the visualizations clear [13]. They rated the toolsas usable, useful, understandable and organized.

A much more simple such visualization is edufeedr [21],where a matrix includes a row for every student that dis-plays his progress along a series of assignments. A nice fea-ture is that such progress can take place on the individualblog of the student, outside of the institutional LearningManagement System (LMS), Virtual Learning Environment(VLE) or even institution provided PLE widgets. Rather,the coherence of the course is maintained through the track-back mechanism between the teacher blog and those of thestudents.

What these visualizations have in common is that they en-able a learner or teacher to obtain an overview of their ownefforts and of those of the (other) learners. This is theessence of our ”dashboard” approach to visualizations forlearning that remedy the ”blind driving”that often occurs onthe side of teachers and learners alike. Similar approacheshave proven to be beneficial in for instance software engi-neering [3] and social data analysis [18].

7. LEARNING RECOMMENDERSBy collecting data about user behavior, learning analyticscan also be mined for recommendations, of resources, activ-ities or people [17]. In this way, we can turn the abundanceof learning resources into an asset, by addressing ”l’embarrasdu choix” that is at the core of ”the paradox of choice”[29]. Of course, similar approaches have been deployed forbooks, music, entertainment, etc. Yet, only by basing rec-ommenders on detailed learning attention metadata can theytake into account the learning specific characteristics and re-quirements of our activities.

In one particular tool, we applied this approach to filter andrank search results when a learner searches for material inYouTube (http://www.youtube.com/), SlideShare (http://www.slideshare.net/) and Globe (http://www.globe-info.org/): as figure 7 illustrates, every search activity in our toolis tracked in the form of attention metadata that are storedin a repository. The user can indicate whether search resultsare relevant or not and that feedback is also stored in theattention metadata repository. Search results are filteredand ranked based on earlier interaction by the user and byother users in her social network, as made available throughOpenSocial. Although we need to do more user evaluations,the first results are very encouraging [11, 20].

8. DATA BASED RESEARCH ON LEARN-ING

On a meta-level, learning analytics provides exciting op-portunities to ground research on learning in data and totransform it from what is currently all too often a collectionof opinions and impossible-to-falsify conceptualizations and

Page 6: Attention Please! Learning Analytics for Visualization and

At the top of the dashboard (label 1), there is the option of filtering per application. The modification of this filter affects all visualizations. The charts are also interlinked. Table 1 presents which actions trigger updates of other visualizations.

Table 1 Actions overview

Section Action triggered Affected visualizations

Sent Information

1 Selecting an application 2,3,4,5 Name of the widget

2 Restricting a period of time

3,4,5 Starting date

Ending date

3 Selecting a day of the week

2,4,5 Day of the week

4 Selecting a type of action 5 Type of Action

5 Selecting a type of item 4

5. USE CASE: XMPP CHAT BEHAVIOR This use case describes the behavior of a specific widget in a PLE environment, deployed during a course at RWTH Aachen University during the period May to July 2010. After this period, the environment was occasionally used in an informal way. In this PLE, four widgets were used. The widgets use Open Social [13] for their communication in a PLE.

• ABC Testing widget. This widget was only used during the first two weeks (this information is also displayed in the dashboard).

• Cam Widget. This widget tracks the Open Social communication and translates this communication to CAM. Users can deactivate or activate tracking of their data.

• Role Web 2.0 Knowledge Map. This widget allows to search for articles by entering keywords.

• XMPP Multiuser Chat. This widget enables chat functionality between different users based on the XMPP technology.

In this use case, we will focus on the XMPP Multiuser Chat widget because it is the most active in terms of event communication providing us more information about its particular characteristics. We will now explain how we can derive the conclusions from: 1. Detect changes on usage patterns: When we select theXMPP

Multiuser Chat in part 1 one of Figure 2 and we obtain an overview of the overall activity (Figure 3). The annotated time line chart (Figure 3) enables us to see that the activity was concentrated during the period from May to July 2010. After this period, the activity was reduced considerably. In the “events per type of action” chart (Figure 3), we can see that people enter to room chats more than sending messages (if we

Figure 3 XMPP Multiuser Chat visualization

Figure 2 CAM Dashboard overview

Figure 5: The CAM dashboard [27]

Visualizing Activities for Self-reflection and Awareness 5

Fig. 1. The user interface with the 3 di↵erent visualizations.

The parallel coordinates in visualization B are a common way to displayhigh-dimensional data [10]. On the vertical axes, we show (i) the total time spenton the course, (ii) the average time spent on a document, (iii) the number ofdocuments used and (iv) the average time of the day that the students work. Astudent is represented as a polyline connecting the vertices on the vertical axes.The position of the vertex on the i-th axis corresponds to the i-th data pointcoordinate. For example, the green line (the logged in user) works on averagein the early evening and is spending an average time in line with the majority.He does not use so many di↵erent documents and on average looks at these fora short time. He scores the worst here. The ‘average’ student of the class (inyellow) is also calculated. This is a much more advanced visualization but canprovide a good overview of the tendencies in the behavior of the students.

Figure 6: The Student Activity Monitor (SAM) [13]

Page 7: Attention Please! Learning Analytics for Visualization and

include search results from content on your personal socialnetworks (e.g. Twitter and Google Buzz). Bing collaborateswith Facebook to provide similar functionalities: searchingin the profiles of your friends and in content liked by yourfriends. IBM also added social features based on its em-ployee directory, tags, bookmarking behavior and ratings toits intranet to show more relevant blogs, wikis, forums andnews to great success [2, 8]. Social search is also a popu-lar research topic. Haystaks [9] extends mainstream searchengines with new features: users can create sharable searchresult folders and the content of these folders is used for rec-ommendations. I-SPY [10] allows people to create searchcommunities and based on the queries and results in thesecommunities it will adaptively re-rank the search results.

To enable search over multiple repositories, one can applyfederated or harvested search, which collects all the meta-data from all repositories in a central repository for faster re-trieval [11]. When working with vast repositories of web 2.0sources (e.g. YouTube), it is impossible to apply harvestedsearch. As far as federated search is concerned, Ariadne [11]focuses on the interoperability between different reposito-ries. The search engine relies on the Standard Query Lan-guage (SQI) to ensure interoperability and to offer a trans-parent search over a network of repositories. Ariadne alsoprovides harvesting. This is applied in the GLOBE networkwith 13 repositories3. MetaLib and WebFeat provide feder-ated search over scientific content [12]. WebFeat sends thequery to all search engines and then shows the results in allthe native user interfaces. In contrast with MetaLib that usesit’s own UI and communicates with the repositories over thestandardized Z39.50 protocol. ObjectSpot [13] is originallya federated search widget for scientific publications, but itis now extended for web 2.0 sources. It uses a cover den-sity algorithm to rank the search results, which can also bemanually re-ordered. Basic recommendations based on theuser selection of search results are also provided. ExtendingObjectSpot with social features was not trivial due to the useof incompatible technology. None of these federated searchengines provides social features or social search results. Justas mainstream search engines are exploring social networksand attention metadata (voting & sharing), we adopted thisstrategy in our federated search engine.

THE DESIGN OF THE WIDGETIn this section, the software architecture behind the searchwidget is explained and the user interface (UI) design andimplementation is discussed.

Software architectureTo enable search over multiple data sources, we employ aclient-server architecture, as shown in Figure 1. When theuser enters a search term, it is sent to the federated searchservice (step 1), which transmits it to all the different datasources concurrently (step 2). Currently it queries YouTube,SlideShare.net, Wikipedia, GLOBE and the OpenScout repos-

3The Global Learning Objects Brokered Exchange (GLOBE) al-liance, http://www.globe-info.org

Figure 1. The client-server architecture of the federated search andrecommendation service.

itory4, but extra sources can be easily added to support dif-ferent learning scenarios. When all the results are returnedfrom the data sources (step 3), the federated search servicere-ranks the results (step 4) based on the metadata with theApache Lucene library [14]. The ranked results are returnedto the widget in the ATOM format [15] (step 5), which en-ables us in the future to make the service OpenSearch com-pliant [16]. OpenSearch allows search engines to publishsearch results in a standard and open format to enable syn-dication and aggregation. In future, we plan to adapt theservice to return results every time the repositories returnthem to improve search speed. Once the widget receives thesearch results, they are presented to the user and the searchresult URLs are sent to the recommendation service (step 6).This service will return recommendations, based on the at-tention metadata stored in the database. The recommenda-tions are sent back to the widget, where they will be pre-sented to the user. The user can interact with the search re-sults, e.g. preview a movie inside the widget or (dis)like it.When some of these interactions happen, they are trackedand the attention metadata (basically the user, the URL ofthe search result and the action) is sent to the recommen-dation service (step 7). The service then stores the attentionmetadata in a database (step 8) to be able to calculate recom-mendations later. The next section describes the recommen-dation algorithm in more detail. The client-server architec-ture enables us to expose repositories not openly accessibleby deploying the service inside the intranet.

User Interface DesignThe main design goal was to provide a simple, clean searchinterface with visually rich search results to enable betterdecision making while selecting a search result. The widgetprovides a simple Google-like search interface over multipleweb 2.0 data sources. Although advanced search settingsare available (see Figure 3), they are not visible by default.Morville et al. [2] advice this as well, because advancedsearch is often used by expert users. Figure 3 shows the ad-vanced search settings where the wanted media types, repos-itories and social recommendations can be configured. Thiscan be operated with the wrench icon.

The search results are presented in a uniform way (see Fig-ure 2): basic metadata and tags together with a screenshot

4The OpenScout repository, http://www.openscout.net/demo

2

Figure 7: Storing attention metadata in federated search [11]

theories [23].

As a precursor to making that happen, it is important thatwe agree on ways to share data sets, in an ”open science”approach [26, 9, 14]. That is why a group of interested re-searchers has started an initiative around ”dataTEL” (http://www.teleurope.eu/pg/groups/9405/datatel/ [34]. Themain objective is to promote exchange and interoperabilityof educational data sets.

9. CONCLUSIONOf course, one of the big problems around learning ana-lytics is the lack of clarity about what exactly should bemeasured to get a deeper understanding of how learningis taking place: typical measurements include time spent,number of logins, number of mouse clicks, number of ac-cessed resources, number of artifacts produced, number offinished assignments, etc. But is this really getting to theheart of the matter?

Moreover, there are serious issues about privacy when de-tailed data of learner interactions are tracked [4]. An in-teresting early approach to deal with these issues was theproposal of the no longer active not for profit ”Attention-Trust” [35]: their guiding principles included

• property: the data about a person’s attention remainthe property of that person;

• mobility: it should be possible to move data about aperson out of one system and into another system -see also google’s recent data liberation initiative (seehttp://www.dataliberation.org/);

• economy: a person should be able to sell data abouthis attention;

• transparency: it should always be clear to a personthat she is being tracked.

Especially that last principle seems key: tools like ghostery(http://www.ghostery.com/) enable a user to know whenshe is being tracked on a web site. As we evolve towards aworld where not only learning activities, but virtually every-thing will be tracked [2], this issue is likely to become evenmore important.

Some people are quite concerned about the ”filter bubble”that personalization and recommendation engines may cre-ate [22]: we agree that there is a certain danger there, butwe also believe that more advanced algorithms and ethicalreflection can help us to address these issues.

In any case, we believe that learning analytics can be usedto put the user in control, not to take control away in anIntelligent Tutoring Systems kind of way, by using attentionto filter and suggest, provide awareness and support sociallinks.

10. ACKNOWLEDGMENTSThis research has received funding from the European Com-munity Seventh Framework Programme (FP7/2007-2013)under grant agreement no. 231396 (ROLE) and no. 231913(STELLAR). Much more importantly, the support, com-ments and feedback from my team and students have thoughtme much more than I will ever be able to teach them.

11. REFERENCES[1] C. Anderson. The End of Theory: The Data Deluge

Makes the Scientific Method Obsolete. WiredMagazine, 16(7), 2008.

[2] G. Bell and J. Gemmel. Your life, uploaded. Plume,2010.

[3] J. Biehl, M. Czerwinski, G. Smith, and G. Robertson.FASTDash: a visual dashboard for fostering awarenessin software teams. In Proceedings of the SIGCHIconference on Human factors in computing systems,pages 1313–1322. ACM, 2007.

[4] D. Boyd. Facebook’s Privacy Trainwreck: Exposure,Invasion, and Social Convergence. Convergence: TheInternational Journal of Research into New MediaTechnologies, 14(1):13–20, 2008.

[5] C.-H. Chang, M. Kayed, M. R. Girgis, and K. F.Shaalan. A Survey of Web Information ExtractionSystems. IEEE Transactions on Knowledge and DataEngineering, 18:1411–1428, 2006.

[6] N. Corthaut, S. Lippens, S. Govaerts, E. Duval, andJ.-P. Martens. The integration of a metadatageneration framework in a music annotation workflow.Oct. 2009.

[7] A. Croll and S. Power. Complete Web Monitoring.O’Reilly Media, Inc., 2009.

[8] E. Duval and K. Verbert. On the role of technicalstandards for learning technologies. IEEE

Page 8: Attention Please! Learning Analytics for Visualization and

Transactions on Learning Technologies, 1(4):229–234,Oct. 2008.

[9] J. Fry, R. Schroeder, and M. den Besten. Open sciencein e-science: contingency or policy? JOURNAL OFDOCUMENTATION, 65(1):6–32, 2009.

[10] M. H. Goldhaber. The Attention Economy and theNet. First Monday, 2(4), Apr. 1997.

[11] S. Govaerts, S. E. Helou, E. Duval, and D. Gillet. AFederated Search and Social RecommendationWidget. In Proceedings of the 2nd InternationalWorkshop on Social Recommender Systems (SRS2011) in conjunction with the 2011 ACM Conferenceon Computer Supported Cooperative Work (CSCW2011), pages 1–8, 2011.

[12] S. Govaerts, K. Verbert, D. Dahrendorf, C. Ullrich,S. Manuel, M. Werkle, A. Chatterjee, A. Nussbaumer,D. Renzel, M. Scheffel, M. Friedrich, J. L. Santos,E. Duval, and E. L.-c. Law. Towards Responsive OpenLearning Environments : the ROLE InteroperabilityFramework. In ECTEL11: European Conference onTechnology Enhanced Learning, Lecture Notes inComputer Science, 2011.

[13] S. Govaerts, K. Verbert, J. Klerkx, and E. Duval.Visualizing Activities for Self-reflection andAwareness. In Proceedings of the 9th internationalconference on Web-based Learning, pages 91–100.Springer, 2010.

[14] T. Hey and A. E. Trefethen. Cyberinfrastructure fore-Science. Science, 308(5723):817–821, 2005.

[15] U. Kirschenmann, M. Scheffel, M. Friedrich,K. Niemann, and M. Wolpers. Demands of ModernPLEs and the ROLE Approach. In M. Wolpers,P. Kirschner, M. Scheffel, S. Lindstaedt, andV. Dimitrova, editors, Sustaining TEL: FromInnovation to Learning and Practice, volume 6383 ofLecture Notes in Computer Science, pages 167–182.Springer, 2010.

[16] X. Ma, G. Chen, and J. Xiao. Analysis of An OnlineHealth Social Network. In Proceedings of the 1st ACMInternational Health Informatics Symposium, pages297–306. ACM, 2010.

[17] N. Manouselis, H. Drachsler, R. Vuorikari,H. Hummel, and R. Koper. Recommender Systems inTechnology Enhanced Learning. In F. Ricci,L. Rokach, B. Shapira, and P. B. Kantor, editors,Recommender Systems Handbook, pages 387–415.Springer US, 2011.

[18] M. McKeon. Harnessing the web informationecosystem with wiki-based visualization dashboards.IEEE transactions on visualization and computergraphics, 15(6):1081–8, 2009.

[19] J. Najjar, M. Wolpers, and E. Duval. Contextualizedattention metadata. D-Lib Magazine, 13(9/10), Sept.2007.

[20] X. Ochoa and E. Duval. Use of contextualizedattention metadata for ranking and recommendinglearning objects. In CAMA06: Proceedings of the 1stinternational workshop on Contextualized attentionmetadata: collecting, managing and exploiting of richusage information, pages 9–16, 2006.

[21] H. Poldoja. EduFeedr-following and supportinglearners in open blog-based courses. In Proceedings of

OpenEd2010, number 2010. Universitat Oberta deCatalunya, 2010.

[22] E. Pariser. The Filter Bubble: What the Internet IsHiding from You. Penguin Press, 2011.

[23] K. Popper. The Logic of Scientific Discovery.Routledge, 1959.

[24] A. S. Rath, D. Devaurs, and S. Lindstaedt. UICO: anontology-based user interaction context model forautomatic task detection on the computer desktop. InProceedings of the 1st Workshop on Context,Information and Ontologies, CIAO ’09, pages8:1—-8:10, New York, NY, USA, 2009. ACM.

[25] C. Romero and S. Ventura. Educational data mining:A survey from 1995 to 2005. Expert Systems withApplications, 33(1):135–146, July 2007.

[26] S. S. Sahoo, A. Sheth, and C. Henson. Semanticprovenance for eScience - Managing the deluge ofscientific data. IEEE INTERNET COMPUTING,12(4):46–54, 2008.

[27] J. L. Santos, K. Verbert, S. Govaerts, and E. Duval.Visualizing PLE Usage. In Proceedings of EFEPLE11:1st Workshop on Exploring the Fitness andEvolvability of Personal Learning Environments.CEUR workshop proceedings, 2011.

[28] H. Schmitz, M. Scheffel, M. Friedrich, M. Jahn,K. Niemann, and M. Wolpers. CAMera for PLE. InU. Cress, V. Dimitrova, and M. Specht, editors,Learning in the Synergy of Multiple Disciplines,volume 5794 of Lecture Notes in Computer Science,pages 507–520. Springer, 2009.

[29] B. Schwartz. The paradox of choice - Why more isless. HarperCollins, 2007.

[30] C. Shirky. Here Comes Everybody: The Power ofOrganizing Without Organizations. Penguin Press,2008.

[31] E. Singer. The Measured Life. Technology Review,2011.

[32] M. Swan. Emerging patient-driven health care models:an examination of health social networks, consumerpersonalized medicine and quantified self-tracking.International journal of environmental research andpublic health, 6(2):492–525, Feb. 2009.

[33] S. Ternier, K. Verbert, G. Parra, B. Vandeputte,J. Klerkx, E. Duval, V. Ordonez, and X. Ochoa. TheAriadne Infrastructure for Managing and StoringMetadata. IEEE Internet Computing, 13(4):18–25,July 2009.

[34] K. Verbert, E. Duval, H. Drachsler, N. Manouselis,M. Wolpers, R. Vuorikari, and G. Beham.Dataset-driven Research for Improving TELRecommender Systems. In 1st InternationalConference on Learning Analytics and Knowledge,Banff, Canada, 2011.

[35] M. Wolpers, J. Najjar, and E. Duval. Workshop reporton the international {ACM} workshop oncontextualized attention metadata: collecting,managing and exploiting rich usage information (cama2006), June 2007.

[36] M. Wolpers, J. Najjar, K. Verbert, and E. Duval.Tracking actual usage: the attention metadataapproach. Educational Technology and Society,10(3):106–121, 2007.