Tag Archives: algorithms

Reflections on the Tweetorial

Last week’s tweetorial was the first time I participated in a so-called “tweetstorm” on a topic. As I am not really an active Twitter user outside of this course it was a new experience for me. Reflecting on it I’ve noticed that the 140 character limit per tweet has very interesting and real consequences for a discussion and my own participation within it. Obviously, the limit causes people to express their opinions in very brief statements which can leave room for interpretation. To counteract  the limitations one can keep sending out tweets to get one’s message across – in my mind a rather inefficient way compared to other mediums such as blog posts. The consequence is that it will likely clog up the twitter feed and potentially drown out other voices. Another way is to think hard about how to best come up with an answer that is deliberately vague and open to interpretation yet still conveys meaning. I wasn’t too comfortable with overshadowing the conversation with too many messages (and I unfortunately couldn’t participate on Friday) but I tried to come up with messages that were appropriate for the medium.

tweet

The Tweet Archivist and Keyhole analyses of our tweetorial show a discrepancy in the number of posts people were willing to share. While I was on the lower end with 6 tweets, the top tweeter by far was my colleague PJ with 59 contributions.

piechart

As I was unavailable for most of Friday I unfortunately missed the peak of the discussion.

timeline

Once I came back, I was feeling overwhelmed by the fragmented nature of tweets and retweets on a variety of topics. People hadn’t just stuck to the questions Sian and Jeremy had prepared but instead switched to other topics as well, such as the topic of learning to code – as seen in this keyword cloud:

topics

Looking at the data sets that these analytical tools generate I can’t help but question their value in terms of how they can help us learn.  The word cloud is the best example of how data needs to be interpreted to create information, let alone to generate knowledge or wisdom. Atomising the conversation and displaying the frequency distribution of words visually might give an outsider a quick overview of the topics discussed but there doesn’t seem to promote much learning. Perhaps analytic algorithms will in the future be able to extract the meaning of such conversations and assist the learner in getting the gist of it but in their current state these analytical tools don’t seem to offer much value in terms of content.

There is, however, an interesting observation that we can glean from the analysis on a meta perspective: the social dynamics of the conversation.

user mentions

Compared to the tweet count from earlier we can see that Sian, even though she only posted half as many tweets as PJ was being addressed the most. As she is the tutor in this course this does not seem all too surprising but considering the scope that learning analytics could be scaled up to, identifying such influencers might turn out to be valuable meta data.

Overall I have to say that the algorithmic snapshots of our tweetorial have not given me any particularly valuable insight that could significantly support me in learning from the tweets. Perhaps more context aware algorithms will one day be able to better distill the meaning of such conversations. For now, the meta data, particularly regarding the social structure of the conversations, are the most useful parts generated by these analytic tools.

Comment on Exploring Algorithms by mkiseloski

Thanks Martyn and Jeremy for your comments!

@ Martyn
Thank you, I will check out Ghostery. It seems like one more useful tool to protect privacy on the internet.

@ Jeremy
You raise some very interesting points here! I wholeheartedly agree that the issues of privacy need to be taken much more seriously in our public discourse, regardless of whether our data are collected from a government entity or a private business. You are right when you say that information like “likes Latin American music” or “likes winter sports” on their own seem rather inconspicuous, but the point to be made here is that over the more such seemingly useless factoids merge to create a stunningly accurate profile. This reminds me of how after the Snowden leaks people tried justify the warrantless NSA surveillance programs citing that they only collected metadata when in fact metadata (who did you talk to, when were you in what place, where did you use your credit card how much money, where did you go regularly, who was with you during those times, etc.) can potentially present a much more accurate description than the content of phone calls.
I read recently that Uber could easily infer from their user data how likely someone was for having an affair with someone (and where) simply by looking at driving patterns of people regularly driving some place in the evening and driving back home in the early hours of the morning. Knowing that some private company can so easily obtain such sensitive information feels quite unsettling for me.
I think that algorithms first discover identity but that as they become more aware of your existing identity and as they form a filter bubble for you they tend to influence you with their suggestions, possibly shaping your identity as you interact with their services more and more. In any case I think that privacy has to be protected if we want to live in a free society.

from Comments for Mihael’s EDC blog http://edc15.education.ed.ac.uk/mkiseloski/2015/03/07/exploring-algorithms/#comment-609
via IFTTT

Exploring Algorithms

Algorithms play an increasing role in our online lives. In their neverending quest to accurately profile their users in order to maximise ad revenues IT companies employ more and more sophisticated data mining methods incorporating information from  your activity history, your friends, your location and even strangers with similar interests to you.

Ever since I’ve watched Gary Kovac’s shocking TED talk “Tracking the Trackers” I’ve become increasingly privacy conscious and have taken several precautions to not be as easily tracked. From activating the ‘Do not track’ option in my browser to installing privacy enhancing browser extension such as Disconnect to opting out of targeted ads within the Google settings.

Thanks to these steps my advertisement profile with Google is now rather unspecific. Without taking such measures however, Google has been able to profile people surprisingly well.

I actively try to avoid the personalised features that sites present me with. In Facebook I never use their EdgeRank algorithm that sorts my feed according to “Top Stories” – I use “Most Recent” instead – simply because most of the Top Stories unsurprisingly is paid content from pages I subscribed to, not posts from my friends. Another reason is that I prefer to keep an open mind and personalised filters tend to create a filter bubble which not only distort people’s view of the outside world according to their own preferences and beliefs, they also do so invisibly.

For this week’s exercise of exploring algorithms I have decided to take a look at YouTube, since I have a long standing history of using that site. My main interaction with the site is wih the “My Subscriptions” tab which is always more relevant (and recent) than the algorithmically populated “What to watch” feature.

Logged into my account, this is what my front page looks like

what_to_watch

I can immediately tell why YouTube is recommending these videos to me. All of these videos are closely related to videos I have watched on YouTube within the last 48 hours. I watched one “CinemaSins” video, one Pink Floyd song, one Kygo song, a fail video and a Strokes song. While the songs that I played were actively sought out the other videos showed up on my subscription feed which made me click them.

If I scroll further down, it seems that YouTube still takes the same 3 or 4 videos from earlier and shows more related videos. Additionally, it suggests videos by la belle musique, a channel I am already subscribed to. what_to_watch2

While the suggested videos generally meet my taste, they don’t necessarily entice me to watch them now, especially since they don’t lead me to interesting new channels which I might want to subscribe to.

If I log out and visit YouTube in an incognito mode I am greeted with the following suggestions.

what_to_watch3

None of these videos have any relevance to my search or watched videos history but looking at the channel names (Ad Council, RadioKRONEHIT) one can assume that these videos have been placed on the front page because someone had paid for it.

Let’s take a look at the comments section of YouTube which has long been famous for its disastrous reputation. Apparently Google sorts the comments according to your Google+ profile which I, however, never use. This shows in the comparison between the two comments pages of a random video I clicked on.

Logged in:
comments1

Logged out:
comments2

As we can see, probably due to the lack of usable profile data Google has from my Google+ account, the comments shown are the same both when logged in and logged out. Again, the combative nature of YouTube comments shines through once more. It seems that no matter how sophisticated the algorithms in place, unless YouTube actively censors comments, it will always fight an uphill battle against the culture that has developed within the YouTube comments universe.

In conclusion, suggestion algorithms like the one used by YouTube can somewhat enhance your user experience to a certain extent, provided that you are okay with sharing enough data about yourself. Given the problems associated with filter bubbles and privacy concerns however, at present I still prefer a carefully manually selected subscriptions list to algorithmically derived suggestions.