jrtom: (safe cat)
http://www.netflixprize.com/

I so don't have time for this right now, but it sounds like a really interesting problem. Not sure I'm eligible to participate (my PhD advisor is apparently one of the judges, which I think explains a conversation that I had with him a couple of weeks ago involving a lot of intentional vagueness on his part), anyway, but I bet that I could do a pretty decent job if I had the time to sink into it. (Especially considering the data that they have but are not using...)

argh.
jrtom: (Default)
From [livejournal.com profile] fdmts: A Face Is Exposed for AOL Searcher No. 4417749

In essence, AOL recently released 20 million anonymized search queries. However, it turns out that it's not that hard to figure out who someone is based on what they're searching for, as the article details.

I deal with this sort of issue on a continuing basis as part of my profession (among other things, I do research on learning models for social network analysis). In some cases, the data is inarguably public: no one really minds if I analyze the network defined by the "writes-a-paper-with" relation. But in other cases, it's been drilled into the heads of researchers--supposedly--that anonymization is required in order to release data, and often in order to get it in the first place.

The problem is, of course, that clearly anonymization isn't sufficient in this case.

It's a tricky problem, of course; we can't do research if we don't have data to work with, and there are valuable things that can be learned from such data that _don't_ involve violating peoples' privacy. I guess the question is, if it's necessary to collect such data in the first place, and to study it, is there anything in addition to anonymization that can be done to prevent this sort of 'reverse-engineering' of someone's identity? (Obviously AOL shouldn't have released the data publically in the first place…but the point is that by current standards they probably thought that it wouldn't do any harm because it was anonymized.) Aggregating it isn't the answer, because then you lose much of the information that made the data valuable in the first place.

*ponder*
jrtom: (Default)
Amygdala: Blue in the Face

This is mostly a placeholder in case I come back to this later, but this blogger suggests that the reason why Bush & co. didn't get the warrants was that they were doing large-scale pattern analysis on the communications of tens of thousands of people (or more) . . . thus making acquiring warrants impractical at best.

This kind of analysis is precisely what I do in my research. I have no doubt whatsoever that I could get a job with the CIA or NSA to simply continue doing what I've been doing. Let me be clear: I don't think that there's anything ethically wrong with the research qua research; the evil, if any, is in how it is used.

But it still itches me.

Profile

jrtom: (Default)
jrtom

May 2011

S M T W T F S
1234567
891011121314
1516 1718192021
22232425262728
29 3031    

Syndicate

RSS Atom

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated 15 October 2025 22:56
Powered by Dreamwidth Studios