[open-science] Big and open data – who should bear data transfer costs?
Lukasz Bolikowski
l.bolikowski at icm.edu.pl
Sat May 17 08:06:46 UTC 2014
Dear all,
when compiling a list of big, open, publicly available, data sets for my
students to use in their projects, I recently stumbled upon an
interesting problem: as the cost of transferring a large data set from A
to B is not negligible and someone has to bear that cost, what does
"open" mean in case of "big data"?
For example, Amazon Web Services offer a treasure trove of data sets,
some on CC-BY or CC-BY-SA licenses:
http://aws.amazon.com/publicdatasets/
Understandably, Amazon charges for data transfers out of its
infrastructure. When you rent Amazon's infrastructure in the same
region in which the interesting data set is located, you're not charged
for the transfer (but you are charged for the machines you use).
In the recent rewrite of the Panton Principles website, initiated by
Michelle Brook (http://goo.gl/cq1SuD) open research data is currently
defined as "data [...] made available on the internet under licenses
that permit anyone to download [...] without financial, legal, or
technical barriers".
The quoted sentence is careful to require lack of financial barriers
only in the license, so charging for data transfers seems to be
compatible with openness.
A practical question: If, as a researcher or a research organization, I
want to publish a large data set and keep the "open" label, can I charge
for data transfers (plus amortization costs of data storage), or do I
have to cover them myself?
What are your thoughts?
Best regards,
Lukasz
--
Dr. Łukasz Bolikowski, Assistant Professor
Centre for Open Science, ICM, University of Warsaw
Contact details: http://www.icm.edu.pl/~bolo/
More information about the open-science
mailing list