Academia.eduAcademia.edu
What’s with the Free Images? A Study of Flickr’s Creative Commons Attribution Images Herkko Hietanen Kumaripaba Athukorala Antti Salovaara Helsinki Institute for Information Technology, HIIT PO BOX 19215 00076, Aalto, Finland Helsinki Institute for Information Technology, HIIT PO BOX 19215 00076, Aalto, Finland kumaripaba.athukorala@hiit.fi Helsinki Institute for Information Technology, HIIT PO BOX 19215 00076, Aalto, Finland antti.salovaara@hiit.fi herkko.hietanen@hiit.fi ABSTRACT Our survey of the Flickr photo site’s Creative Commons attribution licensed images reveals that there is a wide variety of high quality of relevant stock images available. However, searching the images can be demanding since the image metadata is inconsistent. The main problem of finding open images is that the search tools are mostly based on user generated tags. The search results would benefit from human sorting and simple machine vision analysis. These steps might be able to close the gap between commercial stock photo and open image collections. Categories and Subject Descriptors H.3.3 [Information Systems]: Information Storage and Retrieval – information search and retrieval. General Terms Management & Human Factors. Keywords Flickr, Creative Commons, image search, ImageNet, stock photo, metadata, tagging. 1. INTRODUCTION In the not-too-distant past, image searching was a time consuming and costly exercise. Now there are stock photos online on services such as Getty and Corbis Images. Online microstock photo companies have opened the stock photo markets to a new class of amateur and semiprofessional photographers. The increased supply of stock images has reduced their price considerably [8] [1]. The commercial services are no longer the only sources for quality stock images. There are tens of millions of images online which are licensed free of charge with open content licenses. The falling price of the stock photos combined with the growth of open content images raises questions: Is the zero-price the result of the process of commodification of stock images? Can the open images compete with the commercial stock photo services? Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. MindTrek’11, September 28-30, 2011, Tampere, Finland. Copyright 2011 ACM 978-1-4503-0816-8/11/09....$10.00. While the license price of open images is zero, there are transactional costs to finding the right images. These transaction costs could increase the opportunity costs of open image collections and eventually make the paid stock photo services the less expensive alternative. In this research we examine whether the Creative Commons (CC) images are accurately retrievable and how their quality compares to commercial stock images. We do this to determine how much work needs to be invested to improve the open image repositories to the level of professional stock photo sites. 2. THE STATE OF THE ART There exist several commercial photo collections which sell professional quality images by professional photographers. The images are also screened by paid editors before publishing. However, the services have slightly different approaches to selecting the images they sell. Magnum Photos is a photographic cooperative owned by its photographer-members. It has a long tradition of photo journalism. Magnum has an online archive of over 500,000 photos. Photographers retain their copyright to photographs which Magnum licenses. Magnum recently started using crowdsourcing methods to create image tags. Their crowdsourcing tests have shown that workers optimally assign between four and eight keywords per image --- any more resulted in duplication [10]. The online image trade is dominated by relatively new companies like Getty and Corbis Images. Getty has grown aggressively to become one of the biggest image services by acquiring traditional photo agencies and archives, by digitizing their collections and enabling easy online distribution. While much of Getty’s income derives from professional photography, Getty bought one of the biggest microstock photo providers, iStockphoto, in 2007. Unlike Magnum, iStockphoto has very liberal entry requirements for its contributors. At iStockphoto any photographer can choose to upload their images to the Web site. The catalogue of over 6 million files is crowdsourced [4] mainly by amateurs and semiprofessionals. iStockphoto acts as an agency by selling the images to clients. Individual photographers make small profits per download, while iStockphoto takes a portion of the profits. Many of the contributors are motivated by money [1]. In 2011 iStock expects to pay over 2 million dollars in royalties per week to contributors [9]. Amateurs have uploaded nearly 6 billion images to online photo sharing service Flickr. Over 190 million of the images are licensed with one of the Creative Commons (CC) licenses. Each of the six CC licenses has a different set of restrictions. The most restrictive license permits only unmodified use of images for noncommercial usage. The most permissive CC-Attribution (CC-By) license requires licensees merely to attribute the author. Previously photographers have used the more restrictive CClicenses, but Flickr’s data shows that recently users have started to prefer more permissive licenses [7]. In 2011, over 25 million of the CC images are licensed with CC-By licenses. Flickr provides three easy ways to filter search the Creative Commons licensed images: by using the options in the Advanced Search, by browsing through recent images uploaded under each type of licence or through the Flickr API which allows other applications to access the images and their metadata freely. Flickr provides a way to sort the search results according to their relevance and interestingness. Flickr determines the interestingness by determining where the clickthroughs are coming from, who comments on the images and who marks them as favorites and by examining image tags, description, title and several other factors which are constantly changing [2]. It is well known that image labels originating from user tags have a number of problems. Previous research shows that amateur photographers typically optimize their image sharing for browsing and not for searching [6]. Their tags often provide an incomplete description of the visual content, focusing mainly on the interest of the photographer while leaving out many “plain”, yet visible, objects. At the same time, and for the same reason, a large proportion of tags may refer to information not directly visible in the image. There are several initiatives to improve search results which are based on textual user created links. ImageNet [3] is a repository of over 3 million images built upon the backbone of the WordNet ontology. The images are quality controlled and human annotated through crowdsourcing methods. ImageNet is designed for the purpose of supporting object recognition, image classification and automatic object clustering applications. This design goal makes the collection less interesting for regular users for two reasons: 1) Although, ImageNet has clean and full resolution images, most of them are “all rights reserved” images and thus not freely usable. 2) While ImageNet has over 80,000 noun word sets, it does not contain verbs or adjectives. of ImageNet. Unlike ImageNet, we did not restrict our searches to nouns only. We used WordNet’s 147,306 term ontology as a list of search terms to retrieve CC-By licensed images from Flickr. Our importer script retrieved hundred images that Flickr determined to be the most relevant [1] for each search term and their metadata. We gave the retrieved images points according to the order in which they showed up in the search results. We gave the first image 99 points and the second 98 points. This way the first 99 images received a relevancy score. The importer also retrieved all the tag words that the photographer had associated with the image. Those tags were not issued any relevancy score unless they were among the first hundred search results for that given word. Our database is a collection of images with both weighted WordNet terms and additional Flickr tags. 147,306 searches retrieved the total of 5,671,643 images. Thirty parallel importer instances retrieved one million images per week. The searches returned an average of 39 images for each Wordnet keyword searched. Several images showed up multiple times in the search results. Because of this, only 2,866,612 of the retrieved images were unique. The retrieved images had 28,200,124 tags and there were 1,170,349 different terms. The average image had 9.8 tags while the median was 7. Only a few images had more than a thousand tags each. The highest number of tags used in an image was 1886. Originally the image had only 29 tags, but it had an 8894-word essay attached into the image description field. Because Flickr’s searches are not limited only to the tags but analyzed also the image name and description fields, this image showed up in several searches. Figure 1 shows the number of images against the number of tags for images. Many of the images were poorly tagged, with only 30.6% of the images having more than 10 tags. Given that a fairly high percentage of these tags are uninformative (e.g., meaningful only to the photographers), the number of truly useful tags is low. For example the most used tag was “2010” which was associated with 53,806images. There were 6,386 tags that had more than 500 images associated to them. MIRFLICKR is another image retrieval test data set. It is a subset containing 25.000 of Flickr’s CC-By licensed images with their original tags and EXIF image metadata. MIRFLICKR has also been developed for training of content-based image retrieval systems [5]. Unlike ImageNet, MIRFLICKR’s collection is not built on any ontology and it has only ten basic categories tagged and verified by experts. 3. METHOD AND RESULTS Our two tests collected information about the quality of open images. Our automated script made nearly 150,000 searches to the Flickr’s CC-By licensed images. The goal of this experiment was to collect information about the metadata that these open images contain. In the second experiment we conducted a study where participants rated the quality of the search results of different collections. 3.1 Retrieval test In the first step of this research we built a dataset of images and matching metadata. The approach we followed was similar to that Figure 1. Number of Tags per Image 3.2 User study In order to evaluate the retrieval results from Flickr, we carried out a small-scale user study. Ten participants were asked to review results from three image services. We asked the participants to mark suitable and mismatching images, compare the search results and sort them in order. Our user study (N = 10) examine the quality of search results in 3 image services: iStockphoto, Flickr, and Flickr-CC-By. We inspected Flickr’s collection of the week’s most interesting images, and used them to determine a list of 10 keywords (sunset, woman, flower, sea, bird, child, city, tree, car, and landscape). For these keywords, we retrieved the first 100 search results from each of the three services. This resulted in a dataset comprising 3 x 10 x 100 = 3000 images which we printed in full color on thirty A3sized papers, each printout containing the search results from one service and a keyword. These papers showed only the images and the search term – we had removed all the extraneous graphics and texts from the search results. To hide the origins of the images from the participants, we also made sure that the papers did not reveal the name of the service used. We asked our participants to imagine a situation in which they needed to prepare a PowerPoint presentation with images that match our list of keywords. We asked them to inspect the printouts and to perform two tasks: 1) select five images that they would use in a power point presentation and 2) strike out the images that did not match the keyword. We advised the participants to strike out the images that had poor image quality, or poor match with the given keyword. In the first part of the user study, with each participant having rated the retrieved images for all 3 services and 10 keywords, we could compare the answers given by each participant individually and then combine these analyses into the overall result. Therefore, we ran a repeated-measures ANOVA with 3 x 10 within-subjects design, with the number of strike-outs as the dependent variable. The results showed a difference between the results for different keywords (p < .05). This meant that there was a statistically significant difference between search results for different keywords: the retrieved images were better rated on some keywords than others. When comparing the results on the keyword level, we found no general factor that all good-quality or bad-quality search results would have shared in common. For instance, “landscape” and “sunset” images got generally speaking few strike-outs, but “sea”, which is another keyword producing scenery, received high numbers of strike-outs. A statistically stronger and more interesting result was that there was also a difference in retrieval quality between the services (p < .01). Of the 100 images, the numbers of images struck out for iStockphoto, Flickr and Flickr-CC-By were 10.4, 16.4, and 18.3, respectively. Only the difference between iStockphoto and the other two services was statistically significant; no difference was found between Flickr and Flickr-CC-By. In other words, iStockphoto was found the best of the three services, and Flickr and Flickr-CC-By were found equally good. The second part of the study contained a more general rating task. For each keyword, we gave the participants the three clean 100image printouts (with no strike-outs drawn over images), one for each service, and asked them to sort the printouts in order of increasing retrieval quality. Participants could rate the printouts as "equally good", having "slight difference" or having "clear difference". These differences were turned into scores for each service. The lowest-rated service always scored 0, the next service was scored either 1 or 2, depending on the level of difference, and similarly with the third service, when compared to the second one. Each service was therefore given a score between 0 and 4. For each keyword, each service was finally given another score that told the difference of its 0...4 score to the average of the scores for that keyword. We used the same analysis as in the first part of the study. The difference scores were used as dependent variables in a 3 x 10 within-subjects repeated-measures ANOVA. The result from this overall analysis was the same as from the first part of the study: iStockphoto was significantly better than the Flickr services, and again there was no statistical difference between the Flickr services. While evaluating the results in the second part of the study, we asked the participants to tell the reasons for their ordering. Common reasons for low ratings were that search results were too homogenous. For instance, 99/100 of the images retrieved by iStockphoto with the keyword "woman" portrayed women of a Caucasian origin, and a big percentage of the "city" images were skylines typical of American cities. Flickr-CC-By suffered from results coming from the same content provider and the same image set, with similar results. Another commonly stated reason was the keyword being in a secondary role in the retrieved results. iStockphoto’s "flower" pictures mostly pictured flowers in female subjects' hands or hair, for instance. The third common reason for poor rating was the presence of unnecessary content in the images. Flickr and Flickr-CC BY suffered often from this, because their images were mostly unpolished photographs. There were also several duplicate images in both Flickr and Flickr-CCBy image sets. 4. CONCLUSIONS The fact that our 147,306 searches to over 25 million CC-By images returned only under 3 million images suggests that most of the images cannot be easily found through common searches. It means that only 12 percent of the CC-By licensed images are easily findable. One thing that could affect the results is the language. We used only English words. Flickr is currently available in ten languages and many of its users may have used non-English tags, rendering the images inaccessible to us. The other more likely explanation is that even though people choose to share their images with permissive licenses they don’t bother to tag the images well enough for them to be found. Our tests did show that the professional iStockphoto service provides considerably better search results in nearly all of our test searches. However, the difference between 3 million CC-By licensed images and the 6 billion general Flickr images suggest that the problem is not that there are not enough good quality open images available, but that the image search needs refinement. CC-By licensed pictures did not do worse than the All rights reserved pictures in our user evaluation. Therefore, a free image service is a realistic goal. On the other hand, there was a clear difference in search results between Flickr image sets and that of iStockphoto. Efforts should therefore be directed to improve search results and filtering in Flickr image searches. Simple selection of the images either by experts or by employing crowdsourcing methods could help to close this gap. The feedback from our participants also suggests that prioritizing lighter colored images in image retrieval could improve the appeal of the search results. We believe that future work should therefore concentrate on creating models and incentives to harness crowds and machine vision to improve the quality of search results. Over time such improvements could help to attract a critical mass of people to open image collection users. Such crowds could lead to emergent network effects that would unlock the value of open photos. It is still too early to say whether the open content image collections are the final stage of the commodification of stock photos. However, the race to the bottom is fierce, and stock photo companies need to innovate in order to compete with the open image sources. 5. REFERENCES [1] Brabham, D., Moving the crowd at iStockphoto: The composition of the crowd and motivations for participation in a crowdsourcing application, First Monday [Online], Volume 13 Number 6 (21 May 2008), available at http://firstmonday.org/htbin/cgiwrap/bin/ojs/index.php/fm/art icle/view/2159/1969 [2] Butterfield, D., Fake, C., Henderson-Begg, J., Mourachov, S., (inventors) Interestingness ranking of media objects, Assigned to Yahoo! Inc. US Patent Application 20060242139 Published October 26, 2006. [3] Deng, J., Dong, W., Socher, R., Li L.-J., Li. K., Fei-Fei, L. 2009. ImageNet: A large-scale hierarchical image database. Computer Vision and Pattern Recognition, 2009, 248–255. [4] Howe, J., The rise of crowdsourcing, Wired, volume 14, number 6 (June), available at http://www.wired.com/wired/archive/14.06/crowds.html,. [5] Mark J. Huiskes and Michael S. Lew. 2008. The MIR flickr retrieval evaluation. In Proceeding of the 1st ACM international conference on Multimedia information retrieval (MIR '08). ACM, New York, NY, USA, 39-43. DOI=10.1145/1460096.1460104 http://doi.acm.org/10.1145/1460096.1460104 [6] David Kirk, Abigail Sellen, Carsten Rother, and Ken Wood. 2006. Understanding photowork. In Proceedings of the SIGCHI conference on Human Factors in computing systems (CHI '06), Rebecca Grinter, Thomas Rodden, Paul Aoki, Ed Cutrell, Robin Jeffries, and Gary Olson (Eds.). ACM, New York, NY, USA, 761-770. DOI=10.1145/1124772.1124885 http://doi.acm.org/10.1145/1124772.1124885 [7] Linksvayer, M., Creative Commons licenses on Flickr: many more images, slightly more freedom. Creative Commons news post (2010), available at www.creativecommons.org/weblog/entry/20870. [8] Taub, E., When Are Photos Like Penny Stocks? When They Sell. New York Times, June 5, 2007, available at http://www.nytimes.com/2007/06/05/technology/circuits/05s yndicate.html?em&ex=1181188800&en=687225a44f80273c &ei=5087%0A. [9] Thompson, K., Royalty Change Follow up, iStockPhoto forum, Sep 8, 2010, http://www.istockphoto.com/forum_ messages.php?threadid=252322. [10] Wolmurth, P., Magnum Photo’s tagging game, British Journal of Photography, 17 Feb 2011, available at http://www.bjp-online.com/british-journal-ofphotography/report/2027044/ magnum-photos-tagging-game.