1. Introduction
Digital technology is transforming the way in which Cultural Heritage (CH) is produced, presented, and experienced. Accelerated digital evolution in the form of massive digitization and annotation activities along with actions towards multi-modal cultural content generation from all possible sources has resulted in vast amounts of digital content being available through a variety of cultural institutions, such as museums, libraries, archives and galleries. In addition, the evolution of web technologies has contributed to making the Web the core platform for the circulation, distribution and consumption of a broad range of cultural content. It has been estimated that 300 million objects from European heritage institutions have been digitized, representing 10% of the region’s cultural heritage, out of which only about one-third is available online (Europeana Strategy 2020–2025,
https://pro.europeana.eu/page/strategy-2020-2025-summary).
In the last two decades, a number of initiatives at organizational, regional, national and international level have focused on aggregating and facilitating access to digital cultural content, giving rise to a number of thematic, domain-based as well as cross-domain CH hubs and web platforms, such as the European Digital Library Europeana (
www.europeana.eu), the Digital Public Library of America (DPLA,
https://dp.la), and many others. These initiatives aim on one hand to streamline the aggregation process and make it easier for CH Institutions to prepare and share high-quality content and, on the other, to engage users from different audiences-from educators and creatives to researchers and the general public-via a number of added-value services that make content make readily available for browsing, search, study, and reuse.
However, due to the complex, heterogeneous and multi-channel aggregation workflow and shortcomings in the data providing process, many of the digital resources served through the numerous web platforms suffer from poor metadata descriptions. The lack and insufficient quality of structured and rich descriptive metadata highly affects the accessibility, visibility and dissemination range of the available digital content, and limits the potential of added-value services and applications that reuse the available cultural material in innovative ways, consequently limiting the user experience provided by CH platforms.
Metadata quality improvement and enrichment is a major challenge that receives increasing attention in the digital cultural heritage domain and is among the top priorities of Europeana’s strategy. CH Institutions have put significant efforts in improving the quality of their collections’ metadata, however, the efficiency of such efforts is compromised by a problem of scale: improving or even adding new metadata to hundreds of thousands or even millions of records coming from different sources requires significant investment in time, effort, and resources which organizations cannot usually afford. Recent advances in the field of Artificial Intelligence (AI) and the availability of a variety of off-the-shelf AI content and semantic analysis tools offer remarkable opportunities for overcoming the bottleneck of scale, by providing capabilities for analyzing almost any amount and type of data and extracting useful metadata with minimal time and resources needed. For example, there is a plethora of methods are used for Named Entity Recognition and Disambiguation [
1,
2,
3], tools for extracting different kinds of features from audiovisual content (e.g., for object and location recognition [
4], multimedia event recognition [
5], audio analysis [
6] etc.), and many others that can be used for automatic enrichment in different contexts in the CH domain and beyond.
Past and ongoing attempts to take advantage of automatic enrichment tools in the field of CH have demonstrated the great potential of AI techniques, e.g., entity linking for the enrichment of Irish Historical Archives [
7], in the Apollonis (
https://apollonis-infrastructure.gr) project for digital humanities, and the Visual Recognition for Cultural Heritage project (
https://www.projectcest.be/wiki/Bestand:VR4CH_rapport_1-0.pdf). However, resorting to purely automatic methods has also revealed a number of technical and methodological limitations that deter the effectiveness, scalability, and reuse potential of automatic enrichment tools as well as the degree to which these are exploited by the CH institutions. Firstly, the accuracy of results is negatively affected by the fact that the automatic enrichment tools have been trained and tuned on corpora outside the CH sector. The availability of human-annotated data can produce a considerable improvement in accuracy, however, the acquisition of appropriate labeled data is a costly process. Moreover, parties interested in exploiting state-of-the-art AI tools for enriching their datasets, lack a streamlined and scalable process for evaluating the automating enrichments and for deciding which are acceptable for being published and presented on their platforms. Manual validation of all automatic enrichments is costly and evidently does not scale up, with existing tools lacking the necessary functionalities to support this process.
2. Materials and Methods
Figure 1 provides an architectural overview of the main components of the overall CrowdHeritage ecosystem and their interactions, hinting also to the workflows that the ecosystem can support. At the core of the system stands the CrowdHeritage crowdsourcing platform (
crowdheritage.eu), which connects the human actors and the products of their intelligence with the software components. A detailed description of the internal architecture of the CrowdHeritage platform is provided in
Section 2.1.
The CrowdHeritage platform is interconnected with all other components of the ecosystem:
The automatic enrichment tools, which currently include three different AI tools for extracting features from images and for named entity identification. The tools interact with the crowdsourcing platform along two directions: they can supply it with datasets that have been automatically enriched and ask the crowd to validate (down- or up-vote) the automatic annotations; and they can take datasets annotated by humans as ground-truth data. The benefits of this interchange are reciprocal. On the one hand, the human annotators are guided and facilitated in their tasks, which become more focused and less cumbersome. On the other hand, automatic tools can be trained and tweaked on good-quality labeled CH-specific corpora and thus improve their accuracy by taking into account the special characteristics of this domain. More information about the AI tools and their interaction with human intelligence is provided in
Section 2.5.
The MINT open web-based data aggregation platform [
23], which is used by more than 550 CH organizations and 8 Europeana aggregators for the aggregation and management of their metadata records and it is the main component used for ingesting datasets to Europeana. The platform offers a user and organization management system that allows the operation of different aggregation schemes (thematic or cross-domain, international, national or regional) and registered organizations can upload (via HTTP, FTP, or OAI-PMH) their metadata records in XML or csv serialization. MINT offers a visual mapping editor that enables users to map their dataset records to a desired XML target schema. The role of MINT in the CrowdHeritage ecosystem is two-fold: to feed the crowdsourcing platform with datasets of metadata records in the right format to be enriched; and, vice versa, to insert the outcomes of the crowdsourcing process, i.e., annotations (see
Section 2.4), as enrichment-additions to the original metadata records and to ultimately publish the enriched datasets to the CH presentation platforms. It should be mentioned that the data loaded via MINT can be either new datasets provided by content providers or datasets sourced from the Europeana platform, thus enabling the improvement of already published metadata.
The CH presentation platforms, which serve the CH data to the end-user, offering a number of added-value services (e.g., search, collection views, etc.). The CrowdHeritage platform is connected to the Europeana Search and Record API and can source cultural resources stored in Europeana in order to make them available for crowdsourcing. Besides being interlinked with Europeana, interconnection with systems and platforms maintained by CH organizations and especially national and domain aggregators themselves for the custom organization and presentation of their content, such as the EFHA platform (
https://fashionheritage.eu/browse), is also supported via MINT. Through the Validation Editor of the crowdsourcing platform (see
Section 2.3), campaign organizers can review the added annotations and decide which are acceptable for being published and presented on their platforms and on Europeana. The CrowdHeritage workflow supports two different ingestion routes for publishing the enrichments to Europeana: besides publishing them as part of augmented metadata records via MINT as already mentioned, there is also the option to submit them as separate annotations by making use of the Europeana Annotation API (
pro.europeana.eu/page/annotations). The later route is more appropriate in the cases where content providers prefer to keep the added tags decoupled from the original metadata records.
2.1. The Main Platform Internal Components
The development of the CrowdHeritage crowdsoucing platform was conducted in collaboration with stakeholders who were involved in the definition of the user scenarios and evaluation tasks by actively participating in the necessary discussions in order to concretize the functional requirements for the technological platform, identify the content for the campaigns and carefully shape every detail regarding the campaigns’ execution. Cultural Heritage Institutions (CHIs) and associations such as the ModeMuseum Antwerpen (
https://www.momu.be), the Network of European Museum Organisations (
https://www.ne-mo.org) and the Philharmonie de Paris, strongly influenced the objectives and outcomes of CrowdHeritage, and broader target audiences (students, teachers) provided useful feedback for the proper functioning of the CrowdHeritage platform. The development of the platform was implemented in line with the agile principles, where three versions of functional requirements were developed, each one directing the next development sprint and taking into account the evaluation results of the previous iteration. Both the frontend and the backend of the platform are deployed on servers maintained by the National Technical University of Athens, who is also responsible for the overall administration and maintenance of the CrowdHeritage platform. The source code is available on GitHub (
https://github.com/ails-lab/crowdheritage), licensed under the Apache 2 license. For the backend, the Play Framework (
https://www.playframework.com) is used, following the model–view–controller architectural pattern, and the code has been written in Java and Scala. The frontend uses the Aurelia (
https://aurelia.io) JavaScript client framework, along with Node.js (
https://nodejs.org) as a runtime environment. The languages and technologies used for the development are JavaScript, Less.js, and HTML5.
The CrowdHeritage platform consists of three basic internal components: (i) the content aggregation and collection management system; (ii) the Crowdsourcing Web Spaces, where end-users can navigate through selected collections of cultural records and add different annotation types (see
Section 2.2 for more details); and (iii) the administrative user interfaces facilitating the design, customization, and validation process of the campaigns (described in
Section 2.3). The internal platform architecture and the connections between its components is illustrated in
Figure 2.
The backend layer of the platform, and particularly its content aggregation and collection management system, is built on top of the WITHCulture platform (
www.withculture.eu) [
24,
25], which provides access to digital CH resources which provides access to digital cultural heritage items from different repositories and offers a number of added-value services for the structuring and creative reuse and of that content. By making use of the WithCulture capabilities, the CrowdHeritage platform can collect CH-related data from various sources and take advantage of the federated and faceted search services that allow for the simultaneous search of multiple searchable CH repositories such as Europeana, DPLA, DigitalNZ (
www.digitalnz.org), the Rijksmuseum (
www.rijksmuseum.nl), and others, giving access to a huge set of heterogeneous items (images, videos, different metadata schemata etc.). The aggregated CH data is converted into a homogeneous data model compatible with the Europeana Data Model (EDM,
https://pro.europeana.eu/page/edm-documentation), organized into thematic collections by one or more collaborators and stored in a NoSQL database. Opting for a NoSQL database enabled the maintenance of large volumes of sourced data, which are often semi-structured and follow different complex schemas depending on the content provider, while at the same allowing for more flexibility and easier updates to the target schema for representing CH records used by the platform.
The same item stored in the WithCulture database can become part of more than one thematic collection, since each collection consists of a set of references to items along with additional metadata referring to the collection itself (e.g., title, description, creator etc.). A thematic collection can then be made available for crowdsourcing by linking it to a specific campaign setup via the Campaign Editor (see
Section 2.3) and for automatic enrichment with the use of appropriate AI tools as described in
Section 2.5. The same collection can become an input to more than one campaign and a campaign may refer to the enrichment of more than one collections (see also
Section 2.3). Each item can be associated with multiple annotations, coming from different annotators and campaigns, following the Annotation Model described in
Section 2.4, which contains all the necessary information that links each item with all the collected annotations referring to it.
The CrowdHeritage platform has been designed to fully support multilingualism in a dynamic way, both with regard to the platform’s interface, including campaign descriptions and instructions, as well as the with regards to the support for multilingual vocabularies and thesauri. Currently, the interface is available in English, Italian and French and can easily support any language as long as appropriate translations are provided. The annotating process also currently supplies translated tags in the three above languages and the uploading of any multilingual vocabulary can be supported, in order to present the labels and auto-complete functionalities to the user in the language they have selected.
2.2. Functionalities for End Users
Through the platform’s landing page, the unregistered user can have a look at the list of ongoing, completed or upcoming campaigns and browse through each campaign via the dedicated Crowdsourcing Web Spaces. A web space consists of a set of campaign-specific pages, including a summary page that provides an overview of the campaign, presenting its goal, progress, and relevant statistics and collection pages for browsing the involved CH items in a contextual setting, where the item view only showcases annotations concerning the respective campaign along with the original metadata.
By browsing through the Crowdsourcing Web Spaces, the registered CrowdHeritage users are able to contribute to an ongoing campaign via a simple and user-friendly interface and with a quick learning curve. The content served by the platform is organized under thematic collections of cultural records enabling end-users to navigate, choose a collection, browse through the records and their metadata, and select the ones they wish to enrich, by adding annotations to them. The underlying annotation model (see
Section 2.4) is expressive enough to cover a large variety of different representations, from simple textual tags to linking to Web Resources of various formats, while the User Interface (UI) is designed to serve to the end-user the different structures supported by the model in a comprehensive and functional way. In this respect, the UI currently facilitates the semantic annotation of records with terms from controlled online vocabularies and thesauri, color-tagging, and geotagging items.
The annotation process begins with the users browsing through the thematic collections of a campaign and selecting one to contribute to. By choosing the collection they wish to annotate, the users are presented with a grid consisting of the collection’s items, with the option to filter out the ones they have already annotated. The annotating process begins with the user clicking on an item of the grid and work their way in the collection in a serial way. Alternatively, the participants can choose to be presented with a (currently random) selection of items for enrichment. In semantic tagging, users can tag the records by typing the desired tag into the relevant text field, which displays a list of suggested terms derived from the selected thesauri or vocabularies, supported by an auto-complete functionality. Color-tagging is accomplished by selecting the desired color from a palette of available colors and in geotagging users can pinpoint their location on a map. The users can also validate existing annotations by up-voting or down-voting them, depending on whether or not they agree with them, or even delete their own annotations. The annotation process for the end-users is illustrated in
Figure 3.
The CrowdHeritage platform also provides information and statistics about each campaign, like its percentage-based progress depending on the set annotation goal, the total count of the contributors, and gamification elements such as leaderboards consisting of the most active users, with the aim to make crowdsourcing a transparent and engaging experience. The user is encouraged to add more annotations via a point and earning and reward system: usually, they gain two points for new annotations and one point for an up-vote or down-vote, pursuing the gold, silver or bronze badge which, depending on the campaign, can be accompanied with a prize. On every campaign page, CrowdHeritage provides statistics for each user regarding their contribution to the campaign, e.g., the total number of new annotations they have added, down-voted or up-voted, the number of digital cultural objects they have annotated, and their ranking in the campaign leaderboard, based on the awarded points for their contribution and determining the badge they have earned. Furthermore, the user karma points are calculated based on the percentage of upvotes versus downvotes that the tags inserted by the user have received. The karma percentage can be seen as a means of peer-reviewing that provides a quick way of identifying malicious users, who may insert quick and unrelated annotations in order to gain points.
Non logged-in users can browse the list of the available campaigns with basic information: title, description, banner and thumbnail, start date, end date, contributors, annotation target, current number of annotations, and percentage of completion and can filter the list of campaigns according to their status. By opening a campaign, more detailed information and statistics appear as well as a grid with the collections available for crowdsourcing containing their records, visualization (photo, video, sound), metadata, and existing annotations. The campaign leaderboard is also visible, illustrating the most active users of the campaign.
2.3. Administrative Functionalities
The platform also offers administrative functionalities for the campaign organizers, including a custom Campaign Editor and a Validation Editor. By taking advantage of these functionalities, a user who has been granted administrative permissions can launch their own customized crowdsourcing campaign in order to enrich selected digital cultural collections by deriving annotations from the public and finally moderate the campaign results by validating the produced annotations.
The Campaign Editor enables the setup, editing, deletion, and preview of custom crowdsourcing campaigns. It allows the administrative user to specify the appearance of the campaign, the content to be used, and the annotation process to be followed. The user can define a set of features, such as the title, description and duration, choose a banner, and select the content to be used in the campaign either by directly importing a Europeana collection or by searching into Europeana and curating their own collections. Subsequently, they are able to design the desired annotation process by setting a target for the campaign, selecting from a wide variety of vocabularies and thesauri to be used, and choosing the type of desired annotations from semantic tagging, geotagging or color-tagging. They can also compile the instructions for participants and describe the prizes for the top contributors.
New campaigns are set up and launched in a fully configurable and dynamic way. Each campaign initialized through the editor interface, is stored as a separate entry in the database, which includes all the information specified by the organizer. Based on these parameters, the set of campaign-specific HTML pages that constitute the Crowdsourcing Web Space (see
Section 2.2) is automatically constructed and assigned a dedicated URL based on the campaign’s name, e.g.,
https://crowdheritage.eu/en/garment-classification. The implementation also enables the campaign-specific navigation of collections, so that, depending on the URL via which a CH collection is accessed (e.g., via
crowdheritage.eu/en/garment-classification/collection/5daac5aa4c74793bb0b68a40), the end-user is presented only with the annotations resulting from the respective campaign. The new campaign is also added to the list of campaigns on the platform’s landing page.
At the moment, the platform UI supports the use of several widely used Linked Data vocabularies, datasets, and ontologies from which campaign organizers can choose to use as the campaign’s vocabulary. These include the Getty Art and Architecture Thesaurus (AAT,
https://www.getty.edu/research/tools/vocabularies/aat), the General Multilingual Environmental Thesaurus (
www.eionet.europa.eu/gemet), the Musical Instruments Museums Online (MIMO) thesaurus (
www.mimo-db.eu), the Europeana Fashion thesaurus (
http://thesaurus.europeanafashion.eu), the KULeuven Photography Vocabulary (
http://bib.arts.kuleuven.be/photoVocabulary), Wordnet (
wordnet.princeton.edu), and knowledge bases such as Wikidata (
www.wikidata.org), DBpedia (
wiki.dbpedia.org) and Geonames (
www.geonames.org). In all the above vocabularies and datasets, each resource is always accompanied by one or more textual labels, possibly in several languages. These labels provide textual representations for the specific resource and are used for indexing the resources and facilitating term lookup.
In the platform backend, a Thesauri Manager has also been developed to support, through an offline process, the import of more Linked Data vocabularies and datasets that can be subsequently selected as a campaign’s vocabulary. The Thesauri Manager converts the imported vocabularies from their source format (e.g., SKOS thesauri, OWL ontologies, N-triples datasets) to a common JSON EDM-consistent representation, stores them in the database, and indexes them to allow fast search and retrieval.
The Validation Editor provides to campaign organizers access to an interface via which they can review the annotations produced by the crowd or automatic tools (see
Section 2.5) and filter them according to their own acceptance criteria. Moderation is necessary in cases where expert knowledge is required on top of the crowd contribution, ambiguity needs to be resolved (e.g., the dominant color of an outfit) or some correct, yet unhelpful information needs to be removed (e.g., records were tagged with obvious but too general annotations such as “womenswear”). Through the validation interface, the organizers can view the popular tags of campaigns, click on them, find out the records tagged with each term, and un-tag the irrelevant records assuring useful and valid annotations. A visualization of the process is depicted in
Figure 4. Through the Validation Editor the campaign organizers can also view the profile of the campaign participants, including their karma points, and thus providing a means for identifying misbehaving participants.
4. Conclusions
The CrowdHeritage ecosystem offers an end-to-end solution that couples machine-driven enrichment tools with the power of collective human intelligence, mobilized via a user-friendly crowdsourcing platform that supports the organization and launching of engaging campaigns in the CH sector. This way, it provides an efficient way for making the complete workflow of high-quality cultural data supply easier, more scalable, and cost-effective, thus streamlining and simplifying the work of aggregators and data providers. Besides providing better services for CH institutions, the CrowdHeritage ecosystem also contributes a step forward to enhancing the way digital cultural heritage is experienced by end users. The more comprehensive the metadata accompanying a digital heritage object, the more likely it is to be viewed, understood, and used by educators, creatives, culture lovers, researchers and citizen at large. By facilitating the improvement of metadata quality, the CrowdHeritage ecosystem and platform enable users to effectively discover what they are looking for, browse and go deeper into a subject, and understand its context and interconnections with other cultural heritage objects. At the same time, the CrowdHeritage platform provides the technological means to stimulate a more participatory approach to cultural heritage and engage experts as well as the general audience in its improvement.
The current implementation of the CrowdHeritage ecosystem and platform is a first important step that opens up multiple directions for future extensions, improvements, and reuse possibilities. From the implementation-technical perspective, a number of new features and improvements are planned to make the platform more functional, widen the use cases it can support, and enhance user experience. Future work towards such features include a user interface for the dynamic upload of custom vocabularies and thesauri; addition of personalization features so that so that the items and tasks suggested to the participants are selected taking into account the user’s history; extensions to the Validation Editor so as to support more advanced filtering and moderation based on certain criteria to be specified by the user (e.g., based on the automatic confidence levels, the popularity of an annotation, etc.). Furthermore, more consistent efforts have to be invested into the adoption of appropriate user-moderation tools and policies. Although malicious behaviors are usually detected through our karma-points system, we also plan to provide functionality for users to directly report such behaviors. Afterwards, system moderators can manually investigate the validity of each report and act accordingly. Automatic spam and policy-violation detection methods are also being considered for campaigns in which permissible contributions by participants are not controlled via a vocabulary, i.e., allow the entry of free text.
From the more research-oriented perspective, further work and experiments need to be directed towards closing the loop of the active learning cycle [
21]. To this end, a methodology that defines a set of selection criteria for the datasets and tasks to be assigned to humans, quality thresholds etc. is yet to be specified. Elaborating on preliminary results with respect to the fine-tuning of the GEEK entity extraction tool, more extensive experiments need to be conducted by using the acquired CH-specific ground-truth datasets, in order to gain deeper insight on how the tool behaves under different parameterizations and in comparison with the general-purpose datasets. Moreover, we plan to integrate to the ecosystem and consider in our human-in-the-loop methodology more AI tools that focus on extracting different types of elements from different types of content depending on the needs of case studies from the CH domain, e.g., object extraction from image and video, speech to text, Optical Character Recognition, etc.
Considering the possibilities of impact on the CH and citizen engagement field, another interesting direction for future work concerns extending the CrowdHeritage platform to support scenarios that go beyond the enrichment of metadata towards inviting the end-user to contribute with genuine thoughts and content and associate CH collections with interpretations, emotions, and other items that can ultimately lead to richer and multi-vocal perceptions of CH collections.
Author Contributions
E.K. is mainly responsible for the conception and design of the overall framework and has written crucial parts of the current article. M.R. has written most parts of the conference paper on which the current extended article relies on, especially with respect to text describing concrete functionalities. S.B., M.R., and O.M.-M. have been the main developers of the actual tools mentioned herein and have all contributed with text describing the respective functionalities. V.T. and G.S. have been overseeing and guiding the implementation process and have provided essential feedback and made revisions to the current article.All authors have read and agreed to the published version of the manuscript.
Funding
This work has been co-financed through (i) the Connecting Europe Facility (CEF) project CrowdHeritage: Crowdsourcing Platform for Enriching Europeana Metadata (No A2017/1564820); (ii) the Greek Operational Program Competitiveness, Entrepreneurship and Innovation, under the call RESEARCH–CREATE–INNOVATE (T1EDK-01728–ANTIKLEIA); and (iii) from the European Union’s Horizon 2020 research and innovation program under grant agreement No 770158. The sole responsibility of this publication lies with the authors. The European Union is not responsible for any use that may be made of the information contained therein.
Informed Consent Statement
Informed consent was obtained from all subjects involved in the user evaluation.
Data Availability Statement
Not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Ganea, O.; Ganea, M.; Lucchi, A.; Eickhoff, C.; Hofmann, T. Probabilistic Bag-Of-Hyperlinks Model for Entity Linking. In Proceedings of the 25th International Conference on World Wide Web WWW, Montréal, QC, Canada, 11–15 April 2016; pp. 927–938. [Google Scholar]
- Piccinno, F.; Ferragina, P. From TagME to WAT: A new entity annotator. In Proceedings of the First ACM International Workshop on Entity Recognition & Disambiguation ERD’14, Gold Coast, QLD, Australia, 11 July 2014; Carmel, D., Chang, M., Gabrilovich, E., Hsu, B.P., Wang, K., Eds.; ACM: New York, NY, USA, 2014; pp. 55–62. [Google Scholar]
- Mandalios, A.; Tzamaloukas, K.; Chortaras, A.; Stamou, G. GEEK: Incremental Graph-based Entity Disambiguation. In Proceedings of the Linked Data on the Web Co-Located with the 2018 Web Conference-LDOW@WWW, Lyon, France, 23 April 2018. [Google Scholar]
- Varytimidis, C.; Tsatiris, G.; Rapantzikos, K.; Kollias, S. A systemic approach to automatic metadata extraction from multimedia content. In Proceedings of the 2016 IEEE Symposium Series on Computational Intelligence-SSCI, Athens, Greece, 6–9 December 2016; pp. 1–7. [Google Scholar]
- Amid, E.; Mesaros, A.; Palomäki, K.; Laaksonen, J.; Kurimo, M. Unsupervised feature extraction for multimedia event detection and ranking using audio content. In Proceedings of the ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing-Proceedings, Florence, Italy, 4–9 May 2014; pp. 5939–5943. [Google Scholar] [CrossRef]
- Fillon, T.; Simonnot, J.; Mifune, M.F.; Khoury, S.; Pellerin, G.; Le Coz, M.; Amy de la Bretèque, E.; Doukhan, D.; Fourer, D.; Rouas, J.L.; et al. Telemeta: An open-source web framework for ethnomusicological audio archives management and automatic analysis. In Proceedings of the 1st Digital Libraries for Musicology workshop (DLfM 2014), London, UK, 12 September 2014. [Google Scholar]
- Munnelly, G.; Pandit, H.J.; Lawless, S. Exploring Linked Data for the Automatic Enrichment of Historical Archives. In Lecture Notes in Computer Science, Proceedings of the Semantic Web: ESWC 2018 Satellite Events, Heraklion, Greece, 3–7 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11155, pp. 423–433. [Google Scholar]
- Dijkshoorn, C.; Boer, V.; Aroyo, L.; Schreiber, G. Accurator: Nichesourcing for Cultural Heritage. Hum. Comput. 2017, 6. [Google Scholar] [CrossRef]
- Ralli, M.; Bekiaris, S.; Kaldeli, E.; Menis-Mastromichalakis, O.; Sofou, N.; Tzouvaras, V.; Stamou, G. CrowdHeritage: Improving the quality of Cultural Heritage through crowdsourcing methods. In Proceedings of the 15th International Workshop on Semantic and Social Media Adaptation and Personalization (SMAP 2020), Zakynthos, Greece, 29–30 October 2020; pp. 1–6. [Google Scholar]
- Stiller, J.; Petras, V.; Gäde, M.; Isaac, A. Automatic Enrichments with Controlled Vocabularies in Europeana: Challenges and Consequences. In Digital Heritage. Progress in Cultural Heritage: Documentation, Preservation, and Protection; Ioannides, M., Magnenat-Thalmann, N., Fink, E., Žarnić, R., Yen, A.Y., Quak, E., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 238–247. [Google Scholar]
- Navarrete, T. Crowdsourcing the digital transformation of heritage. In Digital Transformation in the Cultural and Creative Industries; Massi, M., Marilena Vecco, Y.L., Eds.; Taylor and Francis: Abingdon, UK, 2020; Chapter 4; pp. 99–116. [Google Scholar]
- Mourelatos, V.; Tzagarakis, M.; Dimara, E. A review of online crowdsourcing platforms. South-East. Eur. J. Econ. 2016, 14, 59–74. [Google Scholar]
- Simpson, R.J.; Page, K.R.; Roure, D.D. Zooniverse: Observing the world’s largest citizen science platform. In Proceedings of the 23rd International World Wide Web Conference, WWW ’14, Seoul, Korea, 7–11 April 2014; pp. 1049–1054. [Google Scholar]
- Ellis, A.; Gluckman, D.; Cooper, A.; Andrew, G. Your Paintings: A nation’s Oil Paintings Go Online, Tagged by the Public. In Museums and the Web Conference; Available online: https://www.museumsandtheweb.com/mw2012/papers/your_paintings_a_nation_s_oil_paintings_go_onl.html (accessed on 3 February 2021).
- Karl, R.; Roberts, J.; Wilson, A.; Möller, K.; Miles, H.; Edwards, B.; Tiddeman, B.; Labrosse, F.; La Trobe-Bateman, E. Picture This! Community-Led Production of Alternative Views of the Heritage of Gwynedd. J. Community Archaeol. Herit. 2014, 1, 23–36. [Google Scholar] [CrossRef] [Green Version]
- Vincent, M.; Flores Gutiérrez, M.; Coughenour, C.; Manuel, V.; Bendicho, V.M.; Remondino, F.; Fritsch, D. Crowd-sourcing the 3D Digital Reconstructions of Lost Cultural Heritage. In Proceedings of the 2015 Digital Heritage, Granada, Spain, 28 September–2 October 2015; pp. 171–172. [Google Scholar]
- Dhonju, H.; Xiao, W.; Mills, J.; Sarhosis, V. Share Our Cultural Heritage (SOCH): Worldwide 3D Heritage Reconstruction and Visualization via Web and Mobile GIS. ISPRS Int. J. Geo-Inf. 2018, 7, 360. [Google Scholar] [CrossRef] [Green Version]
- Chun, S.; Cherry, R.; Hiwiller, D.; Trant, J.; Wyman, B. Steve.museum: An Ongoing Experiment in Social Tagging, Folksonomy, and Museums. In Museums and the Web 2006: Proceedings; Archives and Museum Informatics: Toronto, ON, Canada, 2006. [Google Scholar]
- Oomen, J.; Aroyo, L. Crowdsourcing in the Cultural Heritage Domain: Opportunities and Challenges. In Proceedings of the 5th International Conference on Communities and Technologies, Brisbane, Australia, 29 June–2 July 2011; pp. 138–149. [Google Scholar]
- Gligorov, R.; Hildebrand, M.; Van Ossenbruggen, J.; Aroyo, L.; Schreiber, G. An Evaluation of Labelling-Game Data for Video Retrieval. In Proceedings of the 35th European Conference on Advances in Information Retrieval, Moscow, Russia, 24–27 March 2013. [Google Scholar]
- Gilyazev, R.; Turdakov, D. Active learning and crowdsourcing: A survey of annotation optimization methods. Proc. Inst. Syst. Program. RAS 2018, 30, 215–250. [Google Scholar] [CrossRef]
- Inel, O.; Khamkham, K.; Cristea, T.; Dumitrache, A.; Rutjes, A.; van der Ploeg, J.; Romaszko, L.; Aroyo, L.; Sips, R. CrowdTruth: Machine-Human Computation Framework for Harnessing Disagreement in Gathering Annotated Data. In Lecture Notes in Computer Science, Proceedings of the ISWC 2014-13th International Semantic Web Conference, Riva del Garda, Italy, 19–23 October 2014; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8797, pp. 486–504. [Google Scholar]
- Drosopoulos, N.; Tzouvaras, V.; Simou, N.; Christaki, A.; Stabenau, A.; Pardalis, K.; Xenikoudakis, F.; Kollias, S. A Metadata Interoperability Platform. In Museums and the Web 2012; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
- Chortaras, A.; Christaki, A.; Drosopoulos, N.; Kaldeli, E.; Ralli, M.; Sofou, A.; Stabenau, A.; Stamou, G.; Tzouvaras, V. WITH: Human-Computer Collaboration for Data Annotation and Enrichment. In Proceedings of the Companion of the The Web Conference 2018, Lyon, France, 23–27 April 2018; pp. 1117–1125. [Google Scholar]
- Giazitzoglou, M.; Tzouvaras, V.; Chortaras, A.; Alfonso, E.; Drosopoulos, N. The WITH platform: Where culture meets creativity. In Proceedings of the Posters and Demos Track of the 13th International Conference on Semantic Systems-SEMANTiCS 2017, Amsterdam, The Netherlands, 11–14 September 2017. [Google Scholar]
- Hoffart, J.; Yosef, M.; Bordino, I.; Fuerstenau, H.; Pinkal, M.; Spaniol, M.; Taneva, B.; Thater, S.; Weikum, G. Robust Disambiguation of Named Entities in Text. In Proceedings of the EMNLP 2011-Conference on Empirical Methods in Natural Language Processing, Edinburgh, Scotland, UK, 27–31 July 2011; pp. 782–792. [Google Scholar]
- Menis-Mastromichalakis, O.; Sofou, N.; Stamou, G. Deep Ensemble Art Style Recognition. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar] [CrossRef]
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).