The Cataloging and Conservation of Digital Survey in Archaeology: A Photogrammetry Protocol in the Context of Digital Data Curation

: The technological advancement of software and hardware and the lowering of the prices of instrumentation has made photogrammetry the preferred instrument for surveying activities in archaeological projects. Consequently, archaeological datasets have been enriched with 3D models of archaeological ﬁnds and structures. Each project has developed its work pipeline for raw data acquisition and the elaboration of models and their archiving and dissemination. In most cases, the pipeline is the result of empirical experimentation and is designed to act within the speciﬁc context of the project. To date, we still lack a shared method for a photogrammetric survey that derives from the speciﬁc design and techniques/contexts. This paper aims at proposing an approach for a shared 3D survey workﬂow for photogrammetry in archaeology. The general approach relies on the digital data curation framework for cultural heritage and encompasses several specialized tasks. We describe the general functions and processes and how they can be implemented in a practical workﬂow. As a proof of concept, we show how a preliminary release of the workﬂow has been applied in the context of the BeArchaeo project, for the acquisition, processing, export, modeling, archiving, and indexing of 3D models, resulting from photogrammetric surveys. A long-term aim is a methodological approach for different endeavors of cultural heritage.


Introduction
The objective of this contribution is the presentation of an acquisition methodology in the field of cultural heritage that is independent of the use of specific tools or software and compliant with the methods adopted by researchers in humanities and computer science. Over the last few years, numerous approaches have been based on the use of photogrammetry for the acquisition of artifacts and architectural structures both with field applications, as in the case of Catalhoyuk [1], and through the use of drones for overall environmental acquisitions [2]. The photogrammetric technique is also used in similar fields, such as paleontology [3] or geoarchaeology [4]. What is observed is the creation of operational structures and acquisition paths that are limited by a "contextual dependence", that is, the specific cases for which they were created, i.e., the underlying method relies on the technical tool in use and deploys operational workflows according to the following pipeline: the definition of the tools->preparation of the set->creation of different scenarios->processing of the photos [5]. Although there is a basis to the methodology adopted, it focuses exclusively on the elements of the physical characteristics of the system, such as the management of the lights, the morphological characteristics, and the hardware and software tools to be used [6] Looking for a repeatable and methodologically corroborated acquisition protocol is a pressing necessity, because of the interoperability of models and the exchange of good practices [7]. Here, we focus on the archaeological context, where photogrammetric acquisition has become of paramount importance. The punctual collection of diverse nature

Classification Types: The Absence of Standards
Data classification concerns the methods used for indexing the digital data that result from a photogrammetric acquisition and the consequent elaboration. The lack of a unified standard from the projects makes inter-communication hard, without considering that the end of a project often represents the loss of the interpretation keys of the classifications. Since attempts at the standardization of the classifications are under debate, we focus on the standardization of the classifications of the elements that produce the metadata (in this case, photographs, point clouds, and 3D models). This problem takes on new applicative aspects when the emphasis is placed on the interdisciplinary nature of a project, which is the case of digital projects involving cultural heritage. Photogrammetry is no exception. Archaeology faces a particular challenge with classification since it involves diverse research interests and record-keeping systems that frequently draw from, or intersect with, other academic fields beyond archaeology. This is partly because archaeological projects tend to be "itinerant," covering several domains and constantly evolving as research progresses (with the need to use data from different datasets) or short-term collected materials [25]. Added to this is the regulatory issue that although digital recording (often alongside paper recording) is ubiquitous in the discipline, current regulatory practices do not necessarily promote the retention, re-use, or understanding of data, as observed by Huggett when, analyzing the impact of big data in archaeological practice, he speaks of a third revolution in archeology which is not yet able to interpolate the previous techniques of annotation and data collection of a digital nature with the new tools available [26]. Classification in archaeology raises various issues, such as characterizing typologies and arranging artifacts to gain insights into past cultures [27]. A rigorous, consistent, and replicable classification method replaces heterogeneous artifacts with internally homogeneous data classes. This method reflects cultural concepts and ideas implemented by artisans and is analyzed using methods sensitive to the material form of cultural ideas [28].

Obsolescence
Another problem that has emerged from the use of new technologies in the field of cultural heritage and, in particular, in archeology, is obsolescence, i.e., the progressive aging of machines and software. Obsolescence makes projects unreachable after a short time and irremediably lost after replacements in hardware, software, or in the IT approach. One relevant example of this type is the Electronic Domesday Book, a project initiated by the BBC to celebrate the 900th anniversary of King William's original inventory of post-conquest Britain. In this specific case, the problems concerned the publication of the project on video discs that could only be read using a specific proprietary player-a software designed for personal computers, BBC Master, adopted at the time by the majority of schools and libraries in the United Kingdom and now hopelessly obsolete [29]. The project, which cost 2.5 million pounds, aimed to establish universally applicable standards and anticipate future technological advancements. However, the long-term forecast proved inaccurate, and the project's use was limited due to the obsolescence of the interface devices employed [30]. As early as 2013, Italian researchers sought to address device obsolescence by creating a path to insert photogrammetric surveys of selected artifacts from the excavation of Pyrgos Mavroraki in Cyprus into a virtual museum. The surveys were isolated from other 3D elements to recognize their unique nature and the modeling and processing that generated them [31]. What must be taken into consideration is that a threedimensional model of an archaeological find is substantially different from any other threedimensional model both from a computer graphics point of view and from its generative path that inextricably links it to the photographs that generated it in a relationship of mutual dependence and validation. Despite numerous attempts to address obsolescence in digital technologies, it is evident that a focus on adaptable methodologies is more effective than software-based approaches. Abstract methodologies are better equipped to adapt to technological advancements, whereas software-based solutions inevitably face obsolescence issues.

Interdisciplinarity
Interdisciplinarity introduces a linguistic, terminological, and methodological gap between the parties involved (digital sciences and humanities but also archaeometric disciplines), leading to misunderstandings or loss of information. Interdisciplinary research faces not only institutional obstacles but also cognitive constraints and conceptual barriers that limit collaboration across disciplinary boundaries. Lessons from cognitive science and the studies of scientific practices highlight the importance of domain specificity in science, and philosophers of science play a crucial role in understanding and addressing these barriers. Case studies of collaborations between ecologists and economists and between molecular biologists and systems biologists illustrate the cognitive barriers that hinder interdisciplinary work [32]. This issue has become so central as to push scholars to attempt to formalize a language of communication for "changing knowledge and disciplinary boundaries through integrative research methods in the social sciences and humanities", as established by the STERP project within the Bologna agreements of the 1999 and later described by Ruth Keeling [33]. The identification of communication methods that will be fungible and clear for all the professionals involved from different research fields is not only necessary for correct communication between the parties but also to guarantee the legitimization of the research for the scholars who will verify it. Krebs et al., analyzing the difficulties in interdisciplinary communication, have emphasized the need to identify "quantitative approaches" towards a "scientification of knowledge production" to create a common ground of exchange that does not limit the intrinsic nature of the respective research subjects [34]. Efforts to increase communicability between different domains have in the past caused tension precisely in the attempt to guarantee this legitimacy. On the one hand, the human sciences are considered, in the context of research, to be defined as more reflexive (the concept of reflexivity can be summarized as the tendency to refer internally to the research itself, as opposed to the connection with external elements, characteristic instead of the natural sciences) [35]. On the other hand, however, this has also provoked criticisms of those who observe a dominance of the natural sciences, which push the human sciences to use their methods and languages in an attempt to legitimize themselves [36] or proceeding through some empirical utility testing case by case [37]. What emerges is a need for an upstream adoption of a method that is shared by the entire research community, regardless of individual research areas [38,39]. The idea of abstraction can be useful in indicating formats that are then specialized into different disciplines while keeping a shared core representation.

Traceability: Relations and Legitimacy Features
Finally, traceability, with particular attention to archaeological data, refers to the necessity of accessibility to the history of the data processes. Since the circulation of the excavation results often takes decades, to which is added the non-compliance of the data released for digital processing, also requiring the entry and adaptation of large and expensive data [40]. Within this process that mixes subjectivity and objectivity, accuracy is measured in the traceability of each step and the richness of the recorded testimonies, but since archaeological excavation is not repeatable, the recording operations must be as accurate as possible. Scientific replicability in this field is uniquely guaranteed by access to primary data [41]. At present, in the context of archaeological projects, data acquisition phases are not fully recorded and primary data are rarely, if ever, published in full [42]. This statement also concerns the excavation reports, which are rarely integral in the conservation of primary information [43]. If we add to this the presence of multidisciplinary research teams (see also above), we can observe the production of highly differentiated datasets (paper files, digital files, databases, etc.), of which only a small part is formally disseminated [44]. Although we observe an attempt to organize the activity of the archaeological field and the consequent interpretations, according to the processes of what Londoño calls "reflexivity" [45] and their subsequent applications with contexts, such as those of Çatalhöyük [46,47], the practice is still substantially dominated by traditional data logging methodologies, generating a mixture of analog and digital techniques strongly influenced by limitations and subjectivity at each stage, as reported by Roosevelt et al. in 2015 [43]. This lack of traceability of the 3D model production process reduces the possibility of their verifiability by researchers external to the project that produced them, negatively influencing the legitimacy of the 3D models. This problem represents a concrete obstacle to the realization of that coincidence, or at least intrinsic dependence, between cultural heritage and its digital representation, as indicated by the 1989 UNESCO convention [48] and restated in 2009 [49]. The abstraction of the method is necessary to ensure that the development of the photo-detected 3D model generation protocol is traceable, and therefore every result produced can be verified, thus giving legitimacy to the work performed.
The mapping of the problems presented demonstrates the need for a resolution that takes into account these difficulties, aiming at a general approach. In this paper, we envisage a proposal that relies on an encompassing methodology, as already put forward by some works in the literature.

Towards an Encompassing Solution: The State of the Art
In recent years, the need to produce an encompassing solution has emerged in several operational pipelines, based on greater abstraction [50].
Remondino and colleagues aimed to identify common patterns in surveying methods across different contexts, with the goal of automating these repetitive tasks. However, they acknowledge that full automation may not currently be feasible [51]. The study addresses traceability and classification but only in a technical context, excluding acquisition stages and interdisciplinary concerns. It focuses on computer science methodology.
De Reu et al. propose a detailed methodology based, however, on a specific software (in this case Agisoft PhotoScan, AgiSoft LLC, 2011b), which focuses on the exclusive context of emergency excavations; moreover, no space is given in this context to a reflection on the fate of 3D models in the subsequent archiving and dissemination phases [52]. The work poses the problem of information classification and was also created as a clear and intelligible operational workflow for cultural heritage scholars. However, the proposal builds its methodology on specific software. There is not even a specific definition of the acquisition method and the fate of the original photographic data being archived.
Deseilligny et al. focus again on the analysis of specific software and develop a path that is, on the one hand, methodologically structured but, on the other hand, does not respond to the needs of the variegated landscape of existing projects that cannot adopt a single tool [53]. The work poses the problem of the need for a punctual acquisition methodology, and it is the only work among those analyzed that addresses the problem of the ultimate fate of the data and metadata in the archiving phase; furthermore, as it does not involve any specific software, it is immune to obsolescence issues. However, the entire method does not go into detail regarding the replicability of the method and is instead closely linked to the specific photographic instrument and does not explain the reasons for the methodological choices made. No indication is given on how to account for the photos underlying the survey.
The work of P. Hallot and M. Gil (2019) poses the problem of the acquisition method and the behavior of the camera, according to a pre-established scheme; however, these observations remain once again limited to a specific field of application and are not placed in a linked relationship with the subsequent stages of data development [54] and other solutions proposed for the photograph acquisition phase remain limited to highly contextual scenarios [55]. In other cases, we observe how instead the search for a shared method focuses on the photo-processing phase. In those cases, research has focused on finding and ranking the best software to use, based on the number of photos required, the image quality, or the algorithm used [56], and the comparison between the different software that perform the processing operations and become the central topic and the procedural pipeline that is built on these comparison considerations [57]. However, by operating in this manner, the 3D model is not managed holistically and the data are not generated with consideration for the project's overall function. Other projects have instead focused on the subsequent phase in which the 3D model obtained is manipulated and modeled, but even in these cases, rather than proposing an abstract method, in the literature consulted, this phase is also considered to be part of the exclusive remit of the technical, and the focus is on the definition of which software to select [58]. Koutsoudis et al. propose a workflow that, attempting to broaden the spectrum of possibilities, proposes a series of software that can be used but once again does not produce a theory for the final archiving [59]. The final step in which the model is exported from the editing software has also been analyzed, but again, the comparison with the literature places this phase in a purely technical context and the choice of the format is made in this case exclusively taking into account the final needs of the project (such as the preparation for three-dimensional printing [60]). Even where all the necessary steps are considered, as in the case of Ucakar et al. 2022 [61], the resulting pipeline responds to the specific needs of the project presented and does not attempt to make an abstract work for the definition of a general model applicable to different cases.
In general, these project-related approaches do not provide solutions that can be transversal to the needs of archaeological projects, lacking a general solution. We propose the deployment of the methodological workflow from a shared abstraction. Each phase of the digitization process, from the photographic acquisition to the metadata encoding and the consequent dissemination, is properly recorded. We consider the abstraction of the general processes of cultural heritage management by adopting the digital data curation (DDC) framework. The DDC has been developed to implement terminology exchange (overcoming interdisciplinarity barriers), avoid the dependency on specific software or hardware technicalities (addressing the obsolescence failures), and ultimately yield a consistent method of organizing information (shared classification terms).

Digital Data Curation as a Theoretical Framework for an Encompassing Method
This section introduce the digital data curation framework. After a historical summary, we address the application of the schema of the photogrammetry process.

Before Digital Data Curation
Before the advent of digital data curation, attempts to sequence the phases of photogrammetric surveying had been made since the technique's introduction in the early 2000s [62]. These attempts were particularly prevalent in the archaeological context [63], where photogrammetry's advantages over analog methods were apparent, and efforts were being made to address associated challenges [64]. Initially, a design was attempted for the cataloging, homologation, and indexing operations of original photographic material and 3D models, without considering the Digital Data Center's observations. This design had to address the different classification speeds inherent to archaeologists and IT scientists, who are responsible for designing operational workflows and conducting the surveys and subsequent processing, raising classification issues [65,66]. The first solution, therefore, tended to outsource the digital data of the find, placing it back in the debate on the management of additional metadata [67]. In general, in these proposals, one observes a postulation of an ideal working context that does not take into account the horizontal nature in which archaeological projects normally develop. For this reason, proposals aiming at the definition of a complete and comprehensive acquisition work pipeline (such as the inception protocol) depend on the assumption that the entire development path is "supervised" [68]. A more abstract and theoretical approach allows every single party involved in the project to self-determine its position on the pipeline which, referring to the broader definitions of the DDC, does not depend on a "central coordination".

Historical Development of DDC
The history of digital curation and the formation of its methodological process are intrinsically linked to the three phases of the Digital Curation Centre. The first phase (2004-2007) focused on the holistic and interdisciplinary transversality of data in order to bring together a research community that is not specific to a single discipline and takes into account a wide range of types of digital data creation methods. The second phase (2007-2010) focused on data care and conservation, developing connection networks between institutes, libraries, and museums for shared and interoperable terminology (e.g., case studies from the Digital Preservation Coalition to the SCARP12 project [69]). The last phase, albeit officially completed in 2016, triggered an abstraction of the processes developed by the individual institutes for the encoding of methods and approaches to be adopted by an international community of professionals. The applications work on data management from the beginning of information collection and ensure effective data validation and active use for the entire lifecycle. Therefore, the major parameters for interpreting the Digital Data Curation universe are interdisciplinary and holistic data representation planned from scratch. Effective methods of data validation, conservation, and access for their active use shared the abstraction of processes, which can be implemented in several ways [70]. To fulfill the purpose of maintaining and preserving and enriching digital data, the DDC and the subsequent reflections of the Digital Curation Center have developed a glossary that defines the lifecycle and curation of digital data, and we refer to this glossary in the exposition of our method. Among these terms, it is necessary to remember some that will be used in the exposition phase of the present method: where the glossary speaks of "Conceptualizing", it refers to the conception and planning of the digital object, including the data acquisition methods, such as the archiving options. The term "Create" means the production of digital objects and related metadata. The expression "Access and use" refers to the availability of access to digital data; while "Evaluating and selecting" refers to the curation of digital objects as well as the selection of objects that need such attention. The term "Dispose" refers to the disposal of digital objects or parts thereof not included in the previous selection, while the term "Ingest" refers to the transfer of the digital objects to an archive or any other final destination. The "Conservation actions", as the expression suggests, represents those actions undertaken to guarantee the conservation of the digital object.

Towards a DDC Scheme for Photogrammetry
Digital data curation refers to the active management and preservation of digital data throughout its lifecycle. The fundamental principles of digital data curation include ensuring the authenticity, integrity, and reliability of the data, as well as its accessibility and usability by future users. The literature recognizes various phases in digital data curation, including the acquisition, appraisal, selection, preservation, maintenance, and dissemination of data. These phases involve a range of activities, such as data identification, documentation, organization, metadata creation, and storage in trusted repositories. The aim of digital data curation is to ensure that valuable digital data is preserved in the long-term and is accessible and usable by future generations of researchers.
Among the objectives of the projects that refer to the observations of the Digital Curation Center, there is not only the schematic systematization of the work phases but also the preservation of the stability of the data by avoiding unwanted transformations to place the active nature of the data at the mercy of the curator [71]. In parallel with the creation of a system of categories aimed at managing and understanding data, digital data curation has set itself the goal of abstracting these concepts to ensure the total transparency of data processing procedures. Transparency is linked to the concept of accessibility (following the principle whereby a structure built on a transparent system produces legible and, therefore, accessible structures) and accessibility is necessary to ensure the scientific validity of the data and the resulting metadata [72]. We can observe that photogrammetric survey projects have already been produced that have taken into account the indications of the Digital Curation Center, such as the work carried out on the Church St. Peter and Paul (Calw, Germany). Although focused exclusively on the conservation and acquisition of archaeological structures in elevation, the project has outlined an acquisition methodology, photogrammetric and via laser scanner, which aims to establish data collection classes to guarantee the conservation of information and its correct curatorship, even if the archaeological asset is ruined or lost [73]. The BeArchaeo project aimed to address the issue of active data management in data archiving by considering the principles of digital data curation. The project aimed to ensure reproducibility, reuse, and proper conservation of data, which would facilitate the generation of new data from existing research. To achieve this, the project implemented institutionalized workflows and utilized IT resources to index information accurately. These precautions are necessary because, in the specific field of archeology, the data acquisition and generation processes are characterized by a tendentially non-linear production for which the abstraction of the archaeological activities is necessary to bring the works together towards a correct form of exhibition curating [74]. In general, it is possible to find approaches in the literature that are inspired by the digital data curation, even if they do not refer to it. However, this lack of "declaration of belonging" represents an obstacle to comparing the established behavioral classes and glossaries and defining to what extent they are attributable to the guidelines established. DDC-compliant projects take into account, in a unitary and continuous way, the management of data-generative processes from the early stages (in this case, photos) to their final use [75] or those projects that involve the needs of cultural heritage with data entries and their use in terms of digital data [76], or in addition, they take into account the fact that errors of an epistemological or compilation nature can only cause cascading error propagation if inserted into technically functional but methodologically inconsistent frameworks [77]. Among these, we can mention the Europeana project [78] or Fernandez's proposal of 2019 [79].

A DDC-Compliant Workflow for Photogrammetry
In this section, we address the instantiation of the general digital data curation framework into the photogrammetry workflow. We address a specific workflow, developed for the BeArchaeo project as an instantiation of the DDC schema: each phase is embedded as a component of the high-level DDC operations, adopting a constrained format that descends from the wide design. We proceed by providing a brief description of the project and then analytically on each phase (cf. Figure 1).

The BeArchaeo Case Study: Interdisciplinarity and Internationality
The EU H2020 "Beyond Archaeology" (BeArchaeo) project develops interdisciplinary and multidisciplinary methodological models for the study of the historical relationships between the archaeological areas of Okayama and Shimane [80]. The BEA-PG work phases are sequenced as follows: acquisition, processing, modelling, export, archiving, postarchiving, and indexing. The survey activities were carried out from February 2019 to August 2019 and the elaboration processes took place from August 2019 to May 2022, which led to the exhibition phases of the project, first the Cave Automatic Virtual Environment (CAVE) of the University of Turin (June 2022) and then the creation of BeA-Vir, an interactive system for the Exhibition at Museum of Izumo in Autumn 2022. Considering an average of 50 photos for artifacts and 100 photos for architectural or environmental surveys, we compared a photographic archive of 5850 photographs, which was organized, recovered, and recognized thanks to the specific acquisition methods that we will now illustrate ( Figure 1).
Heritage 2023, 6, FOR PEER REVIEW 9 Considering an average of 50 photos for artifacts and 100 photos for architectural or environmental surveys, we compared a photographic archive of 5850 photographs, which was organized, recovered, and recognized thanks to the specific acquisition methods that we will now illustrate ( Figure 1).

The Conceptualization of Surveys
The process of photogrammetric survey relies on conceptualization as a crucial element throughout its phases, with particular significance during the acquisition phase, where the object of survey is understood and conceptual abstraction begins. Digital data curation outlines six distinct phases for data management-acquisition, appraisal, selection, preservation, maintenance, and dissemination. While these phases may not perfectly align with the proposed phases of the BeA-PG protocol, they serve as an abstract reference for its development, with each phase characterized by the phases of digital data curation. Therefore, these phases should be considered as a defining feature of the BeA-PG protocol. It should be noted that conceptualization is a vital aspect of the entire protocol, as a theoretical and methodological framework. The BeA-PG protocol is composed of several phases, which we will describe in detail. These steps are crucial for producing 3D models of cultural heritage and employing them in archives, exhibitions, and interactive virtual contexts. The phases encompass acquisition, processing, modeling, export, pre-indexing, archiving, and indexing. Although these phases stem from digital data curation, they are not entirely identical.

The Conceptualization of Surveys
The process of photogrammetric survey relies on conceptualization as a crucial element throughout its phases, with particular significance during the acquisition phase, where the object of survey is understood and conceptual abstraction begins. Digital data curation outlines six distinct phases for data management-acquisition, appraisal, selection, preservation, maintenance, and dissemination. While these phases may not perfectly align with the proposed phases of the BeA-PG protocol, they serve as an abstract reference for its development, with each phase characterized by the phases of digital data curation. Therefore, these phases should be considered as a defining feature of the BeA-PG protocol. It should be noted that conceptualization is a vital aspect of the entire protocol, as a theoretical and methodological framework. The BeA-PG protocol is composed of several phases, which we will describe in detail. These steps are crucial for producing 3D models of cultural heritage and employing them in archives, exhibitions, and interactive virtual contexts. The phases encompass acquisition, processing, modeling, export, pre-indexing, archiving, and indexing. Although these phases stem from digital data curation, they are not entirely identical.

Acquisition
The acquisition phase starts with the identification of the object of the acquisition. In the case of an architectural structure, this is identified in an inspection, during which the notable points are identified, i.e., those points which, due to their geometric uniqueness or chromatic discrepancy, can be easily recognized by the software at the time of processing. During these preliminary operations, the recognition of significant points allows the avoidance of the use of targets in the acquisition phase. At this point, an eidotype will be drawn. At the same time as this first phase, a "mask" will be created. With this expression, we refer to a rapid survey of the structure that does not have to coincide with stringent precision parameters but only serves to identify a series of specific problems, such as reflectance, the influence of the sun, the cutting of shadows, and notable points that have escaped the drafting of the eidotype. At this point, we proceed with the photographic acquisition, which must respect the overlapping parameters between one photograph and the next one, at 60/70% [81]. The BeA-PG method enucleates 4 specific acquisition methodologies. The definition of specific categories of acquisition methods is an attempt to summarize the infinite variables that the photogrammetric survey operator normally faces, enclosing the possible acquisition scenarios in a group established based on the context. Furthermore, this approach can guarantee the order in the data collection from the first photographs; in this way, not only will the use of storage space be optimized but future retro-engineering operations will be made much simpler to the extent that it will be possible to repeat the processing of the survey or extrapolate new data as processing technology changes.

•
Straight-line acquisition: an acquisition process to be used in the acquisition of large buildings and facades, proceeding parallel to the structure and turning the camera perpendicular to the object of the survey. • Spiral acquisition: a set of photographs collected by defining an ideal focal point at the center of the object of the survey and pointing the camera towards the center proceed with an overall acquisition of the photographs, first perpendicularly then with an inclination of 45 • , and finally with an inclination of 75 • . This technique is to be used for single finds and artifacts. However, the spiral acquisition can also be used to collect data on a cavity or a structure with a structure similar to a corridor. In this case, the focal length must be placed at the bottom and one must proceed by placing the base of the spiral perpendicular to the ground and not parallel to it, as in the case of finds and other objects.

•
Aerial acquisition: a variant of the straight line acquisition but considering proceeding parallel to the ground and turning the camera downwards. This technique is optimal when used with a drone for the acquisition of a territory and is also perfect for surveying archaeological trenches. • Vertical capture: a straight line acquisition which is used for elevations that develop in height particularly, and in this case, a drone will most likely have to be used. The acquisition of the photographs can be carried out with cameras and mobile phones and there is no preferential indication for the choice of the instrument [82,83].
More generally, from a purely methodological point of view, one must consider the "rule of opposites". This process is designed to increase the inclination of the color points in a group of photographs. Photogrammetry processing facilitates the process of recognition of geometries through the inclination of a theoretical polygon identified by the way a specific area of color obstructs the photograph or not and by its transformation from one photograph to another, defining the trend of the surface [84].
The use of such precise terminology and the definition of the various operational sub-phases is aimed at allowing operations carried out on monuments to also be intelligible for photogrammetric operators of the humanities, as it has a basis that allows them to respond to contextual difficulties and thus the problems of the interdisciplinarity of matter. In the same way, a set of documentation made up of an eidotype, a mask, and the collected photographs make it possible to reconstruct, and therefore to trace, all the choices made by the operator and possibly even question them. Furthermore, by keeping detailed information on acquisition procedures, it is possible to repeat the survey to the extent that new software allows for a greater optimization of the inferable data, thus opposing obsolescence processes. From the point of view of classification, already in this first phase, the nature of the object of the survey is clearly distinguished based on its geometric characterizations.

Processing
Although, as stated, the conceptualization underlies each phase, here we propose that there is a profound link between the processing phase and the term "Create" in the DDC glossary. Processing is when the photographs are processed to give life to the point cloud and subsequently to the three-dimensional model. The choice of a specific software or photographic acquisition tool would seem decisive, but studies on main parameters, such as views/ease of use, primitives/polygon creation tools, editing polygons, complex surface and volume operations, texture maps and materials, and export [85]; the process of structure for motion [86]; the accuracy of return [87]; processing algorithms [88]; or the efficient scalability of the models produced [89], on the other hand, show us how difficult it is to determine a definitive processing software. In the context of also illustrating the operations to be carried out in the point cloud to non-experts, an attempt was made to describe a method that could be applied to any cloud editing software (thus obviating the dependence on specific software and thus curbing the problems inherent in obsolescence). This phase is also used to skim the data, in fact, in the context of the BeArchaeo project, the BeA-PG method required all the operators involved to check the following:

•
The homogeneity of the point cloud, with a consistent average density. • For all surveys, it was necessary to produce a control video in AVI format showing the survey in its entirety for a maximum duration of 1 min, with a black background and double lighting, one fixed and one mobile.

•
The mesh correction of the part where the deformations of the texture coincidence must be eliminated. • All the surveys must be known and exposed at the time of loading on the database in the form of a technical sheet whose structure is illustrated below.
Once again, the decision to provide the verification indications of the point cloud status was made in order not to exclude any of the researchers involved from the job evaluation processes, providing general indications to be able to recognize the qualitative status of the job. Naturally, numerous variables in the context of processing are purely technical, but for this very reason, they are not the object of this abstract analysis and which, if placed as a central topic, as we have seen in previous case studies, risk moving toward software-based paths.

Modeling
Modeling can be defined, from an operational point of view, as a set of operations that are carried out on the obtained three-dimensional model. Modeling is a process that here is distinguished both from processing and from exporting as the conceptual independence of both the phase and the existence of the 3D model, which is claimed in the context of editors (Blender, Maya, Rhino, 3D Studio Max) and detached from the photos that generated it and from the inert model after export. Once again, if one observes the literature consulted, this phase too is considered to be within the exclusive remit of the technical and the focus is on the definition of which software to select [58]; however, even in this phase, we have limited ourselves to providing indications of the verification and cleaning operations to be carried out regardless of the software used to oppose its obsolescence processes as its potential and tools change considerably over the years. Even when a workflow is proposed, if it is built starting from the centrality of this phase over other phases, it is structured by imagining only one possible sequence of technical choices (type of camera, type of processing software, type of software modeling process) and omitting the subsequent data archiving phases [59]. This choice was made to guarantee the traceability of the phases carried out and the modeling, isolated in this specific context, becomes the only manipulative part of the 3D model, thus isolating any actions that need to be traced in a specific moment that is easily recognizable. Yet, the fact that this phase requires speculative processes becomes more and more evident and is compared with the characteristic operations of design. Since the operations carried out in the 3D model point to a specific selection of the contents which are preserved or eliminated (regarding which we can speak of "Selection" and "Dispose" in the context of the DDC) [90]. In light of the aforementioned and remaining faithful to the principles of interdisciplinarity, the activities to be carried out in the context of modeling, proposed here, are as follows: eliminate the areas in which textures and polygons do not coincide, then proceed with the distinction between the shadows due to the absence of data (such as objects that may have stood in the way at the time of acquisition) and shadows generated from processing errors. Then, the average detail trend of the point cloud needs to be determined. A quick method to achieve this is to evaluate the density of the cloud, ensuring that it is consistent throughout the model. Another important aspect is the identification of the object of the work, that is, the element that represents the concrete object of study, in order to proceed with the elimination of non-inherent or misleading aspects. This must also be performed to try to gain storage space and stay within the parameters established in the previous steps. Naturally, the original processing data must be kept in rigid archives that will be part of the project's warehouse material. In any case, each modeling operation carried out must be reported in the notes of the final sheet. Yet, the fact that this phase requires speculative processes becomes more and more evident and is compared with the characteristic operations of "Design" as defined by DDC. Since the operations carried out in the 3D model point to a specific selection of the contents, which are preserved or eliminated [91]. All the manipulation operations carried out on the 3D model must be shown and must be able to fall into one of the exposed categories which serve to justify the actions carried out and thus trace the work activities.

Exporting
An often overlooked problem is the file export format. This phase represents the "point of no return", beyond which the 3D model is "closed" and becomes a defined object [92], and for this reason, in this proposal, we match the export with the "Ingest" and "Preservation Action" of the digital data curation glossary. This is because the 3D model, as long as it is located in the editor for manipulation, is an "open project" and is therefore susceptible not only to modifications but also to problems related to its placement in a specific hard drive. The editors make references to elements external to the software itself and it is often impossible to export the open project to new locations related to the specific use that the project foresees. Since it is not the object of this paper to analyze in detail the advantages and disadvantages of the different formats, we deem it crucial here to emphasize the methodological importance of distinguishing this phase from the previous manipulation phase. Assuming that the user of this methodology retains all the raw data, which are still traceable and classifiable, in this phase, it is necessary to add a note indicating the reason that prompted the use of a specific export format.

Pre-Indexing
From a purely operational point of view, archiving is the process by which data are collected, categorized, indexed, and entered into a database from which they can be easily retrieved. For 3D models, this phase is often transient, as they are often placed in query software, such as virtual interactives, presentations, museum exhibits, and videos, but this would not be possible without the proper archiving of the models. For these reasons, "Archiving", in the digital data curation terminology, coincides both with the "Store", "Access Use", and "Reuse", and due to its fundamental function of not dispersing data, it has probably been the most analyzed phase in the scientific literature, much often to the detriment of the other phases. Having the DDC as its main objective, along with the archiving and management of digital data, and considering the number of studies that are concerned with the specific area of this paper, it would be misleading to delve too far into this topic. What is important here is to represent this phase in the context of the DDC applied to the photogrammetric acquisition, in other words, how the archiving of the 3D models is the result of a reasoned process which respects the parameters set out and which allows for a second phase of the easy reuse of data. Such "second access" to information must meet the content quality verification metrics, as outlined by the DDC. The need for this preliminary pre-indexing step is linked to the consideration that in this phase a considerable amount of data has already been produced for this reason and it becomes essential to carry out the preventive cataloging of the material [93]. This cataloging is not a substitute for the final one, which will compose the database, but is an obligatory step to be able to trace the data at a later time, carry out checks, or submit the three-dimensional models to reverse engineering processes. For this reason, every time the three-dimensional model is exported, a form is filled in and associated with the model. Additionally, for this reason, cataloging follows the elaboration processes, the classification parameters are progressive, and the definition of the categories is clear. The classification of digital materials takes place by indicating the following: the time at which the acquisition was made, the macro-category of reference, and an alphanumeric list that makes the identification code unique.
This method allows the creation of reference tables for which data can be easily recovered even in the measure in which, as occurred during the case study, there is a long time between the first acquisition and the first insertion of the data in the database. Furthermore, this leads to the creation of a table in which to communicate the data created up to that point and as the elements that are inserted into the database it is connected with. This intermediate step is meant to eventually be buried as a junction part of the manufacturing process when the database becomes fully operational. This ensures complete traceability of the work (Figure 2) carried out not only from the point of view of the 3D model alone but also from a temporal point of view. Pre-indexing provides all the information on the period in which the survey was carried out, to which category the find belongs, and to which subgroup it has been assigned ( Figure 3). This categorization is free from the use of a specific database and is the basis on which digital archives must also be built. As we will also observe, in the practical example reported in the conclusion, this allows not only the traceability and classification of the find, but since it does not depend on any specific database, it allows adaptability and transfer to other contexts as well (thus opposing the problems of obsolescence). Furthermore, this system traces the digital object without requiring the help of specific database glossaries which, although shared, can isolate scholars belonging to disciplinary areas that do not use them.

Archiving and Final Indexing
This phase closes the cycle of creating the 3D model by photogrammetry, according to the BeA-PG protocol, and represents the delicate keystone of the entire process as poor data management practices (how data are created and maintained), which can lead to information loss, and undetected errors in data can cause confusion and lead to misinterpretations [94]. The post-archiving and indexing phase is perhaps the most nuanced of the entire process and coincides with "Transformation", according to the DDC glossary. Concretely, it is the moment in which the 3D models collected are finalized for a specific case study. The models can be extracted from the archive to be subjected to new analyzes, or they can be transmitted via a virtual or real exhibition model (utilizing 3D prints). In any case, whatever the final destination, it is at this moment that a second indexing of the models to adapt their wording to the new reception environment often occurs. This last phase as described here as the most speculative and is derived from the assumption that the whole process of coincidence between DDC and the photogrammetric acquisition pipeline is taken for granted. For this reason, any in-depth analysis and methodological comparison is entrusted to the presentation of the method, where reference is made to the

Archiving and Final Indexing
This phase closes the cycle of creating the 3D model by photogrammetry, according to the BeA-PG protocol, and represents the delicate keystone of the entire process as poor data management practices (how data are created and maintained), which can lead to information loss, and undetected errors in data can cause confusion and lead to misinterpretations [94]. The post-archiving and indexing phase is perhaps the most nuanced of the entire process and coincides with "Transformation", according to the DDC glossary. Concretely, it is the moment in which the 3D models collected are finalized for a specific case study. The models can be extracted from the archive to be subjected to new analyzes, or they can be transmitted via a virtual or real exhibition model (utilizing 3D prints). In any case, whatever the final destination, it is at this moment that a second indexing of the models to adapt their wording to the new reception environment often occurs. This last phase as described here as the most speculative and is derived from the assumption that the whole process of coincidence between DDC and the photogrammetric acquisition pipeline is taken for granted. For this reason, any in-depth analysis and methodological comparison is entrusted to the presentation of the method, where reference is made to the following paragraphs. After the temporary pre-organization based on progressive classification notes, which are also useful later when accessing the repository of intermediate forms, from the photos to the 3D model, one can proceed with the insertion of the data in the final database of the project to the extent that one has foreseen. In this phase, all the finds are reclassified to adapt to the new tool aimed at the work of archaeologists or for dissemination purposes, so a database entry that integrates the digital twin metadata with the data input from archaeometry and archaeologists, reporting the global project identifier.
Observing these works, one realizes that attempts at abstraction and theoretical definition, if deduced from a purely technical and non-methodological reflection, risk forcing the method towards the idea of an ideal workspace where specific technologies, generous funds, and ample timescales are available [95].

Verification and Legitimacy: The Case Study
The advantages of having been able to experiment and use the BeA-PG method in a real project has allowed the verification of its ongoing functioning, demonstrating, above all, the main nature of this process, which can be defined as a palindrome, in the sense that it can be utilized in both directions, either starting from the database card with any final destination up to the original photos or starting from the surveys, following its natural process of creating the template. Taking, for example, the pre-indexed exhibit, Alfa01E, as shown in the interdisciplinary workshop on 17 July 2020 for researchers, starting from the simple initials, it was possible to trace the name, the type, the reference culture, the current positioning of the find, the date on which it was found, with which software it was processed, following which algorithm, and how many points the cloud was made up of, as well as accessing a large number of images that perfectly frame the find. Furthermore, by breaking down the identification code and having access to the project hard drives, since the raw data folders are built according to the same methodology, we will know the exact position of the original photos of the survey, i.e., C: Alfa\01\E, and in it we will find the original photos, including those discarded from the survey. All this additional information then flows into the "Database" tab in which the reference code to the survey tab will be queued in order to be able to carry out the reverse path again, from which researchers will be able to access all the existing details on the find, including carrying out a cross-check with the database files relating to the stratigraphic units to which the find is associated and verify which other finds have the same origin. The database in question was created according to the principles of the Semantic Web-this means that a stringent categorization based on known archives has been created, taken from the documentation registers of the Italian Central Institute for Cataloging and Documentation, Getty AAT9 and CIDOC-CRM, and the main semantic properties taken were directly from Web Omeka-S11 [96].
The definitive verification, the so-called litmus test regarding the functioning of this dual-direction system, took place in February 2022 after a long operational pause imposed on the project by the COVID-19 pandemic, when in the archiving phase of the 3D finds, it was possible to recover the original photos, following the classifications indicated by the survey and to be able to repeat the surveys, thus overcoming some problems that had occurred in the first iterations, but above all, it was possible to easily search thousands of photos, recovering exact image required, and, in this way, repeat the surveys of an excavation activity carried out 4 years earlier, obtaining new information on the stratigraphic units and the distribution of the finds in the excavation.

Detailed Case Study
The method described so far, applied in the context of the BeArchaeo project using the times and methods described, has led to the survey of 80 elements, including archaeological finds, architectural structures, and archeozoological finds. To these must be added the survey of the burial chamber of Tobiotsuka Kofun and the survey of the excavation trenches for a total number of further 20 three-dimensional models. Considering an average of 50 photos for artifacts and 100 photos for architectural or environmental surveys, we were confronted with a photographic archive of 5850 photographs, which could be organized, recovered, and recognized thanks to the specific acquisition methods that have been illustrated in the previous paragraphs. The photographs, therefore, led to the elaborations, and even in this phase, for each element, the reference sheet was drawn up with information on the reconstructive algorithm used, the software used, the number of points that make up the surface of the model, and the images of comparison between the 3D model obtained and the original object. In some cases, it was also possible to insert in the cards the threedimensional model of the find at the time of its discovery, from the excavation directly, allowing it to be associated with the reference stratigraphic unit (this was valid for 29 finds found in the excavations carried out in the Sepulchral Chamber of Tobiotsuka Kofun). This allowed an initial classification by observing the sites of origin of the finds and identifying the most represented morphological categories (Figure 4). that have been illustrated in the previous paragraphs. The photographs, therefore, led to the elaborations, and even in this phase, for each element, the reference sheet was drawn up with information on the reconstructive algorithm used, the software used, the number of points that make up the surface of the model, and the images of comparison between the 3D model obtained and the original object. In some cases, it was also possible to insert in the cards the three-dimensional model of the find at the time of its discovery, from the excavation directly, allowing it to be associated with the reference stratigraphic unit (this was valid for 29 finds found in the excavations carried out in the Sepulchral Chamber of Tobiotsuka Kofun). This allowed an initial classification by observing the sites of origin of the finds and identifying the most represented morphological categories (Figure 4).

Overcoming the Paradigm of the Absolute Standard
In conclusion, the proposed methodological protocol does not claim to be revolutionary or resolutive in its entirety but tries to shift the object of the problem from a purely technical to a purely humanities-based methodology, attempting to bring the two visions together. It is the opinion of the authors that many of the problems inherent in the realization of concretely interdisciplinary projects are the difficulties arising from the various parties involved in abandoning their approaches; nevertheless, the increasingly intensive use of technologies in the field of cultural heritage pushes more and more archaeologists to deal with tools such as databases, laser scanners, and photogrammetry and more and more computer scientists to have to apply their techniques and methodologies to the infinite and unique complexities characteristics of archaeological finds and ancient structures. The method has a specific and characterizing focus: to not disperse data. The most dangerous feature of an interdisciplinary system that fails to involve all the various researchers from different subjects is that the data quickly becomes unobtainable, and an archaeological excavation for which concrete and measurable data does not exist may as well have never existed. Furthermore, this method also allows an external observer to verify

Overcoming the Paradigm of the Absolute Standard
In conclusion, the proposed methodological protocol does not claim to be revolutionary or resolutive in its entirety but tries to shift the object of the problem from a purely technical to a purely humanities-based methodology, attempting to bring the two visions together. It is the opinion of the authors that many of the problems inherent in the realization of concretely interdisciplinary projects are the difficulties arising from the various parties involved in abandoning their approaches; nevertheless, the increasingly intensive use of technologies in the field of cultural heritage pushes more and more archaeologists to deal with tools such as databases, laser scanners, and photogrammetry and more and more computer scientists to have to apply their techniques and methodologies to the infinite and unique complexities characteristics of archaeological finds and ancient structures. The method has a specific and characterizing focus: to not disperse data. The most dangerous feature of an interdisciplinary system that fails to involve all the various researchers from different subjects is that the data quickly becomes unobtainable, and an archaeological excavation for which concrete and measurable data does not exist may as well have never existed. Furthermore, this method also allows an external observer to verify the data very easily, how they were produced, and if necessary, repeat the operations that led to the specific product line of three-dimensional materials, making the entire structure of the project available for de facto a peer review that is not limited exclusively to final data of the searches but also to all the intermediate steps. With the publication of the BeArchaeo project database, it will be possible to verify the method and how functional it was in the production of data, and since it is open, anyone can repeat the various operations, even to the extent that if they want to request the raw data, they will always be easily recoverable, thanks to the organization of the folders, which, as already noted, follows the order of the entire structure. In addition, the suggestion of acquisition methodologies can prove to be a very efficient tool to bring even archaeologists closer to a more technical photogrammetric survey methodology and, above all, also adaptable to new software that may become available. Creating a perfect model is impossible due to the very nature of technological development; therefore, our aspiration was not to create a perfect model.
All the activities carried out were made available in real-time by the research team by sharing an Excel file ( Figure 5), where each step was updated and gradually associated with the official classifications that the archaeologists provided for the finds. photogrammetric survey methodology and, above all, also adaptable to new software that may become available. Creating a perfect model is impossible due to the very nature of technological development; therefore, our aspiration was not to create a perfect model.
All the activities carried out were made available in real-time by the research team by sharing an Excel file ( Figure 5), where each step was updated and gradually associated with the official classifications that the archaeologists provided for the finds. At this point, the material was sorted by the various intended uses, including the Interactive Model for CAVE, presentations and conferences, and, above all, the database.
To show the operations carried out in detail, let us take an artifact as an example and observe all the steps that led it from its acquisition to its final results. This process is a true palindrome, in the sense that it can be followed in both directions, either starting from the database card or any final results and moving backwards, up to tracing the original photos, or starting from the initial finds and following the natural process of creating the template. So let us perform both processes. During the interdisciplinary workshop on 17 July 2020, a wild boar jaw was shown in an illustrative video as an example of the three-dimensional surveys carried out in high definition ( Figure 6). The video, according to the archiving procedure provided by the methodology, depicted the wording Alfa01E at the bottom left. This is a very interesting example, as we are not starting from the official database of the project, but from a video that has as its object a find that, not being the result of an excavation activity, will be documented by archaeologists. Anyone who was part of At this point, the material was sorted by the various intended uses, including the Interactive Model for CAVE, presentations and conferences, and, above all, the database.
To show the operations carried out in detail, let us take an artifact as an example and observe all the steps that led it from its acquisition to its final results. This process is a true palindrome, in the sense that it can be followed in both directions, either starting from the database card or any final results and moving backwards, up to tracing the original photos, or starting from the initial finds and following the natural process of creating the template. So let us perform both processes. During the interdisciplinary workshop on 17 July 2020, a wild boar jaw was shown in an illustrative video as an example of the three-dimensional surveys carried out in high definition ( Figure 6). The video, according to the archiving procedure provided by the methodology, depicted the wording Alfa01E at the bottom left. This is a very interesting example, as we are not starting from the official database of the project, but from a video that has as its object a find that, not being the result of an excavation activity, will be documented by archaeologists. Anyone who was part of the project, even with these data only, could access the file on the storage drive. From this, they would learn the name, the type, the reference culture, the current location of the find, the date on which it was detected, with which software it was processed, following which algorithm, and how many points the cloud is made up of, as well as accessing a large number of images that perfectly frame the find. However, let us imagine that anyone who is following this reverse path requires the original data of the find, perhaps in order to carry out new processing. First of all, the same code that allows access to the card also allows the identification the object model of the find within the 3D archive always present in the shared project drive. Furthermore, by breaking down the identification code and having access to the project hard drives, since the raw data folders are built according to the same methodology, we will know the exact position of the original photos of the survey; therefore, C: Alfa\01\E, and through this we would find the original photos, including those discarded from the presentation. carry out new processing. First of all, the same code that allows access to the card also allows the identification the object model of the find within the 3D archive always present in the shared project drive. Furthermore, by breaking down the identification code and having access to the project hard drives, since the raw data folders are built according to the same methodology, we will know the exact position of the original photos of the survey; therefore, C: Alfa\01\E, and through this we would find the original photos, including those discarded from the presentation. Let us now observe how the same structure is conventionally used to move from the survey to the "Database" tab. Let us take as an example a metal fragment found during the funerary excavation of the Tobiotsuka Kofun burial chamber on 3 September 2019 (Figure 7), which, on 7 September 2019, was brought to the excavation laboratory. At this point, the find was assigned the acronym AF137. On 12 September 2019, the find was detected. The photos were saved in the C: \\Beta\01\S\22 folders, from which the threedimensional model saved as Beta01S22, which was obtained, and the reference card was associated with it. Subsequently, the data from the excavation directly were processed, from which the data of the find were obtained at the time of its discovery in association with its reference "Stratigraphic Unit". All this additional information then flows into the "Database" tab, in which the reference code to the survey tab will be saved in order to be able to carry out the reverse path again, and from which researchers will be able to access all the existing details on the find, including carrying out a cross-check with the database files relating to the stratigraphic units to which the find is associated and verify which other finds have the same origin. Let us now observe how the same structure is conventionally used to move from the survey to the "Database" tab. Let us take as an example a metal fragment found during the funerary excavation of the Tobiotsuka Kofun burial chamber on 3 September 2019 (Figure 7), which, on 7 September 2019, was brought to the excavation laboratory. At this point, the find was assigned the acronym AF137. On 12 September 2019, the find was detected. The photos were saved in the C: \\Beta\01\S\22 folders, from which the three-dimensional model saved as Beta01S22, which was obtained, and the reference card was associated with it. Subsequently, the data from the excavation directly were processed, from which the data of the find were obtained at the time of its discovery in association with its reference "Stratigraphic Unit". All this additional information then flows into the "Database" tab, in which the reference code to the survey tab will be saved in order to be able to carry out the reverse path again, and from which researchers will be able to access all the existing details on the find, including carrying out a cross-check with the database files relating to the stratigraphic units to which the find is associated and verify which other finds have the same origin.

Towards the Evolution of Bea-PG: Application of the Method on Another Project
The main strength of the BEA-PG method is the fact that it is completely independent of specific software or technical tools, aiming instead at being a behavioral handbook while remaining faithful to the dictates of digital data curation. Another key advantage is the fact that this method has been actively tested within the BeArchaeo project. What was perhaps the most successful was the possibility of verifying the interdisciplinary breadth of the methodology, which was intelligible for all its users. Even more important was the verification of the palindrome nature of the project. In fact, because of an interruption due to the COVID-19 pandemic, the project was paused for almost two years, at the end of which, there was considerable difficulty in recovering the data; this obstacle, however, did not concern the 3D models of the various finds, which were not only easily recoverable but, in some cases, completely reworked. In fact, at the time of verification, before being included in the exhibition held at the Izumo Museum, during which the models were presented in a virtual interactive environment, by exploiting the production methodology, it was possible to recover the original photos, rework and optimize the exhibits, and reintroduce them into the work pipeline without any problems. On the other hand, the limits of this methodology are evident: desiring to be a bridge between the needs of the field of cultural heritage and the technical requirements of the world of computer sciences, the BEA-PG method lacks the technical specificities of either area. With its high level of abstraction, on the one hand, it aims to be adaptable to multiple different scenarios, even fields external to cultural heritage; however, on the other hand, this risks making the project superficial. For this reason, in order to achieve a concrete evaluation of the method, it must be also applied to other contexts, archaeological or otherwise, and this will be one of the objectives of our work in the coming years.
Heritage 2023, 6, FOR PEER REVIEW 19 Figure 7. The example of exhibit BETA01S22 in the context of retrieving information from the database card.

Towards the Evolution of Bea-PG: Application of the Method on Another Project
The main strength of the BEA-PG method is the fact that it is completely independent of specific software or technical tools, aiming instead at being a behavioral handbook while remaining faithful to the dictates of digital data curation. Another key advantage is the fact that this method has been actively tested within the BeArchaeo project. What was perhaps the most successful was the possibility of verifying the interdisciplinary breadth of the methodology, which was intelligible for all its users. Even more important was the verification of the palindrome nature of the project. In fact, because of an interruption due to the COVID-19 pandemic, the project was paused for almost two years, at the end of which, there was considerable difficulty in recovering the data; this obstacle, however, did not concern the 3D models of the various finds, which were not only easily recoverable but, in some cases, completely reworked. In fact, at the time of verification, before being included in the exhibition held at the Izumo Museum, during which the models were presented in a virtual interactive environment, by exploiting the production methodology, it was possible to recover the original photos, rework and optimize the exhibits, and reintroduce them into the work pipeline without any problems. On the other hand, the limits

Data Availability Statement:
The data relating to the BeArchaeo project database can be consulted at the link https://bearchaeo.unito.it/omeka-s/s/bea_resources_site_ja/page/welcome_bearchaeoresources-site.