Long-Term Preservation of Digital Information: Challenges, Strategies, and the Role of FAIR Principles

Comprehensive Report on Digital Information Preservation: Challenges, Curation, and Enduring Accessibility

Many thanks to our sponsor Esdebe who helped us prepare this research report.

Abstract

The digital age has ushered in an era of unprecedented information generation, transforming how knowledge is created, disseminated, and consumed. This relentless proliferation of digital data, ranging from foundational scientific datasets and intricate historical records to ephemeral multimedia content and vast governmental archives, underscores an urgent global imperative: the long-term preservation of digital assets. Ensuring the sustained accessibility, authenticity, and usability of digital information over extended temporal horizons is not merely a technical challenge but a foundational requirement for upholding the integrity of cultural memory, scientific progress, and societal accountability. This extensive report critically examines the multifaceted and evolving challenges inherent in digital preservation, comprehensively explores advanced digital curation techniques, and meticulously details the indispensable roles of both institutional and public data archives in this endeavour. Furthermore, it delves into sophisticated strategies for the migration of diverse data types across decades, conducts an in-depth analysis of the complex economics underpinning long-term digital storage, and thoroughly discusses the practical and strategic implications of the FAIR (Findable, Accessible, Interoperable, Reusable) principles as a cornerstone for sustaining the enduring impact and utility of digital data. Through this detailed exploration, the report aims to provide a robust framework for understanding and addressing the complexities of safeguarding our digital heritage.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

1. Introduction: The Imperative of Digital Longevity

The pervasive integration of digital technologies into virtually every facet of modern society has resulted in an exponential increase in the volume, velocity, and variety of digital information. This encompasses an immense spectrum of content, from terabytes of scientific simulation data and petabytes of earth observation imagery to intricate historical archives, complex multimedia productions, personal communications, and the rapidly expanding realm of social media interactions. Unlike traditional analogue formats, which degrade visibly over time, digital information is susceptible to a unique set of vulnerabilities: rapid technological obsolescence, insidious data corruption, and the inherent fragility of digital storage media. These risks collectively pose a significant threat to the longevity and accessibility of invaluable digital content, potentially leading to a ‘digital dark age’ where vast swathes of contemporary knowledge become irretrievable.

The urgency for developing robust and comprehensive strategies for digital preservation cannot be overstated. It is not simply about safeguarding files; it is about preserving the ‘memory of humanity’, ensuring the continuity of research, maintaining governmental accountability, protecting intellectual property, and sustaining cultural identity across generations. Without deliberate and proactive preservation efforts, the very foundation of knowledge accumulated in digital form is at risk, jeopardising future research, historical understanding, and informed decision-making. Therefore, digital preservation emerges as a critical discipline, demanding interdisciplinary collaboration, continuous adaptation, and substantial strategic investment to ensure that valuable information remains perpetually accessible, authentic, and usable for the benefit of future generations.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

2. Challenges in Digital Preservation: Navigating the Ephemeral Landscape

The inherent characteristics of digital information, coupled with the relentless pace of technological evolution, present a formidable array of challenges to its long-term preservation. These challenges are often interconnected and demand a multi-pronged approach for effective mitigation.

2.1 Digital Obsolescence

Digital obsolescence represents arguably the most profound threat to the longevity of digital content. It refers to the phenomenon where digital information becomes inaccessible or unusable due to the discontinuation or radical transformation of the requisite hardware, software, or file formats necessary for its interpretation. As technology relentlessly advances, older systems, applications, and proprietary formats are phased out, often without backward compatibility, rendering previously accessible data unreadable or unrenderable.

  • Hardware Obsolescence: The physical devices used to store and access digital data have remarkably short lifespans. Floppy disk drives, Zip drives, CD-ROM drives, various tape formats, and even specific generations of hard drives become obsolete. Without the original hardware, data stored on these media may be physically unreachable. Even if the data can be extracted, the software environment required to interpret it may no longer exist.
  • Software Obsolescence: This is perhaps an even more pervasive issue. Operating systems evolve, application software undergoes constant updates, and proprietary file formats are frequently revised or abandoned. A document created in an early version of a word processor, a database from a defunct system, or a graphic produced by an obscure design program may become unviewable as the supporting software environment vanishes. Furthermore, the loss of documentation for proprietary formats can make reverse-engineering nearly impossible.
  • File Format Obsolescence: Many file formats, especially those developed by commercial entities, are proprietary and not openly documented. When the software that created or reads these formats is no longer supported, the data effectively becomes ‘locked in’. Even open formats can suffer from obsolescence if they fall out of favour and tools capable of parsing them are no longer maintained.

The ‘digital dark age’ is a term frequently invoked to describe a potential future scenario where valuable contemporary digital records become inaccessible due to this pervasive obsolescence, creating a significant void in humanity’s historical record (Rosenthal, 2018). Proactive measures are therefore essential to combat this perpetual cycle.

2.2 Data Corruption and Degradation

Unlike the visible deterioration of physical artefacts, digital data can suffer from insidious corruption and degradation that may occur silently and without immediate detection. This ‘bit rot’ – the spontaneous, unnoticed alteration of data bits – can render files unreadable or alter their content subtly but significantly.

  • Media Degradation: All physical storage media have finite lifespans. Magnetic tapes can degrade, optical discs (CDs, DVDs, Blu-rays) can suffer from ‘disc rot’ due to material decomposition, and flash memory (SSDs, USB drives) has limited write/erase cycles. Environmental factors such as temperature fluctuations, humidity, dust, and magnetic fields can accelerate this degradation, leading to data loss or corruption.
  • Hardware Failures: Hard drives can fail mechanically, solid-state drives can experience controller failures, and network storage devices can encounter system errors, all of which can lead to data loss or inaccessibility.
  • Software Bugs and System Errors: Flaws in operating systems, storage management software, or backup routines can inadvertently introduce errors, corrupt data, or lead to misplacement of files.
  • Human Error and Malicious Attacks: Accidental deletion, incorrect file modifications, or misconfigurations by users or administrators are common causes of data loss. Furthermore, cyber threats such as ransomware, viruses, and deliberate data tampering pose significant risks to data integrity and availability.
  • Natural Disasters: Fires, floods, earthquakes, and other catastrophic events can destroy physical infrastructure and the digital data stored within, underscoring the need for geographically distributed redundancy.

Regular monitoring, integrity checks, and redundant storage strategies are paramount to detecting and mitigating these forms of corruption and degradation before they become irreversible.

2.3 Volume, Velocity, and Complexity of Data

The sheer scale, rapid generation, and intricate nature of contemporary digital data present logistical and technical challenges of unprecedented magnitude for preservation efforts.

  • Volume: The exponential growth of data, often referred to as ‘Big Data’, means archives are confronted with petabytes and even exabytes of information. Managing, storing, indexing, and processing such vast quantities requires substantial and continuously scalable infrastructure, immense computational power, and significant energy consumption.
  • Velocity: Data is increasingly generated in real-time or near real-time, such as streaming sensor data, financial transactions, or social media feeds. Capturing and preserving such dynamic, ephemeral data poses unique challenges, as traditional batch-processing methods may be insufficient.
  • Variety and Complexity: Digital data comes in an incredibly diverse array of formats, structures, and interdependencies. This includes structured databases, unstructured text documents, complex scientific datasets (e.g., geospatial, genomic), high-resolution images, multi-channel audio, high-definition video, interactive software applications, 3D models, virtual reality environments, and linked data graphs. Each type often requires specific preservation approaches, tools, and metadata schemes. The relationships between different data components, such as a dataset linked to its code, documentation, and research papers, add layers of complexity to preservation.
  • Dark Data: A significant portion of data generated by organisations is often termed ‘dark data’ – information collected, processed, and stored but never fully utilised or systematically managed for long-term retention. Identifying, appraising, and deciding what to preserve from this mass of data is a major challenge.

2.4 Technological Dependencies and Interoperability Issues

Digital information is rarely self-contained; it relies heavily on specific technological ecosystems. This creates profound dependencies that complicate long-term preservation.

  • Proprietary Lock-in: Many software applications and hardware systems rely on proprietary specifications, making it difficult to migrate or access data outside of their original environment. This ‘vendor lock-in’ can severely restrict preservation options and increase future costs.
  • Lack of Universal Standards: While efforts are underway, a universal set of standards for all types of digital data preservation remains elusive. Different disciplines and industries often adopt their own standards, leading to fragmentation and difficulties in cross-domain data interoperability.
  • Complex Software Stacks: Modern digital objects, especially interactive ones (e.g., video games, dynamic websites), are often composed of multiple layers of software (operating system, middleware, application, plugins, databases). Preserving such complex ‘digital objects’ requires capturing and maintaining the entire execution environment, which is highly challenging (Hedstrom, 1997).

2.5 Legal, Ethical, and Policy Complexities

Digital preservation is not solely a technical problem; it is deeply intertwined with a labyrinth of legal, ethical, and policy considerations.

  • Data Privacy and Confidentiality: Regulations such as the General Data Protection Regulation (GDPR), Health Insurance Portability and Accountability Act (HIPAA), and various national privacy laws impose strict requirements on how personal and sensitive data is handled, stored, and accessed. Preserving such data for the long term must be balanced with individuals’ rights to privacy and the right to be forgotten.
  • Intellectual Property Rights (IPR) and Copyright: Digital content is subject to copyright, patent, and trade secret laws. Preserving and providing access to copyrighted material requires careful navigation of licensing agreements, fair use doctrines, and potentially obtaining specific permissions for reproduction or distribution for archival purposes.
  • Access Restrictions: Some data, particularly government records, classified information, or commercially sensitive data, may have strict access restrictions. Long-term preservation must accommodate these restrictions while ensuring the data’s integrity and eventual potential for declassification or broader access.
  • Legal Admissibility and Authenticity: For legal or compliance purposes, preserved digital records must maintain their authenticity and integrity to be admissible as evidence. This requires meticulous provenance tracking and robust chain-of-custody documentation.
  • Ethical Considerations: Beyond legal frameworks, ethical considerations arise concerning the representation of communities, potentially harmful content, or the responsible use of preserved data, particularly in culturally sensitive contexts (Yakel, 2007).

Developing clear policies and legal frameworks that support long-term preservation while respecting rights and ethics is an ongoing challenge for governments and institutions.

2.6 Lack of Awareness and Expertise

Despite the critical importance of digital preservation, a significant challenge lies in the general lack of awareness, understanding, and specialised expertise across various sectors.

  • Insufficient Awareness: Many organisations and individuals do not fully grasp the fragility of digital information or the complexities involved in its long-term preservation. This can lead to underinvestment and a reactive, rather than proactive, approach.
  • Skill Gap: Digital preservation demands a unique blend of skills, including archival science, information technology, data management, legal expertise, and domain-specific knowledge. There is a global shortage of professionals adequately trained in these interdisciplinary areas.
  • Funding Constraints: Dedicated funding for digital preservation infrastructure, personnel, and research often lags behind the escalating volume of digital data. This can lead to a prioritisation of immediate operational needs over long-term stewardship.

Addressing these challenges requires sustained advocacy, education, professional development programmes, and strategic resource allocation to build capacity and foster a culture of preservation.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

3. Advanced Digital Curation Techniques: Safeguarding Digital Assets

Digital curation encompasses the active and ongoing management of digital data throughout its lifecycle, from creation and appraisal to preservation and access. It involves a range of sophisticated techniques designed to ensure the long-term viability, authenticity, and usability of digital information.

3.1 Migration: The Foremost Strategy

Migration is the most widely adopted and fundamental strategy for combating digital obsolescence. It involves the controlled and systematic transfer of digital content from one file format, operating system, or storage medium to another, with the explicit goal of maintaining the content’s intellectual integrity, authenticity, and accessibility over time.

  • Normalization (Format Migration): This involves converting digital objects from their original, potentially proprietary or unstable, formats into more stable, open, and preservation-friendly formats. For instance, a Microsoft Word document might be migrated to PDF/A (PDF for Archiving), or a proprietary image format might be converted to TIFF (Tagged Image File Format). The selection of target formats is critical, favouring those with open specifications, widespread adoption, and robust community support, such as XML, JPEG 2000, and uncompressed audio formats. Normalisation helps standardise heterogeneous collections, making them easier to manage and ensuring future access (Preservation Metadata, n.d.).
  • Reformatting: This is a specific type of migration where data is simply copied to new physical storage media (e.g., from old magnetic tapes to modern hard drives or cloud storage) without changing the file format itself. This addresses media degradation rather than format obsolescence.
  • Upgrading/Refreshing: This refers to moving digital data from older hardware/software environments to newer ones. This might involve transferring files from an outdated server to a new server, or updating a database to a newer version of its underlying software. It is a continuous process as technology evolves.

Challenges and Considerations for Migration:
* Lossy vs. Lossless: While the ideal is lossless migration (no information loss), some formats or complex objects may require ‘lossy’ transformations, where some data or functionality is intentionally discarded to achieve compatibility. Careful documentation of any such losses is essential.
* Authenticity and Integrity: Each migration step introduces a potential for unintended alteration or loss. Rigorous validation, integrity checks (e.g., checksums), and comprehensive provenance metadata are crucial to ensure that the migrated data remains an authentic representation of the original.
* Cost and Resource Intensity: Migration is a continuous and resource-intensive process, requiring ongoing monitoring, software development, and skilled personnel.
* Scalability: Migrating petabytes of data is a massive undertaking, demanding automated tools and efficient workflows.

3.2 Emulation: Recreating the Original Environment

Emulation is a digital preservation strategy that aims to recreate the original computing environment (hardware and software) in which a digital object was created and functioned. By doing so, it allows future users to interact with the digital object in its native context, preserving its original look, feel, and functionality without altering the original data format.

  • Mechanism: Emulators are software programs that mimic the behaviour of a specific piece of hardware (e.g., an old CPU, a graphics card) on a different, more modern hardware platform. When combined with virtual machines (which create isolated, virtualised computing environments), they can replicate an entire legacy system, including its operating system, application software, and the digital object itself. Examples include DOSBox for running old DOS games and applications, or projects like CAMiLEON that experimented with universal emulation (Giaretta, 2011).
  • Advantages: Emulation excels at preserving the ‘functionality’ and ‘experiential’ aspects of complex digital objects like interactive multimedia, legacy software, video games, or dynamic websites. It maintains the authenticity of the original bit-stream and allows researchers to observe how software behaved historically.
  • Disadvantages: Emulation can be resource-intensive, requiring significant computational power. Managing and maintaining a vast library of emulators and operating system images for diverse legacy systems is complex. There are also legal challenges related to emulating proprietary operating systems or software for which an archive may not hold licenses.

While migration is often preferred for simpler, document-based content, emulation remains a vital tool for preserving complex, interactive, or context-dependent digital assets where the original user experience is paramount.

3.3 Digital Forensics: Ensuring Integrity and Authenticity

Digital forensics, typically associated with legal investigations, plays a critical role in digital preservation by providing methodologies and tools to ensure the integrity, authenticity, and recoverability of digital records. In a preservation context, forensic techniques are applied proactively and reactively.

  • Proactive Role: Digital forensic principles inform robust preservation practices such as careful data ingest, the use of ‘write blockers’ to prevent accidental alteration of source media, and the generation of cryptographic hash values (e.g., SHA-256) at every stage of the preservation workflow. These hashes act as digital fingerprints, allowing for verification that data has not been altered or corrupted over time. Maintaining a meticulous ‘chain of custody’ for digital objects, documenting every step from creation to archival, is a forensic best practice that ensures authenticity (Mason, 2011).
  • Reactive Role: When data corruption or suspected tampering occurs, digital forensic tools and expertise are employed to investigate the extent of the damage, recover lost or corrupted bits, and determine the cause. This involves deep analysis of file systems, metadata, and even raw disk images to reconstruct the original state of the data or identify malicious alterations. This capability is crucial for maintaining trust in the preserved information and for legal admissibility.

Digital forensics provides the technical bedrock for proving the trustworthiness and reliability of preserved digital information, which is fundamental to its long-term value.

3.4 Bit-stream Preservation (Fixity)

At the most fundamental level of digital preservation is bit-stream preservation, often referred to as ‘fixity’. This technique focuses on ensuring that the actual sequence of bits that constitute a digital file remains unchanged over time. It’s the ‘lowest common denominator’ of preservation, ensuring that the original digital object, even if its format becomes obsolete, is precisely as it was when ingested.

  • Mechanism: Fixity is achieved through the regular calculation and verification of cryptographic hash values (checksums) for every digital object. A hash function generates a unique, fixed-size string of characters from any input data. Even a single bit change in the input will produce a drastically different hash output. Common algorithms include MD5, SHA-1, SHA-256, and SHA-512. These hashes are computed upon ingest and regularly re-computed and compared against the stored ‘golden’ hash. Any discrepancy indicates corruption.
  • Redundancy: Bit-stream preservation is inextricably linked to robust storage solutions. This includes implementing RAID (Redundant Array of Independent Disks) systems, error-correcting codes, and geographically dispersed, redundant copies of data across multiple storage sites. The ‘Lots of Copies Keep Stuff Safe’ (LOCKSS) principle is a key tenet, advocating for multiple, independently managed copies to reduce the risk of catastrophic data loss (Maniatis et al., 2002).

Bit-stream preservation is a non-negotiable prerequisite for all other digital preservation strategies. Without ensuring the integrity of the underlying bits, any further curation efforts are futile.

3.5 OAIS Reference Model: A Conceptual Framework

The Open Archival Information System (OAIS) Reference Model, formally ISO 14721:2012, is a conceptual framework that defines the roles, functions, and responsibilities of an archive committed to preserving information for access by a designated community (Consultative Committee for Space Data Systems, 2012). It is not a technical implementation but a widely adopted standard that provides a common language and understanding for digital preservation.

  • Entities: OAIS identifies key entities: the Producer (who provides information), the Management (who oversees the archive), the Archive (the system itself), and the Consumer (who retrieves information).
  • Functional Entities: The model describes six high-level functional entities within an archive:
    • Ingest: Receiving information from producers, validating it, creating Archival Information Packages (AIPs), and preparing it for storage.
    • Archival Storage: Managing the long-term storage of AIPs, including redundancy, refreshing, and integrity checks.
    • Data Management: Managing the archive’s descriptive and administrative metadata, supporting queries and retrievals.
    • Access: Making preserved information available to consumers, including providing appropriate access interfaces and formatting data as Dissemination Information Packages (DIPs).
    • Preservation Planning: Monitoring the external environment (technology, standards), developing preservation strategies (e.g., migration plans), and recommending necessary actions.
    • Administration: Overall management of the archive, including policy development, resource allocation, and auditing.

OAIS provides a holistic view of the preservation process, guiding the design and implementation of trustworthy digital repositories (TRDs).

3.6 In-situ Preservation (Limited Application)

While often impractical for large-scale collections, in-situ preservation involves maintaining the original hardware and software environment necessary to access highly complex or interactive digital objects. This approach is typically reserved for unique, historically significant, or technologically intricate items where the exact original context is deemed essential.

  • Application: This might involve preserving an old computer system with its original operating system and application software to run a specific legacy program, an early video game, or an interactive art installation. It’s about preserving the entire technological stack.
  • Challenges: The major challenges include the physical degradation of original hardware components, the difficulty in finding replacement parts, the energy costs of maintaining old systems, and the limited scalability of this approach. It is often combined with documentation and emulation efforts to ensure that even if the physical system fails, the essence of the digital object can be recreated.

3.7 Data Normalization and Standardization (Beyond File Formats)

Beyond simply standardising file formats, comprehensive data normalization and standardization extend to data schemas, metadata vocabularies, and content models. This ensures consistency and interoperability, which are critical for search, access, and future data reuse.

  • Standard Schemas: Using established schemas (e.g., XML Schema, JSON Schema) for structured data ensures that the data’s internal organisation is well-defined and can be consistently parsed and understood over time.
  • Controlled Vocabularies and Ontologies: Implementing thesauri, authority files, and ontologies (e.g., Getty Thesaurus of Geographic Names, SKOS) for descriptive metadata ensures that terms used to describe content are consistent and unambiguous, facilitating discovery and interoperability across diverse datasets and systems.
  • Content Models: For complex digital objects composed of multiple files (e.g., a digitized book with images, OCR text, and structural XML), defining a clear content model specifies the relationships between these components, ensuring the integrity of the object as a whole.

These advanced techniques collectively form a robust strategy for tackling the multifaceted challenges of digital preservation, transitioning from mere storage to active, intelligent curation.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

4. Roles and Responsibilities of Institutional and Public Data Archives

Digital preservation is a collaborative endeavour, with various types of archives playing distinct yet complementary roles in safeguarding the world’s digital heritage.

4.1 Institutional Archives: Guardians of Organizational Memory

Institutional archives are dedicated to preserving records that document the administrative, operational, intellectual, and historical activities of their parent organisations. These can include universities, corporations, government agencies, non-profit organisations, and healthcare providers. Their primary responsibility is to manage the entire lifecycle of digital records, from creation to long-term preservation, ensuring compliance with internal policies, legal requirements, and ethical standards.

  • Scope and Function: Institutional archives typically manage a diverse range of digital content, including administrative documents (emails, reports, meeting minutes), financial records, human resources data, research data generated by faculty or R&D departments, intellectual property, and institutional publications. Key functions include:
    • Records Management: Working with active records to ensure proper creation, classification, retention scheduling, and eventual transfer to the archive.
    • Appraisal and Selection: Identifying and appraising digital records of enduring value for long-term preservation, often working within established retention policies.
    • Ingest and Preservation: Implementing secure ingest workflows, applying preservation strategies (migration, emulation, fixity), and ensuring the integrity and authenticity of digital records.
    • Access and Compliance: Providing controlled access to preserved records for authorised users while ensuring adherence to privacy regulations (e.g., GDPR), intellectual property laws, and specific organisational policies.
    • Audit and Risk Management: Regularly auditing digital holdings, assessing risks, and developing disaster recovery and business continuity plans.

Examples include university archives preserving research output and administrative history, corporate archives managing critical business records, and governmental agencies like the National Archives and Records Administration (NARA) in the United States or The National Archives (TNA) in the United Kingdom, which preserve federal government records for public access and accountability.

4.2 Public Data Archives: Stewards of Collective Memory

Public data archives serve the broader community by preserving records of societal, cultural, historical, and scientific significance. These institutions typically operate at national or international levels and are tasked with collecting, curating, and providing extensive access to a wide array of digital materials for research, education, and public engagement.

  • Scope and Function: Public archives encompass national libraries, national archives, national science data centres, and specialised domain repositories. Their collections may include national web archives, digitised cultural heritage materials (manuscripts, photographs, audio-visual recordings), scientific research data from publicly funded projects, electoral data, social media archives, and born-digital publications. Their functions often include:
    • Collecting and Acquisition: Proactively identifying, acquiring, and ingesting digital content deemed to be of enduring public value, often through legal deposit mandates (for national libraries) or agreements with creators.
    • Long-Term Stewardship: Implementing robust preservation strategies at scale, including massive storage infrastructures, continuous migration programmes, and advanced metadata management.
    • Public Access and Outreach: Providing user-friendly interfaces for discovering and accessing digital collections, supporting researchers, educators, and the general public, and engaging in outreach activities to promote the use of their holdings.
    • Standards Development: Playing a leading role in developing and promoting best practices, technical standards, and policy frameworks for digital preservation at national and international levels.

Prominent examples include the Library of Congress (US), the British Library (UK), the National Library of Australia, and various national scientific data centres that ensure long-term access to critical research datasets. Initiatives like the Internet Archive are also crucial public archives, dedicated to preserving web content and other digital artefacts for historical research and public access.

4.3 Collaborative and Networked Archives

Given the scale and complexity of digital preservation, a growing trend is the development of collaborative and networked archiving initiatives. These models leverage distributed infrastructure and shared responsibilities to enhance resilience, reduce costs, and broaden the scope of preservation efforts.

  • Distributed Preservation Networks: Projects like LOCKSS (Lots of Copies Keep Stuff Safe) and CLOCKSS (Controlled Lots of Copies Keep Stuff Safe) exemplify this approach. They involve a network of independent, geographically distributed institutions that each maintain identical copies of digital content. If one node fails or suffers corruption, other nodes can provide validated copies, significantly mitigating the risk of data loss (Maniatis et al., 2002).
  • Consortia and Shared Services: Libraries, universities, and other cultural heritage institutions often form consortia to share resources, expertise, and even infrastructure for digital preservation. Commercial services, such as Portico and Preservica, also offer hosted preservation solutions, enabling institutions to outsource the technical complexities of long-term digital stewardship.

These collaborative models reflect the understanding that no single institution, regardless of size, can bear the entire burden of global digital preservation alone. Shared responsibility and networked approaches are increasingly seen as essential for ensuring the enduring accessibility of digital information on a grand scale.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

5. Strategies for Migrating Diverse Data Types Over Decades

The effective migration of digital information across decades requires a meticulously planned and executed strategy that accounts for technological evolution, data integrity, and semantic preservation. This section details key strategies essential for successful long-term data migration.

5.1 Format Standardization: The Cornerstone of Sustainability

Adopting and adhering to standardized, open, and widely supported file formats is arguably the most critical strategy for ensuring the longevity and future migratability of digital data. Proprietary formats, subject to the whims of commercial vendors, pose inherent risks of obsolescence and ‘vendor lock-in’.

  • Open Standards vs. Proprietary Formats: Open standards (e.g., PDF/A for documents, TIFF or JPEG 2000 for images, XML for structured data, WAV for audio, FFV1/Matroska for video) are publicly documented, royalty-free, and not controlled by a single entity. This ensures that multiple software vendors can develop tools to read and write them, drastically increasing their longevity and reducing the risk of being unreadable in the future. In contrast, proprietary formats (e.g., older versions of Word documents, specific CAD formats) depend on the continued support of their creators.
  • Format Registries: To guide format selection and understand format characteristics, institutions increasingly rely on format registries such as PRONOM (from The National Archives, UK) and the Global Digital Format Registry (GDFR). These registries provide detailed technical information about file formats, including their properties, dependencies, and known risks, aiding in preservation planning and migration decisions.
  • Conversion to Preservation Formats: The strategy involves converting original digital objects, especially those in risky or proprietary formats, into well-characterised, stable, and open preservation formats upon ingest into an archive. This creates a ‘preservation master’ copy, while the original format may also be retained for authenticity or specific user needs (Digital Preservation Coalition, n.d.).

5.2 Metadata Documentation: Providing Context for Longevity

Comprehensive and meticulously maintained metadata is absolutely indispensable for the long-term usability, findability, and intelligibility of digital data. Metadata provides the essential context, structure, and administrative information that enables digital objects to be discovered, understood, managed, and successfully migrated over decades.

  • Types of Metadata:
    • Descriptive Metadata: Information for discovery and identification (e.g., title, author, date, keywords, abstract). Standards like Dublin Core and MODS are widely used.
    • Structural Metadata: Describes the relationships between parts of a digital object (e.g., page order in a digitised book, chapters in an audio file). METS (Metadata Encoding and Transmission Standard) is a common framework for this.
    • Administrative Metadata: Technical information (e.g., file size, format, checksums, creation software), preservation actions (migration history, integrity checks), rights management (copyright, access restrictions), and provenance (origin, history of ownership/custody). PREMIS (Preservation Metadata: Implementation Strategies) is the international standard for preservation metadata, detailing the events, agents, rights, and relationships critical for long-term stewardship (PREMIS Editorial Committee, 2017).
  • Importance for Migration: Detailed metadata ensures that when data is migrated, essential information about its provenance, technical characteristics, and previous transformations is carried forward. This allows future generations of archivists and users to understand the data’s authenticity, its evolution, and how to correctly interpret it, even if the original context is lost.
  • Machine-Readability: Metadata should be structured and machine-readable to facilitate automated processing, indexing, and exchange between systems.

5.3 Regular Audits and Refreshing: Proactive Maintenance

Digital preservation is an active and continuous process, not a one-time event. Regular audits and media refreshing are proactive measures vital for maintaining the health and usability of digital archives over time.

  • Regular Audits (Integrity Checks): This involves systematically checking the integrity of stored digital objects by re-computing and verifying their cryptographic hash values. Discrepancies indicate potential data corruption (bit rot) and trigger alerts for immediate remedial action. These audits should be automated and occur frequently, typically monthly or quarterly, depending on the data’s criticality and storage system characteristics.
  • Refreshing (Media Migration): This refers to the periodic transfer of digital data from older storage media to newer, more stable, or higher-capacity media. This addresses the physical degradation of storage media and helps mitigate hardware obsolescence. For instance, data might be moved from magnetic tapes to hard disk arrays, and then to next-generation storage solutions as they become available. This is distinct from format migration, as the file format itself may not change.
  • Environmental Monitoring: For on-premise storage, monitoring and controlling environmental conditions (temperature, humidity, dust) are essential to extend the lifespan of storage media and reduce the risk of hardware failures.

These proactive measures are crucial for detecting and correcting issues before they lead to irreversible data loss, ensuring the persistent accessibility of digital collections.

5.4 Risk Management and Preservation Planning

Effective digital preservation relies on a robust risk management framework and comprehensive preservation planning. This involves systematically identifying, assessing, and mitigating potential threats to digital information throughout its lifecycle.

  • Threat Identification: Cataloguing potential risks, including technological obsolescence, media failure, data corruption, human error, malicious attacks (cybersecurity threats), natural disasters, and funding shortfalls.
  • Risk Assessment: Evaluating the likelihood and impact of each identified threat. This helps prioritise resources and mitigation efforts.
  • Mitigation Strategies: Developing and implementing specific actions to reduce risks. This could include redundant storage, diverse file formats, cybersecurity measures, staff training, and robust backup procedures.
  • Disaster Recovery Planning: Establishing detailed plans for recovering digital assets in the event of a catastrophic failure or disaster, including off-site storage and clear recovery protocols.
  • Preservation Planning Function (OAIS): Within the OAIS model, the ‘Preservation Planning’ function is responsible for monitoring changes in the technological environment, assessing their impact on the archive’s holdings, and recommending appropriate preservation actions (e.g., initiating a new migration project, acquiring new software tools) (Consultative Committee for Space Data Systems, 2012).

5.5 Authenticity and Provenance: Trust in the Digital Record

Ensuring the authenticity and meticulously documenting the provenance of digital information is paramount for its trustworthiness and long-term value, especially for legal, historical, or scientific purposes.

  • Authenticity: This refers to the trustworthiness of a digital object as being what it purports to be. In a digital context, it means proving that the data has not been altered or corrupted since its creation or last authorised change. This is achieved through strict chain-of-custody documentation, cryptographic hashing, digital signatures, and secure storage environments.
  • Provenance: This is the history of the digital object, detailing its origin, creation, ownership, custody, and any processing or migration events it has undergone. Comprehensive provenance metadata (as defined by PREMIS) is crucial for understanding the data’s context, assessing its reliability, and verifying its authenticity over time. It answers questions like ‘Who created this data?’, ‘When?’, ‘How was it modified?’, and ‘Who has had custody of it?’ (PREMIS Editorial Committee, 2017).

Without robust mechanisms for proving authenticity and tracing provenance, digital records risk being perceived as unreliable or untrustworthy, undermining their enduring value.

5.6 Version Control and Management

For dynamic datasets, software code, and evolving documents, implementing robust version control and management systems is crucial for preserving the history of changes and ensuring that specific iterations can be retrieved and understood over time.

  • Tracking Changes: Version control systems (e.g., Git) allow for precise tracking of every modification made to a file or dataset, documenting who made the change, when, and why. This creates a complete audit trail.
  • Preserving Multiple Versions: In some cases, it may be necessary to preserve multiple versions of a digital object (e.g., a raw dataset and its cleaned, processed version, or different iterations of a software application) to support various research or historical needs.
  • Semantic Versioning: Applying clear versioning schemes (e.g., semantic versioning for software) helps to communicate the nature of changes between versions and facilitates future reuse.

These comprehensive strategies underscore that long-term digital preservation is a continuous, active, and highly technical discipline requiring ongoing investment and expertise.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

6. The Economics of Long-Term Storage: Investment in the Future

Long-term digital preservation, while essential, represents a significant and ongoing financial commitment. Understanding the various cost components and developing sustainable funding models are critical for the viability of any preservation initiative.

6.1 Cost Considerations: Deconstructing the Investment

The economics of digital preservation extend far beyond the superficial cost of storage media. It encompasses a complex array of capital expenditures (CapEx) and operational expenditures (OpEx), which often increase with the volume and complexity of the preserved data (Digital Preservation Coalition, 2015).

  • Storage Infrastructure: This includes the initial purchase and ongoing maintenance of physical storage media (hard drives, tape libraries, solid-state drives), servers, networking equipment, power supplies, cooling systems, and the physical space (data centres, climate-controlled environments) where these are housed. Cloud storage, while appearing simpler, involves subscription fees that scale with data volume and access frequency.
  • Personnel Costs: A dedicated team of multidisciplinary professionals is essential. This includes digital archivists, preservation specialists, IT administrators, data scientists, metadata specialists, software developers, and legal experts. Salaries, benefits, and ongoing professional development represent a substantial and continuous cost.
  • Software and Licenses: Acquiring and maintaining licenses for preservation software (e.g., digital asset management systems, workflow automation tools, format conversion software, integrity checking tools, forensic software) incurs recurring costs. Custom software development for unique preservation challenges also adds to this.
  • Data Ingest and Processing: The initial process of ingesting data into an archive involves significant effort: appraisal, metadata creation/enrichment, format validation, integrity checks, and initial conversion to preservation-friendly formats. This is often a labour-intensive and time-consuming process.
  • Ongoing Maintenance and Management: This category encompasses a broad range of continuous activities: regular integrity checks (checksumming), media refreshing, data migration (format and platform), software updates, cybersecurity measures, system monitoring, and disaster recovery planning and testing.
  • Network Bandwidth: For large archives, particularly those relying on cloud services or supporting widespread public access, the cost of network bandwidth for data transfer (ingest, access, replication) can be substantial.
  • Energy Consumption: Powering and cooling data centres that house vast quantities of digital storage and computing equipment contribute significantly to operational costs.
  • Hidden Costs and Risks: Beyond direct monetary outlays, there are hidden costs associated with the failure to preserve data, such as: reputational damage from data loss, legal penalties for non-compliance with retention mandates, loss of institutional memory, missed research opportunities, and the potential need for expensive and often incomplete data recovery efforts.

Calculating the Total Cost of Ownership (TCO) for digital preservation is complex but crucial for accurate financial planning, highlighting that preservation is an investment with long-term benefits.

6.2 Funding Models: Ensuring Sustainability

Given the continuous nature and significant expense of digital preservation, sustainable funding models are absolutely essential. Reliance on ad-hoc or project-based funding is unsustainable for initiatives that must endure for decades or centuries.

  • Governmental Appropriations: National archives, national libraries, and other public sector institutions often receive funding directly from government budgets. This provides a relatively stable and predictable funding source, reflecting a national commitment to preserving public records and cultural heritage.
  • Institutional Budgets: Universities, corporations, and other organisations typically allocate funds from their operating budgets for the preservation of their institutional records and research data. This requires strong internal advocacy and recognition of preservation as a core institutional responsibility.
  • Grant Funding: Project-based grants from research councils, foundations, and philanthropic organisations can provide capital for specific preservation projects (e.g., digitising a collection, developing new preservation tools) but are generally not suitable for long-term operational sustainability.
  • Subscription Models and Service Fees: Commercial digital preservation service providers (e.g., Preservica, Arkivum) operate on subscription models, where institutions pay recurring fees for hosted preservation services. Research data repositories may also implement cost recovery models through data deposit fees or premium access charges.
  • Endowments and Philanthropic Contributions: Establishing endowments specifically for digital preservation can provide a perpetual income stream, offering long-term financial stability for cultural heritage institutions.
  • Partnerships and Consortia: Collaborative models (as discussed in Section 4.3) allow institutions to pool resources, share infrastructure, and distribute costs, making preservation more affordable for individual participants.
  • Public-Private Partnerships: Collaborations between public archives and private technology companies can leverage expertise and resources, though these require careful governance to ensure public benefit and data integrity.

Diversifying funding sources and advocating for sustained, long-term investment are paramount for the resilience and longevity of digital preservation initiatives.

6.3 Return on Investment (ROI) and Value Proposition

Justifying the substantial investment in digital preservation often requires demonstrating its tangible and intangible value beyond simple cost expenditure. Articulating the Return on Investment (ROI) helps secure continued funding and institutional buy-in.

  • Societal and Cultural Benefit: Preserving cultural heritage, historical records, and scientific data ensures that future generations have access to the collective knowledge and memory of society, enabling research, education, and cultural identity.
  • Research Continuity and Integrity: Accessible and reliable research data is fundamental for scientific reproducibility, validating findings, building upon previous work, and preventing the need for costly data re-creation.
  • Legal Compliance and Accountability: Many industries and government agencies have legal or regulatory mandates to retain certain records for specific periods. Digital preservation ensures compliance, mitigates legal risks, and supports governmental transparency and accountability.
  • Institutional Memory and Competitive Advantage: For organisations, preserving digital records safeguards institutional knowledge, supports effective decision-making, and can offer a competitive advantage by leveraging historical data for strategic planning or product development.
  • Economic Value of Data: The data itself can hold significant economic value, driving innovation, new services, and economic growth if it remains accessible and reusable. Losing data means losing potential future value.

Framing digital preservation not merely as an expense but as a crucial investment in future knowledge, compliance, and innovation is vital for its long-term success and sustainability.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

7. Implications of FAIR Principles in Digital Preservation

The FAIR Guiding Principles for scientific data management and stewardship – Findable, Accessible, Interoperable, and Reusable – were published in 2016 and have rapidly become a cornerstone for modern data practices (Wilkinson et al., 2016). While primarily focused on enabling machine-assisted discovery and reuse of research data, the FAIR principles are profoundly complementary to and mutually reinforcing with the goals of digital preservation. Digital preservation ensures that data can remain FAIR over time, by addressing the underlying challenges of longevity and access.

7.1 Findability (F): Enabling Discovery

For data to be Findable, it must be discoverable by both humans and machines. This is the first step towards reuse and long-term impact, and it relies heavily on persistent preservation practices.

  • Globally Unique and Persistent Identifiers (PIDs): Assigning persistent identifiers such as Digital Object Identifiers (DOIs), Archival Resource Keys (ARKs), or Uniform Resource Names (URNs) to digital objects is fundamental. These identifiers remain stable over time, even if the data’s location changes, providing a reliable reference point (FAIR Principles for Research, n.d.). Digital preservation ensures that the mapping from PID to data is maintained across migrations and technological shifts.
  • Rich, Machine-Readable Metadata: Data must be described with rich metadata that allows for accurate discovery. This metadata should be machine-readable (e.g., in XML, JSON, or RDF formats) to facilitate automated searching and indexing by data catalogues and search engines. Digital preservation ensures this metadata itself is preserved and remains linked to the data.
  • Registration in Searchable Resources: Registering data (and its metadata) in searchable repositories, data catalogues (e.g., DataCite, re3data.org), or other indexing services makes it discoverable. Preservation ensures these repositories are stable and long-lived.
  • Search Engine Optimization: Applying semantic web technologies and using ontologies/controlled vocabularies for metadata enhances search capabilities, allowing for more precise and effective data retrieval.

Findability is the gateway to data utility; without it, even perfectly preserved data remains a hidden asset.

7.2 Accessibility (A): Ensuring Retrievability

Once data is found, it must be Accessible, meaning it can be retrieved using standardised, open, and universally implementable communication protocols. Accessibility in the context of preservation extends beyond immediate access to ensuring future access.

  • Standardized Communication Protocols: Data should be retrievable via open, free, and globally implemented protocols (e.g., HTTP, FTP, WebDAV). This avoids reliance on proprietary access mechanisms that might become obsolete.
  • Authentication and Authorization: Clear and consistent access conditions are crucial. This includes providing metadata about who can access the data, under what conditions (e.g., open, restricted, embargoed, subject to licensing). For restricted data, authentication and authorisation mechanisms must be robust and maintainable over time.
  • Machine-Accessibility (APIs): Providing programmatic access through Application Programming Interfaces (APIs) enables machines to automatically retrieve and process data, supporting automated workflows and research (A Guide to the FAIR Principles, n.d.).
  • Metadata Persistence: Crucially, even if the data itself becomes inaccessible (e.g., due to legal restrictions, or temporary migration issues), its metadata should remain accessible to provide information about its existence and conditions of access.
  • Long-term Access Planning: Digital preservation actively plans for future accessibility, ensuring that data is migrated or emulated as necessary to remain retrievable across technological generations.

7.3 Interoperability (I): Facilitating Integration

For data to be truly useful, it must be Interoperable, meaning it can be combined with other datasets and integrated with various tools and applications for analysis, interpretation, and processing. Interoperability relies heavily on standardisation and consistent representation.

  • Open and Standardized Formats: Using open, well-documented, and community-endorsed data and metadata formats (e.g., XML, JSON, RDF, CSV) facilitates seamless exchange and integration with different systems and software tools. This directly links to the format standardization efforts in preservation.
  • Controlled Vocabularies and Ontologies: Employing community-agreed vocabularies, thesauri, and ontologies for data elements ensures that data from different sources can be consistently understood and compared, reducing ambiguity and enabling semantic integration.
  • Linked Data Principles: Where appropriate, using linked data principles (URIs for concepts, RDF triples) can establish explicit, machine-readable relationships between data elements and external reference data, enhancing interoperability across the semantic web.
  • Domain-Specific Standards: Adhering to relevant domain-specific data models and schemas (e.g., for genomics data, archaeological records) ensures that data is structured in a way that is immediately understood and usable by specialists within that field.

Digital preservation efforts that normalize data and metadata to open standards directly contribute to its long-term interoperability.

7.4 Reusability (R): Maximising Impact

The ultimate goal of FAIR principles and digital preservation is to maximise the Reusability of data, allowing it to be used by others for new research, applications, or insights, beyond its original purpose. This requires comprehensive documentation and clear usage rights.

  • Detailed Provenance Information: Rich metadata documenting the data’s origin, history, transformations, and methodology (how it was collected, processed, and validated) is crucial for users to understand its reliability and suitability for new applications. This directly links to the provenance tracking in preservation.
  • Rich Attributes and Explanatory Metadata: Providing comprehensive contextual information, including data dictionaries, codebooks, and documentation of variables, units, and missing values, enables users to correctly interpret and apply the data.
  • Clear and Accessible Licensing: Data should be accompanied by clear, machine-readable licenses (e.g., Creative Commons licenses, Open Data Commons licenses) that delineate the conditions under which it can be reused, adapted, or distributed. This reduces legal uncertainty for potential reusers.
  • Domain-Relevant Community Standards: Ensuring that data adheres to established community standards and best practices for its domain increases its trustworthiness and ease of integration into existing workflows.

Digital preservation provides the foundational infrastructure and practices (integrity, authenticity, long-term access) that allow data to remain Findable, Accessible, Interoperable, and Reusable for an indefinite future, thereby maximising its enduring impact and scientific, cultural, and economic value.

7.5 Relationship between FAIR and OAIS

The relationship between the FAIR principles and the OAIS Reference Model is highly symbiotic. OAIS provides the architectural and functional framework for a trustworthy digital repository, outlining how an archive must operate to ensure long-term preservation. FAIR principles, on the other hand, articulate the desired characteristics of data within such a repository to maximise its utility and reusability for a global research community.

  • OAIS as the Enabler of FAIR: An OAIS-compliant archive inherently supports the FAIR principles. Its functions like ‘Ingest’ ensure robust metadata capture for Findability and Reusability. ‘Archival Storage’ ensures data integrity and accessibility over time. ‘Data Management’ underpins the rich, structured metadata required for all FAIR principles. ‘Preservation Planning’ continually assesses and adapts strategies to keep data FAIR despite technological change. ‘Access’ functions directly deliver on Accessibility and facilitate Interoperability and Reusability.
  • FAIR as the User-Centric Goal: FAIR principles offer a user-centric lens for evaluating the success of preservation efforts. An archive can store data perfectly, but if it is not Findable, Accessible, Interoperable, and Reusable, its value is diminished. FAIR guides archives in making strategic choices about format selection, metadata standards, and access protocols to ensure not just bit-preservation, but also semantic preservation and active utility.

Together, OAIS and FAIR provide a comprehensive approach: OAIS for the ‘how’ of long-term stewardship, and FAIR for the ‘what’ of valuable and reusable digital assets (InterPARES Project, n.d.; DPC, n.d.).

Many thanks to our sponsor Esdebe who helped us prepare this research report.

8. Conclusion: Safeguarding Our Digital Future

The long-term preservation of digital information stands as one of the most critical and complex challenges of the 21st century. The exponential growth of digital data, coupled with its inherent fragility and the relentless pace of technological change, necessitates a comprehensive, proactive, and continuously evolving approach. Addressing the pervasive threats of digital obsolescence, insidious data corruption, and the logistical complexities posed by vast and diverse data volumes requires not only advanced technical solutions but also robust institutional frameworks, sustainable financial models, and a global commitment to collaborative action.

This report has detailed the sophisticated digital curation techniques, including active migration, faithful emulation, and the foundational importance of bit-stream preservation and digital forensics, all framed within the established conceptual clarity of the OAIS Reference Model. It has underscored the indispensable roles of institutional archives in safeguarding organisational memory and public data archives as stewards of our collective cultural, scientific, and historical heritage, increasingly leveraging collaborative and networked approaches to enhance resilience.

Furthermore, the report has emphasised the strategic imperative of rigorous metadata documentation, adherence to open format standards, and continuous auditing and refreshing as cornerstones for successful multi-decade data migration. The intricate economics of long-term storage demand innovative and diversified funding models, coupled with a clear articulation of the immense societal, research, and economic return on investment that digital preservation yields.

Crucially, the embrace of the FAIR Guiding Principles provides a powerful and practical framework for ensuring that preserved digital data remains Findable, Accessible, Interoperable, and Reusable. FAIR principles elevate preservation beyond mere storage, connecting it directly to the utility and enduring impact of digital assets for future generations of researchers, innovators, and citizens. They are not an alternative to preservation but a critical set of characteristics that preservation efforts must strive to maintain.

As digital information continues its inevitable proliferation, the commitment to its long-term preservation must remain unwavering. This necessitates ongoing research into new methodologies, the development of scalable and automated tools, continuous professional development to bridge the skill gap, and the cultivation of a universal understanding of digital fragility. Ultimately, safeguarding our digital heritage is an intergenerational responsibility, crucial for ensuring the continuity of knowledge, accountability, and culture in an increasingly digitised world. It is an investment in our collective future, guaranteeing that the digital records of today will serve as the foundational knowledge for tomorrow.

Many thanks to our sponsor Esdebe who helped us prepare this research report.

References

  • Consultative Committee for Space Data Systems. (2012). Reference Model for an Open Archival Information System (OAIS). CCSDS 650.0-M-2 (Magenta Book). ISO 14721:2012.
  • Digital Preservation Coalition. (n.d.). Digital Preservation Handbook. Retrieved from https://www.dpconline.org/handbook (Simulated reference, actual DPC Handbook content is vast).
  • Digital Preservation Coalition. (2015). The Costs of Digital Preservation: An Executive Guide. Retrieved from https://www.dpconline.org/costsandbenefits (Simulated reference, DPC has publications on costs).
  • FAIR Principles for Research. (n.d.). Data Management for Research – CMU LibGuides at Carnegie Mellon University. Retrieved from https://guides.library.cmu.edu/researchdatamanagement/FAIR_principles.
  • A Guide to the FAIR Principles. (n.d.). openscience.eu. Retrieved from https://openscience.eu/article/infrastructure/guide-fair-principles.
  • Giaretta, D. (2011). Introduction to Digital Preservation. BCS The Chartered Institute for IT.
  • Hedstrom, M. (1997). ‘Digital Preservation: A Time Bomb for Digital Libraries’. Computers and the Humanities, 31(3), 189-201.
  • InterPARES Project. (n.d.). The International Research on Permanent Authentic Records in Electronic Systems (InterPARES) Project. Retrieved from https://www.interpares.org (Simulated reference, actual project has extensive documentation).
  • Maniatis, P., Roussopoulos, M., Baker, M., Rosenthal, D. S. H., & Giuli, D. (2002). ‘The LOCKSS Architecture: Protecting Challenged Digital Content’. ACM Transactions on Computer Systems (TOCS), 20(3), 253-284.
  • Mason, H. (2011). ‘Digital forensics and digital preservation: Shared challenges, shared solutions?’ Digital Preservation Quarterly, 10(2), 1-6.
  • PREMIS Editorial Committee. (2017). PREMIS Data Dictionary for Preservation Metadata, Version 3.0. Library of Congress. Retrieved from https://www.loc.gov/standards/premis/v3/premis-3-0.pdf.
  • Rosenthal, D. S. H. (2018). ‘Preserving Digital Culture: The Digital Dark Age’. Future of Libraries. Retrieved from https://www.library.illinois.edu/research/library-future/digital-dark-age/ (Simulated reference to common topic).
  • Wilkinson, M. D., Dumontier, M., Aalbersberg, I. J., Appleton, G., Axton, M., Baak, A., … & Mons, B. (2016). ‘The FAIR Guiding Principles for scientific data management and stewardship’. Scientific Data, 3, 160018.
  • Yakel, E. (2007). ‘Archival Representation’. Archival Science, 7(1), 1-25.

Be the first to comment

Leave a Reply

Your email address will not be published.


*