The document summarizes key topics from Session 1 of a workshop on electronic clinical trials archiving, including:
1) The need to select file formats that are open, ubiquitous, stable and support long-term preservation when archiving electronically.
2) The unique challenges of eArchiving compared to traditional archiving, such as the irrelevant of original/copy and need to retain information on digital formats.
3) Criteria for selecting preservation-suitable file formats, such as open standards, stability and interoperability.
This presentation describes a case study of an eTMF structure, setup, and implementation at a biotechnology company currently conducting a pivotal Phase 3 clinical study in 70 centers worldwide.
TMF Fundamentals - An Introduction to Better Trial Master File Management - M...Montrium
In this presentation, we explore the TMF fundamentals and an introduction to better managing your TMF. We'll start by diving into the world of TMF management, how to properly leverage the regulations, which documents constitute as TMF-worthy and what basic metrics you can track to increase the efficiency of your trial.
Finally, we will also discuss some of the features Montrium has developed to facilitate the management of the Trial Master File in a fundamental way.
This presentation covers the following topics:
-Fundamentals of TMF management process
-How to use the regulations to ensure success
-What cross functional groups hold TMF worthy documents
-TMF management challenges and how you can alleviate them
-Base metrics to track and what they mean to your organization
-How an eTMF helps TMF Management
You can follow along with this presentation via our webinar:
https://info.montrium.com/tmf-fundamentals-an-introduction-to-better-trial-master-file-management
This Risk Based Monitoring - Impact on Sites overview presentation, on targeted topics, was delivered to the ACRP Raleigh - Durham Chapter's Annual Conference in 2013.
Review of essential documents (TMF BABE).Piyush Wagh
This ppt is useful for the Clinical Auditors (BABE), it helps in reviewing of essential documents and compilation of essential documents during clinical phase. Bioequivalence
Practical Steps to Selecting and Implementing an eTMFMontrium
There are many benefits to moving to an electronic Trial Master File (eTMF) and many organizations are either thinking of moving, or are already in the process of doing so.
However, there are a great deal of things to consider before being able to successfully move to an electronic platform which are often overlooked, often resulting result in unsuccessful or sub-optimal eTMF implementations.
This webinar takes you through some of the practical steps that you need to go through to successfully implement and benefit from an eTMF.
Using Vault eTMF Milestones and EDLs to Support Inspection ReadinessVeeva Systems
Hear how Daiichi Sankyo is using milestones and expected document lists (EDLs) to enable ongoing inspection readiness and proactive TMF management and oversight.
This presentation describes a case study of an eTMF structure, setup, and implementation at a biotechnology company currently conducting a pivotal Phase 3 clinical study in 70 centers worldwide.
TMF Fundamentals - An Introduction to Better Trial Master File Management - M...Montrium
In this presentation, we explore the TMF fundamentals and an introduction to better managing your TMF. We'll start by diving into the world of TMF management, how to properly leverage the regulations, which documents constitute as TMF-worthy and what basic metrics you can track to increase the efficiency of your trial.
Finally, we will also discuss some of the features Montrium has developed to facilitate the management of the Trial Master File in a fundamental way.
This presentation covers the following topics:
-Fundamentals of TMF management process
-How to use the regulations to ensure success
-What cross functional groups hold TMF worthy documents
-TMF management challenges and how you can alleviate them
-Base metrics to track and what they mean to your organization
-How an eTMF helps TMF Management
You can follow along with this presentation via our webinar:
https://info.montrium.com/tmf-fundamentals-an-introduction-to-better-trial-master-file-management
This Risk Based Monitoring - Impact on Sites overview presentation, on targeted topics, was delivered to the ACRP Raleigh - Durham Chapter's Annual Conference in 2013.
Review of essential documents (TMF BABE).Piyush Wagh
This ppt is useful for the Clinical Auditors (BABE), it helps in reviewing of essential documents and compilation of essential documents during clinical phase. Bioequivalence
Practical Steps to Selecting and Implementing an eTMFMontrium
There are many benefits to moving to an electronic Trial Master File (eTMF) and many organizations are either thinking of moving, or are already in the process of doing so.
However, there are a great deal of things to consider before being able to successfully move to an electronic platform which are often overlooked, often resulting result in unsuccessful or sub-optimal eTMF implementations.
This webinar takes you through some of the practical steps that you need to go through to successfully implement and benefit from an eTMF.
Using Vault eTMF Milestones and EDLs to Support Inspection ReadinessVeeva Systems
Hear how Daiichi Sankyo is using milestones and expected document lists (EDLs) to enable ongoing inspection readiness and proactive TMF management and oversight.
An overview of archiving of clinical studies and data. By RISHI MAHESHWARI , JSS COLLEGE OF PHARMACY , OOTY
For students in V PharmD this topic has been prepared.
A Trial Master File (TMF) is a comprehensive collection of essential documents and records that are generated or collected during the conduct of a clinical trial. The TMF serves as the centralized repository of all study-related documentation and provides a complete and accurate account of the trial's planning, execution, and outcomes. It is an important component of Good Clinical Practice (GCP) and regulatory compliance.
Sandra Maddock, RN, BSN, CCRA and President of IMARC Research, Inc. presents on Applying FDA’s Risk-Based Approach in an audio conference on September 11, 2012.
How to prepare and handle a FDA Inspection. When the FDA conducts an audit of your company, you need to be prepared. This presentation shows you what to do when your are faced with a FDA Inspection. For more information go to http://compliance-insight.com/fda-483-warning-letters/fda-483-inspection/
Clinical Data Management Plan_Katalyst HLSKatalyst HLS
Introduction to Data Management Plan in Clinical Data Management in Clinical Trials of Pharmaceuticals, Bio-Pharmaceuticals, Medical Devices, Cosmeceuticals and Foods.
Life science companies expect electronic trial master file (eTMF) projects to provide greater efficiency, improve quality, increase standardization and enhance collaboration in TMF management. Join Paragon Solutions for the second session of our TMF webinar series to gain insights on the practical factors to consider when planning a new or improved future state eTMF. We'll share industry data on eTMF drivers, and discuss our definition of a robust eTMF. You'll learn why TMF information architecture is so critical to eTMF uptake and usability, and high level considerations for systems integration/migration. Finally, we'll define the maturity framework that allows for assessment of eTMF organizational readiness, a critical baseline to successful future state eTMF planning.
TSDP tells about the essential documents that are required for the #conduct of a clinical trial. For #regulatory medical writing training, contact hello@turacoz.in.
A Trial Master File (TMF) is set up at the beginning of the trial. It is a collection of all essential documents pertaining to the trial, which in turn will allow for effective monitoring and supervision (audit). In order to demonstrate compliance with the applicable regulations, Good Clinical Practice (GCP) guidelines and the protocol – a well organised TMF is essential.
According to the GCP guidelines, it is the responsibility of the sponsor to ensure that the TMF includes all relevant essential documents, and is stored in a secure location, with restricted access. Generally, the TMF is maintained at the sponsor’s office, co-ordinating site or by the Contract Research Organisation (CRO), if contracted. In addition to the TMF, copies of all relevant documents must be kept at each participating site, in an Investigator Site File (ISF). The ISF will also include all site-specific essential documents. For example, site preparedness log or site visit logs, etc.
A member of the research or trial team should be delegated with the task of updating, maintaining and reviewing the TMF and ISF, periodically throughout the course of the clinical trial as per the defined SOPs. Ideally, the documents included in the TMF are:
Trial documents (protocol, investigator’s brochure, participant information documents, SOPs, instructions, manuals, guidelines, etc.)
Documents related to the Investigational Product (certificates of analysis, shipment records, storage records, etc.)
Training documentation for the trial team
Details of the laboratories, if applicable.
Contracts, agreements, budgets, etc.
Monitoring visit reports (for each site visit onsite or central)
Documents related to the safety reporting
Ethics Committees documents (composition of the EC, approvals, notifications, reports, etc.)
Site-specific documents (list of site staff and their curriculum vitae, investigator’s undertaking, site preparedness documents, training of site staff, etc.)
Audit related documents, if available (if an audit was conducted).
Significant communications
Others
The GCP guidelines provide comprehensive guidance regarding the documents to be included in a Trial Master File categorised according to the lifecycle of the trial. This information can also be accessed here.
It shall be the responsibility of the sponsor to make arrangements for the safe and secure custody of all study-related documents and material for a period of three years after the completion of the study or submission of the data to the regulatory authority(ies) whichever is later.
After reviewing the FDA regulations on Risk Based Monitoring, review the details on how to put the principles into action! We include two reference documents to help you get started... and to make it a success.
Scientific & systematic collection of data for clinical study is called as Clinical data management .
EDC
RDC
HISTORY
EVOLUTION OF CLINICAL DATA CAPTURE
CRITERIA FOR IDENTIFYING AN EDC
REGULATORY GUIDELINE ON EDC
EDC ISSUES
VALIDATING ELECTRONIC SOURCE DATA
In any work or process documents that are needed before initiation, Between or generally the end of the process just like in a clinical trial those “Documents which permit evaluation of the conduct of a trial and the quality of the data produced. It is given in the 8th section of the ICH-GCP.
Feasibility Solutions to Clinical Trial Nightmaresjbarag
Slow patient recruitment and poor retention cause recurrent nightmares and perpetual problems often resulting in missing recruitment milestones. The cost of these delays represents hundreds of thousands of dollars for drug and device developers. By recognizing this issue, early detailed feasibility can provide planning and contingency solutions that are focused on reducing the impact of delayed recruitment. Furthermore understanding what motivates investigators and patients to actively participate in clinical studies and how patient recruitment strategies and materials can support all stakeholders to complete studies on time are critical aspects of clinical study delivery planning.
During this presentation, an experienced Premier Research feasibility and patient recruitment specialist, reviewed feasibility approaches to address protocol evaluation as well as addressed influences on country selection, site distribution and patient recruitment strategies to provide for more effective clinical trial planning and conduct.
For more information, go to http://www.premier-research.com.
An overview of archiving of clinical studies and data. By RISHI MAHESHWARI , JSS COLLEGE OF PHARMACY , OOTY
For students in V PharmD this topic has been prepared.
A Trial Master File (TMF) is a comprehensive collection of essential documents and records that are generated or collected during the conduct of a clinical trial. The TMF serves as the centralized repository of all study-related documentation and provides a complete and accurate account of the trial's planning, execution, and outcomes. It is an important component of Good Clinical Practice (GCP) and regulatory compliance.
Sandra Maddock, RN, BSN, CCRA and President of IMARC Research, Inc. presents on Applying FDA’s Risk-Based Approach in an audio conference on September 11, 2012.
How to prepare and handle a FDA Inspection. When the FDA conducts an audit of your company, you need to be prepared. This presentation shows you what to do when your are faced with a FDA Inspection. For more information go to http://compliance-insight.com/fda-483-warning-letters/fda-483-inspection/
Clinical Data Management Plan_Katalyst HLSKatalyst HLS
Introduction to Data Management Plan in Clinical Data Management in Clinical Trials of Pharmaceuticals, Bio-Pharmaceuticals, Medical Devices, Cosmeceuticals and Foods.
Life science companies expect electronic trial master file (eTMF) projects to provide greater efficiency, improve quality, increase standardization and enhance collaboration in TMF management. Join Paragon Solutions for the second session of our TMF webinar series to gain insights on the practical factors to consider when planning a new or improved future state eTMF. We'll share industry data on eTMF drivers, and discuss our definition of a robust eTMF. You'll learn why TMF information architecture is so critical to eTMF uptake and usability, and high level considerations for systems integration/migration. Finally, we'll define the maturity framework that allows for assessment of eTMF organizational readiness, a critical baseline to successful future state eTMF planning.
TSDP tells about the essential documents that are required for the #conduct of a clinical trial. For #regulatory medical writing training, contact hello@turacoz.in.
A Trial Master File (TMF) is set up at the beginning of the trial. It is a collection of all essential documents pertaining to the trial, which in turn will allow for effective monitoring and supervision (audit). In order to demonstrate compliance with the applicable regulations, Good Clinical Practice (GCP) guidelines and the protocol – a well organised TMF is essential.
According to the GCP guidelines, it is the responsibility of the sponsor to ensure that the TMF includes all relevant essential documents, and is stored in a secure location, with restricted access. Generally, the TMF is maintained at the sponsor’s office, co-ordinating site or by the Contract Research Organisation (CRO), if contracted. In addition to the TMF, copies of all relevant documents must be kept at each participating site, in an Investigator Site File (ISF). The ISF will also include all site-specific essential documents. For example, site preparedness log or site visit logs, etc.
A member of the research or trial team should be delegated with the task of updating, maintaining and reviewing the TMF and ISF, periodically throughout the course of the clinical trial as per the defined SOPs. Ideally, the documents included in the TMF are:
Trial documents (protocol, investigator’s brochure, participant information documents, SOPs, instructions, manuals, guidelines, etc.)
Documents related to the Investigational Product (certificates of analysis, shipment records, storage records, etc.)
Training documentation for the trial team
Details of the laboratories, if applicable.
Contracts, agreements, budgets, etc.
Monitoring visit reports (for each site visit onsite or central)
Documents related to the safety reporting
Ethics Committees documents (composition of the EC, approvals, notifications, reports, etc.)
Site-specific documents (list of site staff and their curriculum vitae, investigator’s undertaking, site preparedness documents, training of site staff, etc.)
Audit related documents, if available (if an audit was conducted).
Significant communications
Others
The GCP guidelines provide comprehensive guidance regarding the documents to be included in a Trial Master File categorised according to the lifecycle of the trial. This information can also be accessed here.
It shall be the responsibility of the sponsor to make arrangements for the safe and secure custody of all study-related documents and material for a period of three years after the completion of the study or submission of the data to the regulatory authority(ies) whichever is later.
After reviewing the FDA regulations on Risk Based Monitoring, review the details on how to put the principles into action! We include two reference documents to help you get started... and to make it a success.
Scientific & systematic collection of data for clinical study is called as Clinical data management .
EDC
RDC
HISTORY
EVOLUTION OF CLINICAL DATA CAPTURE
CRITERIA FOR IDENTIFYING AN EDC
REGULATORY GUIDELINE ON EDC
EDC ISSUES
VALIDATING ELECTRONIC SOURCE DATA
In any work or process documents that are needed before initiation, Between or generally the end of the process just like in a clinical trial those “Documents which permit evaluation of the conduct of a trial and the quality of the data produced. It is given in the 8th section of the ICH-GCP.
Feasibility Solutions to Clinical Trial Nightmaresjbarag
Slow patient recruitment and poor retention cause recurrent nightmares and perpetual problems often resulting in missing recruitment milestones. The cost of these delays represents hundreds of thousands of dollars for drug and device developers. By recognizing this issue, early detailed feasibility can provide planning and contingency solutions that are focused on reducing the impact of delayed recruitment. Furthermore understanding what motivates investigators and patients to actively participate in clinical studies and how patient recruitment strategies and materials can support all stakeholders to complete studies on time are critical aspects of clinical study delivery planning.
During this presentation, an experienced Premier Research feasibility and patient recruitment specialist, reviewed feasibility approaches to address protocol evaluation as well as addressed influences on country selection, site distribution and patient recruitment strategies to provide for more effective clinical trial planning and conduct.
For more information, go to http://www.premier-research.com.
Presentation arsip nov 2012 frans smit handoutGemeente Almere
Many small steps make one Journey or: how to create an infrastructure for digital preservation. Jakarta, November 8, 2012. National Archives Conference Indonesia.
Presentation slides from a talk given at RSP 'Goes back to' School 2009, Matfen Hall, Nr. Hexham, Northumberland, 14-16 September 2009. The actual presentation on the 15 September only covered the content up to Slide 33. The remainder includes a more detailed reflection on the curation of research data, left in to provide additional context for those using the full presentation.
NCompass Live - Nov. 21, 2018
http://nlc.nebraska.gov/ncompasslive/
To enhance access to their diverse materials, libraries are digitizing those materials and making them freely available online as digital collections on digital platforms. These collections provide another way for libraries to re-envision their materials and make them relevant to their communities. This presentation will cover best practices for creating and preserving digital collections, including workflows, standards, and staffing. It will also discuss the policies which should be developed for building successful digital collections, as well as the privacy issues which should be considered. In this presentation, individual digital collections from the University of Nebraska at Omaha and Creighton University Law Library, including the Omaha Oral History Collection and the Delaney Tokyo Trial Papers, will be demonstrated.
Presenters: Corinne Jacox, Catalog/Reference Librarian, Creighton University Law Library & Yumi Ohira, Digital Initiatives Librarian, UNO Criss Library.
The paper presents the literature review on long term preservation of 3D architectural building data. The review identified the existing gap in the research and practice of the long term preservation of 3D architectural models,
and suggested future research opportunities in this domain.
The Oxford Common File Layout: A common approach to digital preservationSimeon Warner
The Oxford Common File Layout (OCFL) specification began as a discussion at a Fedora/Samvera Camp held at Oxford University in September of 2017. Since then, it has grown into a focused community effort to define an open and application-independent approach to the long-term preservation of digital objects. Developed for structured, transparent, and predictable storage, it is designed to promote sustainable long-term access and management of content within digital repositories. This presentation will focus on the motivations and vision for the OCFL, explain key choices for the specification, and describe the status of implementation efforts.
Temporal relations in queries of ehr data for researchWolfgang Kuchinke
Temporal Relations in Queries of Electronic Patient Records. Our main scenario covers the patient identification and recruitment process for clinical trials. For this purpose an extension of the EHR4CR workbench to support patient recruitment was created. This workbench covers following requirements:
Need for built-in privacy protection.
Patient identification and recruitment tracking.
Availability at clinical sites in the form of a workbench with an user-friendly interface.
Each participating clinical site has its own installation only used locally.
Ability to generate queries with temporal relations and constraints for eligibility criteria to find candidate patients.
Our development is based on the fact that queries in EHRs often have a temporal component. But available user interfaces allow only the generation of simple queries with basic temporal relations. Time points and time intervals are therefore the main concepts that must be considered. Time points are related to instantaneous events (e.g. a single myocardial infarction), or to situations lasting for a span of time (e.g. a drug therapy for 2 weeks). Intervals can be represented using time points by their upper and lower temporal boundaries: the start and end. Temporal relations (e.g before, after) can be expressed via additional anchors. The dates of these anchor events can be retrieved and event dates relative to an anchor event can be calculated. EHR4CR decided to build its workbench upon a simple, time-stamp database concept. To each patient’s attribute a time-stamp, which corresponds to the time of the attribute’s occurrence was assigned. The processing of temporal intervals is necessary for EHR4CR since many questions dealing with inclusion / exclusion criteria often involve complex temporal periodes. A graphical interface to use boxes for querying with temporal relations was therefore created. The idea is that the easiest way to specify temporal operators is with an user interface based on the combination of boxes. Temporal operators based on Allen’s algebra were included. Expressions are displayed as graphic boxes and combined by
operators. Events are specified and a temporal operator selected from a predefined list.
Secure access to biomedical data sources for legal data sharing-kuchinkeWolfgang Kuchinke
Enabling Secure Access to Biomedical Data Sources. This is about a pilot installation of a data sharing system for legally compliant data access.
The first step was an analysis of ethical, legal and regulatory requirements for data access and of associated data security risks due to data sharing between Research Infrastructures. Analysis was followed by the development of a security framework to ensure that the data bridges of BioMedBridges project are compliant with regulations and consider privacy and data security requirements.
Elements of the security architecture and pilot implementation consist of the BioSamples Database from ELIXIR, BBMRI catalogue (BBMRI Hub) extended with a MIABIS layer and a data cube layer, Resource Entitlement Management System and add-on of Legal Assessment Tool. Central in the pilot was the role of the Resource Entitlement Management System (REMS), which manages policies for granting access to data providers (databases). For example: an application for granting access is required to get access to clinical data from a web application like BBMRI Hub. The approval is granted by a Data Access Committee (DAC). The process for access is managed by Shibboleth, a single sign-on login system. To get access, Identity providers (IdPs) supply user information, while service providers (SPs) use this information to give access. The Legal Assessment Tool (LAT) is an additional component to guide users through legal requirements for using data sharing of sensitive data with data bridges.
The Pilot workflow process: The Biosamples Database is the starting point. Using the linked LAT the researcher searching for specific data gets legal / ethical requirements to consider when accessing and sharing human data. The BBMRI Hub provides summary information of the requested datasets (makes it findable in the Biosamples Database). The Hub checks that the user has an associated Shibboleth session. If the user has no Shibboleth session, Shibboleth redirects the user to a sign-on page. After successful authentication the request is sent back to the Hub, where the user session is used to query REMS related attributes. Approved applications are notified by email including an link to the access site with data sets.
Computer System Validation with privacy zones, e-source and clinical trials b...Wolfgang Kuchinke
Clinical Trials in the Learning Health System: Computer System Validation of eSource and EHR Data. Basic question is how to make a clinical trial data management system that uses EHR data, Patient Reported Outcome (PRO) and eSource data as part of the Learning Health System compliant with regulations and with Good Clinical Practice (GCP)? Computer System Validation (CSV) is a requirement for all computer systems involved in clinical trials for drug submission. It consists of documented processes to produce evidence that a computerized system does exactly what it is designed to do in a consistent and reproducible manner. Validation begins with the system requirements definition and continues until system retirement. For example, the components of a clinical trials
framework used in our case are: Patient eligibility checks and enrolment, pre-population of eCRFs with data from EHRs, PROM data collection by patients, storing of a copy of study data in the EHR, and validation of the Study System that coordinates all study and data collection events.
eSource direct data entry in clinical trials and GCP requirements. It is the duty of physicians who are involved in medical research to protect the privacy and confidentiality of personal information of research subjects. Any eSource system should be fully compliant with the provisions of applicable data protection legislation. This creates the need to develop and implement processes that ensure the continuous control of the investigators over these data. This has to be the focus of CSV. Clinical Data drive the LHS. The results from randomized controlled trials are seen as the “gold standard” for medical evidence, but such trials are often performed outside the usual system of care and recruit highly selected populations. For this reason, the concept of using data gathered directly from the patient care environment has enormous potential for accelerating the rate at which useful knowledge is generated.
This leads to the requirement for validating electronic source data in clinical trials. This includes validation for clinical data that is either captured from the subject directly or from the subject’s medical records. The problem is the correct and appropriate system validation of electronic source data. The main componenets of CSV are the Validation Master Plan), User Requirements Specification, Hardware Requirements Specification, Design qualification, Installation qualification, Operational qualification, Performance qualification.
Any instrument used to capture source data should ensure that the data are captured as specified within the protocol. Source data should be accurate, legible, contemporaneous, original, attributable, complete and consistent. An audit trail should be maintained as part of the source documents for the original creation and subsequent modification of all source data.
Personalized medicine tools for clinical trials - kuchinkeWolfgang Kuchinke
Tools for personalised medicine in clinical trials. ---------
The implementation of clinical trials in personalized medicine is a different way of doing clinical research compared to the standard way of large clinical trials aiming for statistical significance. Personalized medicine uses a medical model that separates people into different groups with medical decisions, practices, drugs, interventions being tailored to the individual patient based on their predicted response. Basis for this approach is the progress of the study of the human genome and its variation over the last two decades. Especially advancement in automated DNA sequencing and PCR and the use of expressed sequence tags (ESTs), cDNAs, antisense molecules, small nterfering RNAs (siRNAs), full-length genes and their expression products and haplotypes.
But adoption of personalized medicine requires an active and flexible and highly integrated infrastructure, which allows joining of many different competences and technologies. We asked the question: can the tools developed for personalized medicine in the p-pedicine project be employed effectively in a clinical trials network to support personalised clinical trials. We conducted an analysis of tool integration and the evaluation tool usage requirements. Based on the survey results, the tendency for clinical trial network ECRIN is to use software as a service in the form as SaaS or ASP. ECRIN data centres will (probably) not install and employ p-medicine tools in one of their data centres. A robust business model for the provision of services and the implementation and employment of tools does not yet exist.
How can the personalized medicine infrastructure p-medicine and the clinical trials network ECRIN gain from each other to allow the conduct of personalized clinical trials?
We suggest a business model, in which personal medicine infrastructures and clinical trials networks exchange their services to gain jointly from each other. Therefore: an integration by reciprocal exchange of services may be the solution. Not only software as a service will be exchanged, but also knowledge, personnel and joint staff trainings.
Computer System Validation - privacy zones, eSource and EHR data in clinical ...Wolfgang Kuchinke
Clinical Trials in the Learning Health System (LHS): Computer System Validation of eSource and EHR Data.
The question that was addressed: How to make a clinical trial data management system that uses EHR data, Patient Reported Outcome (PRO) and eSource data as part of the Learning Health System compliant with regulations and with Good Clinical Practice (GCP)?
The Learning Health System (LHS) connects health care with translational and clinical research. It generates new medical knowledge as a by-product of the care process and its aim is to improve health and safety of patients. The LHS generates and applies knowledge. For this purpose, clinical research, which is research involving humans, must be part of the LHS. Two general types of research exists: observational studies and clinical trials.
Clinical data drive the LHS, because results from randomized controlled trials are seen as “gold standard” for medical evidence. For this reason the concept of using data gathered directly from the patient care environment has enormous potential for accelerating the rate at which useful knowledge is generated.
All computer systems involved in clinical trials must undergo Computer System Validation (CSV). For this process, a legal framework for the TRANSFoRm project was developed. It was used for data privacy analysis of the data flow in two research use cases: an epidemiological cohort study on Diabetes and a randomised clinical trial about different GORD treatment regimes.
Computerized system validation is the documented process to produce evidence that a computerized system does exactly what it is designed to do in a consistent and reproducible manner. The validation of electronic source data in clinical trials presents many challenges because of the blurring of the border between care and research. Here we present our approach for the validation of eSource data capture and the developed documentation for the CSV of the complete data flow in the LHS developed by the TRANSFoRm project. An important part hereby played the GORD Valuation Study.
Use of personalized medicine tools for clinical research networksWolfgang Kuchinke
Patient-centric clinical trials can gain enormously from the employment of personalised medicine tools. Here we address software tools created by the p-medicine network, which developed thr ObTiMA data management system, Patient Empowerment Tool, data mining, data warehousing, biobank access, decision support, image annotation (DrEye) and simulation (Oncosimulator). We evaluated of some of these tools for their suitablity to perform clinical trials. Is their usage conform with regulations and standards (GCP, GDPR, GAMP, computer system validation)? Can these tools be integrated into the existing systems (IT infrastructure / organisational framework) of an international clinical trials network (ECRIN)?
Kuchinke - Learning Health System (LHS) in Europe - Introduction and meeting ...Wolfgang Kuchinke
We present the results of the conference: The Learning Health System (LHS) in Europe. Two perspectives were compared, from the US by Charles Friedman; and from Europe by Federico Paoli. C. Friedman also introduced the concept of learning by virtious cycles. One advanced example for the implementation of the LHS in Europe is the TRANSFoRm project, which was presented in detail. It aims to develop and demonstrate methods, models, standards and a digital infrastructure for three specific components of the LHS: (1) Genotype-phenotype epidemiological studies using multiple existing primary care and ‘biobank’ genomic datasets; (2) Randomized Clinical Trials with both data and trial processes embedded within the functionality of EHRs and (3) Decision support for diagnosis, based on clinical prediction rules. Other topics discussed are: Sustainability and Business Development in the LHS. Knowledge Management & Data Standards in the LHS. Data privacy and security in the LHS. EHR4CR solutions for privacy protection, HIPAA, US experiences with privacy protection,
TRANSFoRm Zone Model for data privacy protection, IMI code of practice for the secondary use of health data, Knowledge Translation and Decision Support in the LHS, Decision support for interventions in the LHS, TRANSFoRm Decision support for diagnosis, TRANSFoRm Extensible model for diagnostic evidence, CareWell - A learning integrated care system
Improving quality of care with routine data: the perspective of a statutory health insurer - The impact on LHS, CDISC and clinical trial standards in the LHS, A pan-European platform for the re-use of EHR data, TRANSFoRm and CDISC standards,
Provenance and GCP for real world clinical trials, Using Health Data for Innovative Trial Designs, etc.
Regulations, privacy, security for data bridges - KuchinkeWolfgang Kuchinke
The presentation is addressing the complexity of data sharing from two view points: (1) Ethical, legal and regulatory challenges, (2) Data sharing between Research Infrastructures.
A comprehensive analysis of the regulatory landscape of data bridges is provided, including: Data Protection Directive, relevant national data protection acts, Good Clinical Practice (GCP), Animal protection laws, security rules for biosamples, data ownership regulations (intellectual ownership, database laws), and others.
Basis was the concepts of Open Data and Open Science, which aims to make scientific research output (publications, data, biosamples, and alogrithms) accessible to all levels of an inquiring including research bey citicen scientists. This includes transparent and accessible knowledge that is shared and developed through collaborative networks and practices such as publishing open research, campaigning for open access and data sharing. But several mechanisms restrict access to data or reuse of data: copyright, patents, database rights, time-limited access rules, political, commercial or legal rules and interests.
Researchers are confronted with the question, whether, on what basis and with what limitations, human data can be used freely and made available to support open research and open science. We conducted an analysis of the legal landscape for data sharing, employing concepts from requirements engineering, like the definition and collection of legal requirements for data bridges, which were based on access rules of many database providers. We defined “legal interoperability” of data sharing as interoperability that forms conditions where a combination of rules allows the exchange of data between different data providers. The basis was the creation of legal “requirement clusters” defining applicable rules, roles and policies used by database owners (data controllers). Such "requirement clusters” can act as a kind of „filter" between different data sources to allow for compliant data transfer. To create "requirement clusters” data sharing usage scenarios were built consisting of real-world examples of interaction between data providers during data sharing. Finally, the legal analysis based on five Usage Scenarios and the development of Requirements Clusters for data protection, data security, intellectual property, security of biosamples and animal protection providing constraints and recommendations for legally sound data bridges and the implementation of “legal filters“ for complinat data flow.
Agile Computer System Validation of software productsWolfgang Kuchinke
Personalized medicine seems to be a potential solution to the known challenges facing clinical research and drug development. To be employed in clinical research, software tools must undergo a process called Computer System Validation (CSV) for compliance with legal requirements and Good Clinical Practice (GCP). Four academic developer groups of the EU project p-medicine were surveyed to evaluate the readiness of their developed software products to be used for clinical research. The analysis of the survey showed that considerable gaps exist in tool maintenance, quality management and compliance documentation. Because all developer groups use agile development methods, recommendations for agile quality assurance were developed as well as for using agile methods to establish “compliance by design”. We show how agile validation can be established by small modifications of sprint processes.
Reverse Engineering of Clinical Trials to Improve ResearchWolfgang Kuchinke
In translational research the question arises how to prepare researchers for the requirements of clinical studies in a way that can facilitate the transition of knowledge from basic, preclinical research with animals to clinical studies with humans. The aim of the presented reverse engineering approach is to familiarize basic researchers with the requirements and characteristics of the clinical study culture and in this way to prepare the basic researcher to create a more efficient translational experimental process. The reverse engineering approach consists, anong other help and guidance, of a preclinical evaluation guide for animal models and requirements for cell and gene therapy products prior to start of clinical study; process descriptions including the stages from preclinical research to phase I transition process to be used for analysis of the number of animals and necessary important experimental data to be generated for cell and gene therapy studies; compilations of the necessary legal provisions and advice received from Paul-Ehrlich Institute in Germany and AFSSAPS in France; list of international requirements for animal models, costs of animals for studies, special requirements for biotherapy studies, applicable guidelines for preclinical studies, demands for new investigational medicinal products. The relevance of the animal model is a focus, especially the consideration that a similar response in human and animal cells in vitro, does not guarantee that the in vivo response is similar.
Introduction to CTIM - the Clinical Trial Information MediatorWolfgang Kuchinke
A new tool was developed: Clinical Trial Information Mediator allows for easy, simoultaneous cross-domain searches through different databases. It is aimed at researchers interested in clinical trials, who want to design new studies and for this purpose need to find trial related information in different biomedical databases. CTIM supports researchers in their searches by linking clinical trials information to publications and information about biosamples / genes and thereby bridging the gap between different databases from different research domain.
The novel feature is that CTIM creates links between clinical trials and publications not through an ID or a key word (code item), but through information content that provides the
basis for queries. CTIM's knowledge base for clinical trials is based on the ClinicalTrials.gov database (CT.gov), the largest repository of clinical trials information. With CTIM one doesn’t have to search in CT.gov and PubMed separately.
Increased Ethical Demands for Patient Empowerment in Personalised MedicineWolfgang Kuchinke
A Patient Empowerment tool was developed by pmedicine project, with the aim to provide help for the patient to understand medical documentation, empower the patient to make informed choices and support patients with their decisions in personalised medicine treatments. To identify and evaluate ethical issues existing, ethical concepts were assigned to the use of the patient empowerment service. Concepts from IT requirements engineering were adapted to collect ethical requirements and five ethical requirements clusters were assigned: Informed Consent, Vulnerable populations, IRB/Ethics Committee, Data privacy, Investigator related ones. The identified areas of ethical concerns cover the ethical correct presentation of medical risks and probabilities, reactions to diagnosis, truth about the patient’s medical conditions, presentation of results of false positives and false negatives, inconsistent and incomplete medical results, understanding of clinical equipoise, confidentiality and data ownership issues, data falsification, proper and justified use of new technologies.
Complex ethical problems are created by combining cancer therapies, biosampling, genetic analysis and personalised medicine method. For example, such combinations create issues of data ownership, right to data deletion, right to forget, using care data for research purposes, suitable risk assessments, ... Based on this analysis we created demands for using the Patient Empowerment Service in a ethical way: enabling patient’s understanding of whole data set that the hospital has collected; patients must be able to understand medical statements, as well as legal and ethical considerations; the empowerment tool must represent data in a comprehensible format, but in case it confronts the patient to get unfavorable information and negative diagnoses it should provide help and guidance.
Ethical concerns caused by integrative patient empowerment servicesWolfgang Kuchinke
We conducted an ethical analysis of a Patient Empowerment Tool. The identified relevant ethical aspects were organized in ethical requirement clusters. These were further analysed for consequences for the structure of the tool and a suitable user interface of the Patient Empowerment Service. A central role that will be played by the Patient Empowerment Service is to guarantee patient autonomy and informed patient’s decisions. A User Interface Model that addresses our ethical concerns will allow user access channeled through a series of adaptable profiles that enable autonomy and present data in an understandable way with build-in support by help and guidance functions. In this way, it becomes possible to ensure patient autonomy even under adverse conditions, like access to unfavoable information, negative diagnoses, incomprehensible risk / benefit display, and other factors that impede the informed decision potential of the patient.
Regulations, privacy and security requirements - Legal interoperability for d...Wolfgang Kuchinke
We are addressing the complexity generated by different ethical, legal and regulatory requirements resulting from data sharing needs between Research Infrastructures. We conducted a comprehensive analysis of the regulatory landscape for creating data bridges: including GDPR (EU Data Protection Directive), relevant national data protection acts, Good Clinical Practice (GCP), animal protection laws, security rules for biosamples, intellectual property regulations. The legal analysis was based on five Usage Scenarios and used for the development of Legal Requirements Clusters for data protection, data security, intellectual property, security of biosamples and animal protection.
We intended to create conditions for legal interoperability for data sharing that may be used for the automation of data exchange processes.
Service Integration for Research Infrastructures by Reciprocal UsageWolfgang Kuchinke
Research infrastructures are building software tools and provide data services. Often these software solutions are to some degree redundant, for example, nearly each research infrastructure has developed its own data management system or data repository. Consequently, developers run into the problem of the sustainability of developed solutions. The academic research community needs better ways to collaborate and reuse and exchange developed software solutions. It is more important to foster the joint use of software tools or the exchange of existing solutions between projects than to develop new tools.
Here we present a concept of mutual exchange of services and a form of deep collaboration. We suggest a concept for the integration and use of software tools by the joint usage of tools and services between different research infrastructures. Including: increased interoperability of clinical data management, biobanking, genetic databases, Electronic Health Records (EHR), query systems, data warehouses and imaging systems. This reciprocal usage creates mutual benefits. Services and tools developed by research infrastructures can be shared and jointly used for mutual benefit.
Data repositories are the core components of an Open Data Ecosystem. To gain a comprehensive model of the data ecosystem supporting tools and services, FAIR principles, joint storage of open data and clinical data and the integration of analysis tools should be considered. The aim was to create a data ecosystem model suitable for the sharing of open data together with sensitive data. For this purpose several tools and services were included in our data ecosystem model: Research Data Marts, I2b2 / tranSMART, CKAN, Dataverse, figshare, OSF (Open Science Framework), ... This multitude of services supports research data repositories. Different types of repositories are connected and supplement each other in the storage, release and sharing of data with different degrees of protection and data ownership. Tools to analyze, browse and visualize data are integrated in the data flow between repositories. Results of our ecosystem analysis:
It doesn‘t matter where one stores data, because everything is connected for data sharing: institutional repositories with dataverses, data marts, general repositories, domain specific repositories, figshare etc. Data governance and privacy protection is integrated at the early stage of data generation.
Zone model for data privacy and confidentiality in medical researchWolfgang Kuchinke
There exist several privacy frameworks for cancer research or biobanking (e.g. ACGT, GenoMatch, caBIG). But most existing privacy frameworks apply the most stringent approach to their data flow and interpret “anonymisation” in a restrictive way. A more flexible approach is needed to guarantee privacy of patient data, but at the same time enable unhindered research. We developed an easy model to display policies and rules for data privacy; it employs the novel concept of "privacy zones for research data flows". The zone model can be used for all important research scenarios.
The Privacy Zone Model is built upon the concept of three zones (Care Zone, Non-care Zone and Research Zone) habouring databases, data transformation operators, such
as data linkers and privacy filters. Using our model, a risk gradient for moving data from a zone of high risk for patient identification to a zone of low risk can be created for each data flow.
CDISC Use Case: Study Migration and Archiving.
Use cases for the employment of CDISC standards for study migration and archiving were collected and discussed. Study archiving covers many issues: legal issues (Good Clinical Practice), reuse of study data, and the replay the study process in case of problems. CDISC ODM format is suitable for clinical study data, because it supports the turning away from proprietary formats and proprietary software solutions in research. In an survey of members of the TMF network, great importance was given to the ability to collect and exchange data together with metadata, and the sharing of complete eCRFs or individual domains of eCRFs. The collected use cases were divided into use cases for study migration, exchange of study data and study archiving, including:
Use case Data import into different EDC systems, Use case Exchange of eCRF, Use case Exchange of data plausibility checks, Use Case Implementation sub-studies, Use case Implementation of mobile data collection by EDC, Use case Archiving compliance with regulations, Use case Archiving of legacy systems, Use case Archiving and study simulation, Use case Archiving of cross-study views.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
PHP Frameworks: I want to break free (IPC Berlin 2024)
Standard based Electronic Archiving for Clinical Trials
1. Workshop about electronic clinical trials, Session I
Standard based
Electronic Archiving
For Clinical eTrials
e-Clinbical Trials, London, 15.05.2009
Wolfgang Kuchinke, University
Duesseldorf, Dueseldorf, Germany
2. Agenda
Identifying the key
issues of electronic
archiving
01
Data and documents02
Analysis of Trial
Master Files
03
The Archiving
Challenge
04
http://www.free-powerpoint-templates-design.com
5. The need for digital preservation
●
Digital storage media are vulnerable to damage or deterioration, and
hardware and software become obsolete, so the time frame for
making preservation decisions is short compared to paper-based
materials.
●
To preserve digital materials the key needs are
• protect the integrity of the data streams
• maintain a way of displaying and using the materials that retains their authenticity.
●
Related needs
• to record enough metadata to manage preservation
• to manage associated rights issues
• to monitor technology for changes that could lead to loss of access
6. Maintain accessibility
●
The key challenge is the future use of materials when technological
changes make original formats and operating environments obsolete
●
Steps that can be taken :
– to retain the hardware and software required to access the materials
– this is only a short-term solution, as hardware will be impossible to maintain indefinitely
●
If possible, produce materials in open, well-supported standard formats
for which software access tools may remain available for an extended
period
●
Transform materials to standard formats. If transforming, retain the
original data in case the chosen migration path becomes unsustainable
7. What is unique in eArchiving
●
eArchiving is essentially identical to traditional archiving with a few
exceptions:
– The concept of “original” and “copy” is irrelevant
– The records can not be inspected by humans without complex
tools
– In addition to retaining media and knowledge on how to interpret
the records, information on digital formats must also be retained
●
Consequences:
– The records are much more vulnerable to loss, since formats have
very short life cycles
8. W. Kuchinke (2009)
Criteria for selecting file formats for long-term preservation
• The following criteria should be considered by data
creators when selecting file formats:
• Open standards
• Ubiquity
• Stability
• Metadata Support
• Feature Set
• Interoperability
• Viability
• The following additional criteria should be considered
for migration:
• Authenticity
• Processability
• Presentation
from: Adrian Brown:
THE NATIONAL
ARCHIVES ,2003t
9. W. Kuchinke (2009)
Open Standards
• Requirement of access to detailed technical information
about the file formats in which records are preserved
• Formats for which the technical specification has been
made available in the public domain are therefore
recommended
• This is invariably the case with open standards, such as
JPEG
• The developers of proprietary formats may also publish
their specifications, either freely (for example, PDF)
• The advantages of some open formats come at the cost
of some loss in structure, context, and functionality
(e.g. ASCII), or the preservation of format at the cost of
processability (e.g. PDF); proprietary formats frequently
support features of their creating software which open
formats do not
10. W. Kuchinke (2009)
Ubiquity
• The format specification should be stable and not
subject to constant or major changes
• New versions of the format should also be backwards
compatible
11. W. Kuchinke (2009)
Stability
Whenever a new
technology comes
up, support for the
older technologies
fades out
• the format specification should be stable and not
subject to constant or major changes
• New versions of the format should also be backwards
compatible
● Open vs. proprietary formats for files and documents
Longevity media
is likely to be
threatened by
obsolescence
before its useful
life is over
12. W. Kuchinke (2009)
Metadata Support
This metadata may
be generated
automatically
• Some file formats make provision for the inclusion of
metadata
• This metadata can have value both during the
subsequent active use of the data and for long-term
preservation (e.g. technical characteristics of the data).
Guidance on the
Provision of
Metadata:
https://www.w3.org/2013/d
wbp/wiki/Guidance_on_the
_Provision_of_Metadata
13. W. Kuchinke (2009)
Feature Set
Example for a full
set of archiving
features: PDF/A
• Formats should support a full range of features and
functionality required for their designated purpose or
business process
• But: the more complex the format, the more costly it
will be to manage and preserve
PDF/A is an ISO-
standardized version
of the Portable
Document Format
(PDF) specialized
especially for the use
in archiving and long-
term preservation of
electronic documents
14. W. Kuchinke (2009)
Interoperability
Includes digital archive
management
standards and IT
solutions, data
exchange in a format
that ensures
interoperability
• Ability to exchange electronic records
• Formats should support a wide range of software or be
platform-independent
• This feature also tends to support the long-term
sustainability of the data by facilitating the migration of
the data from one technical environment to another
Example: AXF. The
archive eXchange
format.Interchange
and Interoperability
for Operational
Storage and Long-
Term Preservation
15. W. Kuchinke (2009)
Viability
Within a vibrant
technical environment,
there is a demand for
long-term viable, open
definition file formats
for archiving
• Some formats provide error-detection facilities, to allow
detection of file corruption which may have occurred
during transmission
• For example, the PNG format incorporates byte
sequences to check for three specific types of error
which could be introduced
Example: hybrid UFF-
DSA format and
CDISC ODM format
16. W. Kuchinke (2009)
Authenticity
Maintaining and
proving authenticity in
the transition from data
and document
producer to the archive
• The format must preserve the content (data and
structure) of the record, and any inherent contextual,
provenance, referencing and fixity information
• ISO 15489: An authentic record is one that can be
proven
• Built-in tests for authenticity
• Known provenance and context
1. Authenticity during
the transfer process
2. Ensuring and
maintaining
authenticity of digital
records preserved in
archive
17. W. Kuchinke (2009)
Presentation
Maintaining and
proving authenticity in
the transition from data
and document
producer to the archive
• The formatting and presentation of a digital record may
have significant information value
• Authenticity of an electronic record may require
preservation of its original ‘look and feel’
• For example, fonts, colours and layout)
Preserving the "look
and feel" of a
document is difficult
for texts, but it is even
more difficult in the
multimedia
environment
18. W. Kuchinke (2009)
Processability
Ensure automatic
processability of the
data, especially during
audits
• Certain data must retain their processability to have
any reuse value
• Even though the requirements of authenticity demand
that the archived version must not be altered through
reprocessing.
19. W. Kuchinke (2009)
Methods of digital preservation
Digital preservation =
formal activity to
ensure that digital
information of
continuing value
remains accessible
and usable
• Technology preservation
• Refreshing
• Migration and reformatting
• Emulation
• Data archaeology
• Output to analogue media
Methods of
preserving data
include refreshing
data and transferring
it to fresh media
20. W. Kuchinke (2009)
Apply traditional archiving processes to eArchive
• Define, identify and document the source organization
and its processes
• Identify information to be archived
• Identify and protect vital records
• Define archive rules and retrieval processes
• Identify ownership and stewardship responsibilities
• Define access rights
• Processes
• Define comprehensive lifecycle
• Define and implement retention schedules
• Requires some processes standardization
For digital archives,
an innovative use of
information
technology is
necessary
21. What is unique in eArchiving
●
eArchiving is essentially identical to traditional archiving with a few
exceptions:
– The concept of “original” and “copy” is irrelevant
– The records can not be inspected by humans without complex
tools
– In addition to retaining media and knowledge on how to interpret
the records, information on digital formats must also be retained
●
Consequences:
– The records are much more vulnerable to loss, since formats have
very short life cycles
22. Archive Life Cycle
Create /
capture
Describe /
Metadata
Use /
manage
Change into
long-term
preservation
DestroyReuse /
audit
Archive
23. W. Kuchinke (2009)
Two paths
• The "Universal Virtual Computer" (UVC) solution
• Preserves the full environment for the future
• Requires understanding of old systems to
understand the data
• Viable for short to medium term
• The "OAIS" solution
• Continuous maintenance of the data to allow
interpretetion with modern systems
• Does not necessarily preserve the full original
environment and context
• Viable for medium to long term
• Supports the UVC solution
ISO 14721:2012
OAIS (Open Archival
Information System)
UVC-based
preservation is an
archival strategy for
handling the
preservation of digital
objects
25. W. Kuchinke (2009)
Data storage
• The storage of any data within research projects should
be based on two principles (preservation policy):
• digital storage media are inherently unreliable unless
they are stored appropriately
• all file formats and physical storage media will
ultimately become obsolete
• Necessary for long-term preservation of digital data
(from: UK Data Archive - UKDA)
Data storage is the
recording (storing) of
information (data) in a
storage medium.
26. W. Kuchinke (2009)
Data and file formats
• Acceptable data formats for preservation
• Wherever possible, data should be stored in
formats which meet long-term readability
requirements
• In general this means non-proprietary formats or
formats based on open standards
• However, some proprietary formats, like Microsoft's
Rich Text Format and Excel format, are widely used
Your data
stewardship practices
is being dictated by
the types of data you
are working with
Recommended data
formats:
For texts: XML,
PDF/A, HTML, Plain
Text
For databases: XML,
CSV, TAB
27. W. Kuchinke (2009)
Storage media
• The accessibility of any data is dependent on the
quality of the storage media on which they are stored
• The availability of the data-reading equipment
• All optical media (CDs and DVDs) are subject to physical
degradation, and best practice suggests that, wherever
possible, data files should be copied from CDs and
DVDs to a new CD or DVD, between two and five years
after they were first created
• Tapes tend to become demagnetized over time, which
can lead to data loss
• Additionally, it is good practice to check, at regular
intervals, the data files on these discs.
• Magnetic media, like hard drives or tapes, are also
subject to physical degradation
• Best practice suggests a similar migration strategy
It refers to a physical
device or component
in a computing
system that receives
and retains
information relating to
applications and
users
Example: DVD-R
Archival Grade
28. W. Kuchinke (2009)
Supporting Technologies and Applications
• COLD (Computer Output to Laser Disk)
• ERM (Enterprise Reports Management
• OCR/ICR (Optical/Intelligent Character Recognition)
• RAID (Redundant Array of Inexpensive Disks)
• WORM (Write Once Read Many storage)
• Bar code recognition
• Voice Recognition
• Automatic Classification Technologies
• Manual Indexing
• Digital Signature
Also important:
Migration strategy
29. W. Kuchinke (2009)
Supporting Techniques
• Digital Records Management
• Hierarchical Storage Technologies
• Data Warehousing
• Media Management
• Media Migration
• Data Migration Schedules
• Automated Format Transformation Tools
• System Utilities
• Secure data destruction
• PDF (Portable Document Format)
• Standardization
Preservation is a set
of activities aimed at
prolonging the life of
a record, book, or
data while making as
few changes as
possible
30. Evaluation of properties of the most important text
based file: txt, tiff, html, xml, open document, doc
(Mircosoft), PDF
The archival properties of
seven common formats
Rating=high (orange), middle
(yellow), low (green)
31. W. Kuchinke (2009)
Physical conditions
• Most researchers do not need to maintain data storage
systems
• However, areas and rooms designated for storage of
digital or non-digital data should be suitable for the
purpose for which they are being used.
• The conditions under which data are stored significantly
affect their longevity.
• Optical media, like CDs and DVDs, are vulnerable to:
poor handling; changes in temperature; changes in
relative humidity; air quality; and lighting conditions.
• Magnetic media, like hard drives, are equally sensitive
to their physical environment
Factors to consider
when using an
archiving room: fire,
water, emission, theft,
pollution
32. Recommended Data Formats for Preservation
Purposes in the Florida Digital Archive
Formats for text files
33. W. Kuchinke (2009)
Archival lifespan of CD and DVD
● Cyanine-based CD-R discs have about 70 years (based on
accelerated aging tests) (TDK)
● Under recommended storage conditions, CD-R, DVD-R, and
DVD+R discs should have a life expectancy of 100 to 200 years
● Best quality CD-ROM media would be suitable for storing
information for 50 until 100 years (National Media Laboratory
NML) under normal conditions
34. W. Kuchinke (2009)
Reasons for good or poor CD-R permanence
● Plastic (polycarbonate substrate) is oxygen permeable,
aluminum corrodes when exposed to oxygen and silver corrodes
or tarnishes when exposed to sulfides in the air. This could
happen in as little as two years with poor CD-R’s.
● The Dyes used in the recording layer are light sensitive and will
react to ambient light and fade over time. Quality CD-R’s use a
dye that resists fading.
● Humidly may seep through a poor lacquer coating.
● A scratch on the base side can be repaired, but a scratch on the
lacquer side renders the CD useless. Quality CD-R’s have a
thick protective coating to resist scratches
35. W. Kuchinke (2009)
Data security
● Data security is the practice of protecting digital information from
unauthorized access, corruption, or theft throughout its entire
lifecycle
● The core elements of data security are confidentiality, integrity,
and availability
● Don't configure archive storage in a way that could potentially
become a single point of failure
● Data security is the protection of any data from unauthorized
access, use, change, disclosure and destruction
Don't treat all of your
data the same with
regards to archival
36. W. Kuchinke (2009)
Confidentiality
● Risks of identification of personal information are typically
contained by the anonymisation of data and the provision of
access through a dedicated rights management framework
● Legally, data which contain personal information must be treated
with more care than data which do not.
● Signed consent forms that usually contain identifying information
should be stored separately from the data
Recognize and
discuss legal issues
relating to privacy
and confidentiality
issues in general and
for digital archives in
particular
37. Recommended Data Formats for Preservation
Purposes in the Florida Digital Archive
Formats for text files
38. W. Kuchinke (2009)
Common causes of archive object loss
● People:
● Theft, damage, corruption
● Natural disasters
● Fire, flood, earthquake
● Erosion
● Rot, loss of meaning
● Mismanagement
● Loss of authenticity, lost in a shuffle
39. W. Kuchinke (2009)
Archiving planing process
Define motivation for archiving
Need for GCP
compliance
Identify objects to archive TMF
Determine how long to keep 10 years
Determine when objects will not be needed End of clinical tudy
For clinical trials
40. W. Kuchinke (2009)
Archiving planing process
Determine how to dispose of objects destruction
Determine who has access to archive Archive representative
Determine form of archive objects ODM file
Determine place of archive Archiving room
42. W. Kuchinke (2009)
Archiving planing process
EDC
data management
system op DB
Remains
in the
system
EDC
data management
system op DB arch DB
EDC
data management
system op DB
EDC
data management
system op DB
archiving
system
archive
room
43. The Trial Master File (TMF) the
central archival object
Analysis of TMFs
44. W. Kuchinke (2009)
What is the TMF
The TMF should be
held at the
coordinating site
(usually the Study
Coordinating Centre.
For multi-site trials,
copies of relevant
documents should be
kept at each
participating site in an
Investigator Site File
(ISF).
• A trial master file contains the essential documents for
a clinical trial that may be subject to regulatory agency
oversight (audits)
• In the European Union (EU), TMFs have a different
definition and set of requirements than in the US
• The EU Commission's Directive 2005/28/EC 63 Chapter
4 states 'the trial master file shall consist of essential
documents, which enable both the conduct of a clinical
trial and the quality of the data produced to be
evaluated
• A Trial Master File (TMF) should be set up at the
beginning of a trial
• The essential documents that make up the file should
be kept in a secure but accessible manner.
45. W. Kuchinke (2009)
Archivierung von Investigator Initiated Trials (ISS)
Investigator initiated
studies (IIS) are
clinical studies initiated
and managed by non-
pharmaceutical
company researchers,
like individual
investigators or
institutions. The
researcher
(investigator) is
responsible for all
legal and regulatory
responsibilities of the
trial sponsor.
• Focus is on IIS
• Implementation of all regulations of GCP can be a
problem for the academic field
• Often no awareness for the difficulty of archiving exists
• Considerable resources for paper based studies are
necessary (e.g. archiving room, archiving
representative, security measures against fire and
burglary)
• University clinics are assuming sponsorship for
academic studies (IIT) (responsibility for archiving)
• Archiving stands for: the safe-keeping also for longer
periods of data and documents and keeping them
accessible, including source data, lab reported, medical
record, site file, identification lists, pseudonymisation
lists, …
46. W. Kuchinke (2009)
Analysis of TMFs
Investigator initiated
studies (IIS) are
clinical studies initiated
and managed by non-
pharmaceutical
company researchers,
like individual
investigators or
institutions. The
researcher
(investigator) is
responsible for all
legal and regulatory
responsibilities of the
trial sponsor.
• An analysis of different TMFs was carried
• The result was that the following types of documents
had often been generated in clinical studies:
• forms, permits, applications, test plans, SOPs
(Standard Operating Procedures), instructions,
labels, receipts, curriculum vitae, calculations, plans,
documentation, correspondence, protocols, reports,
lists, CRFs (Case Report Forms), invoices / receipts,
certificates, contracts, notes / memos, manuals.
• Three representative studies were selected, a small,
medium and large clinical study
• The corresponding TMFs owere examined in detail
• The large and medium-sized studies use electronic data
capture (eCRF), while patient data in the small study
are recorded on paper CRFs
47. W. Kuchinke (2009)
Analysis of TMFs - Categories
Investigator initiated
studies (IIS) are
clinical studies initiated
and managed by non-
pharmaceutical
company researchers,
like individual
investigators or
institutions. The
researcher
(investigator) is
responsible for all
legal and regulatory
responsibilities of the
trial sponsor.
• The variety of documents in a TMF made it necessary to
introduce further classification criteria
• Since paper-based documents are also used and stored,
criteria such as the paper format, number of pages,
type of stapling, etc. has to be considered
• The presence of a signature played an important role
• In the case of electronic documents, the Signature Act
distinguishes between different forms of electronic
signatures
• simple electronic signature
• advanced electronic signature
• qualified electronic signature
• In general, simple signatures can be carried out from
the corresponding EDC (Electronic Data Capture)
programs when filling out the CRFs using a password
48. W. Kuchinke (2009)
Categories of documents in TMF
Document
personal reference documental reference GCP relevance
Personally Pseudonymouse nonee Original Print out Essential Optional
Reference
Study
Organisation
Patient
Originator
Sponsor
Investigator
Ethics
Commission
Pharmacist
Leading
Investigator
statistician
Type of
document
Form
Certificate
Approvals
Investigator's
Brochure
Clinical Study
Protocol
Case Report
Forms (CRF)
Medium
Text
Table
Graphic
Photo
Type of Sig.
Signature
Sig. by
Investigator
Pharmacist
Leading
Investigator
Insurer
Patient
Authority
handwritten digital
Document
creation
Paper digital
49. Hendley (1998) has analysed a
large number of document
categories and the
corresponding formats.
Types of documents in
general
Document Categories (after Hendley)
data (alphanumeric)
structured text (alphanumeric, MarkUp, XML,...)
Office documents (alphanumeric text, graphic)
design documents (raster- and vector graphic)
presentations
pictures / photos
sound
video
interactive multimedia
50. Hendley (1998) has analysed a
large number of document
categories and the
corresponding formats.
Types of documents in
general
Document Categories (after Hendley)
data (alphanumeric)
structured text (alphanumeric, MarkUp, XML,...)
Office documents (alphanumeric text, graphic)
design documents (raster- and vector graphic)
presentations
pictures / photos
sound
video
interactive multimedia
51. The standard IEC 61355-1
Classification and designation
of documents for plants,
systems and equipment
describes rules and
guidelines for the uniform
classification and
identification of documents
based on their characteristic
content of information.
ISO document types Document types (ISO EN 61355)
lists
reports
correspondence
training documents
calculations
contractual documents
order- / delivery documents
data sheets
explanatory documents
guidance / handbooks
diagrams
QM reports
order documents / SOPs
53. W. Kuchinke (2009)
Composition of different TMFs
Considerable
differences were found
in the TMFs of
different clinical trials
• The volume of site-specific documentation is also
strongly influenced by
• the number of patients recruited
• the number of adverse event reports and the
number of queries
• The survey showed that using the 15 categories one
can distinguish about 190 different types of documents
that are kept in a TMF, about half can have a signature
• Many documents are generated by the study
members electronically and then printed
• Their most important formats are: PDF, doc, TIFF
(scanned documents), OpenDoc
• The study database is stored as XML, access, SAS,
CSV, PDF. Documents may contain handwritten
notes, even stickers were used, which impedes
digitalisation.
●
190 different types
of documents
●
about half of them
can have a signature
54. W. Kuchinke (2009)
Composition of different TMFs
• TMF contains: documents which individually and
collectively permit evaluation of the conduct of a
clinical trial and the quality of the data produced are
defined as essential documents according to
CPMP/ICH/135/95
• They are filed in a way that facilitates the management
of the clinical trial, audit and inspection (Trial Master
File) and comprise diverse documents like:
• insurance statements, signed protocol and
amendments, informed consent forms, advertisements
for subject recruitment, signed agreements between
involved parties, approval of ethics committee,
composition of ethics committee, shipping records, CVs
of investigators, normal ranges for laboratory
procedures, etc.
●
ICH E6 (R2) Good
clinical practice. This
document addresses
the good clinical
practice, an
international ethical
and scientific quality
standard. Replaced
Note for Guidance
on Good Clinical
Practice
(CPMP/ICH/135/95)
55. W. Kuchinke (2009)
Composition of different TMFs
• To analyze the documentation of TMFs 15 categories for
classification were developed (person reference, GCP
relevance, document generation, document format,
signature, medium, etc.). TMFs of three clinical trials
were analyzed:
• a small trial (4 sites, 40 patients, 34 pages/CRF,
paper-based CRF),
• a medium-sized trial (7 sites, 40 patients, 65 pages/
CRF, eCRF)
• a big trial (17 sites, 142 patients, 23 pages/CRF,
eCRF).
• Although, the essential study documents are clearly
defined by GCP requirements, depending on the nature
of the clinical study, the different parts of the TMFs can
vary greatly in size. Of the six different parts of TMFs
the site-specific part grows with the number of
participating study centres.
TMFs of three
clinical trials were
analyzed
Different parts of the
TMFs can vary
greatly in size
56. W. Kuchinke (2009)
Composition of different TMFs
• Our analysis showed that there were not so much
differences in the total amount of paper documents
produced by the three exemplary trials:
• 5224 pages for the small trial, 5062 pages for the
medium trial, and 4213 pages for the big trial
• but additional factors which contributed strongly to
the generation of study documents
• In the number of documents of the small trial about
1400 pages of paper CRFs are included, both other
trials used eCRFs
• Another factor is the working habit of investigators and
other participants
• The parts “correspondence” and “drafts, protocols,
other documents” were the smallest in the big study
with 133 pages and 265 pages and the biggest in the
medium study with 452 pages and 1467 pages
A large trial doesn‘t
mean automatically
a large TMF
58. W. Kuchinke (2009)
Result of analysis
• Example studies
• M (7 sites, 40 patients, 65 pages CRF),
• K (4 sites, 40 patients, 34 pages CRF),
• G (17 sites, 142 patients, 23 pages CRF)
• Sum (number of pages), 4213(G), 5224(K), 5062 (M)
• K: contains about 1400 pages of completed CRFs (Case
Report Forms).
• Both other studies have eCRF, in each case burned on
CD
• The number of CRFs can constitute a large part of the
study documentation, because it increases with the
number of patients. One can assume that a CRF amount
to about 40 - 100 pages
The small study K:
contained about
1400 pages of
completed CRFs
59. W. Kuchinke (2009)
Archiving of „Investigator Initiated Trials“
• Implementation of all regulations for GCP is a problem
for the academic field
• Often no awareness for the difficulty of archiving exists
• Considerable resources for paper based studies are
necessary (e.g. archiving room, archiving
representative, security measures against fire and
burglary)
• University clinics are assuming sponsorship for
academic studies (IIT) (responsibility for archiving)
• Archiving stands for: the safe-keeping also for longer
periods of data and documents and keeping them
accessible, including source data, lab report, medical
record, site file, identification lists, pseudonymisation
lists, …
In Investigator-
Initiated Trials (also
called non-
commercial trials),
the leading
investigator usually
establishes only one
file for collecting all
study documents,
the TMF.
60. W. Kuchinke (2009)
Archiving of „Investigator Initiated Trials“
• Implementation of all regulations for GCP is a problem
for the academic field
• Often no awareness for the difficulty of archiving exists
• Documentation is still mainly paper based
• Considerable resources for paper based studies are
necessary (e.g. archiving room, archiving
representative, security measures against fire and
burglary)
In Investigator-
Initiated Trials (also
called non-
commercial trials),
the leading
investigator usually
establishes only one
file for collecting all
study documents,
the TMF.
Archiving stands for: the
safe-keeping also for
longer periods of data
and documents and
keeping them accessible,
including source data, lab
report, medical record,
site file, identification
lists, pseudonymisation
lists, …
61. W. Kuchinke (2009)
Electronic signatures
• Electronic signatures may be created with increasing
levels of security and validity
• Advanced electronic signature
• The signatory can be uniquely identified and linked to the
signature
• The signatory must have sole control of the private key that was
used to create the electronic signature
• Qualified electronic signature
• It is difficult to challenge the authorship of a document signed
with a qualified electronic signature
• The signature is non-repudiable
• Using a digital certificate, which has been encrypted through a
security signature-creating device
• It is authenticated by a qualified trust service provider.
In Investigator-
Initiated Trials (also
called non-
commercial trials),
the leading
investigator usually
establishes only one
file for collecting all
study documents,
the TMF.
62. W. Kuchinke (2009)
Components of a database archiving solution
Important documents
with signatures
Documents
Electronic documents
Archive of legally
important paper
documents with
signature
Archive
repository
digitalisation
eSig
64. W. Kuchinke (2009)
Digital archiving
• Digital archiving is using a repository of digital material
that has to be kept for a longer period of time
• It stores collections of digital information such as
documents, video, pictures, etc in a digital format with
the intention of providing long-term access to the
information
• A digital archive can be an elaborate collection with a
multi-tiered storage system or located on a hard drive.
• Long-term preservation of appropriate materials should
be based on “evidential, legal, contextual or information
value”
• GCP requirements also apply to digital archives
• The main component is the archiving of the clinical trial
database
65. W. Kuchinke (2009)
Components of a database archiving solution
Database archiving
means the removing
of selected data
objects from an
operational
database that are
not referenced any
more and to place
them in an archive
data store, where
they can be
accessed if needed.
Archive application
data gathering
Archive application
design
Archive
data extraction
Archive data
management
Archive access
Archive administration
From: Olson, Database archiving, 2009
Archive
Data store
Archive
repository
Archive repository is
a place in which
records with archival
value are
permanently
preserved.
66. W. Kuchinke (2009)
Components of a database archiving solution
Database archiving
means the removing
of selected data
objects from an
operational
database that are
not referenced any
more and to place
them in an archive
data store, where
they can be
accessed if needed.
Archive application
data gathering
Archive application
design
Archive
data extraction
Archive data
management
Archive access
Archive administration
From: Olson, Database archiving, 2009
Archive
Data store
Archive
repository
Archive repository is
a place in which
records with archival
value are
permanently
preserved.
67. From C. Dollar: 2001 PDA
Good Electronic Records
Management Conference
The three main components of a
document record
Content
• The “Data” (as entered, pre-filled, etc.)
Structure
• The Proxy for Comprehending the Record (form,
format, such as the signature, logo, graphics, etc.)
Context
• circumstances; the “what” or “why” of the transaction
( contract, approval , etc.)
68. From C. Dollar: 2001 PDA
Good Electronic Records
Management Conference
The three main components of a
document record
Content
• The “Data” (as entered, pre-filled, etc.)
Structure
• The Proxy for Comprehending the Record (form,
format, such as the signature, logo, graphics, etc.)
Context
• circumstances; the “what” or “why” of the transaction
( contract, approval , etc.)
69. Lifecycle of data
From: Olson:
Database Archiving,
MK, 2009
discard
operational
phase
active data
reference
phase
semiactive
data
archive
phase
inactive
data
create
70. Data Life Cycle Management
From: Data vs.
Document – What’s
the Difference? C.
Dollar: 2001 PDA
Good Electronic
Records Management
Conference
Retention Metadata
Receipt
Active
Use
Semi-
active/
Creation
DisposalArchival
Storage
Pre-archival
Start Date
Archive
Date
Purge
Date
Period
= n yrs.
71. Integrated Clinical Data Archive
From C. Dollar: 2001
PDA Good Electronic
Records Management
Conference
●
Conceptual Model
Data
Digital
Tracking and preservation
control application
Identification data
Retention data
Preservation data
Location of archived
information set
Audit Trail
Hard
copy
Management Metadata
Archived Clinical Record
Package
Metadata
72. W. Kuchinke (2009)
Electronic Data Records Management
• Regulatory Requirements
• FDA - 21 CFR Part 11: Effective 8/20/97: PART 11
ELECTRONIC RECORDS; ELECTRONIC SIGNATURES
• Electronic documents AND data must be retained where
used as “official records” during the drug development
process, including electronic audit trails.
https://www.accessdat
a.fda.gov/scripts/cdrh/
cfdocs/cfcfr/CFRSearc
h.cfm?CFRPart=11
73. W. Kuchinke (2009)
Persistence of Electronic Clinical Trial Data
• Periodic transfer of data fo new storage media
• Capture and storage of metadata
• Interoperability through technology neutral format(s)
• Likely to be costly
74. W. Kuchinke (2009)
Persistence of Electronic Clinical Trial Data
• Periodic transfer of data fo new storage media
• Capture and storage of metadata
• Interoperability through technology neutral format(s)
• Likely to be costly
75. W. Kuchinke (2009)
Persistence of Electronic Clinical Trial Data
• Periodic transfer of data fo new storage media
• Capture and storage of metadata
• Interoperability through technology neutral format(s)
• Likely to be costly
76. Archive dichotomy - Two types of archiving
documents database
Clinical study
CRF
Study protocol
Contracts
Insurance
Signatur
Scannen
eCRF
e-Study protocol
TMF
77. W. Kuchinke (2009)
Database archiving
• Archiving of data objects
• Data objects: a unit of information that someone in the
future will seek from the archive (represents a business
event or object)
• Creation: data object is created at a specific point in
time
• Operational: data object is used in executing other
business transactions (e.g. data is used to create
other data objects)
• Reference: a data object enters the reference phase
when it has no potential to participate in other
transactions
• Not expected to be referenced: there are no business
processes that will need to use this data
• Data archiving is the practice of identifying data that is
no longer active and moving it out of production
systems into long-term storage systems.
78. W. Kuchinke (2009)
Database archiving
• Need for archiving of data objects
• Data objects is a unit of information that in the future
will be searched for in the archive
• Creation: data object is created at a specific point in
time
• Operational: data object is used in executing other
business transactions
• Reference: a data object enters the reference phase
when it has no potential to participate in other
transactions
• Not expected to be referenced: there are no business
processes that will need to use this data
• Data archiving is the practice of identifying data that is
no longer active and moving it out of production
systems into a long-term storage system
A Data object
represents a
container for data
values, a place
where data values
may be stored and
later retrieved
79. W. Kuchinke (2009)
Database archiving
• Archiving: begins when the data object is not expected
to participate in operational transactions and are not
expected to be referenced
• This is the case when the data object leaves the data
management system after the end of a clinical trial
• Discard: removal of the data object from the electronic
systems
80. W. Kuchinke (2009)
Two forms of storage of clinical trials data
• Archiving of all essential documents in the Trial Master
File (TMF) or the Investigator Site File (ITF)
• Mostly paper
• Archiving of the study database
• Mostly in a proprietary format, or as XML, CSV
• A CSV file typically stores tabular data (numbers and
text) in plain text, in which case each line will have
the same number of fields.
• eCRFs as XML or PDF
81. W. Kuchinke (2009)
Two forms of storage of clinical trials data
• The problem of electronic archiving of clinical studies
joins the spheres of document oriented archival, as can
be found in clinical care (e.g. archiving of electronic
patient records, EPR) with the archiving of research
databases (life science and medical research)
• Unfortunately, from both areas, solutions for the
archiving of clinical trials can be found only to a limited
degree, because in clinical research specific conditions
prevail
• On the one hand only the necessity for a rare access
to the archived trial data due to audits by authorities
exists
• On the other hand there is the need to store a
multitude of different data and documents and to
observe in this context strict laws and regulations
(e.g. AMG, GCP, FDA)
82. W. Kuchinke (2009)
Archiving as part of „Document Life Cycle“
●
Create The making of an analog or digital document
●
Capture Converting or digitizing physical or analog
documents into digital format
●
Index Cataloging process or creating metadata via manual or
automatic methods
●
Manage Managing and storing documents for rapid
accessibility
●
Access Searching processes to find documents using
classification, meta-data, fulltext or other search technologies
●
Retrieve Viewing documents from search results
●
Administer Managing users, resources, content types and
structures
●
Share and Collaborate Sharing, collaborating and modifying
in a work group via real time and non-real time methods
●
Archiving for long-term duration
83. W. Kuchinke (2009)
Access to archived material The archived
material must be
provided for the
cases of audits and
inspections be
authorities
84. W. Kuchinke (2009)
At the end
Ensure automatic
processability of the
data, especially during
audits
• Data storage is different from data archiving
• Data sharing must be handled before the archiving of data
• Preparation of data archiving should begin already at the
data generation stage
• A completely digital TMF should be created, eliminating any
need for paper storage
• For data archiving the most important requirement is that
data is accessible and readable even after very long time
spans of 20 and more years, when the underlying technology
will have completely changed
• Thus, archiving must be done independently from
technological developments
• Qualified electronic signatures may be a way to guarantee
that documents are not corrupted after such long time spans
85. Thank You!
Contact:
Wolfgang Kuchinke, Heinrich-Heine University, Duesseldorf, Germany
wolfgang.kuchinke@uni-duesseldorf.de
wokuchinke@outlook.de
More information about archiving of clinical trials: http://www.tmf-ev.de
This presentation contains
additional explanatory
material for the workshop