In this 2019 paper published in Journal of Medical Biochemistry, Lippi and Mattiuzi examine the primary drivers and outcomes of implementing project management strategies in the field of laboratory medicine. Citing funding and personnel shortages in healthcare services, as well as an increasing volume and complexity of analytical tests in the lab, the authors point out the importance of proper project management in the various aspects of the clinical laboratory and its organization and operation in order to better ensure more rapid and accurate diagnoses and improved clinical outcomes. They lay out four primary steps and address additional drivers to consider when implementing project management into laboratory planning and operations, including defining the laboratory environment, planning technical resources, managing staff, and interacting with hospital administrators and other important stakeholders. They conclude that through these activities, as well as some final last steps, clinical laboratories—and by extension the patients they serve—can benefit greatly from incorporating thoughtful project planning into their activities.
Information technology and medical technology personnel’s perception regarding segmentation of medical devices: A focus group study
Cybersecurity is an increasingly relevant and critical topic for most industries, including the healthcare industry. In hospitals and other clinical settings, patient monitoring systems, medical imaging systems, and laboratory instruments are increasingly networked to get the most out of data management and operational efficiency. However, these devices are at risk of cyber attack, with potentially significant consequences. Johansson et al. acknowledge this problem and attempt to address the perceptions of information and medical technology experts on the topic of securing in-house medical devices through the practice of network segmentation. The authors present their discussion results with 18 stakeholders from the Swedish Region Skåne, and they conclude that the stakeholders “were positively receptive to the increase in cybersecurity provided by network segmentation but concerned about the increase in the administration that it will entail for medical devices.” They note that these and other opinions stated in the research are vital for health networks to take into account when considering improving their cybersecurity.
Ontologies are important to biological and biomedical researchers working with multiple data sources because, when implemented well, ontologies ensure heterogeneous, multimodal data across multiple domains and they aid with more effective data mining and discoveries. Ontologies also assist in improving data quality. In this 2019 paper, Jackson et al. highlight the importance of ontologies with their ROBOT open-source library, which aids with the automation of ontology development. They describe the various means which ROBOT “provides ontology processing commands for a variety of tasks, including commands for converting formats, running a reasoner, creating import modules, running reports, and various other tasks.” They conclude that such automation allows developers to easily “configure, combine, and execute individual tasks in comprehensive, automated workflows.” The authors also demonstrate how ROBOT can help catch logical errors and provide quality control mechanisms to ontology creation and updating.
In this brief article by Schmieder et al., the authors provide an example of applying laboratory informatics to research activities for drug discovery and personalized medicine. In particular, the authors look at the in vitro nephrotoxicity assay and the use of transepithelial electrical resistance (TEER) to evaluate the barrier function of cellular layers. Citing issues such as the significant multi-step processes involved with this type of testing, the authors attempt to demonstrate how a laboratory information management system (LIMS), laboratory execution system (LES), and automation standards such as SiLA can be used together to decrease the workflow complexity of such cell-based assays.
Researchers in a wide variety of fields are gradually realizing the benefits of open-source hardware (OSH) in their scientific activities, though not easily. Projects associated with OSH often fail for lack of technological know-how, long-term funding, and strong organization. In this 2019 article, Hill et al. suggest the benefits of “a provisional framework for developing and sustaining the life cycle of not‐for‐profit OSH,” in this case applying it to their biological and conservation science. The authors describe their six-phase framework spread across two management teams, one for engineering and one for logistics (with both sharing social media and outreach duties). They present as a case study the AudioMoth project for the development of a low-cost, full-spectrum acoustic logger, which was largely crowd-sourced. They conclude by emphasizing the importance of frameworks such as theirs, stating that without them “existing OSH will continue to have short life spans, and remain out of reach for the majority of conservation biologists.” More broadly, they also note the need “to push access to OSH toward communities outside of the pockets of wealth and high opportunity that the framework may initially serve.”
In this 2018 article published in The Journal of Supercomputing, Ibrahim et al. present the results of their application of the National Institute of Standards and Technology’s (NIST) Cybersecurity Framework to a custom assessment tool for gauging the cybersecurity preparedness of an organization. In this paper, the authors used their assessment tool to assess a Western Australian government. After discussing the formula involved with the assessment tool and visualization of results using Microsoft Power BI, the authors conclude that using their tool and the NIST Cybersecurity Framework, organizations can better “identify the specific people, processes, and technology areas that require improvement” and improve threat mitigation. The authors also cite the user-friendly nature of the framework.
In this early 2019 journal article published in the journal Database, Grand et al. of the Candiolo Cancer Institute present the fine details of their laboratory information management system (LIMS) Laboratory Assistant Suite (LAS) for cancer and other genomic research. Citing “a substantial mismatch between the LIMS solutions on offer and the functional requirements dictated by research practice,” the authors describe the requirements they had for a LIMS in their institution and how they went about creating it. After describing the data models, functionalities, modular architecture, and its usage, the authors conclude that their LAS, in conjunction with a custom data management module, allows researchers to “execute complex queries without any knowledge of query languages or database structures, and easily integrate heterogeneous data stored in multiple databases,” while also resulting in an improvement in data quality, a reduction in data entry and retrieval, and new insights with the enabled data interconnections.
While the definitions surrounding “open-source” software have largely sufficed for such software, the term as applied to tangible hardware products has been insufficient, according to arguments by Bonvoisin et al. in this 2017 paper published in the Journal of Open Hardware. Their work turns to analysis of 132 proclaimed open-source ” non-electronic and complex open-source hardware products.” After lengthy background information and discussion of their methods and results, the authors conclude: “The empirical results strongly indicate the existence of two main usages of open-source principles in the context of tangible products: publication of product-related documentation as a means to support community-based product development and to disseminate privately developed innovations. It also underlines the high variety of interpretations and even misuses of the concept of open-source hardware.”
The topic of making bioinformatics applications more approachable to researchers and students has been discussed off and on for years, and some efforts have even been made in that regard. Another step forward for bioinformatics applications is offered by Joppich and Zimmer of Ludwig-Maximilians-Universität München, with their open-source bioGUI. The software attempts to address two problems of bioinformatics applications that rely heavily on the command line: many of them work on Unix-based systems but not Microsoft Windows, and researchers have a tendency to shy away from complex command-line apps despite their utility. The authors present in detail their framework and its use cases, showing how a graphical user interface or GUI can make many such command-line apps more approachable. They conclude that providing a GUI and easy-to-use install modules for bioinformatics apps using the command line makes “execution and usage of these tools more comfortable” while allowing scientists to better analyze their data.
ChromaWizard: An open-source image analysis software for multicolor fluorescence in situ hybridization analysis
In this brief article published in Cytometry Part A, researchers at the Austrian Centre for Industrial Biotechnology present their open-source multiplex fluorescence in situ hybridization (M‐FISH) software for chromosome painting. The tool—ChromaWizard—acts as a free and open-source option for hybridization analysis, integrating “image processing, multicolor integration, chromosome separation, and visualization with false color assignments.” The software can handle images in TIFF, PNG, and JPEG formats and provides robust visualization tools. The authors conclude that ChromaWizard “allows direct inspection of the original hybridization signals and enables either manual or automatic assignment of colors, making it a functional and versatile tool that can also be used for other multicolor applications.”
Open-source software has been a topic of discussion for decades, both as a model of software development and distribution and as a potential for what could be done with it. But the concept of open-source hardware, particularly in the field of science, has been more challenging to address. In this brief perspective article by University of Tübingen’s André Maia Chagas, he discusses the benefits of open science hardware to address the growing concern of “haves and have nots” in the scientific research community. Citing high prices, the overall closed-source nature of equipment, businesses that can close without warning, and poor customer support, Chagas attempts to demonstrate that advances in modern design—such as the smartphone—and organizational efforts to implement and promote open hardware philosophies provide opportunity for more people to engage in scientific endeavors. He concludes we “need to reassess our relationship to knowledge and technology, how it determines our role in society, and how we want to spend grant money entrusted to us by the people,” and by focusing on making hardware more open and accessible, we’ll shrink the divide and improve scientific research as a result.
In this brief article published in BMC Bioinformatics, Wang and LaFramboise address the topic of cytogenetic data and their genomic coordinates, which the authors describe as “precisely [specifying] a chromosomal location according to its distance from the end of the chromosome.” The authors note that despite changes in techniques over the years, the use of karyotype and cytogenetic nomenclature was the primary way of characterizing aberrations in chromosomes, and those methods are still being used today. Additionally, archival data used karyotypes and cytogenetic nomenclature. Given the lack of a maintained, robust tool for converting that nomenclature to genomic coordinates, the authors address their creation, CytoConverter, and explain how it accomplishes such conversions. They conclude that the tool should have “considerable value to the community for analyzing archival patient samples, as well as samples for which higher-resolution copy number data is unavailable.”
Implementing a novel quality improvement-based approach to data quality monitoring and enhancement in a multipurpose clinical registry
The clinical registry is an important tool in assisting those affected by rare diseases. Using observational study methods, the registry collects and assists with analysis of data related to a particular disease or condition, helping to serve researchers and their policies towards helping the afflicted. While studies have been performed on the “fit for use” nature of registry data, data quality studies of registries using specific quality improvement approaches have been fewer. Pratt et al. give their take on a quality improvement scheme for registries in this 2019 paper, tapping into the significant data generated by the ImproveCareNow (ICN) Network and its goal of improving outcomes of pediatric and adolescent inflammatory bowel diseases. After going through their methodology, the authors conclude that data quality improvement campaigns, particularly those that include support for training and tools, can improve “the completeness and consistency” of registry data, and by extension resulting “in a higher level of confidence when accessing the data for various purposes, including clinical decision making.”
In this 2019 article published in Molecules, Mandrioli et al. detail the process they developed for detecting 10 cannabinoids in Cannabis sativa L. using reversed-phase high-performance liquid chromatography with an ultraviolet detector (RP-HPLC-UV). Citing numerous other published methods, many of them requiring more expensive mass spectrometry detectors, the authors sought to “identify and titrate cannabinoids in a simple way” using a method that optimally would be “fast, easy, robust, and cost-efficient,” making the process more approachable by not only research laboratories but also small businesses focused on quality control. Using their method, the authors conclude that the method can produce results in eight minutes, with high sensitivity and simple methodology.
We use our mobile phones daily, and many of don’t give consideration to whether or not those devices are tracking or monitoring our activities. At the root of this cybersecurity issue is, most frequently, the permissions given to one or more applications on the device to access one or more sensors contained in the device. The lackadaisical attitude of the average user towards the cybersecurity of their mobile device can be attributed to a variety of aspects, including poor education regarding smartphone use, low public awareness, and ignorance due to developers’ stealthy or “permission hungry” methodologies. Mehrnezhad and Toreini discuss these issues and more at length in this 2019 paper published in Informatics, concluding that while “teaching about general aspects of sensors might not immediately improve people’s ability to perceive the risks,” over time users may “successfully identify over-privileged apps” and make more informed decisions about “modifying the app permissions, uninstalling, or keeping it as-is.”
In this 2019 review paper published in Frontiers in Oncology, Bhattacharya et al. describe the state of collaborative, artificial-intelligence-based computational cancer research within various agencies and departments of the United States. The researchers point to three major initiatives that aim “to push the frontiers of computing technologies in specific areas of cancer research” at the cellular, molecular, and population levels. They present details concerning the three initiatives, enacted as pilot programs with specific goals: Pilot One “to develop predictive capabilities of drug response in pre-clinical models of cancer,” Pilot Two “on delivering a validated multiscale model of Ras biology on a cell membrane,” and Pilot Three “to leverage high-performance computing and artificial intelligence to meet the emerging needs of cancer surveillance.” Additionally, emerging opportunities and challenges that continue to arise out of these pilots are also addressed, before concluding that “opportunities for extreme-scale computing in AI and cancer research extend well beyond these pilots.”
Genomic and sequencing data are inherently complex and have significant storage requirements. They require a robust infrastructure with well-considered policies to make the most of their potential. While North America and Europe have helped lead the way in this goal, Africa is behind them in the adoption of genomic technologies. Parker et al., of the Human Hereditary and Health in Africa (H3Africa) program, have taken on the challenge of provisioning and managing the infrastructure required to meet the goals of various Africa-based genetic research projects. This paper describes the H3Africa Data Archive, “the first formalized human genomic data archive on the continent.” The authors discuss their process and findings, also noting various challenges that presented during the implementation process, as well as recognizing the various benefits from such a project.
Process variation detection using missing data in a multihospital community practice anatomic pathology laboratory
In this brief journal article by Ochsner Health System’s Gretchen Galliano, a case is made for a programmatic approach to analyzing missing data in various laboratory information systems (LIS) and determining potential correlations with procedural and systemic processes in the health system. Using the R programming language, homemade scripts, and existing R packages, Ochsner Health Systems visualized and analyzed data from more than 70,000 cases of missing timestamp data, splitting the various cases into five pools. Galliano concludes that the process of “evaluating cases with missing predefined process timestamps” has the potential for improving “the ability to detect other data variations and procedure noncompliance in the AP workflow in a prospective fashion.” She added that as an additional benefit, “[p]eriodically evaluating data patterns can give AP LIS teams and operations teams insight into user–LIS interactions and may help identify areas that need focus or updating.”
Development and validation of a fast gas chromatography–mass spectrometry method for the determination of cannabinoids in Cannabis sativa L
In this 2018 paper published in the Journal of Food and Drug Analysis, Cardenia et al. discuss their development of a “routine method for determining cannabinoids” in the flowers of Cannabis sativa L. using fast gas chromatography coupled to mass spectrometry (fast GC-MS), with appropriate derivatization approaches that take into account potential decarboxylation. The authors discuss the various problems with other methods then present their materials and methods. After considerable discussion of their results, they conclude that the procedure is fast (within seven minutes), has good resolution (R > 1.1), and remains cost-effective. Sensitivity was also high, with “a high repeatability and robustness in both cannabinoid standard mixtures and hemp inflorescence samples.”
Clinical data research networks (CDRNs)—consisting of a variety of health care delivery organizations that share deidentified clinical data for clinical research purposes—constitute yet another collaboratory mechanism for scientific researchers to pool data and make new discoveries. However, one of the faults of CDRN data is that it typically comes from electronic health records (EHRs), which contain data with a lean more towards supporting “clinical operations rather than clinical research.” This means data quality is of the utmost importance when pooling and putting to effective use such disparate data sources. In this research, Khare et al.propose a systematic workflow for making quality assessments of CRDN’s data before use, a workflow that includes hundreds of systematic data checks and a GitHub-based reporting system to track and correct issues in a more timely fashion. They conclude that their publicly available toolkit definitively has value, though implementers should be advised that “sufficient resources should be dedicated for investigating problems and optimizing data” due to the time-intensive nature of the entire process.