N. Ramesh, T. Tasdizen. Semi-supervised learning for cell tracking in microscopy images, In 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), IEEE, April, 2018.
This paper discusses an algorithm for semi-supervised learning to predict cell division and motion in microscopy images. The cells for tracking are detected using extremal region selection and are depicted using a graphical representation. The supervised loss minimizes the error in predictions for the division and move classifiers. The unsupervised loss constrains the incoming links for every detection such that only one of the links is active. Similarly for the outgoing links, we enforce at-most two links to be active. The supervised and un-supervised losses are embedded in a Bayesian framework for probabilistic learning. The classifier predictions are used to model flow variables for every edge in the graph. The cell lineages are solved by formulating it as an energy minimization problem with constraints using integer linear programming. The unsupervised loss adds a significant improvement in the prediction of the division classifier.
In this paper we introduce a novel approach for enhancing the sampling convergence for properties predicted by molecular dynamics. The proposed approach is based upon the construction of a multi-fidelity surrogate model using computational models with different levels of accuracy. While low fidelity models produce result with a lower level of accuracy and computational cost, in this framework they can provide the basis for identification of the optimal sparse sampling pattern for high fidelity models to construct an accurate surrogate model. Such an approach can provide a significant computational saving for the estimation of the quantities of interest for the underlying physical/engineering systems. In the present work, this methodology is demonstrated for molecular dynamics simulations of a Lennard-Jones fluid. Levels of multi-fidelity are defined based upon the integration time step employed in the simulation. The proposed approach is applied to two different canonical problems including (i) single component fluid and (ii) binary glass-forming mixture. The results show about 70% computational saving for the estimation of averaged properties of the systems such as total energy, self diffusion coefficient, radial distribution function and mean squared displacements with a reasonable accuracy.
Informal family caregivers (FCG) are an integral and crucial human component in the cancer care continuum. However, research and interventions to help alleviate documented anxiety and burden on this group is lacking. To address the absence of effective interventions, we developed the electronic Support Network Assessment Program (eSNAP) which aims to automate the capture and visualization of social support, an important target for overall FCG support. This study seeks to describe the preliminary efficacy and outcomes of the eSNAP intervention.
Forty FCGs were enrolled into a longitudinal, two-group randomized design to compare the eSNAP intervention in caregivers of patients with primary brain tumors against controls who did not receive the intervention. Participants were followed for six weeks with questionnaires to assess demographics, caregiver burden, anxiety, depression, and social support. Questionnaires given at baseline (T1) and then 3-weeks (T2), and 6-weeks (T3) post baseline questionnaire.
FCGs reported high caregiver burden and distress at baseline, with burden remaining stable over the course of the study. The intervention group was significantly less depressed, but anxiety remained stable across groups.
With the lessons learned and feedback obtained from FCGs, this study is the first step to developing an effective social support intervention to support FCGs and healthcare providers in improving cancer care.
A. Rodenhauser, W.W. Good, B. Zenger, J. Tate, K. Aras, B. Burton, R.S. Macleod.
PFEIFER: Preprocessing Framework for Electrograms Intermittently Fiducialized from Experimental Recordings, In The Journal of Open Source Software, Vol. 3, No. 21, The Open Journal, pp. 472. Jan, 2018.
Preprocessing Framework for Electrograms Intermittently Fiducialized from Experimental Recordings (PFEIFER) is a MATLAB Graphical User Interface designed to process bioelectric signals acquired from experiments.
PFEIFER was specifically designed to process electrocardiographic recordings from electrodes placed on or around the heart or on the body surface. Specific steps included in PFEIFER allow the user to remove some forms of noise, correct for signal drift, and mark specific instants or intervals in time (fiducialize) within all of the time sampled channels. PFEIFER includes many unique features that allow the user to process electrical signals in a consistent and time efficient manner, with additional options for advanced user configurations and input. PFEIFER is structured as a consolidated framework that provides many standard processing pipelines but also has flexibility to allow the user to customize many of the steps. PFEIFER allows the user to import time aligned cardiac electrical signals, semi-automatically determine fiducial markings from those signals, and perform computational tasks that prepare the signals for subsequent display and analysis.
U. Ruede, K. Willcox, L. C. McInnes, H. De Sterck, G. Biros, H. Bungartz, J. Corones, E. Cramer, J. Crowley, O. Ghattas, M. Gunzburger, M. Hanke, R. Harrison, M. Heroux, J. Hesthaven, P. Jimack, C. Johnson, K. E. Jordan, D. E. Keyes, R. Krause, V. Kumar, S. Mayer, J. Meza, K. M. Mrken, J. T. Oden, L. Petzold, P. Raghavan, S. M. Shontz, A. Trefethen, P. Turner, V. Voevodin, B. Wohlmuth,, C. S. Woodward.
Research and Education in Computational Science and Engineering, In SIAM Review, Vol. 60, No. 3, SIAM, pp. 707--754. Jan, 2018.
This report presents challenges, opportunities and directions for computational science and engineering (CSE) research and education for the next decade. Over the past two decades the field of CSE has penetrated both basic and applied research in academia, industry, and laboratories to advance discovery, optimize systems, support decision-makers, and educate the scientific and engineering workforce. Informed by centuries of theory and experiment, CSE performs computational experiments to answer questions that neither theory nor experiment alone is equipped to answer. CSE provides scientists and engineers with algorithmic inventions and software systems that transcend disciplines and scales. CSE brings the power of parallelism to bear on troves of data. Mathematics-based advanced computing has become a prevalent means of discovery and innovation in essentially all areas of science, engineering, technology, and society; and the CSE community is at the core of this transformation. However, a combination of disruptive developments—including the architectural complexity of extreme-scale computing, the data revolution and increased attention to data-driven discovery, and the specialization required to follow the applications to new frontiers—is redefining the scope and reach of the CSE endeavor. With these many current and expanding opportunities for the CSE field, there is a growing demand for CSE graduates and a need to expand CSE educational offerings. This need includes CSE programs at both the undergraduate and graduate levels, as well as continuing education and professional development programs, exploiting the synergy between computational science and data science. Yet, as institutions consider new and evolving educational programs, it is essential to consider the broader research challenges and opportunities that provide the context for CSE education and workforce development.
A. Sanderson, A. Humphrey, J. Schmidt, R. Sisneros. Coupling the Uintah Framework and the VisIt Toolkit for Parallel In Situ Data Analysis and Visualization and Computational Steering, In High Performance Computing, June, 2018.
Data analysis and visualization are an essential part of the scientific discovery process. As HPC simulations have grown, I/O has become a bottleneck, which has required scientists to turn to in situ tools for simulation data exploration. Incorporating additional data, such as runtime performance data, into the analysis or I/O phases of a workflow is routinely avoided for fear of excaberting performance issues. The paper presents how the Uintah Framework, a suite of HPC libraries and applications for simulating complex chemical and physical reactions, was coupled with VisIt, an interactive analysis and visualization toolkit, to allow scientists to perform parallel in situ visualization of simulation and runtime performance data. An additional benefit of the coupling made it possible to create a "simulation dashboard" that allowed for in situ computational steering and visual debugging.
A. Sanderson, X. Tricoche. Exploration of periodic flow fields, In 18th International Symposium on Flow Visualization, 2018.
One of the difficulties researchers face when exploring flow fields is understanding the respective strengths and limitations of the visualization and analysis techniques that can be applied to their particular problem. We consider in this paper the visualization of doubly periodic flow fields. Specifically, we compare and contrast two traditional visualization techniques, the Poincaré plot and the finite-time Lyapunov exponent (FTLE) plot with a technique recently proposed by the authors, which enhances the Poincaré plot with analytical results that reveal the topology. As is often the case, no single technique achieves a holistic visualization of the flow field that would address all the needs of the analysis. Instead, we show that additional insight can be gained from applying them in combination.
The use of a limited set of signatures in nuclear forensics and nuclear safeguards may reduce the discriminating power for identifying unknown nuclear materials, or for verifying processing at existing facilities. Nuclear proliferomics is a proposed new field of study that advocates for the acquisition of large databases of nuclear material properties from a variety of analytical techniques. As demonstrated on a common uranium trioxide polymorph, α-UO3, in this paper, nuclear proliferomics increases the ability to improve confidence in identifying the processing history of nuclear materials. Specifically, α-UO3 was investigated from the calcination of unwashed uranyl peroxide at 350, 400, 450, 500, and 550 °C in air. Scanning electron microscopy (SEM) images were acquired of the surface morphology, and distinct qualitative differences are presented between unwashed and washed uranyl peroxide, as well as the calcination products from the unwashed uranyl peroxide at the investigated temperatures. Differential scanning calorimetry (DSC), UV–Vis spectrophotometry, powder X-ray diffraction (p-XRD), and thermogravimetric analysis-mass spectrometry (TGA-MS) were used to understand the source of these morphological differences as a function of calcination temperature. Additionally, the SEM images were manually segmented using Morphological Analysis for MAterials (MAMA) software to identify quantifiable differences in morphology for three different surface features present on the unwashed uranyl peroxide calcination products. No single quantifiable signature was sufficient to discern all calcination temperatures with a high degree of confidence; therefore, advanced statistical analysis was performed to allow the combination of a number of quantitative signatures, with their associated uncertainties, to allow for complete discernment by calcination history. Furthermore, machine learning was applied to the acquired SEM images to demonstrate automated discernment with at least 89% accuracy.
B. Summa, N. Faraj, C. Licorish, V. Pascucci.
Flexible Live‐Wire: Image Segmentation with Floating Anchors, In Computer Graphics Forum, Vol. 37, No. 2, Wiley, pp. 321-328. May, 2018.
We introduce Flexible Live‐Wire, a generalization of the Live‐Wire interactive segmentation technique with floating anchors. In our approach, the user input for Live‐Wire is no longer limited to the setting of pixel‐level anchor nodes, but can use more general anchor sets. These sets can be of any dimension, size, or connectedness. The generality of the approach allows the design of a number of user interactions while providing the same functionality as the traditional Live‐Wire. In particular, we experiment with this new flexibility by designing four novel Live‐Wire interactions based on specific primitives: paint, pinch, probable, and pick anchors. These interactions are only a subset of the possibilities enabled by our generalization. Moreover, we discuss the computational aspects of this approach and provide practical solutions to alleviate any additional overhead. Finally, we illustrate our approach and new interactions through several example segmentations.
While convolutional neural networks (CNN) produce state-of-the-art results in many applications including biomedical image analysis, they are not robust to variability in the data that is not well represented by the training set. An important source of variability in biomedical images is the appearance of objects such as contrast and texture due to different imaging settings. We introduce the neighborhood similarity layer (NSL) which can be used in a CNN to improve robustness to changes in the appearance of objects that are not well represented by the training data. The proposed NSL transforms its input feature map at a given pixel by computing its similarity to the surrounding neighborhood. This transformation is spatially varying, hence not a convolution. It is differentiable; therefore, networks including the proposed layer can be trained in an end-to-end manner. We demonstrate the advantages of the NSL for the vasculature segmentation and cell detection problems.
S. Thomas, J. Silvernagel, N. Angel, E. Kholmovski, E. Ghafoori, N. Hu, J. Ashton, D.J. Dosdall, R.S. MacLeod, R. Ranjan.
Higher contact force during radiofrequency ablation leads to a much larger increase in edema as compared to chronic lesion size, In Journal of Cardiovascular Electrophysiology, Wiley, June, 2018.
Reversible edema is a part of any radiofrequency ablation but its relationship with contact force is unknown. The goal of this study was to characterize through histology and MRI, acute and chronic ablation lesions and reversible edema with contact force.
2 Methods and results
In a canine model (n = 14), chronic ventricular lesions were created with a 3.5‐mm tip ThermoCool SmartTouch (Biosense Webster) catheter at 25 W or 40 W for 30 seconds. Repeat ablation was performed after 3 months to create a second set of lesions (acute). Each ablation procedure was followed by in vivo T2‐weighted MRI for edema and late‐gadolinium enhancement (LGE) MRI for lesion characterization. For chronic lesions, the mean scar volumes at 25 W and 40 W were 77.8 ± 34.5 mm3 (n = 24) and 139.1 ± 69.7 mm3 (n = 12), respectively. The volume of chronic lesions increased (25 W: P < 0.001, 40 W: P < 0.001) with greater contact force. For acute lesions, the mean volumes of the lesion were 286.0 ± 129.8 mm3 (n = 19) and 422.1 ± 113.1 mm3 (n = 16) for 25 W and 40 W, respectively (P < 0.001 compared to chronic scar). On T2‐weighted MRI, the acute edema volume was on average 5.6–8.7 times higher than the acute lesion volume and increased with contact force (25 W: P = 0.001, 40 W: P = 0.011).
With increasing contact force, there is a marginal increase in lesion size but accompanied with a significantly larger edema. The reversible edema that is much larger than the chronic lesion volume may explain some of the chronic procedure failures.
Osteoarthritis of the hip can result from mechanical factors, which can be studied using finite element (FE) analysis. FE studies of the hip often assume there is no significant loss of fluid pressurization in the articular cartilage during simulated activities and approximate the material as incompressible and elastic. This study examined the conditions under which interstitial fluid load support remains sustained during physiological motions, as well as the role of the labrum in maintaining fluid load support and the effect of its presence on the solid phase of the surrounding cartilage. We found that dynamic motions of gait and squatting maintained consistent fluid load support between cycles, while static single-leg stance experienced slight fluid depressurization with significant reduction of solid phase stress and strain. Presence of the labrum did not significantly influence fluid load support within the articular cartilage, but prevented deformation at the cartilage edge, leading to lower stress and strain conditions in the cartilage. A morphologically accurate representation of collagen fibril orientation through the thickness of the articular cartilage was not necessary to predict fluid load support. However, comparison with simplified fibril reinforcement underscored the physiological importance. The results of this study demonstrate that an elastic incompressible material approximation is reasonable for modeling a limited number of cyclic motions of gait and squatting without significant loss of accuracy, but is not appropriate for static motions or numerous repeated motions. Additionally, effects seen from removal of the labrum motivate evaluation of labral reattachment strategies in the context of labral repair.
We present a novel approach for improving the shape statistics of medical image objects by generating correspondence of skeletal points. Each object's interior is modeled by an s-rep, i.e., by a sampled, folded, two-sided skeletal sheet with spoke vectors proceeding from the skeletal sheet to the boundary. The skeleton is divided into three parts: the up side, the down side, and the fold curve. The spokes on each part are treated separately and, using spoke interpolation, are shifted along that skeleton in each training sample so as to tighten the probability distribution on those spokes' geometric properties while sampling the object interior regularly. As with the surface/boundary-based correspondence method of Cates et al., entropy is used to measure both the probability distribution tightness and the sampling regularity, here of the spokes' geometric properties. Evaluation on synthetic and real world lateral ventricle and hippocampus data sets demonstrate improvement in the performance of statistics using the resulting probability distributions. This improvement is greater than that achieved by an entropy-based correspondence method on the boundary points.
racing neurons in large-scale microscopy data is crucial to establishing a wiring diagram of the brain, which is needed to understand how neural circuits in the brain process information and generate behavior. Automatic techniques often fail for large and complex datasets, and connectomics researchers may spend weeks or months manually tracing neurons using 2D image stacks. We present a design study of a new virtual reality (VR) system, developed in collaboration with trained neuroanatomists, to trace neurons in microscope scans of the visual cortex of primates. We hypothesize that using consumer-grade VR technology to interact with neurons directly in 3D will help neuroscientists better resolve complex cases and enable them to trace neurons faster and with less physical and mental strain. We discuss both the design process and technical challenges in developing an interactive system to navigate and manipulate terabyte-sized image volumes in VR. Using a number of different datasets, we demonstrate that, compared to widely used commercial software, consumer-grade VR presents a promising alternative for scientists.
As simulations grow in scale, the need for in situ analysis methods to handle the large data produced grows correspondingly. One desirable approach to in situ visualization is in transit visualization. By decoupling the simulation and visualization code, in transit approaches alleviate common difficulties with regard to the scalability of the analysis, ease of integration, usability, and impact on the simulation. We present libIS, a lightweight, flexible library which lowers the bar for using in transit visualization. Our library works on the concept of abstract regions of space containing data, which are transferred from the simulation to the visualization clients upon request, using a client-server model. We also provide a SENSEI analysis adaptor, which allows for transparent deployment of in transit visualization. We demonstrate the flexibility of our approach on batch analysis and interactive visualization use cases on different HPC resources.
F. Wang, W. Li, S. Wang, C.R. Johnson.
Association Rules-Based Multivariate Analysis and Visualization of Spatiotemporal Climate Data, In ISPRS International Journal of Geo-Information, Vol. 7, No. 7, MDPI AG, pp. 266. July, 2018.
Understanding atmospheric phenomena involves analysis of large-scale spatiotemporal multivariate data. The complexity and heterogeneity of such data pose a significant challenge in discovering and understanding the association between multiple climate variables. To tackle this challenge, we present an interactive heuristic visualization system that supports climate scientists and the public in their exploration and analysis of atmospheric phenomena of interest. Three techniques are introduced: (1) web-based spatiotemporal climate data visualization; (2) multiview and multivariate scientific data analysis; and (3) data mining-enabled visual analytics. The Arctic System Reanalysis (ASR) data are used to demonstrate and validate the effectiveness and usefulness of our method through a case study of "The Great Arctic Cyclone of 2012". The results show that different variables have strong associations near the polar cyclone area. This work also provides techniques for identifying multivariate correlation and for better understanding the driving factors of climate phenomena.
Atrial fibrillation (AF) is the most prevalent form of cardiac arrhythmia. Current treatments for AF remain suboptimal due to a lack of understanding of the underlying atrial structures that directly sustain AF. Existing approaches for analyzing atrial structures in 3D, especially from late gadolinium-enhanced (LGE)-MRIs, rely heavily on manual segmentation methods which are extremely labor-intensive and prone to errors. As a result, a robust and automated method for analyzing atrial structures in 3D is of high interest. We have therefore developed AtriaNet, a 16-layer convolutional neural network (CNN), on 154 3D LGE-MRIs with a spatial resolution of 0.625 mm × 0.625 mm × 1.25 mm from patients with AF, to automatically segment the left atrial (LA) epicardium and endocardium. AtriaNet consists of a multi-scaled, dual pathway architecture that captures both the local atrial tissue geometry, and the global positional information of LA using 13 successive convolutions, and 3 further convolutions for merging. By utilizing computationally efficient batch prediction, AtriaNet was able to successfully process each 3D LGE-MRI within one minute. Furthermore, benchmarking experiments showed that AtriaNet outperformed state-of-the-art CNNs, with a DICE score of 0.940 and 0.942 for the LA epicardium and endocardium respectively, and an inter-patient variance of <0.001. The estimated LA diameter and volume computed from the automatic segmentations were accurate to within 1.59 mm and 4.01 cm³ of the ground truths. Our proposed CNN was tested on the largest known dataset for LA segmentation, and to the best of our knowledge, it is the most robust approach that has ever been developed for segmenting LGE-MRIs. The increased accuracy of atrial reconstruction and analysis could potentially improve the understanding and treatment of AF.
Z. Yang, D. Sahasrabudhe, A. Humphrey, M. Berzins. A Preliminary Port and Evaluation of the Uintah AMT Runtime on Sunway TaihuLight, In 9th IEEE International Workshop on Parallel and Distributed Scientific and Engineering Computing (PDSEC 2018), IEEE, May, 2018.
The Sunway TaihuLight is the world's fastest supercomputer at the present time with a low power consumption per flop and a unique set of architectural features. Applications performance depends heavily on being able to adapt codes to make best use of these features. Porting large codes to novel architectures such as Sunway is both time-consuming and expensive, as modifications throughout the code may be needed. One alternative to conventional porting is to consider an approach based upon Asynchronous Many Task (AMT) Runtimes such as the Uintah framework considered here. Uintah structures the problem as a series of tasks that are executed by the runtime via a task scheduler. The central challenge in porting a large AMT runtime like Uintah is thus to consider how to devise an appropriate scheduler and how to write tasks to take advantage of a particular architecture. It will be shown how an asynchronous Sunway-specific scheduler, based upon MPI and athreads, may be written and how individual taskcode for a typical but model structured-grid fluid-flow problem needs to be refactored. Preliminary experiments show that it is possible to obtain a strong-scaling efficiency ranging from 31.7% to 96.1% for different problem sizes with full optimizations. The asynchronous scheduler developed here improves the overall performance over a synchronous alternative by at most 22.8%, and the fluid-flow simulation reaches 1.17% the theoretical peak of the running nodes. Conclusions are drawn for the porting of full-scale Uintah applications.
Spectral Element and hp Methods, In Encyclopedia of Computational Mechanics Second Edition, John Wiley & Sons, Ltd, pp. 1--43. 2018.Y. Yu, R.M. Kirby, G.E. Karniadakis.
Spectral/hp element methods provide high‐order discretization, which is essential in the longtime integration of advection–diffusion systems and for capturing dynamic instabilities in solids. In this chapter, we review the main formulations for simulations of incompressible and compressible viscous flows as well as for solid mechanics and present several examples with some emphasis on fluid–structure interactions and interfaces. The first generation of (nodal) spectral elements was limited to relatively simple geometries and smooth solutions. However, the new generation of spectral/hp elements, consisting of both nodal and modal forms, can handle very complex geometries using unstructured grids and can capture strong shocks by employing discontinuous Galerkin methods. New flexible formulations allow simulations of multiphysics problems including extremely complex geometries and multiphase flows. Several implementation strategies have also been developed on the basis of multilevel parallel algorithms that allow dynamic p ‐refinement at constant wall clock time. After three decades of intense developments, spectral element and hp methods are mature and efficient to be used effectively in applications of industrial complexity. They provide the capabilities that standard finite element and finite volume methods do, but, in addition, they exhibit high‐order accuracy and error control.
Y.Y. Yu, S.Y. Elhabian, R.T. Whitaker. Clustering With Pairwise Relationships: A Generative Approach, In CoRR, 2018.
Semi-supervised learning (SSL) has become important in current data analysis applications, where the amount of unlabeled data is growing exponentially and user input remains limited by logistics and expense. Constrained clustering, as a subclass of SSL, makes use of user input in the form of relationships between data points (e.g., pairs of data points belonging to the same class or different classes) and can remarkably improve the performance of unsupervised clustering in order to reflect user-defined knowledge of the relationships between particular data points. Existing algorithms incorporate such user input, heuristically, as either hard constraints or soft penalties, which are separate from any generative or statistical aspect of the clustering model; this results in formulations that are suboptimal and not sufficiently general. In this paper, we propose a principled, generative approach to probabilistically model, without ad hoc penalties, the joint distribution given by user-defined pairwise relations. The proposed model accounts for general underlying distributions without assuming a specific form and relies on expectation-maximization for model fitting. For distributions in a standard form, the proposed approach results in a closed-form solution for updated parameters.