Abstract
Manipulating the three-dimensional (3D) structures of cells is important for facilitating to repair or regenerate tissues. A self-assembly system of cells with cellulose nanofibers (CNFs) and concentrated polymer brushes (CPBs) has been developed to fabricate various cell 3D structures. To further generate tissues at an implantable level, it is necessary to carry out a large number of experiments using different cell culture conditions and material properties; however this is practically intractable. To address this issue, we present a graph neural network-based simulator (GNS) that can be trained by using assembly process images to predict the assembly status of future time steps. A total of 24 (25 steps) time-series images were recorded (four repeats for each of six different conditions), and each image was transformed into a graph by regarding the cells as nodes and the connecting neighboring cells as edges. Using the obtained data, the performances of the GNS were examined under three scenarios (i.e. changing a pair of the training and testing data) to verify the possibility of using the GNS as a predictor for further time steps. It was confirmed that the GNS could reasonably reproduce the assembly process, even under the toughest scenario, in which the experimental conditions differed between the training and testing data. Practically, this means that the GNS trained by the first 24 hour images could predict the cell types obtained three weeks later. This result could reduce the number of experiments required to find the optimal conditions for generating cells with desired 3D structures. Ultimately, our approach could accelerate progress in regenerative medicine.
Supplementary materials
Title
Supporting Information: Graph Networks Simulate Multicellular Dynamics Driven by Concentrated Polymer Brush-Modified Cellulose Nanofibers
Description
Detailed (wet-lab) experimental procedure and results
Actions