InferLoop: Leveraging Single-Cell Chromatin Accessibility for the Signal of Chromatin Loop
Briefings in Bioinformatics(2023)
Shanghai Jiao Tong Univ Sch Med
Abstract
Deciphering cell-type-specific 3D structures of chromatin is challenging. Here, we present InferLoop, a novel method for inferring the strength of chromatin interaction using single-cell chromatin accessibility data. The workflow of InferLoop is, first, to conduct signal enhancement by grouping nearby cells into bins, and then, for each bin, leverage accessibility signals for loop signals using a newly constructed metric that is similar to the perturbation of the Pearson correlation coefficient. In this study, we have described three application scenarios of InferLoop, including the inference of cell-type-specific loop signals, the prediction of gene expression levels and the interpretation of intergenic loci. The effectiveness and superiority of InferLoop over other methods in those three scenarios are rigorously validated by using the single-cell 3D genome structure data of human brain cortex and human blood, the single-cell multi-omics data of human blood and mouse brain cortex, and the intergenic loci in the GWAS Catalog database as well as the GTEx database, respectively. In addition, InferLoop can be applied to predict loop signals of individual spots using the spatial chromatin accessibility data of mouse embryo. InferLoop is available at https://github.com/jumphone/inferloop.
MoreTranslated text
Key words
chromatin structure,single-cell chromatin accessibility,single-cell 3D genome structure,single-cell multi-omics,spatial chromatin accessibility
求助PDF
上传PDF
View via Publisher
AI Read Science
AI Summary
AI Summary is the key point extracted automatically understanding the full text of the paper, including the background, methods, results, conclusions, icons and other key content, so that you can get the outline of the paper at a glance.
Example
Background
Key content
Introduction
Methods
Results
Related work
Fund
Key content
- Pretraining has recently greatly promoted the development of natural language processing (NLP)
- We show that M6 outperforms the baselines in multimodal downstream tasks, and the large M6 with 10 parameters can reach a better performance
- We propose a method called M6 that is able to process information of multiple modalities and perform both single-modal and cross-modal understanding and generation
- The model is scaled to large model with 10 billion parameters with sophisticated deployment, and the 10 -parameter M6-large is the largest pretrained model in Chinese
- Experimental results show that our proposed M6 outperforms the baseline in a number of downstream tasks concerning both single modality and multiple modalities We will continue the pretraining of extremely large models by increasing data to explore the limit of its performance
Upload PDF to Generate Summary
Must-Reading Tree
Example

Generate MRT to find the research sequence of this paper
Related Papers
2009
被引用306 | 浏览
2013
被引用1125 | 浏览
2015
被引用39 | 浏览
2015
被引用1695 | 浏览
2015
被引用88 | 浏览
2016
被引用8186 | 浏览
2017
被引用890 | 浏览
2018
被引用4242 | 浏览
2019
被引用169 | 浏览
2019
被引用603 | 浏览
2019
被引用194 | 浏览
2019
被引用518 | 浏览
2021
被引用129 | 浏览
2021
被引用71 | 浏览
2022
被引用1 | 浏览
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper