Question: De Novo Transcriptome Assembly of multiple tissues: To pool or not to pool, that's the question?!
gravatar for giorgiocasaburi
3.1 years ago by
United States
giorgiocasaburi90 wrote:

Hi all,

I have a dilemma and in the current literature there are controversial opinions. The nature protocol for Trinity also is not really clear about how to treat samples derived from different tissues, although it does suggest to pool together (prior to assembly) biological and technical replicates.  I would like to hear a general opinion based on experience when it comes to assemble reads from different tissues.

A) Do you pool all the reads together and then run Trinity?

B) Do you run Trinity on every distinct data set (tissue) and then merge together the outputs with another assembler (e.g. cap3)?

I think it would be interesting to know what people thinks about this and what their general results.

Thanks in advance,


ADD COMMENTlink modified 2.8 years ago by • written 3.1 years ago by giorgiocasaburi90

I am interested in what others have to say. I too have pooled reads from different tissues and assembled with velvet/oases. I then did FPKM predictions with cufflinks using reads from different tissues separately on the pooled assembly.

ADD REPLYlink written 3.1 years ago by apelin20470
gravatar for Ginsea Chen
2.8 years ago by
Ginsea Chen110
Chinese Academy of Tropical Agricultural Sciences, Danzhou, China
Ginsea Chen110 wrote:


Two methods you said (A and B in your question) are practicable. If you have a large dataset, I suggest you to assembly them firstly for each tissue and then merged them together for timing saving. In Trinity pipeline, the auto-run scripts told us that they merged all reads firstly and then run the software. You can read it in trinity website.

ADD COMMENTlink modified 2.8 years ago • written 2.8 years ago by Ginsea Chen110
gravatar for
2.8 years ago by
Philadelphia, PA wrote:

If the end goal is to perform a DEG analysis on the samples, then Trinity says to combine samples, assemble, and then one can perform DEG analysis. Otherwise you will have a difficult time creating the abundance tables with RSEM and analysing with edgeR. When you run the wrapper scripts for edgeR you can specifiy the sample conditions/tissues in your 'samples_described.txt' while specifying replicates.

I think you will get the best results merging your samples. If you have a lot of samples or really deep coverage, you can use the 'in silico read normalization' parameter to normalize your reads, which will help in processing time.

If you haven't done any QC on your samples, Trinity also supports Trimmomatic.


ADD COMMENTlink modified 2.8 years ago • written 2.8 years ago by
Please log in to add an answer.


Use of this site constitutes acceptance of our User Agreement and Privacy Policy.
Powered by Biostar version 2.3.0
Traffic: 864 users visited in the last hour