Question: How to determine the total average insert size from differently divided from one fastq data pool to each reference?
0
gravatar for bioinfo
14 months ago by
bioinfo10
bioinfo10 wrote:

Hi, I have difficulty in calculation for average insert size and standard deviation of RNA-seq NGS data to submit GEO database. I generated two fastq files of paired end read(Read1, Read2) using illumina NGS sequencer. And, I used bacteria genome as reference which has 3 chromosomes. For analysis, I used bowtie2 pipeline to align paired end data to each chromosomes. To get average insert size, such as 'samtools stats' is good using alignment. But I think using generated each sam file ,which is divided from fastq data, for calculation is not accurate. Is there any way to get the average insert size and standard deviation from total fastq? thank you in advance.

rna-seq size calculation • 1.1k views
ADD COMMENTlink modified 14 months ago by genomax73k • written 14 months ago by bioinfo10
0
gravatar for genomax
14 months ago by
genomax73k
United States
genomax73k wrote:

Use instructions in this post using BBMap suite: C: Target fragment size versus final insert size

There are two methods to do this. Either by alignment to a reference or by merging (if R1/R2 reads show sequence overlap).

ADD COMMENTlink modified 14 months ago • written 14 months ago by genomax73k

Oh, thanks a lot. But my fastq files are not overlap between reads because read length is shorter than inner distance. So in my case, what is the best option for bbmap? Is it right to calculate after merging the three chromosomses fasta file?

ADD REPLYlink modified 14 months ago • written 14 months ago by bioinfo10

You should ideally do it using the original fastq data if you have it. But if not, fasta should work (I think).

ADD REPLYlink written 14 months ago by genomax73k
Please log in to add an answer.

Help
Access

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.
Powered by Biostar version 2.3.0
Traffic: 3214 users visited in the last hour