Page 3 of 5 FirstFirst 12345 LastLast
Results 21 to 30 of 41

Thread: Online service for fastq and BAM conversion

  1. #21
    Veteran Member
    Join Date
    Jul 2019
    Last Online
    03-11-2024 @ 04:25 PM
    Ethnicity
    Unknown
    Country
    Antarctica
    Gender
    Posts
    3,911
    Thumbs Up
    Received: 3,471
    Given: 1,541

    0 Not allowed!

    Default

    Quote Originally Posted by Tomenable View Post
    Are there, currently, any interesting samples in FASTQ worth converting ???
    These 3 Hungarian studies are available on ENA and not yet converted:
    Maternal lineages from 10-11th century commoner cemeteries of the Carpathian Basin

    Mitogenomic data indicate admixture components of Central-Inner Asian and Srubnaya origin in the conquering Hungarians

    Early medieval genetic data from Ural region evaluated in the light of archaeological evidence of ancient Hungarians

    Polish data
    https://ftp.cngb.org/pub/gigadb/pub/...0310/Raw_data/

  2. #22
    Veteran Member Apricity Funding Member
    "Friend of Apricity"


    Join Date
    Oct 2016
    Last Online
    @
    Ethnicity
    me
    Country
    European Union
    Y-DNA
    R1a > YP1337 > R-BY160486*
    mtDNA
    H3*
    Gender
    Posts
    6,066
    Thumbs Up
    Received: 7,243
    Given: 2,623

    1 Not allowed!

    Default

    30 GB bam was produced for about day. I guess using fastqbam on my computer it would takes a week

  3. #23
    Veteran Member Apricity Funding Member
    "Friend of Apricity"


    Join Date
    Oct 2016
    Last Online
    @
    Ethnicity
    me
    Country
    European Union
    Y-DNA
    R1a > YP1337 > R-BY160486*
    mtDNA
    H3*
    Gender
    Posts
    6,066
    Thumbs Up
    Received: 7,243
    Given: 2,623

    0 Not allowed!

    Default

    Quote Originally Posted by Lucas View Post
    30 GB bam was produced for about day. I guess using fastqbam on my computer it would takes a week
    OMG I attained limit of 250 GB on acount.

  4. #24
    Veteran Member
    Join Date
    Jul 2019
    Last Online
    03-11-2024 @ 04:25 PM
    Ethnicity
    Unknown
    Country
    Antarctica
    Gender
    Posts
    3,911
    Thumbs Up
    Received: 3,471
    Given: 1,541

    0 Not allowed!

    Default

    Quote Originally Posted by Lucas View Post
    OMG I attained limit of 250 GB on acount.
    After which step?
    Did you extract commercial company snps in the second step? The file should be much smaller then.

    Also you should delete datasets from the previous steps you don't need anymore, but it's a bit tricky, you need to use "delete hidden datasets" and "purge deleted datasets" in the history settings after deleting them, or if you have more than 1 history there is also something like "purge deleted histories"
    Last edited by vbnetkhio; 09-12-2021 at 08:19 AM.

  5. #25
    Veteran Member Apricity Funding Member
    "Friend of Apricity"


    Join Date
    Jun 2014
    Last Online
    03-13-2024 @ 06:31 PM
    Location
    Helsinki
    Ethnicity
    Finnish
    Country
    Finland
    Y-DNA
    I1
    mtDNA
    H39
    Politics
    Ugly history as it is. Don't blame me.
    Gender
    Posts
    4,729
    Thumbs Up
    Received: 3,437
    Given: 1,436

    0 Not allowed!

    Default

    I use BWA to align reads.

  6. #26
    Veteran Member Apricity Funding Member
    "Friend of Apricity"


    Join Date
    Jun 2014
    Last Online
    03-13-2024 @ 06:31 PM
    Location
    Helsinki
    Ethnicity
    Finnish
    Country
    Finland
    Y-DNA
    I1
    mtDNA
    H39
    Politics
    Ugly history as it is. Don't blame me.
    Gender
    Posts
    4,729
    Thumbs Up
    Received: 3,437
    Given: 1,436

    0 Not allowed!

    Default

    Quote Originally Posted by Tomenable View Post
    Are there, currently, any interesting samples in FASTQ worth converting ???
    Some studies release only Fastq-files in ENA or in other archives.. You can search by study project names from ENA.

  7. #27
    Veteran Member Apricity Funding Member
    "Friend of Apricity"


    Join Date
    Oct 2016
    Last Online
    @
    Ethnicity
    me
    Country
    European Union
    Y-DNA
    R1a > YP1337 > R-BY160486*
    mtDNA
    H3*
    Gender
    Posts
    6,066
    Thumbs Up
    Received: 7,243
    Given: 2,623

    0 Not allowed!

    Default

    Quote Originally Posted by vbnetkhio View Post
    After which step?
    Did you extract commercial company snps in the second step? The file should be much smaller then.

    Also you should delete datasets from the previous steps you don't need anymore, but it's a bit tricky, you need to use "delete hidden datasets" and "purge deleted datasets" in the history settings after deleting them, or if you have more than 1 history there is also something like "purge deleted histories"
    What you said before was enough and I found it by myself that must use "purge.."

    It was fault of very big fastq, which produced 30GB bam. Then vcf was few times bigger. And I did it two times simultaneously. So I reached maximum of capacity. OK now I'm downbloading those bams and will convert them in WGS as usual. But probably only those older version would work for them.
    Last edited by Lucas; 09-12-2021 at 06:19 PM.

  8. #28
    Member
    Join Date
    Sep 2020
    Last Online
    02-14-2024 @ 06:30 PM
    Ethnicity
    Ukrainian
    Country
    Ukraine
    Gender
    Posts
    171
    Thumbs Up
    Received: 62
    Given: 171

    0 Not allowed!

    Default

    Quote Originally Posted by vbnetkhio View Post
    first upload your file, then in bcftools mpileup select this file under "input BAM/CRAM", and select hg19 under "Select reference genome", and then you can execute the algorythm. When it finishes it will output a BCF file and it will be added to your history.

    then after that run "bcftools call" on that BCF file output from the first step, just select the file and run it, and you'll get another BCF file. in this step you can already choose to output a VCF file but it will probably be too big and take too long to download and convert.

    because of that I also run "bcftools filter" on the second BCF file. select the second bcf file, then upload this file : https://easyupload.io/769eny , then under Restrict To > Regions select "Operate on Regions specified in a history dataset", and select this dataset you uploaded (new.tsv), and for your output type select "uncompressed VCF".

    then you can download this VCF file, and convert it to 23andme with DNAKitStudio.

    for this file, I started with the fastq file, not BAM. in this case there is one extra step at the begining. convert the fastq to BAM with "map with bwa-mem", you also need to select hg19 reference genome, "single" under "Single or Paired-end reads", and then select your fastq file and run it. then you'll have a BAM and you can do the mpileup and the rest.
    1. With the help of "bcftools call" from the BCF file I produce the second BCF file (more compressed). But when I run "bcftools filter" and try to produce VCF - the error occurs:
    [E::bcf_sr_regions_init] Could not parse the file /galaxy-repl/main/files/061/373/dataset_61373762.dat, using the columns 1,2[,-1]
    Failed to read the regions: /galaxy-repl/main/files/061/373/dataset_61373762.dat
    This occurs also if I try to run the first BCF (61 - original BCF).

    2. Also "bcftools call" does not make uncompressed VCF from the second BCF(more compressed), only from the first BCF. The error:
    Note: none of --samples-file, --ploidy or --ploidy-file given, assuming all sites are diploid
    Wrong number of PL fields? nals=1 npl=-3

  9. #29
    Veteran Member
    Join Date
    Jul 2019
    Last Online
    03-11-2024 @ 04:25 PM
    Ethnicity
    Unknown
    Country
    Antarctica
    Gender
    Posts
    3,911
    Thumbs Up
    Received: 3,471
    Given: 1,541

    0 Not allowed!

    Default

    Quote Originally Posted by smd555 View Post
    1. With the help of "bcftools call" from the BCF file I produce the second BCF file (more compressed). But when I run "bcftools filter" and try to produce VCF - the error occurs:

    This occurs also if I try to run the first BCF (61 - original BCF).

    2. Also "bcftools call" does not make uncompressed VCF from the second BCF(more compressed), only from the first BCF. The error:
    There is probably something wrong with your .dat file, could you post the first few lines from it?

    Btw you can skip filtering and just output an uncompressed vcf from bcftools call but it will probably be very big.

  10. #30
    Veteran Member Apricity Funding Member
    "Friend of Apricity"


    Join Date
    Oct 2016
    Last Online
    @
    Ethnicity
    me
    Country
    European Union
    Y-DNA
    R1a > YP1337 > R-BY160486*
    mtDNA
    H3*
    Gender
    Posts
    6,066
    Thumbs Up
    Received: 7,243
    Given: 2,623

    0 Not allowed!

    Default

    Quote Originally Posted by vbnetkhio View Post
    There is probably something wrong with your .dat file, could you post the first few lines from it?

    Btw you can skip filtering and just output an uncompressed vcf from bcftools call but it will probably be very big.
    It is also possible to download vcf and convert using plink. But maybe this kind of vcf needs additional processing before?

Page 3 of 5 FirstFirst 12345 LastLast

Thread Information

Users Browsing this Thread

There are currently 1 users browsing this thread. (0 members and 1 guests)

Similar Threads

  1. Replies: 98
    Last Post: 02-12-2022, 11:21 AM
  2. Replies: 8
    Last Post: 10-24-2018, 04:32 PM
  3. Online daters are racist!!! Whites Prefer Whites When Dating Online
    By Savant in forum Dating and Relationships
    Replies: 97
    Last Post: 08-25-2014, 07:21 PM
  4. Conversion
    By Loddfafner in forum Christianity
    Replies: 32
    Last Post: 06-20-2010, 09:06 PM

Bookmarks

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •