help='number of reads to use for AGS estimation (default = 1e6)')
	parser.add_argument('-l', dest='read_length', type=int,
						help='trim reads to this length (default = median read length)',
						choices=[50, 60, 70, 80, 90, 100, 110, 120, 130, 140, 150, 175, 200, 225, 250, 300, 350, 400, 450, 500])
	parser.add_argument('-f', dest='file_type', type=str,
						help='file type (default = autodetect)',
						choices=['fasta', 'fastq'])
	parser.add_argument('-c', dest='fastq_format', type=str,
						help='quality score encoding (default = autodetect)',
						choices=['fastq-sanger', 'fastq-solexa', 'fastq-illumina'])
	parser.add_argument('-t', dest='threads', type=int, default=1,
						help='number of threads to use for database search (default = 1)')
	parser.add_argument('-q', dest='min_quality', type=int, default=-5,
						help='minimum base-level PHRED quality score (default = -5; no filtering)')
	parser.add_argument('-m', dest='mean_quality', type=int, default=-5,
						help='minimum read-level PHRED quality score (default = -5; no filtering)')
	parser.add_argument('-d', dest='filter_dups', action='store_true', default=False,
						help='filter duplicate reads (default = False)')
	parser.add_argument('-u', dest='max_unknown', type=int, default=100,
						help='max percent of unknown bases (default = 100 percent; no filtering)')
	parser.add_argument('-k', dest='keep_tmp', action='store_true', default=False,
						help='keep temporary files (default = False)')
	parser.add_argument('-s', dest='quiet', action='store_true', default=False,
						help='suppress printing program\'s progress to stdout (default = False)')
	args = vars(parser.parse_args())

	# run pipeline
	est_ags, args = microbe_census.run_pipeline(args)

	# write output
	microbe_census.report_results(args, est_ags)
Example #2
0
 def setUp(self):
     self.expected = 3530599.61
     self.infile = os.path.join(data_dir, 'metagenome.fa.gz')
     self.observed = microbe_census.run_pipeline({'seqfiles':[self.infile]})[0]
						help="file type (default = autodetect)",
						choices=["fasta", "fastq"])
	type.add_argument('-c', dest='fastq_format', type=str,
						help="quality score encoding (default = autodetect)",
						choices=["fastq-sanger", "fastq-solexa", "fastq-illumina"])

	qc = parser.add_argument_group('Quality control (optional)')
	qc.add_argument('-l', dest='read_length', type=int,
						help="all reads trimmed to this length; reads shorter than this discarded (default = median read length)",
						choices=[50, 60, 70, 80, 90, 100, 110, 120, 130, 140, 150, 175, 200, 225, 250, 300, 350, 400, 450, 500])
	qc.add_argument('-q', dest='min_quality', type=int, default=-5,
						help="minimum base-level PHRED quality score (default = -5; no filtering)")
	qc.add_argument('-m', dest='mean_quality', type=int, default=-5,
						help="minimum read-level PHRED quality score (default = -5; no filtering)")
	qc.add_argument('-d', dest='filter_dups', action='store_true', default=False,
						help="filter duplicate reads (default = False)")
	qc.add_argument('-u', dest='max_unknown', type=int, default=100,
						help="max percent of unknown bases per read (default = 100 percent; no filtering)")

	args = vars(parser.parse_args())
	args['seqfiles'] = args['seqfiles'].split(',')

	# run pipeline
	from microbe_census import microbe_census
	est_ags, args = microbe_census.run_pipeline(args)
	if not args['no_equivs']: count_bases = microbe_census.count_bases(args)
	else: count_bases = None

	# write output
	microbe_census.report_results(args, est_ags, count_bases)