TARGET EcoTox Challenge Technology Advancing Rapid Gene Expression based Testing (TARGET) A US Federal Challenge Competition, Sponsored by the U.S. Environmental Protection Agency U.S. EPA and partners are seeking high quality, low cost, technologies for measuring global gene expression in samples from non-human organisms including fish, invertebrates, and plants/algae. The technology will serve as a foundation for the next generation of high throughput ecological toxicity tests for use in environmental safety evaluation of chemicals. $300,000 will be awarded for the submission demonstrating the best technical performance at the lowest cost. ------- Problem Chemical Safety Landscape Tens of thousands of chemicals are currently in use and hundreds more are introduced to the market every year. Only a small fraction has been thoroughly evaluated for potential risks to human health and the environment. High Throughput Screening To aid hazard evaluation, U.S. EPA and partners have generated high throughput screening data for thousands of chemicals via the ToxCast and Tox21 programs. Data Gap However, to date, these programs have focused on human health. Coverage of biological pathways and physiological functions that are not conserved with mammals (for example photosynthesis in plants; molting in invertebrates) is lacking. This leaves critical gaps in the current high throughput screening battery relative to identifying ecological hazards and protecting the environment. Proposed Solution Coupling global gene expression profiling with high throughput assays that employ representatives of plant and animal diversity can address these gaps, allowing us to better protect the environment from harmful effects of chemicals. ------- Challenge Overview What we need: U.S. EPA and partners seek high quality, low cost, technologies/platforms for evaluating global gene expression in samples (RNA or tissue homogenates) from non-human organisms. The technology must be capable of the following: • Precisely quantifying the expression individual genes as reflected by relative mRNA abundance. • Providing complete or near-complete pathway coverage for the expressed genome. This may be achieved through direct measurement of all expressed genes, or using validated sentinel gene sets whose expression can be confidently used to infer the response of the rest of the genome. • Facilitating gene-specific annotation suitable for linking target expression with target functions, and readily accommodating updates as annotations evolve and improve over time. • Performing analyses with limited sample masses necessary to make them compatible with high throughput testing protocols with small organisms or cells (e.g., <0.5 ng total RNA per sample). • Generating data in formats compatible with a standardized and automated quality assurance and data analysis workflow that can be used for dose-response modeling and differentially expressed gene/pathway analysis. • Meeting a level of quality, performance, and transcriptome coverage that is also economically and commercially viable for high throughput screening applications. ------- Challenge Process for Solvers: Register Develop technology/platform /informatics for four 1 Pimephales promelas (fish) ¦ Daphnia magna (crustacean) 1 Chironomous dilutus (insect) 1 Raphidocelis subcapitata (algae) Analyze reference samples ' 36 reference RNA samples will be provided !• 9 samples per species Submit for evaluation • Data • Technology description template Challenge Details Step 1. Register To be eligible to compete for the award, prospective solvers must register at: https://www.epa.gov/innovation/ecotox-target-challenge. no later than March 16, 2020. . Eligibility: • Eligible: Individuals, or teams from private companies, academic institutions, non- governmental organizations, or independent research or technological institutes. The competition is open to both US and foreign citizens/organizations. • Not eligible: U.S. or foreign government organizations. • Not eligible: Individuals involved in development of award selection criteria or reference sample generation. ------- Terminology: • Solver(s): Eligible individuals, groups, or organizations competing for the award. • Sponsors: The U.S. EPA and partners involved in the development, implementation, and administration of the Challenge. Required Information: 1. Technical point of contact for application (name, position, title, and contact information). This is the individual that will manage communications and coordination between the challenge sponsors (e.g., US EPA and partners) and the solver(s). 2. Team members (including affiliations) and partner organizations (as appropriate) that will be contributing to a registered application and will share the prize for that solution (if awarded). 3. Financial point of contact: Name, full address, and contact information for the organization or individual that will receive the award money and manage distribution of the funds if your solution is selected as the winner. • Note: If you receive the award, a single lump sum will be provided to a single point of contact, identified above. The financial point of contact will be responsible for any further dispersal or distribution of the award money to the Solver's institution and/or identified team members. Solvers are encouraged to establish and agree to an award distribution plan, in writing, prior to identifying the financial point of contact. 4. Acceptance of the terms of participation. • Solvers will not receive compensation for resources or time invested in addressing the challenge. Only the winning solution will receive a cash award. • Only the top-ranked solution will receive the award. • Solvers retain their rights to all intellectual property (e.g., details and design of their technology) that may be disclosed to the sponsors over the course of the challenge. Technical details and designs will not be disclosed or published without permission from the technical point of contact named in the registration. • Sponsors retain the right to disclose reference sample data, performance criteria, and other evaluation criteria summarized in the technology description template (see p. 8) to provide a transparent reporting of how the winning solution was selected. • Sponsors retain the right to publish, present, and/or otherwise publicize results of the challenge competition that does not involve the disclosure of intellectual property of the Solver(s). Solvers will be afforded opportunity to review publications, presentation, or other publicity to protect against unwanted disclosure of intellectual property. ------- • Solvers reserve the right to remove themselves from the competition at any time, up to final submission of results for evaluation, by notifying the sponsor in writing. The technical point of contact must make the request in writing on behalf of his/her team. • Registration for the challenge does not confer any obligation to deliver results. However, any solvers removing themselves from the competition prior to evaluation forfeit the rights to publish results obtained for the reference samples supplied for the competition unless they obtain written consent from the challenge sponsors. • Solvers that do not submit their results and technology description template by the submission deadline will be automatically removed from the competition and subject to the same terms as if they had forfeited in writing. The submission deadline may be extended at the discretion of the sponsors, but any extension will apply to all registered solvers. Challenge Kick-off: All registered Solvers will be invited to an on-line Challenge kick-off meeting/webinar on March 19, 2020. The technical point of contact and/or appropriate designees should plan to attend. The kick-off meeting will provide an overview of the Challenge process and time-line and provide Solvers with an opportunity to ask questions. The kick-off meeting will serve as the official start date for initiation of the Challenge. Registered Solvers are welcome to begin development of their solution at any time. However, the time-line for availability of method development and reference samples from the sponsors will be set relative to the kick-off date. All registered Solvers will receive reference samples at the same time (within the same 7 d period to the extent that shipping times allow). Step 2. Develop Each registered solver will develop and/or adapt their transcriptomic assay technology for use with the following four species: ------- Pimephales promelas Daphnia magna Chironomous dilutus Raphidocelis subcapitata An assay platform should consist of any means necessary to facilitate global gene expression analysis of an RNA sample input including sample pre-processing, analysis and data collection/extraction. The assay platform, consisting of supplies, reagents, processes, instrumentation, and software can remain proprietary Solvers are challenged to provide complete or near-complete pathway coverage of the expressed transcriptome for all four species. This may be achieved through direct measurement of all expressed genes or using validated sentinel gene sets whose expression can be confidently used to infer the response of the rest of the genome. Technology development should include development of an automated and standardized pipeline or process for annotating the transcripts detected and quantified with respect to their identity (e.g., gene name or unique identifier) and biological function (if known). Sponsors will furnish up to 1 |ig of pooled RNA from each species for the Solver to use for method development. Pooled RNA for method development will be supplied within 60 d of Challenge kick-off. The technical point of contact for the Solver(s) should provide an overnight shipping address and desired date(s) for delivery of the method development samples. The assumed starting material for each technology/solution is purified RNA. At the request of the Solver, alternative sample matrices such as crude homogenates may be provided for Solvers that may want to demonstrate unique capabilities of their technology. However, evaluation will be based on reference purified RNA samples, unless such samples are incompatible with the solution developed. { 6 } ------- Genomic/Transcriptomic Resources for Species of Interest: For annotation and development purposes, the sponsors recommend the following sources: Species Source(s) for Genome/Transcriptome Information Pimephales promelas A reference genome assembly and annotations are available from US EPA; contact Biales.Adam@epa.gov Daphnia magna Genome assemblies: Strain Xinb3 - NCBI BioProject PRJNA298946 Strain SK - NCBI BioProject PRJNA490418 Other resources: Wfleabase.org Chironomous dilutus Chironomous tentans (Chironomous dilutus) transcriptome NCBI BioProject PRJEB1888 Kutsenko A, Svensson T, Nystedt B, Lundeberg J, Bjork P, Sonnhammer E, Giacomello S, Visa N, Wieslander L. The Chironomus tentans genome sequence and the organization of the Balbiani ring genes. BMC Genomics. 2014 Sep 27;15:819. doi: 10.1186/1471-2164-15-819. Raphidocelis subcapitata Raphidocelis subcapitata strain NIES-35 NCBI BioProject PRJDB5653 Suzuki S, Yamaguchi H, Nakajima N, Kawachi M. Raphidocelis subcapitata (=Pseudokirchneriella subcapitata) provides an insight into genome evolution and environmental adaptations in the Sphaeropleales. Sci Rep. 2018 May 23;8(1):8058. doi: 10.1038/s41598-018-26331-6 Solvers are welcome to use al ternative or additional genomic and transcriptomic resources at their discretion. Step 3. Analyze reference samples Around 180 days after kick-off, the Sponsors will provide each registered Solver with a set of 36 reference samples. Nine reference samples will be provided for each species. Samples will be provided blind with respect to experimental conditions used to generate each sample pool but will be labeled with respect to species of origin. All registered solvers will receive aliquots from the same stock reference sample pools and each reference sample pool will be pre-qualified by RNAseq prior to distribution to the Solvers. ------- Each reference sample will consist of 1 |ag purified total RNA. Solvers are encouraged to test smaller amounts of RNA if they wish to demonstrate the capability of their technology to function with smaller sample masses/volumes. Solvers should analyze each of the 36 samples using their technology. Processed data should be provided as a single table with a column for each test sample and a row for each gene or measurement feature (e.g. probe), with values representing a measure of absolute or relative expression of each gene/feature in each sample, intended for differential expression and signaknoise analysis. Solvers must also provide a clear, unambiguous mapping from columns to provided sample IDs/method variants, and from rows to gene IDs in the transcriptome annotation for each species. This table should be provided as either an Excel document, tab- delimited text, or CSV file. Solvers should also document in detail the steps taken to generate the final processed data table from the raw data and note any software or internal data that is proprietary (e.g. that solver would not include when publishing results from this platform; see Technology Description below). Solvers should be willing to furnish raw data upon request by the Sponsors. Step 4. Submit for Evaluation To be considered for the award, Solvers must submit processed data file(s) for the reference samples, annotation files, and a technology description document by the Challenge end date (June 14, 2021). Late submissions will not be granted, unless the deadline has been extended for all eligible Solvers. File submission process Data files to be used for Judging and Award selection should must be submitted to the Sponsors by the end date (June 14, 2021). Sponsors will communicate with the technical contact for each project to determine a viable mechanism for file transmission based on file size and IT security restrictions and policies of the participating organizations. Solvers are responsible for contacting the Sponsors at least 1 month before the challenge end date (i.e., no later than May 14, 2021) to work out the appropriate mechanism for file submission. There can be no exceptions for late submission due to file transmission problems, unless contact was made on or before May 14, 2021. Technology Description Template Along with the data submission, each Solver needs to submit a technology description that includes the following information: ------- 1 Transcriptomic assay platform Title [50 characters or less] • This title will be used for reporting results of the Challenge and for publicity relation to the winning Solution. 2 Transcriptomic assay platform Description [1-2 pages] • Briefly describe the conceptual/theoretical basis of the assay approach. o For example, is a targeted or non-targeted approach employed? o What is the transcript detection/and quantification technology? o How are data collected/extracted (e.g., image analysis; fluorescence detection, other) o Whether data output are in absolute quantities or relative values compared to reference o Emphasize any unique technological approaches/capabilities. • Use of one or more figures to help illustrate the conceptual/theoretical basis of the assay approach is recommended. • Note the details of your assay platform description that should remain proprietary/confidential. 3 Quality control features • Describe quality control features that are incorporated into the assay platform and how they are used to assure data quality. • Include features for assuring data quality both within and across samples 4 Sample requirements • Solvers should report the amount of reference RNA used to generate the data submitted for evaluation. • Solvers have the option to submit supplementary data demonstrating assay performance trade-offs if smaller sample amounts were to be used, but this is not required. • If the platform can accommodate or is designed for sample types other than purified RNA, that should be noted here as well. 5 Per sample cost • Solvers should track and report the per sample cost associated with processing and analysis from the receipt of reference samples to the output of the data submitted for evaluation. • This cost should include the cost of supplies, labor, equipment wear and tear, etc. and should reflect a viable commercial cost per sample charge if these assay were to be conducted as a contract service. 6 Downstream data analysis requirements (if appropriate) • Solvers should also document in detail the steps taken to generate the final processed data table from the raw data. • Note any software or internal data that is proprietary (e.g. that the Solver would not include when publishing results from this platform). • If proprietary software is required, the Solver should provide the name of the software and licensing costs (if any). These will be factored into evaluation of the overall cost per sample. ------- Commercial viability • Future implementation of an ecological high throughput transcriptomics- based chemical screening program is expected to generate thousands or even tens of thousands of samples per year for transcriptomic analyses. Consequently, the winning Solution will require a commercially viable level of throughput that can meet the potential sample demand, including production of supplies, reagents, and the infrastructure necessary for sample processing and analysis. Solvers must describe a pathway and timeline to a commercially viable level of throughput that can reasonably meet HTP sample demand. ------- Step 5. Judging and Award Selection Solver submissions will be evaluated within 120 days of the Challenge end date. Submissions will be judged based on the following criteria: 1. Quality and performance of the platform and associated data (40%) 2. Economic and commercial viability of the platforms (30%) 3. Transcriptome coverage (30%) Scoring Overview Scoring will be based on the weighted (% of total score) criteria provided in the tables below. Each criterion is scored based on either a nominal or fractional scale of 0-5 with 0 being the lowest and 5 being the highest, or by pass/fail with 5 being pass and 0 being fail. 1. Quality and performance Points Weighting Quality control: Does the platform contain a quality control system that addresses consistency within and between samples and is consistent with current standards used within various platforms for transcriptomic analyses? o E.g., Microarray chip-based platforms should contain hybridization controls, redundant positional controls to evaluate edge effects, etc. o E.g., RNA-seq platforms provide number of reads per sample, base quality score by cycle, nucleotide distribution by cycle, GC content, etc. o Note - scoring for this category will require subjectivity from the judging panel Data collection/extraction: Are the data collection/extraction methods and expression normalization/quantification methods described in adequate detail? Are they compatible with the ToxCast high throughput transcriptomics data analysis pipeline? Precision: Precision will be determined by evaluating 1) coefficients of variation of gene expression values across unblinded technical duplicates, 2) correlation analysis of fold-change profiles between selected reference samples, resulting in a metric of concordance, and 3) clustering unblinded reference samples when analyzed together with all conditions. Results from these three analyses will be normalized and merged into a multiplier between 0-1 that will be used to determine the total score between 0-5. Accuracy: Accuracy will be determined by evaluating 1) the percent concordance between fold-change values determined by Solver data compared to the fold-change values determined by pre-qualification. Results to determine the total score between 0-5. Quantity of RNA: Was the quantity of reference RNA used per analysis tracked and reported (Y = points awarded; N=0 points)? 0 to 5 10% 0 to 5 5% multiplier x 5 10% 0 to 5 (Oto 5) 10% 5% ------- o Results generated using < 1 |ag total RNA = 2 pts o Results generated using < 0.25 |ag total RNA = 1 pt o Results generated using < 0.1 |ag total RNA = 1 pt o Results generated using < 0.01 |ag total RNA = 1 pt 2. Economic and commercial viability Points Weighting • Economic viability (i.e., cost per sample, including downstream data analysis cost) o Is the cost of sample preparation and per sample supply and reagent costs for conducting the sample analysis and generating the data provided? If proprietary downstream data analysis software is required, include a per-sample adjustment to overall sample cost based on software license cost. 0 to 5 20% o Total per-sample cost is i. $20 or less = 5 pts ii. >$20-$30 = 4 pts. iii. >$30-$50 = 3 pts. iv. >$50-$75 = 2 pts. v. >$75-$100= 1 pt. vi. >$100 = Opts. • Commercial viability and throughput capability o Is there a reasonable demonstration/description of how and 0 to 5 10% when the Solver would be able to meet the potential throughput requirements of HTP sample generation? 3. Coverage Points Weighting • Approach and annotation o Is the approach taken for detection and quantification of transcript expression adequately described? (e.g., whether the platform employs a targeted or non-targeted analysis and the general means by which the platform detects and quantifies transcript presence and abundance) o Are annotation files provided with each platform/species that contain the required information and link to the data files? 0 to 5 10% • Transcriptome coverage o What proportion of transcriptome coverage do the platforms have in relation to the pre-qualification standards? The mean percent coverage will be calculated across the four species' platforms and used as a multiplier to determine point value. multiplier x 5 20% ------- • Species coverage o Did the solvers provide a platform and reference sample data for all four species? *Eligible submissions will include platforms, data and associated required information for all Notification Sponsors will notify the Solvers, via the technical point of contact within 120 days of the Challenge end date. Solvers will receive scores for their submission along with a blinded summary of the scores attained for the competing submissions. The winning Solution will be announced at the same time. After all Solvers have been notified, Sponsors will publicly announce the Award and the winning solution. The announcement will be accompanied by a press release, social media, and other public promotion by the Sponsors. By participating in the Challenge, the Solvers agree to public disclosure of all team members that contributed to the winning Solution, and their affiliations. Step 6. Reporting Following selection of the winning Solution, the Sponsors will prepare a written report summarizing the results of the TARGET Challenge. The number of registered Solvers, and Solvers submitting data and documents for the competition will be reported along with summaries of the scores obtained. The rationale for selection of the winning Solution, based on the scoring criteria will be provided. Solvers will have the opportunity to review the report to ensure non-disclosure of confidential intellectual property prior to dissemination or publication. Data for the winning Solution will be made public. Data from non-winning submissions may be made public as required to facilitate peer review and publication of the Challenge report. Following selection of the winning Solution, reference samples and experimental details associated with the reference samples will be disclosed to all Solvers that participated in the competition. Solvers are welcome to publish results and conclusions related to their analysis of the reference sample set but should include relevant individuals from the Sponsoring organization as co-authors to ensure that experimental details and associated conclusions are presented accurately. YorN *100% four species. ------- |