Team:Michigan Software/Project

From 2014.igem.org

(Difference between revisions)
Line 5: Line 5:
<div class="project_describe">  
<div class="project_describe">  
<h1>Description</h1>
<h1>Description</h1>
 +
<font size = "3">
<p>
<p>
Line 22: Line 23:
To attempt to address these problems, we set out to build a database that integrates a crowdsourced ratings and comments system to clearly document, rate, elaborate, review, and organize variants of experimental protocols. Such a tool serves as a curator for protocol variants and enables investigators to compare protocol efficacies, acceptance via crowdsourced ratings and provides an avenue for transfer of experiential knowledge through protocol comments. In all, we hope these tools will help students and investigators document, organize, and compare protocols to assist with scientific experimentation.
To attempt to address these problems, we set out to build a database that integrates a crowdsourced ratings and comments system to clearly document, rate, elaborate, review, and organize variants of experimental protocols. Such a tool serves as a curator for protocol variants and enables investigators to compare protocol efficacies, acceptance via crowdsourced ratings and provides an avenue for transfer of experiential knowledge through protocol comments. In all, we hope these tools will help students and investigators document, organize, and compare protocols to assist with scientific experimentation.
</p>
</p>
 +
</font>
</div>
</div>
</html>
</html>

Revision as of 00:54, 16 October 2014

Michigan Software


Description

Choosing reliable protocols for new experiments is a problem laboratories routinely face. Even when drawing from research publications, it is difficult to determine which protocols will produce the best results. Experimental practices differ immensely across laboratories, and precise details of these practices may be lost or forgotten as skilled members leave the lab. These two realities give rise to a vast number of experimental protocols without their original curators around to describe them. Furthermore, no tool yet exists to allow wet lab investigators to measure and compare the efficacy of protocols before executing them.

Such fragmentation in protocol methods and their documentation often hampers scientific progress. Indeed, there are few well-defined protocols that are generally agreed upon by the scientific community, in part due to the lack of a system that measures a protocol’s success. In turn, the lack of commonly accepted protocols and inadequate documentation affects experimental reproducibility through method inconsistencies across laboratories.

To attempt to address these problems, we set out to build a database that integrates a crowdsourced ratings and comments system to clearly document, rate, elaborate, review, and organize variants of experimental protocols. Such a tool serves as a curator for protocol variants and enables investigators to compare protocol efficacies, acceptance via crowdsourced ratings and provides an avenue for transfer of experiential knowledge through protocol comments. In all, we hope these tools will help students and investigators document, organize, and compare protocols to assist with scientific experimentation.

Contents

Aims

Methods

Django package for Python

SQLite

Concept map.png

Success

Future Directions

contact info: igemsoftware-requests@umich.edu.