Cohort models debated

Two prominent scientists argue the merits of forming a new 500,000-person cohort versus pooling data from existing studies

| 3 min read

Register for free to listen to this article
Listen with Speechify
0:00
3:00
Share
Francis Collins and Walter Willett are debating different models for constructing a huge cohort study resource to investigate the influence of genes and environment: On the pages of this week's Nature, Collins outlines his support for a made-from-scratch cohort of half a million people, while Willett backs a cheaper, faster model that combines existing cohort studies. "If it only cost a few hundred million dollars to start a new cohort, there wouldn't be much of an issue at all," Willett, Chairman of the Department of Nutrition at the Harvard School of Public Health (HSPH), told The Scientist. But with a projected cost of $3 billion or more for a study that includes subjects representative of the country's ethnicities and walks of life, "then it's going to displace a lot of other research at a time when research is being hugely squeezed," he said. The National Institutes of Health, where Collins directs the National Human Genome Institute, originally envisioned the massive cohort project as providing a vast database of genetic and biomarker data linked to information about participants' diet, geography, smoking habits, and many other factors.Collins and his co-author, Teri Manolio, were not available for comment before deadline.According to Collins' commentary, a new cohort study holds several advantages over a cohort that combines existing studies. For instance, researchers could apply exactly the same methods to gather data from each subject, include enough minorities and other groups without statistical adjustments, collect data on entire lifespans, and employ the latest technology, such as mass spectrometers.One advantage of larger cohorts that was not explored in the commentaries is that their standardized approach makes it easier to combine data on subjects, while some detail gathered by small cohorts could be lost when they are pooled, said Willett. When researchers examine pooled data, they must use a "common denominator" of data, potentially ignoring variables that had been gathered by some studies, he said. But this drawback is partly countered by another advantage of pooled studies: time. In addition to its high cost, a new, large-scale cohort study will take as long as 10 years to produce any results, due to time-consuming planning and enrollment, as well as the time needed to research common diseases, argues Willett.In contrast, combining existing cohorts would take much less time, given that most of the data already exist, Willett says. He lists 13 studies -- such as the Nurse's Health Study -- that he feels are suited for pooling, including about 1.4 million subjects and 845,000 biological samples. Four of these are associated with HSPH, accounting for more than a quarter of the total patients and biological samples, putting Willett in significant control of any combined cohort project if it moves forward.In a ballpark estimate, Willett told The Scientist that the kind of cohort study model he envisions might require a budget of $20 million to $50 million per year, with perhaps 50-75 percent coming from existing funding from smaller constituent cohort studies. A pooled cohort project is "certainly a much more cost-efficient strategy," said Roberta Ness, chair of the University of Pittsburgh department of epidemiology. "It won't get you everything -- there will be real limitations with respect to the age composition of the cohorts, the baseline data collection, [and standardization of] the specimen-collection techniques," she said.A pooled cohort's shortcomings will hinder the study of diseases that develop early in life, such as asthma and major psychoses, because it would lack adequate data from children, argues Collins. He is also concerned that research into changing environmental factors, such as emerging infections, will suffer, because researchers using a pooled cohort might not concentrate on updating their information past a number of basic variables. This isn't an insurmountable problem, insisted Carolyn Williams, chief of epidemiology in the division of AIDS at the National Institute of Allergy and Infectious Diseases, who told The Scientist that gaps like this are routinely filled in with existing pooled cohorts, where researchers check study subjects for a new set of variables once the cohorts have been pooled. The NHGRI has already taken expensive steps toward a large cohort study, including its $40 million Genes and Environment Initiative, which is developing genotyping facilities, data analysis infrastructure, disease biomarkers, and better tools for measuring environmental exposures.Chris Womack mail@the-scientist.comLinks within this article:F. Collins, "Necessary but Not Sufficient," Nature, January 18, 2007. http://www.Nature.comW. Willett, "Merging and Emerging Cohorts," Nature, January 18, 2007. http://www.Nature.comWalter Willett http://www.hsph.harvard.edu/facres/wlltt.htmlSpivey, A. "Gene-Environment studies: What, How, When, Where?" Environ. Health Perspect. 114, A466?A467 (2006). http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1551984Francis Collins http://www.genome.gov/10000779F. Collins, "Delivering on the Dream: biomedical Research in the Genomic Era," The Scientist, v20: 2, p 46. http://www.the-scientist.com/article/display/23067/M. Anderson, "A 500,000-person study?" The Scientist, May 26, 2004 http://www.the-scientist.com/article/display/22202/Genes and Environment Initiative http://www.gei.nih.gov
Interested in reading more?

Become a Member of

The Scientist Logo
Receive full access to more than 35 years of archives, as well as TS Digest, digital editions of The Scientist, feature stories, and much more!
Already a member? Login Here

Meet the Author

  • Chris Womack

    This person does not yet have a bio.
Share
May digest 2025 cover
May 2025, Issue 1

Study Confirms Safety of Genetically Modified T Cells

A long-term study of nearly 800 patients demonstrated a strong safety profile for T cells engineered with viral vectors.

View this Issue
iStock

TaqMan Probe & Assays: Unveil What's Possible Together

Thermo Fisher Logo
Meet Aunty and Tackle Protein Stability Questions in Research and Development

Meet Aunty and Tackle Protein Stability Questions in Research and Development

Unchained Labs
Detecting Residual Cell Line-Derived DNA with Droplet Digital PCR

Detecting Residual Cell Line-Derived DNA with Droplet Digital PCR

Bio-Rad
How technology makes PCR instruments easier to use.

Making Real-Time PCR More Straightforward

Thermo Fisher Logo

Products

fujirebio-square-logo

Fujirebio Receives Marketing Clearance for Lumipulse® G pTau 217/ β-Amyloid 1-42 Plasma Ratio In-Vitro Diagnostic Test

The Scientist Placeholder Image

Biotium Launches New Phalloidin Conjugates with Extended F-actin Staining Stability for Greater Imaging Flexibility

Leica Microsystems Logo

Latest AI software simplifies image analysis and speeds up insights for scientists

BioSkryb Genomics Logo

BioSkryb Genomics and Tecan introduce a single-cell multiomics workflow for sequencing-ready libraries in under ten hours