Responses:

2,313 useable and qualified responses were received. Responses were rejected if the respondent did not identify him or herself as a tenured, permanent, or tenure-track scientist working in a non-commercial organization, if the respondent's institution was not identified or identifiable, or if the response was a duplicate, based on e-mail address and other criteria.

Analysis:

Respondents were asked to assess their working environment according to 41 criteria in 8 different areas by posing positive statements with which the respondent was asked to agree or disagree. Answers were scored on a 1 - 5 scale with 5 = "Strongly agree", 1 = "Strongly disagree" and 3 = "Neither agree nor disagree". Respondents were also asked to rank how important each factor was important to them.

Identification of Institutions:

As far as possible institutions were identified and names were standardized.

Thresholds:

US organizations that received fewer than 5 responses were omitted from the rankings. Non-US organizations with fewer than 4 responses were eliminated. We ranked 73 institutions - 54 from the US and 19 from the rest of the world.

Scoring:

Scores for each statement were averaged by institution and country.

Ranking:

In order to calculate the overall rankings of institutions, we first weighted each factor based on the importance ranking. Because several factors that ranked as important in the United States are valued less elsewhere and vice versa, we used different factor weightings to rank US and non-US institutions. The overall rankings were based on the average score per institution on all factors weighted as described.

Institutions were also ranked based on all factors, unweighted. In addition, we ranked institutions based on unweighted average scores for the 8 categories of statements included in the survey. These categories are:

  • Job Satisfaction
  • Peers
  • Infrastructure and Environment
  • Research Resources
  • Pay
  • Management and Policies
  • Teaching and Mentoring
  • Tenure

Caveats:

  • The sample of respondents, while large, was self selected, which may introduce some bias into the results.
  • The scoring of results is not standardized and standards may fluctuate between individuals, institutions and countries.
  • In some cases, small sample responses may have led to bias in the results.
  • No attempt has been made to measure the statistical significance of the results. The difference between, say a 10th ranked and a 20th ranked institution may be insignificant.

Survey Development and Analysis:

The survey development and data analysis were carried out by AMG Science Publishing (www.amgpublishing.com).

Interested in reading more?

Magaizne Cover

Become a Member of

Receive full access to digital editions of The Scientist, as well as TS Digest, feature stories, more than 35 years of archives, and much more!