Writing science questions: sources and topics
Posted: Tue May 18, 2010 12:07 am
While editing the science for ACF Nationals, Susan and I noticed some fairly widespread problems with questions that were submitted, so we want to offer some suggestions. An earlier thread (viewtopic.php?f=9&t=8822&p=159369#p158855) covered some similar issues, and might also be worth reading. And of course you should read Jerry's "How to Write Questions" (http://www.acf-quizbowl.com/documents/howtowrite.php) on the ACF website. I'll lead off with some comments, and Susan will add more.
What sources to use
The best sources to use when writing science questions, especially if you're not an expert on the topic you're writing about, are widely-used textbooks. This guarantees that you're writing on a well-known topic and that the clues you choose are things that knowledgeable people can reasonably be expected to know about. The more you know about a topic, the more free you should feel to venture to more specialized sources, but good general textbooks are always a good place to start. Which isn't to say that a library book chosen at random is a good place; try looking at syllabi for courses to get a sense of what books people are most likely to be familiar with.
Wikipedia on some topics is very good, on others less so; it can be a good place to do some reading and get a quick sense of what might be important about a topic. But writing an entire question from Wikipedia (as some people pretty clearly did) is a bad idea. One issue that arose from this, which we managed to catch, were nearly-identical tossups written for the NAQT ICT and for ACF Nationals by different people. I can't tell you definitively that both writers were using Wikipedia, but the selection and ordering of clues was identical, and the lead-in was a fairly recent development that featured prominently in the beginning of the Wikipedia article.
Using primary literature is dangerous, although it can work well if you're familiar with the subject you're writing about or willing to spend some time reading. The trouble is that the scientific literature is vast, and even experts in a field usually can't read more than small fraction of the new papers in that field. If you grab one of those papers at random and use it for your lead-in, the result will be a lead-in that no quizbowler will know, and perhaps only a handful of people in the entire world will recognize. This isn't an exaggeration -- some of the submitted questions were guilty of it. Just as textbooks are good guides to what is well-known, a well-written review article on a subject can be a good place to get a sense of what is important and what isn't, and might be a better source of clues than a random journal article. If you're just Googling a topic and clicking at random, you'll end up with nonsense. Google Scholar displays moderately accurate lists of how many times a paper has been cited, which can be a rough proxy for how interesting it is -- and, at minimum, whether anyone has ever heard of it. Individual fields might have their own more useful databases of papers (in my field, particle physics, the SLAC SPIRES database is best, although often sluggish).
To give some examples, without naming the guilty parties or quoting the questions: one submitted physics tossup began with a clue taken from a 1996 paper that has never been cited (according to searches on Google Scholar and the SLAC SPIRES database), by a little-known author with a marginal publication record. Another began with two references to recent literature: a 2009 paper with no citations according to Google Scholar (and one of the author's names spelled incorrectly in the question!) and a 2007 paper that has been cited once. In the latter case, the clue arose from a misreading of the abstract and was not correct. A common feature of questions with lead-ins from recent literature was phrasing copied almost directly from the abstract of the paper; these questions showed no evidence that the writer understood what they were writing about, and often showed evidence that they didn't. If you don't understand what you're writing, and you can't give an explanation of why the result is significant or relevant, it's probably a sign that you need to find a different clue.
There's nothing inherently wrong with using recent literature as a clue; for instance, I wrote a tossup mentioning the Kerr black hole/CFT correspondence, a 2008 result of Strominger and collaborators that has been cited 96 times (and it's relevant for several different subfields of physics, so it's fairly broadly known). The threshold for looking at a citation count and deciding if a result is well-known varies by field, so it's hard to declare a sharp cutoff. But in general, if you find something with zero or only a handful of citations, it is probably not well-known even to experts, much less to the quizbowl community. Such a clue won't reward expertise, it will just confuse even the experts.
Distribution
Many, many teams completely failed to follow the expected distribution. Some submitted packets didn't include any chemistry at all. Others included oddly slanted subdistributions, like a heavy emphasis on math or CS in all of the "other science" questions. The ACF distribution is:
* Biology 1/1
* Chemistry 1/1
* Physics 1/1
* Math or computer science: 1 question
* Astronomy, earth science/geology, or other science not covered above: 1 question
* Any science: 1/1 (you can write a math question here if you wrote computer science above, or earth science here if you wrote astronomy above, but don't write a second question on any of the "minor" fields. You can also just write more biology, chemistry, or physics here.)
Note that, e.g., 1/1 math or 1/1 CS is never acceptable, and that it's perfectly OK to use the "any science" distribution to write more bio or physics or chem.
I'll end this here and let Susan add her comments, though I might have more to say later.
What sources to use
The best sources to use when writing science questions, especially if you're not an expert on the topic you're writing about, are widely-used textbooks. This guarantees that you're writing on a well-known topic and that the clues you choose are things that knowledgeable people can reasonably be expected to know about. The more you know about a topic, the more free you should feel to venture to more specialized sources, but good general textbooks are always a good place to start. Which isn't to say that a library book chosen at random is a good place; try looking at syllabi for courses to get a sense of what books people are most likely to be familiar with.
Wikipedia on some topics is very good, on others less so; it can be a good place to do some reading and get a quick sense of what might be important about a topic. But writing an entire question from Wikipedia (as some people pretty clearly did) is a bad idea. One issue that arose from this, which we managed to catch, were nearly-identical tossups written for the NAQT ICT and for ACF Nationals by different people. I can't tell you definitively that both writers were using Wikipedia, but the selection and ordering of clues was identical, and the lead-in was a fairly recent development that featured prominently in the beginning of the Wikipedia article.
Using primary literature is dangerous, although it can work well if you're familiar with the subject you're writing about or willing to spend some time reading. The trouble is that the scientific literature is vast, and even experts in a field usually can't read more than small fraction of the new papers in that field. If you grab one of those papers at random and use it for your lead-in, the result will be a lead-in that no quizbowler will know, and perhaps only a handful of people in the entire world will recognize. This isn't an exaggeration -- some of the submitted questions were guilty of it. Just as textbooks are good guides to what is well-known, a well-written review article on a subject can be a good place to get a sense of what is important and what isn't, and might be a better source of clues than a random journal article. If you're just Googling a topic and clicking at random, you'll end up with nonsense. Google Scholar displays moderately accurate lists of how many times a paper has been cited, which can be a rough proxy for how interesting it is -- and, at minimum, whether anyone has ever heard of it. Individual fields might have their own more useful databases of papers (in my field, particle physics, the SLAC SPIRES database is best, although often sluggish).
To give some examples, without naming the guilty parties or quoting the questions: one submitted physics tossup began with a clue taken from a 1996 paper that has never been cited (according to searches on Google Scholar and the SLAC SPIRES database), by a little-known author with a marginal publication record. Another began with two references to recent literature: a 2009 paper with no citations according to Google Scholar (and one of the author's names spelled incorrectly in the question!) and a 2007 paper that has been cited once. In the latter case, the clue arose from a misreading of the abstract and was not correct. A common feature of questions with lead-ins from recent literature was phrasing copied almost directly from the abstract of the paper; these questions showed no evidence that the writer understood what they were writing about, and often showed evidence that they didn't. If you don't understand what you're writing, and you can't give an explanation of why the result is significant or relevant, it's probably a sign that you need to find a different clue.
There's nothing inherently wrong with using recent literature as a clue; for instance, I wrote a tossup mentioning the Kerr black hole/CFT correspondence, a 2008 result of Strominger and collaborators that has been cited 96 times (and it's relevant for several different subfields of physics, so it's fairly broadly known). The threshold for looking at a citation count and deciding if a result is well-known varies by field, so it's hard to declare a sharp cutoff. But in general, if you find something with zero or only a handful of citations, it is probably not well-known even to experts, much less to the quizbowl community. Such a clue won't reward expertise, it will just confuse even the experts.
Distribution
Many, many teams completely failed to follow the expected distribution. Some submitted packets didn't include any chemistry at all. Others included oddly slanted subdistributions, like a heavy emphasis on math or CS in all of the "other science" questions. The ACF distribution is:
* Biology 1/1
* Chemistry 1/1
* Physics 1/1
* Math or computer science: 1 question
* Astronomy, earth science/geology, or other science not covered above: 1 question
* Any science: 1/1 (you can write a math question here if you wrote computer science above, or earth science here if you wrote astronomy above, but don't write a second question on any of the "minor" fields. You can also just write more biology, chemistry, or physics here.)
Note that, e.g., 1/1 math or 1/1 CS is never acceptable, and that it's perfectly OK to use the "any science" distribution to write more bio or physics or chem.
I'll end this here and let Susan add her comments, though I might have more to say later.