Connectionists: Best practices in model publication
Brad Wyble
bwyble at gmail.com
Tue Jan 28 08:25:04 EST 2014
Thanks Randal, that's a great suggestion. I'll ask my colleagues in
physics for their perspective as well.
-Brad
On Mon, Jan 27, 2014 at 11:54 PM, Randal Koene <randal.a.koene at gmail.com>wrote:
> Hi Brad,
> This reminds me of theoretical physics, where proposed models are
> expounded in papers, often without the ability to immediately carry out
> empirical tests of all the predictions. Subsequently, experiments are often
> designed to compare and contrast different models.
> Perhaps a way to advance this is indeed to make the analogy with physics?
> Cheers,
> Randal
>
> Dr. Randal A. Koene
> Randal.A.Koene at gmail.com - Randal.A.Koene at carboncopies.org
> http://randalkoene.com - http://carboncopies.org
>
>
> On Mon, Jan 27, 2014 at 8:29 PM, Brad Wyble <bwyble at gmail.com> wrote:
>
>> Thank you Mark, I hadn't seen this paper. She includes this other point
>> that should have been in my list:
>>
>> "From a practical point of view, as noted the time required to build
>> and analyze a computational model is quite substantial and validation may
>> require teams. To delay model presentation until validation has occurred
>> retards the development of the scientific field. " ----Carley (1999)
>>
>>
>> And here is a citation for this paper.
>> Carley, Kathleen M., 1999. Validating Computational Models. CASOS Working
>> Paper, CMU
>>
>> -Brad
>>
>>
>>
>>
>> On Mon, Jan 27, 2014 at 9:48 PM, Mark Orr <mo2259 at columbia.edu> wrote:
>>
>>> Brad,
>>> Kathleen Carley, at CMU, has a paper on this idea (from the 1990s),
>>> suggesting the same practice. See
>>> http://www2.econ.iastate.edu/tesfatsi/EmpValid.Carley.pdf
>>>
>>> Mark
>>>
>>> On Jan 27, 2014, at 9:39 PM, Brad Wyble wrote:
>>>
>>> Dear connectionists,
>>>
>>> I wanted to get some feedback regarding some recent ideas concerning the
>>> publication of models because I think that our current practices are
>>> slowing down the progress of theory. At present, at least in many
>>> psychology journals, it is often expected that a computational modelling
>>> paper includes experimental evidence in favor of a small handful of its
>>> own predictions. While I am certainly in favor of model testing, I have
>>> come to the suspicion that the practice of including empirical validation
>>> within the same paper as the initial model is problematic for several
>>> reasons:
>>>
>>> It encourages the creation only of predictions that are easy to test
>>> with the techniques available to the modeller.
>>>
>>> It strongly encourages a practice of running an experiment, designing a
>>> model to fit those results, and then claiming this as a bona fide
>>> prediction.
>>>
>>> It encourages a practice of running a battery of experiments and
>>> reporting only those that match the model's output.
>>>
>>> It encourages the creation of predictions which cannot fail, and are
>>> therefore less informative
>>>
>>> It encourages a mindset that a model is a failure if all of its
>>> predictions are not validated, when in fact we actually learn more from a
>>> failed prediction than a successful one.
>>>
>>> It makes it easier for experimentalists to ignore models, since such
>>> modelling papers are "self contained".
>>>
>>> I was thinking that, instead of the current practice, it should be
>>> permissible and even encouraged that a modelling paper should not include
>>> empirical validation, but instead include a broader array of predictions.
>>> Thus instead of 3 successfully tested predictions from the PI's own lab, a
>>> model might include 10 untested predictions for a variety of different
>>> experimental techniques. This practice will, I suspect, lead to the
>>> development of bolder theories, stronger tests, and most importantly,
>>> tighter ties between empiricists and theoreticians.
>>>
>>> I am certainly not advocating that modellers shouldn't test their own
>>> models, but rather that it should be permissible to publish a model without
>>> testing it first. The testing paper could come later.
>>>
>>> I also realize that this shift in publication expectations wouldn't
>>> prevent the problems described above, but it would at least not reward
>>> them.
>>>
>>> I also think that modellers should make a concerted effort to target
>>> empirical journals to increase the visibility of models. This effort
>>> should coincide with a shift in writing style to make such models more
>>> accessible to non modellers.
>>>
>>> What do people think of this? If there is broad agreement, what would be
>>> the best way to communicate this desire to journal editors?
>>>
>>> Any advice welcome!
>>>
>>> -Brad
>>>
>>>
>>>
>>> --
>>> Brad Wyble
>>> Assistant Professor
>>> Psychology Department
>>> Penn State University
>>>
>>> http://wyblelab.com
>>>
>>>
>>>
>>
>>
>> --
>> Brad Wyble
>> Assistant Professor
>> Psychology Department
>> Penn State University
>>
>> http://wyblelab.com
>>
>
>
--
Brad Wyble
Assistant Professor
Psychology Department
Penn State University
http://wyblelab.com
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20140128/5ee75078/attachment.html>
More information about the Connectionists
mailing list