I run some experiments on fish behavior , and extracted some variables from each test (tests are replicated twice). One of the variable is "latency to exit a shelter" which is a classical operational definition for boldness and/or exploratory behavior.

Although most of my fish left the shelter during the test time (1200 seconds), some of them never left in either replicate 1 and replicate 2 (very shy fish I guess). When estimating repetabilities, I wonder how to treat these maximum latencies (1200 s). I read in the literature that sometimes those values are simply included in the dataset together with the "real" latencies from the other fish. But my concern is that this could overestimate repetabilities since the match between the first and the second replicate is assumed to be perfect (1200 s in both replicates).

Is there any way to account for this or this is just the way to go?

Edited: a further concern of including these values is that the distribution of the data could be difficult to handle. For example not considering those values I have a clear poisson or gamma distributions (a lot of values close to zero). But if I include these maximum latencies, then I have an accumulation of data on the right end of the distribution, which complicate thigns I guess.

Thanks!

David

More David Villegas-Ríos's questions See All
Similar questions and discussions