Refine
Has Fulltext
- no (33)
Year of publication
Document Type
- Article (29)
- Review (2)
- Monograph/Edited Volume (1)
- Preprint (1)
Language
- English (33) (remove)
Is part of the Bibliography
- yes (33)
Keywords
- Numerical distance effect (3)
- Activation suppression model (2)
- Mental number line (2)
- RT models (2)
- Simon effect (2)
- numerical distance effect (2)
- 2AFC (1)
- Bayesian inference (1)
- Bundesliga (1)
- Categorization (1)
- Category effect (1)
- Conflict task (1)
- Consciousness (1)
- Decision making (1)
- Decision-making (1)
- Delta plot (1)
- Delta plots (1)
- Diffusion model (1)
- Diffusion models (1)
- Libet (1)
- Line perception (1)
- Muller-Lyer illusion (1)
- Neuroscience (1)
- Number-size congruency effect (1)
- Number-size congruity effect (1)
- Numerical cognition (1)
- Numerical magnitude (1)
- Paired comparison (1)
- Psychometric and chronometric function (1)
- ROC curve (1)
- Reafference principle (1)
- Response bias (1)
- Sensitivity (1)
- Signal detection theory (1)
- Size-weight illusion (1)
- Visual search (1)
- biased decisions (1)
- bimodality (1)
- cognitive aging effects (1)
- compound detection or discrimination (1)
- diffusion model (1)
- display size effect (1)
- error latency (1)
- hypothesis testing (1)
- individual differences (1)
- measurement error (1)
- non-normality (1)
- null effects (1)
- numerical comparison (1)
- posterior statistical power (1)
- probability of replication (1)
- probability summation (1)
- psychological models (1)
- psychometric functions (1)
- random walk model (1)
- random walk models (1)
- random-effects model (1)
- redundancy gain (1)
- regression towards the mean (1)
- repeated measures (1)
- saccadic amplitude (1)
- saccadic latency (1)
- signal prevalence (1)
- skewed distributions (1)
- speed-accuracy trade-off (1)
- sports statistics (1)
- statistical estimation (1)
- strategic termination (1)
- visual search (1)
We compared individual-participant and jackknife-based methods for scoring the onset latencies of event-related potential (ERP) components using a diffusion process as a model for an ERP. We studied "ramp-like" components in which the true ERP increases or decreases monotonically, except for noise. If the growth rates of such components vary across participants, the jackknife-based measure can easily have only 10%-20% as much error variance as the traditional method, and this advantage is magnified with more participants. We also studied boolean AND-shaped or "bump-like" components. Jackknifing generally yielded smaller error variances with these components too, especially when the component's peak amplitude varied across participants, but less so if the component's peak latency varied. These results help illuminate the reasons for the superiority of jackknife-based onset latency measures over traditional measures in recent simulations.
R. Sekuler, P. Tynan, and E. Levinson (1973) found that when 2 characters are presented side-by-side with a short onset asynchrony, subjectively, they often appear in a "first-left, then-right" order. The authors of this article conducted 6 experiments in which observers judged the temporal order (TOJs) in which 2 digits were presented. They found a consistent TOJ benefit (larger d') when the numerically smaller digit was presented first, even though this semantic information was irrelevant to the task and unrelated to the correct response. They concluded that digits located to the left of the mental number line are transmitted faster to a central comparison stage, which represents an "internal counterpart" to the Sekuler et al. (1973) finding regarding external locations. A corresponding benefit is found for letters pairs (e.g., A-Z) and also for mixed digit-letter pairs (e.g., I-Z).
We consider the specific transformation of a Wiener process {X(t), t >= 0} in the presence of an absorbing barrier a that results when this process is "time-locked" with respect to its first passage time T-a through a criterion level a, and the evolution of X(t) is considered backwards ( retrospectively) from T-a. Formally, we study the random variables defined by Y(t) = X(T-a - t) and derive explicit results for their density and mean, and also for their asymptotic forms. We discuss how our results can aid interpretations of time series "response-locked" to their times of crossing a criterion level.
Many perceptual and cognitive tasks permit or require the integrated cooperation of specialized sensory channels, detectors, or other functionally separate units. In compound detection or discrimination tasks, 1 prominent general mechanism to model the combination of the output of different processing channels is probability summation. The classical example is the binocular summation model of Pirenne (1943), according to which a weak visual stimulus is detected if at least 1 of the 2 eyes detects this stimulus; as we review briefly, exactly the same reasoning is applied in numerous other fields. It is generally accepted that this mechanism necessarily predicts performance based on 2 (or more) channels to be superior to single channel performance, because 2 separate channels provide "2 chances" to succeed with the task. We argue that this reasoning is misleading because it neglects the increased opportunity with 2 channels not just for hits but also for false alarms and that there may well be no redundancy gain at all when performance is measured in terms of receiver operating characteristic curves. We illustrate and support these arguments with a visual detection experiment involving different spatial uncertainty conditions. Our arguments and findings have important implications for all models that, in one way or another, rest on, or incorporate, the notion of probability summation for the analysis of detection tasks, 2-alternative forced-choice tasks, and psychometric functions.
Neuroscientific studies have shown that brain activity correlated with a decision to move can be observed before a person reports being consciously aware of having made that decision (e.g., Libet, Gleason, Wright, & Pearl, 1983; Soon, Brass, Heinze, & Haynes, 2008). Given that a later event (i.e., conscious awareness) cannot cause an earlier one (i.e., decision-related brain activity), such results have been interpreted as evidence that decisions are made unconsciously (e.g., Libet, 1985). We argue that this interpretation depends upon an all-or-none view of consciousness, and we offer an alternative interpretation of the early decision-related brain activity based on models in which conscious awareness of the decision to move develops gradually up to the level of a reporting criterion. Under this interpretation, the early brain activity reflects sub-criterion levels of awareness rather than complete absence of awareness and thus does not suggest that decisions are made unconsciously.
Delta plots (DPs) graphically compare reaction time (RT) quantiles obtained under two experimental conditions. In some research areas (e.g., Simon effects), decreasing delta plots (nDPs) have consistently been found, indicating that the experimental effect is largest at low quantiles and decreases for higher quantiles. nDPs are unusual and intriguing: They imply that RT in the faster condition is more variable, a pattern predicted by few standard RT models. We describe and analyze five classes of well-established latency mechanisms that are consistent with nDPs-exhaustive processing models, correlated stage models, mixture models, cascade models, and parallel channels models-and discuss the implications of our analyses for the interpretation of DPs. DPs generally do not imply any specific processing model; therefore, it is more fruitful to start from a specific quantitative model and to compare the DP it predicts with empirical data.
We present three experiments in which observers searched for a target digit among distractor digits in displays in which the mean numerical target-distractor distance was varied. Search speed and accuracy increased with numerical distance in both target-present and target-absent trials (Exp. 1A). In Experiment 1B, the target 5 was replaced with the letter S. The results suggest that the findings of Experiment 1A do not simply reflect the fact that digits that were numerically closer to the target coincidentally also shared more physical features with it. In Experiment 2, the numerical distance effect increased with set size in both target-present and target-absent trials. These findings are consistent with the view that increasing numerical target-distractor distance affords faster nontarget rejection and target identification times. Recent neurobiological findings (e.g., Nieder, 2011) on the neuronal coding of numerosity have reported a width of tuning curves of numerosity-selective neurons that suggests graded, distance-dependent coactivation of the representations of adjacent numbers, which in visual search would make it harder to reject numerically closer distractors as nontargets.
Aggregate and individual replication probability within an explicit model of the research process
(2011)
We study a model of the research process in which the true effect size, the replication jitter due to changes in experimental procedure, and the statistical error of effect size measurement are all normally distributed random variables. Within this model, we analyze the probability of successfully replicating an initial experimental result by obtaining either a statistically significant result in the same direction or any effect in that direction. We analyze both the probability of successfully replicating a particular experimental effect (i.e., the individual replication probability) and the average probability of successful replication across different studies within some research context (i.e., the aggregate replication probability), and we identify the conditions under which the latter can be approximated using the formulas of Killeen (2005a, 2007). We show how both of these probabilities depend on parameters of the research context that would rarely be known in practice. In addition, we show that the statistical uncertainty associated with the size of an initial observed effect would often prevent accurate estimation of the desired individual replication probability even if these research context parameters were known exactly. We conclude that accurate estimates of replication probability are generally unattainable.
Using a large representative database (12,902 matches from the top professional football league in Germany), I show that the number (441) of two-penalty matches is larger than expected by chance, and that among these 441 matches there are considerably more matches in which each team is awarded one penalty than would be expected on the basis of independent penalty kick decisions (odds ratio=11.2, relative risk=6.34). Additional analyses based on the score in the match before a penalty is awarded and on the timing of penalties, suggest that awarding a first penalty to one team raises the referee's penalty evidence criterion for the same team, and lowers the corresponding criterion for the other team.