21 |
Using Virtual Reality to Assess Reading Fluency in Children
|
|
|
|
In: ISSN: 2504-284X ; Frontiers in Education ; https://hal-amu.archives-ouvertes.fr/hal-03257346 ; Frontiers in Education , Frontiers, 2021, 6, ⟨10.3389/feduc.2021.693355⟩ (2021)
|
|
BASE
|
|
Show details
|
|
22 |
Word representation and processing in deaf readers: Evidence from ERPs and eye-tracking
|
|
|
|
BASE
|
|
Show details
|
|
23 |
Building a multimodal corpus to study the development of techno-semio-pedagogical competence across different videoconferencing settings and languages
|
|
|
|
In: https://halshs.archives-ouvertes.fr/halshs-03476577 ; 2021 (2021)
|
|
BASE
|
|
Show details
|
|
24 |
Children’s Generalization of Novel Relational Nouns in Comparison Contexts ; Children’s Generalization of Novel Relational Nouns in Comparison Contexts: An Eye Tracking Analysis
|
|
|
|
In: ISSN: 1069-7977 ; Proceedings of the Annual Conference of the Cognitive Science Society ; Comparative Cognition : Animal Minds ; https://hal-univ-bourgogne.archives-ouvertes.fr/hal-03481949 ; Comparative Cognition : Animal Minds, Jul 2021, Vienne, Austria. Proceedings of the Annual Conference of the Cognitive Science Society, 43, pp.2808-2814, 2021, Proceedings of the Annual Meeting of the Cognitive Science Society ; https://cognitivesciencesociety.org/cogsci-2021/ (2021)
|
|
BASE
|
|
Show details
|
|
25 |
Influence of Actor's Congruent and Incongruent Gaze on Language Processing ...
|
|
|
|
BASE
|
|
Show details
|
|
26 |
Integration of Social Context vs. Linguistic Reference During Situated Language Processing ...
|
|
|
|
BASE
|
|
Show details
|
|
27 |
An Eye Tracking Study of the Differential Effects of Mobile Phone, Tablet, Computer, and Paper Media on Incidental Second Language Vocabulary Acquisition and Second Language Reading Comprehension ...
|
|
|
|
BASE
|
|
Show details
|
|
28 |
Eye movement patterns of primary-school aged children with developmental dyslexia learning to read alphabetic orthographies ...
|
|
|
|
BASE
|
|
Show details
|
|
29 |
Onomatopoeia and gesture processing in bilingual preschoolers ...
|
|
|
|
BASE
|
|
Show details
|
|
30 |
The influence of verb tense on mental simulation during literary reading ...
|
|
|
|
BASE
|
|
Show details
|
|
31 |
Sample Stimuli Presentation for a Remote Speech Segmentation Study ...
|
|
|
|
BASE
|
|
Show details
|
|
32 |
Sample Stimuli Presentation for a Remote Speech Segmentation Study ...
|
|
|
|
BASE
|
|
Show details
|
|
33 |
Everyday Language Exposure Shapes Prediction of Specific Words in Listening Comprehension: A Visual World Eye-Tracking Study ...
|
|
|
|
BASE
|
|
Show details
|
|
34 |
The impact of facial expressions in consumer purchase decision
|
|
|
|
BASE
|
|
Show details
|
|
35 |
Investigating reliability and construct validity of a source-based academic writing test for placement purposes
|
|
|
|
In: Graduate Theses and Dissertations (2021)
|
|
Abstract:
Source-based writing, in which writers read or listen to academic content before writing, has been considered to better assess academic writing skills than independent writing tasks (Read, 1990; Weigle, 2004). Because scores resulting from ratings of test takers’ source-based writing task responses are treated as indicators of their academic writing ability, researchers have begun to investigate the meaning of scores on source-based academic writing tests in an attempt to define the construct measured on such tests. Although this research has resulted in insights about source-based writing constructs and the rating reliability of such tests, it has been limited in its research perspective, the methods for collecting data about the rating process, and the clarity of the connection between reliability and construct validity. This study aimed to collect and analyze evidence regarding the reliability and construct validity of a source-based academic English test for placement purposes, called the EPT Writing, and to show the relationship between these two parts of the study by presenting the evidence in a validity argument (Kane, 1992, 2006, 2013). Specifically, important reliability aspects, including the appropriateness of the rating rubric based on raters’ opinions and statistical evidence, the performance of the raters in terms of severity, consistency, and bias, as well as test score reliability, were examined. Also, the construct of academic source-based writing assessed by the EPT Writing was explored by analysis of the writing features that raters attended to while rating test takers’ responses. The study employed the mixed-methods multiphase research design (Creswell & Plano Clark, 2012) in which both quantitative and qualitative data were collected and analyzed in two sequential phases to address the research questions. In Phase 1, quantitative data, consisting of 1,300 operational ratings provided by the EPT Office, were analyzed using Many-Facets Rasch Measurement (MFRM) and Generalizability theory to address the research questions related to the rubric’s functionality, raters’ performance, and score reliability. In Phase 2, 630 experimental ratings, 90 stimulated recalls collected with assistance from records from eye-tracking technology, as well as nine interviews from nine raters were analyzed to address the research questions pertaining to raters’ opinions of the rubric and the writing features that attracted raters’ attention during rating. The findings were presented in a validity argument to show the connection between the reliability of the ratings and the construct validity, which needs to be taken into account in research on rating processes. Overall, the raters’ interviews and MFRM analysis of the operational ratings showed that the rubric was mostly appropriate for providing evidence of variation in source-based academic writing ability. Regarding raters’ performance, MRFM analysis revealed that while most raters maintained their comparability and consistency in terms of severity, and impartiality towards the writing tasks, some of them were significantly more generous, inconsistent, and biased against task types. The score reliability estimate for a 2-task x 2-rater design was found below the desired level, suggesting that more tasks and raters are needed to increase reliability. Additionally, analysis of the verbal reports indicated that the raters attended to the writing features aligned with the source-based academic writing construct that the test aims to measure. The conclusion presents a partial validity framework for the EPT Writing, in addition to implications for construct definition of source-based academic writing tests, cognition research methods, and language assessment validation research. Recommendations for the EPR Writing include a clearer definition of the test construct, revision of the rubric, and more rigorous rater training. Suggested directions for future research include further research investigating raters’ cognition in source-based writing assessment and additional validation studies for other inferences of the validity framework for the EPT Writing.
|
|
Keyword:
construct validity; eye tracking; mixed methods; reliability; source-based writing; validity argument
|
|
URL: https://lib.dr.iastate.edu/cgi/viewcontent.cgi?article=9580&context=etd https://lib.dr.iastate.edu/etd/18573
|
|
BASE
|
|
Hide details
|
|
36 |
The Role of Literal Features During Processing of Novel Verbal Metaphors ...
|
|
|
|
BASE
|
|
Show details
|
|
37 |
A cross-linguistic investigation of retroactive similarity-based interference in sentence comprehension. ...
|
|
|
|
BASE
|
|
Show details
|
|
38 |
Auditory distraction while reading in different languages ...
|
|
|
|
BASE
|
|
Show details
|
|
39 |
Children's processing of written irony: An eye-tracking study ...
|
|
|
|
BASE
|
|
Show details
|
|
40 |
Perceptual narrowing in face and speech perception during infancy ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|