Frontline Learning Research Vol.4 No.5 (2016)
1-33
ISSN 2295-3159
aJönköping University,
Jönköping, Sweden
bUniversidade Salgado de Oliveira, Rio de Janeiro,
Brazil
cInternational Baccalaureate, The Hague, Netherlands
dUniversity of Twente, Enschede, Netherlands
Article received 13 February / revised 5 April / accepted 25 September / available online 15 December
Qualitative research supports a developmental dimension in views on teaching and learning, but there are currently no quantitative tools to measure the full range of this development. To address this, we developed the Epistemological Development in Teaching and Learning Questionnaire (EDTLQ). In the current study the psychometric properties of the EDTLQ were examined using a sample (N= 232) of teachers from a Swedish University. A confirmatory factor and Rasch analysis showed that the items of the EDTLQ form a unidimensional scale, implying a single latent variable (eg epistemological development). Item and person separation reliability showed satisfactory levels of fit indicating that the response alternatives differentiate appropriately. Endorsement of the statements reflected the preferred constructivist learning-teaching environment of the response group. The EDTLQ is innovative since is the first quantitative survey to measure unidimensional epistemological development and it has a potential to be used as an apt tool for teachers to monitor the development of students as well as to offer professional development opportunities to the teachers.
Keywords: Rasch modelling, unidimensional development, epistemological development, good teaching, application, good textbook
In the post-industrial society, global issues such as climate
change, sustainable economic development and increasing wellbeing
worldwide have highlighted the need for education that promotes
the ability to think critically and for problem solvers who can
use conflicting evidence to understand and address complex or
so-called wicked problems. Such problem solvers need to be able to
tolerate ambiguity, take different perspectives and acquire new
knowledge throughout life, skills associated with higher order
thinking and complex meaning making (e.g. Kegan, 1994; Van Rossum
& Hamer, 2010), developed epistemic thinking (e.g. Barzilai
& Eshet-Alkalai, 2015), and flexible performance associated
with deep understanding (Perkins, 1993). The need for more of such
problem solvers underlies the current focus on characteristics and
skills currently referred to as 21st century skills which include
“communication skills, creativity, critical reflection (and
self-management), thinking skills (and reasoning), information
processing, leadership, lifelong learning, problem solving, social
responsibility (ethics and responsibility) and teamwork”
(Strijbos, Engels & Struyven, 2015, p20). Although one can
safely argue these skills are not unique or particularly novel to
the 21st century (e.g. Mishra & Kereluik, 2011; Voogt et al.,
2013), the consensus is that the complexity and global nature of
the major problems that confront humanity now and in the future
increased the need for this type of thinking and problem solving.
Formal education, and higher education (HE) in particular, is an
essential tool to ensure a society will include psychologically
mature citizens (e.g.Piaget, 1954; Kohlberg, 1984; Schommer, 1998)
with a well-developed ability to think critically. For this, it
needs to develop students’ understanding of how knowledge is
created, its scope and what constitutes justified beliefs and
opinions: in short it needs to focus on ensuring students’
epistemological development. However, currently many HE graduates
have not developed towards the lowest level of critical thinking
and reflective judgment in the sense of routinely using
established procedures and assumptions within one discipline or
system to evaluate knowledge claims and solve ill-structured
problems well (Kegan, 1994; King & Kitchener, 2004; Van Rossum
& Hamer, 2010; Arum & Roksa, 2011), let alone that such
graduates have developed a theory of knowledge that accommodates
differences in methods and procedures across systems and
disciplines that would be necessary to address the global issues
and wicked problems facing us. Indeed, a range of studies indicate
that education focusing on reproduction of established knowledge
and procedures can lead to poorer learning strategies (e.g. Newble
& Clarke, 1986; Gow & Kember, 1990; Linder, 1992) and
occasionally to regression in epistemic beliefs (Perry, 1970) and
decreased student well-being (e.g. Van Rossum & Hamer, 2010;
Lindblom-Ylläne & Lonka, 1999; Yerrick, Pedersen, &
Arnason, 1998). To increase education’s success rate, curricula at
all age levels would need to address and assess development in
students’ epistemic beliefs, choosing instruction methods that
emphasize the complexity of understanding, that stimulate
students' development of value systems and epistemic reflection.
Whilst literature points towards a relationship between teachers’
epistemological development and identity development, and their
success in encouraging similar development in their students (see
for a review Van Rossum & Hamer, 2010), as yet it is not easy
to establish which teachers are more successful in modelling
complex meaning making and so in teaching students complex value
systems.
With regard to the further life span, research on adult
development shows that epistemological development does not need
to end after adolescence or leaving formal education. Adults may
develop increasingly complex meaning and value systems (Commons,
1989, 1990; Demick & Andreoletti, 2003; Fischer & Pruyne,
2003; Hoare, 2006; Loevinger & Blasi, 1976), with convincing
evidence that the later phases of personality development lead to
higher levels of critical and meta systemic thinking (i.e. the
ability to see how different systems interact with each other),
responsibility, positive valuation of human rights, extended time
horizon and perspective consciousness (Commons & Ross, 2008a,
2008b; Kjellström & Ross, 2011; Sjölander, Lindstöm, Eriksson
& Kjellström, 2014; Kjellström & Sjölander et al, 2014).
A wealth of research exists describing students’ and teachers’
views on the nature and source of knowledge, measured both
quantitatively and qualitatively, proposing a unidimensional model
of epistemological sophistication (e.g. Perry, 1970; Baxter
Magolda, 1992, 2001; Kegan, 1994; Kuhn, 1991; see also Van Rossum
& Hamer, 2010 for a review linking qualitative and
quantitative models). Student and teacher views on the nature of
knowledge have been linked to views on learning, good teaching,
understanding and application (Van Rossum & Hamer, 2010). The
first major problem is that the existing measurements in this
field are time consuming both for students (participants) and
teachers (or researchers). The students need to be interviewed or
write long essays on dilemmas. Secondly, the researchers’ analysis
procedures are time consuming and/or require scoring skills that
need to be trained and often are only honed over time. Thirdly,
many of these models or qualitative empirical studies have not be
designed with a large enough sample size to capture the relatively
rare sophisticated epistemic perspectives and belief structures
(Kegan, 1994; Van Rossum & Hamer, 2010), for instance the
recent research regarding conceptions of understanding (e.g.
Irving & Sayre, 2013) has not progressed beyond the flexible
performance interpretation championed by Perkins (1993) in
Teaching for understanding, whilst at least two more complex
interpretations of understanding have now been empirically
observed (Hamer & Van Rossum, 2016). Existing quantitative
measures, such as the EBQ (Schommer, 1994), ILS (Vermunt, 1996) or
the dilemma or scenario based assessment tool explored by Barzilai
and Weinstock (2015), on the other hand, also do not include the
more sophisticated views on learning and knowledge (see for an
extensive discussion of the ILS and EBQ Van Rossum & Hamer,
2010, chapter 4).
Therefore, a questionnaire measuring the more sophisticated views
on teaching and learning concepts would be a valuable tool for
teachers to monitor the effectiveness of a curriculum designed to
encourage students’ cognitive identity development, complex
meaning making and their adoption of a complex value system.
Similarly, given that the more complex epistemological approach
the teacher has, the more likely it is that students develop
similar capabilities (Van Rossum & Hamer, 2010). Such a tool
would be useful to educational management to know when to offer
professional development opportunities to teachers, thereby
ensuring that their staff is indeed able to shape the
developmental educational environment necessary to support
students’ epistemological development to the higher levels of
complex thinking. However, as many of the current models and data
collection tools do not include these higher levels of epistemic
thinking, to construct such a tool requires a review of models
that have sufficient evidence and examples of student, teacher and
adult thinking reflecting these more elusive levels of complex
thinking.
In preparation towards developing such a questionnaire a review of
existing models was undertaken, merging evidence of student and
teacher thinking with that regarding adult development models
(e.g. Baxter Magolda, 1992, 2001; Belenky, Clinchy, Goldberger
& Terule, 1986; Dawson, 2006; Kegan, 1994; King &
Kitchener, 1994; Kuhn, 1991; Perry, 1970; Van Rossum & Hamer,
2010; West, 2004). From this review, the six stage developmental
model of learning-teaching conceptions developed by Van Rossum and
Hamer (2010) and expanded upon in a number of follow up studies
(Van Rossum & Hamer, 2012, 2013; Hamer & Van Rossum, 2016)
was selected as the primary source for constructing items and
scales designed to access a broader range of epistemological
development and through this psychological maturity within a
teaching and learning environment (see Table 1). An important
consideration here was that as of 2016, this model is supported by
the narratives of more than 1,200 students, including repeated
measurements and longitudinal student data evidencing the
developmental aspect of the model. Further, although it only
covers approx. 70 teacher narratives, the model is proposed to
model both student and teacher thinking (Richardson, 2012) and
would therefore also model adult thinking to some degree at least.
The result is the Epistemological Development in Teaching and
Learning Questionnaire (EDTLQ) that aims to measure stages of
development in the domain of teaching and learning. A number of
items (statements) was constructed, covering five issues/domains
in learning and teaching. Domains included views on ‘good’
teaching, understanding, application, good classroom discussion
(Van Rossum & Hamer, 2010; Hamer & Van Rossum, 2016) and a
good textbook (Van Rossum & Hamer, 2013). A sixth scale was
constructed based upon developmental responsibility research
(Kjellström, 2005; Kjellström & Ross, 2011) and unpublished
preliminary results from an ongoing longitudinal study on
responsibility for learning. The items were selected to represent
a sequence of learning and teaching conceptions, ranging from
level 2 (L2, Memorizing) to level 6 (L6, Growing self-awareness)
(Van Rossum & Hamer, 2010, see and Table 1), which the authors
believed should appeal in different ways to various teachers and
students as appropriate to their level of thinking within the
domain of epistemological development (see Table 2). The different
scales developed here refer to specific and shared experiences in
the teaching-learning environment, and so aim to address a
weakness in other measurement tools regarding “the use of abstract
and ambiguous words [or] items which are too general or vague to
allow for a consistent point of reference” (Barzilai &
Weinstock, 2015, p. 143). In the questionnaire items from each
scale were presented in a random order. The Swedish and English
version were created in a continuous process, were the items were
translated back and forth several times by native speakers among
the current authors.
The construction of different scales representing different
aspects of students’ views on the learning-teaching environment
may be perceived to point towards a multi-dimensional
interpretation regarding students’ epistemic beliefs and
epistemological development. Indeed traditionally, in developing a
tool to measure differences in epistemic beliefs, respondents
would be asked to indicate the extent to which each item reflected
their beliefs. Factor analysis of the responses ideally would then
identify which sets of items are strongly correlated, i.e. load
significantly on one or another factor. Each factor then would be
identified as representing a learning-teaching conception or level
of development. This approach underlies existing tools, such as
the ILS (Vermunt, 1996) and the more recently approach to measure
epistemic thinking by Barzilai and Weinstock (2015). In the EBQ
originally developed by Schommer (1990, 1994) and expanded upon by
a range of followers (e.g. Qian & Alverman, 2000; Qian &
Pan, 2002; Paulsen & Feldman, 2005; Bråten & Strømsø,
2004; Schreiber & Shinn 2003), the factors refer to (possibly
orthogonal) trajectories of epistemic development. The scales
developed in the current study are however proposed to represent
different contexts in which a student’s underlying epistemic
belief system will express itself. The items present descriptions
of aspects of different epistemological ecologies (Van Rossum
& Hamer, 2010): profiles or constellations of beliefs that are
closely linked and which change more or less simultaneously when a
respondent’s perspective shifts to a more complex epistemic level.
In this sense, Van Rossum and Hamer’s 2010 model is one of the
family of models that “suggest that there is … an underlying
trajectory explaining development across domains and topics [i.e.]
… a more global development” (Barzilai & Weinstock, 2015, p.
144), an assumption that is often disputed. The aim of this study
is to explore the assumption of an underlying unidimensional
development trajectory or “more global development” (Barzilai
& Weinstock, 2015).
Traditionally, respondents are asked to endorse each item as it is
presented if it is a part of their epistemic position, i.e. how
they view (their own) learning or experience teaching. In that
case, the hierarchical inclusiveness of the developmental model
underlying the item construction implies that items representing
less sophisticated views on learning would be endorsed by more
respondents, whilst at least in theory, the items reflecting the
more sophisticated levels are expected to be endorsed by fewer
respondents. In this sense the items reflecting the more
sophisticated levels would seem ‘more difficult’ to endorse.
However, participants completing the items of the EDTLQ were
instructed to rate the statements in accordance with how important
the statements were to teaching and learning on a five-point
ordinal scale, ranging from least important (1) to most important
(5). By rating the items to the respondents’ perception of most to
least importance to learning and teaching, the qualitative
empirical evidence used by Van Rossum and Hamer (2010; Hamer &
Van Rossum, 2016) points towards a different response pattern that
fits Kegan’s consistency hypothesis. Kegan (1994) states that
people strongly prefer to function at their highest level of
epistemic thinking and when circumstances prevent the expression
of this highest level they feel unhappy. The consistency
hypothesis then would predict that respondents will prefer items,
i.e. rate as most important or find easier to endorse, those that
reflect their own current epistemic beliefs, and will reject all
others. The items reflecting the lower levels of development,
which in a traditional approach would be relatively easy to
endorse, will now be rejected – i.e. difficult to endorse –
because they reflect a view on learning-teaching that the
respondent has moved beyond. The items that reflect thinking that
is more sophisticated than that of the respondent are rejected, as
they do not reflect a current aspect of the respondent’s thinking
about learning, teaching and knowing. This means that the choice
to ask respondents to rate the items by importance will lead to an
answer pattern that will reflect the level of development, a
learning-teaching conception, of the majority of the respondent
group.
The empirical qualitative evidence of over 1200 HE students and 70
HE teachers (Hamer & Van Rossum, 2016) suggests that a
respondent group of freshmen or sophomore HE students will result
in the endorsement of items on levels 2, 3 in a traditional
learning environment, or in a more constructivist
teaching-learning environment a preference for 3 and 4 level
items, with items reflecting level 1, 5 and 6 being rejected by
most respondents, i.e. being most difficult to endorse. Again
depending on the level of traditional or constructivist teaching
practices, a sample of teachers will either demonstrate a similar
pattern to students (in predominantly traditional teaching
environments or secondary school level), whilst in a predominantly
constructivist teaching environment respondents would find items
at levels 3, 4 and perhaps 5 to be most easy to endorse, rejecting
those reflecting level 1, 2 and 6. For teachers however there is a
complicating factor to interpreting their responses. Teacher
self-generated responses regarding good teaching or their
expectations regarding student learning are often seen to include
large quantities of academic discourse, occasionally to an extent
that it obscures the formulation of their own views on teaching or
learning (Säljö, 1994, 1997), implying that their self-generated
responses (e.g. in interviews or narratives) may reflect social
desirable answering patterns to a greater degree than that of
students. It is unknown to what extent similar teacher response
bias or patterns may express themselves in the EDTLQ.
Table 1
Epistemological profiles within Van Rossum & Hamer six
stage developmental model of learning-teaching conceptions
Table 2
EDTLQ items, codes and scoring levels according to Van Rossum
& Hamer (2010) and Hamer & Van Rossum (2016)
The aim of the current paper is to investigate the psychometric properties of EDTLQ via confirmatory factor analysis and the Rasch model for dichotomous responses, in order to answer the following questions:
The study was designed as an online survey involving both staff and students in higher education, with the aim to develop a quantitative tool to measure psychological maturity using a variety of existing and newly developed survey items. In this study we report on the results of survey among staff, i.e. university teachers and researchers. The analysis focuses on the psychometric properties of the newly developed EDTLQ with regard to the fit of the items and scales developed.
The participants comprise a convenience sample of staff which included teachers and researchers working in one of the four schools at Jönköping University: The School of Health and Welfare, the School of Education and Communication, Jönköping International Business School and the School of Engineering. The official language is Swedish at three of the schools and English at the remaining school. About two thirds of the respondents approached opened the questionnaire (N=340, 68%) resulting in 232 full responses.
Table 3
Participant demographics
The majority of the respondents were well established teachers: half of the respondents had 11 years or more teaching experience, and fifty-eight per cent were aged between 30 and 59 years old. Seven per cent have no, or less than a year teaching experience which could mean that they were recently employed, but all have studied at university level which means that all are familiar with education and learning concepts. As expected from university teachers and researchers, they were highly educated with almost half (48%) having achieved the Swedish licentiate, doctorate, associate or full professorship. More detailed demographics of the respondents are provided in Table 3.
The data were collected by two of the current authors in 2014. A web-based questionnaire was sent to all respondents by e-mail. The invitation e-mail included information about the study and a booklet about the study. The questionnaire and all supporting information were provided in both English and Swedish, with respondents given the option to responding to the survey in either language. Two reminders were sent to the participants.
All respondents were informed that participation was entirely voluntary, of their right to withdraw at any time, and the confidential treatment of their responses during analysis and in reporting. The study was approved by Regional Ethical Review Board in Linköping.
The full survey comprised two smaller sections: one covering socio-demographic information (i.e., gender, age, education), and the other comprising the EDTLQ. The items of the EDTLQ are presented in Table 2 above. Table 4 shows how the items are expected to cluster to represent the different levels of thinking or ways of knowing in order of complexity according to developmental theory and the model of Van Rossum and Hamer (2010, 2013; Hamer & Van Rossum, 2016).
Table 4
Items of the EDTLQ by assumed level of epistemological
development and complexity of thinking (Van Rossum & Hamer,
2010, 2013; Hamer & Van Rossum, 2016)
The developmental literature (Commons et al., 2008a, 2008b;
Dawson, Goodheart, Draney, Wilson, & Commons, 2010; Golino,
Gomes, Commons & Miller, 2014) uses the Rasch family of
statistical models to assess the quality of its instruments, as
well as the expected equivalence and order of items (Andrich,
1988; Rasch, 1960). The benefits of using the Rasch family of
models for measurement involve the construction of objective and
additive scales, with equal-interval properties (Bond & Fox,
2001; Embreston & Reise, 2000). It also produces linear
measures, gives estimates of precision, allows the calculation of
quality of fit indexes and enables the parameters’ separation of
the object being measured and of the measurement instrument
(Panayides, Robinson & Tymms, 2010). Rasch modeling enables
the reduction of all items of a test into a common developmental
scale (Demetriou & Kyriakides, 2006), collapsing in the same
latent dimension person’s abilities and item’s difficulty (Bond
& Fox, 2001; Embreston & Reise, 2000; Glas, 2007), as well
as enables verification of the hierarchical sequences of both item
and person, which is especially relevant to developmental stage
identification (Golino et al, 2014; Dawson, Xie & Wilson,
2003).
The simplest model is the dichotomous Rasch Model (Rasch,
1960/1980) for binary (i.e. one or the other) responses. It
establishes that the right/wrong scored response Xvi, depends upon
the performance β of that person and on the difficulty δ of the
item. Andrich (1978) extended the classic Rasch Model for
dichotomous responses to accommodate polytomous responses with the
same number of categories, also called the rating scale model
(RSM). The rating scale model constrains the category-intersection
distances (thresholds) to be equal across all items (Andrich,
1978; Mair & Hatzinger, 2007a, 2007b). In the rating scale
model the item difficulty is interpreted as the resistance to
endorsing a rating scale response category.
Although some authors point out that the Rasch models are the
simplest forms of Item Response Theory (IRT; Hambleton, 2000),
Andrich (2004) argues this is an irrelevant argument either way,
as IRT and Rasch models differ in nature and epistemological
approach. In IRT one chooses the model to be used (one, two or
three parameters) according to which better accounts for the data,
while in the Rasch paradigm the model is used because “it arises
from a mathematical formalization of invariance which also turns
out to be an operational criterion for fundamental measurement”
(Andrich, 2004, p. 15). Instead of data modeling, the Rasch’s
paradigm focuses on the verification of data fit to a fundamental
measurement criterion, compatible with those found in the physical
sciences (Andrich, 2004).
A confirmatory factor analysis was conducted prior to the rating
scale model, using the lavaan package (Rosseel, 2012), in order to
verify if the questionnaire presented an adequate fit to a
unidimensional model. Verifying the questionnaire dimensionality
is relevant given the unidimensionality assumption of the rating
scale model. The estimator used was the robust weighted least
squares (WLSMV). The data fit to the model was verified using the
root mean-square error of approximation (RMSEA), the comparative
fit index (CFI: Bentler, 1990), and the non-normed fit index
(NNFI: Bentler & Bonett, 1980). A good data fit is indicated
by a RMSEA shorter or equal than 0.05 (Browne & Cudeck, 1993),
or a stringent upper limit of 0.07 (Hu & Bentler, 1999), a CFI
equal to or greater than 0.95 (Hu & Bentler, 1999), and a NNFI
greater than 0.90 (Bentler & Bonett, 1980). The reliability of
the general factor was calculated using the semTools package
(Pornprasertmanit, Miller, Schoemann, & Rosseel, 2013).
In order to apply the rating scale model, the eRm package (Mair,
Hatzinger, & Maier, 2014) from the R software (R Core Team,
2013) was used. The R software is a free and open source system
for statistical computing and graphics creation that was
originally developed by Ross Ihaka and Robert Gentleman at the
Department of Statistics, University of Auckland, New Zealand
(Hornik, 2014). Being free and open source, makes R an attractive
tool to use as it facilitates replication and verification by
other researchers. Most importantly however is that it is
flexible: the user can perform various analyses, implement
different techniques of data processing, as well as generate
graphs of multiple types on a single platform to a greater degree
than in many commercial software packages.
The application of the rating scale model through the eRm package
(Mair, Hatzinger, & Maier, 2014) followed a four step
procedure described below. First, the dataset was subset in order
to exclude cases with missing data in all items. Secondly, the
rating scale model was applied using the RSM function of the eRm
package. Then, a graphic analysis was made to detect issues in the
categories. Andrich (2011) points out a large difference between
the traditional IRT paradigm and the Rasch paradigm when dealing
with rating scales. The former takes the ordering of the
categories for granted, while in the latter the order is a
hypothesis that needs to be checked (Andrich, 2011). If the
categories do not follow the hypothesized order (e.g.: reversed
categories C1 - C3 - C4 - C2), then they will need to be examined
further and improved experimentally (Andrich, 2011). Also, it is
necessary to identify whether each category has a probability of
being chosen (or marked) greater than the other categories in a
specific range of the latent continuum. Graphically, it means that
CX’s category characteristic curve cannot be contained inside any
other category curve. Finally, the fourth step in the rating scale
analysis involved checking the fit to the model. The
information-weighted fit (Infit) mean-square statistic was used.
It represents “the amount of distortion of the measurement system”
(Linacre, 2002. p.1). Values between 0.5 and 1.5 logits are
considered productive for measurement. Values smaller than 0.5 and
between 1.5 and 2.0 logits are not productive for measurement, but
do not degrade it (Wright, Linacre, Gustafson & Martin-Lof,
1994).
It is important to note that respondents were instructed to rate
the item statements in accordance with how important they were to
learning and teaching on a five-point ordinal scale, where 1 is
least important and 5 is most important. This affects the
interpretation of the item parameters: Items with high difficulty
estimates (i.e. with high resistance to endorsing a rating scale
response category) are those which fewer people considered
important to their views and opinions, while items with low
difficulty estimates (i.e. with low resistance to endorsing a
rating scale response category) are those that people considered
very important to their views. Thus, to reflect increasing
psychological maturity the interpretation of the latent dimension
must be reversed in the present study: from statements most
important to a person’s view (easier to endorse) to statements
least important to a person’s view (more difficult to endorse).
A first analysis focused on the first research question
regarding the items of the EDTLQ fitting a unidimensional model.
The confirmatory factor analysis showed that the unidimensional
model, with one latent variable (f1 = factor1) explaining the 36
questionnaire items, presented an adequate data fit (χ2 (594) =
816, p = 0.000, RMSEA = 0.05, CFI = 0.95, NNFI = 0.94). The
standardized factor loadings ranged from 0.16 (item un2) to 0.77
(item un4). The reliability of the general factor was 0.91
calculated using both the Cronbach’s alpha and the coefficient
omega (Raykov, 2001). The standardized estimates of the
unidimensional model were plotted using the semPlot package
(Epskamp, 2014).
The factor structure is represented in Figure 1 as a weighted
network, where the higher the standardized factor loading the
thicker and more saturated is the arrow. The items un2 (‘you know
something by heart’), di2 (‘when the teacher answers the students’
questions), un6 (‘able to answer test or exam questions
correctly'), ap1 (‘pass one’s exams’) and di6 (‘hear all the
different views that people have’) all have a factor loading
smaller than 0.3 and reflect the less sophisticated levels of
epistemological development (see Table 2). The items with the
largest factor loadings (> 0.55) in majority reflect the more
sophisticated levels of epistemological thinking of the model of
Van Rossum and Hamer (2010, 2013), i.e. di3 (‘different solutions
are illustrated from different perspectives’), bt3 (‘focus on
application on what is learnt’), un5 (‘realise how different
perspectives influence what you see and understand’), un4 (‘see
the underlying assumptions and how these lead to particular
conclusions’) and un3 (‘to see connections within a larger
context’).
Figure 1. Unidimensional factor structure of the Teaching and Learning Questionnaire
The relevance of conducting a confirmatory factor analysis prior
to the Rasch analysis lies in the assessment of the
unidimensionality assumption. This is one of the available
strategies, i.e. one can verify if the data fits a unidimensional
model using confirmatory factor analysis and then proceed to the
Rasch analysis.
The rating scale model results indicated an adequate fit of the
items to the unidimensional model (see Table 5), with mean infit
meansquare of 0.97 (Min = 0.69, Max = 1.30). These values are
within the range considered productive for measurement (Wright,
Linacre, Gustafson & Martin-Lof, 1994). The location parameter
of Table 5 is basically the item difficulty (i.e. the resistance
to endorsing a rating scale response category), while the
threshold are the points where the category curves intersect. So,
threshold 1 is the point in the latent variable where the category
1 (‘least important’, value = 0) intersects with category 2 (value
= 1; Figure 1). This means that people with ability or proficiency
in the latent variable equal to threshold 1 (-0.41) will have
equal probability of choosing item sb1 to reflect category 1 or
category 2.
Table 5
Item Infit, location and thresholds of items in the EDTLQ
Another indicator of the instrument quality is the person separation reliability and the items separation reliability. Both have the same interpretation as the reliability calculated using Cronbach's alpha: The closer to one, the greater the reliability of the measurement. With regard to the interpretation this means the closer the value to 1, the better the pattern of people’s responses, or items endorsements fits the measurement structure (Hibbard, Collins, Baker & Mahoney, 2009). In other words, the separation reliability of people indicates how sure we can be that a person with an estimated ability (in this case perhaps it is better to use affinity to the item) of 2 has, indeed, a greater ability (i.e. affinity to the item) than another person who has an estimated ability (i.e. affinity to the item) of 1. Similarly, the items separation reliability indicates the confidence that an estimated item difficulty 2 has, indeed, a greater difficulty (i.e. lower endorsement level) than another item with estimated difficulty of 1. Both are calculated using a relationship between the standard error’s variance and the mean square error (MSE):
The separation reliability of the EDTLQ items was 0.99 and the separation reliability of persons was 0.87.
Figure 2. Category characteristic curves for items (statements) from the question “a good study book”. Category one = black line; category two = red line; category three = green line; category four = blue line; category five = light blue line
Figure 2 shows the category characteristic curves for each item (statement) of the “a good study book” question. The hypothesized ordering of the categories matches the empirical evidence that the first category (value = 1, “least important”) is less difficult to endorse than the second category (value = 2), which is less difficult to endorse than the third category (value = 3), and so on, until the fifth category, the most difficult to endorse (value = 5, “most important”; see also Table 2). Each category, in every item, presents a greater probability of being endorsed than the other categories in at least a small range of the latent variable. Thus, there was no issue related to the ordering of the scales. The same scenario was found in every item of the questionnaire, as can be verified in Table 5. This result means that the response categories of the questionnaire (ranging from “least important” to “most important”) are empirically supported, and do not need to be altered. Figure 3 shows the parameter distribution of all the items in the EDTLQ. In this figure items ap4 is the easiest to endorse, followed by bt2, un3 and a small cluster of un5, un4, di3, sb4, re2 and re6. Theoretically these items would mostly reflect epistemological level 4 and 5, implying that these levels of epistemological thinking regarding application, good teaching, understanding and discussion correspond closest to what the respondents in the current study feel is important in learning and teaching. At the other end of the scale, the most difficult to endorse statements reflect both the least and the most sophisticated ways of thinking, including the items
Figure 3. Person-item map, all items of the EDTLQ
This seems to point towards a preference of the respondent group (teachers and staff) as a whole for epistemological thinking at level 4 and to a lesser degree 5, and a rejection of both the most sophisticated level 6, as well as the least sophisticated levels 1 and 2, matching the expected pattern under the consistency hypothesis. To examine in more detail if this pattern is consistent for all the scales, each scale is plotted and discussed separately below. First we will discuss four of the six scales that seem to behave in a way that confirms the consistency hypothesis where items reflecting the extremes of epistemic thinking are relatively difficult to endorse. These are the scale regarding discussion (di) – discussed in more detail below – and the scales regarding application (ap) and understanding (un) and good teaching (bt). Figure 4 shows the parameter distribution of the items regarding discussions during a course. Considering only the discussion during course items, item di3 (“different solutions are illustrated from multiple perspectives”) was the least difficult to endorse, while item di2 (“the teacher answers students’ questions”) was the most difficult to endorse. Table 6 links the discussion in course items ordered by the difficulty parameter of the Rasch model to the respective epistemological development level of the items, as per Van Rossum and Hamer’s theory. In this scale the items reflecting the least sophisticated levels of thinking (di1, di6 and di2) are on average the most difficult to endorse, reflecting the pattern found for all the items together.
Figure 4. Person-item map, subset of the discussion items (di).
Table 6
Discussion in course items by order of endorsement and
epistemological level (Van Rossum & Hamer, 2010).
A similar pattern can be seen for the items regarding being able to apply knowledge (Figure 5). Items ap4, ap2 and ap5 (reflecting levels 4 and 5) are the least difficult to endorse, while those items reflecting the least sophisticated levels (ap1, ap3; level 2 and 3) and ap6 reflecting the most sophisticated epistemic thinking level were the most difficult to endorse, again reflecting the pattern expected under the consistency hypothesis. The third scale that displays the response pattern expected under the consistency hypothesis is the scale with items regarding the “to understand something means to” issue (Figure 6). Here we see again that the items reflecting higher mid-level epistemic thinking (un3, un5 and un4, reflecting respectively level 4 and level 5 twice, see Table 2) are relatively easy to endorse, whilst those reflecting the less sophisticated epistemological positions (un1, un6 and un2, reflecting respectively level 3, 2 and 1) are rejected, i.e. difficult to endorse.
Figure 5. Person-item map, subset of the application items (ap)
Figure 6. Person-item map, subset of the understanding items (un)
Finally, Figure 7 shows the parameter distribution of the fourth scale that seems to behave as expected – the items regarding the issue “The best teaching ought to …”. Item bt2 (“inspire the students so that they are motivated to learn”) was the least difficult to endorse, while item bt6 (“make sure that the whole content of the course is covered”) was the most difficult to endorse. Examining the theoretically linked level of epistemic thinking to the level of endorsement (Table 7), again we see that those items reflecting both the most sophisticated as well as the least sophisticated levels of Van Rossum and Hamer’s model are difficult to endorse, a pattern that is as expected given the rank ordering of the items in combination with the consistency hypothesis.
Figure 7. Person-item map, subset of the best teaching items (bt).
Table 7
Best teaching items by order of endorsement and epistemological
level (van Rossum & Hamer, 2010)
The two remaining scales, reflecting the views on a good study book (Figure 8 and Table 8) and responsibility for learning (Figure 9 and Table 9) both do not seem to fit the expectation of the consistency hypothesis.
Figure 8.Person-item map, subset of the good textbook items (sb)
Table 8
Good text book items by order of endorsement and
epistemological level (Van Rossum & Hamer, 2013)
In examining the good study book items, Figure 8 shows that item sb4 (“evokes a critical attitude and invites reflection”) was the least difficult to endorse, while item sb2 (“provides many up-to-date examples and examples from practice”) was the most difficult to endorse. Examining the level of endorsement by theoretically expected level of epistemic thinking (Table 8) the parameters seem to indicate a clustering inconsistent with the consistency hypothesis. Level 4 thinking represents what the respondent group feels is most important in a good textbook the closest, closely followed by a preference for clear structure (level 2) and clarification of alternative perspectives (level 5). Study books that challenge current thinking and require reflection on the nature of knowledge seem to be more difficult to endorse, although books with many examples from practice (level 3) also seem to relatively unimportant to the teachers and researchers in this respondent group. The items regarding the nature of a good study book (Van Rossum & Hamer, 2013) are based on student responses only, and have not been similarly confirmed in multiple studies as views reflected in the previous four scales on discussion (di), application (ap), understanding (un) and good teaching (bt). The final scale is shown in Figure 9 which gives the parameter distribution of the items regarding the issue “For students to take responsibility for learning means that students…”. Item re2 (“have an active interest in the course and be motivated to learn”) was the least difficult to endorse, while item re4 (“follow their own interests in searching for knowledge”) was the most difficult to endorse. Table 9 shows the responsibility items ordered by the difficulty parameter of the Rasch model and the respective epistemological development level of the items, as per Van Rossum and Hamer’s theory. Item re2 and re6, reflecting “an active interest and motivation to learn” and “fitting what is learned into prior knowledge” are easiest to endorse for the respondent group, whilst items that refer to personal reflection, planning and interests seem to be difficult to endorse and thus seem less important for the teachers and researchers in the respondent group. The response pattern of this scale also does not fit with the expectation under the consistency hypothesis.
Figure 9. Person-item map, subsetof the responsibility for learning items (re)
Table 9
Responsibility items by order of endorsement and assumed
epistemological level
*) Please note: the order of these items are not empirically
derived but based developmental research literature and work on
responsibility issues (Kjellström, 2005; Kjellström & Ross,
2011)
For the overall picture we return to Figure 3 which shows the
person-item map including all items, sorted by relative
endorsement difficulty, we can construct a similar table for all
the items together. The figure shows a level wide range of
endorsement with infit measures indicating good fit to the
unidimensional Rasch model. The distribution of the levels of the
statement is presented in Table 10, and shows that when all
statements are taken together, levels 4 and 5 are the easiest to
endorse and levels 1 and 2 the most difficult. This response
pattern implies that the preferred level of epistemic thinking for
this group of teachers is constructivist (level 4 and 5) and
teaching by telling and focusing on recall and reproduction is
actively rejected.
Table 10
All statements by order of endorsement, based upon figure 6
Summarizing the results of the analysis of psychometric
properties of the six scales comprising the EDTLQ, the
confirmatory factor analysis supports the assumption that the six
scales together represent a single latent developmental dimension
underlying epistemological development that explains the
variability in the items, with sufficient infit statistics for all
the items. If the assumption was incorrect, the analysis may have
resulted in up to six dimensions, one for each issue covered in a
scale of the EDTLQ. As this was not the case, the first research
question can be answered positively. Examining the factor
parameters, with items representing the least sophisticated way of
thinking clustering at one end of the dimension and those
reflecting more sophisticated epistemological thinking at the
other end, the results further seem to imply that the underlying
dimension is somewhat similar to the surface – deep level
dimension discussed in earlier works regarding learning strategies
(e.g. Entwistle & Ramsden, 1983; Van Rossum & Schenk,
1984).
The rated statements comprising the EDTLQ have satisfactory levels
of separation, meaning that the items are characterized by clearly
different levels of endorsement and therefore separate the
respondents to an acceptable degree, answering the second research
question positively. The EDTLQ seems to be a satisfactory tool to
measure epistemological views on teaching and learning. The
analysis of the category characteristic curves for all of the
items in the EDTLQ indicates that there are no response categories
that completely overlap, nor are category curves in an
unanticipated sequence, indicating that it is not necessary to
change the number or sequence of the response categories per item.
The design decision to offer a five-point rating scale ranging
from “least important” to “important” offers clearly separated
response categories. This means that the third research question
of this study results in a positive answer as well.
The at first glance puzzling finding from a unidimensional
epistemological development perspective, is the fact that the
response group of university staff as a whole endorse relatively
sophisticated ways of thinking, namely the items reflecting
epistemological development levels within the constructivist
learning paradigm (levels 4 and 5, see Van Rossum & Hamer,
2010). This result indicates that for this response group, the
items reflecting constructivist learning reflect their views on
what is important in learning and teaching the best. As discussed
in the introduction, a hierarchical inclusive model would predict
that items reflecting less sophisticated levels of thinking would
be easier to endorse than those reflecting sophisticated ways of
thinking, and items reflecting the most sophisticated levels of
thinking being endorse by the fewest of the response group.
However, contrary to expectations, in this study it is the items
that reflect the least sophisticated levels of thinking that are
rejected, i.e. are not endorsed, by the larger majority of the
response group. Whilst this outcome seems to indicate that the
underlying model of epistemological development is incorrect, and
so implies that answer to the fourth research question may be
negative, there is reason to believe that the response
instructions – to indicate how well the statement reflected an
important aspect of learning and teaching – led to a different
response pattern. As introduced before, Kegan (1994) discusses the
consistency hypothesis and states that when circumstances prevent
respondents to express their highest level of epistemic beliefs,
the pressure felt to “regress” is experienced as unhappiness
because “we do not feel ‘like ourselves’” (Kegan, 1994, p. 372).
Something similar is discussed by Van Rossum & Hamer (2010)
when they describe what they refer to as Disenchantment and its
counterpart Nostalgia. Disenchantment refers to the feeling of
disillusionment that students experience when they are exposed to
a teaching-learning environment characterized by significantly
less epistemological sophistication than their own. This can lead
to rebellion or despair (e.g. Yerrick, Pedersen & Arnason,
1998; Lindblom-Ylänne & Lonka, 1999). Nostalgia refers to the
wistful hankering to a more traditional learning-teaching
environment expressed by students when they are over asked and
required to function at an epistemic level too far beyond their
understanding (Van Rossum & Hamer, 2010, pp 415 - 426).
Considering the consistency hypothesis, and the request to rate
items to reflect the most important level of epistemological
development in learning and teaching, it is then in fact not
surprising to see that respondents reject items reflecting less
sophisticated ways of thinking. Indeed, in Van Rossum and Hamer
(2010) many examples are given of students expressing their active
rejection of a way of knowing or perception of learning that they
feel they have outgrown. Something similar may be taking place in
the mind of teachers. However, this does mean that the EDTLQ
results of a group reflect the epistemological sophistication
level that the majority of the respondent group feel is important
to learning and teaching, and when used in a way similar to here
does not in fact reflect the epistemological development of a
single respondent. To establish what individual respondents feel
is important to learning and teaching, reflecting the different
preferred levels of epistemological development present in a
classroom or lecture hall, will require additional study and
analysis, and perhaps different response instructions to those
completing the questionnaire.
Examining the six scales of the EDTLQ introduced here, which
reflect six different contexts in which an underlying epistemic
development may express itself, it seems that four of these five
scales behave as can be expected under the consistency hypothesis.
These are the scales referring to discussion, application,
understanding and good teaching. The scales for views on a good
study book and the responsibility for learning scale do not seem
to fit the expected pattern.
In the case of the good study book scale, there are two possible
explanations that need to be considered and require further study.
Firstly, the underlying empirical data for this scale is
relatively recent and the analysis of the original data has not
been confirmed by further data. Secondly, this scale may provide
some insight into the extent to which the teacher responses in
this study may be affected by discourse and social desirability.
In the pedagogical literature aimed at teachers, significant
attention is given to describing constructivist interpretations of
application, understanding, the role of discussion in arriving at
new knowledge and perspectives and characteristics, providing a
rich vein of discourse potentially affecting teacher responses. To
the current authors knowledge there is very little of similar
discourse available defining characteristics of a good study book.
If so, the mixed pattern may reflect the teachers’ epistemic
thinking more faithfully than the other scales.
Regarding the responsibility scale, again there are at least two
reasonable explanations. Firstly, the scale was developed by the
use developmental responsibility research from partly another
domain (namely responsibility for health) which may illustrate how
difficult it is to transfer ideas in one domain to another, and
that there may be domain specific patterns (Fischer & Pruyne,
2003). Secondly, and perhaps more importantly, the responsibility
items were allocated developmental levels using a model to which
they have no empirical link. Hence, the items were interpreted by
two of the authors and were allocated a developmental level that
may not reflect the actual position on the assumed epistemological
development trajectory. Without empirical evidence linking the
responsibility items to any of the other scales, it is less clear
which level to assign as previous research has shown that a
particular statement may be attractive to people different stages
of development (Ross, 2008; Loevinger & Hy, 1994).
Remains then the issue that the respondent group of teachers seems
to prefer items reflecting relatively sophisticated levels of
epistemological development that is usually not observed to this
degree and is quite sophisticated even for teachers in higher
education (for a review of literature see Van Rossum & Hamer,
2010 chapter 5, 2012). Whilst this seems to indicate that the
teachers in this study are characterized by a relatively
sophisticated view on learning and teaching, there is another
possible explanation: the interpretation of the instructions for
rating the statements. Where the example used to illustrate how
respondents were to rate the statements referred to a personal
view (‘I know I have done a good job when...’) the items of the
five scales were formulated in more general way (‘Discussion
during a course are good when …’). This means that we cannot be
sure the instruction was interpreted by the teachers as a request
to rate each statement in importance to their personal view on
learning, and therefore that their ratings reflect their personal
preference. It is entirely possible that the results reflect ‘more
socially constructed discourse and less personal … understanding’
(Van Rossum & Hamer, 2010, p. 449). In other words,
qualitative responses of teachers are often thick with socially
acceptable discourse that at times obscures their personal
understanding or preference of the teaching-learning environment
(Säljö, 1994, 1997). The pilot of the EDTLQ was performed in
Sweden, where the official written curricula focus on skills such
as discussing, analyzing and synthesizing rather than on facts and
learning by heart. The request to rate statements with regard to
their importance to learning and teaching in general may well have
activated a preference for items reflecting these written and
therefore socially acceptable curricula, above a preference for
items reflecting their own beliefs or the reality of the taught
curricula.
If this latter interpretation is correct, reformulating the
instructions towards an importance to a respondent’s own teaching
practice may result in a different outcome. Based on theory and
earlier experimental findings (Van Rossum & Hamer, 2012,
2013), the findings for the teachers as a group may move slightly
towards levels 3 and 4. In addition, the results for students
would reflect a significantly less sophisticated preference. As
students are less well versed in the academic discourse with
regard to teaching, the ambiguous instructions may not have the
same effect, which means that the preferred items for the freshmen
students in particular would differ considerably with a
significant shift towards items reflecting levels 2 and perhaps 3.
Although the results of this study are promising, further analyses
as well as improvements to the survey, in particular to the rating
instruction, and additional data collection are necessary to
definitively answer the final research question in this study.
The current study indicates that the unidimensional Rasch model fits the set of scales regarding epistemological preference for a group of university staff. A similar analysis needs to be undertaken for the two groups of students that participated, freshmen and senior (third year undergraduate) students separately, as these two groups may not interpret the scale items in the same way as the teachers and researchers who participated in this study. To establish if the EDTLQ is a viable tool to measure individual epistemological development, further analysis is required to identify specific response patterns for individual participant or participant groups. To clarify if the results reflect the personal preference in teaching in learning it is necessary to clarify the instructions regarding the rating of the statements. Whilst the categories ranging from least important to most important work well, the instruction above each set of items should include a reference to this personal preference, e.g. “Please rate each statement from 1 (least important) to 5 (most important) to your own personal views regarding an ideal learning and teaching environment.” Remains the issue of the potential social desirable bias in the teacher responses. When surveying teachers in particular, using inventory statements it will remain difficult to disentangle the use of social acceptable discourse from personal understanding, and as such it is recommended to supplement the EDTLQ with other data collection methods that provide more opportunity for teachers to structure their own response and so provide a more clear window into their personal views. An interesting finding is that the response group of university teachers seems to be attracted to the more advanced statements, and we probably would not have the same results with high school teachers who may prefer items that reflect levels 2 and 3 (e.g. Brickhouse, 1989, 1990; Martens, 1992; Maor and Taylor, 1995; Hashweh, 1996) which fits the experience in the adult development research field, but hitherto has not been shown statistically. In this sense this result is the most interesting as it enhances the knowledge base within adult development regarding how the statements are endorsed within each item and the whole collection of items. Whilst the main finding, that the items that are easiest to endorse within each scale reflect the midrange levels of epistemological sophistication of the six stage developmental model of learning-teaching conceptions developed by Van Rossum and Hamer (see Table 10) seems contradictory for readers not familiar with adult development theories, both the consistency hypothesis (Kegan, 1994) and the experience in adult development of people endorsing values at their own stage and those slightly more advanced would, is, given the respondent group, exactly what one might expect. On a more detailed level, there may be issues with individual items or scales. In Table 10 we see that level 4-5 statements are most easily endorsed. Examining the individual items more closely, it is interesting to note in this table is that among the level 5 statements some seem easy to endorse and others more difficult. Three statements that are all about taken perspectives are easily endorsed: “realize how different perspectives influence what you see and understand”, “see the underlying assumptions and how these lead to particular conclusions”, and “different solutions are illustrated form multiple perspectives”. Two level 5 statements that require a more critical awareness and questioning, are more difficult do endorse: “make students questions current view of the world” and “help students realize the limits of current knowledge and understanding”. A possible explanation would be that the easier to endorse items can both be endorsed by respondents viewing learning and teaching from a multiplistic perspective (L3) where all opinions are equally valid (Perry, 1970), and by respondents viewing learning and teaching from a relativist perspective (L5), where opinions can differ in validity and in range of supporting evidence, but judging opinions is less acceptable, as in this way of thinking the focus is not on uncovering the most valid opinion, but on empathically understanding the point of view of others (Van Rossum & Hamer , 2010). This means that the easily endorsed items that describe multiple perspective taking need amending, to make them more clearly either level 3 or level 5. Further, there are some individual items that seem to not fit the model. In particular this is the case for the items regarding the responsibility for learning in particular. As is indicated above, these items were not developed using empirical data from Van Rossum and Hamer (2010). These items seem to reflect the Swedish written curriculum and as such are probably part of the socially acceptable discourse for the respondents in this study. In addition, some items are currently assigned to the lowest level where they apply, e.g. re5 “are aware of their strengths, monitoring their progress and making adjustments when necessary” reflecting meta-cognition and self-monitoring, whilst these items could well be endorsed by respondents with more sophisticated levels of thinking. Finally, there could be cultural differences in understanding the items. Van Rossum and Hamer collected their data from Dutch students. Although in their 2010 review, Van Rossum and Hamer explicitly address cultural differences with regard to their model, the understanding of individual items could of course differ. Such cultural differences could include the afore mentioned emphasis in the written curricula on skills such as discussing, analyzing and synthesizing rather than focusing on facts and recall, but perhaps there are other less obvious cultural influences that will only become manifest if the EDTLQ is piloted in other countries and with other respondent groups. An obvious choice given the countries represented by the authors would be an English language pilot, and perhaps a translation into Portuguese. In addition, respondent groups could include teachers and students in secondary education.
Andrich, D. (1978). A rating formulation for ordered
response categories. Psychometrika 43(4), 561–574.
Andrich, D. (1988). Rasch Models for Measurement.
Newbury Park, CA: Sage.
Andrich, D. (2004). Controversy and the Rasch model: A
characteristic of incompatible paradigms? Medical Care,
42(1), 7 – 16.
Andrich, D. (2011). Rating scales and Rasch measurement. Expert
Review Pharmacoeconomics Outcomes Reserach, 11(5),
571–585. doi: 10.1586/erp.11.59
Arum, R., & J. Roksa, (2011). Academically adrift:
Limited learning on college campuses. Chicago, city, 2
letter state: University of Chicago Press.
Barzilai, S., & Eshet-Alkalai, Y. (2015). The role of
epistemic perspectives in comprehension of multiple author
viewpoints. Learning and Instruction, 36, 86 – 103.
doi: 10.1016/j.learninstruc.2014.12.003
Barzilai, S., & Weinstock, M. (2015). Measuring epistemic
thinking within and across topics: A scenario-based approach,
Contemporary Educational Psychology, 42, 141-158.
doi: 10.1016/j.cedpsych.2015.06.006
Baxter Magolda, M.B. (1992). Knowing and reasoning in
college. San Francisco, CA: Jossey-Bass.
Baxter Magolda, M.B. (2001). Making their own way.
Sterling, VA: Stylus Publishing.
Belenky, M.F., Clinchy, B.M., Goldberger, N.R., & Tarule,
J.M. (1986 ed. 1997). Women’s ways of knowing: The
development of self, voice and mind. New York: Basic
Books.
Bentler, P.M. (1990). Comparative fit indexes in structural
models. Psychological Bulletin, 107(2), 238-246.
doi: dx.doi.org/10.1037/0033-2909.107.2.238
Bentler, P.M., & Bonett, D.G. (1980). Significance tests
and g.oodness of fit in the analysis of covariance structures.
Psychological Bulletin, 88, 588–606. doi:
10.1037/0033-2909.88.3.588
Bond, T. G., & Fox, C. M. (2001). Applying the Rasch
model: Fundamental measurement in the human sciences.
Mahwah, NJ US: Lawrence Erlbaum Associates Publishers. doi:
10.1111/j.1745-3984.2003.tb01103.x.
Bråten, I. and Strømsø, H.I. (2004). Epistemological beliefs
and implicit theories of intelligence as predictors of
achievement goals. Contemporary Educational Psychology, 371-388.
doi: 10.1016/j.cedpsych.2003.10.001
Brickhouse, N.W. (1989). The teaching of the philosophy of
science in secondary classrooms: case studies of teachers
personal theories. International Journal of Science
Education, 11(4), 437-449. doi:
10.1080/0950069890110408.
Brickhouse, N.W. (1990). Teachers’ Beliefs About the Nature of
Science and Their Relationship to Classroom Practice. Journal
of Teacher Education, 41(3), 53-62. doi:
10.1177/002248719004100307.
Browne, M.W., & Cudeck, R. (1993). Alternative ways of
assessing model fit. In: Bollen, K. A., Long, J. S. (Ed.)
Testing structural equation models (pp.136 -162). Newbury
Park: Sage.
Commons, M.L., & Ross, S. N. (2008a) Special Issue:
Postformal Thought and Hierarchical complexity. World
Future: Journal of General Evolution, 64(5-7): p.
297-562. Doi: 10.1080/02604020802301121.
Commons, M.L., & Ross, S. N. (2008b), What postformal
thought is, and why it matters. World Future: Journal of
General Evolution, 64 (5-7): p. 321-329.
doi: 10.1080/02604020802301139.
Commons, M.L. (1989) Adult development. Vol. 1,
Comparisons and applications of developmental models.
New York: Praeger. xii.
Commons, M.L. (1990) Adult development. Vol. 2, Models
and methods in the study of adolescent and adult thought.
New York: Praeger.
Dawson, T.L. (2006), The meaning and measurement of conceptual
development in adulthood. In C.H. Hoare Handbook of adult
development and learning, Oxford University Press:
Oxford. p. 433-454. Doi: 10.1558/jate.v6i1.98
Dawson, T., Goodheart, E., Draney, K., Wilson, M., &
Commons, M. (2010). Concrete, abstract, formal, and systematic
operations as observed in a 'Piagetian' balance-beam task
series. Journal of Applied Measurement, 11(1),
11-23.
Dawson, T. L., Xie, Y., & Wilson, M. (2003).
Domain-general and domain-specific developmental assessments:
Do they measure the same thing? Cognitive Development, 18,
61-78. Doi: 10.1016/S0885-2014(02)00162-4.
Demick, J., & Andreoletti, C. (2003) Handbook of Adult
Development. The Springer Series in Adult Development and
Aging. 2003, New York: Springer.
Demetriou, A., & Kyriakides, L. (2006). The functional and
developmental organization of cognitive developmental
sequences. British Journal of Educational Psychology, 76(2),
209-242. doi: 10.1348/000709905X43256
Embretson, S.E., & Reise, S. P. (2000). Item response
theory for psychologists. London: Erlbaum.
Entwistle, N.J., & Ramsden, P. (1983). Understanding
Student Learning. New York, NY: Nichols.
Epskamp, S. (2014). semPlot: Path diagrams and visual analysis
of various SEM packages' output. R package version 1.0.1. http://CRAN.R-project.org/package=semPlot
Fischer, K.W., & Pruyne, E, (2003) Reflective thinking in
adulthood: Emergence, development, and variation. In J. Demick
& C. Andreoletti Handbook of adult development,
Kluwer Academic/Plenum: New York. p. 169-198.
Glas, C.A. (2007). Multivariate and Mixture Distribution
Rasch Models. New York: Springer-Verlag.
Gow, L., & D. Kember, D. (1990). Does higher education
promote independent learning? Higher Education, 19,
p. 307-322. doi: 10.1007/BF00133895.
Golino, H.F., Gomes, C.M.A., Commons, M.L., & Miller, P.
(2014). The construction and validation of a developmental
test for stage identification: Two exploratory studies. Behavioral
Development Bulletin, 19, 37-54. Doi: 10.1037/h0100589
Hambleton, R. K. (2000). Response to Hays et al and McHorney
and Cohen: Emergence of Item Response Modeling in Instrument
Development and Data Analysis. Medical Care, 38(9),
II-60.
Hamer, R., & Van Rossum, E.J. (2010). De zes talen in het
onderwijs: communicatie en miscommunicatie [The six languages
in education: communication and miscommunication]. Onderwijsvernieuwing,
17 (November).
Hamer, R., & van Rossum, E.J. (2016). Students’ conception
of understanding and its assessment. In E. Cano & G. Ion
(Eds.), Innovative Practices for Higher Education
Assessment and Measurement. (pp. 141-162). Heshey, PA:
IGI Global.
Hasweh, M.Z. (1996). Effects of Science Teachers’
Epistemological Beliefs in Teaching. Journal of Research
in Science Teaching, 33(1). 47-63.
Hibbard, J., Collins, P., Mahoney, E., & Baker, L. (2010).
The development and testing of a measure assessing clinician
beliefs about patient self-management. Health
Expectations: An International Journal of Public
Participation in Health Care & Health Policy,
13(1), 65 72. Doi: 10.1111/j.1369-7625.2009.00571.x.
Hoare, C.H., (2006). Handbook of adult development and
learning. Oxford, UK: Oxford University Press. xviii, 579 p.
Hornik, K. (2014). The R FAQ. Retrieved from
http://CRAN.R-project.org/doc/FAQ/R-FAQ.html
Hu, L.T., & Bentler, P.M. (1999). Cutoff criteria for fit
indexes in covariance structure analysis: conventional
criteria versus new alternatives. Structural Equation
Modeling, 6(1), 1-55. doi: 10.1080/10705519909540118.
Irving, P.W., & Sayre, E.C. (2013). Upper-level Physics
Students’ Conceptions of Understanding. Paper presented at
2012 Physics Education Research Conference. In AIP
Conference Proceedings (vol. 1513, pp. 98-201).
Doi:10.1063/1.4789686.
Kegan, R. (1994). In over our Heads: The mental demands
of modern life. Cambridge, MA: Harvard University
Press.
King, P.M., & Kitchener, K.S. (1994). Developing
Reflective Judgment. San Fransisco, CA: Jossey-Bass
Publishers.
King, P.M., & Kitchener, K.S. (2004). Reflective judgment:
Theory and research on the development of epistemic
assumptions through adulthood. Educational Psychologist,
39(1), p. 5-18. doi: 10.1207/s15326985ep3901_2.
Kjellström, S. (2005). Responsibility Health and The
Individual [Ansvar, hälsa och människan – en studie av idéer
om individens ansvar för sin hälsa]. Linköping Studies in Arts
and Science, nr 318. Linköping University.
Kjellström, S. & S.N. Ross (2011). Older Persons’
Reasoning about Responsibility for Health: Variations and
Predictions”, International Journal of Aging and Human
Development, 73(2), 99-124.
Kjellström, S. & Sjölander, P. (2014). The level of
development of nursing assistants’ value system predicts their
views on paternalistic care and personal autonomy. International
Journal of Ageing and Later Life, 9(1), 35-68. Doi:
10.3384/ijal.1652-8670.14243
Kohlberg, L. (1984). Essays on moral development: Vol. 2. The
psychology of moral development San Francisco: Harper &
Row.
Kuhn, D. (1991). The Skills of Argument. Cambridge,
UK: Cambridge University Press:
Linacre J. M. (2002). What do infit and outfit, mean-square
and standardized mean? Rasch Measurement Transactions, 16 (2),
878.
Lindblom-Ylänne, S., & Lonka, K. (1999). Individual ways
of interacting with the learning environment – are they
related to study success? Learning and Instruction, 9,
1-18. doi: 10.1016/S0959-4752(98)00025-5.
Linder, C.J. (1992). Is teacher-reflected epistemology a
source of conceptual difficulty in physics? International
Journal of Science Education, 14(1), 111-121. doi:
10.1080/0950069920140110.
Loevinger, J., & Blasi, A. (1976). Ego development:
Conceptions and theories (1. ed.). San Francisco: Jossey-Bass.
Loevinger, J., & Hy, L. X. (1996). Measuring ego
development (2. ed.). Mahwah, N.J.: L. Erlbaum
Associates.Francisco: Jossey-Bass.
Mair P., & Hatzinger, R. (2007a). Extended Rasch Modeling:
The eRm package for the application of IRT models in R. Journal
of Statistical Software, 20(9), 1–20.
Mair P., & Hatzinger, R. (2007b). CML based estimation of
extended Rasch models with the eRm package in R. Psychology
Science, 49, 26–43.
Mair, P., Hatzinger, R., & Maier, M.J. (2014). eRm:
Extended Rasch Modeling. R package version 0.15-4. http://erm.r-forge.r-project.org/.
Maor, D. and Taylor, P.C. (1995). Teacher Epistemology and
Scientific Inquiry in Computerized Classroom Environments. Journal
of Research in Science Teaching, 32(8). 839-354. doi:
10.1002/tea.3660320807.
Martens, M.L. (1992). Inhibitors to Implementing a
Problem-Solving Approach to Teaching Elementary Science: Case
Study of a Teacher in Change. School Science and
Mathematics, 92(3). 150-156.
Mishra, P., & Kereluik, K. (2011). What 21st century
learning? A review and a synthesis. In C. D. Maddux, M.
J.Koehler, P. Mishra, & C. Owens (Eds.), Proceedings of
Society for Information Technology & Teacher Education
International Conference 2011 (pp. 3301–3312). Chesapeake, VA:
AACE.
Newble, D.I., & Clarke, R.M. (1986). The approaches to
learning of students in a traditional and an innovative
problem-based medical school, Medical Education,20,
267-273. doi: 0.1111/j.1365-2923.1986.tb01365.x.
Panayides, P., Robinson, C., & Tymms, P. (2010). The
assessment revolution that has passed England by: Rasch
measurement. British Educational Research Journal, 36(4),
611 626. doi: 10.1080/01411920903018182.
Paulsen, M.B. and Feldman, K.A. (2005). The Conditional and
Interaction Effects of Epistemological Beliefs on the
Self-regulated Learning of College Students: Motivational
Strategies. Research in Higher Education, 46(7),
731-768. doi: 10.1007/s11162-004-6224-8.
Perkins, D. (1993). Teaching for Understanding. American
Educator: The Professional Journal of the American
Federation of Teachers, 17 (3), 28-35. Document
available online
http://www.exploratorium.edu/IFI/resources/workshops/teachingforunderstanding.html
, 24 July 2011.
Perry, W.G. (1970). Forms of intellectual and ethical
development in the college years: A scheme. New York,
NY: Holt, Rinehart & Winston.
Piaget, J. (1954). The construction of reality in the child.
New York: Basic Books.
Pornprasertmanit, S., Miller, P., Schoemann, A., &
Rosseel, Y. (2014). semTools: Useful tools for structural
equation modeling.. R package version 0.4-6.
http://CRAN.R-project.org/package=semTools
Qian, G. and Alvermann, D.E. (2000). Relationship between
Epistemological Beliefs and Conceptual Change Lea.rning. Reading
&Writing Quarterly, 16, 59-74. doi:
10.1080/105735600278060.
Qian, G. and Pan, J. (2002). A Comparison of Epistemological
Beliefs and Learning from Science Text Between American and
Chinese High School Students. In. B.K. Hofer and P.R. Pintrich
(Eds.). Personal Epistemology. (pp. 365-385).
Mahwah, New Jersey U.S.A.: Lawrence Erlbaum Ass.
R Core Team (2014). R: A language and environment for
statistical computing. R Foundation for Statistical Computing,
Vienna, Austria. URL http://www.R-project.org/.
Rasch, G. (1960). Probabilistic Models for some
Intelligence and Attainment Tests. Danish Institute for
Educational Research, Copenhagen.
Raykov, T. (2012). Scale construction and development using
structural equation modeling. In R. H. Hoyle (Ed.), Handbook
of structural equation modeling (pp. 472-494). New York:
Guilford.
Richardson, J.T.E. (2012). Erik Jan van Rossum and Rebecca
Hamer – The meaning of learning and knowing. Book review. Higher
Education, 64 (5), pp 735-738. doi:
10.1007/s10734-012-9518-3.
Ross, S. N. (2008). Postformal (mis)communications. World
Future: Journal of General Evolution, 64 (5-7),
530-535. doi: 10.1080/02604020802303952.
Rosseel, Y. (2012). lavaan: An R Package for Structural
Equation Modeling. Journal of Statistical Software, 48(2),
1-36. URL http://www.jstatsoft.org/v48/i02/.
Säljö, R. (1994). Minding action; Conceiving of the world
versus participating in cultural practices. Nordisk
Pedagogik, 14(2), 71-80.
Säljö, R. (1997). Talk as Data and Practice – a critical look
at phenomenographic inquiry and the appeal to experience. Higher
Education Research & Development, 16, 173-190. doi:
10.1080/0729436970160205.
Schommer, M. (1990). Effects of Beliefs About the Nature of
Knowledge on Comprehension. Journal of Educational
Psychology, 82(3), 498-504. doi:
10.1037/0022-0663.82.3.498.
Schommer, M. (1994). An Emerging Conceptualization of
Epistemological Beliefs and Their Role in Learning. In Garner,
R and Alexander, P.A. (Eds.). Beliefs about Text and
Instruction with Text (pp. 25-40). Hillsdale:
Lawrence Erlbaum Associates.
Schommer, M. (1998). The influence of age and education on
epistemological beliefs. British Journal of Educational
Psychology, 68, 551-561. doi:
10.1111/j.2044-8279.1998.tb01311.x.
Schreiber, J.B. and Shinn, D. (2003). Epistemological Beliefs
of Community College Students and their Learning Processes. Community
College Journal of Research and Practice, 27, 699-709.
doi: 10.1080/713838244.
Sjölander, P., N. Lindström, A. Ericsson, S. Kjellström, 2014,
A pattern recognition method for disclosing different levels
of value system from questionnaire data, Behavioral
Development Bulletin, 19(3), 114-127. doi:
10.1037/h0100596.
Strijbos, J., Engels, N. & Struyven, K. (2015). Criteria
and standards of generic competences at bachelor degree level:
A review study. Educational Research Review, 14, 18-32).
doi:10.1016/j.edurev.2015.01.001.
Van Rossum, E.J., & Hamer, R.N. (2010). The meaning
of learning and knowing. Rotterdam, Netherlands: Sense
Publishers.
Van Rossum, E.J., & Hamer R., (2012). Analysing higher
education teachers’ learning–teaching conceptions with a model
of student thinking. Paper presented at the AERA Conference,
Vancouver, Canada.
Van Rossum, E.J., & Hamer, R. (2013). The relationship
between students’ conception of good teaching and their views
on a good textbook. Paper presented at the EARLI Conference,
Munich, Germany.
Van Rossum, E.J., & Schenk, S.M. (1984). The relationship
between learning conception, study strategy and learning
outcome. British Journal of Educational Psychology
54, 73-83. doi:10.1111/j.2044-8279.1984.tb00846.x.
Vermunt, J.D. (1996). Metacognitive, cognitive and affective
aspects of learning styles and strategies: A phenomenographic
analysis. Higher Education, 31, 25-50. doi:
10.1007/BF00129106.
Voogt, J, Erstad, O., Dede, C., & Mishra P. (2013)
Challenges to learning and schooling in the digital networked
world of the 21st century. Journal of Computer Assisted
Learning, 29, 403–413, doi: 10.1111/jcal.12029.
West, E.J., (2004) Perry's legacy: models of epistemological
development. Journal of adult development, 112,
61-70. doi: 10.1023/B:JADE.0000024540.12150.69.
Wright, B. D., Linacre, J. M., Gustafson, J. E., &
Martin-Lof, P. (1994). Reasonable mean-square fit values.
Rasch measurement transactions, 8(3), 370.
Yerrick, R.K., Pedersen, J.E., & Arnason, J. (1998).
“We're just spectators”: A case study of science teaching,
epistemology, and classroom management, Science
Education, 82(6), 619-648. doi:
10.1002/(SICI)1098-237X(199811)82:6<619::AID-SCE1>3.0.CO;2-K.