INDIVIDUAL ASIGNMENT OF LANGUAGE ASSESSMENT
By
:
1.
Fitri Yunita NPM. 12.1.01.08.0215P
ENGLISH EDUCATION
DEPARTMENT
UNIVERSITY OF NUSANTARA
PGRI KEDIRI
Jl. K. H. Achmad Dahlan
No. 76 Kediri, Phone : (0354) 771576
Kediri , 2013
1.TESTING,
ASSESSING, TEACHING
·
What is a test ?
A test,
in simple terms, is a method of measuring a person’s ability; knowledge, or
performance in a given domain. Let’s look at the components of this definition.
A test is first a method. It is an instrument- a set of techniques, procedures,
or items that requires performance on the part of test-taker. Second, a test
must measure. Some tests measure general ability, while others determine a
general ability level; a quiz on recognizing correct use of definite articles
measure specific knowledge.
Next,
a test measures an individual’s ability; knowledge, or performance. Testers
need to understand who the test-takers are. What their previous experience and
background? Is the test appropriately matched to their abilities? How should
test-takers interpret their scores? Finally, a test measures in given domain.
In the case of proficiency test, even though the actual performance on the test
involves only a sampling of skills, that domain is overall proficiency in a
language-general competence in all skill of a language.
A
well-constructed test is an instrument that provides an accurate measure of the
test-takers’ ability within a particular domain.
·
Assessment
and teaching
Assessment
is popular and sometimes misunderstood term in current educational practice.
You might be to think of testing and assessing as synonymous terms, but they
are not. Tests are prepared
administrative procedures that occur at identifiable times in a curriculum when
learners muster all their faculties to offer peak performance, knowing that
their responses are being measured and evaluated.
Assessment,
on the other hand, is an ongoing process that encompasses a much wider domain.
Whenever a student responds to a question, offers a comment, or tries out a new
word or structure, the teacher subconsciously makes an assessment of the
student’s performance.
At
the same time, during these practice activities, teachers are indeed observing
students’ performance and making various evaluations of each learner: How did
the performance compare to previous performance? Which aspects of the
performance were better than others? Is the learner performing up to an
expected potential? How does the performance compare to that of others in the
same learning community? In the ideal classroom, all these observations feed
into the way the teacher provides instruction to each student.
·
Informal
and Formal Assessment
One
way to begin untangling the lexical conundrum created by distinguishing among
tests, assessment, and teaching is to distinguish between informal and formal
assessment. Informal assessment can
take a number of forms, Starting with incidental, unplanned comments and
responses, along with coaching and other impromptu feedback to the students. A
good deal of teacher’s informal assessment is embedded in classroom tasks
designed to elicit performance without recording results and making fixed
judgment about a student’s competence.
On
the other hand, formal assessments
are exercises or procedures specifically designed to tap into a storehouse of
skills and knowledge. They are systematic, planned sampling techniques
constructed to give teacher and students an appraisal of students’ achievement.
·
Formative
and summative assessment
Another
useful distinction to bear in mind is the function of assessment: how is the
procedure to be used? Two functions are commonly identified in the literature
formative and summative assessment. Most our classroom assessment is formative assessment: evaluating
students in the process of “forming” their competencies and skills with the
goal of helping them to continue that growth process. The key to such formation
is the delivery (by the teacher) and internalization (by the students) of
appropriate feedback on performance, with an eye toward the future continuation
(or formation) of learning.
Summative assessment
aims to measure, or summarize, what a student has grasped, and typically occurs
at the end of a course or unit of instruction. A summation of what a students’
has learned implies looking back and taking stock of how well the student has
accomplished objectives, but does not necessarily point the way to future
progress.
·
Norm-Referenced
and Criterion-Referenced tests
Another
dichotomy that is important to clarify here and that aids in sorting out common
terminology in assessment is the distinction between norm-referenced and
criterion-referenced testing. In
norm-referenced tests, each test-takers’ score is interpreted in relation
to mean (average score), median (middle score), standard deviation (extent of
variance score), and/or percentile rank. The purpose in such tests is to place
test-takers along in the form of numerical score.
Criterion-referenced test,
on the other hand, are designed to give test-takers feedback, usually in the
form of grades, on specific course or lesson objectives. Classroom tests
involving the students in only one class, and connected to a curriculum, are typical
of criterion-referenced testing. In a criterion-referenced test, the
distribution of students’ score across a continuum may be of little concern as
long as the instrument assesses appropriate objectives.
2. APPROACHES TO TEACHING,
LEARNING, AND LANGUAGE TESTING
A
brief history of language testing over the past half-century will serve as a
backdrop to an understanding of classroom-based testing. Historically,
language-testing trends and practices have followed the shifting sands of
teaching methodology. For example, in the 1950s an era of behaviorism and
special attention to contrastive analysis, testing focused on specific language
elements such as the phonological, grammatical, and lexical contrast between
two languages.
·
Discrete-point
and Integrative Testing
Discrete-point tests
are constructed on the assumption that language can be broken down into its
component parts and that those parts can be tested successfully. These
components are the skills of listening, speaking, reading, and writing and
various units of language (discrete language) of phonology, graphology,
morphology, lexicon, syntax, and discourse. It was claimed that an overall
language proficiency test, then should sample all four skills and as many
linguistic discrete points as possible.
What
does an integrative test look like?
Two types of tests have historically been claimed to be examples or integrative
tests: cloze tests and dictations. A
cloze test is a reading passage (perhaps 150 to 300 words) in which roughly
every sixth or seventh word has been deleted the test-taker is required to
supply words that fit into those blanks. Dictation
is familiar language-teaching technique that evolved into a testing technique.
Essentially, learners listen to passage of 100 to 150 words read aloud by an
administrator (or audiotape) and write what they hear, using correct spelling.
Supporters
argue that dictation is an integrative test because it taps into grammatical
and discourse competencies required careful listening, reproduction in writing
of what is heard, efficient short-term mem
·
Communicative
Language Testing
A
new theory of language and language use that exerted significant influence on
language teaching and therefore language testing from the early 1970s was the theory
of communicative competence.Communicative
competence is a linguistic term which refers to a
learner’s second language ability. It refers to a learner’s ability to: apply
and use grammatical rules,form
correct utterances,use
these utterances appropriately
The
term was coined by Dell Hymes (in 1966) who was inspired by Noam Chomsky‘s
distinction on linguistic competence and performance. According to Chomsky
(1965) a speaker’s language ability comprised two components: linguistic competence and linguistic performance. Hymes proposed that
knowing a language entailed knowing more than its grammar and rules. According to Hymes
there culturally specific rules that created a relationship between:the
language used,features
of the communicative context. e.g. What is appropriate
language for communication with a sibling may not be appropriate for
communication with an employer or lecturer.
Before
the theory of communicative competence, language was often described from a
psychological perspective, but this theory marked a profound shift in how
language was perceived as it presented language as an internal phenomenon.
With
the appearance of the communicative competence theory the focus shifted to a
more sociological one, where the focus was on external, social functions of
language.
The
relevance of Hymes’ theory to language testing was almost immediately
recognized when it appeared.
However,
it was a decade later that its actual impact was felt on practice with the
development of communicative language testing
Communicative
language tests (CLT) are distinguished by two main features:
a. CLTs
are performance tests and therefore require assessment to be carried out when
the learner or candidate in engaged in an extended (receptive/productive) act
of communication
b. CLTs pay attention to
the social roles candidates would assume and hence considers the roles that
candidates would assume in the real world on passing the test and offers a
means of specifying the demands of such roles in detail
Models
of communicative ability
Thought
it was a challenge to shift perspectives not to mention focuses of language
tests, there was a continuous theoretical engagement with the idea of
communicative competence and its implications for the performance requirements
of communicative language testing since the advent of the theory of
communicative competence.
A
number of writers have tried to specify the components of communicative
competence in second languages and their role in performance.
The
purpose of this is to provide a comprehensive framework for: test development,testing research , an interpretation of test
performance.
The
first such models specified the components of knowledge of language without
dealing in detail with their role in performance.
In
1980, Michael Canale and Merrill Swain published a paper that specified four
components of communicative competence:
Grammatical
competence- knowledge of systematic features of grammar, lexis and phonology
Sociolinguistic
competence - knowledge of rules of language use in terms of what is appropriate
in different contexts
Strategic
competence - ability to compensate for incomplete or imperfect linguistic
resources in a second language by using (other) successful communication
strategies
Discourse
competence - ability to deal with extended use of language in context (cohesion
and coherence)ory; and to an extent.
·
Performance-Based
Assessment
Performance-based assessment
of language typically involves oral production, written production, open-ended
responses, integrated performance, group performance and other interactive
tasks. To be sure, such assessment is time-consuming and therefore expensive,
but those extra efforts are paying off in the form of more direct testing
because students are assessed as they perform actual or stimulated real-world
tasks. In technical terms, higher content validity is achieved because learners
are measures in the process of performing the targeted linguistic acts.
3. SOME
CHARACTERISTICS OF
A GOOD ENGLISH TEST
·
Practicality:
a practical test is easy
to administer and to score without wasting too much time or effort.
·
Reliability
: a
test is considered reliable if it is taken again by the same students under the
same circumstances and the score average is almost the constant , taking into
consideration that the time between the test and the retest is of reasonable
length.
·
Validity: a test is considered as
valid when it measures what it is supposed to measure.
·
Authenticity: the language of a test
should reflect everyday discourse.
·
Washback
Any
language test or piece of assessment must have positive washback (backwash), by
which I mean that the effect of the test on the teaching must be beneficial.
This should be held in mind by the test constructors; it is only too easy to
construct a test which leads, for example, to candidates learning material by
heart or achieving high marks by simply applying test-taking skills rather than
genuine language skills
4. DESIGNING AUTHENTIC CLASSROOM LANGUAGE ASSESSMENT
·
Types of Authentic
Assessment
The
first task you face in designing a test for your students is to determine the
purpose for the test. Defining your purpose will help you choose the right kind
of test and it will also help you to focus on the specific objectives of the
test. Two types to create as a classroom teacher – language aptitude tests and
language proficiency tests, and three types that you will almost certainly need
to create – placement test, diagnostic tests, and achievement tests.
a.
Language
Aptitude Tests
A
language aptitude test is designed to measure capacity or general ability to
learn a foreign language and ultimate success in that undertaking. Language
aptitude tests are ostensibly designed to apply to the classroom learning of
any language.
Task
in the modern language aptitude test
1.
Number learning : Examinees must learn a set of numbers through aural input and
then discriminate different combination of those numbers.
2.
Phonetic script : Examinees must team a set of correspondences between speech
sounds and phonetic symbols.
3.
Spelling dues : Examinees must need words that are spelled some what
phonetically
4.
Word in sentence : Examinees are given a key word in a sentence and are then
asked to select a word in second sentence that performs the same grammatical
action as the key word.
5.
Paired associates : Examinees must quickly team a set of vocabulary words from
another language and memorize their English meaning.
b.
Proficiency
Tests
A
proficiency test is not limited to any one course, curriculum, or single skill
in the language, rather, is tests overall ability. Proficiency tests have
traditionally consisted of standardized multiple. Choice items on grammar
vocabulary, reading comprehension and aural comprehension. Some times also
writing is added.
Proficiency
test are almost always summative and norm – referenced. For example test of
English as a foreign language (TOEFL) produced by the educational testing
service.
c.
Placement
Tests
Certain
proficiency tests can act in the role of placement tests, the purpose of which
is to place a student into a particular level or section of language curriculum
or school. Placement tests come in many varieties; Assessing comprehension and
production, responding through written and oral performance, ended and limited
responses, selection (multiple - choice) and gap-filling formats, depending on
the nature of a program and its needs.
d.
Diagnostic
Tests
A
diagnostic tests is designed to diagnose specified aspects of a language. A
test in pronunciation, for example. Might diagnose the phonological features of
English that are difficult for learners and should there fore become part of a
curriculum.
There
is also a fine line of difference between a diagnostic test and a general
achievement test. Achievement test analyze the extent to which students have
acquired language features that have already been taught, diagnostic tests
should elicit information on what students need to work on in the future. In a
curriculum that has a form – focused phase, for example, a diagnostic test
might offer information about a learner’s acquisition of verb tense, modal
auxiliaries, definite articles, relative clause and the like.
e.
Achievement
Tests
An
achievement test is related directly to classroom lessons, units, or even a
total curriculum achievement tests are (or should be) limited to particular
material addressed in curriculum within a particular time frame and are offered
a course has focused on the objectives in question.
Achievement
tests are often summative because they are administered at the end of a unit or
term of study. The specifications for an achievement test should be determined
by :
•
The objectives of the lesson, unit, or course being assessed
•
The relative importance (or weight) assigned to each objective
•
The tasks employed in classroom lessons during the unit of time.
Achievement
tests range from five or ten – minute quizzes to three hour final examinations,
with an almost in finite variety of item types and formats. Here is the outline
for a midterm examination offered at the high intermediate level of an
intensive English program in the US:
Section
A Vocabulary
Part
1 (5 items) : Match words and definitions
Part
2 (5 items) : use the words in a sentence
Section
B Grammar
(10
sentences) : error detection (Underline or circle the error)
Section C Reading Comprehension
(2
one paragraph passage) : Four short – answer items for each
Section
D Writing
Respond
to a two-paragraph article on Native American culture
5. SOME PRACTICAL STEPS TO TEST CONSTRUCTION
·
Designing Multiple-Choice Test Items
This
is a back-to-basics article about the undervalued and little-discussed multiple
choice question. It’s not as exciting as discussing 3D virtual learning
environments, but it might be just as important. If you need to use tests, then
you want to reduce the errors that occur from poorly written items.
The
rules covered here make tests more accurate, so the questions are interpreted
as intended and the answer options are clear and without hints. Just in case
you’re not familiar with multiple choice terminology, it’s explained in the
visual below.
Here
are the ten rules. If you have any others, please add them through the Comments
form below.
Rule 1: Test comprehension and critical thinking,
not just recall
Multiple
choice questions are criticized for testing the superficial recall of
knowledge. You can go beyond this by asking learners to interpret facts,
evaluate situations, explain cause and effect, make inferences, and predict
results.
Rule 2: Use simple sentence structure and precise
wording
Write
test questions in a simple structure that is easy to understand. And try to be
as accurate as possible in your word choices. Words can have many meanings
depending on colloquial usage and context.
Rule 3: Place most of the words in the question stem
If
you’re using a question stem, rather than an entire question, ensure that most
of the words are in the stem. This way, the answer options can be short, making
them less confusing and more legible.
Rule 4: Make all distractors plausible
All
of the wrong answer choices should be completely reasonable. This can be very
hard to accomplish, but avoid throwing in those give-away distractors as it
detracts from the test’s validity. If you’re really stuck, get help from your
friendly SME. (BTW, this word can also be spelled as “distracter.”)
Rule 5: Keep all answer choices the same length
This
can be difficult to achieve, but expert test-takers can use answer length as a
hint to the correct answer. Often the longest answer is the correct one. When I
can’t get all four answers to the same length, I use two short and two long.
Rule 6: Avoid double negatives
No
big news here, right? Don’t use combinations of these words in the same
question: not, no, nor, the -un prefix, etc. For example, this type of question
could confuse test-takers: ‘Which of the following comments would NOT be
unwelcome in a work situation?’ Flip it around and write it in the positive
form: ‘Which of the following comments are acceptable in a work situation?’
Rule 7: Mix up the order of the correct answers
Make
sure that most of your correct answers aren’t in the “b” and “c” positions,
which can often happen. Keep correct answers in random positions and don’t let
them fall into a pattern that can be detected. When your test is written, go
through and reorder where the correct answers are placed, if necessary.
Rule 8: Keep the number of options consistent
Did
you ever have to convince a SME that he or she can’t have answer choices that
go to ‘h’ in one question and ‘c’ in the next? It’s something of a user
interface issue. Making the number of options consistent from question to
question helps learners know what to expect. Research doesn’t seem to agree on
whether 3 or 4 or 5 options is best. Personally, I like to use 4 options. It
feels fair.
Rule 9: Avoid tricking test-takers
As
faulty as they are, tests exist to measure knowledge. Never use questions or
answer options that could trick a learner. If a question or its options can be
interpreted in two ways or if the difference between options is too subtle,
then find a way to rewrite it.
Rule 10: Use ‘All of the Above’ and ‘None of the Above’
with caution
I
hate this rule because when you run out of distractors, All of the Above and
None of the Above can come in handy. But they may not promote good instruction.
Here’s why. All of the Above can be an obvious give-away answer when it’s not
used consistently. Also, the All of the Above option can encourage guessing if
the learner thinks one or two answers are correct. In addition, the downside to
None of the Above is that you can’t tell if the learner really knew the correct
answer.
Cloze test:
A gap-filling task, where words are deleted at fixed stages in a text and the
candidate has to replace them. For example, a cloze test may have every 6th
word deleted. Cloze tests are easy to prepare, but because of the random effect
of the deletion of every nth word, different cloze tests behave very
differently from one another. They should, therefore, undergo item analysis
before they are given to candidates.
Composition/ Essay:
A task where candidates have to produce at least a paragraph of their own
written language. Such tasks are marked subjectively (see analytic and holistic
marking scales).
Tidak ada komentar:
Posting Komentar