Assessment Handbook - PowerPoint PPT Presentation

1 / 27
About This Presentation
Title:

Assessment Handbook

Description:

Fayetteville State University ... University leaders want and need information about such topics as factors ... who entered the University after July 1, 1990, ... – PowerPoint PPT presentation

Number of Views:168
Avg rating:3.0/5.0
Slides: 28
Provided by: unc54
Category:

less

Transcript and Presenter's Notes

Title: Assessment Handbook


1
Assessment Handbook
A Guide for Establishing Awareness of the
Assessment Process At Fayetteville State
University
35 copies of this public document were printed at
a cost of 49.35 or 1.41 per copy
2
Introduction
  • One of the most frequently utilized, yet seldom
    understood processes for making educational
    decisions about student learning on the
    Fayetteville State University campus is the
    Assessment Process. Over the years, the
    assessment process has experienced few changes.
    Primarily, assessment has been focused in the
    general education division, namely, the
    University College, and has included placement
    testing, pre and post testing in core areas such
    as reading, mathematics, and critical thinking,
    and satisfaction inventories to gauge student
    opinions concerning the programs and services
    provided by the university. Since 1990, rising
    junior testing has emerged as one of the
    assessment initiatives used to improve student
    learning and development.
  •  
  • With increased emphasis on data-driven
    decision-making, assessment needs at FSU are
    growing rapidly. University leaders want and
    need information about such topics as factors
    influencing retention, learning outcomes, and a
    multiplicity of aspects of the student
    experience. Demands for information from various
    accreditation bodies, the General Administration
    of the university system, and the campus
    community are constantly increasing. The purpose
    of this Manual is to present the vision,
    structure, activities, and assessment measures
    that are an integral part of the educational
    program at Fayetteville State University.
  •  
  • The Manual concentrates on specific assessment
    activitiesdescribing assessment methods, use of
    surveys and questionnaires, using institutional
    data. A glossary of frequently used assessment
    terms is included to facilitate understanding the
    language of assessment. The Manual was written
    by the staff of University Testing Services as a
    resource for faculty and administrators
    responsible for assessment. It is our hope that
    the Manual will evolve over time and will be used
    as the basis for developing long-range assessment
    plans. We welcome your comments and suggestions
    to increase its usefulness.

3
The Vision
It is easier to go down a hill than up, but the
view is from the top
The motivation for this Manual is based on a
simple philosophy assessment should be
deliberately designed to improve student
performance. The only way we can properly judge
where we are is directly proportional to where we
would like to be. Our vision, in short, must be
that assessment is a functional rather than a
symbolic process, and that the connections
between the data we have and practical
implications are fully apparent.
4
Testing Dates
5
Nine Principles of Good Practice for Assessing
Student Learning
  • The assessment of student learning begins with
    educational values. Assessment is not an end in
    itself but a vehicle for educational improvement.
    Its effective practice, then, begins with and
    enacts a vision of the kinds of learning we most
    value for students and strive to help them
    achieve. Educational values should drive not only
    what we choose to assess but also how we do so.
    Where questions about educational mission and
    values are skipped over, assessment threatens to
    be an exercise in measuring what's easy, rather
    than a process of improving what we really care
    about.
  • Assessment is most effective when it reflects an
    understanding of learning as multidimensional,
    integrated, and revealed in performance over
    time. Learning is a complex process. It entails
    not only what students know but what they can do
    with what they know it involves not only
    knowledge and abilities but values, attitudes,
    and habits of mind that affect both academic
    success and performance beyond the classroom.
    Assessment should reflect these understandings by
    employing a diverse array of methods, including
    those that call for actual performance, using
    them over time so as to reveal change, growth,
    and increasing degrees of integration. Such an
    approach aims for a more complete and accurate
    picture of learning, and therefore firmer bases
    for improving our students' educational
    experience.
  • Assessment works best when the programs it seeks
    to improve have clear, explicitly stated
    purposes. Assessment is a goal-oriented process.
    It entails comparing educational performance with
    educational purposes and expectations -- those
    derived from the institution's mission, from
    faculty intentions in program and course design,
    and from knowledge of students' own goals. Where
    program purposes lack specificity or agreement,
    assessment as a process pushes a campus toward
    clarity about where to aim and what standards to
    apply assessment also prompts attention to where
    and how program goals will be taught and learned.
    Clear, shared, implementable goals are the
    cornerstone for assessment that is focused and
    useful.

6
Nine Principles of Good Practice for Assessing
Student Learning cont.
  • Assessment requires attention to outcomes but
    also and equally to the experiences that lead to
    those outcomes. Information about outcomes is of
    high importance where students "end up" matters
    greatly. But to improve outcomes, we need to know
    about student experience along the way -- about
    the curricula, teaching, and kind of student
    effort that lead to particular outcomes.
    Assessment can help us understand which students
    learn best under what conditions with such
    knowledge comes the capacity to improve the whole
    of their learning.  
  • Assessment works best when it is ongoing not
    episodic. Assessment is a process whose power is
    cumulative. Though isolated, "one-shot"
    assessment can be better than none, improvement
    is best fostered when assessment entails a linked
    series of activities undertaken over time. This
    may mean tracking the process of individual
    students, or of cohorts of students it may mean
    collecting the same examples of student
    performance or using the same instrument semester
    after semester. The point is to monitor progress
    toward intended goals in a spirit of continuous
    improvement. Along the way, the assessment
    process itself should be evaluated and refined in
    light of emerging insights.
  • Assessment fosters wider improvement when
    representatives from across the educational
    community are involved. Student learning is a
    campus-wide responsibility, and assessment is a
    way of enacting that responsibility. Thus, while
    assessment efforts may start small, the aim over
    time is to involve people from across the
    educational community. Faculty play an especially
    important role, but assessment's questions can't
    be fully addressed without participation by
    student-affairs educators, librarians,
    administrators, and students. Assessment may also
    involve individuals from beyond the campus
    (alums, trustees, employers) whose experience can
    enrich the sense of appropriate aims and
    standards for learning. Thus understood,
    assessment is not a task for small groups of
    experts but a collaborative activity its aim is
    wider, better-informed attention to student
    learning by all parties with a stake in its
    improvement.

7
Nine Principles of Good Practice for Assessing
Student Learning cont.
  • Assessment makes a difference when it begins with
    issues of use and illuminates questions that
    people really care about. Assessment recognizes
    the value of information in the process of
    improvement. But to be useful, information must
    be connected to issues or questions that people
    really care about. This implies assessment
    approaches that produce evidence that relevant
    parties will find credible, suggestive, and
    applicable to decisions that need to be made. It
    means thinking in advance about how the
    information will be used, and by whom. The point
    of assessment is not to gather data and return
    "results" it is a process that starts with the
    questions of decision-makers, that involves them
    in the gathering and interpreting of data, and
    that informs and helps guide continuous
    improvement.
  • Assessment is most likely to lead to improvement
    when it is part of a larger set of conditions
    that promote change. Assessment alone changes
    little. Its greatest contribution comes on
    campuses where the quality of teaching and
    learning is visibly valued and worked at. On such
    campuses, the push to improve educational
    performance is a visible and primary goal of
    leadership improving the quality of
    undergraduate education is central to the
    institution's planning, budgeting, and personnel
    decisions. On such campuses, information about
    learning outcomes is seen as an integral part of
    decision-making, and avidly sought.
  • Through assessment, educators meet
    responsibilities to students and to the public.
    There is a compelling public stake in education.
    As educators, we have a responsibility to the
    publics that support or depend on us to provide
    information about the ways in which our students
    meet goals and expectations. But that
    responsibility goes beyond the reporting of such
    information our deeper obligation -- to
    ourselves, our students, and society -- is to
    improve. Those to whom educators are accountable
    have a corresponding obligation to support such
    attempts at improvement.

8
Assessment Requirements
Assessment is the systematic collection, review,
and use of information about educational programs
undertaken for the purpose of improving student
learning and development (Palomba Banta). 
  Assessment involves the systematic process of
gathering and using data for evaluating and
improving programs and services. Basic
statistical measures are used to gain realistic
and concrete evidence about how well FSU is
achieving its strategic goals. In all cases
where individual student scores are on file, they
will be made available to interested students and
where possible, will be used in the academic
advisement process. Individual results will be
used for research purposes only. University
Testing Services maintains confidentiality of
individual test results. All students at FSU are
expected to participate in appropriate assessment
activities. FSU is a public university and is
therefore accountable to the state, and is
expected to prove, by demonstrating student
performance outcomes, that funds are being spent
appropriately. The University Assessment
Council, formerly known as the Assessment Policy
Advisory Committee, supports these assessment
requirements and will monitor the assessment
strategies and plans outlined in the recent SACS
self-study.
9
FSU Assessment Policy
  • All new students are required to take profile
    exams before registering for classes at FSU.  
    The only exceptions will be transfer students who
    transfer in 30 or more hours of credit in
    university-level courses, including both English
    and mathematics, and those students who are
    non-degree seeking students such as those
    enrolling for teacher certification or life
    enhancement or students who have permission from
    other institutions to enroll in FSU courses.
  • Students who entered the University after July 1,
    1990, as first time students with fewer than 57
    hours of credit, and/or students who have
    completed between 45-74 credit hours who are not
    teacher education majors, are required to take
    the rising junior examination before being
    unconditionally admitted to the upper division.
    Students will be notified by University Testing
    Services, which semester they are expected to
    fulfill the requirements. Should a student fail
    to meet the requirements as scheduled,
    registration for future courses may be withheld.

10
Assessment Plan Academic Year 2002-2003
Type Key 1 Outcomes 2 Expectations 3
Satisfaction 4 State of Affairs
11
Assessment TerminologyA Glossary of Useful
Terms The following list of terms may be useful
to describe current educational assessment
practices
  • Accountability  - The demand by a community
    (public officials, employers, and taxpayers) for
    school officials to prove that money invested in
    education has led to measurable learning.
    "Accountability testing" is an attempt to sample
    what students have learned, or how well teachers
    have taught, and/or the effectiveness of a
    school's principal's performance as an
    instructional leader. School budgets and
    personnel promotions, compensation, and awards
    may be affected. Most school districts make this
    kind of assessment public it can affect policy
    and public perception of the effectiveness of
    taxpayer-supported schools and be the basis for
    comparison among schools.
  •  
  • Achievement Test - A standardized test designed
    to efficiently measure the amount of knowledge
    and/or skill a person has acquired, usually as a
    result of classroom instruction. Such testing
    produces a statistical profile used as a
    measurement to evaluate student learning in
    comparison with a standard or norm.

12
Glossary cont.
  • Action Research - School and classroom-based
    studies initiated and conducted by teachers and
    other school staff. Action research involves
    teachers, aides, principals, and other school
    staff as researchers who systematically reflect
    on their teaching or other work and collect data
    that will answer their questions.
  • Affective - Outcomes of education involving
    feelings more than understanding likes,
    pleasures ideals, dislikes annoyances, values.
  • Alternative Assessment- Many educators prefer the
    description "assessment alternatives" to describe
    alternatives to traditional, standardized, norm-
    or criterion-referenced traditional paper and
    pencil testing. An alternative assessment might
    require students to answer an open-ended
    question, work out a solution to a problem,
    perform a demonstration of a skill, or in some
    way produce work rather than select an answer
    from choices on a sheet of paper. Portfolios and
    instructor observation of students are also
    alternative forms of assessment.
  • Analytic Scoring - A type of rubric scoring that
    separates the whole into categories of criteria
    that are examined one at a time. Student writing,
    for example, might be scored on the basis of
    grammar, organization, and clarity of ideas.
    Useful as a diagnostic tool.
  • Aptitude Test - A test intended to measure the
    test-taker's innate ability to learn, given
    before receiving instruction.
  • Assessment - In an educational context, the
    process of observing learning describing,
    collecting, recording, scoring, and interpreting
    information about a student's or one's own
    learning. At its most useful, assessment is an
    episode in the learning process part of
    reflection and autobiographical understanding of
    progress. Traditionally, student assessments are
    used to determine placement, promotion,
    graduation, or retention. In the context of
    institutional accountability, assessments are
    undertaken to determine the principal's
    performance, effectiveness of schools, etc. In
    the context of school reform, assessment is an
    essential tool for evaluating the effectiveness
    of changes in the teaching-learning process.

13
Glossary, cont.
  • Authentic Assessment- Evaluating by asking for
    the behavior the learning is intended to produce.
    The concept of model, practice, feedback in which
    students know what excellent performance is and
    are guided to practice an entire concept rather
    than bits and pieces in preparation for eventual
    understanding. A variety of techniques can be
    employed in authentic assessment. Authentic
    assessment implies that tests are central
    experiences in the learning process, and that
    assessment takes place repeatedly. Patterns of
    success and failure are observed as learners use
    knowledge and skills in slightly ambiguous
    situations that allow the assessor to observe the
    student applying knowledge and skills in new
    situations over time.
  •  
  • Benchmark - Student performance standards (the
    level(s) of student competence in a content
    area.) An actual measurement of group performance
    against an established standard at defined points
    along the path toward the standard. Subsequent
    measurements of group performance use the
    benchmarks to measure progress toward
    achievement.
  • Capstone Course - A capstone is a project
    planned and carried out by the student during the
    final semester as the culmination of the
    educational experience. These projects typically
    require higher-level thinking skills,
    problem-solving, creative thinking, and
    integration of learning from various sources.  A
    capstone course is designed to integrate the
    knowledge, concepts, and skills associated with a
    portion of or the entire sequence of study in a
    program (University of Colorado at Boulder).
  • Cohort - A group whose progress is followed by
    means of measurements at different points in
    time.

14
Glossary, cont.
  • Competency Test - A test intended to establish
    that a student has met established minimum
    standards of skills and knowledge and is thus
    eligible for promotion, graduation,
    certification, or other official acknowledgement
    of achievement.
  •  
  • Criterion Referenced Tests - A test in which the
    results can be used to determine a student's
    progress toward mastery of a content area.
    Performance is compared to an expected level of
    mastery in a content area rather than to other
    students' scores. Such tests usually include
    questions based on what the student was taught
    and are designed to measure the student's mastery
    of designated objectives of an instructional
    program. The "criterion" is the standard of
    performance established as the passing score for
    the test. Scores have meaning in terms of what
    the student knows or can do, rather than how the
    test-taker compares to a reference or norm group.
    Criterion referenced tests can have norms, but
    comparison to a norm is not the purpose of the
    assessment. Criterion referenced tests have also
    been used to provide information for program
    evaluation, especially to track the success or
    progress of schools and student populations that
    have been involved in change or that are at risk
    of inequity. In this case, the tests are not used
    to compare teachers, teams or buildings within a
    district but rather to give feedback on progress
    of groups and individuals.
  • Curriculum-embedded or Learning-embedded
    Assessment - Assessment that occurs
    simultaneously with learning such as projects,
    portfolios and "exhibitions." Occurs in the
    classroom setting, and, if properly designed,
    students should not be able to tell whether they
    are being taught or assessed. Tasks or tests are
    developed from the curriculum or instructional
    materials.
  • Cut Score - Score used to determine the minimum
    performance level needed to pass a competency
    test.

15
Glossary, cont.
  • Descriptor - A set of signs used as a scale
    against which a performance or product is placed
    in an evaluation.
  • Dimension - Aspects or categories in which
    performance in a domain or subject area will be
    judged. Separate descriptors or scoring methods
    may apply to each dimension of the student's
    performance assessment.
  • Essay Test - A test that requires students to
    answer questions in writing. Responses can be
    brief or extensive. Tests for recall, ability to
    apply knowledge of a subject to questions about
    the subject, rather than ability to choose the
    least incorrect answer from a menu of options.
  • Evaluation - Both qualitative and quantitative
    descriptions of pupil behavior plus value
    judgments concerning the desirability of that
    behavior.
  • Formative Assessment - Observations which allow
    one to determine the degree to which students
    know or are able to do a given learning task, and
    which identifies the part of the task that the
    student does not know or is unable to do.
    Outcomes suggest future steps for teaching and
    learning. (See Summative Assessment.)
  • Grade Equivalent A score that describes student
    performance in terms of the statistical
    performance of an average student at a given
    grade level. A grade equivalent score of 5.5, for
    example, might indicate that the student's score
    is what could be expected of a average student
    doing average work in the fifth month of the
    fifth grade.

16
Glossary, cont.
  • High Stakes Testing - Any testing program whose
    results have important consequences for students,
    teachers, schools, and/or districts. Such stakes
    may include promotion, certification, graduation,
    or denial/approval of services and opportunity.
  • Holistic Method - In assessment, assigning a
    single score based on an overall assessment of
    performance rather than by scoring or analyzing
    dimensions individually. The product is
    considered to be more than the sum of its parts
    and so the quality of a final product or
    performance is evaluated rather than the process
    or dimension of performance.
  • I. Q. Tests - Traditional psychologists believe
    that neurological and genetic factors underlie
    "intelligence" and that scoring the performance
    of certain intellectual tasks can provide
    assessors with a measurement of general
    intelligence. There is a substantial body of
    research that suggests that I.Q. tests measure
    only certain analytical skills, missing many
    areas of human endeavor considered to be
    intelligent behavior.
  • Item Analysis - Analyzing each item on a test to
    determine the proportions of students selecting
    each answer. Can be used to evaluate student
    strengths and weaknesses may point to problems
    with the test's validity and to possible bias.
  • Mean - One of several ways of representing a
    group with a single, typical score. It is figured
    by adding up all the individual scores in a group
    and dividing them by the number of people in the
    group. Can be affected by extremely low or high
    scores.
  • Median - The point on a scale that divides a
    group into two equal subgroups. Another way to
    represent a group's scores with a single, typical
    score. The median is not affected by low or high
    scores as is the mean. (See Norm.)

17
Glossary, cont.
  • Multidimensional Assessment - Assessment that
    gathers information about a broad spectrum of
    abilities and skills (as in Howard Gardner's
    theory of Multiple Intelligences.
  • Multiple Choice Tests - A test in which students
    are presented with a question or an incomplete
    sentence or idea. The students are expected to
    choose the correct or best answer/completion from
    a menu of alternatives.
  • Norm - A distribution of scores obtained from a
    norm group. The norm is the midpoint (or median)
    of scores or performance of the students in that
    group. Fifty percent will score above and fifty
    percent below the norm.
  • Norm Group - A random group of students selected
    by a test developer to take a test to provide a
    range of scores and establish the percentiles of
    performance for use in establishing scoring
    standards.
  • Norm Referenced Tests - A test in which a student
    or a group's performance is compared to that of a
    norm group. The student or group scores will not
    fall evenly on either side of the median
    established by the original test takers. The
    results are relative to the performance of an
    external group and are designed to be compared
    with the norm group providing a performance
    standard. Often used to measure and compare
    students, schools, districts, and states on the
    basis of norm-established scales of achievement.
  • Normal Curve Equivalent - A score that ranges
    from 1-99, often used by testers to manipulate
    data arithmetically. Used to compare different
    tests for the same student or group of students
    and between different students on the same test.
    An NCE is a normalized test score with a mean of
    50 and a standard deviation of 21.06. NCEs should
    be used instead of percentiles for comparative
    purposes.
  • On-Demand Assessment - A test for which the
    scoring procedure is an assessment process that
    takes place as a scheduled event outside the
    normal routine. An attempt to summarize what
    students have learned that is not embedded in
    classroom activity.

18
Glossary, cont.
  • Outcome - An operationally defined educational
    goal, usually a culminating activity, product, or
    performance that can be measured.
  •  
  • Percentile - A ranking scale ranging from a low
    of 1 to a high of 99 with 50 as the median score.
    A percentile rank indicates the percentage of a
    reference or norm group obtaining scores equal to
    or less than the test-taker's score. A percentile
    score does not refer to the percentage of
    questions answered correctly, it indicates the
    test-taker's standing relative to the norm group
    standard.
  • Performance-Based Assessment - Direct,
    systematic observation and rating of student
    performance of an educational objective, often an
    ongoing observation over a period of time, and
    typically involving the creation of products.
    Performance-based assessment is a test of the
    ability to apply knowledge in a real-life
    setting. Performance of exemplary tasks in the
    demonstration of intellectual ability.  
  • Performance Criteria - The standards by which
    student performance is evaluated. Performance
    criteria help assessors maintain objectivity and
    provide students with important information about
    expectations, giving them a target or goal to
    strive for.
  • Portfolio - A systematic and organized collection
    of a student's work that exhibits to others the
    direct evidence of a student's efforts,
    achievements, and progress over a period of time.
  • Portfolio Assessment - Portfolios may be
    assessed in a variety of ways. Each piece may be
    individually scored, or the portfolio might be
    assessed merely for the presence of required
    pieces, or a holistic scoring process might be
    used and an evaluation made on the basis of an
    overall impression of the student's collected
    work.

19
Glossary, cont.
Primary Trait Method - A type of rubric scoring
constructed to assess a specific trait, skill,
behavior, or format, or the evaluation of the
primary impact of a learning process on a
designated audience. Process - A generalized
method of doing something, generally involving
steps or operations which are usually ordered
and/or interdependent. Process can be evaluated
as part of an assessment, as in the example of
evaluating a student's performance during
pre-writing exercises leading up to the final
production of an essay or paper. Product - The
tangible and stable result of a performance or
task. An assessment is made of student
performance based on evaluation of the product of
a demonstration of learning. Proficiency Level
The equivalent of a cut score (on a forced-choice
assessment) but for a performance/complex
assessment. The proficiency level for a
performance assessment is set by determining the
required performance criteria (such as the
required level on a rubric) for a specific grade
level. Such a proficiency level could be
achievement of all the criteria required for a
scoring level, or it could be a set number of
points achieved by combining scores for each
feature on the rubric. Profile - A graphic
compilation of the performance of an individual
on a series of assessments. Quartile - The
breakdown of an aggregate of percentile rankings
into four categories the 0-25th percentile,
26-50th percentile, etc. Quintile - The
breakdown of an aggregate of percentile rankings
into five categories the 0-20th percentile,
21-40th percentile, etc.  
20
Glossary, cont.
Rating Scale - A scale based on descriptive words
or phrases that indicate performance levels.
Qualities of a performance are described (e.g.,
advanced, intermediate, novice) in order to
designate a level of achievement. The scale may
be used with rubrics or descriptions of each
level of performance. Reliability - The measure
of consistency for an assessment instrument. The
instrument should yield similar results over time
with similar populations in similar
circumstances. Rubric - Some of the definitions
of rubric are contradictory. In general a rubric
is a scoring guide used in subjective
assessments. A rubric implies that a rule
defining the criteria of an assessment system is
followed in evaluation. A rubric can be an
explicit description of performance
characteristics corresponding to a point on a
rating scale. A scoring rubric makes explicit
expected qualities of performance on a rating
scale or the definition of a single scoring point
on a scale. Sampling A way to obtain
information about a large group by examining a
smaller, randomly chosen selection (the sample)
of group members. If the sampling is conducted
correctly, the results will be representative of
the group as a whole. Sampling may also refer to
the choice of smaller tasks or processes that
will be valid for making inferences about the
student's performance in a larger domain. "Matrix
sampling" asks different groups to take small
segments of a test the results will reflect the
ability of the larger group on a complete range
of tasks. Scaled Scores - Scores based on a
scale ranging from 001 to 999. Scale scores are
useful in comparing performance in one subject
area across classes, schools, districts, and
other large populations, especially in monitoring
change over time.
21
Glossary, cont.
  • Scoring Criteria - Rules for assigning a score
    or the dimensions of proficiency in performance
    used to describe a student's response to a task.
  • Self-Assessment - A process in which a student
    engages in a systematic review of a performance,
    usually for the purpose of improving future
    performance. May involve comparison with a
    standard, established criteria. May involve
    critiquing one's own work or may be a simple
    description of the performance.
  • Senior Project - Extensive projects planned and
    carried out during the senior year of high school
    as the culmination of the secondary school
    experience, senior projects require higher-level
    thinking skills, problem-solving, and creative
    thinking. They are often interdisciplinary, and
    may require extensive research. Projects
    culminate in a presentation of the project to a
    panel of people, usually faculty and community
    mentors, sometimes students, who evaluate the
    student's work at the end of the year.
  • Standardized Test - An objective test that is
    given and scored in a uniform manner. Scores are
    often norm-referenced. A standardized test is a
    measure of student learning (or other ability)
    that has been widely used with other students.
    These are also sometimes called achievement
    tests. Examples are the SAT, GRE, GMAT, LSAT,
    MCAT, etc. 
  • Summative Assessment - Evaluation at the
    conclusion of a unit or units of instruction or
    an activity or plan to determine or judge student
    skills and knowledge or effectiveness of a plan
    or activity. Outcomes are the culmination of a
    teaching/learning process for a unit, subject, or
    year's study.
  • Validity - The test measures the desired
    performance and appropriate inferences can be
    drawn from the results. The assessment accurately
    reflects the learning it was designed to measure.

22
University Assessment Council
  • Dr. Akbar Aghajanian, Professor of Sociology
  • Dr. Booker T. Anthony, Chair, Interim Associate
    Vice Chancellor for Academic Affairs
  • Mrs. Brenda Freeman, Director of Institutional
    Research
  • Ms. Patricia Heath, Director of Public Education
    Outreach Testing Services
  • Dr. Stanley Johnson, Assistant Professor of
    History
  • Dr. Jon Young, Associate Vice Chancellor for
    Academic Affairs
  • Mr. Deon Winchester, President, SGA
  • College/Schools Representatives

23
University Testing Services
  • University Testing Services (UTS) provides test
    administration, scanning and scoring of
    performance and survey evaluations, consultation,
    and other assessment services to the FSU
    community, including students, faculty, and
    staff. University Testing Services has a variety
    of scanning options, which will allow faculty to
    gather information in the way that best suits
    your research or testing needs, including a
    Scantron stand-alone scanner and an NCS Optical
    Mark Reader.

24
University Testing Services, cont.
  • Five general types of assessment are conducted
    through UTS (1) national certification,
    admissions and matriculation tests for graduates
    and undergraduates (2) interest and personality
    assessments (3) performance evaluations, i.e.
    placement testing and rising junior testing (4)
    survey evaluations and (5) distance education
    testing.
  • A variety of national testing programs is offered
    at the university, including the CLEP tests,
    Graduate Record Exam (GRE), Graduate Management
    Admission Test (GMAT), Praxis Series, National
    Board for Professional Teaching Standards
    (NBPTS), Pharmacy Technician Certification Tests
    (PTCE), Miller Analogies Test, Law School
    Admissions Test (LSAT), Foreign Service Written
    Exam (FSWE), TOEFL, and many others. Contact the
    office at (910) 672-1301 or (910) 672-1299 to
    check on the availability of specific tests.

25
University Testing Services
The Testing Services Center is located on the 1st
floor of the Collins Administration Building,
room 134.  University Testing Services provides
the information necessary to evaluate the quality
of the educational programs and to obtain student
perceptions of the University's programs and
services.
26
References
Douglas, Karen, and La Voy Sharon The Campus
Assessment Working Group A Collaborative
Approach to Meeting Escalating Assessment Needs
in Higher Education Assessment Update, May-June
2002. Younger, Donna, McGrury, Susan, and
Fiddler, Morris Interpreting Principles of Good
Practice in Assessment Assessment Update, May
June 2001. Wiggins, Grant. Educative Assessment.
1st Ed. Jossey-Bass, San Francisco, Ca. 1998
27
35 copies of this public document were printed at
a cost of 49.35 or 1.41 per copy
Write a Comment
User Comments (0)
About PowerShow.com