The Academy of Process Educators continues to increase its
membership and work to drive transformational change in education by generating,
disseminating, and archiving research based on the principles of Process
Education, including the sponsorship of a yearly Process Education Conference.
Academy membership is open to anyone passionate about Process Education
and interested in engaging, supporting, and collaborating with a community of
similarly dedicated educators. Visit the Academy online at:
www.processeducation.org.
The Academy of Process
Educators is a registered non-profit corporation.
Now for a few Academy news items:
Academy Store
This is a one-stop shopping area for
Academy membership (new or renewals), the print version of the International
Journal of Process Education, subscriptions to Dialogues, and free subscriptions
to the Pacific Crest Newsletter and the Bi-weekly Insight. Donations are also
accepted through this site as the Academy is a registered non-profit 501(c)
corporation.
Member Portal on
Academy Website
We've created a special area where
members have direct access to (1) e-Faculty Guidebook, (2) electronic version of
the International Journal of Process Education, (3) Academy directory, (4)
Insights and Dialogues, (5) Pacific Crest website, (6) Academy meeting minutes,
and (7) Process Education learning objects that you can use to support
professional and classroom work.
International
Journal of Process Education
This is our peer-reviewed journal that serves as an archive
for research on best practices in process education. Issues are unveiled each
year at our Annual Conference where attendees (as well as those who separately
renew their membership) receive a hard copy. Is there an interesting curriculum
or pedagogy topic on which you've done literature analysis and/or about which
you've collected data? Why not submit an article for review and development. The
deadline for draft paper submissions for the 2010 issue that will appear at next
year's conference is January 15, 2010.
www.processeducation.org/ijpe/
Learning to Learn
Camp Research
A
sub-group of Academy members has begun a multi-year effort
to assess the long-term impact of pre-college Learning to
Learn Camps. Two research questions have been posed based
on an extensive literature review:
-
How do learner efficacy in personal identity, learning
skills growth, and retention and persistence statistics change for Learning to
Learn Camp participants compared to a control group of students who have not
taken the Camp?
-
How do educator
effectiveness and practices in mentoring, facilitation of learning, and
attitude toward students change as a result of the Learning to Learn Camp?
Institutions collaborating on this
initiative are Hinds Community College (Libby Mahaffy, Vickie Kelly and Tom
Kelly), Gaston Community College (Virgil Cox), North Carolina Central University
(Bernice Johnson and Janice Harper), Lamar University (Brenda Nichols), and
Grand Valley State University (Ed Baum). These institutions are interested in
holding Camps that will use common data collection procedures for follow-up
measurement of the students from the camps. Joann Horton of Pacific Crest is
leading the team by coordinating biweekly phone conferences. Steve Beyerlein and
Barb Williams, both of the University of Idaho, are providing assistance with
research design, and Carol Nancarrow is offering her experience as a facilitator
of many Learning to Learn Camps. The near-term goal of the group is a
completed proposal to investigate the first research question before December
15th.
Writing Rubric
Research
We are checking
out our tools! Academy members and Pacific Crest have developed many rubric
measures that are used in the field by faculty across the country. The next step
in measurement development for process educators requires research to establish
the inter-rater reliability of these existing measures. One prominent example is
the Writing Rubric published in the Faculty Guidebook, 4th
edition, by Kathy Burke and Carol Nancarrow. A research team consisting of Cy
Leise (Bellevue University), Tris Utschig (Georgia Tech), and Kathy Burke (SUNY
Cortland) are gathering data on a sample of about 50 papers for each of three
types of writing, e.g., scholarly papers, project reports, and student
self-growth reports. To study the reliability of the tool, each paper will be
read by the research team and the rating results will be analyzed for
consistency. Initial work is focusing on rating issues arising from pilot work
and the development of an electronic means of rating individual papers and
collecting the data for later statistical analysis. This project will help (1)
determine to what degree the Writing Rubric can produce consistent results and
(2) guide further development of this useful tool. |
|