>
Volume: | 8 | 7 | 6 | 5 | 4 | 3 | 2 | 1 |
Permission is granted to distribute this article for nonprofit, educational purposes if it is copied in its entirety and the journal is credited. Please notify the editor if an article is to be used in a newsletter. |
Gail S. MacColl & Kathleen D. White Parents, educators, school board members, and legislators all want to know "what works"
and "what doesn't" in terms of educational programs and innovations. The reasons for their
interest are obvious and worthwhile: first, they want to be sure that tax money is being spent
on educational programs that provide a positive return in terms of student progress; second,
they want to stay informed of trends in education so they know that their school districts are
keeping up with the latest practices and programs. This digest describes some of the problems in communicating with these audiences; it then
provides helpful information on how researchers can best present data on educational
practices that work and those that don't to these various audiences for maximum
effectiveness, impact, and influence and to keep communication with these audiences open
and valuable. Problems in effective communication to general, nontechnical audiences Accessibility Most research on effective educational practices does not filter down to the people who
contribute to or control funding. The main reason for this is that research reports on
educational practices almost universally appear only in professional and academic journals
or through other specialized sources. The average reader wanting to learn about successful innovations in education is generally
unable to locate such information, even after expending considerable effort. These kinds of
reports are usually unavailable through popular periodicals or bookstore chains and rarely
through more "serious" bookstores; in addition, they are not often found in or through local
libraries, including those in large metropolitan areas. Readability In the rare event that a general reader gains access to materials about workable educational
programs, three stylistic characteristics of these reports often make them unappealing:
organization, terminology, and presentation of statistical data. First, research studies are often organized in such a way as to hide major findings and
conclusions in the text or present them only at the end. A related problem is that abstracts
and introductions do not provide findings. Even diligent readers become discouraged by
these factors because the usefulness of a report or study is not readily apparent. Second, although the use of technical terminology often simplifies communication within a
discipline, it creates an obstacle for policy makers, parents, and other interested readers,
who usually are not trained in research or statistical techniques needed to understand an
esoteric research study. Third, many research studies use complex tables to summarize statistical data. These tables,
like research jargon, are often difficult for interested parents and program funders to
interpret, even with considerable effort. Difficulties in reporting "what doesn't work" Researchers often have difficulty in reporting on educational practices that don't work,
generally for one or more of the following reasons: Technical weaknesses that limit usefulness Assuming that the other problems are overcome, several weaknesses can occur in the
research itself to limit its value to those funding, evaluating, or deciding on the use of new
educational programs: How to increase the value of research studies to a wider audience Most of the recommendations for making research data more useful to more people are
simple, relatively easy to accomplish, and based on common sense. In general, the primary things to focus on are the needs of the audience. Researchers must
remember that, in order for their data to be most useful, they have to be accessible and
understandable to people with vested interests in the education process: parents, teachers,
legislators, school board members. These audiences either pay for, deliver, or fund education
programs, and each wants the best ones available. Researchers uncovering and reporting on programs and practices that work need to
distribute their findings as widely, clearly, and efficiently as they can; otherwise their efforts
do not create the levels of benefits for the discipline of education that they might. More specifically, when reporting study results to nontechnical audiences, researchers
should keep the following suggestions in mind: To accomplish these goals, researchers will have to learn how to creatively present their
findings not only to reach more general readers but to appeal to them too. This requires
several steps: References Good, R. G. & Wandersee, J. H. (1991). "No Royal Road: More on Improving the Quality of Published
Educational Research." Educational Researcher, 20 (8), 24-25. Newmann, F. M. & Wehlage, G. G. (1995). Successful School Restructuring: A Report to the Public and Educators
by the Center on Organization and Restructuring of Schools. Madison, WI: Center on Organization and
Restructuring of Schools. U. S. General Accounting Office. (1995). Program Evaluation: Improving the Flow of Information to the Congress
(GAO/PEMD-95-1). Washington, DC. U. S. General Accounting Office. (1996). Schools and Workplaces: An Overview of Successful and Unsuccessful
Practices (GAO/PEMD-95-28). Washington, DC.
Sitemap 1 - Sitemap 2 - Sitemap 3 - Sitemap 4 - Sitemape 5 - Sitemap 6
| |||||||||||||
Descriptors: *Communication (Thought Transfer); *Educational Research; Higher Education; *Readability; *Research Reports; *Researchers; Technical Writing; *Writing for Publication |
Sitemap 1 - Sitemap 2 - Sitemap 3 - Sitemap 4 - Sitemape 5 - Sitemap 6