Agenda Exercise About the Study 12 Lessons and Insights Applying the Lessons © Copyright. Saul Carliner. 2000-2002. Setting Expectations Fun But no formula © Copyright. Saul Carliner. 2000-2002. Clarifying “We” Helen Constantinedes Kirk St.Amant Catherine Walstad STC © Copyright. Saul Carliner. 2000-2002. Exercise 1: The Request You’re Director of Technical Communication, reporting to the VicePresident of Development for a telecommunications firm. Your staff: numbers 76, including 5 managers, 50 writers, 5 planners, 5 editors, 5 programmers, 5 graphic illustrators, and the Director. Your work: mostly PDF files—about 75,000 pages per year. Also maintains 12 technical web sites (average of 37 topics each) , participate in the design of user interfaces for some of the software, and produce online marketing materials (prepared one such project last year). By his own admission, the new Vice-President has a limited understanding of technical communication and has suggested that you track metrics for your organization. What do you track? © Copyright. Saul Carliner. 2000-2002. Some Definitions Productivity Effectiveness Metrics Assessment Evaluation Proof © Copyright. Saul Carliner. 2000-2002. About the Studies Personal desire to find out how well I’m doing Best practices often emerge from non-competitors © Copyright. Saul Carliner. 2000-2002. Details: Part One Beginning of a larger study Purpose – Identify metrics used by related professions – Assess their transferability to technical communication Reviewed literature in software engineering, training, marketing and technical communication, public relations © Copyright. Saul Carliner. 2000-2002. Details: Part Two Studied larger technical communication departments Questioned about project, people, and business management Insufficient response for statistical validity but.. Some responses were abundantly clear © Copyright. Saul Carliner. 2000-2002. 12 Best Lessons and Insights about Metrics © Copyright. Saul Carliner. 2000-2002. 1. It’s about Evaluation It’s not about measurement © Copyright. Saul Carliner. 2000-2002. 2. Metrics Involve Several Measures No single measure suffices Differences in assessing Types of products Individuals Types of projects Departments Effectiveness Productivity © Copyright. Saul Carliner. 2000-2002. 3. Simple Measures Don’t (in fact, no standard practices across organizations in technical communication) Page counts Topics written Lines of codes written Student days © Copyright. Saul Carliner. 2000-2002. Issues Producing better often means producing less As number of media grow, no single count works © Copyright. Saul Carliner. 2000-2002. 4. You Cannot Wait Until Completion to Begin Showing a change in performance usually requires a before:after comparison: – Effectiveness of communication product – Productivity of development effort To be credible, measurement cannot begin after-the-fact © Copyright. Saul Carliner. 2000-2002. In Technical Communication Evaluation is not a standard business activity. Until that changes, measures unlikely to have credibility. © Copyright. Saul Carliner. 2000-2002. 5. Metrics Assume You Follow a Documented Process Preferably, an industry-standard one Software Capability Maturity Model Instructional Systems Design Process Analysis-Design-Placement Approach © Copyright. Saul Carliner. 2000-2002. In Technical Communication A standard process seems to exist Focused on project management rather than design and development issues Requirements-Development- “Checks”Release © Copyright. Saul Carliner. 2000-2002. 6. (starting it, anyway) Why do organizations commission us to write information for users? Hint: It’s not for usability. © Copyright. Saul Carliner. 2000-2002. Organizations Hire Us to Produce Business Results Why we’re commissioned Generate revenue Contain expenses Comply with regulations © Copyright. Saul Carliner. 2000-2002. 6. To Assess Effectiveness, Must Begin with Clear Goals State goals in observable and measurable terms: For intended business results For the content © Copyright. Saul Carliner. 2000-2002. In an Interesting Paradox in Technical Communication “Users performing intended tasks” and “business results” are commonly stated measures of effectiveness Projects do not begin with content or business objectives Usability testing is rarely performed © Copyright. Saul Carliner. 2000-2002. 7. To Assess Productivity, You Must Benchmark Against the Industry Industry surveys provide benchmarks: TRAINING Industry Survey IDG Surveys Merrill Lynch and other investment firms AD Week AC Nielsen © Copyright. Saul Carliner. 2000-2002. Unfortunately We have no comparable survey in technical communication. © Copyright. Saul Carliner. 2000-2002. 8. Industry Standard Models Build Acceptance They also improve methods Examples – Kirkpatrick Model – Direct response measures – cpm © Copyright. Saul Carliner. 2000-2002. Question You receive a resume. The applicant is one of the best you’ve seen. But you find a visible typo on the resume. Would you hire the candidate or not? © Copyright. Saul Carliner. 2000-2002. In Technical Communication Most common method of assessing effectiveness is technical reviews Use of Readers’ Comment Forms is low © Copyright. Saul Carliner. 2000-2002. 9. Measurements of Technique Are Not Persuasive They assess neither effectiveness nor productivity Although research might indicate which characteristics correlate with effectiveness, they do not guarantee it © Copyright. Saul Carliner. 2000-2002. Examples White space Typographical errors Number of index entries per page Number of illustrations per page Choice of type font © Copyright. Saul Carliner. 2000-2002. When Possible, Use Financially Based Measures Sales generated Billings Registrations Billings © Copyright. Saul Carliner. 2000-2002. 10. Intangibles Matter to Executives Perceptions Feelings Frustration Name recognition Also collect descriptive data in addition to numeric data © Copyright. Saul Carliner. 2000-2002. Use of Data Numerical data indicates the extent Descriptive data tells the impact © Copyright. Saul Carliner. 2000-2002. 11. Collect Data on an Ongoing Basis Devote between 2 and 5 percent of a project budget to evaluation © Copyright. Saul Carliner. 2000-2002. Issues One-time studies Desperation studies Metrics for tyranny rather than as a tool for continuous improvement © Copyright. Saul Carliner. 2000-2002. 12. On Their Own, Metrics Prove Nothing Metrics build perceptions Data supports those perceptions Data promotes informed decision-making Part of a larger system of building and maintaining relationships with sponsors © Copyright. Saul Carliner. 2000-2002. In Other Words Metrics are a value system. We choose to measure things we believe to be important. © Copyright. Saul Carliner. 2000-2002. A Successful Metrics Program Tracks measurements that are important to us as technical communicators and meaningful to our sponsors © Copyright. Saul Carliner. 2000-2002. Take-Aways © Copyright. Saul Carliner. 2000-2002.
© Copyright 2026 Paperzz