Modal Unit-Response Rates and Strategy: A Study of Historical Response Rate Trends by Mode of Collection for Multiple Surveys Michael Zabelsky U.S. Census Bureau, 4600 Silver Hill Rd., Suitland, MD, 20746 Disclaimer This report is released to inform interested parties of research and to encourage discussion. Any views expressed on statistical, methodological, technical, or operational issues are those of the author and not necessarily those of the U.S. Census Bureau. Abstract In survey processing, often times research and operations are viewed as two completely separate disciplines. This study intends to bridge one aspect of the research and operations gap, using historical, modal, statistical data analysis and response rate data visualization to make recommendations that impact future operations decisions and objectives. Key words: Statistical data analysis, data visualization, operations decisions 1. Thesis This report suggests that, to synchronize the introduction of web-collection instrument enhancements, online data resources, and online data visualization tools to coincide with sample reselection cycles will ultimately help maximize electronic response rates over time. Increases in electronic response are highly correlated with increases in overall response for this paper’s featured survey, the Quarterly Services Survey. Integrated web applications and sample selection redesigns both involve long, and often expensive, development lifecycles. As a result, numerous strategic benefits could result by pairing the development and release of both of these critical activities. In addition, this paper will explore how statistical agencies can use historical, modal, operations statistics to determine future direct-mail package contents and establish multi-mode, first-contact operations based on evolving reporting preferences. Ultimately, it is expected that when a survey design adapts to respondent’s preferences or needs, an increase in first-contact response rates can occur. This increase will also directly lead to decreases in the cost and size of follow-up operations and a reduction in overall response burden. If measured consistently across different surveys, trends in modal response can be used as a method of evaluation for making program comparisons and reconciliations. Tracking unit-level completion rates by operation within one statistical period can give managers a measure of the contribution of maintaining each mode of collection, and of the effectiveness of individual initial and follow-up efforts. More accurately forecasting response trends, and therefore job sizes of cost drivers, could also result in reduced processing center shop floor staffing and cost estimates. 2. External Environment – Risk is an Opportunity The United States Postal Service (USPS) is an integral part of collecting Census Bureau survey data. Over the past three decades, the federal government and the information technology industry in America have dramatically changed the way people communicate. Because of these societal, technological, communication efficiencies, and further fueled by the recession of 2007-2009, the USPS is now 1 threatened by decreased mail volume. An overall decline in mail volume has forced the unit cost of mailing to higher levels. That trend will likely continue as USPS reacts to a changing external environment. The pressures forcing corporate rebranding and restructuring may eventually result in overall service changes. Even with imminent reforms already in place or looming, the Government Accountability Office (GAO) is projecting future losses for the United States Postal Service in the hundreds of billions of dollars through 20201. Using this time horizon, the increasing probability of a high-impact event affecting both institutions warrants the Census Bureau’s consideration. 3. Motivating the Research At a strategic level, the Census Bureau must continue to monitor survey collection practices in the wake of USPS uncertainty and with the advent of new technologies. As part of a long-term risk mitigation strategy, the Census Bureau should continue to reposition itself by designing more progressive and technologically savvy online data collection tools. By investing in new collection practices, the Census Bureau can better position itself for future mail shocks or changes in respondents’ attitudes towards the usage of the mail system. In addition, these new processes and technologies will be focused on helping to: Improve first contact response rates Set tough internal standards for improving electronic response rates Strengthen internal controls that direct investment in technology and staffing devoted to creating public-facing web content Make strategically-timed enhancements to online reporting instruments and data resources to increase user satisfaction and overall online usage Develop dynamic response strategies that recognize and leverage modal response histories to influence the type of survey outreach materials designed and method of delivery 4. Measurements and Participation This study identified all current, multi-mode, Census Bureau Economic Directorate surveys that have maintained historical response records and were willing to participate. Modal response data were collected at the closeout of each statistical period. In this study, the source of all electronic response rates originates from one or a combination of Census Taker and Centurion. Census Taker and Centurion are the primary Internet data collection software applications used by the Economic Directorate at the Census Bureau. In addition, an internal survey of Census managers was conducted to identify changes in each program that may have influenced modal response rates over time. 5. Reliability and Limitations of Data Various factors affect response, including exogenous factors such as socio-political and economic reasons, and endogenous factors such as response follow-up strategies and sample variation. This analysis does not specifically investigate the effect that sample variation has on response rates. Interested parties may refer to www.census.gov/econ to learn more about data limitations and reliability of survey data for surveys included in this report. The analyses in this report focus on the option and level of Internet reporting as it is associated with other modes of collection and overall response, but do not test for cause and effect and how an Internet reporting option, by itself, affects response. 1 U.S. Postal Service: Financial Crisis Demands Aggressive Action: GAO-10-538T: http://www.gao.gov/assets/130/124292.pdf 2 6. Results Data was only collected from surveys after the Internet became a response option. Because of the staggered implementation and availability of electronic reporting instruments to individual programs over time, the number of statistical periods of collected data varies. The surveys specifically included in this study cover three different data collection frequencies: Annual Annual Retail Trade Survey (ARTS) Information and Communication Technology Survey (ICT) Annual Wholesale Trade Survey (AWTS) Annual Capital Expenditures Survey (ACES) Services Annual Survey (SAS) Quarterly Quarterly Services Survey (QSS) -FEATURED SURVEYQuarterly Plant Capacity Utilization Survey (QPC) Quarterly Financial Report, Manufacturing / Mining / Wholesale (QFR Man/Min) Quarterly Financial Report, Retail (QFR Retail) Quarterly Births Survey (SQ-Class) Monthly Monthly Wholesale Trade Survey (MWTS) Over the past eight years a significant phase shift in response preferences occurred, as shown in Figures 1 & 2: Figure 1: Figure 2: Figures 1 & 2: Figure 1 depicts the Internet response rate when electronic reporting in Census Taker or Centurion was debuted for each survey. Debuts in Census Taker and Centurion take place as early as 2004. Figure 2 shows the Internet response rate for these same surveys as of 2011. The two figures help to depict the shift of electronic reporting in the Economic Directorate at the Census Bureau. Respondents, representing a variety of economic industries, are steadily migrating to online survey reporting. Increases have been observed across the board in every survey studied. In nine of eleven economic surveys studied, web-collection has met or outpaced response rates for all other modes of collection combined, as depicted in Figure 3. Some surveys saw this dramatic shift in as few as two statistical periods since the introduction of Internet reporting as a response option. 3 Figure 3: Figure 4: Figures 3 & 4: The first four graphs (Figures 1-4) are tied together using a color scheme. Debut Census Taker and Centurion electronic reporting response rates are visualized in light shading in Figures 1 & 3. The darker shaded bars categorize present day Internet response percentages in Figure 2, and also display Internet response rate gains since an online reporting instrument’s introduction, in Figures 3 & 4. The average annual percent change line in Figure 4 helps quantify Internet response rate growth for making survey comparisons, and can also give the reader a clear understanding of the length of time an electronic reporting option has been available for each survey by comparing the average annual growth rate to total growth since debut. Note: SQ-Class is a quarterly survey, but only has three statistical periods of recorded Internet modal data. This is the reason the average annual increase line is higher than the total percent change bar in Figure 4. Through a more detailed analysis of modal response histories for individual surveys, even more exciting is the finding that under the right circumstances, increases in online response rates can be highly correlated with increases in overall response. No other mode of collection can boast a positive correlation with increases in overall response. As such, when it is offered, the Internet has become the only mode of collection with a promising upward growth potential. The Quarterly Services Survey (QSS) was selected as the featured survey for a number of reasons. Of all surveys studied, it has the oldest online data collection operation2, established in 2004. It has maintained four modes of collection since survey inception: Mail, Internet, Fax, and Phone. Of the study surveys, QSS has the most statistical periods of recorded modal data that includes electronic reporting. Because so many modal data points have been recorded, meaningful regression analysis can be performed comparing mode-to-mode unit response rates, and mode-to-total unit response rates to show changing relationships in response over time. There is also continuity in the QSS response strategy over many statistical periods; 2 Census Taker or Centurion online collection instruments only 4 it is a perfect example of the “traditional approach” of multi-mode response strategies defined in section 7.1 of this report. In addition, QSS has tracked historical, modal response statistics before and after one sample reselection and four sample coverage expansions, allowing for the impacts of these macroprogram changes to be measured against existing trends in modal response data. All modal response rates at collection-closeout since 2004Q1 are displayed in Figure 5: Figure 5: For historical, time-series representations of modal response rates for a single survey, data are being displayed as a stacked bar chart, as modal percentages of the total response rate percentage as a function of sample universe size. See section 6.2 for an explanation of the why there are two 2006Q3 statistical periods being graphed. For context, the major program changes occur during the following statistical periods: Classification Sample 1, named BSR2K: Statistical Period(s) 2004Q1 – 2006Q3 Sample Reselection: 2006Q3, two periods of data displayed Sample 2, named BSR06: 2006Q3 – 2011Q1 Sample Coverage Expansions: 2009Q1, 2009Q2, 2009Q3, 2010Q2 6.1. QSS Sample 1; BSR2K, 2004Q1 -2006Q3 The most apparent trend measured during the first sample, BSR2K, is the conversion of a large percentage of responses from the paper to the electronic mode. Figure 5 shows this electronic reporting growth between quarters 2004Q1 -2006Q3. From the BSR2K sample’s origin to maturity, a 19% increase in electronic response rates was measured, with an almost equal and opposite decrease in mailback forms being received. It appears that once a BSR2K sample respondent reported electronically for the first time, not only did they continue to complete the survey, but they became conditioned to report 5 via the Internet. This established a low electronic reversion rate, observed as steady increases in the blue Internet response rate bars in Figure 5. A low reversion rate laid the foundation for Internet response rate growth to build on previous gains during almost every statistical period until the sample reselection in 2006Q3. While a very promising tendency, this trend indicates that the electronic instrument had not yet met its full growth potential. At the start of the sample, a time-sensitive opportunity existed to increase overall online usage. Based on collection records, Census knows that 36% of all BSR2K firms were eventually willing to report electronically. But Census must also note that there was unfulfilled growth potential based on two measured factors: 1) the continuous increases in electronic response until a major program change that introduced new, unconditioned respondents to the survey, combined with 2) a maximum electronic response rate being achieved during the last statistical period of BSR2K, or at sample maturity. Could the Census Bureau have done more to secure each electronic response sooner in the sample’s maturation? If so, would the continual increases in electronic response have driven the BSR2K sample to finish at a higher rate? Even with the strong, negative correlation between mail and Internet response rates in BSR2K, it is during this sample’s data collection that a positive correlation begins to appear between increases in electronic response and increases in overall response, as shown in Figure 6. For QSS, there are two important factors to consider when reviewing the linear regression for mode-to-mode or mode-to-total response rates. Meaningful data analysis only takes place within each sample. When response data are plotted across samples, no relationship appears to exist, as shown in Figure 7. A second, important factor that contributes to higher overall response rates is record management within a sample by analytical staff that, in effect, shrinks the sample size of cases being contacted in future statistical periods. More problem cases are being removed from each subsequent statistical period’s collection efforts or imputed, improving the quality of the average case being contacted in a sample through to a sample’s maturity. Figure 6: Correlation = .81, R-squared = .66 Figure 7: Correlation = .01, R-squared = .1 6.2 QSS Sample Reselection, 2006Q3 The Quarterly Services Survey (QSS) underwent a sample reselection in statistical period 2006Q3, from which a fraction of cases were dropped completely from the sample. A fraction of cases remained exactly as they were in the sample or experienced a NAICS coverage change, and some corporate entities remained with a change to individual firms being selected to participate. The sample is rounded out by completely new firms being selected to participate. It is from the latter two categories that we find our most noteworthy response event for the sample reselection during 2006Q3. Internet response rates are immediately slashed by 10%. That decrease is felt in overall response rates too, which drop 11%. The difference between the two 2006Q3 bars in Figure 5 is the substitution of BSR2K sample “drops” for 6 BSR06 sample “adds”. The decreases in both Internet and overall response rates are most likely attributed to a respondent seeing a QSS form for the first time. 6.3 QSS Sample 2; BSR06 – Pre-expansion, 2006Q3 – 2008Q4 Prior to the coverage expansions that start in 2009Q1, the BSR06 sample sees electronic response associated with new, overall response rate gains. It is during this period of time that the electronic reporting instrument eclipses the joint contributions of all other modes of collection relative to overall response. Unlike in BSR2K, a strong, negative relationship does not exist between electronic and mail response during BSR06. Mail response can be described as a steady, moderate decline between 2006Q3 – 2008Q4, while electronic response mirrors large gains attained in BSR2K, as shown in Figure 5. As a result of the absence of a strong, negative correlation between electronic and mail response during BSR06, the gains in electronic reporting are now more highly correlated with increases in overall response rates. This correlation is shown in Figure 8. Online data collection still does not appear to have reached its maximum growth potential by 2008Q4. The systematic pattern of electronic response rate gains over both samples proves that respondents were satisfied with reporting using the online instrument. Figure 8: Correlation = .91, R-squared = .83 6.4 Tactical Planning Recommendations The Census Bureau has an opportunity to maximize Internet response rates and improve overall response rates when making sample modifications. Based on what was measured over the course of the BSR2K sample and during the sample reselection, a time-sensitive opportunity exists to aggressively target new respondents during a reselection period and promote the electronic reporting instrument. Eye-catching promotional materials, and a “year 1” strategic communications plan are encouraged. The goal is to design and deliver survey materials that compel a respondent to initiate the Internet reporting log-in process because of the known, high retention rate of online responders. A second planning recommendation states that, for surveys that experience five-year sample reselections, all major online instrument redesigns should be timed to coincide with those reselection periods in order to maximize the effect of the improvement. Survey designers should ensure marketable changes that enhance the overall online experience. For the most part, new respondents making their first impression of a survey are introduced in large groups during periods of sample reselection or coverage expansion. The timing of reselection cycles are known far in advance, and are less likely to be delayed because of budgetary constraints compared to proposed coverage expansions. Therefore, for planning purposes, it may be easier to sync the timing of Internet reporting enhancements with reselection cycles, not coverage expansions. 7 A third planning initiative recommends experimenting with the transition of broadening a respondent’s online experience beyond that of just reporting for a survey; make the sharing of information mutually beneficial. Statistical agencies spend considerable resources getting survey forms into an audience’s hands to generate response, and even more resources to securely offer and promote an electronic mode of collection as a response option. If agencies do not in turn highlight applicable statistical products already being created as part of a respondent’s online experience, or even go so far as to integrate statistical products into the electronic reporting experience itself, then a valuable opportunity to self-promote the work of the organization to the general public during their visit to the agency’s websites is being lost. 6.5 Sample 2; BSR06 including Expansions, 2006Q3 - Present The QSS response rate has fluctuated greatly during the survey’s five major program changes. After all of the expansions, the QSS sample size more than tripled. The overall response rate of seventy five percent is actually the same in the first and last period of collected response data. What has changed over that period of time is the importance of the electronic reporting option that contributes to overall response. In eight years it has become the preferred method of reporting, evolving from around twenty percent of all responses to now contribute over sixty percent of all responses each statistical period, as seen in Figures 5 and 10. Figure 9: Figure 10: Figures 9 & 10: Both figures display response rate changes from the baseline rates established at the inception of the survey, in Statistical Period 2004Q1. Figure 9 shows period-to-period fluctuations in overall response. Figure 10 shows period-to-period fluctuations in Internet response. The resilience of the electronic reporting option is evident, displayed by how quickly the Internet response rates rebound following expansion quarters, as seen in Figure 5. Once again, the final statistical period of recorded modal data in 2011Q1 has Internet finishing at an all-time high for the survey, indicating the electronic reporting instrument has most likely not reached its full growth potential. 6.6 Respondent Type, Feature Driven Not all surveys conduct sample reselections every five years. The Quarterly Financial Report (QFR) is an example of a survey that continually refreshes its sample. The survey form is completely different than 8 QSS; it has many more data elements and is more time consuming for respondents to complete. However, a positive correlation still exists between increases in electronic response and increases in overall response. High electronic reporting retention rates can be achieved by knowing a survey’s audience and tailoring features of the reporting tool to fit their style and needs. Accountants have the ability to print and file a proof of response certificate after successfully submitting a response electronically. No other QFR collection option offers tangible evidence indicating a completed survey. The importance of this type of evidence increases in line with a survey’s complexity and response burden. Even though there are stark contrasts between reporting for the QSS and QFR surveys, by building an online environment with desired features that reinforce the preferred mode of collection, a survey can create and exploit opportunities to improve overall response rates. 7. Solution Vision and Scope Modernizing multi-mode response strategies using paradata is critical to this paper’s thesis. Defining existing and new response strategies is an initial step towards using historical, modal, operations statistics to determine future direct-mail package contents and establish multi-mode, first-contact operations based on evolving reporting preferences for a sample survey. The current Quarterly Services Survey (QSS) response strategy is defined as an example of a typical, existing, multi-mode data collection approach. An overview of the new 2012 Commodity Flow Survey (CFS) response strategy will be introduced as a way to redefine a multi-mode response strategy that leverages response history and previous collection operations. The new CFS response strategy can be applied to other sample surveys. Section 7.2 will propose a new utility for electronic data collection and a new channel for data dissemination. Building on section 7.2, an envisioned, future state of origin-destination surveys is being proposed in Section 7.3 as an example of how to expand on electronic reporting objectives defined in a Census Bureau Director’s ICES-III presentation. The four objectives from the Census Director’s ICES-III presentation appear at the start of the next three subsections, 7.1, 7.2, and 7.3. 7.1 Two Types of Multi-Mode Response Strategies, Defined Objective: facilitate and accelerate reporting3 7.1.1 Quarterly Service Survey (QSS), existing response strategy Traditional Approach - a series of separate, single-mode operations conducted sequentially over the course of one statistical period’s data collection, resulting in multi-mode data collection for the overall statistical period. The traditional approach is typically only reactive based on direct respondent requests. 7.1.2 Commodity Flow Survey (CFS), new adaptive response strategy Responsive Approach - multi-mode operations conducted simultaneously over the course of one statistical period’s data collection, where multiple modes of collection are being used to satisfy the same first contact or follow-up respondent contact dates. Language in outreach materials, tools and scripts are being designed to specifically address each of these unique response scenarios and encourage electronic response, and respondents are being qualified for different Census outreach operations based on past response history and contact information collected in previous statistical periods. 7.1.3 Recommending the adaptive response strategy This paper recommends more multi-mode Census Bureau surveys adopt the responsive approach. Benefits of the traditional approach are in its simplicity and continuity. Stable, consistent response environments have allowed for response rate measurements to become trends over time. Compared to the 3 Lee, Ronald; Mesenbourg, Thomas, The Electronic Reporting Option – Does It Make Sense or Cents? ICES-III Presentation Slide 3 9 responsive approach, the traditional approach requires fewer total operations per statistical period. However, though the responsive approach requires more total operations, the additional operations would take place within the same work units, would represent smaller job sizes, and would be a natural way to break-up outreach operation work for large surveys at Census’s processing center. Major costs associated with planning, programming, and testing would still only be incurred once, increasing the value per operation contributed by each work unit, while transitioning to more personalized outreach operations. The responsive approach can still foster a stable reporting environment over time, just with more response scenarios. An advantage of implementing the responsive approach comes when a survey achieves high levels of electronic response. The subsequent, letter-only, direct-mail operations targeted to past electronic reporters in a sample cost less to print, are easier and faster to assemble, and represent lower and more uniform package weights which reduces postage costs. In fact, for the Commodity Flow Survey (CFS), almost all electronic reporters provided a fax number as part of a previous response. Faxing letter-only outreach materials to generate electronic response is even cheaper than mailing them. For the 2012 CFS, a complete response analysis comparing the effectiveness of form versus letter-only mail and fax operations is not possible at this time. The survey is still conducting data collection. However, what has been measured, is that while overall response rates have declined since the first quarter of data collection, perhaps in part because of a significant reporting burden, the proportion of firms responding electronically has increased over the course of conducting the survey. 7.2 Infusing Data Visualization through Electronic Reporting and Dissemination Objective: transform to an E-Gov environment4 To broaden influence and attract new data users, combine: Informing the public using data visualizations Online survey reporting x The creation and publishing of new data visualizations costs time, resources, and money. However, statistical agencies can market these visualizations for free using existing touch points with respondents. One way is to tap into web traffic flows from respondents reporting for surveys electronically by showcasing relevant visualizations for each particular reporting group. When a respondent logs into the Internet to report for a survey, statistical agencies can use metadata associated with a survey’s User IDs and Passwords to drive custom web creation and personalize web content. Statistical agencies can further improve data dissemination to respondents by grouping data visualizations at national, state, or local levels to help create profiles or special reports that integrate: Demography Industry (if applicable to survey) Geography Organization Scope (if applicable to survey) For instance, if the Census Bureau is asking questions for a manufacturing survey only, then upon completion of electronic data collection, the user can be directed to industry, demographic, and geographic data that pertains to the scope of what they do as an organization. A respondent can be 4 Lee, Ronald; Mesenbourg, Thomas, The Electronic Reporting Option – Does It Make Sense or Cents? ICES-III Presentation Slide 3 10 converted into a data user by introducing statistical products catered to their specific industry, geography, or other characteristics. Adding value at the point of reporting could compel future respondents to report to Census surveys not only out of legal requirement or obligation to country, but also out of direct interest in the work of an organization. Dissemination using the electronic reporting tool could add more context to the reporting experience for a survey, and could be a natural bridge to show respondents how Census or sponsor agencies are using their responses or how a survey applies to them. If executed effectively, and high electronic response rates are achieved and maintained, this strategy paired with the responsive outreach approach in 7.1 could translate into significant processing cost savings. Implementing an integrated electronic reporting data visualization marketing strategy can act as a vehicle to deliver innovative forms of statistical products to new users. However, within the scope of this recommendation, the true catalyst to spark a cultural shift in the way an agency conducts business must include infusing data visualization consistently through data collection, internal data review and analysis, and data dissemination. The following survey concept is one example of how this can be achieved, and is an immediate opportunity to expand the applicability of an existing, custom, mapping tool more broadly through new functionality that affects different business processes. 7.3 An Origin-Destination Survey Concept Objectives: provide value-added services to businesses; improve data quality and reduce processing costs5 Currently, major Economic Directorate components and functions for individual surveys, such as electronic reporting, internal data review and analysis, and data dissemination, are often completed in separate systems. In the Economic Programs Area, one prospect to pilot an initiative where electronic reporting, internal data review and analysis, and data dissemination are all components housed in the same software application is the Commodity Flow Survey (CFS). The CFS paper and electronic data collection tools ask respondents to report for total number of shipments originating from an establishment location. In addition, the CFS collects a sample of these total shipments in the form of domestic and foreign shipment characteristics by destination and mode of transport. Reporting burden can be significant, approximately two hours per form, as respondents are instructed to select and document a sample of up to 40 shipments. An electronic data collection instrument has recently been developed for 2012; this is the first time online reporting is being offered for the CFS. A challenge during development was incorporating the two-page shipments table into a user-friendly, online format. Prior to 2012, reporting for this section of the survey could only happen by respondents filling out the two-page shipments table found within the paper form. Online reporting for the 2012 Commodity Flow Survey gave respondents the option of either uploading a spreadsheet of data, or navigating 40 individual shipment screens on the Internet reporting tool, Centurion. Both online options have already been tested and proven to be effective in the field, and it is not being suggested they be replaced moving forward. However, a new reporting option is suggested, that could give the respondent the opportunity to map their shipments interactively while building the full shipments table off-screen. At log-in, metadata associated with an establishment’s User ID and Password positions and zooms a “You Are Here” on a map. The user then proceeds to enter different domestic and foreign shipment destinations, creating single and multi-mode shipment paths. After each shipment path is locked-in, a pop-up box allows the user to fill out remaining shipment characteristics before proceeding to draw the next shipment path. Respondents would repeat for the full sample of shipments on one map, and then be prompted to review their answers prior finalizing a submission. 5 Lee, Ronald; Mesenbourg, Thomas, The Electronic Reporting Option – Does It Make Sense or Cents? ICES-III Presentation Slide 3 11 Developing an interactive electronic collection instrument could make online reporting for the survey easier and more enjoyable. More experienced Internet users might find the design more user-friendly. Many respondents may already be familiar with this type of interaction through using services like Google Maps to get driving directions. If executed and marketed correctly, the instrument design could potentially increase online response rates over the course of the four quarters of this quinquennial survey. Real data quality benefits could be achieved by more effectively engaging each respondent, receiving more pre-edited data, and allowing Census analysts to review web-maps of shipment data as they were reported in addition to data tables from edits or queries, as a separate component of the same software mapping application. Using a sidebar of all shipment characteristic variables, an analyst could select one or a combination of variables to assist in data review during data collection. After variable selection of, say, a shipment value range ($), mode of transport, and city, an animation could be run, skipping from one ID shipment map to the next, in a string of short time increments. The new reporting tool could also be used during internal data analysis. Data quality edits built into the mapping tool serve as another way of isolating groups of records for analyst review. The review process could simply take place within one firm’s responses to a survey, or it could span survey response for groups of firms with similar characteristics, such as region or commodity classification. The review process could also isolate all records failing the same edit within a given timeframe, identify outliers, or find failures for any combination of inter-related survey variables tied together by edit checks. Paper data capture records could be loaded to the tool to take advantage of the interactive review environment. Individual shipment paths overlaid on one map could be another way to review a limited amount of reported data at the same time. But as the volume of shipment records being visualized on one map increases, the visualization feature utilized by the interactive tool would have to change. Heat maps, with colored shades representing quantity ranges, would be a good way to visualize, say, total weights of all reported domestic railroad shipments originating in Los Angeles by destination location during April 2012. In this example, the geographical destination quantities would be displayed as color-shaded states, cities, or counties where rail shipments originating from Los Angeles were eventually delivered. A month-tomonth animation of this query could be useful in displaying the established rail distribution channels stemming from L.A. A small change to the query could show the pattern of individual commodity groupings being transported over the course of a full year by rail, and give an analyst insight into types and volumes of commodities being imported by sea to the port of Los Angeles6. Colored, shaded, heat maps displaying data by selecting survey variables are the most natural bridge from shipment path micro data review to statistical dissemination with the envisioned interactive tool. Incorporating data visualization into the means by which micro- and macro-data is reviewed could have quality implications, and by augmenting and enhancing the abilities of analysts during this step, one could increase the speed of the review process, discovering new patterns in the reported, edited, and publication data, and giving analysts a deeper understanding of the survey data. A robust, front-end, interactive, mapping, electronic data collection instrument and review tool with the power of time-series animation and variable series selection can also be used seamlessly for data dissemination. With open architecture software development7 and a sound marketing strategy in place connecting the Census data provider base, historical, sponsor agency dissemination products could be pulled in reverse to 6 “California’s Ports: The fickle Asian Container” The Economist: http://www.economist.com/node/21543551 “the confluence of business and technical practices yielding modular, interoperable systems that adhere to open standards with published interfaces.” 7 12 service the very people providing the source of Census data on a just-in-time basis, using the same frame of reference as the online reporting tool. Development costs could be shared across the agencies sponsoring and administering the survey. So, instead of maintaining one tool for electronic reporting, converting and transferring data to use a different system for editing, and transferring and repackaging data again for dissemination, a coordinated effort could produce a multi-faceted interface with functionality that spans multiple business processes. A singular focal point could eventually produce a tool with quality features unmatched by components developed and maintained in isolation. The breadth of this development work would span three or more directorates within the Census Bureau and at least one sponsoring agency. There are much simpler ways to use the electronic reporting instrument as a dissemination tool using static web-content at no additional cost to a sponsor. Static content can still result in a personalized online experience using any variable in a Centurion input file. Whether the content is static or dynamic, this could be one aspect of a broader, agency-wide dissemination strategy. 8. Conclusions Statistical agencies must build the competency to directly connect with and hold the interest of all respondents. At the same time, these agencies must also become better at promoting their statistical products and increasing the reach of their dissemination efforts. These strategies are critical in the uncertain future of government funding, respondent tendencies, and public need for vital data. The individual strategies discussed in this report do not come without risk. For instance, disseminating poor quality data visualizations using the electronic reporting instrument, or prematurely launching an interactive tool being developed for public use could give respondents a negative impression of an agency’s usage of their survey responses. These mistakes could negatively affect future response rates. Furthermore, a greater dependency on Internet reporting and electronic data dissemination can lead to increased IT security threats. Census data and corresponding visualizations would likely come under increasing public scrutiny. Allowing respondents to comment in a public forum on Census visualizations would require internal public relations oversight. In-house, Census might want to establish an expert team of graphics designers and editors to coordinate the standards for and creation of the new visuals meant for public consumption. Overcoming these obstacles may be justified, because as budgets for all government agencies come under closer review by Congress, there is a growing need to prove the value of Census data to the general public. But the greater motivation, and the responsibility that Census has as a body of public servants, is to make statistics relevant and meaningful to the various survey groups who support the mission of the Bureau with their time and responses. Evolving social norms underscore the need for increased Internet reporting options. Statistical agencies must continue to improve these response options. Added benefits of online reporting that reinforce this progression are: Quicker response and faster turn-around time to exclude cases from follow-up contact operations Data entry is done by the respondents and not by Census Bureau staff Built-in edits can help correct data that is entered incorrectly by respondents Cleaner flow of questions and built-in skip patterns Entered data can be stored more easily for future reference by respondents After management has made the commitment to support an electronic reporting option as a mode of collection for a survey, then consider the cost to develop the online instrument as money spent. The future of generating processing cost savings for multi-mode surveys with an electronic reporting option will happen jointly by: 1) identifying respondents’ preferred mode of reporting by referencing individual response histories, and translating that preference into a categorical assignment of future first-contact 13 outreach operations, while 2) reducing the volume of mail-back, fax-back, telephone follow-up, or personal visits by maximizing electronic response rates earlier in an observation’s lifespan in a sample and survey. By doing so, survey managers will begin to minimize the hand touching of paper forms, the additional processing steps associated with routing, capturing, and storing paper records, the redundancy of outreach efforts that result from not recognizing and targeting preferred collection modes, and the costly human capital required to personally follow-up with individual respondents. Processing cost savings can be monitored and reinvested into web-development during previously frozen periods of online content design and funding, in preparation for dissemination that coincides with macro-program change periods for each survey. The end goal is to not completely eliminate all modes of collection other than the Internet, rather, to use the other modes to drive the maximum amount of traffic to the web for reporting, and to spark public interest in the usefulness of statistics. 14 Acknowledgments A special thanks to Nancy Piesto for giving me the authority to conduct the study, and for being instrumental in collecting the modal response rate data from Census managers. Thank you to Michael Kornbau, Michelle Karlsson, Michael Padgett, Ian Hull, and William Davie Jr. for reviewing the paper, providing comments, helping to reorganize the materials, and assisting in the finalization of the report and graphics for publication. A special thanks to Ian Hull for editing the paper, and to Michael Padgett for producing the motion charts in the conference presentation. References “California’s Ports: The fickle Asian container.” The Economist. China and the paradox of prosperity. 2012, January 28th – February 3rd. Retrieved January 28, 2012, from http://www.economist.com/node/21543551. Guertin, Nick (2010). “Naval Open Architecture Contract Guidebook For Program Managers, Version 2.0, June 30, 2010, Washington, Navy Yard, D.C.: Department of the Navy: 2. Herr, Phillip (2010). “GAO-10-538T, U.S. Postal Service: Financial Crisis Demands Aggressive Action.” Testimony Before the Subcommittee on Financial Services and General Government, Committee on Appropriations, U.S. Senate, March 18, 2010, Washington, D.C.: United States Government Accountability Office: 1-9. Lee, Ronald; Mesenbourg, Thomas (2007). “The Electronic Reporting Option – Does It Make Sense or Cents?” Proceedings of the Third International Conference of Establishment Surveys, June 18-21, 2007, Montreal, Canada: American Statistical Association: Presentation Slide 3. 15
© Copyright 2026 Paperzz