首页    期刊浏览 2025年03月02日 星期日
登录注册

文章基本信息

  • 标题:Is it the program or the interpreter? Modeling the influence of program characteristics and interpreter attributes on visitor outcomes.
  • 作者:Powell, Robert B. ; Stern, Marc J.
  • 期刊名称:Journal of Interpretation Research
  • 印刷版ISSN:1092-5872
  • 出版年度:2013
  • 期号:July
  • 语种:English
  • 出版社:National Association for Interpretation
  • 摘要:Much has been written regarding the interpretive techniques that should be employed to enhance visitor outcomes (e.g., Ham, 1992, 2013; Moscardo, 1999; Knudson, Cable, & Beck, 2003; Brochu & Merriman, 2002; Lewis, 2005). These interpretive techniques, which we refer to as program characteristics, are the focus of training efforts offered by organizations such as the National Park Service (see NPS, 2003a,b,c,d,e,f) and the National Association for Interpretation (NAI) as well as college courses offered around the world. These characteristics are believed to improve the quality of interpretive communications and to contribute to reaching desired outcomes, such as inspiring audiences to form intellectual and emotional connections with interpreted resources, influencing attitudes, and in some cases motivating behaviors. Researchers and field interpreters also recognize that there are other factors, such as the attributes of the interpreter (confidence, charisma, enthusiasm, passion, apparent knowledge, etc.) that may influence the effectiveness of an interpretive program (e.g., Ham & Weiler, 2002a). However, these attributes are often overlooked in research and interpretive training. Stern and Powell (article 1, this issue) and Powell and Stern (article 3, this issue) investigated live interpretation programs provided by the U.S. National Park Service (NPS) to examine the relationship between 56 different interpretive practices and interpreter attributes and visitor satisfaction, enjoyment and appreciation for resources, and behavioral intentions. The results suggest that only certain program characteristics and interpreter attributes were significantly related with these outcomes. This study seeks to extend these findings by modeling the relative influence of these program characteristics and interpreter attributes on visitor outcomes using structural equation modeling (SEM).
  • 关键词:Interpretive programs (Parks and museums);Regression analysis

Is it the program or the interpreter? Modeling the influence of program characteristics and interpreter attributes on visitor outcomes.


Powell, Robert B. ; Stern, Marc J.


Introduction

Much has been written regarding the interpretive techniques that should be employed to enhance visitor outcomes (e.g., Ham, 1992, 2013; Moscardo, 1999; Knudson, Cable, & Beck, 2003; Brochu & Merriman, 2002; Lewis, 2005). These interpretive techniques, which we refer to as program characteristics, are the focus of training efforts offered by organizations such as the National Park Service (see NPS, 2003a,b,c,d,e,f) and the National Association for Interpretation (NAI) as well as college courses offered around the world. These characteristics are believed to improve the quality of interpretive communications and to contribute to reaching desired outcomes, such as inspiring audiences to form intellectual and emotional connections with interpreted resources, influencing attitudes, and in some cases motivating behaviors. Researchers and field interpreters also recognize that there are other factors, such as the attributes of the interpreter (confidence, charisma, enthusiasm, passion, apparent knowledge, etc.) that may influence the effectiveness of an interpretive program (e.g., Ham & Weiler, 2002a). However, these attributes are often overlooked in research and interpretive training. Stern and Powell (article 1, this issue) and Powell and Stern (article 3, this issue) investigated live interpretation programs provided by the U.S. National Park Service (NPS) to examine the relationship between 56 different interpretive practices and interpreter attributes and visitor satisfaction, enjoyment and appreciation for resources, and behavioral intentions. The results suggest that only certain program characteristics and interpreter attributes were significantly related with these outcomes. This study seeks to extend these findings by modeling the relative influence of these program characteristics and interpreter attributes on visitor outcomes using structural equation modeling (SEM).

The interpretive techniques promoted by professional associations and organizations have evolved over many decades and are based on experience, expert consensus, theory, and peer-reviewed research (Skibins et al., 2012). However, the empirical support for many of these best practices is largely anecdotal, because few studies to-date have attempted to isolate the influence of particular practices on outcomes through the use of experimental (or quasi-experimental) designs or comparative approaches (Skibins et al., 2012). The isolation of the influence of particular practices is challenging even with these designs, as program outcomes inevitably emerge from a dynamic interaction between the interpreter, the audience, the content, the setting/context, and the delivery (Powell et al., 2009, 2012; Archer & Wearing, 2003; Wearing & Wearing, 2001). Accounting for all factors seems a near impossibility. In this paper, we explore the relative influence of two of these elements, interpreter characteristics and program characteristics, on visitor outcomes. We define interpreter characteristics as those that may be entirely unique to the individual interpreter in any given context. These elements might include their mood, personality, or particular style of presentation. While program characteristics may also be highly dependent upon the interpreter, they could also be incorporated by design into a pre-packaged program, such as the sequence, content, theme, or logistics of the program.

We use structural equation modeling (SEM) for two reasons. First, the models give us a sense of the relative strength of influence of interpreter and program characteristics on visitor outcomes. The models can reveal the percentage of the observed variance in each outcome that can be explained by the predictors (Byrne, 2006). Second, the models allow for an examination of the interactions between interpreter characteristics and program characteristics. SEM also reveals the most parsimonious causal models for each outcome. As such, only the most predictive combination of variables remains in the final models. Examining which variables are present in the final models and their inter-relationships allows for consideration of the relative influence of program design vs. interpreter attributes. For example, if only interpreter characteristics are present in the final models, we would consider them dominant drivers of visitor outcomes. If both interpreter and program characteristics are present, it would support a view that outcomes are produced more by the interactions between interpreter and program design rather than by one or the other.

Methods

Selection of sites

We observed 376 diverse interpretive programs provided by 24 NPS park units across six regions of the NPS that generally reflected the current makeup of the NPS system (see Stern and Powell, this issue). The criteria for selecting NPS units included:

* Annual visitation greater than 35,000

* Geographic distribution across the county

* Variable distances from urban centers (urban, urban proximate, remote)

* Resource-base (cultural, natural, mixed)

* The ability to observe multiple programs in a short period of time

* Willingness to participate

The 24 selected units varied widely in terms of visitation, resource base, and locations, providing a reasonable sample from which to make generalizations regarding interpretation provided across the NPS system.

Sampling and data collection

Four researchers collected field data. Prior to each program one researcher conducted a short interview with the interpreter to collect demographic and background information regarding the program. During the program, this same researcher monitored 56 different program and interpreter characteristics and recorded these details on standardized observation sheets. After the program, we surveyed attendees that were age 15 or older using a standardized questionnaire. For programs with fewer than 50 participants, we attempted a census of all eligible attendees. In programs with more than 50 attendees, we systematically sampled attendees. From the 376 programs, we collected 3,603 surveys from visitors (for more detail, see Stern & Powell, this issue).

Data cleaning

Post-program surveys and program audits were coded and entered into Microsoft Access Database and Microsoft Excel to facilitate data entry. Data were then transferred to SPSS and EQS v6.1 software (Bentler, 2005) for screening and analyses. The visitor survey data were first screened for cases missing more than 50% of the items per factor (Tabachnick & Fidell, 2007). A total of 118 respondents were removed as a result. Data were then screened for univariate and multivariate outliers on outcome variables following Tabachnick and Fidell (2007) using Mahalanobis Distance (MAH) and studentized deleted residuals (SDRESID). A total of 58 cases were removed for exceeding +/- 3 standard deviations, or the criterion Mahalanobis Distance value (Fox 1991). This reduced our sample to 3,427 individual surveys from 376 interpretive programs.

Next we reviewed the number of valid respondents per individual interpretive program. Prior theory and research suggest that programs with a low number of attendees may be inherently different than programs servicing a larger number of attendees (Forist, 2003; McManus, 1987, 1988; Moscardo, 1999; Stern & Powell, this issue). We observed 272 programs with five or more attendees (see Stern & Powell, article 1, this issue for more extensive description). We chose this sample for the analyses conducted herein because it is most representative of programs in general and it provides a sample large enough to conduct structural equation modeling (Byrne, 2006). Because the program was our unit of analysis, our final step in data preparation included aggregating individual data at the program level by calculating the mean score of each visitor outcome for each program. For SEM purposes, all data was then grand mean centered (Tabachnick & Fidell, 2007)

Dependent variables: outcomes

Based on extensive input from the NPS and a review of literature, we developed three dependent variables (Table 1). The first dependent variable served as a measure of visitor satisfaction with the program on a scale from 0 to 10, with 0=Terrible and 10=Excellent. Two indexes were developed from other survey items following procedures outlined by DeVellis (2003) to represent visitor experience and appreciation and behavioral intentions. The items comprising each index were measured using a five-point Likert-type scale, with answer choices: Not at all (1), A little (2), Somewhat (3), A moderate amount (4), and A great deal (5). Composite scores were created for each of the scales by taking the mean of all items (for more detail, see Stern & Powell, this issue).

Program and interpreter characteristics

The independent variables used in this SEM analyses included both interpreter and program characteristics that met two criteria. We included ordinal variables that were correlated (p < 0.01) to the particular outcome in question in any context (See Stern & Powell, article 1, and Powell and Stern, article 3, this issue). We also included categorical variables with at least "moderate" effect size in association with the particular outcome in question in any context (Cohen's d > 0.5). The program characteristics (Table 2) were originally drawn from an extensive literature review aimed at identifying best practices i n the field (see Skibins et al., 2012). The interpreter characteristics were developed from the communications and education literature, though many of these factors are also referenced in the interpretation literature (Table 3). The tables also contain descriptive statistics. For more detail, see Stern and Powell (this issue).

Structural equation modeling

We used structural regression modeling (a.k.a. path analysis), a form of SEM, to examine the influence of different program and interpreter characteristics on three outcomes. We used SEM for this analysis because it is confirmatory (as opposed to exploratory) in nature and requires the researcher to have an explicit hypothesized model; it can model measurement error, which reduces inaccuracies; it allows for the analysis of a complete multivariate model including direct and indirect effects and in this case it can assess causal relationships between independent variables and a dependent variable (Byrne, 2006; Kline, 2005). In this study, all independent variables are formative (as opposed to reflective). That is, they were observed and represent a specific practice or attribute that is thought to directly influence the dependent variables (see Kline, 2005; Diamantopoulis et al., 2008; Diamantopoulis & Winklhofer, 2001; Jarvis et al., 2003, Padsakoff et al., 2007 for further explanation).

We used the EQS v6.1 software (Bentler 2005) to perform the statistical analyses, which progressed in several stages. First, the data were screened for univariate and multivariate deviations from normality. Next, we used structural regression modeling to assess the causal relationships between independent variables and each dependent variable (three separate models). For each outcome, we began with a model that contained all interpreter and program characteristics that met the criteria described above for that outcome. The starting list of program practices and interpreter attributes used in the hypothesized models are in Table 4. To develop the final structural regression models we used an iterative process in which diagnostics (modification indices: Lagrange Multiplier Test (LM), Wald Test) indicated potential modifications, including removal of independent variables from the model, to improve fit and parsimony.

Structural regression analysis provides multiple statistics that can be used to evaluate the "fit" of a specified model (Byrne, 2006). In this paper we report the Satorra-Bentler Scaled Chi-Square (S-B [chi square]), Robust Comparative Fit Index (CFI), Standardized Root Mean Square Residual (SRMR), the Robust Root Mean Square Error of Approximation (RMSEA) and its associated 90% confidence interval (Bentler & Yuan, 1999; Byrne, 2006). The S-B [chi square], which should be interpreted like a [chi square], is reported because it corrects for the degree of kurtosis in the data (Satorra & Bentler, 1994). The Robust CFI accounts for non-normality in the data and is an "incremental or comparative fit index" that evaluates the change in fit between the hypothesized model and the "independence model" (Byrne, 2006, 97; Bentler, 1990; Kline, 2005, 140). The independence model assumes that all the variables in the model are unrelated. The CFI represents the total covariation in the data and is measured on a scale of 0 to 1 with values greater than .9 indicating an acceptable fit and values greater than .95 indicating an excellent fit (Byrne, 2006; Hu & Bentler, 1999). The SRMR statistic provides the average difference between the sample and the predicted correlation matrices and thus is not susceptible to non-normality (Byrne, 2006). The SRMR uses standardized values with the range of scores between 0 and 1; values less than .1 are considered acceptable and less than .05 are considered a good fit (Hu & Bentler, 1995; Kline, 2005). The Robust RMSEA also accounts for non-normality in the data and is based on the average lack of fit per degree of freedom; therefore, as the fit improves, the RMSEA decreases. As such, this measure is sensitive to the degrees of freedom and the complexity of the model (Byrne, 2006). Like the SRMR, the scores range between 0 and 1, with values of .05 to .08 deemed acceptable and values less than .05 considered excellent (Browne & Cudeck, 1993; Hu & Bentler, 1999).

Beta weights in structural regression models reflect the effect size of an independent variable on the dependent variable. [R.sup.2] values gauge the predictive validity of the structural model, explaining the proportion of the total observed variance in the dependent variable explained by the model. It is recommended to assess [R.sup.2] values independently of fit indices, as the latter do not pertain to predictive validity (Kline, 2005).

Results

Three models were created based on the list of variables in Table 4. All independent variables (interpreter and program characteristics) were first entered as direct predictors of each outcome. In each case, the initial fit of each model was deemed unacceptable (Byrne, 2006). Through an iterative process, we adjusted the models using diagnostics that indicate potential model changes that would improve fit and parsimony. This generally involves removing variables one at a time based on statistical indicators produced at each stage of the modeling process. As the iterative modeling continues, it also can include adding or changing the nature of relationships between variables. In the end, a single "best fit" model is produced that represents the most parsimonious and predictive model for each outcome. The resulting models are displayed in Figures 1, 2, and 3.

Figure 1 represents the final model pertaining to how the interpreter and program characteristics influenced visitor satisfaction. Fit indices for the final "satisfaction" model (SB[chi square] = 5.39, p < .07; CFI = .99; SRMR = .029; RMSEA = .08) indicated the model was an acceptable representation of the relationships present in the data (Byrne, 2006; Kline, 2005). Authentic emotion was a strong predictor of interpreter's confidence ([beta] = .388, p < .05) and a weaker predictor of visitor satisfaction ([beta] = .171, p < .05). Organization was also a strong predictor of interpreter's confidence (([beta] = .300, p < .05), but not a direct predictor of visitor satisfaction. Confidence was a strong predictor of visitor satisfaction ([beta] = .307, p < .05). Appropriate for the audience was also a significant predictor of visitor satisfaction ([beta] = .183, p < .05). The model accounted for 35% ([R.sub.2]) of the variance in confidence and 27% ([R.sub.2]) of the variance in visitor satisfaction.

[FIGURE 1 OMITTED]

[FIGURE 2 OMITTED]

The final structural regression model for visitor experience and appreciation had the same structure as the final visitor satisfaction model (Figure 2). Fit indices for the model (SB[chi square] = 4.45, p < .1; CFI = .99; SRMR = .027; RMSEA = .069) indicated the model was an acceptable fit of the data (Byrne, 2006; Kline, 2005). The only structural differences between this model and the satisfaction model involved the relative strength of confidence and appropriateness for the audience. Appropriate for audience ([beta] = .288, p < .05) was the strongest predictor of visitor experience and appreciation, followed by authentic emotion ([beta] = .139, p < .05) and confidence ([beta] = .068, p < .05). The model accounted for 35% ([R.sub.2]) of the variance in confidence and 16% ([R.sub.2]) of the variance in visitor experience and appreciation.

[FIGURE 3 OMITTED]

The model in Figure 3 represents how interpreter and program characteristics predicted intentions to change behaviors. Fit indices for the model in Figure 3 (SB[chi square] = 7.38, p < .05; CFI = .96; SRMR = .040; RMSEA = .03) indicated the model was an acceptable representation of the relationships present in the data. Having a goal to influence behavior ([beta] = .145, p < .05), appropriate logistics ([beta] = .153, p < .05), and humor quality ([beta] = .223, p < .05) were significant positive predictors of intentions to change behaviors. Use of sarcasm ([beta] = -.170, p < .05) was a significant but negative predictor of intentions to change behaviors. The model accounted for 10% ([R.sub.2]) of the variance in intentions to change behaviors.

Discussion: Is it the interpreter or the program?

We used structural equation modeling to examine the relative influence of interpreter and program characteristics upon visitor outcomes at live interpretation programs across the U.S. National Park Service. The resulting models reveal three main lessons. First, it appears in each case that both interpreter and program characteristics influenced visitor outcomes. Second, depending on outcome, certain program practices and interpreter attributes provided the best model fit and predictive power. Third, the final models accounted for a relatively low percentage of the overall variance in visitor outcomes. We explain each finding and some important limitations in the interpretation of the analyses below.

In each model, both interpreter and program characteristics influenced outcomes. The satisfaction and the visitor experience and appreciation models each contained authentic emotion and charisma, organization, confidence, and appropriate for the audience. In each model, authentic emotion and charisma and organization were mediated by confidence. In other words, the model suggests that authentic emotion and charisma and organization each help to create interpreter confidence, which in turn enhances visitor outcomes. Authentic emotion and charisma also served as a direct causal predictor of each outcome, as did the appropriate for the audience variable.

The final structural regression model of intentions to change behaviors suggests that humor quality, appropriate logistics, and intending to influence behaviors through a program positively influenced intentions to change stewardship behaviors. The use of sarcasm was associated with weaker intentions to change stewardship behaviors. In other words, interpreters that successfully employed humor, ensured that their audience's needs were met, and explicitly intended to influence their audience's behaviors were more successful at doing so. Meanwhile, overly sarcastic interpreters were less likely to influence changes in behavioral intentions. Interestingly, only 7% of all interpreters interviewed in the study explicitly intended to influence audience behaviors (Table 3). Ham (2013) reminds interpreters that outcomes, such as behavior change, do not happen magically; instead a program should be planned and developed with an outcome in mind. When focusing on behavior change, numerous techniques may increase the likelihood of influencing specific behaviors (Ham et al., 2007; Powell & Ham, 2008; Stern & Powell, this issue).

Certain limitations in the data and analyses are important to consider when interpreting these findings. First, structural equation modeling explicitly aims to produce the most parsimonious predictive or in this case causal model for selected outcomes. As such, independent variables that may be strongly related to outcomes are commonly removed during the modeling process due to their relationships with other variables. For example, the connection variable is highly correlated with organization, authentic emotion and charisma, and confidence (see Stern et al., this issue). As a result, it may be removed from a model because it explains a redundant proportion of the variance in the outcome as the other independent variables. This is the case with many of the program practices and interpreter characteristics tested in this analysis. It would be inappropriate to assume that their absence in the final models reduces their significance in influencing more positive outcomes.

Second, the models accounted for 10% to 27% of the variance in the outcomes. The strongest model accounted for 27% of the variance in satisfaction. The weakest model accounted for 10% of the variance in behavioral intentions. This suggests that much more is at play than simply the interpreter and the program elements. Interpretive programs are complex phenomena, and audience outcomes can be influenced by characteristics of the individual audience members, the makeup of the group, and the location and context of the program, in addition to characteristics of the program and the interpreter (Powell et al., 2009). Past research into communications (see Ajzen, 1992, for more) suggests that few consistent trends emerge when attempting to examine the range of source (interpreter) factors, receiver (audience) factors, channel (program) factors, and message (content) factors that influence outcomes resulting from communications. These factors vary with each program and produce an almost unlimited number of interactions and potential combinations (Falk, 2004). We examine a small portion of these additional factors in a separate article in this issue (Powell & Stern, this issue).

Relatively low [R.sub.2] values may also be a product of the lack of variance observed in satisfaction and visitor experience and appreciation scores. We discuss this issue in greater depth in a separate article in this issue as well (Stern et al., this issue). Predictive ability may be particularly low for behavioral change for a number of reasons. As noted earlier, few programs actually targeted behavioral change as an outcome. As such, changing behavioral intentions may have been more of a side effect than an intended outcome of a program. Moreover, many interpretive program goers may already perform many of the behaviors discussed in interpretive programs, leaving little room for change (see Stern & Powell, this issue, for a more detailed discussion).

Despite the limitations, the results suggest that outcomes are influenced by both program and interpreter characteristics and that these characteristics interact and influence each other. For example, confidence may ultimately emerge from an interpreter's passion for the resource and careful planning, which leads to good organization. Because most prior research and formal training have focused on what we have categorized as "program characteristics" (Skibins et al., 2012), we urge future researchers, trainers, and practitioners to give some meaningful attention to interpreter attributes and delivery styles. Training programs might add elements that could improve interpreters' abilities to project confidence and authentic emotion. Some lessons for doing so might be found in the formal education field, where "affinity-seeking" and immediacy behaviors have garnered some attention (e.g., Finn et al., 2009). These practices involve efforts to ingratiate teachers with their students by reducing the social distance between them (see also Stern & Powell, this issue; Stern et al., in press). Interpretive organizations might also consider these findings in light of the role of the individual interpreter in program development. If organizations can provide opportunities for creating and sustaining authentic connections between interpreters and the resources they interpret, they might enhance interpreters' abilities to convey their own passions to their audiences. Finally, we urge researchers to consider how different program and interpreter characteristics may function differently in varying contexts.

References

Ajzen, I. (1992). Persuasive communication theory in social psychology: A historical perspective. In M. J. Manfredo (Ed.), Influencing human behavior (pp. 1-27). Champaign, IL: Sagamore Publishing.

Archer, D., & Wearing, S. 2003. Self, space, and interpretive experience: The interactionism of environmental interpretation. Journal of Interpretation Research, 8 (1): 7-23.

Beck, L., & Cable, T. T. (2002). Interpretation for the 21st century: Fifteen guiding principles for interpreting nature and culture (2nd ed.). Champaign: Sagamore.

Bentler, P.M. 1990. Comparative fit indexes in structural models. Psychological Bulletin, 107: 238-46.

Bentler, P.M. 2005. Eqs 6 structural equations program manual. Encino, CA: Multivariate Software (www.mvsoft.com).

Bentler, P.M. and K.H. Yuan. (1999). Structural equation modeling with small samples: Test statistics. Multivariate Behavioral Research, 34: 181-97.

Brochu, L., & Merriman, T. (2002). Personal interpretation: Connecting your audience to heritage resources: Fort Collins, CO: InterpPress.

Browne, M.W., & Cudeck, R. (1993). Alternative ways of assessing model fit. In Testing structural equation models, ed. K.A. Bollen and J.S. Long, 445-55. Newbury Park, CA: Sage.

Byrne, B.M. (2006). Structural equation modeling with eqs: Basic concepts, applications and programming. Second ed. Mahwah, NJ: Erlbaum.

DeVellis, R.F. (2003). Scale development: Theory and applications Applied social research methods. 2nd ed. Thousand Oaks, CA: Sage Publishing.

Diamantopoulos, A., Riefler, P., & Roth, K. P. (2008). Advancing formative measurement models. Journal of Business Research, 61(12), 1203-1218.

Diamantopoulos, A., & Winklhofer, H.M. (2001). Index construction with formative indicators: An alternative to scale development. Journal of Marketing Research, 38(2); 269-277

Falk, J. (2004). The director's cut: Toward an improved understanding of learning from museums. Science Education, 88(S1), S83-S96. doi: 10.1002/sce.20014

Finn, A. N., Schrodt, P., Witt, P. L., Elledge, N., Jernberg, K. A., & Larson, L. M. (2009). A meta-analytical review of teacher credibility and its associations with teacher behaviors and student outcomes. Communication Education, 58(4), 516-537.

Forist, B. (2003). Visitor Use and Evaluation of Interpretive Media. A Report on Visitors to the National Park System. National Park Service Visitor Services Project. http:// nature.nps.gov/socialscience/ docs/Visitor_Use_and_Evaluation.pdf. Accessed May 24, 2012.

Fox, J. 1991. Regression diagnostics. Newbury Park, CA: Sage.

Frauman, E., & Norman, W.C. (2003). Managing visitors via "mindful" information services: One approach in addressing sustainability. Journal of Park and Recreation Administration, 21 (4): 87-104.

Ham, S. (2009). From interpretation to protection: Is there a theoretical basis? Journal of Interpretation Research 14(2): 49-57.

Ham, S. H. (1992). Environmental interpretation: A practical guide for people with big ideas and small budgets. Golden, CO: Fulcrum Publishing.

Ham, S. H. (2013). Interpretation: Making a difference on purpose. Golden, CO: Fulcrum Publishing.

Ham, S. H., Brown, T. J., Curtis, J., Weiler, B., Hughes, M., & Poll, M. (2007). Promoting persuasion in protected areas: A guide for managers. Developing strategic communication to influence visitor behavior. Southport, Queensland, Australia: Sustainable Tourism Cooperative Research Centre.

Ham, S.H., & Weiler, B.M. (2002a). Toward a theory of quality in cruise-based nature guiding. Journal of Interpretation Research, 7 (2): 29-49.

Ham, S H., & Weiler, B.M. (2002b). Tour guide training: a model for sustainable capacity building in developing countries. Journal of Sustainable Tourism, 10(1): 52-69.

Hu, L.-T., & P.M. Bentler. (1995). Evaluating model fit. In Structural equation modeling: Concepts, issues, and applications, ed. R. H. Hoyle, 76-99. Thousand Oaks, CA: Sage.

Hu, L.-T., & P.M. Bentler. (1999). Cutoff criteria for fit indices in covariance structure analysis: Guidelines, issues, and alternatives. Structural Equation Modeling, 6: 1-55.

Jacobson, S. K. (1999). Communication skills for conservation professionals. Washington, D.C.: Island Press.

Jarvis, C.B., MacKenzie, S.B., & Podsakoff, P.M. (2003). A critical review of construct indicators and measurement model misspecification in marketing and consumer research. Journal of Consumer Research, 30(2); 199-218.

Kline, R.B. (2005). Principles and practice of structural equation modeling. 2nd ed. New York: The Guilford Press.

Knapp, D., & Benton, G. M. (2004). Elements to successful interpretation: A multiple case study of five national parks. Journal of Interpretation Research, 9(2), 9-25.

Knapp, D. & Yang, L. (2002). A phenomenological analysis of long-term recollections of an interpretive program. Journal of Interpretation Research, 7(2), 7-17.

Knudson, D. M., Cable, T. T., & Beck, L. (2003). Interpretation of cultural and natural resources, (2nd ed.). State College: Venture Publishing.

Larsen, D. L. (2003). Meaningful Interpretation: How to Connect Hearts and Minds to Places, Objects, and Other Resources. Fort Washington: Eastern National.

Lewis, W. J. (2005). Interpreting for park visitors (9th ed.). Fort Washington: Eastern National.

Madin, E. M. P., & Fenton, D. M. (2004). Environmental interpretation in the Great Barrier Reef Marine Park: An assessment of programme effectiveness. Journal of Sustainable Tourism, 12(2), 121-137.

McCroskey, J.C., Richmond, V.P., & Stewart, R.A. (1986). One on one: The foundations of interpersonal communication. Englewood Cliffs, NJ: Prentice-Hall.

McManus, P.M. (1987). It's the company you keep ... The social determination of learning-related behaviour in a science museum. The International Journal of Museum management and Curatorship, 6: 263-270.

McManus, P. M. (1988). Good companions: More on the social determination of learning-related behaviour in a science museum. International Journal of Museum Management and Curatorship, 7(1), 37-44.

Moscardo, G. (1999). Making visitors mindful: Principles for creating quality sustainable visitor experiences through effective communication. Champaign: Sagamore.

National Park Service. (2003a). Interpretive Development Program. Module 101: Fulfilling the NPS mission: The process of interpretation.

National Park Service. (2003b). Interpretive Development Program. Module 103: Preparing and presenting an effective interpretive talk.

National Park Service. (2003c). Interpretive Development Program. Module 210: Prepare and present an effective conducted activity.

National Park Service. (2003d). Interpretive Development Program. Module 220: Prepare and present an interpretive demonstration or other illustrated program.

National Park Service. (2003e). Interpretive Development Program. Module 230: Interpretive writing.

National Park Service. (2003f). Interpretive Development Program. Module 311: Interpretive media development.

Podsakoff, P.M., MacKenzie, S.B., Lee, J.Y., & Podsakoff, N.P. (2007). Common method biases in behavioral research: A critical review of the literature and recommended remedies. Journal of Applied Psychology, 88(5) 879-903.

Powell, R.B., Brownlee, M.T.J., Kellert, S. R. & Ham, S.H. (2012) From awe to satisfaction: Immediate affective responses to the Antarctic tourism experience. Polar Record. 48(2): 145-156.

Powell, R.B., Kellert, S. R., & Ham, S.H. (2009). Interactional theory and the sustainable nature-based tourism experience. Society and Natural Resources, 22(8): 761-776.

Powell, R. B., & Ham, S. H. (2008). Can ecotourism interpretation really lead to pro-conservation knowledge, attitudes and behavior? Evidence from the Galapagos Islands. Journal of Sustainable Tourism, 16(4): 467-489.

Regnier, K., Gross, M., & Zimmerman, R. (1992). The Interpreter's Guidebook: Techniques for Programs and Presentations. Interpreter's Handbook Series. Stevens Point: UWSP Foundation Press.

Satorra, A., & Bentler, P. M. (1994). Corrections to test statistics and standard errors on covariance structure analysis. In A. von Eye & C. C. Clogg (Eds.), Latent variables analysis (pp. 399-419). Thousand Oaks, CA: Sage.

Skibins, J.C., Powell, R.B., & Stern, M.J. (2012). Exploring empirical support for interpretation's best practices. Journal of Interpretation Research. 17(1): 25-44.

Sharpe, G. W. (1976). Interpreting the environment. New York: John Wiley & Sons.

Stokols, D., & Altman, I. eds. (1987). Handbook of environmental psychology. New York: John Wiley & Sons.

Tabachnick, B.G. and L.S. Fidell. (2007). Using multivariate statistics (5th ed.). Needham Heights, MA: Allyn & Bacon.

Tilden, F. (1957). Interpreting our heritage (3rd ed.). Chapel Hill: The University of North Carolina Press.

Veverka, J. A. (1998). Interpretive master planning: the essential planning guide for interpretive centers, parks, self guided trails, historic sites, zoos, exhibits and programs (2nd ed.). Tustin: Acorn Naturalists.

Wearing, S., & Wearing, B. (2001). Conceptualizing the selves of tourism. Leisure Studies, 7: 11-23.

Ward, C.W., & Wilkinson, A. E. (2006). Conducting meaningful interpretation: A field guide for success. Golden: Fulcrum.

Robert B. Powell

Department of Parks, Recreation and Tourism Management and School of Agricultural

and Forest Environmental Sciences, Clemson University

Marc J. Stern

Department of Forest Resources and Environmental Conservation, Virginia Tech
Table 1. Outcome (dependent) variables with descriptive statistics.

Outcomes                                         N    Mean   St. Dev.

Satisfaction: 0 to 10 scale                     272   8.94     0.64

Visitor experience and appreciation             272   4.41     0.32
  ([alpha] = .89): 1 to 5 scale

* Made my visit: to this park more enjoyable          4.55     0.30
* Made my visit to this park more meaningful          4.49     0.32
* Enhanced my appreciation for this park              4.36     0.37
* Increased my knowledge about the                    4.45     0.34
  program's topic
* Enhanced my appreciation for the National           4.27     0.36
  Park Service

Behavioral intentions ([alpha] = .94):          272   2.92     0.64
  1 to 5 scale

* Changed the way I will behave while I'm             2.92     0.67
  in this park
* Changed the way I will behave after I               2.92     0.61
  leave this park

Table 2. Program characteristics, their definitions, and descriptive
statistics.

Program          Definition
characteristic

Organization     Equally weighted composite mean score of 6 program
([alpha] =       characteristics:
0.82)
Scale:           * Quality of the introduction (Brochu and Merriman,
1 to 5           2002; Ham, 1992; Jacobson, 1999): Degree to which
Mean: 3.34       the introduction captured the audience's attention
S.D.: 0.71       and oriented (or pre/disposed) the audience to the
                 program's content and/or message.

                 * Appropriate sequence (Beck and Cable, 2002; Ham,
                 1992; Jacobson, 1999; Larsen, 2003): Degree to
                 which the program followed a logical sequence.

                 * Effective transitions (Beck and Cable, 7742;
                 Brochu and Merriman, 7742; Ham, 3342; Jacobson,
                 1999; Larsen, 2003): Degree to which program used
                 appropriate transitions that kept the audience
                 engaged and did not detract from the program's
                 sequence.

                 * Holistic story (Beck and Cable, 2002; Larsen,
                 2003; Tilden, 1957): Degree to which the program
                 aimed to present a Holistic story (with characters
                 and a pilot) as opposed to disconnected pieces of
                 information.

                 * Clarity of theme (Beck and Cable, 2002; Brochu
                 and Merriman, 2002; Ham, 1992; Jacobson, 1999;
                 Knudson, Cable, and Beck, 2003; Larsen, 2003;
                 Lewis, 2005; Moscardo, 1999; Sharpe, 1976; Veverka,
                 1998; Ward and Wilkinson, 2006): Degree to which
                 the program  had a clearly communicated theme(s). A
                 theme is  defined as a single sentence (not
                 necessarily  explicitly stated) that links
                 tangibles,  intangibles, and universals to organize

                 and develop  ideas.

                 * Link between introduction nnd conclusion (Beck
                 and Cable, 2002; Brochu and Merriman, 2002; Larsen,
                 2003): Degree to which program connected conclusion
                 back to the introduction in an organized or
                 cohesive way (i.e., program "came full circle.")

Connection       Equally weighted composite mean score of 5 program
([alpha] =       characteristics
0.88)
Scale: 1 to 5    * Link tangibles to intangible meanings and
Mean: 2.77       universal concepts (NPS Module 101; Beck and Cable,
S.D.: 0.78       2002; Brochu and Merriman, 2002; Ham, 1992;
                 Knudson, et al, 2003; Larsen, 2003; Lewis, 2005;
                 Moscardo, 1999; Tilden, 1957; Ward and Wilkinson,
                 2006): Communication connected tangible resources
                 to intangibles and universal concepts.

                 * Cognitive engagement (Knudson, et al., 2003;
                 Moscardo, 1999; Sharpe, 1976; Tilden, 1957;
                 Veverka, 1998)): Degree to which the program
                 cognitively engaged audience members in a
                 participatory experience beyond simply listening;
                 i.e. calls to imagine something, reflect, etc.

                 * Relevance to audience (Beck and Cable, 2002;
                 Brochu and Merriman, 2002; Ham, 1992; Jacobson,
                 1999; Knapp ami Benton, 2004; Lewis, 2005;
                 Moscardo, 1999; NPS Module 101; Sharpe, 1976;
                 Tilden, 1957; Veverka, 1998): Degree to which the
                 program explicitly communicated the relevance of
                 the subject to the lives of the audience.

                 * Affective messaging (Jacobson, 1999; Lewis, 2005;
                 Madin and Fenton, 2004; Tilden, 1957; Ward and
                 Wilkinson, 2006): Degree to which the program
                 communicated emotion (in terms of quantity, not
                 quality).

                 * Provocation (Beck and Cable, 2002; Brochu and
                 Merriman, 2002; Knudson, et al., 2003; Tilden,
                 1957): Degree to which the program explicitly
                 provoked participants to personally reflect on
                 content and its deeper meanings.

Appropriate      Degree to which basic audience and program needs
logistics        were met (i.e., restrooms, weather, technology,
Scale: 1 to 4    accessibility, shade, etc). (Jacobson, 1999;
Mean: 3.71       Knudson et al., 2003)
S.D.: 0.93

Appropriate      Degree to which the program aligned with audience's
for audience;    ages, cultures, and level of knowledge, interest,
Scale: 1 to 5    and experience. (Beck and Cable, 2002; Jacobson,
Mean: 3.93       1999; Knudson et al 2003)
S.D.: 0.70

Multisensory     Degree to which the program intentionally and
Scale: 1 to 3    actively engaged more than just basic sight and
Mean: 2.39       sound. (Beck and Cable, 2002; Knudson et al., 2003;
S.D.: 0.71       Lewis, 2005; Moscardo, 1999; Tilden, 1957; Veverka,
                 1998; Ward and Wilkinson, 2006)

Physical         Degree to which the program physically engaged
engagement       audience members in a participatory experience;
Scale: 1 to 4    i.e., through touching or interacting with
Mean: 1.42       resource. (Beck and Cable, 2002; Knudson, et al,
S.D.: 0.39       2003; Lewis, 2005; Moscardo, 1999; NPS Module 101;
                 Sharpe, 1976; Tilden, 1957)

Verbal           Degree to which the program verbally engaged
engagement       audience members in a participatory experience;
Scale: 1 to 5    i.e, dialogue (a two-way discussion). (Knudson, et
Mean: 2.71       al, 2003; Moscardo, 1999; Sharpe, 1976; Tilden,
S.D.: 1.42       1957; Veverka, 1998)

Fact-based       Program communicated only fact-based information.
messaging        (Frauman and Norman, 2003; Jacobson, 1999; Lewis,
Binary: 27%      2005; Tilden, 1957; Ward and Wilkinson 2006)

Clear message    Degree to which program's message (s) was clearly
Scale: 1 to 4    communicated; i.e., the "so what?" element of the
Mean: 2.70       program. (Beck and Cable, 2002; Brochu and
S.D.: 0.94       Merriman, 2002; Ham, 1992; Jacobson, 1999)

Consistency      Degree to which the program's tone and quality were
Scale: 1 to 3    consistent throughout the program. (Beck and Cable,
Mean: 2.78       2002; Ham, 1992)
S.D.: 0.37

Table 3. Interpreter characteristics observed in the study, their
definitions, and descriptive statistics for cases analyzed in this
paper.

Interpreter      Definition
characteristic

Confidence       Equally weighted composite mean score of 3
(a = 0.70)       interpreter characteristics:
Scale: 1 to 4
Mean: 3.28       * Comfort of the Interpreter (Lewis 2005; Moscardo,
S.D.: 0.49       1999; Ward and Wilkinson, 2006): Degree to which
                 the interpreter presenting the program seems
                 comfortable with the audience and capable of
                 successfully presenting the program without
                 apparent signs of nervousness or self-doubt.

                 * Apparent knowledge (Ham and Weiler, 2002a; Lewis,
                 2005; Ward and Wilkinson, 2006): The degree to
                 which the interpreter appears to know the
                 information involved in the program, the answers to
                 visitors questions, and has local knowledge of the
                 area and its resources.

                 * Eloquence (Lewis, 2005): The extent to which the
                 interpreter spoke clearly and articulately, and did
                 not mumble or frequently use filler words such as
                 "um" or-"like"

Authentic        Equally weighted composite mean score of 3
emotion and      interpreter characteristics:
charisma
(a = 0.85)       * Passion (Beck and Cable, 2002; Ham and Weiler,
Scale: 1 to 5    2002b; Moscardo, 1999): The interpreter's apparent
Mean: 3.57       level of enthusiasm for the material, as opposed to
S.D.: 0.85       a bored or apathetic attitude toward it. The
                 overall vigor with which the material is presented.

                 * Charisma (Ward and Wilkinson, 2006): A general
                 sense of the overall likeability/charisma of the
                 interpreter, commonly recognized by seemingly
                 genuine interaction with the visitors, including
                 smiling, looking people in the eye, and having an
                 overall appealing presence.

                 * Sincerity (Ham, 2009): The degree to which the
                 interpreter seems genuinely invested in the
                 messages he or she is communicating, as opposed to
                 reciting information, and seems sincere in the
                 emotional connection they may exude to the message
                 and/or the resource. In other words, the extent to
                 which the interpretation was delivered through
                 authentic emotive communication.

Responsiveness   The extent to which the interpreter interacts with
Scale: 1 to 3    the; audience, collects information about their
Mean: 2.81       interests and backgrounds, and responds to their
S.D.: 0.41       specific questions and requests or non-verbal cues.
                 (Jacobson, 1999; Knudson et al., 2003; Lewis,
                 2005))

Humor quality    How funny is the interpreter overall? Does the
Scale: 1 to 4    audience react positively to the interpreter's use
Mean 2.08        of humor and seem to enjoy it? (Ham and Weiler,
S.D 0.73         2002b; Knapp and Yang, 2002; Regnier et al., 1992)

Sarcasm          The degree to which the interpreter used sarcasm
Scale: 1 to 3    (the use of mocking, contemptuous, or ironic
Mean: 1.23       language or tone) or self-deprecation that was not
S.D.: 0.46       meant to be serious, as a part of presenting their
                 program.

Audibility       The extent to which the interpreter could be
Scale: 1 to 3    clearly heard and understood by the audience.
Mean: 2.86
S.D.: 0.36

Impatience       Exhibition of explicit impatience toward audience
Binary: 1.8%     members.

Goal: Behavior   Intention of the interpreter for the program to
Change           influence audience's behavior. (Ham, 2013)
Binary: 7%

Table 4: Variables included in hypothesized models for each outcome.

                                             Visitor
                                            Experience    Behavioral
Variable                    Satisfaction       and        Intentions
                                           Appreciation

Interpreter
characteristics

Audibility                       X              X
Authentic emotion and            X              X             X
  charisma
Confidence                       X              X             X
False assumption about           X              X
  audience
Goal: Behavior change                                         X
Humor quality                    X              X             X
Impatience                       X
Responsiveness                   X              X
Sarcasm                                                       X

Program characteristics

Appropriate for audience         X              X             X
Appropriate logistics            X              X             X
Clear message                    X              X             X
Consistency                      X              X
Connection                       X              X
Multisensory engagement          X
Organization                     X              X
Verbal engagement                X              X             X
Fact-based messaging             X
联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有