Scientific Process and Research Design 1 Topics of this Session 1. 2. 3. 4. The scientific process Types of empirical studies Purposes of research design Research ethics 2 Four sources of knowledge or truth Mythical Authoritative Logical Scientific 3 Goal of Science: Seek truth or create knowledge Science is the pursuit of truth, of explanation, prediction, and control a phenomenon. Truth obtained from the scientific method contains both logic and evidence that are consistent with each other. Science is about the creation of knowledge, not the application of knowledge Science cannot settle debate about values (good or bad, right or wrong, e.g. stem cell research) objectivity Ultimate goal of science is to better the human condition 4 Methodology The conduct of scientific inquiry Methodology is the science of finding out knowledge (truth). The standards for determining truth is rigor in logic and methods The scientific community determines the standards of rigor and professional competence. Each scientist is accountable to the scientific community for adhering to standards of professional competence and norms. The visible part of our scientific conduct and results is “reconstructed logic” The ethics of science is in the conduct of the scientist. 5 Reality (Epistemology) Three views in social sciences Premodern: Only one reality, no individual experience matters or allowed. Modern: There is an objective reality (e.g., the temperature) but we have different subjective experience of it (e.g., warm, cold). Agreement in subjective experiences could indicate objective reality. Postmodern: No objective reality, only images of reality or subjective experiences, and all are true. Agreement is neither necessary or desirable to define reality. 6 Normal Science based on the modern view of reality Normal science is empirical science – data, evidences or observations are necessary to support theory. Determination of significant fact, matching of facts with theory, and articulation of theory – basic paradigm of normal science that we accept. Research based on shared paradigm is committed to the same rules and standards for scientific practice. That commitment and the apparent consensus it produces are both the genesis and the continuation of a research tradition. Paradigms involve different assumptions of truth or reality – normal science is based on a modern view of reality. 7 An Example Inter-firm level: firm to firm, firm to environment Firm level: Strategy, structure, culture, process Group level: process, structure, dynamics Individual level: attitudes, behavior, decision, perception Cross-level: firm firm individual group individual individual group firm 8 Commonly Used Research Methods in Management Survey Research Laboratory experiments Quasi experiments Secondary data design Qualitative and case methods 9 Scientific Research Scientific research is a systematic, controlled, empirical, and critical investigation of natural or social phenomena (using commonly accepted methodology) that either (a) is guided by theory and hypotheses about the presumed (true) relations among such phenomena, or b) results in theory and propositions about the possible (true) relationships among such phenomena. 10 Two Types of Empirical Research Theory testing empirical research Surveys (interview, mail, internet, phone) Experiments (university setting) Quasi experiments (university or company) Secondary data (financial, operational, personnel) Theory building empirical research Qualitative research (interview, observation, text) Case methods (interviews, text, secondary data) 11 A Simplified Scientific Research Process and Types of Research 1. Research Question Deductive research 2. Literature review 4. The empirical study Inductive research Theoretical research 3. Theory and Hypotheses Descriptive research 12 Elements of the Scientific Process Theories Concept formation Propositions Induction Empirical Generalization Measurement, sample summarization, parameter estimation Logical Deduction Logical Inference Accept or reject Hs Tests of hypotheses Observations Hypotheses Research Design, instrumentation, scaling, sampling From W.L. Wallace, The Logic of Science in Sociology, Figure 1 13 The Scientific Process - Review Curiosity or puzzle Research Question Literature review Yes Theory? Deductive Design Observation No Inductive Design 14 The Truth @@@###@@# Research is “messy”. The “reconstructed story” does not reveal the “messiness” in the “construction process”. Research is “perseverance” through iterations of induction and deduction Thomas Edison – 2000 times before he succeeded – 2000 step journey 15 Research design: Purpose 1 - variance control Systematic variance Extraneous variance Error variance A good research design should: maximize systematic variance, control extraneous variance, and minimize error variance 16 Approaches to variance control Maximize systematic variance Through good experimental control Through a strong theory Through systematic sampling Control extraneous variance Through randomization or matching of subjects Through including meaningful control variables Minimize error variance Through controlled conditions Through valid measurement 17 Variance Control in EOR study (Tsui, et al 1997, AMJ) Systematic: Independent variable (EOR) Sample selection – multiple industries (5), multiple firms (10) – 85 jobs Measurement of I.V. – 4 types at the job level QSC – 31%, UI – 18%, OI – 18%, MI – 33% Extraneous: Control variables individual, job, company – 8 total Error: Precision in measurement (EFA and CFA) 18 Variance Control in IJV control study (Yan & Gray, 1994, AMJ) Systematic: Independent variable (Management Control) Sample selection – theoretical sampling using criteria of IJV age, size, industries, ownership (4). Measurement or data – interviews and archives Extraneous: Through sampling – match of age, foreign country, etc. Error: Triangulation of data from multiple sources 19 Variance control in social capital study (Xiao & Tsui, 2007, ASQ) Systematic variance in the IV – industry experts nominate firms that vary on commitment Extraneous variance – through 9 control variables Error variance – through valid measurement 20 Research Design: Purpose 2 - Enhance Validity Validity is the probable truth or falsity of an assertion or inference. Four types of validity: Construct validity Present when measure produces results consistent with alternative (valid) measures of same construct Internal validity Absence of alternative explanations External validity Present when control variables do not interact with causal variables, i.e., results would hold at other times, in other settings, and for other individuals Statistical conclusion validity Significant presence of co-variation between variables 21 Construct Validity - Does the variable measure what it purports to measure? Content validity When the measure is absent of deficiency or contamination Criterion-related validity When the measure relates to other constructs as expected Convergent validity When the measure relates to other measures of the same construct Discriminant validity When the measure does not relate to other constructs as expected. 22 Construct validity in EOR study (Tsui et al., 1997, AMJ) 1 independent and 7 dependent variables Assess internal consistency of multiple items for the same construct (alpha = .76 to .96) Factor analysis to ensure that items load on the intended factor (CFA discriminant validity) Agreement between multiple assessment of EOR dimensions in the same job (ANOVA F, R-2 = .33 and .50 for the two dimensions) 23 Construct validity in IJV study (Yan & Gray, 1994, AMJ) Triangulation of data from multiple sources Replication of discovered constructs across cases Case examples to illustrate constructs 24 Construct validity in the social capital study (Xiao & Tsui, 1997, ASQ) 3 main constructs: structural holes, high commitment organization, career success 2 of the 3 are based on current measures High commitment – firm level construct measured at the individual level (n=88, 117, 102, 128 employees in four firms) EFA, Rwg=.83 to .84 (4 firms), ICC1=.14, ICC2=.95 25 Internal Validity: Confidence in inference, absence of alternative explanations Strong statistical validity Strong theory (internal validity) Strong controls (extraneous variance) Valid measurement (construct validity) Appropriate sample (external validity) Strong inference of causation (internal validity) No artificial covariance due to design or attribution 26 Internal validity of EOR study (Tsui et al., 1997, AMJ) Strong theoretical foundation No serious common method variance problem Inclusion of 8 control variables Construct validity of EOR measure – only 7 items (3 for contribution and 4 for inducement) - content deficiency? 27 Internal validity in IJV study (Yan & Gray, 1994, AMJ) Replication of relationships in different cases Consistency of patterns across cases Coders should be unaware of the initial model or hypotheses Competing explanations (e.g., IJV changes) are incorporated into the revised model 28 Internal validity in the social capital study (Xiao & Tsui, 1997, ASQ) Controlling for reverse causality Career data (DV) collected 6 months after IV Networks ties within performance period excluded Using 9 control variables that may also influence the dependent variable 3 ways to test the moderating hypothesis Subgroup analysis Interaction of firm level commitment score Interaction of individual level commitment score 29 Contributing to International Relations Knowledge (how to publish in international journals) Address an interesting and important “puzzle” Contribute to a “current conversation” Connect to current theories and constructs Use current normal science method Select a type of international relations research Engage in appropriate “contextualization” 30 The “Puzzle” Asking interesting questions “Uninteresting” questions in any study: Obvious questions Irrelevant questions Absurd questions Definitional questions Affirmation questions e.g. “Uninteresting” Chinese studies: Replicating a published study Uncritical application of an existing theory Affirmation of similarities to Western samples Outside-in and literature-driven approach in selecting problems or research questions (Davis, 1971) 31 The “Conversation” (Huff, 1999) What are the old and new topics in the conversation? What are the current debates? How familiar are current conversationalists about your topic and context (sample)? How to make the connection to the unfamiliar? Why should “they” listen to you? 32 Connect to Current Theories Apply current theories Extend or modify Create and connect Making the novel appear familiar Making the familiar appear novel Borrow and return something better Vision limited by borrowed lens Whetten, 2002; 2008 33 Use of Current Methods Meet criteria of validity – construct, internal, external Contextualization of measurement and data collection procedure Develop new methods of data collection, measurement, or analyses to supplement current methods 34 Contextualization (Tsui, 2006) The Phenomena – from “outside in” to “inside out” The Theory – from “application” to “creation” The Measurement – from “translation” to “indigenization” The Methodology – from “sharpening old tools” to “developing new instruments” 35
© Copyright 2026 Paperzz