簡易檢索 / 詳目顯示

研究生: 鄭仲棠
Cheng, Chung-Tang
論文名稱: The Methodology of Randomized Experiments in Economics
經濟學中隨機控制實驗的方法論分析
指導教授: 趙相科
Chao, Hsiang-Ke
口試委員: 吳世英
Wu, Shih-Ying
毛維凌
Mao, Wei-Lin
學位類別: 碩士
Master
系所名稱: 科技管理學院 - 經濟學系
Department of Economics
論文出版年: 2013
畢業學年度: 101
語文別: 英文
論文頁數: 83
中文關鍵詞: 隨機控制實驗因果關係田野實驗內部有效性外部有效性小班教學實驗
外文關鍵詞: randomized controlled trial, RCT, causal relation, field experiment, internal validity, external validity, Project STAR
相關次數: 點閱:3下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 摘要

    二十世紀六零年代,隨機控制實驗 (Randomized Controlled Trial,簡稱RCT) 於醫藥實驗取得了成功,成為了科學方法上的黃金標準 (gold standard)。自八零年代以來,RCT亦逐步地被應用在經濟學的田野實驗 (field experiment) 中,並一定程度地影響了實證經濟學的研究走向。奠基於醫藥實驗的成功上,一些實證經濟學家認為:RCT作為一種黃金標準,也能夠對社會政策的制定提供客觀證據。本論文從思想史及科學方法論的角度,試圖探討:(一)為何RCT在實證經濟學家之間逐漸蔚為主流?(二)RCT真的能夠確證因果關係嗎?其中的理論限制為何?(三)有沒有任何實作 (practice) 能夠映證第二個問題的解答?

    茲陳述本論文的研究結果如下。首先,在RCT的架構下實證經濟學家解決了傳統計量經濟學「被動觀察」 (passive observation) 的問題,並且為政策工具找到了客觀依據。其次,在「理想RCT→真實RCT→真實世界」的方法論圖像中,我們認為,實證經濟學家或其他科學實作者於田野RCT將面臨理想化缺口(idealization gap) 和一般化缺口 (generalization gap) 此兩類方法上的限制,各包含了實作時可能造成之偏誤 (bias) ,進而分別威脅其內部有效性 (internal va-lidity) 和外部有效性 (external validity)。最後在案例研究中,我們了回顧田納西州的小班實驗與加州的小班政策。結果顯示,此案例存在偌大的一般化缺口,至於理想化缺口是否出現則仍有待驗證。


    Abstract

    In the 1960s, Randomized Controlled Trial (RCT) was established as a ‘gold standard’ in clinical trial. Since the 1980s, RCT-method has gradually adopted in field experi-ments in economics and greatly influenced empirical economics. Based on the success of clinical trial, some empirical economists feel that RCT is also a ‘gold standard’ that provides objective evidence in policy-making procedures. This thesis, will consider (a) why the RCT-method becomes popular among empirical economists? (2) Can RCTs in economic research truly confirm causal hypotheses? Are there any theoretical limi-tations in the RCT methodology? (3) Does any real practice correspond to these limi-tations?
    Three brief conclusions are, first, RCT solves the problem of passive observation, which bothered traditional econometrics, and RCT as a policy tool is proved to be ‘objective’. Second, in the methodological figure we proposed, practicing economists and scientists may confront ‘idealization gaps’ and ‘generalization gaps’ that threaten the internal validity and external validity of field RCTs, respectively. Lastly, we take two class-size deduction cases: the Project STAR in Tennessee and the CSR Program in California as examples. The results indicate that the ‘generalization gaps’ are huge, and the ‘idealization gaps’ in these are still to be verified.

    摘要 iii Abstract iv Acknowledgements vii List of Abbreviations viii Chapter 1. Introduction 1 Chapter 2. Randomized Controlled Trial: A Brief History in Science 4 Chapter 3. A Historical Review of RCTs in Economics 19 Chapter 4. The Methodology of RCTs in Economics 38 Chapter 5. Case Study: Project STAR and Its Evaluations 59 Chapter 6. Conclusion 74 References 77

    References

    Abadie, A. 2005. “Semiparametric Difference-in-Differences Estimators.” Review of Economic Studies, 72(1): 1-19.

    Angrist, J.D. 1990. “Lifetime Earnings and the Vietnam Era Draft Lottery: Evidence from Social Security Administrative Records.” The American Economic Review, 80(3): 313-336.

    Angrist, J.D. and Lavy, V. 1999. “Using Maimonides’ Rule to Estimate the Effect of Class Size on Scholastic Achievement.” Quarterly Journal of Economics, 106(4):979-1014.

    Ackermann, R. 1989. “The New Experimentalism.” British Journal for the Philoso-phy of Science 40: 185-90.

    Averett, S.L. and McLennan, M.C. 2004. “Exploring the Effect of Class Size on Student Achievement: What Have We Learned Over the Past Two Decades?” in International Handbook on the Economics of Education, Geraint Johnes and Jill Johnes, eds. Edward Elgar Publishers.

    Angrist, J.D. and Krueger, A.B. 2001. “Instrumental Variables and the Search for Identification: From Supply and Demand to Natural Experiments.” The Journal of Economic Perspectives, 15(4): 69-85.

    Angrist, J.D., Imbens, G.W. and Rubin, D.B. 1996. “Identification of Causal Ef-fects Using Instrumental Variables.” Journal of the American Statistical Associa-tion, 90(430): 431-442.

    Angrist, J.D. and Pischke, J. 2009. Mostly Harmless Econometrics: An Empiricist’s Companion. Princeton and Oxford: Princeton University Press.

    Angrist, J.D. and Pischke, J. 2010. “The Credibility Revolution in Empirical Eco-nomics: How Better Research Design Is Taking the Con out of Economet-rics.” Journal of Economic Perspectives, 24(2): 3-30.

    Ashenfelter, O. 1987. “The Case for Evaluating Training Programs with Randomized Trials.” Economics of Education Review, 6(4): 333-338.

    Ashenfelter, O. and Card, D. 1985. “Using the Longitudinal Structure of Earnings to Estimate the Effect of Training Programs.” The Review of Economics and Statis-tics, 67(4): 648-660.

    Bertrand, M. and Mullainathan S. 2004. “Are Emily and Greg More Employable than Lakisha and Jamal? A Field Experiment on Labor Market Discrimination.” The American Economic Review, 94(4): 991-1013.

    Blair, G., Iyengar, R.K. and Shapiro, J.N. 2012. “Sample Bias in the Experimental Movement in Economics and Political Science.” Working Paper.

    Box, J.F. 1978. R. A. Fisher: The Life of a Scientist, New York: Wiley.

    Boumans, M. 2010. “The Problem of Passive Observation.” History of Political Economy, 42(1):75-110.

    Cameron, A.C. and Trevedi, P.K. 2005. Microeconometrics. Cambridge: Cambridge University Press.

    Campbell, D.T. and Stanley, J.C. 1963. Experimental and Quasi-experimental de-signs for Research. Chicago: RandMcNally.

    Card, D. 1990 “The Impact of the Mariel Boatlift on the Miami Labor Market.” In-dustrial and Labor Relations Review, 43(2): 245-257.

    Carpenter, D. 2010. Reputation and Power: Organizational Image and Pharmaceu-tical Regulation at the FDA. Princeton: Princeton University Press.

    Cartwright, N. 1989. Nature’s Capacities and Their Measurement. UK: Oxford Uni-versity Press.

    Cartwright, N. 1995. “Causal Structures in Econometric Models”, in Little, Daniel (ed.) The Reliability of Economic Models. Dordrecht: Kluwer Academic Pub-lishers.

    Cartwright, N. 2007a. “Are RCTs the Golden Standard?” BioSocieties, 2: 11-20.

    Cartwright, N. 2007b. Hunting Causes and Using Them: Approaches in Philosophy and Economics. UK: University of Cambridge Press.

    Cartwright, N. 2009. “What is this thing called efficacy?” in Mantzavinos, C., (ed.) Philosophy of the Social Sciences: Philosophical theory and scientific practice. Cambridge University Press, Cambridge, UK.

    Cartwright, N. 2009. “Evidence-Based Policy: What’s to Be Done about Relevance? For the 2008 Oberlin Philosophy Colloquium.” Philosophical Studies: An Inter-national Journal for Philosophy in the Analytic Tradition, 143(1): 127-136.

    Cartwright, N. 2013 “Will This Policy Work for You? Predicting Effectiveness Better: How Philosophy Helps.” Philosophy of Science, 79: 973-989.

    Case, A. and Deaton, A. 1999. “School Inputs And Educational Outcomes In South Africa.” The Quarterly Journal of Economics, 114(3): 1047-1084.

    Cox, D.R. 1958. Planning of Experiments. New York: Wiley.
    Deaton, A.S. 2009. “Instruments of Development: Randomization in the Tropics, and the Search for the Elusive Keys to Economic Development.” NBER Working Paper 14690.

    Deaton, A. 2010. “Instruments, Randomization, and Learning about Develop-ment.” Journal of Economic Literature, 48(2): 424-55.

    Doll, R. and Hill, A.B. 1950. “Smoking and the Carcinoma of the Lung: Preliminary Report.” British Medical Journal, 30(3): 739-948.

    Doolittle, F.C. and Traeger, L. 1990. Implementation of the National JTPA Study. New York: Manpower Demonstration Research Corporation.

    Duflo, E. 2006. “Field Experiments in Development Economics.” Unpublished paper.

    Duflo, E. and Saez, E. 2003. “The Role of information and Social Interactions in Re-tirement plan Decisions: Evidence from a Randomized Experiment.” Quarterly Journal of Economics, 118(3): 815-842.

    Duflo, E. and Kremer, M. 2004. “Use of Randomization in the Evaluation of De-velopment Effectiveness,” in Evaluating Development Effectiveness (World Bank Series on Evaluation and Development, Volume 7, Osvaldo Feinstein, Gregory K. Ingram and George K. Pitman, editors, Transaction Publishers: New Brunswick, NJ, 2004, pp. 205-232.

    Duflo, E., Glennerster, R. and Kremer, M. 2008. “Using Randomization in Devel-opment Economics Research: A Toolkit,” in: Schultz, T. and Strauss, J., (eds.), Handbook of Development Economics, vol. 4: 3896-3962. Amsterdam: North-Holland.

    Finn, J.D. and Achilles, C.M. 1990. “Answers and Questions About Class Size: A Statewide Experiment.” American Educational Research Journal, 27(3): 557-577.

    Finn, J.D., Gerber, S.B. and Boyd-Zaharias, J. 2005 “Small Classes in the Early Grades, Academic Achievement, and Graduating From High School.” Journal of Educational Psychology, 97(2): 214-223.

    Fisher, R.A. and Mackenzie, W.A. 1923. “Studies in Crop Variation II: The Manuri-al Response of Different Potato Varieties.” The Journal of Agricultural Science, 13(3): 311-320.

    Fisher, R.A. 1935(1966). The Design of Experiment. Eighth Edition. Edinburgh: Oli-ver and Boyd.

    Fisher, R.A. 1959. Smoking : The Cancer Controversy: Some Attempts to Assess the Evidence. Edinburgh: Oliver and Boyd.

    Francis, Jr., T., Napier, J.A., Voigt, R.B. et al. 1957. Evaluation of the 1954 field trial of poliomyelitis vaccine: Final report. Ann Arbor: University of Michigan.

    Guala, F. 2005. The Methodology of Experimental Economics. Cambridge; New York: Cambridge University Press.

    Haavelmo, T. 1944. “The Probability Approach in Econometrics.” Econometrica, 12: iii-115.

    Hacking, I. 1983. Representing and Intervening: Introductory Topics in the Philosophy of Natural Science. Cambridge: Cambridge University Press.

    Hacking, I. 1988. “Telepathy: Origins of Randomization in Experimental Design.” Isis, 79(3): 427-451.

    Hanushek, E.A. 1999. “The Evidence on Class Size.” in Susan E. Mayer and Paul E. Peterson (eds.), Earning and Learning: How Schools Matter, Chapter 7, Wash-ington, DC: Brookings Institution Press and New York, NY: Russell Sage Foun-dation, 131-68.

    Harré, R. 2009. Pavlov's Dogs and Schrödinger's Cat: Scenes from the Living La-boratory. Oxford: Oxford University Press.

    Hausman, J.A. and Wise, D.A. 1985. Social Experimentation. Chicago: University of Chicago Press.

    Heckman, J.J. 1979. “Sample Selection as a Specification Error.” Econometrica, 47(1): 153-161.

    Heckman, J.J. 1992. “Randomization and Social Policy Evaluation.” In Evaluating Welfare and Training Programs, ed. Charles F. Manski and Irwin Garfinkel, 201-30. Cambridge and London: Harvard University Press.

    Heckman, J.J. and Smith, J.A. 1995. “Assessing the Case for Social Experiments.” The Journal of Economic Perspectives, 9(2): 85-110.

    Hendry, D.F. and Morgan M.S. 1995. The Foundations of Econometric Analysis. Cambridge: Cambridge University Press.

    Hill, A.B. and Medical Research Council. 1948. “Streptomycin treatment of pul-monary tuberculosis.” British Medical Journal, 2:769-783.

    Hill, A.B. 1952. “The Clinical Trial.” New England Journal of Medicine, 247: 113-119.

    Holland, P. 1986. “Statistics and Causal Inference.” Journal of the American Statis-tical Associations, 81(396):947.

    Hoover, K.D. 1990. “The Logic of Causal Inference.” Economics and Philosophy, 6: 207-234.

    Hoover, K.D. 1994. “Six Queries About Idealization in an Empirical Context.” Ide-alization VI: Idealization in Economics (ed. by Hamminga, B. and De Marchi, D.B.), Amsterdam: Rodopi.

    Hoover, K.D. 2005. “The Methodology of Econometrics.” Prepared for the Palgrave Handbooks of Econometrics, volume 1: Theoretical Econometrics.

    Hoover, K.D. 2006. “Causality in Economics and Econometrics.” An Entry for the New Palgrave Dictionary of Economics.

    Hoxby, C. 2000. “The Effects of Class Size on Student Achievement: New Evidence from Population Variation.” The Quarterly Journal of Economics, 115(4): 1239-1285.

    Imbens, G.W. and Angrist, J.D. 1994. “Identification and Estimation of Local Av-erage Treatment Effects.”
    Econometrica, 62(2):467-475.

    Imbens, G. and Rubin, D.B. 2008. “Rubin Causal Model.” in: The New Palgrave Dictionary of Economics. Second Edition. Eds. Steven N. Durlauf and Lawrence E. Blume. Palgrave Macmillan, 2008.

    Imbens, G.W. and Wooldridge, J.M. 2009. “Recent Development in the Economet-ric of Program Evaluation.” Journal of Economic Literature, 47(1): 5-86.

    Imbens, G.W. 2010. “Better LATE than Nothing: Some Comments on Deaton (2009) and Heckman and Urzua (2009).” Journal of Economic Literature, 48(2): 399-423.

    Keynes, J.N. 1890(1917). The Scope and Method of Political Economy. Fourth Edi-tion. London: Macmillan & Company.

    Kremer, M. 2003. “Randomized Evaluations of Educational Programs in Developing Countries: Some Lessons.” The American Economic Review, 93(2): 102-106.

    Krueger, A.B. 1999. “Experimental Estimates of Education Production Functions.” Quarterly Journal of Economics, 114(2): 497-532.

    Lakatos, I. 1974. “The Role of Crucial Experiments in Science." Studies in History and Philosophy of Science Part A, 4(4): 309-325.

    LaLonde, R.J. 1986. “Evaluating the Econometric Evaluations of Training Programs with Experimental Data.” The American Economic Review, 76(4):604-620.

    The Lancet. 2004. “The World Bank is Finally Embracing Science.” The Lancet, 364(9436): 731- 732.

    Leamer, E.E. 2010. “Tantalus on the Road to Asymptopia.” Journal of Economic Perspectives, 24(2): 31-46.

    Lee, L. -F. 2001. “Self-Selection,” in A Companion to Theoretical Econometrics, B. Baltagi (ed.), 383-409, Oxford: Blackwell.

    List, J.A. and Lucking-Reiley, D. 2002. “The Effects of Seed Money and Refunds on Charitable Giving: Experimental Evidence from a University Capital Cam-paign.” Journal of Political Economy, 110(1): 215-233.

    Lucas, R. 1976. “Econometric Policy Evaluation: A Critique.” in Brunner, K.; Meltzer, A., The Phillips Curve and Labor Markets, Carnegie-Rochester Conference Series on Public Policy, 1:19-46. New York: American Elsevier.

    Manning, W.G., Newhouse, J.P., Duan, N., Keeler, E.B. and
    Leibowitz, A. 1987. “Health Insurance and the Demand for Medical Care: Evidence from a Random-ized Experiment.” The American Economic Review, 77(3): 251-277.

    Manski, C. and Garfunkel, I. 1992. Evaluating Welfare and Training Programs. Harvard: Harvard University Press.

    Marks, H.M. forthcoming. “The 1954 Salk Poliomyelitis Vaccine Field Trial.” 100 Landmark Clinical Trials (ed. by Steven N. Goodman, Harry M. Marks, Karen Robinson), New York: John Wiley and Sons.

    McCall, W.A. 1923. How to Experiment in Education. New York: The Macmillan Company.

    Meyer, B.D. 1995. “Natural and Quasi-Experiments in Economics.” Journal of Busi-ness & Economic Statistics, 13(2): 151-161.

    Miguel, E. and Kremer, M. 2004. “Worms: Identifying Impacts on Education and Health in the Presence of Treatment Externalities.” Econometrica, 72(1): 159-217.

    Mosteller, F. 1995. “The Tennessee Study of Class Size in the Early School Grades.” Critical Issues for Children and Youths, 5(2): 113-127.

    Nigenda, G. 2005. “Lessons Offered by Latin American Cash Transfer Programmes, Mexico’s Oportunidades and Nicaragua’s SPN. Implications for African Coun-tries.” Center of Social and Economic Analysis, Mexican Health Foundation

    Nye, B. and Hedges, L.V. 2001. “The Long-Term Effects of Small Classes in Early Grades: Lasting Benefits in Mathematics Achievement at Grade 9.” The Journal of Experimental Education, 69(3): 245-257.

    Nye, B., Hedges, L.V. and Konstantopoulos, S. 1999. “The Long-Term Effects of Small Classes: A Five-Year Follow-Up of the Tennessee Class Size Experiment.” Educational Evaluation and Policy Analysis, 21(2): 127-142.

    Parker, S.W., Rubalcava, L. and Teruel, G. 2008. “Evaluating Conditional School-ing and Health Programs,” in: Schultz, T. and Strauss, J., (eds.), Handbook of Development Economics, vol.4: 3896-3962. Amsterdam: North-Holland.

    Peirce, C.S. and Jastrow, J. 1885. “On Small Differences of Sensation.” Memoirs of the National Academy of Sciences, 3: 75-83.

    Rockoff, J. 2009. “Field Experiments in Class Size from the Early Twentieth Century.” Journal of Economic Perspectives, 23(4): 211-230.

    Rosenberger, W.F. and Lachin, J.M. 2002. Randomization in Clinical Trials: Theory and Practice. New York: Wiley.

    Rosenberg, A. 2000. Philosophy of Science: A Contemporary Introduction. First Edi-tion. London and New York: Routledge.

    Rosenzweig, M.R. and Wolpin, K.I. 2000. “Natural ‘Natural Experiments’ in Eco-nomics.” Journal of Economic Literature, 38(4): 827-874.

    Rubin, D.B. 1974. “Estimating causal effects of treatments in randomized and non-randomized studies.” Journal of Educational Psychology, 66(5): 688-701.

    Rubin, D.B. 1978. “Bayesian Inference for Causal Effects: The Role of Randomiza-tion.” Annals of Statistics, 6: 34-58.

    Salsburg, D. 2001. The Lady Tasting Tea: How Statistics Revolutionized Science in the Twentieth Century. New York: W.H. Freeman and Company.

    Schulz, K.F. and Grimes, D.A. 2006. The Lancet Handbook of Essential Concepts in Clinical Research. Edinburgh: Elsevier.

    Shadish, W.R., Cook, T.D. and Campbell, D.T. 2002. Experimental and Qua-si-Experimental Designs for Generalized Causal Inference. Boston: Houghton Mifflin.

    Simon, J.L. 1966. “The Price Elasticity of Liquor in the U.S. and a Simple Method of Determination.” Econometrica, 34(1): 193-205.

    Stafford, F. 1985. “Income-Maintenance Policy and Work Effort: Learning from Ex-periments and Labor-Market Studies,” in: Hausman, J. and Wise, D. (eds.), Social Experimentation. Chicago: University of Chicago Press.

    Stecher, B.M., Bohrnstedt, G.W., Kirst, M., McRobbie, J. and Williams, T. 2001. “Class-Size Reduction In California.” Phi Delta Kappan, 82(9): 670.

    Stecher, B.M. and Bohrnstedt, G.W. (ed.) 2002. Class size reduction in California: Findings from 1999–00 and 2000–01.
    Sacramento, CA: California Department of Education.

    Steel, D.P. 2008. Across the Boundaries: Extrapolation in Biology and Social Science. Oxford and New York: Oxford University Press.

    Stigler, S.M. 1992. “A Historical View of Statistical Concepts in Psychology and Educational Research.” American Journal of Education, 101(1): 60-70.

    Stock, J.H. and Trebbi, F. 2003. ”Who Invented Instrumental Variables Regression?” The Journal of Economic Perspectives, 17(3):177-194.

    Stock, J.H. and Watson, M.W. 2003(2011). Introduction to Econometrics. Third Edition. Addison Wesley Longman.

    Tiera, D. and Reiss, J. 2013. “Causality, Impartiality and Evidence-Based Policy.” Towards the Methodological Turn in the Philosophy of Science: Mechanism and Causality in Biology and Economics (ed. by Hsiang-Ke Chao, Szu-Ting Chen and Roberta Millstein), New York: Springer.

    White, H. 1980. “A Heteroskedasticity-Consistent Covariance Matrix Estimator and a Direct Test for Heteroskedasticity.” Econometrica, 48(4): 817-838.

    無法下載圖示 全文公開日期 本全文未授權公開 (校內網路)
    全文公開日期 本全文未授權公開 (校外網路)

    QR CODE