382C Empirical Studies in Software Engineering. Lecture 1. Introduction. Dewayne E Perry ENS present, Dewayne E Perry

Similar documents
CSC2130: Empirical Research Methods for Software Engineering

Underlying Theory & Basic Issues

Who? What? What do you want to know? What scope of the product will you evaluate?

Experimental Design. Dewayne E Perry ENS C Empirical Studies in Software Engineering Lecture 8

MODULE 3 APPRAISING EVIDENCE. Evidence-Informed Policy Making Training

Hypothesis-Driven Research

In this chapter we discuss validity issues for quantitative research and for qualitative research.

What Case Study means? Case Studies. Case Study in SE. Key Characteristics. Flexibility of Case Studies. Sources of Evidence

What Is Science? Lesson Overview. Lesson Overview. 1.1 What Is Science?

Key Ideas. Explain how science is different from other forms of human endeavor. Identify the steps that make up scientific methods.

Unit 2, Lesson 5: Teacher s Edition 1. Unit 2: Lesson 5 Understanding Vaccine Safety

What is Psychology? chapter 1

High-level Vision. Bernd Neumann Slides for the course in WS 2004/05. Faculty of Informatics Hamburg University Germany

The Nature of Science

Political Science 15, Winter 2014 Final Review

Research Approach & Design. Awatif Alam MBBS, Msc (Toronto),ABCM Professor Community Medicine Vice Provost Girls Section

Course summary, final remarks

Core Competencies for Peer Workers in Behavioral Health Services

What Science Is and Is Not

Evaluation: Scientific Studies. Title Text

Science is a way of learning about the natural world by observing things, asking questions, proposing answers, and testing those answers.

RESEARCH METHODS. Winfred, research methods, ; rv ; rv

RESEARCH METHODS. Winfred, research methods,

What s Unique about Health Dialog? 2004 AAAI Fall Symposium: Dialog Systems for Health Communication

Applying the Experimental Paradigm to Software Engineering

Applying GRADE-CERQual to qualitative evidence synthesis findings paper 4: how to assess coherence

Scientific Thinking Handbook

The Conference That Counts! March, 2018

Research Seminar: Introduction to Empirical Science

Expert judgements in risk and reliability analysis

Selecting a research method

Group Assignment #1: Concept Explication. For each concept, ask and answer the questions before your literature search.

Math 140 Introductory Statistics

Experimentation Design in Software Engineering & Computer Science. Ways to Acquire Knowledge. Jeff Offutt.

Narrative Analysis Handout

Study of the Toyota Brand Impression Resulting From the Recall Crisis

THE CUSTOMER SERVICE ATTRIBUTE INDEX

Research Questions and Survey Development

Why is science knowledge important?

Behaviorism: An essential survival tool for practitioners in autism

Positive Psychologists on Positive Psychology: Alex Linley

The Current State of Our Education

PSYC1024 Clinical Perspectives on Anxiety, Mood and Stress

Empirical Approaches, Questions & Methods

04/12/2014. Research Methods in Psychology. Chapter 6: Independent Groups Designs. What is your ideas? Testing

Science and Engineering Practice (SEP) Rubric

Experimental Research in HCI. Alma Leora Culén University of Oslo, Department of Informatics, Design

Evidence Informed Practice Online Learning Module Glossary

SEMINAR ON SERVICE MARKETING

Case studies. Course "Empirical Evaluation in Informatics" Prof. Dr. Lutz Prechelt Freie Universität Berlin, Institut für Informatik

Doing High Quality Field Research. Kim Elsbach University of California, Davis

Critical Thinking Rubric. 1. The student will demonstrate the ability to interpret information. Apprentice Level 5 6. graphics, questions, etc.

VPS PRACTICUM STUDENT COMPETENCIES: SUPERVISOR EVALUATION VPS PRACTICUM STUDENT CLINICAL COMPETENCIES

School of Social Work

Master of Science Psychology (Conversion)

Integrating Theory, Research, and Practice in Vocational Psychology: Perspectives of a Journal of Vocational Behavior Editorial Board Member

Conduct Rigorous and Scientific Research. Chang Xu Department of Computer Science and Technology Nanjing University Sep 21, 2018

Lecturer: Dr. Emmanuel Adjei Department of Information Studies Contact Information:

Lab 2: The Scientific Method. Summary

Supplementary Material for Malle, Knobe, and Nelson (2007). Actor-observer asymmetries in explanations of behavior: New answers to an old question.

THE SCIENTIFIC WORLD VIEW

Stat 13, Intro. to Statistical Methods for the Life and Health Sciences.

Proposal Writing. Lynda Burton, ScD Johns Hopkins University

Pediatrics Milestones and Meaningful Assessment Translating the Pediatrics Milestones into Assessment Items for use in the Clinical Setting

Getting ready for propensity score methods: Designing non-experimental studies and selecting comparison groups

What Constitutes a Good Contribution to the Literature (Body of Knowledge)?

Measuring impact. William Parienté UC Louvain J PAL Europe. povertyactionlab.org

The Scientific Method

Experimentation Design in Software Engineering & Computer Science. Ways to Acquire Knowledge

Insight Assessment Measuring Thinking Worldwide

Science & Software Engineering

Seminar 3: Basics of Empirical Research

Assignment 4: True or Quasi-Experiment

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 1.1-1

SOCIAL WORK PROGRAM Field Education Coordinator s Evaluation of Practicum Agency

CHECKLIST FOR EVALUATING A RESEARCH REPORT Provided by Dr. Blevins

SOCIAL WORK PROGRAM Field Education Director s Evaluation of Practicum Agency

Work, Employment, and Industrial Relations Theory Spring 2008

Recovering Families: A Tool For Parents in Recovery August 1, 2016

Research Methods and Validadion

Conducting Research in the Social Sciences. Rick Balkin, Ph.D., LPC-S, NCC

10 30% This section must achieve a minimum mark of 5/10 independently to pass the assessment overall. A score of 4 or

2 Critical thinking guidelines

Lessons Learned about Mixed Methods Research Strategies in Systems Engineering: Evidence from PhD Dissertations

Overview of cognitive work in CBT

Lecture 9 Internal Validity

A Framework for Conceptualizing, Representing, and Analyzing Distributed Interaction. Dan Suthers

One slide on research question Literature review: structured; holes you will fill in Your research design

Chapter 02 Developing and Evaluating Theories of Behavior

Gifted/Talented Students from Poverty-Varied Faces

What is Science 2009 What is science?

Previous Example. New. Tradition

Critical Thinking and Reading Lecture 15

Ideas RESEARCH. Theory, Design Practice. Turning INTO. Barbara Fawcett. Rosalie Pockett

Current Reality Trees A Collaborative Tool for Continuous Improvement. Your Coach: Paul Hodgetts

Status of Empirical Research in Software Engineering

Causal inference: Nuts and bolts

Psychology (Conversion)

Definitions of Nature of Science and Scientific Inquiry that Guide Project ICAN: A Cheat Sheet

Name Class Date. 1. How does science differ from other kinds of human endeavors such as art, architecture, and philosophy?

Transcription:

Introduction Dewayne E Perry ENS 623 perry@mail.utexas.edu 1

Course Goals Create/evaluate empirical studies/data Understand underlying scientific and theoretical bases Understand fundamentals of experimental design Independent and dependent variables Variable manipulation/data gathering Various issues of validity Empirical logic and reasoning Analytic tools for reasoning about the experiment Understand and control confounding variables Random and systematic biases Alternative explanations Understand and apply appropriate data analysis techniques Understand and create the underlying logics of empirical studies in reasoning out conclusions 2

Experimenter/Evaluator Goals Responsible skepticism Look for Failures in experimental designs Failures of observations Gaps in reasoning Alternative explanations Compare new evidence against old Raise counter objections/hypotheses Question grounds for doubt as well Accumulate weight of evidence 3

Good Research Practices Enthusiasm an enjoyable endeavor Open-mindedness keen, attentive, inquisitive Common sense avoid looking under the lamppost Inventiveness creative Not only in experimental work In resource management as well Confidence in ones own judgment Despite detractors when right Know when you are wrong Consistency and care about detail No substitute for accuracy - keep complete records, organize and analyses accurately/carefully Ability to communicate Writing is a superb, essential research technique Make your discoveries known to others Honesty integrity and scholarship 4

Review of SWE Factors in software engineered products Theory (basis for product) CS Core Domain specific theory Experience (basis for judgment) Feedback Engineering experiments (prototypes) Empirical studies Observations Correlations Causal connections Process (basis for production) Methods and techniques Technology Organizational structures Teams Projects Cultures 5

Review of SWE SW development processes Phases Requirements Architecture Design Construction Deployment and maintenance Integral to all phases Documentation Measurement & Analysis Evolution Teamwork Management of artifacts 6

Daily Life of SWE: Decisions Determine what users want/need Make architecture, design and implementation decisions Evaluate/compare architecture and design choices Evaluate functional characteristics Evaluate non-functional properties Evaluate/compare technologies For supporting tools For product support For process support Determine what went wrong Determine good resource allocations etc 7

Some Illustrative Examples Most frequently stolen car Language/Solution comparisons Inspection reading techniques Exit Interviews 8

Most Frequently Stolen Car Honda Accord Study presented in media Honda most frequently stolen automobile What does this study tell us? Actually very little What can we infer from this? Shouldn t buy a Honda? Buy a Mercedez instead? Very misleading More recently Frequency relative to the total number of cars Claims per 1000 Different story Lincoln Navigator 12.2; Cadillac Escalade 10.3; etc All Escalades recovered GPS system Honda and Camry didn t make it anywhere near the top 10 Percentage of thefts claims is low compared to the number on road 9

Language/Solution Comparison From a summer project Comparing Java/C 2 different computers used Pentium III 600MHz 128M C 1.4 times faster than Java AMD 1GHz 256M Java 1.09 times faster than C Conflicting evidence How do we account for it? What are the differences? How do we resolve these differences? What can we conclude, if anything? 10

Preview of Confounding Variables 11

Inspection Reading Techniques Experiment: evaluate reading techniques for object oriented code inspections Inconclusive results Possibilities: There are no differences Poor experimental design Insufficient data Poor analyses and reasoning 12

Surveys Exit Polls Used to predict outcome 2004 election Significant disagreement with final vote Early calls retracted Pollster s response Results with in margin of error Data selectively reported Possible biases Questions asked slanted, non-uniform (ie general vs specific), etc Time of polls 7am - men on the way to work 10am soccer mom s after kids are in school Place of polls East Austin versus Westlake Nature of volunteers 13

Current State of SE Empirical Work Implementation oriented Fenton: poor statistical designs, don t scale Basili: differences in projects make comparisons difficult Johnson: practitioners resist measurement Need to be more requirements oriented Think hard about what experiments really are How they can be most effectively used Core problem: conceptualizing and organizing a body of work as a scientific basis 14

Software Development Little hard evidence to inform decisions Correlations suggestive but not sufficient in all cases Many times don t know exception cases Do not know fundamental mechanisms Software tools Methods and techniques Processes Empirical studies are the key Show mechanisms Eliminate alternative explanations Empirical validation is standard in some fields Quality of empirical studies in SE is rising Funding agencies recognizing value of empirical studies Increasing number of tutorials, panels, SOTAs, papers, etc Key consciousness raising papers (Tichy etal, Zelkowitz) Several key organizations: SEL, ISERN 15

Systematic Problems Research ideas are not empirically validated Should retroactively validate Should proactively directed Search for perfect study Instead of focus on credibility Study the obvious OK, but we need deeper insights Lots of data Not enough should answer important questions Lack useful hypotheses Lack conclusions from data 16

Empirical Software Engineering Studies Individual programmer studies have credibility due to well understood techniques from psychology and statistics. Large software development studies with the addition of large population social factors are not well established or credible. General goal: Establish a spectrum of empirical techniques that are robust to large variances from social factors present. 17

Challenges Create better empirical studies Establish principles that are Causal: correlated, temporally ordered, testable theory Actionable: causal agent effectively controllable General: widely applicable Answer important questions Family of focused studies illuminate related aspects Cost effective and reproducible Credible interpretations Degree of confidence we have in conclusions Eliminate alternative explanations Provide a compelling logic in the discussion Validity is critical: construct, internal, external Hypothesis is critical: ask important questions Resolutions appropriate to the intent of the study Make the data public 18

Some Concrete Steps Designing studies Ask significant questions Knight-Leveson, N-version programming Families of studies Schneiderman et al, on the value of flowcharts Build partnerships Takes time; multi-person effort; interdisciplinary, industry Long running in vivo/situ experiments Subparts; subject rights; know when to stop Collecting data Retrospective artifact analysis Eg, version management systems Simulation and modeling Eg, integration studies of Solheim and Rowland Involving others Meta-analysis eg Porter and Johnson Educational laboratories Teach empirical studies basics Populate lab with appropriate data/designs/equipment 19

Goals for SWE Empirical Studies Some help Look to other empirical disciplines Adapt what is useful here Goals Perform better empirical studies Focus on causal mechanisms Generate theories Iterate and improve Good empirical studies enable us to Encode knowledge more rapidly Prune low payoff ideas rapidly Recognize and value high payoff ideas Exploit important practical ideas 20

What is Critical Fundamentals Credible interpretations Repeatability Understanding validity limits Identifying underlying mechanism Practical significance Non-fundamentals Whether qualitative or quantitative Both have their place and usefulness Identical results Want congruent results Correlation studies Important precursor, but not the goal Opportunistic studies 21

Structure of an Empirical Study Research context Problem definition Research review Hypothesis Abstract about the world Concrete about the design Experimental design Variables independent and dependent Plan to systematically manipulate variables Control operational context Threats to validity: construct, internal, external Data analysis and presentation Quantitative: hypothesis testing, power analysis Qualitative Results and conclusions Limits, influences Explain how answered question and its practical significance Sufficient information for repeatability 22