As we reported a few days ago, one of our contributors, Greg Wilson, gave a keynote at the MSR Vision 2020 workshop in Kingston on August 20. In that, he explored why there's still a gulf between software engineering researchers and the people who actually build software for a living (see the slides or the discussion on Reddit for details). He also said that:
there's no easy way to close that gap, because most of the people in industry that researchers want to collaborate with have never encountered empirical software engineering studies, and therefore don't understand their scope or value; so
researchers—many of whom are professors—should pivot the software engineering classes they teach to focus on how to analyze real-world data, and what past analyses have told us, so that the next generation of developers will understand (and listen, and want to collaborate).
To make this more concrete, Greg asked the workshop participants to make up some assignments and exam questions for such a course. Some of the suggestions are listed below; we would welcome other ideas as well (please post them as comments). We'd also like to know who'd be interested in trying to teach such a course at their institution, and what you think the prerequisites would have to be: statistics, obviously, but would a database course that introduced students to SQL be necessary? What about a natural language processing course? Or something else we haven't thought of?
Give two examples of success stories in studies of the social aspects of software engineering.
Reorganization based on social structures
Identifying the "big players" in a software project
What are three sources of social interaction in software projects?
source code comments
Name three challenges in preprocessing emails.
fake/multiple email addresses
identifying email headers and inline replies
use of multiple languages
You are given a dataset A of OSS projects and a subset of it B. Evaluate whether a hypothesis H can be rejected on A and B. Design the question in such a way that H is significant (at 0.05 level) at A and not B. Discuss the discrepancy.
Given a dataset and a specific question, perhaps from exisitng MSR papers, discuss which data mining approach is best suited for that question.
Given a specific question (e.g., bug finding) what repositories should you use to solve it? Illustrate it with Bugzilla. How do you adapt this to Jira?
Given that two variables A and B correlate, can you say "A causes B"? Why or why not?
Repeat an existing analysis from an MSR paper. Do you get the same results? Vary a number of variables. How different are the results?
What is wrong with this claim: "Files with a large number of committers/authors have more defects/bugs, so we conclude that more authors cause more bugs, and we recommended that the number of commiters be reduced."
A tool is 99% accurate in detecting defective lines of code. Should developers use the tool? Why or why not?
What are the internal validity issues and external validity issues with this method? "Researcher X finds that a lack of modularity leads to more defects in Windows, and Y is going to apply that predict defects in Eclipse."
Design a study to see whether people who go to lunch together have fewer build defects in their software.
Which would product fewer false positives: 90% recall and 10% precision, or 10% precision and 90% recall?
Given a table of bug reports with severity, etc. and another table of users with qualifications, etc., determine whether experience and bug report frequency are correlated, and if so, how strongly.
Define: evolutionary coupling, tokenizing, word nets, stemming, n-gram, entropy.
List 10 sources of data that could be mined to estimate the risks to a software projects, and describe the limitations of each.
Interpretation and Actionability
Your boss has asked you to generate documentation for a legacy system that doesn't have any. What approach(es) would you use to automatically generate some useful documentation for each class and method?
Given a set of version control logs, how would you tell which commits were bug fixes (vs. adding new features)?
What technique(s) would you use to correlate email messages from a mailing list archive with related version control commits?
Given a data set (mailing list archive, bug reports, and version control log), anonymize it so that it can be shared without risk.
Is it ethical to do an experiment to find out whether one race or gender produces more bugs than another? Justify your answer. How about graduates of one university vs. another?