specification mining with few false positives
DESCRIPTION
Specification Mining With Few False Positives. Claire Le Goues Westley Weimer University of Virginia March 25, 2009. Slide 0.5: Hypothesis. We can use measurements of the “trustworthiness” of source code to mine specifications with few false positives. . Slide 0.5: Hypothesis. - PowerPoint PPT PresentationTRANSCRIPT
![Page 1: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/1.jpg)
1
Specification Mining With Few False Positives
Claire Le GouesWestley WeimerUniversity of VirginiaMarch 25, 2009
![Page 2: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/2.jpg)
2
Slide 0.5: HypothesisWe can use measurements of
the “trustworthiness” of source code to mine
specifications with few false positives.
![Page 3: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/3.jpg)
3
Slide 0.5: HypothesisWe can use measurements of
the “trustworthiness” of source code to mine
specifications with few false positives.
![Page 4: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/4.jpg)
4
Slide 0.5: HypothesisWe can use measurements of
the “trustworthiness” of source code to mine
specifications with few false positives.
![Page 5: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/5.jpg)
5
Slide 0.5: HypothesisWe can use measurements of
the “trustworthiness” of source code to mine
specifications with few false positives.
![Page 6: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/6.jpg)
6
Outline•Motivation: Specifications•Problem: Specification Mining•Solution: Trustworthiness•Evaluation: 3 Experiments•Conclusions
![Page 7: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/7.jpg)
Specifications
7
![Page 8: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/8.jpg)
8
Why Specifications?•Modifying code, correcting defects, and
evolving code account for as much as 90% of the total cost of software projects.
•Up to 60% of maintenance time is spent studying existing software.
•Specifications are useful for debugging, testing, maintaining, refactoring, and documenting software.
![Page 9: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/9.jpg)
9
Our Definition (Broadly)
A specification is a formal description of some aspect of legal program behavior.
![Page 10: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/10.jpg)
10
What kind of specification?•We would like specifications that are simple and machine-readable
•We focus on partial-correctness specifications describing temporal properties ▫Describes legal sequences of events,
where an event is a function call; similar to an API.
•Two-state finite state machines
![Page 11: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/11.jpg)
11
Example Specification
Event A: Mutex.lock()Event B: Mutex.unlock()
![Page 12: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/12.jpg)
12
Example: Locks
1 21
Mutex.lock()
Mutex.unlock()
![Page 13: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/13.jpg)
13
Our Specifications•For the sake of this work, we are talking
about this type of two-state temporal specifications.
•These specifications correspond to the regular expression (ab)*▫More complicated patterns are possible.
![Page 14: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/14.jpg)
The Problem
14
![Page 15: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/15.jpg)
15
Where do formal specifications come from?•Formal specifications are useful, but there aren’t as many as we would like.
•We use specification mining to automatically derive the specifications from the program itself.
![Page 16: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/16.jpg)
16
Mining 2-state Temporal Specifications•Input: program traces – a sequence of
events that can take place as the program runs.▫Consider pairs of events that meet certain
criteria.▫Use statistics to figure out which ones are
likely true specifications.•Output: ranked set of candidate
specifications, presented to a programmer for review and validation.
![Page 17: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/17.jpg)
17
Problem: False Positives Are Common
Event A: Iterator.hasNext()Event B: Iterator.next()
•This is very common behavior.•This is not required behavior.
▫Iterator.hasNext() does not have to be followed eventually by Iterator.next() in order for the code to be correct.
•This candidate specification is a false positive.
![Page 18: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/18.jpg)
18
Previous WorkBenchmark LOC Candidate
SpecsFalse Positive
RateInfinity 28K 10 90%Hibernate 57K 51 82%Axion 65K 25 68%Hsqldb 71K 62 89%Cayenne 86K 35 86%Sablecc 99K 4 100%Jboss 107K 114 90%Mckoi-sql 118K 156 88%Ptolemy2 362K 192 95% *
Adap
ted
from
Wei
mer
-Nec
ula,
TAC
AS 2
005
![Page 19: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/19.jpg)
19
Previous WorkBenchmark LOC Candidate
SpecsFalse Positive
RateInfinity 28K 10 90%Hibernate 57K 51 82%Axion 65K 25 68%Hsqldb 71K 62 89%Cayenne 86K 35 86%Sablecc 99K 4 100%Jboss 107K 114 90%Mckoi-sql 118K 156 88%Ptolemy2 362K 192 95% *
Adap
ted
from
Wei
mer
-Nec
ula,
TAC
AS 2
005
![Page 20: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/20.jpg)
Our Solution: Trustworthiness
20
![Page 21: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/21.jpg)
21
The Problem (as we see it)•Let’s pretend we’d like to learn the rules
of English grammar.•…but all we have is a stack of high school
English papers.•Previous miners ignore the differences
between A papers and F papers.•Previous miners treat all traces as though
they were all equally indicative of correct program behavior.
![Page 22: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/22.jpg)
22
Solution: Code Trustworthiness•Trustworthy code is unlikely to exhibit API policy violations.
•Candidate specifications derived from trustworthy code are more likely to be true specifications.
![Page 23: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/23.jpg)
23
What is trustworthy code?Informally…•Code that hasn’t been changed recently•Code that was written by trustworthy
developers•Code that hasn’t been cut and pasted all
over the place•Code that is readable•Code that is well-tested•And so on.
![Page 24: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/24.jpg)
24
Can you firm that up a bit?•Multiple surface-level, textual, and
semantic features can reveal the trustworthiness of code▫Churn, author rank, copy-paste
development, readability, frequency, feasibility, density, and others.
•Our miner should believe that lock() – unlock() is a specification if it is often followed on trustworthy traces and often violated on untrustworthy ones.
![Page 25: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/25.jpg)
25
A New Miner•Statically estimate the trustworthiness of
each code fragment.•Lift that judgment to program traces by
considering the code visited along the trace.
•Weight the contribution of each trace by its trustworthiness when counting event frequencies while mining.
![Page 26: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/26.jpg)
26
Incorporating Trustworthiness•We use linear regression on a set of
previously published specifications to learn good weights for the different trustworthiness factors.
•Different weights yield different miners.
![Page 27: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/27.jpg)
Evaluation
27
![Page 28: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/28.jpg)
28
Experimental Questions•Can we use trustworthiness metrics to
build a miner that finds useful specifications with few false positives?
•Which trustworthiness metrics are the most useful in finding specifications?
•Do our ideas about trustworthiness generalize?
![Page 29: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/29.jpg)
29
Experimental Questions•Can we use trustworthiness metrics to build a miner that finds useful specifications with few false positives?
•Which trustworthiness metrics are the most useful in finding specifications?
•Do our ideas about trustworthiness generalize?
![Page 30: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/30.jpg)
30
Experimental Setup: Some Definitions•False positive: an event pair that
appears in the candidate list, but a program trace may contain only event A and still be correct.
•Our normal miner balances true positives and false positives (maximizes F-measure)
•Our precise miner avoids false positives (maximizes precision)
![Page 31: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/31.jpg)
31
Experiment 1: A New MinerNormal Miner
Precise Miner
WN
Program
False
Violations
False
Violations False Viola
tions
Hibernate
53% 279 17% 153 82% 93
Axion 42% 71 0% 52 68% 45
Hsqldb 25% 36 0% 5 89% 35
jboss 84% 255 0% 12 90% 94
Cayenne 58% 45 0% 23 86% 18
Mckoi-sql
59% 20 0% 7 88% 69
ptolemy 14% 44 0% 13 95% 72
Total 69% 740 5% 265 89% 426
On this dataset:• Our normal miner produces 107 false positive specifications.• Our precise miner produces 1• The previous work produces 567.
![Page 32: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/32.jpg)
32
More Thoughts On Experiment 1•Our normal miner improves on the false positive
rate of previous miners by 20%. •Our precise miner offers an order-of-magnitude
improvement on the false positive rate of previous work.
•We find specifications that are more useful in terms of bug finding: we find 15 bugs per mined specification, where previous work only found 7.
•In other words: we find useful specifications with fewer false positives.
![Page 33: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/33.jpg)
33
Experimental Questions•Can we use trustworthiness metrics to
build a miner that finds useful specifications with few false positives?
•Which trustworthiness metrics are the most useful in finding specifications?
•Do our ideas about trustworthiness generalize?
![Page 34: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/34.jpg)
34
Experiment 2: Metric Importance
• Results of an analysis of variance (ANOVA).• Shows the importance of the trustworthiness metrics.• F is the predictive power (1.0 means no power).• p is the probability that it had no effect (smaller is better).
Metric F pFrequency 32.3 0.0000Copy-Paste 12.4 0.0004Code Churn 10.2 0.0014Density 10.4 0.0013Readability 9.4 0.0021Feasibility 4.1 0.0423Author Rank 1.0 0.3284Exceptional 10.8 0.0000Dataflow 4.3 0.0000Same Package 4.0 0.0001One Error 2.2 0.0288
![Page 35: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/35.jpg)
35
More Thoughts on Experiment 2
• Statically predicted path frequency has the strongest predictive power.
Metric F pFrequency 32.
30.0000
Copy-Paste 12.4 0.0004Code Churn 10.2 0.0014Density 10.4 0.0013Readability 9.4 0.0021Feasibility 4.1 0.0423Author Rank 1.0 0.3284Exceptional 10.8 0.0000Dataflow 4.3 0.0000Same Package 4.0 0.0001One Error 2.2 0.0288
![Page 36: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/36.jpg)
36
More Thoughts on Experiment 2
• Statically predicted path frequency has the strongest predictive power.• Author rank has no effect on the model.
Metric F pFrequency 32.3 0.0000Copy-Paste 12.4 0.0004Code Churn 10.2 0.0014Density 10.4 0.0013Readability 9.4 0.0021Feasibility 4.1 0.0423Author Rank 1.0 0.3284Exceptional 10.8 0.0000Dataflow 4.3 0.0000Same Package 4.0 0.0001One Error 2.2 0.0288
![Page 37: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/37.jpg)
37
More Thoughts on Experiment 2
Metric F pFrequency 32.3 0.0000Copy-Paste 12.4 0.0004Code Churn 10.2 0.0014Density 10.4 0.0013Readability 9.4 0.0021Feasibility 4.1 0.0423Author Rank 1.0 0.3284Exceptional 10.
80.0000
Dataflow 4.3 0.0000Same Package
4.0 0.0001
One Error 2.2 0.0288
• Statically predicted path frequency has the strongest predictive power.• Author rank has no effect on the model.• Previous work falls somewhere in the middle.
![Page 38: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/38.jpg)
38
Experimental Questions•Can we use trustworthiness metrics to
build a miner that finds useful specifications with few false positives?
•Which trustworthiness metrics are the most useful in finding specifications?
•Do our ideas about trustworthiness generalize?
![Page 39: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/39.jpg)
39
Experiment 3: Does it generalize?•Previous work claimed that more input is
necessarily better for specification mining.
•We hypothesized that smaller, more trustworthy input sets would yield more accurate output from previously implemented tools.
![Page 40: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/40.jpg)
40
Experiment 3: Generalizing
![Page 41: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/41.jpg)
41
Experiment 3: Generalizing
![Page 42: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/42.jpg)
42
Experiment 3: Generalizing
![Page 43: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/43.jpg)
43
Experiment 3: Generalizing
![Page 44: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/44.jpg)
44
Experiment 3: Generalizing• The top 25% “most trustworthy” traces make for a much more accurate miner; the opposite effect is true for the 25% “least trustworthy” traces.• We can throw out the least trustworthy 40-50% of traces and still find the exact same specifications with a slightly lower false positive rate.• More traces != better, so long as the traces are trustworthy.
![Page 45: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/45.jpg)
45
Experimental Summary• We can use trustworthiness metrics to Build a
Better Miner: our normal miner improves on the false positive rate of previous work by 20%, our precise miner by an order of magnitude, while still finding useful specifications.
• Statistical techniques show that our notion of trustworthiness contributes significantly to our success.
• We can increase the precision and accuracy of previous techniques by using a trustworthy subset of the input.
![Page 46: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/46.jpg)
Conclusions
46
![Page 47: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/47.jpg)
47
Summary•Formal specifications are very useful.•The previous work in specification mining
yields too many false positives for industrial practice.
•We developed a notion of trustworthiness to evaluate the likelihood that code adheres to two-state temporal specifications.
![Page 48: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/48.jpg)
48
ConclusionA specification miner that
incorporates notions of code trustworthiness can mine useful specifications with a much lower
false positive rate.
![Page 49: Specification Mining With Few False Positives](https://reader035.vdocuments.site/reader035/viewer/2022062310/56816931550346895de07fdd/html5/thumbnails/49.jpg)
49
The End(questions?)