1 / 14

Software Release Readiness Metric ShipIT [0,1]

Software Release Readiness Metric ShipIT [0,1]. Piyush Ranjan Satapathy Department of Computer Science & Engineering University of California Riverside. Outline. Why Release Metric ? Stages & Factors Considered Metrics Considered Formulation of ShipIT

summer-kerr
Download Presentation

Software Release Readiness Metric ShipIT [0,1]

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Software Release Readiness MetricShipIT [0,1] Piyush Ranjan Satapathy Department of Computer Science & Engineering University of California Riverside

  2. Outline • Why Release Metric ? • Stages & Factors Considered • Metrics Considered • Formulation of ShipIT • Justification & Elaboration of the formula • Validation of the Formula • Conclusion • References • Q & A

  3. Why Release Metric ? • Cost:Penalty for late delivery? • Confidence:accurate estimate? Complexity? • Quality: Danger of breaking something else? How much testing is needed? • Schedule: Holidays? Vacation? Customer availability? Hard deadline? • Relationship: Unhappy customer? Reference site? • Workaround: Manual? Send staff on site? • Quantify: How much loss? second? • Accuracy:performance test environment? • Alternatives: Optional download? Automatic download? • Support: Ease of download? Increased support load? • Competition:Market leader? Slipping? • Usability: lose customers if slow?

  4. Phases or Factors Considered • RequirementAnalysisDesign Phase • Coding or Implementation Phase • Testing Phase • Quality Assurance Phase • Manuals & Documentation • Early Deployment • Early Support

  5. Metrics Considered 1.RequirementAnalysisDesign Phase  Planned vs Implemented features  List of unimplemented features  List of New features coming on… 2. Coding Phase 2.1 Creating Source  System Modules ( Already Implementged no. Vs Total planned no)  Application Modules ( Planned Vs Implemented no)  GUI Modules (Planned vs Implemented no.) 2.2 Creating Object (Use another way..)  KSLOC (Thousands of Source Line of Codes)  Number of Functional Points (determines Complexity)  Known Anomalies in the code 2.3 Building Process  Compilation Time  No of warnings during compilation  Incremental Build time as per the platform dependency  Incremental build time as per the compiler dependency

  6. Metrics Considered… 3. Testing Phase 3.1 Testing (Finding Bugs)  Unit testing ( Planned Vs Completed no of L1 testcases)  Integration testing (Planned vs Complted no of L2 testcases)  System testing (Planned vs completed no of L3 testcases) 3.2 Debugging  No of open issues  Line Coverage 4. Quality Assurance ( Regression testing)  Total no of Test hours planned  Debugged faults till date  Acceptable no of Faults 5. Manuals & Documentation  Requirement Documentation (% Completed)  Design Documentation (% Completed)  Implementation and Usability Documentation (% Completed)  Test Plan Documentation (% Completed)  User guide Documentation (% Completed)

  7. Metrics Considered… 6. Supervision (Early Deployment) 6.1 Installation Process  Distribution of softwares ( Planned vs Completed)  Installation of software (Planned No. vs Completed No.)  Acceptance testing (No of Planned Vs executed Testcases) 6.2 Training process  Developing Training Materials ( % Completion)  Validating Training program (% Completion)  Implementing Training program (% Completion) 7. Support (Early Customer Feedback)  Handling Beta Customer Bugs  Reapplying the software development cycle  Major Metrics : Maintainability Index desired and Maintainability Index reached

  8. Formulation of ‘ShipIT’ ShipIT = [(WRAD x RAD) + (WCODE x CODE) + (WTEST x TEST) + (WQA x QA) + (WMD x MD) + (WSV x SV) +(WSP x SP)] / 100 Where, RAD = Factor of contribution towards the completion of software development from RequirementsAnalysisDesign Stage So are the CODE, TEST, QA, MD, SV and SP respectively… And WRAD, WCODE,, WTEST , WQA, WMD , WSV, WSP are all Є [0,100] and all sum up to 100.

  9. Justification & Elaboration (1) Assumptions: 1. Perfect Waterfall Model ( No coding or testing phase until requirement analysis done) 2. Target is only Major release 3. Release after deploying in Customers site and obtaining the proper MI. 4. New requirements until the end of Detailed Design phase..Not after that..if so then for next version of release.. Elaboration: RAD = [( WR x R) + (WA x A) + (WD x D)] / 100; Where WR +WA + WD = 100 and R,A,D Є [0,1]. CODE = [(Wsource x Source) + (Wobject x Object) + (WBuild x Build)] / 100 TEST = [(WBugfinding x Bugfinding) + (WDebugging x Debugging)] / 100 QA = [(Pseudo Test Hours Completed) / (Total test hours planned) MD = [(WRD x RD) + (WDD x DD) + (WID x ID) + (WTD x TD) + (WUD xUD) ] / 100 SV = [(WIP x IP) + (WTP x TP)]/ 100 SP = [Maintainability Index reached / Maintainability Index desired]

  10. Justification & Elaboration (2) Used Methods & Models: 1. COCOMO Prediction Model Effort = a (KSLOC)b Time = a (effort)b 2. Halstead’s Metrics Model Effort = n1N2NlogN/ 2n2 T = E/18 Sec. 3. Albrecht’s Function Points Model LOC = SourceStatement x FP FP = UFC x TCF 4. Zero failure Method 5. Stopping Rules Method 6. Maintainability Index Method

  11. Validation of the Formula ShipIT = [(22 x RAD) + (19 x CODE) + (30 x TEST) + (8 x QA) + (7 x MD) + (9 x SV) + (5 x SP) ] / 100(From Research Data) [ref 1.] RAD = [( 30 x R) + (20 x A) + (50 x D)] / 100 CODE = [(40 x Source) + (20 x Object) + (40 x Build)] / 100 TEST = [(65 x Bugfinding) + (35 x Debugging)] / 100 QA = QA MD = [(15 x RD) + (15 x DD) + (15 x ID) + (25 x TD) + (30 xUD) ] / 100 SV = [(50 x IP) + (50 x TP)]/ 100 SP = SP Source = [(57 x Sm) + (28 x Am) + (15 x Gm)] / 100 (From Ref 2) Build = [(40x CT) + (30 x HT) + (15 x BPT) + (15 x BCT)] (from Ref2) Bugfinding = [(35 xL1) + (35 xL2) + (30 xL3)] /100

  12. Conclusions • Software Release Readiness Metric Important for Market Vs Features Vs Quality • no one tool or method should be relied on to arbitrarily make the final determination of whether a software product should be released • Detecting the measurable factors in software development life cycle is a skill and comes from Experience… • Considering the most practical metrics and under certain assumptions the formula defined for “ShipIT” holds true !!!!!!!

  13. References …..24 Research papers…Can’t put all.. Ref1. Robert B. Grady, Hewlett-Packard, “Successfully Applying Software metrics”, IEEE Trans. Soft Engr., September 1994 (Vol. 27, No. 9), pp. 18-25 Ref2. Gregory A. Hansen, GAPI, “Simulating Software Development Processes”, IEEE Software, January 1996 (Vol. 29, No. 1), pp. 73-77

  14. Q & A

More Related