Anda di halaman 1dari 10

Manual Testing

1. Getting Started 1.1 What is Software? 1.2 Different Kinds of Software 1.3 What are bugs in Software? 1.4 Why does software have bugs? 2. Software Testing 2.1 Basic Concepts 2.2 Psychology of Testing 2.3 Software Testing Definitions 2.4 Motivations of Software Testing: Why? 2.5 Evolution of Software Testing: Where? 2.6 Testing Concepts 2.7 Career in Testing 3. Quality Fundamentals 3.1 Definitions of Quality 3.2 Views of Quality Gurus 3.3 Shewharts PDCA or Demings Wheel 3.4 Quality Assurance (QA) vs. Quality Control (QC) 3.5 Verification and Validation (V&V) 3.6 The Cost of Quality 3.7 Managing Quality 4 Software Development Life Cycle Models 4.1 Phases of Software Project 4.1.1 Requirement gathering and analysis 4.1.2 Planning 4.1.3 Design 4.1.4 Development or Coding 4.1.5 Testing 4.1.6 Deployment and Maintenance 4.2 Life Cycle Models 4.2.1 Waterfall Model

4.2.2 Modified Waterfall Model 4.2.3 Prototype Model 4.2.4 Iterative Enhancement Model 4.2.5 Spiral Model 4.2.6 V-Model 5. Test Case Design - Black Box Testing approach 5.1 What is Black Box Testing? 5.2 Why Black Box Testing? 5.3 When to do Black Box Testing? 5.4 How to do Black Box Testing? 5.4.1 Equivalence Partitioning 5.4.2 Boundary Value Analysis 5.4.3 Cause effect graphing using Decision Tables 6. Defect/Incident Reporting and Tracking 6.1 Outcome of Test Execution 6.2 Defect Log 6.3 Defect Tracking Workflow 6.4 Defect Metrics 7. Levels of Testing 7.1 Unit Testing 7.1.1 Unit test plan 7.1.2 Unit test case 7.1.3 Defect reporting and tracking 7.2 Integration Testing 7.2.1 Big Bang Testing 7.2.2 Bottom up Integration 7.2.3 Top down Integration 7.2.4 Sandwich Integration 7.2.5 Types of checks in Integration 7.2.6 Integration test plan 7.2.7 Integration test case 7.2.8 Defect reporting and tracking 7.3 System Testing 7.3.1 Business Process Testing

7.3.2 Performance Testing 7.3.3 Volume Testing 7.3.4 Stress Testing 7.3.5 Security Testing 7.3.6 Recovery Testing 7.3.7 Compatibility Testing 7.3.8 Scalability Testing 7.3.9 Usability Testing 7.4 Acceptance Testing 7.5 Regression Testing 8. Test Case Design - White Box Testing approach 8.1 What is White Box Testing? 8.2 Classification of White Box Testing 8.3 Structural Testing 8.3.1 Code Coverage Testing 8.3.1.1 Statement Coverage 8.3.1.2 Decision Coverage 8.3.1.3 Path Coverage 8.3.1.4 Function Coverage 9. Testing Process 9.1 Process Approach and ETVX Model 9.2 Test Strategy 9.3 Test Planning 9.4 Test Design 9.5 Test Execution 9.6 Defect Reporting and Tracking 9.7 Metrics and Analysis 9.8 Deliverables and Quality Records 10. Software Project Environment 10.1 Software Project Management 10.2 Software Quality Management 10.3 Software Configuration Management 10.4 Software Risk Management 10.5 Test Organization Structure

11. Test Automation 11.1 Motivation for Automating Tests 11.2 Code Coverage and Memory Checkers 11.3 Functional/Regression, Performance and Management Testing Tools 11.4 Comparison of Automation Vs Manual 12. Human Aspects of Testing 12.1 Roles and Responsibilities in Software Testing 12.2 Attitudes and Competencies of a Successful Tester

Quick Test Professional


1. Introduction to QuickTest 1.1 Preparing to Automate 1.2 QuickTest Benefits in Automation 1.3 Overall QuickTest Workflow 2. Creating a Basic Test 2.1 Preparing to Record 2.2 Planning the Initial and End Conditions 2.3 Determine the data to be used 2.4 Understanding the QuickTest workflow 2.5 Add-in Manager 2.6 Understanding the QuickTest Window 2.7 Understanding the Keyword View

2.8 Understanding the Expert View 2.9 Recording a Test 2.10 Replaying a Test 2.11 Analyzing the QuickTest Result 2.12 Test Directory Structure 3. Working with Objects 3.1 QuickTest Object Identification 3.2 Defining QuickTest Objects and Classes 3.3 Analyzing the Object Property : Value Pairs 3.4 Using the Object Spy 3.5 Object Repository Stored Test Objects 3.6 Defining a New Test Object 3.7 Renaming the Logical Name 3.8 Using the Object Repository Tool bar 4 Synchronizing Scripts 4.1 Understanding Synchronization 4.2 Adding Synchronization 5. Verifying with Standard Checkpoints 5.1 Understanding Checkpoints 5.2 Inserting a Checkpoint During Recording 5.3 Inserting a Checkpoint After Recording 5.4 Renaming and Modifying a Checkpoint 6. Verifying with Other Checkpoints 6.1 Verifying with Text Checkpoints 6.2 Verifying with Text Area Checkpoints 6.3 Inserting a Bitmap Checkpoint 6.4 Using Regular Expressions 7. Using Parameters 7.1 Parameter Types 7.2 Input Parameters for Data Driven Tests

7.3 Using Design Time Data Table 7.4 Using Global and Local Data Sheets 7.5 Using Random Number Parameter 7.6 Using Environment Parameters 7.7 Using Built in User Defined Variables 7.8 Using Output Parameters 7.9 Inserting Output Parameter While Recording 7.10 Inserting Output Parameter After Recording 7.11 Reusing Output as Input 8. Creating Tests with Multiple Actions 8.1 A Test with Multiple Actions 8.2 Creating a New Action 8.3 One Action Used Many Times 8.4 Using External Actions 8.5 Using Global and Local Data Sheets in a Test 8.6 Managing Global and Local Iteration Settings 9. Creating Test on a Web Application 9.1 Exploring a Web Application 9.2 Identifying Web Objects 9.3 Specifying Record and Run Setting for a Web 9.4 Specifying Time out for a Web Page 9.5 Using Checkpoints in a Web Page 10. Creating Database Checkpoints 10.1 Defining a Database 10.2 Introduction to SQL Statements 10.3 Using the SELECT, FROM and WHERE clauses 10.4 Inserting a Database Checkpoint 10.5 Viewing the Database Checkpoint Properties 10.6 Analyzing the Database Checkpoint Properties 11. Using Shared Object Repository 11.1 Types of Object Repositories 11.2 Local Object Repository

11.3 Shared Object Repository 11.4 Merging Object Repositories 12. Handling Object Recognition Problems 12.1 The Normal Recording Mode 12.2 Performing Low-level Recording 12.3 Performing Analog Recording 12.4 Creating Virtual Objects 13. Introduction to Expert View 13.1 An Overview of Expert View 13.2 An Overview of VB Script 13.3 Objects and Classes in QuickTest 13.4 Adding New Steps in Expert View 13.5 Categories in Step Generator 13.6 Inserting Steps Manually 13.7 Adding Conditional Statements in the Test 13.8 Adding If Else Statement 13.9 Adding the ForNext Statement 14. Working with Object Properties 14.1 Retrieving Test Object Properties 14.2 Using the GetROProperty and GetTOProperty 14.3 Setting Object Properties 15. Using Recovery Scenario for Exception Handling 15.1 Introduction to Recovery Scenario 15.2 Selecting the Trigger Event 15.3 Specifying the Recovery Operation 15.4 Specifying the Post Recovery Operation 15.5 Adding an Existing Recovery Scenario to a Test 15.6 Viewing the Test Results

International testing

Quality Center
1. Test Management concepts and process 1.1 Requirements Management 1.2 Test Planning 1.3 Task Allocation 1.4 Test Execution 1.5 Defect Tracking 2. Site Administrator Features 2.1 Create project 2.2 Create Global users 2.3 Ping project database 3. Project Administrator Features 3.1 Create User Groups 3.2 Modify Access rights 3.3 Create Users 3.4 Manage project entities and look-up values 3.5 Manage system fields and user defined fields 3.6 Define alerts 3.7 Define access rights at field levels 4. Requirements Management 4.1 Add macro requirement 4.2 Add micro requirements 4.3 Maintain versions 4.4 Change status of requirements 4.5 Different views of requirements, filters 4.6 Generate reports and graphs 5. Test Planning 5.1 Manage plan folders 5.2 Manage test cases manual and automated

5.3 Attach documents 5.4 Manage test steps and test scripts 5.5 Map requirements to test cases 5.6 Check requirements coverage 5.7 Generate reports and graphs 5.8 Tree view and grid view, filters 6. Test Lab Execution 6.1 Define test set folders and test sets 6.2 Assign test cases and testers to test rounds 6.3 Execute tests and log results 6.4 Generate reports and graphs 6.5 Different views and filters 7. Defect Tracking 7.1 Add defects from test lab 7.2 Add defects independently 7.3 Defect tracking work flow 7.4 Defect history watch, filters 7.5 Generate reports and graphs 8. Review of specific issues, Evaluation

Load Runner
1. Load Testing Concepts 1.1 Load, Stress and Scalability tests 1.2 Load test planning 1.3 Performance Counters 1.4 Protocols 1.5 Load test environment 2. Virtual User Generator 2.1 Protocol Selection 2.2 Recording actions

2.3 Commenting, Logging, Think Time 2.4 Transaction Processing 2.5 Rendezvous Points 2.6 Synchronization Points 2.7 Data Parameterization 2.8 Run and Debug 2.9 Script and Tree Views 3. Scenario Controller 3.1 Manual and goal oriented Scenarios 3.2 Script Mapping 3.3 Configuring a run 3.4 Single run, Iteration run 3.5 Think time, timeout Manipulation 3.6 Network Speed, browser emulations 3.7 Runtime settings 3.8 Scheduled runs 3.9 Ramp-up and ramp-down runs 3.10 Distributed runs using agent process 3.11 Configuring performance counters 4. Analyzer 4.1 Vuser and run detail graphs 4.2 Throughput and his graphs 4.3 User wise performance 4.4 Page wise performance 4.5 Page break down graphs 4.6 Bottle neck identification

Anda mungkin juga menyukai