| Cart Total:

Certified Reliability Engineer Question Bank

Format Web-based

Simulate the CRE exam from the convenience of your home or office. ASQ’s online certification preparation course teaches you the concepts you’ll see on your test, and provides practice exams so you can review the material. This gives you the best chance to learn effectively and pass your exam.

Increase the effectiveness of your study for ASQ's Certified Reliability Engineer (CRE) exam through access to hundreds of practice questions developed and based on the CRE Body of Knowledge (BOK)*. Use these questions to identify your strengths and weaknesses, so you know where to concentrate your studies.Study by topic or simulate the exam experience with at least three completely unique sets of questions drawn from the CRE BOK. Take a different random practice exam as many time as you want during the 365-day access period.*Please note that the questions in this product are not actual ASQ exam questions. Your performance on the practice exams is for study purposes only, and may or may not reflect performance on an actual certification exam.

Course Data

  • CEU Hours: 0
  • Length: 0 Hours
  • ASQ RU: 0
  • Audience: Practitioner; Professional
  • Provider: ASQ - Online
Course Overview

Learning Objectives:

Who Should Attend:

The CRE Question Bank is perfect for anyone interested in preparing for and pursuing the CRE certification.

Reliability Engineer Body of Knowledge

The topics in this Body of Knowledge include additional detail in the form of subtext explanations and the cognitive level at which the questions will be written. This information will provide useful guidance for both the Examination Development Committee and the candidates preparing to take the exam. The subtext is not intended to limit the subject matter or be all-inclusive of what might be covered in an exam. It is intended to clarify the type of content to be included in the exam. The descriptor in parentheses at the end of each entry refers to the highest cognitive level at which the topic will be tested. A more comprehensive description of cognitive levels is provided at the end of this document.

    1. Strategic management
      1. Benefits of reliability engineering
        Describe how reliability engineering techniques and methods improve programs, processes, products, systems, and services. (Understand)
      2. Interrelationship of safety, quality, and reliability
        Define and describe the relationships among safety, reliability, and quality. (Understand)
      3. Role of the reliability function in the organization
        Describe how reliability techniques can be applied in other functional areas of the organization, such as marketing, engineering, customer /product support, safety and product liability, etc. (Apply)
      4. Reliability in product and process development
        Integrate reliability engineering techniques with other development activities, concurrent engineering, corporate improvement initiatives such as lean and six sigma methodologies, and emerging technologies. (Apply)
      5. Failure consequence and liability management
        Describe the importance of these concepts in determining reliability acceptance criteria. (Understand)
      6. Warranty management
        Define and describe warranty terms and conditions, including warranty period, conditions of use, failure criteria, etc., and identify the uses and limitations of warranty data. (Understand)
      7. Customer needs assessment
        Use various feedback methods (e.g., quality function deployment (QFD), prototyping, beta testing) to determine customer needs in relation to reliability requirements for products and services. (Apply)
      8. Supplier reliability
        Define and describe supplier reliability assessments that can be monitored in support of the overall reliability program. (Understand)
    2. Reliability program management
      1. Terminology
        Explain basic reliability terms (e.g., MTTF, MTBF, MTTR, availability, failure rate, reliability, maintainability). (Understand)
      2. Elements of a reliability program
        Explain how planning, testing, tracking, and using customer needs and requirements are used to develop a reliability program, and identify various drivers of reliability requirements, including market expectations and standards, as well as safety, liability, and regulatory concerns. (Understand)
      3. Types of risk
        Describe the relationship between reliability and various types of risk, including technical, scheduling, safety, financial, etc. (Understand)
      4. Product lifecycle engineering
        Describe the impact various lifecycle stages (concept/design, introduction, growth, maturity, decline) have on reliability, and the cost issues (product maintenance, life expectation, software defect phase containment, etc.) associated with those stages. (Understand)
      5. Design evaluation
        Use validation, verification, and other review techniques to assess the reliability of a product’s design at various lifecycle stages. (Analyze)
      6. Systems engineering and integration
        Describe how these processes are used to create requirements and prioritize design and development activities. (Understand)
    3. Ethics, safety, and liability
      1. Ethical issues
        Identify appropriate ethical behaviors for a reliability engineer in various situations. (Evaluate)
      2. Roles and responsibilities
        Describe the roles and responsibilities of a reliability engineer in relation to product safety and liability. (Understand)
      3. System safety
        Identify safety-related issues by analyzing customer feedback, design data, field data, and other information. Use risk management tools (e.g., hazard analysis, FMEA, FTA, risk matrix) to identify and prioritize safety concerns, and identify steps that will minimize the misuse of products and processes. (Analyze)

    1. Basic concepts
      1. Statistical terms
        Define and use terms such as population, parameter, statistic, sample, the central limit theorem, etc., and compute their values. (Apply)
      2. Basic probability concepts
        Use basic probability concepts (e.g., independence, mutually exclusive, conditional probability) and compute expected values. (Apply)
      3. Discrete and continuous probability distributions
        Compare and contrast various distributions (binomial, Poisson, exponential, Weibull, normal, log-normal, etc.) and their functions (e.g., cumulative distribution functions (CDFs), probability density functions (PDFs), hazard functions), and relate them to the bathtub curve. (Analyze)
      4. Poisson process models
        Define and describe homogeneous and non-homogeneous Poisson process models (HPP and NHPP). (Understand)
      5. Non-parametric statistical methods
        Apply non-parametric statistical methods, including median, Kaplan-Meier, Mann-Whitney, etc., in various situations. (Apply)
      6. Sample size determination
        Use various theories, tables, and formulas to determine appropriate sample sizes for statistical and reliability testing. (Apply)
      7. Statistical process control (SPC) and process capability
        Define and describe SPC and process capability studies (Cp, Cpk, etc.), their control charts, and how they are all related to reliability. (Understand)
    2. Statistical inference
      1. Point estimates of parameters
        Obtain point estimates of model parameters using probability plots, maximum likelihood methods, etc. Analyze the efficiency and bias of the estimators. (Evaluate)
      2. Statistical interval estimates
        Compute confidence intervals, tolerance intervals, etc., and draw conclusions from the results. (Evaluate)
      3. Hypothesis testing (parametric and non-parametric)
        Apply hypothesis testing for parameters such as means, variance, proportions, and distribution parameters. Interpret significance levels and Type I and Type II errors for accepting/rejecting the null hypothesis. (Evaluate)

    1. Reliability design techniques
      1. Environmental and use factors
        Identify environmental and use factors (e.g., temperature, humidity, vibration) and stresses (e.g., severity of service, electrostatic discharge (ESD), throughput) to which a product may be subjected. (Apply)
      2. Stress-strength analysis
        Apply stress-strength analysis method of computing probability of failure, and interpret the results. (Evaluate)
      3. FMEA and FMECA
        Define and distinguish between failure mode and effects analysis and failure mode, effects, and criticality analysis and apply these techniques in products, processes, and designs. (Analyze)
      4. Common mode failure analysis
        Describe this type of failure (also known as common cause mode failure) and how it affects design for reliability. (Understand)
      5. Fault tree analysis (FTA) and success tree analysis (STA)
        Apply these techniques to develop models that can be used to evaluate undesirable (FTA) and desirable (STA) events. (Analyze)
      6. Tolerance and worst-case analyses
        Describe how tolerance and worst-case analyses (e.g., root of sum of squares, extreme value) can be used to characterize variation that affects reliability. (Understand)
      7. Design of experiments
        Plan and conduct standard design of experiments (DOE) (e.g., full-factorial, fractional factorial, Latin square design). Implement robust-design approaches (e.g., Taguchi design, parametric design, DOE incorporating noise factors) to improve or optimize design. (Analyze)
      8. Fault tolerance
        Define and describe fault tolerance and the reliability methods used to maintain system functionality. (Understand)
      9. Reliability optimization
        Use various approaches, including redundancy, derating, trade studies, etc., to optimize reliability within the constraints of cost, schedule, weight, design requirements, etc. (Apply)
      10. Human factors
        Describe the relationship between human factors and reliability engineering. (Understand)
      11. Design for X (DFX)
        Apply DFX techniques such as design for assembly, testability, maintainability environment (recycling and disposal), etc., to enhance a product’s producibility and serviceability. (Apply)
      12. Reliability apportionment (allocation) techniques
        Use these techniques to specify subsystem and component reliability requirements. (Analyze)
    2. Parts and systems management
      1. Selection, standardization, and reuse
        Apply techniques for materials selection, parts standardization and reduction, parallel modeling, software reuse, including commercial off-the-shelf (COTS) software, etc. (Apply)
      2. Derating methods and principles
        Use methods such as S-N diagram, stress-life relationship, etc., to determine the relationship between applied stress and rated value, and to improve design. (Analyze)
      3. Parts obsolescence management
        Explain the implications of parts obsolescence and requirements for parts or system requalification. Develop risk mitigation plans such as lifetime buy, backwards compatibility, etc. (Apply)
      4. Establishing specifications
        Develop metrics for reliability, maintainability, and serviceability (e.g., MTBF, MTBR, MTBUMA, service interval) for product specifications. (Create)

    1. Reliability modeling
      1. Sources and uses of reliability data
        Describe sources of reliability data (prototype, development, test, field, warranty, published, etc.), their advantages and limitations, and how the data can be used to measure and enhance product reliability. (Apply)
      2. Reliability block diagrams and models
        Generate and analyze various types of block diagrams and models, including series, parallel, partial redundancy, time-dependent, etc. (Create)
      3. Physics of failure models
        Identify various failure mechanisms (e.g., fracture, corrosion, memory corruption) and select appropriate theoretical models (e.g., Arrhenius, S-N curve) to assess their impact. (Apply)
      4. Simulation techniques
        Describe the advantages and limitations of the Monte Carlo and Markov models. (Apply)
      5. Dynamic reliability
        Describe dynamic reliability as it relates to failure criteria that change over time or under different conditions. (Understand)
    2. Reliability predictions
      1. Part count predictions and part stress analysis
        Use parts failure rate data to estimate system- and subsystem-level reliability. (Apply)
      2. Reliability prediction methods
        Use various reliability prediction methods for both repairable and non-repairable components and systems, incorporating test and field reliability data when available (Apply)

  5. RELIABILITY TESTING (24 Questions)
    1. Reliability test planning
      1. Reliability test strategies
        Create and apply the appropriate test strategies (e.g., truncation, test–to-failure, degradation) for various product development phases. (Create)
      2. Test environment
        Evaluate the environment in terms of system location and operational conditions to determine the most appropriate reliability test. (Evaluate)
    2. Testing during development
      Describe the purpose, advantages, and limitations of each of the following types of tests, and use common models to develop test plans, evaluate risks, and interpret test results. (Evaluate)
      1. Accelerated life tests (e.g., single-stress, multiple-stress, sequential stress, step-stress)
      2. Discovery testing (e.g., HALT, margin tests, sample size of 1),
      3. Reliability growth testing (e.g., test, analyze, and fix (TAAF), Duane)
      4. Software testing (e.g., white-box, black-box, operational profile, and  fault-injection)
    3. Product testing
      Describe the purpose, advantages, and limitations of each of the following types of tests, and use common models to develop product test plans, evaluate risks, and interpret test results. (Evaluate)
      1. Qualification/demonstration testing (e.g., sequential tests, fixed-length tests)
      2. Product reliability acceptance testing (PRAT)
      3. Ongoing reliability testing (e.g., sequential probability ratio test [SPRT])
      4. Stress screening (e.g., ESS, HASS, burn-in tests)
      5. Attribute testing (e.g., binomial, hypergeometric)
      6. Degradation (wear–to-failure) testing

    1. Management strategies
      1. Planning
        Develop plans for maintainability and availability that support reliability goals and objectives. (Create)
      2. Maintenance strategies
        Identify the advantages and limitations of various maintenance strategies (e.g., reliability-centered maintenance (RCM), predictive maintenance, repair or replace decision making), and determine which strategy to use in specific situations. (Apply).
      3. Availability tradeoffs
        Describe various types of availability (e.g., inherent, operational), and the tradeoffs in reliability and maintainability that might be required to achieve availability goals. (Apply)
    2. Maintenance and testing analysis
      1. Preventive maintenance (PM) analysis
        Define and use PM tasks, optimum PM intervals, and other elements of this analysis, and identify situations in which PM analysis is not appropriate. (Apply)
      2. Corrective maintenance analysis
        Describe the elements of corrective maintenance analysis (e.g., fault-isolation time, repair/replace time, skill level, crew hours) and apply them in specific situations. (Apply)
      3. Non-destructive evaluation
        Describe the types and uses of these tools (e.g., fatigue, delamination, vibration signature analysis) to look for potential defects. (Understand)
      4. Testability
        Use various testability requirements and methods (e.g., built in tests (BITs), false-alarm rates, diagnostics, error codes, fault tolerance) to achieve reliability goals (Apply)
      5. Spare parts analysis
        Describe the relationship between spare parts requirements and reliability, maintainability, and availability requirements. Forecast spare parts requirements using field data, production lead time data, inventory and other prediction tools, etc. (Analyze)

  7. DATA COLLECTION AND USE (18 Questions)
    1. Data collection
      1. Types of data
        Identify and distinguish between various types of data (e.g., attributes vs. variable, discrete vs. continuous, censored vs. complete, univariate vs. multivariate). Select appropriate data types to meet various analysis objectives. (Evaluate)
      2. Collection methods
        Identify appropriate methods and evaluate the results from surveys, automated tests, automated monitoring and reporting tools, etc., that are used to meet various data analysis objectives. (Evaluate)
      3. Data management
        Describe key characteristics of a database (e.g., accuracy, completeness, update frequency). Specify the requirements for reliability-driven measurement systems and database plans, including consideration of the data collectors and users, and their functional responsibilities. (Evaluate)
    2. Data use
      1. Data summary and reporting
        Examine collected data for accuracy and usefulness. Analyze, interpret, and summarize data for presentation using techniques such as trend analysis, Weibull, graphic representation, etc., based on data types, sources, and required output. (Create)
      2. Preventive and corrective action
        Select and use various root cause and failure analysis tools to determine the causes of degradation or failure, and identify appropriate preventive or corrective actions to take in specific situations. (Evaluate)
      3. Measures of effectiveness
        Use various data analysis tools to evaluate the effectiveness of preventive and corrective actions in improving reliability. (Evaluate)
    3. Failure analysis and correction
      1. Failure analysis methods
        Describe methods such as mechanical, materials, and physical analysis, scanning electron microscopy (SEM), etc., that are used to identify failure mechanisms. (Understand)
      2. Failure reporting, analysis, and corrective action system (FRACAS)
        Identify the elements necessary for a FRACAS to be effective, and demonstrate the importance of a closed-loop process that includes root cause investigation and follow up. (Apply)

Levels of Cognition
based on Bloom’s Taxonomy – Revised (2001)

In addition to content specifics, the subtext for each topic in this BOK also indicates the intended complexity level of the test questions for that topic. These levels are based on “Levels of Cognition” (from Bloom’s Taxonomy – Revised, 2001) and are presented below in rank order, from least complex to most complex.

Recall or recognize terms, definitions, facts, ideas, materials, patterns, sequences, methods, principles, etc.

Read and understand descriptions, communications, reports, tables, diagrams, directions, regulations, etc.

Know when and how to use ideas, procedures, methods, formulas, principles, theories, etc.
Break down information into its constituent parts and recognize their relationship to one another and how they are organized; identify sublevel factors or salient data from a complex scenario.

Make judgments about the value of proposed ideas, solutions, etc., by comparing the proposal to specific criteria or standards.

Put parts or elements together in such a way as to reveal a pattern or structure not clearly there before; identify which data or information from a complex set is appropriate to examine further or from which supported conclusions can be drawn.


This program is an Internet-based, self-paced training course. Web-based courses require a computer and Internet access. You set the pace making you responsible for your own learning experience.


The learner can take a different random practice exam as many times as they want during the 365-day access period. After attempting each question, you'll see the correct answer and an explanation specific to your choice. Upon completion of a practice exam, you'll get a graphical summary of your scores on each main topic of the BOK.

Additional Notes

  • Please note that the questions in this product are not actual ASQ exam questions. Your performance on the practice exams is for study purposes only, and may or may not reflect performance on an actual certification exam.
System Requirements
Here is what you need to participate in an ASQ web-based, self-paced, or instructor-led virtual WebEx course.

Registration Email

After course registration, you will receive an email with access instructions. If you do not, please check your email junk, spam, or clutter folders. If the email is not there, please contact ASQ. For virtual WebEx courses, you will receive another email 24 hours prior to the start of class containing additional access information.

Technical Requirements

  • Computer – PC or Macintosh
  • High-speed internet connection
  • Adobe Flash Player 10.3 or later
  • Windows: Internet Explorer 8 or later, Google Chrome (latest version), Firefox (latest version)
  • Mac: Safari 7 or later, Google Chrome (latest version), Firefox (latest version)
  • For instructor-led virtual WebEx courses, a speakerphone or telephone headset in a designated room or space with no distractions (preferred)

NOTE: Refer to product description for any additional system requirements.

WebEx – Virtual Course: Advance Preparation and Student Information

Join a Test Session

We strongly suggest that you join a WebEx test session in advance of your scheduled virtual training, using your preferred web browser. This can help prevent problems launching content the day your training begins.

If you do not have administrative privileges on the SAME computer you will be using the day of your training (i.e., your workstation within your organization) you may need the assistance of your local IT/IS help desk staff for the platform to engage correctly.

To join a test session (meeting), click here and enter your name and email address to join. If successful, you will see a screen that says "Congratulations! Your system is now set up properly ..."

Deactivate Pop-up Blockers

You should deactivate any pop-up blockers, spam filters, and company firewalls that could prevent the WebEx client platform or web-based course from working properly. Third-party toolbars such as Bing, Yahoo, and Google should also be disabled due to their own pop-up blocking capabilities.

Cancellation Policy
ASQ reserves the right to cancel or re-schedule courses and to change instructors. Please be advised that in the event of a course cancellation, ASQ is not responsible for airfare penalties or other travel related expenses you may incur.
  • If you need to cancel, we will refund your paid registration fee as noted below.
    • Requests for cancellations/transfers received at least 5 business days before the start of the course receive a full refund/transfer.
    • Requests received within 5 business days of the course starting incur a $150 processing fee.
    • After the course starts, there are no refunds or transfers.
    • Registrants who fail to attend without advance notice are liable for the entire course fee.
    • If you cannot find a substitute, we can transfer your course fees to another ASQ course of your choice.
  • You must successfully complete the entire course or program before IACET CEUs and a Certificate of Completion can be awarded.

Enrollment Details

Format: Classroom
Format: Web-based

Internet-based, self-paced training modules, which may involve combinations of text, visuals, audio, interactive simulations and quizzes (see specific courses for features and tools).  These web-based courses require a computer and Internet access.

QBCREASQ Self-Paced English 365 Days List: $109 Member: $99



ASQ’s customized on-site training programs are the cost-effective way for you to train employees on your schedule. Training on-site minimizes the cost-per-student, reduces travel expenses, and ensures that the content is specific to your exact needs.

5 Benefits of On-Site Training

  1. Expertise to your location: Saves time and travel expenses
  2. Convenient scheduling: Train groups of five or more
  3. Customizable courses: Training tailored to your groups’ needs
  4. Immediate results: Employees complete courses ready to apply what they’ve learned
  5. Value-add support: Instructors available to answer questions after session