nQuery Release Notes | Ver 8.2 | April 2018

Release Notes
nQuery Advanced 8.2 | April 2018 Update

The nQuery April 2018 release will add a wide range of sample size tables ranging from extensions of pre-existing tables for a better and clearer user experience to the those based on the latest academic research and user feedback.

In the April 2018 release, we added 52 new sample size tables to the core product and 20 Bayesian tables. This release summary will provide an overview of what areas have been targeted in this release along with the full list of tables being added.

nQuery Advanced Tables

In the April 2018 release, three main overarching areas were targeted for significant improvement. These were:

  1. Epidemiological Methods
  2. Non-inferiority and Equivalence Tests
  3. Correlation and Diagnostic Testing (ROC) Methods

There is also a number of tables which do not fall into these categories related to areas such as inequality testing for log-normal data, testing of variance and standard deviations and non-parametric tests. These are described at the end of this document.

We will provide background on each of these below and a list of the sample size tables which will be added in the April release. References for each method are provided at the end of this article

1. Epidemiology

Epidemiology is the branch of medicine which primarily studies the incidence, distribution, and possible control of diseases and other factors relating to health. Epidemiological studies are cornerstone of research into areas such as public health, health policy or preventative medicine (e.g. vaccines).

Due to often having to study the effect of medicines and interventions at a more complex society-wide level, processes and methods for epidemiology often adjust for problems that are less prominent in well-controlled clinical trials. These issues include being unable to individually randomise, relying on observational data or attempting to extract causal relationships from highly complex data.

Due to these and other issues, the study designs and statistical methods used by epidemiologists will often have to include adjustments for exogenous effects and have a greater reliance on processes such as pair-matching. While statistical methods for clinical trials provide a useful starting point for getting adequate sample size estimates, there is a growing desire for methods which have found traction in the epidemiological field.

In the April release, 12 new tables will be added with the main areas of focus in the Epidemiology upgrade being the following:

  1. Case-Control Studies (Observational and Prospective)
  2. Vaccine Efficacy Studies
  3. Cluster Randomized Trials (CRT)
  4. Mendelian Randomization Studies

These areas and the tables in each category are explored below.

Case-Control Studies

Case-Control studies are those where the analysis assumes that the effect of a treatment or intervention or prognostic factor can be modelled by comparing the effect on a paired cases and controls. In the epidemiological context, this is most commonly associated with retrospective studies attempting to find a relationship between a risk factor and a disease of interest (e.g. effect of smoking on lung cancer rates) using pre-existing sources such as health databases.

In this context, the nQuery April 2018 release adds an additional four tables which should add additional flexibility when planning a case-control study using nQuery. These are:

    1. Test for Binary Covariate in Logistic Regression
    2. Case-Control Test with M Controls per Case (McNemar Extension)
    3. Conditional Logistic Regression with Binary Risk Factor
    4. Conditional Logistic Regression with Continuous Risk Factor

These tables complement our pre-existing nQuery tables for chi-squared tests, exact tests, correlated proportions and logistic regression and add options for conditional logistic regression to allow for greater flexibility when exploring sample estimates for case-control studies.

Vaccine Efficacy

Vaccine efficacy studies face significant challenges compared to other clinical trials. These include having a much larger scale due often being nation or region wide campaigns, dealing with rare diseases or conditions and the challenges of doing work in the field rather than in a fully controlled setting. For reasons such as these, vaccine efficacy designs and statistical methods have developed their own approaches and terminology to help the relevant researchers and public or private bodies of interest.

For vaccine efficacy, the nQuery April 2018 release adds two additional tables tailored for finding the sample size for the precision of an estimate of the vaccine efficacy. These are:

    1. Confidence Interval for Vaccine Efficacy in a Cohort Study
    2. Confidence Interval for Vaccine Efficacy in a Case-Control Study

In conjunction with our wide range of pre-existing tables for binomial proportions and survival rates, these tables will give researchers in vaccine research more tailored options for their study.

Cluster Randomized Trials

Cluster randomized trials are studies where the unit of randomization is a cluster or group rather than the individual subject. This is a common design when there are natural blocks or clusters such as schools or hospitals. By assigning the same treatment to all subjects within a cluster, the administrative and financial cost of field trials can be reduced significantly. For this reason and others, this design type is very commonly seen in public health policy studies.

For cluster randomized trials, the nQuery April 2018 release includes four additional tables which expand upon our pre-existing options for cluster randomized trials. These are:

    1. Cluster Randomized Trial for Inequality Test of Two Means (unequal sample sizes)
    2. Cluster Randomized Trial for Non-inferiority Test of Two Means
    3. Cluster Randomized Trial for Equivalence Test of Two Means
    4. Cluster Randomized Trial for Superiority by a Margin of Two Means

These options expand upon our pre-existing tables for cluster randomized trials comparing means, proportions, incidence rates and survival curves and for alternative cluster randomized trials such as the matched-pair design.

Mendelian Randomization

Mendelian Randomization is form of randomization which takes advantage of the growing availability and understanding of genetic information to make causal claims about potential treatments without using the common fully randomized approach. By using well characterised relationships between genes and phenotypes with a known secondary effect on an outcome of interest, mendelian randomization offers the opportunity to use genetic information as a instrumental variable to find the causal relationship between a risk factor of interest and a disease outcome.

For studies which use Mendelian Randomization, the nQuery April 2018 release provides two new tables. These are:

    1. Mendelian Randomization for Continuous Outcome
    2. Mendelian Randomization for Binary Outcome

These provide the first tables in nQuery which account for this novel design and innovative approaches such as this will be of active interest in the area

2. Non-inferiority and Equivalence Testing

Non-inferiority and equivalence testing are used to statistically evaluate how similar a proposed treatment is to a pre-existing standard treatment. This is a very common objective in areas such as generics and medical devices. This is particularly important if using a placebo group would be required otherwise.

As non-inferiority and equivalence testing will typically involve evaluation against a well-defined treatment (e.g. RLD), there is a lower incidence of the large parallel studies typically seen in Phase III clinical trials. One-sample, paired samples or cross-over designs are common as these will generally require a lower cost and sample size.

In the April release, we will be adding an additional 20 (22 in the CRT Means reference above are included) tables for non-inferiority and equivalence testing. These are focused on expanding the options available for the non-inferiority and equivalence testing of continuous data, binary data and incidence rates. The focus areas are as follows:

  1. Continuous Outcome Studies
  2. Binary Outcome Studies
  3. Incidence Rate Outcome Studies

These areas and the tables in each category are explored below.

Continuous Outcome Studies

In the context of non-inferiority and equivalence testing, the comparison of continuous outcomes using means is the most common situation to encounter. A wide range of design types and statistical methods are available for comparing this type of data depending on the assumptions and constraints relevant to proposed study. Common design types in this context would be one-sample, paired, cross-over and parallel studies. The most common statistical methods are based on assuming either that the data is normally distributed and comparing the difference in means (Additive model) or that the data is log-normally distributed and analysing the ratio of (geometric) means (Multiplicative model).

For non-inferiority and equivalence testing of continuous data, the nQuery April 2018 release adds an additional 12 tables. These are as follows:

    1. Non-inferiority for One Normal Mean
    2. Non-inferiority for One Log-Normal Mean  
    3. Non-inferiority for Paired Means Ratio
    4. Equivalence for One Mean
    5. Equivalence for Paired Means
    6. Equivalence for One Log-Normal Mean
    7. Equivalence for Paired Means Ratio
    8. Non-inferiority for cross-over design
    9. Non-inferiority for Two-sample Mean Ratio on Log-scale
    10. Non-inferiority for Cross-over Mean Ratio on Log-scale
    11. Non-inferiority for Two-sample Mean Ratio on Original Scale
    12. Non-inferiority for Cross-over ratio on Original Scale

These tables expand upon the large number of pre-existing tables for non-inferiority and equivalence testing for means to give the largest number of options available to find the sample size.

Binary Outcome Studies

In the context of non-inferiority and equivalence testing, the comparison of binary is less common but has grown in more recently as additional statistical methods have become popularised. Common design types in this context would be one-sample, paired and parallel studies. In the context of binary data, one of the most noticeable aspects is wide variety of options available ranging from relatively simple normal approximation tests to more complex exact methods and sample size methods have followed this trend in regard to binary data analyses generally.

For non-inferiority and equivalence testing of binary data, the nQuery April 2018 release adds an additional 2 tables. These are as follows:

    1. Non-inferiority for a Single Binary Proportion
    2. Equivalence Test for Two Independent Proportions

Note that both these tables integrate more exact binomial enumeration methods as an option in addition to the typical normal approximation methods. They also include options for a wide range of proposed statistics with the main categories being chi-squared tests (including option for continuity correction), Z and t-test approximations and several likelihood score statistics (Miettinen and Nurminen, Gart & Nam, Farrington and Manning). These tables expand upon the large number pre-existing tables for the non-inferiority and equivalence testing of binary proportions.

Incidence Rates Studies

In the context of non-inferiority and equivalence testing, the comparison of incidence rates is a relatively uncommon scenario. Incidence rates are where the outcome of interest is the number of events which occur on average in a given time period (i.e. the event rate). The wider availability of software to analyse incidence rates directly rather than relying on normal approximations has seen a growth of interest in methods such as Poisson and Negative Binomial regression. This has naturally extended to the case of non-inferiority and equivalence testing of incidence rate data. The time-dependent nature of incidence rates means that models can integrate greater flexibility for time dependencies and this is reflected in the rapidly growing literature for sample size in the area.

For non-inferiority and equivalence testing of incidence rates data, the nQuery April 2018 release adds 6 tables. These are as follows:

    1. Non-inferiority for Two Rates using Poisson Model
    2. Equivalence for Two Rates using Poisson Model
    3. Equivalence for Negative Binomial Model (Equal Follow-up, Dispersion)
    4. Non-inferiority for Negative Binomial Model (Equal Follow-up, Dispersion)
    5. Equivalence for Negative Binomial Model (Unequal Follow-up, Dispersion)
    6. Non-inferiority for Negative Binomial (Unequal Follow-up, Dispersion)

These tables expand upon the pre-existing options for analysing incidence rate data in the context of inequality testing. These methods represent the latest research and in the last two tables can integrate the effects of dispersion and unequal follow-up on the sample size estimate.

3. Correlation and Diagnostic Testing (ROC) Methods

Correlation, agreement and ROC methods are interested in characterising the strength of the relationship between a predictor (e.g. presence of treatment) and outcome (disease progression) of interest. These measures are often used in conjunction with models and statistical testing to characterise the nature of the relationship of interest. These methods are of interest when attempting to communicate the strength of a model or a relationship.

These types of measures are seen throughout statistical practise but are particularly prominent in areas such as diagnostic testing, the social sciences and biomarker studies.

In the nQuery April release, we will be adding an additional 9 additional tables in this area which fall in the following main categories:

  1. Correlation and Agreement Measures
  2. Diagnostic Screening Measures

These are summarised below.

Correlation and Agreement Measures

Correlation measures are used to characterise the strength of relationship between continuous and/or ordinal outcomes and measures such as Pearson’s correlation are ubiquitous in statistical practise. Agreement measures are used to analyse the strength of the ability of more than one rater (e.g. tester or test) to agree and correctly diagnose the condition of one or more entities (e.g. subject disease status). Both of these are common outcomes of interest in a wide variety of settings.

Due to the ubiquity of these methods, a wide range of measures have been proposed to adjust for scenarios which diverge from the most common correlation and agreement measures (e.g. Pearson correlation and Cohen’s Kappa). Common complications adjusted for are the presence of ordinal instead of continuous variables or divergences from common distributional assumptions (e.g. Normal)

In the nQuery April release, we are adding four additional options in this area. These are as follows:

    1. Confidence Interval for Kendall-Tau Correlation
    2. Confidence Interval for Spearman Correlation
    3. Test for One Intra-cluster Correlation
    4. Test for One Cronbach Alpha

These add to the options available for other common correlation and agreement measures such as the Pearson Correlation, Lin’s Concordance Coefficient and Cohen’s Kappa Coefficient.

Diagnostic Screening Measures

Diagnostic screening measures are very common in clinical research. These measures are used to assess the performance of a diagnostic test to accurately predict a condition of interest in the population(s) of interest. Areas where this type of analysis have become particularly popular are biomarker studies, machine learning and predictive genetic tests.

Commonly, this strength is characterised by the Area under the Curve (AUC) of the Receiving Operating Curve (ROC) which provides a useful summary measure of screening performance over all potential cut-off points for a screening measure. However, a large number of other statistics may be of interest at specific cut-offs such as sensitivity (a.k.a. recall), specificity and positive predictive value (PPV, a.k.a. precision), among other.

In the nQuery April release, we are adding 5 new tables in this area. These are as follows:

  1. Confidence Interval for Area under the Curve (AUC)
  2. Confidence Interval for One Sensitivity
  3. Confidence Interval for One Specificity
  4. Simultaneous Interval for One Sensitivity and One Specificity
  5. Test for Paired Sensitivity

These add to the pre-existing tables already present in nQuery for the testing of AUC values under varying design types and sensitivity and specificity.

4. Miscellaneous Tables

In the April release of the nQuery, 11 tables do not fit into the above categorisations. These cover areas such as the testing of log-normal means, the testing of variances and standard deviations and non-parametric tests. These tables are as follows:

  1. Confidence Interval for One Variance
  2. Confidence Interval for One Variance using Tolerance Probability
  3. Confidence Interval for One Variance using Relative Error
  4. Confidence Interval for One Standard Deviation
  5. Confidence Interval for One Standard Deviation using Tolerance Probability
  6. Confidence Interval for One Standard Deviation using Relative Error
  7. Confidence Interval for Ratio of Two Variances
  8. Confidence Interval for Ratio of Two Variances using Relative Error
  9. One Sample t-test for Log-Normal data
  10. Paired t-test for Mean Ratio (log scale)
  11. One Sample/Paired Sample Wilcoxon Sign-Rank Test

These options expand upon nQuery’s pre-existing options in these areas. 

Bayesian Tables

Below are the Bayesian tables that have been added nQuery in this release:

  • Mixed Bayes/Likelihood for Two Means Equal Variance 
  • Unequal Variance Means - Known Precision 
  • Unequal Variance Means - Mixed 
  • Difference in Two Proportions - Unknown Precision 
  • Difference in Two Proportions - Mixed 
  • Bayesian Reassessment Methods (Phase 1) 
  • Superiority Assurance for one Proportion 
  • Bayesian Assurance for One Group Test of Normal Mean 
  • Bayesian Assurance for One Group Test - Uniform Prior 
  • Bayesian Assurance for Two Proportions - Beta prior for P1 
  • Bayesian Assurance for Two Proportions - Beta prior for P2 
  • Bayesian Assurance for Two Proportions - Beta prior for P1 and P2 
  • Bayesian Assurance for Mean Difference with Common Variance - Prior for δ and σ² (Simulation) 
  • Bayesian Assurance for Mean Difference with Uncommon Variance - Prior for δ and each σ² (Simulation) 
  • Bayesian Assurance for Two Group Non-Inferiority Test of Normal Means - Uniform Prior 
  • Bayesian Assurance for Two Group Non-Inferiority Test of Normal Means - Uniform Prior 
  • Bayesian Assurance for Two Group Test of Survival Curves - Normal Prior for log(HR) 
  • Bayesian Assurance for Two Group Test of Survival Curves - Lognormal Prior for HR 
  • Bayesian Assurance for Two Group Test of Survival Curves - Inverse Gamma Prior for λt 
  • Bayesian Assurance for Two Group Test of Survival Curves - Inverse Gamma Prior for λc

 

Have a question or want to find out more?