1. Trang chủ
  2. » Tất cả

Design of pilot studies to inform the construction of composite outcome measures

6 2 0
Tài liệu đã được kiểm tra trùng lặp

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 6
Dung lượng 461,64 KB

Nội dung

Design of pilot studies to inform the construction of composite outcome measures Q1 Q12 Q2 Q4 Q5 Q3 Alzheimer’s & Dementia Translational Research & Clinical Interventions (2017) 1 6 1 2 3 4 5 6 7[.]

1 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 Alzheimer’s & Dementia: Translational Research & Clinical Interventions - (2017) 1-6 Featured Article Q1 Design of pilot studies to inform the construction of composite outcome measures Q12 Steven D Edland*, M Colin Ard, Weiwei Li, Lingjing Jiang Q2 - Abstract Background: Composite scales have recently been proposed as outcome measures for clinical trials For example, the Prodromal Alzheimer’s Cognitive Composite (PACC) is the sum of z-score normed component measures assessing episodic memory, timed executive function, and global cognition Alternative methods of calculating composite total scores using the weighted sum of the component measures that maximize signal-to-noise ratio of the resulting composite score have been proposed Optimal weights can be estimated from pilot data, but it is an open question as how large a pilot trial is required to calculate reliably optimal weights Methods: We describe the calculation of optimal weights and use large-scale computer simulations to investigate the question as how large a pilot study sample is required to inform the calculation of optimal weights The simulations are informed by the pattern of decline observed in cognitively normal subjects enrolled in the Alzheimer’s Disease Cooperative Study Prevention Instrument cohort study, restricting to n 75 subjects aged 75 years and older with an ApoE E4 risk allele and therefore likely to have an underlying Alzheimer neurodegenerative process Results: In the context of secondary prevention trials in Alzheimer’s disease and using the components of the PACC, we found that pilot studies as small as 100 are sufficient to meaningfully inform weighting parameters Regardless of the pilot study sample size used to inform weights, the optimally weighted PACC consistently outperformed the standard PACC in terms of statistical power to detect treatment effects in a clinical trial Pilot studies of size 300 produced weights that achieved nearoptimal statistical power and reduced required sample size relative to the standard PACC by more than half Discussion: These simulations suggest that modestly sized pilot studies, comparable to that of a phase clinical trial, are sufficient to inform the construction of composite outcome measures Although these findings apply only to the PACC in the context of prodromal Alzheimer’s disease, the observation that weights only have to approximate the optimal weights to achieve near-optimal performance should generalize Performing a pilot study or phase trial to inform the weighting of proposed composite outcome measures is highly cost-effective The net effect of more efficient outcome measures is that smaller trials will be required to test novel treatments Alternatively, second generation trials can use prior clinical trial data to inform weighting, so that greater efficiency can be achieved as we move forward Ó 2017 The Authors Published by Elsevier Inc on behalf of the Alzheimer’s Association This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/ 4.0/) Keywords: Alzheimer’s disease; Phase clinical trial; Phase clinical trial; Composite endpoint; Cognitive decline; Secondary prevention; Power; Sample size Q4 Q5 Q3 *Corresponding author Tel.: -; Fax: - E-mail address: sedland@ucsd.edu http://dx.doi.org/10.1016/j.trci.2016.12.004 2352-8737/ Ó 2017 The Authors Published by Elsevier Inc on behalf of the Alzheimer’s Association This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/) FLA 5.4.0 DTD  TRCI66_proof  23 January 2017  4:06 pm  ce 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 S.D Edland et al / Alzheimer’s & Dementia: Translational Research & Clinical Interventions - (2017) 1-6 Introduction 2.2 Prodromal AD PI cohort Composite endpoints have received increasing attention as potential outcome measures for clinical trials in Alzheimer’s disease (AD) Composites can be defined as the sum of items taken from component instruments of a cognitive battery [1] Or, more simply, composites can be defined as the sum of established cognitive instruments One such composite is the Preclinical Alzheimer’s Cognitive Composite or Prodromal Alzheimer’s Cognitive Composite (PACC) [2] The PACC is constructed from component measures assessing episodic memory, timed executive function, and global cognition and is the primary outcome measure for a major ongoing trial [3] We have described how the performance of a composite endpoint depends on the weighting used and how optimal weights can be derived if the multivariate distribution of change scores on component measures is known [4] The multivariate distribution of change scores of the component measures is typically not known but can be estimated if pilot data are available, for example, from a prior trial or from a prior representative registry study using the component instruments An important consideration is whether prior data are sufficient to inform weighting parameters for a composite outcome measure and, in particular, how large sample size would be required to meaningfully inform calculation of weights In this article, we use data from a completed registry trial to describe calculation of optimal weights and to investigate the question of what size pilot study is sufficient to inform calculation of optimal weights Pilot study longitudinal data for the PACC to inform instrument behavior and clinical trial design are not available [2] However, roughly comparable component instruments are available from the PI protocol conducted by the Alzheimer’s Disease Cooperative Study [9] The PI protocol performed annual neuropsychometric and functional assessments of 644 cognitively normal older persons (age 75 years and older) Although there was no randomization to treatment, the PI enrollment and assessment procedures mimicked that of a clinical trial, with primary purpose to assess the utility of the components of the assessment battery as potential endpoints for an Alzheimer prevention trial, and these data were used in the initial description of the PACC [2] The PACC components that were not assessed in the PI study were the MMSE and the Logical Memory test Comparable domain-specific instruments used in their stead were the modified MMSE [10] substituting for the MMSE, and the New York University Paragraph delayed recall test [11] substituting for the Logical Memory test When the distinction is relevant, we call the resulting composite the PI-PACC to distinguish it from the PACC constructed from the MMSE, FCSRT, Digit Symbol, and Logical Memory test Donohue et al [2] restricted their analysis to subjects with an ApoE E4 risk allele, and we follow suit Subjects aged 75 years and older with this genetic risk profile have with high likelihood an underlying Alzheimer neurodegenerative process, and hence these subjects are an approximate representation of clinically normal, AD biomarker positive subjects that are the target of contemporary secondary prevention trials [2] We call this subset of the PI cohort the PI Prodromal AD cohort Baseline through month 36 data are available for 75 of these subjects (mean age at baseline 78.5 years [standard deviation 2.9 years], 59% female), and these longitudinal data are used to inform the simulations reported here Q6 Methods In overview, we use simulations informed by data from a completed registry trial, the Alzheimer’s Disease Cooperative Study Prevention Instrument (PI) trial, to demonstrate optimal weighting and investigate the question as how large a pilot study is required to determine weights that improve the performance of the PACC In the text that follows we briefly describe the PACC and the PI trial and then formally characterize optimal weights and computer simulation procedures 2.1 Preclinical Alzheimer’s Cognitive Composite We use the PACC [2] to demonstrate the influence of weighting on characteristics of the composite scale The PACC is a weighted sum of well recognized and validated component instruments, the Mini-Mental Status Examination (MMSE) assessing global cognition function [5], the Free and Cued Selective Reminding task (FCSRT) assessing episodic memory [6], and the WAIS-R Digit Symbol task (Digit Symbol), a timed test of processing speed and memory function [7], and the WMS-R Logical Memory story delayed recall task (Logical Memory) [8] 2.3 Optimal weights We assume the primary analysis is mixed model repeated measure (MMRM) comparing change first to last in treatment versus change first to last in control [2] To simplify presentation, we assume complete data for all simulations Including missing values in simulations would reduce power given a total sample size, but would not appreciably impact the relative efficiency of trial designs and endpoints, which is the focus of this article We further make the usual assumption that an effective treatment would shift the mean change but not affect the variability of change (constant variance of change in treatment and control arms) Under these assumptions, optimal weights for constructing a composite endpoint are a simple function of two sets of parameters, the expected change and the covariance of change of the component measures [4] Given the vector m of expected change scores of FLA 5.4.0 DTD  TRCI66_proof  23 January 2017  4:06 pm  ce 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 S.D Edland et al / Alzheimer’s & Dementia: Translational Research & Clinical Interventions - (2017) 1-6 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 component measures and covariance matrix S of change scores, weights that maximize the signal-to-noise ratio of the composite (and therefore statistical power of clinical trials using the composite) are Optimal weights5c  21 X m0 The c is an arbitrary scalar constant—any nonzero value of c will produce equally optimal weights A useful convention is to set c so that the weights sum in absolute value to The distribution of component change scores is typically unknown, but can be estimated, for example, from prior clinical trials that included the component measures or from registry trials specifically designed to investigate properties of potential outcome measures 2.4 Computer simulations We used computer simulations to investigate the properties of weights estimated from pilot registry study data performed before a formal randomized clinical trial We simulated 40,000 pilot study–clinical trial dyads, using pilot study sample sizes of 100 to 300 persons, and clinical trial sample sizes of 100 to 1600 subjects per arm The pilot study component of the dyad could be a prior nonintervention study registry trial or the placebo arm of a previously completed trial with comparable inclusion criteria Simulations assumed multivariate normality of component change scores with the mean and covariance structure observed in the PI prodromal AD cohort A 25% shift in mean change was added to the treatment arm to simulate data from a trial with an effective treatment For each dyad, we calculated the standard PACC [2] and optimal PACC from the simulated component scores, with weights for the optimal PACC estimated from the simulated pilot study and weights for the standard PACC calculated from baseline data of the clinical trial, reflecting how these endpoints would be calculated in practice An MMRM model testing the hypothesis that the mean 3-year decline was different in the treatment and control arms was fit to the respective composite measures Statistical power of the PACC and optimal PACC was calculated as the percentage of simulations for which a statistically significant difference was observed at a 0.05 significance level All data simulations and statistical analyses were performed using the R statistical programming language, with model fitting performed using the nlme package [12] Results Baseline characteristics and 3-year change observed in the PI prodromal AD cohort are summarized in Table The ratio of mean change to the standard deviation of change (the mean to standard deviation ratio (MSDR), aka the signal-to-noise) for each component instrument of the PIPACC is also summarized in Table Instruments with high MSDR are more sensitive to change and are more Table Mean (standard deviation) of component item scores at baseline and year visit, mean to standard deviation ratio of the component scores, and component weights used to construct the weighted sum composite scores Q11 FCSRT mMMSE Mean (SD) Baseline 47.88 (0.47) 95.97 (2.84) Year 46.63 (4.18) 91.88 (15.44) Change 21.27 (4.11) 24.09 (15.02) Mean to standard deviation ratio (MSDR) 0.31 0.27 Item weights PACC 0.72 0.12 Optimal 0.25 0.06 PACC NYU Paragraph Digit Symbol 7.39 (2.49) 41.29 (12.04) 5.69 (3.25) 38.64 (11.10) 21.69 (3.15) 22.65 (9.33) 0.54 0.28 0.14 0.65 0.03 0.04 Abbreviations: Digit Symbol, WAIS-R Digit Symbol task; FCSRT, Free and Cued Selective Reminding task; mMMSE, modified Mini-Mental Status Examination; NYU Paragraph, New York University Paragraph delayed recall test; PACC, Prodromal Alzheimer’s Cognitive Composite powerful endpoints for clinical trials [13] Among the components of the PI-PACC, the paragraph recall test has the greatest MSDR (Table 1) PACC and optimal PACC weights, standardized to sum in absolute value to 1, are summarized in the bottom two rows of the table Both composites give relatively lower weight to the modified MMSE and the Digit Symbol test A primary difference between the PACC and the optimal PACC is a greater weight to the FCSRT by the PACC and greater weight to the paragraph recall test by the optimal PACC (Table 1) Power to detect treatment effects as a function of sample size is summarized in Fig As a reference point, the theoretical maximum power achievable if the true covariance of component change scores was known is also plotted in the figure A three year clinical trial using weights informed by a three year pilot study of size 300 subjects achieves near-optimal power, with obtained power deviating from optimal power by less than 1% in the critical region of the power curve with power of 80% and higher (Fig 1) Power of pilot study–clinical trial dyads decreases if smaller pilot studies are used to inform weights, but only modestly Power obtained was within 1.2 percentage points of the theoretical maximum achievable power when pilot sample size is 200 subjects, and within 2.4 percentage points of the theoretical maximum when pilot sample size is 100 subjects Nonetheless, it is important to note that there is some loss of power, and a modest inflation of estimated sample size would be prudent if the pilot study data used to estimate optimal weights were also used to estimate sample size for a future clinical trial, on the order of 3% (pilot N 300) to 10% (pilot N 100) for the scenarios reported here Discussion The optimal weighting formula as implemented here assumes a treatment effect that shifts the mean change from FLA 5.4.0 DTD  TRCI66_proof  23 January 2017  4:06 pm  ce 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 S.D Edland et al / Alzheimer’s & Dementia: Translational Research & Clinical Interventions - (2017) 1-6 web 4C=FPO Fig Power to detect a 25% slowing in cognitive decline as a function of sample size per arm and outcome measure used For optimal composites, power is also a function of the size of pilot study used to inform optimal weights (Clinical trial with equal allocation to arm, two-sided hypothesis testing, and type I error rate a 0.05.) baseline to last visit but assumes a constant variance of change in treatment and control This is the usual assumption used in power calculations (e.g., [14–16]) Alternative treatment effects may be plausible For example, instead of assuming a percentage shift in mean location, we could assume a percentage decrease in rate of decline in all subjects, so that the variance of change scores would be decreased; for example, under this assumed treatment effect, a 25% shift in mean would be accompanied by a 0.75 squared 0.5625 reduction in variance in the treatment arm and accompanying increase in power We prefer the more conservation mean shift assumption for several reasons First, given the general uncertainty in parameter estimates used to inform power calculations, conservative assumptions provide some margin of error in sample size calculations Second, an alternative scenario that is plausible and even likely is that response to treatment will be variable subject to subject within the Q7 treatment arm In other words, the variance of change in the treatment arm will be the sum of variance in rate of decline plus the variance of response to treatment Under this plausible and likely scenario, the total variance will be larger than the variance in the placebo arm, meaning the percent shift hypothesis would be highly anticonservative and result in underestimates of required sample size and underpowered trials The MMRM analysis plan typically includes baseline level of the outcome variable as a covariate [17] When this term was added to the MMRM model fits to each simulated data set the power increased slightly, less than one percentage point for most of the range of sample sizes simulated for both the PACC and optimal PACC Hence, the observations regarding relative efficiency of the PACC and optimally weighted PACC are unchanged by inclusion on the baseline covariate term Conclusions We have investigated the magnitude of sample size required to estimate weights that optimize the performance of a cognitive composite endpoint and found that pilot studies of as small as 100 to 300 subjects are sufficient to inform composite weighting and achieve near-optimally powerful composite endpoints In other words, trials of the size of a typical phase trial are sufficient to estimate weighting parameters for defining an optimal weighted composite endpoint This finding is similar to previously reported findings in Ard et al [4] for a two-component composite instrument Ard et al used computer simulations to document near-optimal composite performance with weights estimated from pilot studies as small as 100 subjects for the two-component composite The current article replicates and meaningfully extends those results by (1) assessing the prospective performance of a composite currently in use in a major Alzheimer clinical trial, and (2) using data from a completed registry trial to determine realistic simulation parameters A related concern is the representativeness of the pilot study used to train weights—weights optimal in one clinical trial target population may not be optimal in a different population Raghavan et al [18] addressed this latter question and found substantial robustness of cognitive composites to the training data set They found that weights estimated FLA 5.4.0 DTD  TRCI66_proof  23 January 2017  4:06 pm  ce 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 S.D Edland et al / Alzheimer’s & Dementia: Translational Research & Clinical Interventions - (2017) 1-6 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 from longitudinal data obtained relatively earlier or later in the prodromal AD spectrum were comparable and consistently improved trial efficiency regardless of the prodromal AD stage recruited to the ultimate clinical trial As we observed in our investigation of pilot study sample size, even approximate information about the distribution of change scores was sufficient to inform the calculation of optimal weights and improve the efficiency of composite scales On the basis of these observations we speculate that, within the context of prodromal AD trials, weights optimal in one sample will be optimal or near-optimal for future trials with similar design and inclusion criteria, and that an optimal PACC defined using optimal weights estimated from a single registry trial (or completed clinical trial) would be an appropriate endpoint for future trials with similar design and inclusion criteria In contrast, the PACC as originally described is redefined on a trial-by-trial basis—it is the sum of z-score normed component instruments, with z-score normative values estimated from baseline visit data of the respective clinical trial [2] In other words, the PACC is measured on a different scale and has a different interpretation for each clinical trial A single established optimally weighted PACC would have the dual advantages of improved statistical power and of being comparable study to study, so that future pooled meta-analyses would be possible The clear tradeoff and downside of optimal endpoints is that a pilot study is required, a real cost in terms of both time and resources For the “PI-PACC” assuming the distribution of change scores observed in the PI Prodromal AD cohort, the optimal PACC is relatively cost efficient even considering the time and cost of a pilot registry trial— assuming this distribution of change scores, a trial with 80% power to detect a 25% slowing of decline using the optimal PACC would require 600 subjects per arm (1200 subjects total), whereas a trial powered to detect the same percentage slowing in the PI-PACC would require more than 2500 subjects Given the critical importance of statistical power in clinical trials, any method of improving power and trial efficiency should be seriously considered More power means there is less likelihood of false negative trials missing effective treatments or conversely more power means that we can perform smaller trials with equivalent power, so that we may perform more clinical trials and test more treatments with the limited study subject pool available for prodromal AD studies In the long run, more efficient trials will shorten the time till effective treatments are identified and we begin to make meaningful progress against the epidemic of AD Acknowledgments This work is supported by the NIH NIA R03 AG047580, NIH NIA P50 AG005131, NIH NIA R01 AG049810, and Q8 NIH UL1 TR001442 RESEARCH IN CONTEXT Systematic review: Composite scales, typically defined as the weighted sum of established component assessment scales, have recently been proposed as outcome measures for clinical trials Composite scales can be severely inefficient endpoints if suboptimal weights are used to construct the composite Optimal weights can be estimated from pilot data, but it is an open question as how large a pilot trial is required to calculate reliably optimal weights Interpretation: We demonstrated with large-scale computer simulations that pilot trials of size 100 to 300 subjects, the size of typical phase clinical trials, are sufficient to determine optimal weights that maximize the sensitivity and statistical power of composite outcomes to detect treatment effects Future directions: The potential utility of optimally weighted composites has been well demonstrated A practical demonstration of utility using data from completed trials would further validate this approach to clinical trial endpoint development References [1] Langbaum JB, Hendrix SB, Ayutyanont N, Chen K, Fleisher AS, Shah RC, et al An empirically derived composite cognitive test score with improved power to track and evaluate treatments for preclinical Alzheimer’s disease Alzheimers Dement 2014;10:666–74 [2] Donohue MC, Sperling RA, Salmon DP, Rentz DM, Raman R, Thomas RG, et al The preclinical Alzheimer cognitive composite: measuring amyloid-related decline JAMA Neurol 2014;71:961–70 [3] NCT02760602 A Study of solanezumab (LY2062430) in participants with prodromal Alzheimer’s disease (expeditionPRO) Available at: ClinicalTrials.gov Q9 [4] Ard MC, Raghavan N, Edland SD Optimal composite scores for longitudinal clinical trials under the linear mixed effects model Pharm Stat 2015;14:418–26 [5] Folstein MF, Folstein SE, McHugh PR “Mini-mental state” A practical method for grading the cognitive state of patients for the clinician J Psychiatr Res 1975;12:189–98 [6] Grober E, Buschke H, Crystal H, Bang S, Dresner R Screening for dementia by memory testing Neurology 1988;38:900–3 [7] Wechsler D Wechsler Adult Intelligence Scale-Revised New York, NY: Psychological Corp; 1981 [8] Wechsler D WMS-R: Wechsler Memory Scale–Revised: manual San Antonio, TX: Psychological Corp; 1987 [9] Ferris SH, Aisen PS, Cummings J, Galasko D, Salmon DP, Schneider L, et al ADCS Prevention Instrument Project: overview and initial results Alzheimer Dis Assoc Disord 2006;20(Suppl 3):S109–23 [10] Teng EL, Chui HC The modified Mini-Mental State (3MS) examination J Clin Psychiatry 1987;48:314–8 [11] Kluger A, Ferris SH, Golomb J, Mittelman MS, Reisberg B Neuropsychological prediction of decline to dementia in nondemented elderly J Geriatr Psychiatry Neurol 1999;12:168–79 FLA 5.4.0 DTD  TRCI66_proof  23 January 2017  4:06 pm  ce 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 S.D Edland et al / Alzheimer’s & Dementia: Translational Research & Clinical Interventions - (2017) 1-6 [12] Pinheiro J, Bates D Mixed-effects models in S and S-PLUS New York, NY: Springer; 2000 [13] Edland S, Ard MC, Sridhar J, Cobia D, Martersteck A, Mesulam MM, et al Proof of concept demonstration of optimal composite MRI endpoints for clinical trials Alzheimers Dement (N Y) 2016;2:177–81 [14] Lu K, Luo X, Chen PY Sample size estimation for repeated measures analysis in randomized clinical trials with missing data Int J Biostat 2008;4:Article [15] Beckett LA, Harvey DJ, Gamst A, Donohue M, Kornak J, Zhang H, et al The Alzheimer’s Disease Neuroimaging Initiative: annual change in biomarkers and clinical outcomes Alzheimers Dement 2010;6:257–64 [16] Ard MC, Edland SD Power calculations for clinical trials in Alzheimer’s disease J Alzheimers Dis 2011;26 Suppl 3:369–77 [17] Mallinckrodt CH, Lane PW, Schnell D, Peng Y, Mancuso JP Recommendations for the primary analysis of continuous endpoints in longitudinal clinical trials Drug Inf J 2008;42:303–19 [18] Raghavan N, Wathen K Optimal composite cognitive endpoints for pre-symptomatic Alzheimer’s disease: considerations in bridging across studies Alzheimers Dement (N Y) 2016 Q10 FLA 5.4.0 DTD  TRCI66_proof  23 January 2017  4:06 pm  ce 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 ... investigation of pilot study sample size, even approximate information about the distribution of change scores was sufficient to inform the calculation of optimal weights and improve the efficiency of composite. .. S of change scores, weights that maximize the signal -to- noise ratio of the composite (and therefore statistical power of clinical trials using the composite) are Optimal weights5c  21 X m0 The. .. response to treatment will be variable subject to subject within the Q7 treatment arm In other words, the variance of change in the treatment arm will be the sum of variance in rate of decline plus the

Ngày đăng: 24/11/2022, 17:41

TÀI LIỆU CÙNG NGƯỜI DÙNG

TÀI LIỆU LIÊN QUAN

w