Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2023 Nov 17;13:20137. doi: 10.1038/s41598-023-46935-x

Validity and reliability of “My Jump app” to assess vertical jump performance: a meta-analytic review

Cebrail Gençoğlu 1, Süleyman Ulupınar 1,, Serhat Özbay 1, Murat Turan 1, Buğra Çağatay Savaş 1, Selim Asan 1, İzzet İnce 2
PMCID: PMC10656545  PMID: 37978338

Abstract

This systematic review and meta-analysis aims to investigate the validity and reliability of the My Jump smartphone application in measuring vertical jump height, specifically using flight-time-based measures. To identify potential studies for inclusion, a comprehensive search strategy was employed in PubMed, Web of Science, Scopus, and EBSCO host databases. Validity was assessed in two ways: (1) mean and standard deviations of My Jump measurements were compared to criterion methods to assess the agreement of raw scores; (2) correlation coefficients evaluated the within-group consistency of rankings between My Jump and criterion methods. Reliability was assessed using intraclass correlation coefficients (ICC). Heterogeneity was evaluated via Cochrane’s Q statistic, its p-value, I2 value, and tau2 value. Publication bias was explored through funnel plot symmetry and confirmed with extended Egger’s test. Following the search, 21 studies met the inclusion criteria. Results showed no significant difference in raw scores between My Jump and criterion methods, indicating high agreement. High correlation was also found for within-group rankings, suggesting consistency. The My Jump application demonstrated nearly perfect reliability scores. The My Jump application appears to be a valid and reliable tool for sports scientists and strength and conditioning practitioners, offering a cost-effective and accessible means for accurately assessing vertical jump performance in various settings. However, it should be noted that these results are specific to flight-time-based measures, and further research is needed to validate these findings against gold-standard take-off velocity methods.

Subject terms: Computational biology and bioinformatics, Public health

Introduction

The emergence of novel devices (e.g. My Jump smartphone application, GymAware, PUSH Band) measuring athletic performance is quickly gaining momentum as these devices increase in popularity as potential alternatives to expensive laboratory equipment1,2. Their main advantage is that these novel devices are easily portable (especially in the case of software applications that are integrated into tablets and smartphones); they have the potential to offer an excellent solution to the problems of many laboratory-based measurement methods such as the high cost of laboratory equipment, the difficulties of transporting the devices to the field, or people to the laboratory, and the need for periodic maintenance and complex interfaces24. However, to take advantage of all these facilitating aspects, it is necessary to ensure that the measurements made with these methods give valid and reliable outputs.

Validity of the measurement of athletic characteristics requires that the movement pattern is close to the mode and profile observed during competition, that there is an output that represents specific proficiency, that the measurements are associated with a proven gold standard or a criterion measurement, that the evaluations can predict the actual competition performance, and that the results can distinguish between the successful and unsuccessful athletes57. Similarly, reliability of the measurement of athletic characteristics requires that the measurements can be replicated, that the within-group ranking and agreement between the raw scores can be maintained from test to retest, and that successive measurements give the same output when performed in such a short time that no actual performance improvement is possible810. Hence, a novel device or method designed to measure a component of physical fitness is expected to meet all these requirements regarding validity and reliability.

The need to easily measure, evaluate, and monitor athletic performance has inspired sports engineering professionals to design or produce novel smartphone applications1,2. Mobile applications have become very popular due to many beneficial features such as being portable, inexpensive, and easily accessible. For example, the My Jump smartphone application based on high-speed video technology provides an extremely practical way of measuring vertical jump performance and has been used in numerous scientific studies since its introduction to the literature1113. However, although there are many advantages provided by My Jump, there are several potential sources of bias that may impact the measurement of jump height. These include the need to manually identify the take-off and landing frames of the video, the potential for knee and ankle flexion (dorsiflexion) just before landing to artificially extend the measured flight time, the possibility of one foot making contact with the ground before the other, which could be misleadingly used as the frame to determine the end of the flight time, and the consideration that a slight contact by one foot may not accurately represent the displacement of the center of mass1416. Therefore, sports scientists, conditioners, coaches, and athletes need to be confident that this smartphone application can present consistent and repeatable outputs.

In the realm of jump height calculations, two distinct definitions based on the Centre of Mass (COM) displacement are commonly used16,17. The first measures the COM displacement from the initial flat-footed standing position to the apex of the jump18. The second focuses on the COM displacement from the moment of take-off to the jump’s peak16,17. The My Jump smartphone application predominantly employs the latter method, offering greater flexibility in measuring jump performance, as it does not require a specific starting position3,19. While traditional Force Platforms and 3D motion capture systems are considered the ‘gold standard’ for jump height measurement due to their high levels of accuracy and reliability16,20, they come with challenges such as high costs, complex setup procedures, and limited portability1,3,19. Hence, practitioners frequently turn to more cost-effective and portable alternatives like My Jump2,13,21. However, users should be aware that the application comes with potential sources of error. These include the risk of artificially extending the flight time due to knee and ankle flexion (dorsiflexion) just before landing, among other factors.

The calculation methods for jump height are generally categorized into two main groups: Indirect and Direct methods16,21. Indirect methods, such as the Flight Time Method, Impulse-Momentum Method, and Double Integration Method, involve complex mathematical calculations and are often used in conjunction with force platforms to provide vertical ground reaction force data16,17,19,22. Direct methods, on the other hand, provide jump height values directly from vertical jump systems or are derived from motion capture systems16,23,24. The choice of calculation method and equipment can significantly impact the reliability, validity, and accuracy of the jump height measurements2,3,5. In this context, the Impulse-Momentum Theorem provides a reliable and objective measurement methodology15. This theorem is logically equivalent to Newton’s second law of motion, stating “Impulse (I) is equal to mass (m) times velocity (v)”. Force plates calculate jump height using this velocity value. Specifically, the square of the take-off velocity is divided by 9.81, and this result is then halved. These calculations are based on the Law of Conservation of Energy and yield highly reliable and valid jump height values. Therefore, the Impulse-Momentum Theorem offers an effective method for accurately measuring vertical jump height and minimizes the possibilities of ‘gaming the system’ or manipulation in competitive settings16. However, Due to their intricate configurations and substantial costs, traditional methods like force platforms are often unsuitable for field use, paving the way for the increasing popularity of My Jump smartphone application that offer instant results without the need for complex setups or specialized training. While they may lack the capability to provide vertical ground reaction force data, their user-friendly interfaces and instant feedback features make them highly practical for field use3,16,25.

The vertical jump is a dynamically appropriate movement pattern as the movement mechanics required in the actual competition can be repeated during tests and training26,27. As a result of all these positive aspects, jump tests have become indispensable for globally used test batteries. These include EUROFIT, a European test battery for assessing physical fitness; FITNESSGRAM, an American health-related physical fitness assessment; and ALPHA, a test battery designed to assess the lifestyle and fitness of young people2830. The My jump is the most popular technology developed to measure jump performance in the field of sports sciences2,3,19. There is a consensus that the My Jump provides many benefits, such as being user-friendly, accessible, portable, and affordable. Some studies report that this smartphone application yields valid and reliable results, whereas there are also concerns that the smartphone application may produce somewhat doubtable results1,2,14. Therefore, there is a necessity for a pooled evaluation of the research findings reporting the validity and reliability of the My Jump smartphone application in different populations. This study, which is composed of a systematic review and meta-analysis, aims to investigate whether the My Jump smartphone application (versions 1 and 2) produces valid and reliable results when measuring vertical jump height.

Methods

This systematic review and meta-analysis was conducted following the guidelines of the Cochrane handbook31 and following the recommendations of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA)32. We registered in the PROSPERO database (International Prospective Register of Systematic Reviews) with reference number CRD42022295759.

Data sources and search strategy

To identify potential studies for inclusion in this systematic review and meta-analysis, a search strategy was implemented in PubMed, Web of Science, Scopus, and EBSCO host databases from the date of their inceptions through to the 31st of September 2023. The search term was set to “My Jump” AND (“Validity” OR “Reliability”). Additionally, reference lists of the included studies, previous systematic reviews, and meta-analyses were reviewed for any other relevant studies. The search was limited to the English language. Three authors (CG, MT and SU) screened the titles and abstracts, and articles with the potential to be included in the study were read in full for further examination. Relevant studies were assessed against the inclusion and exclusion criteria described below. Inter-author disagreements were resolved by consensus decisions or by the senior author's final decision (İİ).

Inclusion criteria

The inclusion criteria for the articles in the systematic review and meta-analysis were focused on the types of studies, testing methods, participants, variables, statistical analysis, and reported outputs. To be included, the studies had to: (1) be original research; (2) be published in a peer-reviewed scientific journal; (3) have the full text available in English; (4) investigate vertical jump performance; (5) have included human participants such as athletes, untrained individuals, adults, elderly, children, etc.; (6) have investigated validity or reliability scores of the My Jump app; (7) report the Pearson r correlation, regression, and intraclass correlation coefficients (ICCs), or means and standard deviations.

Exclusion criteria

The following types of studies were excluded from the present meta-analysis: (1) studies in a language other than English; (2) unpublished studies, reviews, book chapters, editorials, non-peer-reviewed texts, case studies, abstracts, theses; (3) studies not reporting any validity or reliability statistics; (4) studies that focused on animal experiments; (5) studies using an application other than My Jump; (6) studies examining a physical fitness characteristic other than the vertical jump; (7) studies examining the effect of an exercise intervention on vertical performance.

Data extraction

The extracted data included the authors, year of publication, sample subject characteristics (age, body mass, height), criterion, type of vertical jump, means and standard deviations of the My Jump and criterion measurements, and the validity and reliability outputs (Table 1). Three authors (CG, MT and SÖ) independently extracted the data from the selected articles using a pre-defined form created in Microsoft Excel (Microsoft Corporation, Redmond, WA, USA). If there were any disagreements between the authors about the extracted data, the accuracy of the information was re-checked to reach a consensus.

Table 1.

Descriptive information of included studies.

Study Sample characteristics Criterion Activity patterns Study design My Jump results (mean ± SD) Criterion results (mean ± SD) Validity outputs Reliability outputs
Alias et al. (2021) N = 25, recreational athletes None CMJ Participants performed CMJ for five attempts and was recorded by using an iPhone 7 Plus in 240 frames per second. The videos were then rated by three raters and rated again 7 days later using the My Jump 2 mobile application 23.7 ± 5.7 None None

ICC = 1.00 (inter-rater)

ICC = 1.00 (intra-rater)

ICC = 1.00 (within session-device)

Balsalobre-Fernández et al. (2015) N = 20, recreationally active, healthy, sport sciences male students, age = 22.1 ± 3.6 years, height = 181 ± 8 cm, body mass = 74.0 ± 10.4 kg Force platform (Kistler 9287 BA, Kistler Instruments Ltd., Hook, UK) CMJ Participants performed five CMJs on the force platform while being recorded with an iPhone 5 s. Each jump was separated by a 2-min passive rest period Not reported Not reported Pearson r = 0.995 ICC = 0.997 (within session-device); 0.999 (inter-rater)
Barbalho et al. (2021) N = 11, male soccer players, age = 18.2 ± 1.3 years, height = 174 ± 7 cm, body mass = 69.9 ± 9.5 kg Force platform (BIOMEC400; EMG System do Brasil, Ltda, SP, Brazil) DJ30 The data collection was done in a single session. The My Jump 2 app and force plate data were recorded simultaneously from the 3 jumps that were performed by each athlete 33.1 ± 11.5 33.4 ± 11.6 Not reported

ICC = 1.00 (inter-rater)

ICC = 1.00 (intra-rater)

Bishop et al. (2022a) N = 27, postgraduate sport science students, age = 26.3 ± 5.1 years, height = 178 ± 6 cm, body mass = 82.8 ± 11.7 kg Force platform (Hawkin Dynamics, Westbrook, ME, USA) CMJ Participants performed three practice trials of the CMJ, at their perceived maximal effort. An iPad device was mounted to a tripod at a height of 0.75 m at a distance of 3 m from the front of the force plates 37.0 ± 8.0 37.0 ± 7.0 Pearson r = 0.98 ICC = 0.997 (within session-device)
Bishop et al. (2022b) N = 30, national-level youth basketball athletes, age = 17.7 ± 1.3 years, height = 181 ± 10 cm, body mass = 73.3 ± 13.3 kg None

CMJSL

DJSL

Participants performed single-leg countermovement jumps and single-leg drop jumps all assessed using the My Jump 2 app. To analyze the jump tests, a trained sports scientist with 2 years of experience in slow-motion video apps recorded a video of each test for its analysis using the My Jump 2

CMJSL_L1 = 13.7 ± 5.4

CMJSL_L2 = 13.7 ± 5.9

CMJSL_R1 = 14.2 ± 5.3

CMJSL_R1 = 13.3 ± 4.7

DJSL_L1 = 11.6 ± 6.0

DJSL_L2 = 11.0 ± 5.7

DJSL_R1 = 12.1 ± 5.9

DJSL_R1 = 11.3 ± 5.7

None None

ICC = 0.965 (within session-device for CMJSL_L); 0.940 (within session-device for CMJSL_R)

ICC = 0.975 (within session-device for DJSL_L); 0.965 (within session-device for DJSL_R)

Bogataj et al. (2020a) N = 44, recreationally active and had membership in the local gym in Subotica, Serbia, 26 male (age = 30 ± 10 years, height = 178 ± 16 cm, body mass = 86 ± 24 kg) and 18 female (age = 29 ± 5 years, height = 170 ± 6 cm, body mass = 60 ± 9 kg) Photocell system (Optojump photocell system; Microgate, Bolzano, Italy)

CMJ

CMJAS

SQJ

Participants performed three jumps (in three styles) on the photocell system while being recorded with an iPhone X. Each jump was separated by a 2 min rest period

CMJ: 31.9 ± 6.6

SQJ: 29.6 ± 6.0

CMJAS: 39.4 ± 9.7

CMJ: 32.5 ± 7.1

SQJ: 30.0 ± 6.3

CMJAS: 39.7 ± 9.5

CMJ, Pearson r = 0.98 (for males); 0.96 (for females)

Squat Jump, Pearson r = 0.95 (for males); 0.97 (for females)

CMJ with arm swing, Pearson r = 0.98 (for males); 0.94 (for females)

CMJ, ICC = 0.96 (absolute agreement for males); 0.97 (absolute agreement for females)

Squat Jump, ICC = 0.93 (absolute agreement for males); 0.94 (absolute agreement for females)

CMJ with arm swing, ICC = 0.97 (absolute agreement for males); 0.97 (absolute agreement for females)

Bogataj et al. (2020b) N = 48, primary school children from Subotica, Serbia, 26 male (age = 12.3 ± 0.8 years, height = 159 ± 13 cm, body mass = 51.8 ± 18.3 kg) and 22 female (age = 11.8 ± 0.8 years, height = 157 ± 10 cm, body mass = 50.6 ± 11.6 kg) Photocell system (Optojump photocell system; Microgate, Bolzano, Italy)

CMJ

CMJAS

SQJ

Participants performed three jumps (in three styles) on the photocell system while being recorded with an iPhone X. Each jump was separated by a 2 min rest period

CMJ: 24.5 ± 4.7

SQJ: 22.3 ± 4.1

CMJAS: 27.0 ± 5.8

CMJ: 24.6 ± 4.3

SQJ: 22.2 ± 4.5

CMJAS: 27.2 ± 5.8

CMJ, Pearson r = 0.97

Squat Jump, Pearson r = 0.97

CMJ with arm swing, Pearson r = 0.99

CMJ, ICC = 0.96 (absolute agreement)

Squat Jump, ICC = 0.88 (absolute agreement)

CMJ with arm swing, ICC = 0.93 (absolute agreement)

Brooks et al. (2018) N = 26 (14 male and 12 female), healthy and recreationally active adults, age = 23.2 ± 3.5 years, height = 170 ± 10 cm, body mass = 74.0 ± 10.4 kg

(1) AMTI AccuPower force platform (Advanced Mechanical Technology Inc., MA, USA)

(2) Yardstick jump-and-reach apparatus

CMJAS Participants completed the test protocol on two occasions, separated by a minimum of seven days. Participants performed three jumps on the force platform while being recorded with an iPhone 5 s. Each jump was separated by a 2 min passive rest period Not reported Not reported Pearson r = 0.98 (Force platform); 0.94 (Yardstick) ICC = 0.97 (inter-session); 0.99 (intra-rater)
Carlos-Vivas et al. (2018) N = 40 (29 male and 11 female), recreationally active, healthy, sport sciences students, age = 21.4 ± 1.9 years, height = 174 ± 7 cm, body mass = 68.7 ± 8.4 kg Force platform (BioWare v. 5.3.0.7, Kistler holding aG, Winterthur, Switzerland) CMJ Participants performed five CMJs on the force platform while being recorded with an iPhone 6 28.6 ± 7.2

28.7 ± 7.2 (time in the air method)

28.4 ± 6.8 (velocity at takeoff method)

ICC = 1.00 (inter-instruments reliability for the time in the air method); 0.996 (inter-instruments reliability for the velocity at takeoff method2); ICC = 0.983 (within session-device)
Chow et al. (2023) N = 30, physically active college students, 14 males and 16 females, age = 23.0 ± 1.7 years, height = 168 ± 6 cm, body mass = 61.9 ± 9.8 kg Takei Vertical Jump Meter (Takei Scientific Instruments Co., Niigata, Japan), HomeCourt (homecourt.ai) CMJ Participants visited the laboratory twice, with two days in between, and performed three jumps each day. All jumps were recorded by My Jump 2, HomeCourt, and the Takei Vertical Jump Meter simultaneously 40.9 ± 7.9

46.1 ± 7.6 (HomeCourt)

42.0 ± 8.1 (Takei Vertical Jump Meter)

Pearson r = 0.85 (My Jump 2—HomeCourt)

Pearson r = 0.93 (My Jump 2—Takei Vertical Jump Meter)

ICC = 0.86 (within session-device for day 1)

ICC = 0.88 (within session-device for day 2)

Cruvinel-Cabral et al. (2018) N = 41, elderly people, 12 male (age: 73.2 ± 6.4 years, body mass = 68.3 ± 12.7 kg) and 29 female (age: 69.4 ± 8.9 years, body mass = 64.7 ± 12.6 kg) Contact mat (Chronojump, version. 1.6.2; Boscosystem, Barcelona, Spain) CMJ Participants performed three CMJs. Contact mat and My Jump, were used simultaneously to assess VJ height 10.2 ± 5.1 10.01 ± 5.0 Pearson r = 0.999 ICC = 0.948 (within session-device)
Driller et al. (2017) N = 61 (30 male and 31 female), recreational to highly trained athletes, age = 20 ± 4 years, body mass = 76.4 ± 15.2 kg Force plate (Dual-Axis Force Platform, PASCO, California, USA) CMJ Participants performed three CMJs on the force platform while being recorded with an iPhone 6 s. Each jump was separated by 5 s 25.9 ± 7.9 25.1 ± 7.5 Pearson r = 0.96 ICC = 0.97 (inter-rater)
Gallardo-Fuentes et al. (2016) N = 21 (14 male and 7 female), track and field national and international-level competitive athletes, age = 22.1 ± 3.6 years, height = 181 ± 8 cm, body mass = 74.0 ± 10.4 kg Contact platform (Ergotester, Globus, Codogne, Italy)

CMJ

SQJ

DJ40

Participants completed the test protocol on two occasions, separated by 48 h. Participants performed five jumps in three styles on the contact platform while being recorded with an iPhone 5 s. Each jump was separated by a 2-min passive rest period

Test results: CMJ: 37.0 ± 10.3 SQJ: 35.7 ± 9.1 DJ40: 31.6 ± 5.9

Re-test results: CMJ: 36.8 ± 9.5 SQJ: 34.6 ± 9.1 DJ40: 31.6 ± 7.0

Test results: CMJ: 37.1 ± 9.9 SQJ: 35.8 ± 8.8 DJ40: 31.5 ± 5.9

Re-test results: CMJ: 36.8 ± 9.3 SQJ: 34.9 ± 8.6 DJ40: 31.7 ± 7.0

CMJ, Pearson r = 0.99 (between two instruments, first day); 0.99 (between two instruments, second day)

Squat Jump, Pearson r = 0.99 (between two instruments, first day); 0.99 (between two instruments, second day)

Drop Jump, Pearson r = 0.99 (between two instruments, first day); 0.99 (between two instruments, second day)

CMJ, ICC = 0.99 (within session-device for the first day); 0,99 (within session-device for the second day) Pearson r = 0.95 (inter-session)

Squat Jump, ICC = 0.99 (within session-device for the first day); 0,98 (within session-device for the second day) Pearson r = 0.90 (inter-session)

Drop Jump, ICC = 0.99 (within session-device for the first day); 0,98 (within session-device for the second day) Pearson r = 0.87 (inter-session)

Gür and Ayan (2023) N = 24 (13 male and 11 female), healthy sedentary individuals, age = 22.3 ± 1.1 years, height = 170 ± 9 cm, body mass = 64.0 ± 10.7 kg Smart Jump (Fusion Sport, Queensland, Australia) CMJ After the aerobic and dynamic stretching warm-up protocol, each participant performed four CMJ jumps with maximal effort. Participants were given a 2-min passive recovery between each trial. Jumps were recorded simultaneously with the My Jump 2 mobile app and Fusion Sport brand Smart Jump splash mat device

33.3 ± 7.3 (jump 1)

33.9 ± 7.0 (jump 2)

33.6 ± 6.4 (jump 3)

33.9 ± 6.2 (jump 4)

33.5 ± 7.3 (jump 1)

33.4 ± 6.9 (jump 2)

34.2 ± 6.2 (jump 3)

34.5 ± 6.2 (jump 4)

Pearson r = 0.99 ICC = 0.99 (within session-device)
Haynes et al. (2019) N = 14, male sports sciences student, age = 29.5 ± 9.9 years, height = 178 ± 10 cm, body mass = 81.4 ± 14.1 kg Force platform (FP8, Hurlab, Finland)

DJ20

DJ40

Participants completed the test protocol on two occasions, separated by seven days. Participants performed three jumps (drop heights of 20 and 40 cm) on the force plate while being recorded with an iPhone 5 s. Each jump was separated by a 2 min passive rest period

DJ20: 23.8 ± 7.3

DJ40: 22.6 ± 5.6

DJ20: 23.3 ± 6.2

DJ40: 23.3 ± 5.3

Drop jump from 20 cm, Pearson r = 0.812

Drop jump from 40 cm, Pearson r = 0.959

Drop jump from 20 cm ICC = 0.803 (within session-device)

Drop jump from 40 cm ICC = 0.958 (within session-device)

Jimenez-Olmedo et al. (2022) N = 39, active adult athletes, 25 male (age = 22.2 ± 2.7 years, height = 180.1 ± 4.4 cm, body mass = 77.6 ± 6.8 kg) and 14 female (age = 23.2 ± 1.8 years, height = 170.7 ± 4.4 cm, body mass = 66.2 ± 4.0 kg) two iPhone 7 units (Apple Inc., Cupertino, CA, USA) CMJ Two identical smartphones recorded 195 countermovement jump executions at heights 30 and 90 cm, which were randomly assessed by three experienced observers. The videos were randomly analyzed in regard to the observation heights, jump trials, and participants Not reported Not reported Not reported

CMJ, ICC = 0.99 (between-observer)

ICC = 0.99 (within-observer)

Patiño-Palma et al. (2022) N = 119, high-performance athletes from different sports disciplines, age = 18.5 ± 1.3 years, height = 174 ± 6 cm, body mass = 67.4 ± 6.0 kg Chronojump Boscosystem (Barcelona, Spain), OptoGait (Bolzano, Italy), and Wheeler Jump (Wheeler Sports Tech, FL, USA) CMJ Jump performance was evaluated through the CMJ in a training session using the Chronojump Boscosystem contact platform, the OptoGait photoelectric system, and the My Jump 2 mobile application as measurement tools, comparing the results with the values obtained with the Wheeler Jump sensor CMJ: 42.8 ± 6.9 CMJ: 39.3 ± 7.1 Rho = 0.994 ICC = 0.993 (within session-device)
Plakoutsis et al. (2023) N = 34, 22 male and 12 female, collegiate athletes, age = 21.6 ± 5.7 years KForce Plates (K-Invent, Montpellier, France) CMJ Participants performed three maximal CMJs while standing on a portable force platform. The jumps were recorded with a portable KForce plates system and My Jump 2 through iPhone 13 at the same time. Each participant repeated the testing procedure after seven days in order to assess the reliability of the measurements (ICC) Not reported Not reported Pearson r = 1.00 Not reported
Soares et al. (2023) N = 21 (15 male and 6 female), healthy, judo athletes, age = 26.4 ± 5.4 years, height = 172 ± 8 cm, body mass = 72.6 ± 12.9 kg Chronojump Boscosystem (Barcelona, Spain) CMJ Participants performed two countermovement jumps on the Chronojump platform (42 jumps). Simultaneously, the videos of the jumps were captured using recommendations in the app and were later processed and analyzed independently by two evaluators

CMJ: 26.4 ± 8.9 (Evaluator 1)

CMJ: 26.3 ± 8.1 (Evaluator 2)

CMJ: 26.7 ± 8.1

ICC = 0.94 (Evaluator 1 vs Chronojump)

ICC = 0.97 (Evaluator 2 vs Chronojump)

ICC = 0.95 (inter-rater)
Stanton et al. (2017) N = 29 (10 male and 19 female), healthy, recreationally active adults, age = 26.4 ± 5.4 years, height = 172 ± 8 cm, body mass = 72.6 ± 12.9 kg AMTI BP400 800–2000 force plate (Advanced Mechanical Technology Inc., Watertown, MA) CMJ DJ30 Participants completed the test protocol on two occasions, separated by seven days. Participants performed three jumps in two styles on the force plate while being recorded with an iPhone 5 s. Each jump was separated by a 2-min passive rest period

CMJ: 20.6 ± 8.5

DJ30: 19.4 ± 8.4

CMJ: 20.4 ± 7.6

DJ30: 20.3 ± 8.3

Pearson r = 0.997

Pearson r = 0.998

CMJ, ICC = 0.99 (within session-device); 0.99 (intra-rater)

Drop Jump, ICC = 0.99 (within session-device); 0.99 (intra-rater)

Yingling et al. (2018) N = 135, healthy adults, 94 male (age: 18–29 years, height = 177 ± 8 cm, body mass = 72.8 ± 9.9 kg) and 41 female (age: 18–39 years, height = 167 ± 8 cm, body mass = 63.5 ± 9.3 kg) Vertec linear position transducers (JUMPUSA.com, Sunnyvale, CA, USA) CMJAS Participants performed three jumps. Vertec and My Jump, were used simultaneously to assess VJ height 43.05 ± 12.13 51.93 ± 14.36 Pearson r = 0.813 ICC = 0.813 (consistency); 0.665 (absolute agreement)

CMJ: countermovement jump; CMJSL: single leg countermovement jump; SQJ: squat jump; CMJAS: countermovement jump with arm swing; DJ: drop jump; DJSL: single leg drop jump; r: correlation coefficient; ICC: intraclass correlation coefficient.

Methodological quality assessment

The methodological quality of each included study was assessed using a modified Downs and Black assessment scale33. A total of 8 domains were identified to evaluate the quality of reporting for studies included in this review: (1) the hypothesis/aim described; (2) whether the participants were representative of the target population; (3) the participant characteristics detailed; (4) the intervention procedure detailed; (5) the use of an appropriate reference test/criterion; (6) the use of appropriate statistical tests; (7) the main outcomes reported; (8) if the outcome measures valid and reliable. Each criterion was evaluated as low quality, moderate quality, high quality, inadequate, or unclear.

Meta-analyses

Meta-analyses were conducted using comprehensive meta-analysis software, version 2 for Windows (CMA, Biostat company, Englewood, NJ, USA)34. The meta-analysis of validity was performed in two ways: (1) the means and standard deviations were compared between the My Jump and criterion measurements to assess the agreement of raw scores; (2) the correlation coefficients were used to determine the consistency of the rankings within-group in the My Jump and criterion measurements. Additionally, a meta-analysis of reported ICCs was performed to confirm reliability. However, the types of ICCs or Pearson r coefficients reported in the studies were varied (inter-rater, intra-rater, within-subject, between the devices, between two test days, between the consecutive jump performances of the same participant, etc.). Therefore, when a study reported multiple Pearson r coefficients or ICCs from the same sample group, the study was used as a single unit of analysis to avoid the overestimation of its contribution to the pooled result due to double counting. The pooled correlation values were interpreted according to a random-effects model in case of any heterogeneity between studies (when the p-value of the Q statistic was less than 0.1)35. For validity, pooled correlations were classified as follows: 0–0.19, “no significant correlation”; 0.2–0.39, “low correlation”; 0.4–0.59, “moderate correlation”; 0.6–0.79, “moderately high correlation”; and ≥ 0.8, “high correlation”36,37. The scale designed by Landis and Koch reliability strength thresholds was applied, as follows: 0.01–0.20, “mild reliability”; 0.21–0.40, “fair reliability”; 0.41–0.60, “moderate reliability”, 0.61–0.80, “substantial reliability”; and 0.81–1.00, “nearly perfect reliability”38. Sub-analyses were meticulously conducted to delve into the nuances of various factors affecting the outcomes. These sub-analyses were categorized based on three key parameters: the type of jump performed (CMJ, SQJ, and DJ), the criterion device used for measurement (Force Plates and other devices), and the type of reliability assessed (inter-rater and intra-rater reliability, inter-session, and within-session reliability).

Heterogeneity was determined by Cochrane’s Q statistic and its p-value, I-squared value, and tau-squared value35,39,40. The Q-value (and its p-value), which indicates whether all the studies have shared a common effect size, and the I-squared value, which refers to the proportion of the observed variance when the sampling error is eliminated (i.e. when observing the true effect size for all studies in the analysis), are the most common heterogeneity indicators37,39,41,42. The I-squared values of < 25%, 25–75%, and > 75% were considered to represent low, moderate, and high levels of heterogeneity, respectively43. The Tau-squared value is a measure of the variance of true effects representing a concrete and reliable heterogeneity31,35,41.

The risk of publication bias was explored using funnel plot symmetry, and asymmetries were confirmed using the extended Egger’s test44. Egger’s test regresses the standardized effect sizes on a measure of precision (i.e. standard errors of the correlation coefficients). A significant coefficient for Egger’s test means that the effect sizes and sampling variance for each study are related and indicates that a publication bias is present. In the case of evidence of a publication bias, Duval and Tweedie’s “trim and fill” procedure was applied to determine whether estimates required adjustment based on missing studies37. Additionally, sensitivity analyses were conducted by removing a study14 with validity concerns to assess the robustness of the pooled estimates.

Results

Study selection

We initially found a total of 74 potential research articles related to the My Jump smartphone application published until September 2023. After excluding the 44 duplicates and 4 studies based on their titles and abstracts, 26 studies were reviewed as full texts. Following the identification of studies meeting the inclusion criteria of this paper, a total of 21 studies consisting of 839 accumulated participants were included in the present meta-analysis (Fig. 1).

Figure 1.

Figure 1

Flow chart of the review process.

Methodological quality

Using a modified Downs and Black assessment scale33, eight risk domains for the 21 individual research articles (a total of 168 scores) were evaluated. There were only nine items scored as low quality and one item scored as inadequate, while all other items were rated as moderate or high quality; therefore, the overall methodological quality was considered as moderate-to-high (Table 2).

Table 2.

Methodological quality assessments of original studies included in meta-analyses.

Studies/items 1 2 3 4 5 6 7 8
Alias et al. (2021) L M L M M M M M
Balsalobre-Fernández et al. (2015) H L H M H H L M
Barbalho et al. (2021) H L H M H M M M
Bishop et al. (2022a) H M H M H H M M
Bishop et al. (2022b) H M H M M H M M
Bogataj et al. (2020a) H M H M M M M M
Bogataj et al. (2020b) H M H M M M M M
Brooks et al. (2018) H L H M H H L M
Carlos-Vivas et al. (2018) H M H M H M M M
Chow et al. (2023) H M H M M H M M
Cruvinel-Cabral et al. (2018) H M M M M M M M
Driller et al. (2017) I H H M H M M M
Gallardo-Fuentes et al. (2016) H L H M M H M M
Gür and Ayan (2023) M M H M M M M M
Haynes et al. (2019) M L H M H M M M
Jimenez-Olmedo et al. (2022) H M H M M H M M
Patiño-Palma et al. (2022) H H H M M H M M
Plakoutsis et al. (2023) H M H M H H M M
Soares et al. (2023) M M H M M M M M
Stanton et al. (2017) H M H M H H H M
Yingling et al. (2018) H H H M L M M M

L low quality, M moderate quality, H high quality, I inadequate.

Meta-analysis results

Heterogeneity and publication bias outputs

The heterogeneity statistics and publication bias were assessed for three main categories: Mean differences, Reliability analysis (ICC values), and Validity analysis (r values). For mean differences, the Cochran Q statistic was 46.67 (p < 0.001), with an I2 value of 70.0% and a tau2 of 0.062. Egger’s test for publication bias was not significant (p = 0.860). For reliability analysis, the Cochran Q statistic was notably high at 512.5 (p < 0.001). The I2 value was 96.5%, and tau2 was 0.397. Egger’s test indicated a p-value of 0.156. For validity analysis, the Cochran Q statistic was 959.2 (p < 0.001), with an I2 value of 98.3% and a tau2 of 0.864. Egger's test showed a p-value of 0.436. High I2 values, such as those observed in the Reliability and Validity analyses, indicate substantial heterogeneity among the included studies. An I2 value above 75% is generally considered to represent considerable heterogeneity. This suggests that the observed variations in effect sizes are not solely due to sampling error but may be attributed to other factors, such as methodological differences or population characteristics among the studies (Table 3).

Table 3.

Summary statistics related to the heterogeneity and publication bias.

Validity and reliability analyses Heterogeneity statistics Publication bias
Q p I2 Tau2 Egger p
Mean differences 46.67  < 0.001 70.0 0.062 0.29 0.860
Reliability analysis (ICC values) 512.5  < 0.001 96.5 0.397 4.61 0.156
Validity analysis (r values) 959.2  < 0.001 98.3 0.864 4.16 0.436

ICC intraclass correlation coefficients. Q Cochran Q statistic for homogeneity test, I2: the proportion of total variation caused by heterogeneity rather than within‐study sampling error (%), Tau2: the variance in true effect sizes observed in different studies, Egger: Egger’s regression test.

Validity outputs

The meta-analysis conducted for the agreement between raw scores showed that there was no significant difference between My Jump and the criteria (Hedge’s g = − 0.047; p = 0.21, Fig. 2). Further analyses showed a significant heterogeneity (Q = 46.67; p < 0.001; tau2 = 0.062), with an I2 value indicating 70.0% of effect size variance accounted for across the individual studies (Table 3). Because of the significant heterogeneity, the pooled effect size was conducted according to the random-effects model. Additionally, the risk of publication bias was explored using funnel plot symmetry and confirmed using the extended Egger’s test (Table 3). Egger’s test did not show any potential asymmetry (p = 0.860).

Figure 2.

Figure 2

Forest plot of differences between My Jump and related criterion measures. Values shown are Hedge’s g with 95% confidence intervals. The size of the plotted squares represents the relative weight of the study.

The sub-analysis for CMJ included 13 studies. The fixed effect model yielded a Hedge’s g value of − 0.060 (95% CI − 0.151 to 0.032, p = 0.202), while the random effect model indicated a Hedge's g value of − 0.047 (95% CI − 0.233 to 0.139, p = 0.622). The studies exhibited varying degrees of relative weight, ranging from 3.04 to 14.07%. In the sub-analysis for Squat SQJ, three studies were included. The fixed and random effect models both showed a Hedge’s g value of − 0.020 (95% CI − 0.257 to 0.217, p = 0.869). The studies in this category had relative weights of 31.38%, 32.83%, and 35.78%. The sub-analysis for DJ comprised four studies. Both the fixed and random effect models indicated a Hedge’s g value of − 0.034 (95% CI − 0.295 to 0.226, p = 0.795). The relative weights for the studies in this sub-analysis ranged from 10.50 to 37.79% (Table 4).

Table 4.

Sub-validity analyses for vertical jump types based on Hedge’s g values.

Study Hedge’s g Standard error Variance Lower limit Upper limit Z-value p-value Relative weight (%)
Random Fixed
Sub-analysis for countermovement jump (CMJ)
 Bishop et al. (2022a) 0.000 0.268 0.072 – 0.53 0.53 0.00 1.000 5.79 3.04
 Bogataj et al. (2020a) – 0.049 0.150 0.023 – 0.34 0.24 – 0.33 0.742 8.49 9.71
 Bogataj et al. (2020b) – 0.028 0.144 0.021 – 0.31 0.25 – 0.20 0.845 8.64 10.59
 Carlos–Vivas et al. (2018) 0.011 0.157 0.025 – 0.30 0.32 0.07 0.946 8.31 8.84
 Chow et al. (2023) – 0.394 0.183 0.034 – 0.75 –0.03 – 2.15 0.032 7.69 6.52
 Cruvinel-Cabral et al. (2018) 0.027 0.219 0.048 – 0.40 0.46 0.13 0.901 6.85 4.57
 Driller et al. (2017) 0.103 0.180 0.032 – 0.25 0.46 0.57 0.567 7.76 6.75
 Gallardo-Fuentes et al. (2016) – 0.005 0.216 0.047 – 0.43 0.42 – 0.02 0.981 6.91 4.68
 Gür and Ayan, (2023) – 0.034 0.144 0.021 – 0.32 0.25 – 0.24 0.813 8.64 10.59
 Patiño-Palma et al. (2022) 0.498 0.131 0.017 0.24 0.76 3.80 0.000 8.94 12.71
 Soares et al. (2023) – 0.042 0.216 0.047 – 0.466 0.382 – 0.195 0.845 6.91 4.68
 Stanton et al. (2017) 0.026 0.259 0.067 – 0.482 0.534 0.099 0.921 5.97 3.26
 Yingling et al. (2018) – 0.666 0.125 0.016 – 0.911 –0.422 – 5.342 0.000 9.09 14.07
 Fixed effect model – 0.060 0.047 0.002 – 0.151 0.032 – 1.276 0.202
 Random effect model – 0.047 0.095 0.009 – 0.233 0.139 – 0.494 0.622
Sub-analysis for squat jump (SQJ)
 Bogataj et al. (2020a) – 0.064 0.211 0.045 – 0.48 0.35 – 0.30 0.760 32.83 32.83
 Bogataj et al. (2020b) 0.023 0.202 0.041 – 0.37 0.42 0.11 0.909 35.78 35.78
 Gallardo-Fuentes et al. (2016) – 0.022 0.216 0.047 – 0.45 0.40 – 0.10 0.917 31.38 31.38
 Fixed effect model – 0.020 0.121 0.015 – 0.257 0.217 – 0.165 0.869
 Random effect model – 0.020 0.121 0.015 – 0.257 0.217 – 0.165 0.869
Sub-analysis for drop jump (DJ)
 Barbalho et al. (2021) – 0.028 0.410 0.168 – 0.83 0.78 – 0.07 0.945 10.50 10.50
 Gallardo-Fuentes et al. (2016) 0.000 0.216 0.047 – 0.42 0.42 0.00 1.000 37.79 37.79
 Haynes et al. (2019) – 0.019 0.264 0.069 – 0.54 0.50 – 0.07 0.943 25.43 25.43
 Stanton et al. (2017) – 0.102 0.259 0.067 – 0.61 0.41 – 0.39 0.695 26.28 26.28
 Fixed effect model – 0.034 0.133 0.018 – 0.295 0.226 – 0.259 0.795
 Random effect model – 0.034 0.133 0.018 – 0.295 0.226 – 0.259 0.795

The sub-analysis focused on studies utilizing Force Plates included six studies. The fixed and random effect models both indicated a Hedge’s g value of 0.015 (95% CI − 0.154 to 0.184, p = 0.863). The relative weights for the studies in this sub-analysis ranged from 4.41 to 29.95%. In the sub-analysis focused on studies not utilizing Force Plates, nine studies were included. The fixed effect model showed a Hedge’s g value of − 0.073 (95% CI − 0.166 to 0.020, p = 0.124), while the random effect model indicated a Hedge’s g value of − 0.075 (95% CI − 0.302 to 0.151, p = 0.515). The studies in this category had relative weights ranging from 4.69 to 16.27% (Table 5).

Table 5.

Sub-validity analyses for the criterion device based on Hedge’s g values.

Study Hedge’s g Standard error Variance Lower limit Upper limit Z-value p-value Relative weight (%)
Random Fixed
Sub-analysis focused on studies utilizing force plates for criterion measurement
 Barbalho et al. (2021) – 0.028 0.410 0.168 – 0.83 0.78 – 0.07 0.945 4.41 4.41
 Bishop et al. (2022a) 0.000 0.268 0.072 – 0.53 0.53 0.00 1.000 10.31 10.31
 Carlos-Vivas et al. (2018) 0.011 0.157 0.025 – 0.30 0.32 0.07 0.946 29.95 29.95
 Driller et al. (2017) 0.103 0.180 0.032 – 0.25 0.46 0.57 0.567 22.87 22.87
 Haynes et al. (2019) – 0.019 0.264 0.069 – 0.54 0.50 – 0.07 0.943 10.68 10.68
 Stanton et al. (2017) – 0.040 0.184 0.034 – 0.40 0.32 – 0.22 0.830 21.79 21.79
 Fixed effect model 0.015 0.086 0.007 – 0.154 0.184 0.173 0.863
 Random effect model 0.015 0.086 0.007 – 0.154 0.184 0.173 0.863
Sub-analysis focused on studies not utilizing force plates for criterion measurement
 Bogataj et al. (2020a) – 0.050 0.123 0.015 – 0.29 0.19 – 0.41 0.685 11.98 14.91
 Bogataj et al. (2020b) – 0.013 0.118 0.014 – 0.24 0.22 – 0.11 0.915 12.12 16.27
 Chow et al. (2023) – 0.394 0.183 0.034 – 0.75 – 0.03 – 2.15 0.032 10.28 6.70
 Cruvinel-Cabral et al. (2018) 0.027 0.219 0.048 – 0.40 0.46 0.13 0.901 9.26 4.69
 Gallardo-Fuentes et al. (2016) – 0.010 0.126 0.016 – 0.26 0.24 – 0.08 0.939 11.91 14.24
 Gür and Ayan (2023) – 0.034 0.144 0.021 – 0.32 0.25 – 0.24 0.813 11.41 10.87
 Patiño-Palma et al. (2022) 0.498 0.131 0.017 0.24 0.76 3.80 0.000 11.76 13.05
 Soares et al. (2023) – 0.042 0.216 0.047 – 0.47 0.38 – 0.20 0.845 9.34 4.81
 Yingling et al. (2018) – 0.666 0.125 0.016 – 0.91 – 0.42 – 5.34 0.000 11.93 14.45
 Fixed effect model – 0.073 0.047 0.002 – 0.166 0.020 – 1.538 0.124
 Random effect model – 0.075 0.116 0.013 – 0.302 0.151 – 0.651 0.515

The meta-analysis conducted for identifying the consistency of the rankings within-group showed a high correlation (r = 0.989) between My Jump and criteria while individual studies reported correlations ranging from 0.813 to 0.999 (Fig. 3). Further analyses showed a significant heterogeneity (Q = 959.2; p < 0.001; tau2 = 0.864), with an I2 value indicating 98.3% of effect size variance accounted for across the individual studies (Table 3). The risk of publication bias was explored using funnel plot symmetry and confirmed using the extended Egger’s test (Table 3). Additionally, the risk of publication bias was explored using funnel plot symmetry and confirmed using the extended Egger’s test (Table 3). Egger's test did not show any potential asymmetry (p = 0.436).

Figure 3.

Figure 3

Forest plot for correlations between My Jump and related criterion measures. Values shown are correlation coefficient with 95% confidence intervals. The size of the plotted squares represents the relative weight of the study.

For CMJ, the fixed effect model showed a correlation of 0.990 (95% CI 0.988–0.991, p < 0.001), while the random effect model indicated a correlation of 0.992 (95% CI 0.981–0.996, p < 0.001). For SQJ, the fixed effect model indicated a correlation of 0.984 (95% CI 0.978–0.988, p < 0.001), and the random effect model showed a correlation of 0.989 (95% CI 0.913–0.999, p < 0.001). For DJ, the fixed effect model revealed a correlation of 0.996 (95% CI 0.994–0.997, p < 0.001), and the random effect model indicated a correlation of 0.994 (95% CI 0.923–1.000, p < 0.001). These results suggest that the different types of jumps are highly correlated with the criterion measures, indicating strong validity across the board (Table 6).

Table 6.

Sub-validity analyses for vertical jump types based on correlation values.

Study Correlation Lower limit Upper Limit Z-value p-value Relative weight (%)
Random Fixed
Sub-analysis for countermovement jump (CMJ)
 Balsalobre-Fernández et al. (2015) 0.995 0.987 0.998 12.347 0.000 6.36 2.05
 Bishop et al. (2022a) 0.980 0.956 0.991 11.26 0.000 6.52 2.89
 Bogataj et al. (2020a) 0.969 0.958 0.977 26.51 0.000 6.87 19.74
 Bogataj et al. (2020b) 0.983 0.974 0.988 22.48 0.000 6.82 10.83
 Brooks et al. (2018) 0.965 0.939 0.980 13.69 0.000 6.71 5.54
 Carlos-Vivas et al. (2018) 0.999 0.999 1.000 34.66 0.000 6.79 8.90
 Chow et al. (2023) 0.897 0.831 0.938 10.71 0.000 6.74 6.50
 Cruvinel-Cabral et al. (2018) 0.999 0.998 0.999 23.43 0.000 6.67 4.57
 Driller et al. (2017) 0.960 0.934 0.976 14.82 0.000 6.75 6.98
 Gallardo-Fuentes et al. (2016) 0.999 0.998 0.999 22.80 0.000 6.65 4.33
 Gür and Ayan (2023) 0.990 0.977 0.996 12.13 0.000 6.46 2.53
 Patiño-Palma et al. (2022) 0.994 0.991 0.996 31.27 0.000 6.84 13.96
 Plakoutsis et al. (2023) 0.999 0.998 1.000 21.16 0.000 6.61 3.73
 Soares et al. (2023) 0.958 0.920 0.978 11.49 0.000 6.65 4.33
 Stanton et al. (2017) 0.997 0.994 0.999 16.57 0.000 6.55 3.13
 Fixed effect model 0.990 0.988 0.991 75.834 0.000
 Random effect model 0.992 0.981 0.996 12.865 0.000
Sub-analysis for squat jump (SJ)
 Bogataj et al. (2020a) 0.961 0.941 0.975 17.767 0.000 33.65 50.31
 Bogataj et al. (2020b) 0.970 0.947 0.983 14.04 0.000 33.28 27.61
 Gallardo-Fuentes et al. (2016) 0.999 0.998 0.999 22.80 0.000 33.07 22.09
 Fixed effect model 0.984 0.978 0.988 30.692 0.000
 Random effect model 0.989 0.913 0.999 4.786 0.000
Sub-analysis for drop jump (DJ)
 Gallardo-Fuentes et al. (2016) 0.999 0.998 0.999 22.801 0.000 33.56 42.86
 Haynes et al. (2019) 0.911 0.806 0.960 7.19 0.000 33.13 26.19
 Stanton et al. (2017) 0.998 0.996 0.999 17.61 0.000 33.30 30.95
 Fixed effect model 0.996 0.994 0.997 28.404 0.000
 Random effect model 0.994 0.923 1.000 4.343 0.000

The correlation analyses were conducted to evaluate the validity of different criterion measures, specifically force plates, and non-force plates, in assessing jump performance. The fixed effect model for studies using force plates showed a correlation of 0.994 (95% CI 0.992–0.995, p < 0.001), while the random effect model indicated a correlation of 0.992 (95% CI 0.968–0.998, p < 0.001). On the other hand, for studies not using force plates, the fixed effect model revealed a correlation of 0.981 (95% CI 0.978–0.983, p < 0.001), and the random effect model showed a correlation of 0.985 (95% CI 0.951–0.995, p < 0.001). These results highlight the robustness and high validity of both types of criterion measures in assessing different types of jumps, as evidenced by the consistently high correlations across studies (Table 7).

Table 7.

Sub-validity analyses for criterion device based on correlation values.

Study Correlation Lower limit Upper Limit Z-value p-value Relative weight (%)
Random Fixed
Sub-analysis focused on studies utilizing force plates for criterion measurement
 Balsalobre-Fernández et al. (2015) 0.995 0.987 0.998 12.347 0.000 12.16 5.25
 Bishop et al. (2022a) 0.980 0.956 0.991 11.26 0.000 12.37 7.41
 Brooks et al. (2018) 0.965 0.939 0.980 13.69 0.000 12.62 14.20
 Carlos-Vivas et al. (2018) 0.999 0.999 1.000 34.66 0.000 12.73 22.84
 Driller et al. (2017) 0.960 0.934 0.976 14.82 0.000 12.68 17.90
 Haynes et al. (2019) 0.911 0.806 0.960 7.19 0.000 12.32 6.79
 Plakoutsis et al. (2023) 0.999 0.998 1.000 21.16 0.000 12.48 9.57
 Stanton et al. (2017) 0.998 0.996 0.999 24.17 0.000 12.65 16.05
 Fixed effect model 0.994 0.992 0.995 51.984 0.000
 Random effect model 0.992 0.968 0.998 7.843 0.000
Sub-analysis focused on studies not utilizing force plates for criterion measurement
 Bogataj et al. (2020a) 0.966 0.957 0.974 31.900 0.000 11.29
 Bogataj et al. (2020b) 0.979 0.971 0.985 26.46 0.000 11.24
 Chow et al. (2023) 0.897 0.831 0.938 10.71 0.000 11.09
 Cruvinel-Cabral et al. (2018) 0.999 0.998 0.999 23.43 0.000 10.99
 Gallardo-Fuentes et al. (2016) 0.999 0.999 0.999 39.49 0.000 11.22
 Gür and Ayan (2023) 0.990 0.977 0.996 12.13 0.000 10.72
 Patiño-Palma et al. (2022) 0.994 0.991 0.996 31.27 0.000 11.23
 Soares et al. (2023) 0.958 0.920 0.978 11.49 0.000 10.97
 Yingling et al. (2018) 0.813 0.747 0.863 13.05 0.000 11.24
 Fixed effect model 0.981 0.978 0.983 68.954 0.000
 Random effect model 0.985 0.951 0.995 8.019 0.000

Reliability outputs

The meta-analysis conducted for identifying the reliability of the My Jump smartphone application showed nearly perfect reliability scores (r = 0.986) while individual studies reported correlations ranging from 0.748 to 1.0 (Fig. 4). Further analyses showed a significant heterogeneity (Q = 512.5; p < 0.001; tau2 = 0.397), with an I2 value indicating 96.5% of effect size variance accounted for across the individual studies (Table 3). The risk of publication bias was explored using funnel plot symmetry and confirmed using the extended Egger’s test (Table 3). Egger’s test did not show any potential asymmetry (p = 0.386).

Figure 4.

Figure 4

Forest plot for ICCs of My Jump measures. Values shown are ICC with 95% confidence intervals. The size of the plotted squares represents the relative weight of the study. ICC intraclass correlation coefficients.

The ICC was used to assess the reliability of different types of jumps, including CMJ, SQJ, and DJ. For CMJ, the fixed effect model showed an ICC of 0.969 (95% CI 0.965–0.972, p < 0.001), and the random effect model indicated an ICC of 0.982 (95% CI 0.961–0.992, p < 0.001). For SJ, the fixed effect model revealed an ICC of 0.965 (95% CI 0.953–0.974, p < 0.001), and the random effect model showed an ICC of 0.961 (95% CI 0.889–0.987, p < 0.001). In the case of DJ, the fixed effect model indicated an ICC of 0.972 (95% CI 0.964–0.979, p < 0.001), while the random effect model showed an ICC of 0.987 (95% CI 0.940–0.997, p < 0.001). These results suggest that the methods used for assessing different types of jumps are highly reliable, as evidenced by the consistently high ICCs across studies (Table 8).

Table 8.

Sub-reliability analyses for vertical jump types based on correlation values.

Study ICC Lower limit Upper Limit Z-value p-value Relative weight (%)
Random Fixed
Sub-analysis for countermovement jump (CMJ)
 Alias et al. (2021) 0.999 0.998 0.999 30.873 0.000 5.61 5.23
 Balsalobre-Fernández et al. (2015) 0.998 0.997 0.999 20.56 0.000 5.50 2.69
 Bishop et al. (2022a) 0.997 0.993 0.999 15.92 0.000 5.41 1.90
 Bishop et al. (2022b) 0.954 0.923 0.973 13.79 0.000 5.58 4.28
 Bogataj et al. (2020a) 0.953 0.936 0.965 23.80 0.000 5.68 13.00
 Bogataj et al. (2020b) 0.947 0.921 0.965 17.10 0.000 5.64 7.13
 Brooks et al. (2018) 0.983 0.969 0.990 16.07 0.000 5.56 3.65
 Carlos-Vivas et al. (2018) 0.983 0.968 0.991 14.47 0.000 5.52 2.93
 Chow et al. (2023) 0.870 0.789 0.922 9.81 0.000 5.58 4.28
 Cruvinel-Cabral et al. (2018 0.948 0.904 0.972 11.17 0.000 5.52 3.01
 Driller et al. (2017) 0.970 0.950 0.982 15.93 0.000 5.59 4.60
 Gallardo-Fuentes et al. (2016) 0.990 0.983 0.994 19.45 0.000 5.58 4.28
 Gür and Ayan (2023) 0.990 0.977 0.996 12.13 0.000 5.36 1.66
 Jimenez-Olmedo et al. (2022) 0.990 0.984 0.994 22.46 0.000 5.62 5.71
 Patiño-Palma et al. (2022) 0.993 0.990 0.995 30.43 0.000 5.66 9.19
 Soares et al. (2023) 0.950 0.879 0.980 7.77 0.000 5.31 1.43
 Stanton et al. (2017) 0.990 0.983 0.994 19.09 0.000 5.58 4.12
 Yingling et al. (2018) 0.748 0.690 0.797 15.74 0.000 5.70 20.92
 Fixed effect model 0.969 0.965 0.972 73.738 0.000
 Random effect model 0.982 0.961 0.992 11.761 0.000
Sub-analysis for squat jump (SJ)
 Bogataj et al. (2020a) 0.970 0.954 0.980 18.947 0.000 34.11 45.30
 Bogataj et al. (2020b) 0.880 0.795 0.931 9.23 0.000 32.69 24.86
 Gallardo-Fuentes et al. (2016) 0.984 0.973 0.991 17.74 0.000 33.20 29.83
 Fixed effect model 0.965 0.953 0.974 27.043 0.000
 Random effect model 0.961 0.889 0.987 7.004 0.000
Sub-analysis for drop jump (DJ)
 Barbalho et al. (2021) 1.000 1.000 1.000 14.006 0.000 18.05 4.28
 Bishop et al. (2022b) 0.970 0.950 0.983 15.427 0.000 20.68 28.88
 Gallardo-Fuentes et al. (2016) 0.913 0.856 0.948 11.358 0.000 20.68 28.88
 Haynes et al. (2019) 0.908 0.799 0.959 7.10 0.000 19.95 11.76
 Stanton et al. (2017) 0.990 0.983 0.994 18.53 0.000 20.63 26.20
 Fixed effect model 0.972 0.964 0.979 29.210 0.000
 Random effect model 0.987 0.940 0.997 6.424 0.000

For studies that utilized Force Plates for Criterion Measurement, the fixed effect model showed an ICC of 0.989 (95% CI 0.986–0.991, p < 0.001), and the random effect model indicated an ICC of 0.993 (95% CI 0.981–0.997, p < 0.001). For studies that did not utilize Force Plates for Criterion Measurement, the fixed effect model showed an ICC of 0.960 (95% CI 0.955–0.964, p < 0.001), and the random effect model indicated an ICC of 0.972 (95% CI 0.936–0.988, p < 0.001). These findings suggest that both methods, whether utilizing force plates or not, are highly reliable for the measurements they aim to assess, as evidenced by the consistently high ICCs across studies (Table 9).

Table 9.

Sub-reliability analyses for criterion device based on correlation values.

Study ICC Lower limit Upper Limit Z-value p-value Random Fixed
Sub-analysis focused on studies utilizing force plates for criterion measurement
 Balsalobre-Fernández et al. (2015) 0.998 0.997 0.999 20.556 0.000 12.71 10.30
 Barbalho et al. (2021) 1.000 1.000 1.000 14.01 0.000 10.60 2.42
 Bishop et al. (2022a) 0.997 0.993 0.999 15.92 0.000 12.40 7.27
 Brooks et al. (2018) 0.983 0.969 0.990 16.07 0.000 12.92 13.94
 Carlos-Vivas et al. (2018) 0.983 0.968 0.991 14.47 0.000 12.78 11.21
 Driller et al. (2017) 0.970 0.950 0.982 15.93 0.000 13.04 17.58
 Haynes et al. (2019) 0.908 0.799 0.959 7.10 0.000 12.30 6.67
 Stanton et al. (2017) 0.990 0.985 0.993 26.60 0.000 13.25 30.61
 Fixed effect model 0.989 0.986 0.991 47.149 0.000
 Random effect model 0.993 0.981 0.997 11.337 0.000
Sub-analysis focused on studies not utilizing force plates for criterion measurement
 Alias et al. (2021) 0.999 0.998 0.999 30.873 0.000 8.37 5.08
 Bishop et al. (2022b) 0.963 0.947 0.975 20.66 0.000 8.45 8.31
 Bogataj et al. (2020a) 0.959 0.948 0.968 30.37 0.000 8.53 18.92
 Bogataj et al. (2020b) 0.930 0.904 0.950 19.29 0.000 8.48 10.38
 Chow et al. (2023) 0.870 0.789 0.922 9.81 0.000 8.32 4.15
 Cruvinel-Cabral et al. (2018) 0.948 0.904 0.972 11.17 0.000 8.21 2.92
 Gallardo-Fuentes et al. (2016) 0.976 0.967 0.982 28.03 0.000 8.50 12.46
 Gür and Ayan (2023) 0.990 0.977 0.996 12.13 0.000 7.93 1.62
 Jimenez-Olmedo et al. (2022) 0.990 0.984 0.994 22.46 0.000 8.39 5.54
 Patiño-Palma et al. (2022) 0.993 0.990 0.995 30.43 0.000 8.46 8.92
 Soares et al. (2023) 0.950 0.879 0.980 7.77 0.000 7.83 1.38
 Yingling et al. (2018) 0.748 0.690 0.797 15.74 0.000 8.53 20.31
 Fixed effect model 0.960 0.955 0.964 70.066 0.000
 Random effect model 0.972 0.936 0.988 9.656 0.000

The sub-analysis revealed high levels of reliability across different contexts. For inter-rater reliability, the fixed effect model showed an ICC of 0.993 (95% CI 0.991–0.994, p < 0.001), and the random effect model indicated an ICC of 0.996 (95% CI 0.987–0.999, p < 0.001). In terms of intra-rater reliability, the fixed effect model revealed an ICC of 0.995 (95% CI 0.993–0.996, p < 0.001), and the random effect model showed an ICC of 0.997 (95% CI 0.990–0.999, p < 0.001). For inter-session reliability, both the fixed and random effect models showed an ICC of 0.970 (95% CI 0.954–0.981, p < 0.001). Lastly, within session/device reliability had a fixed effect model ICC of 0.946 (95% CI 0.940–0.952, p < 0.001) and a random effect model ICC of 0.973 (95% CI 0.945–0.987, p < 0.001). These consistently high ICCs across studies suggest that the methods used for assessing different aspects of reliability are highly robust (Table 10).

Table 10.

Sub-reliability analyses for reliability types based on correlation values.

Study ICC Lower limit Upper Limit Z-value p-value Relative weight (%)
Random Fixed
Sub-analysis for inter-rater reliability
 Alias et al. (2021) 0.999 0.998 1.000 17.825 0.000 14.20 7.53
 Balsalobre-Fernández et al. (2015) 0.998 0.997 0.999 20.56 0.000 14.61 11.64
 Barbalho et al. (2021) 1.000 1.000 1.000 14.01 0.000 12.47 2.74
 Driller et al. (2017) 0.970 0.950 0.982 15.93 0.000 14.94 19.86
 Jimenez-Olmedo et al. (2022) 0.990 0.981 0.995 15.88 0.000 14.65 12.33
 Patiño-Palma et al. (2022) 0.993 0.990 0.995 30.43 0.000 15.18 39.73
 Soares et al. (2023) 0.950 0.879 0.980 7.77 0.000 13.96 6.16
 Fixed effect model 0.993 0.991 0.994 48.014 0.000
 Random effect model 0.996 0.987 0.999 10.712 0.000
Sub-analysis for intra-rater reliability
 Alias et al. (2021) 0.999 0.998 1.000 17.825 0.000 20.17 15.94
 Barbalho et al. (2021) 1.000 1.000 1.000 14.01 0.000 17.68 5.80
 Jimenez-Olmedo et al. (2022) 0.990 0.981 0.995 15.88 0.000 20.82 26.09
 Brooks et al. (2018) 0.990 0.977 0.996 12.69 0.000 20.24 16.67
 Stanton et al. (2017) 0.990 0.983 0.994 18.53 0.000 21.10 35.51
 Fixed effect model 0.995 0.993 0.996 34.821 0.000
 Random effect model 0.997 0.990 0.999 9.726 0.000
Sub-analysis for inter-session reliability
 Brooks et al. (2018) 0.970 0.933 0.987 10.034 0.000 29.87 29.87
 Gallardo-Fuentes et al. (2016) 0.970 0.949 0.982 15.38 0.000 70.13 70.13
 Fixed effect model 0.970 0.954 0.981 18.361 0.000
 Random effect model 0.970 0.954 0.981 18.361 0.000
Sub-analysis for within session/device reliability
 Alias et al. (2021) 0.999 0.998 1.000 17.825 0.000 7.33 1.95
 Gallardo-Fuentes et al. (2016) 0.978 0.969 0.985 23.45 0.000 7.93 9.55
 Stanton et al. (2017) 0.990 0.983 0.994 19.09 0.000 7.75 4.60
 Bishop et al. (2022a) 0.997 0.993 0.999 15.92 0.000 7.39 2.12
 Bishop et al. (2022b) 0.963 0.947 0.975 20.66 0.000 7.93 9.55
 Bogataj et al. (2020a) 0.959 0.948 0.968 30.37 0.000 8.02 21.75
 Bogataj et al. (2020b) 0.930 0.904 0.950 19.29 0.000 7.96 11.94
 Carlos-Vivas et al. (2018) 0.983 0.968 0.991 14.47 0.000 7.62 3.27
 Chow et al. (2023) 0.870 0.789 0.922 9.81 0.000 7.77 4.77
 Cruvinel-Cabral et a. (2018) 0.948 0.904 0.972 11.17 0.000 7.64 3.36
 Gür and Ayan, (2023) 0.990 0.977 0.996 12.13 0.000 7.30 1.86
 Haynes et al. (2019) 0.908 0.799 0.959 7.10 0.000 7.33 1.95
 Yingling et al. (2018) 0.748 0.690 0.797 15.74 0.000 8.03 23.34
 Fixed effect model 0.946 0.940 0.952 60.414 0.000
 Random effect model 0.973 0.945 0.987 11.489 0.000

Discussion

In this systematic review and meta-analysis, the validity and reliability findings of the My Jump smartphone application, which is designed to measure vertical jump performance, were summarized using meta-analytical methods. This review summarized the findings of 21 studies consisting of 839 accumulated participants. Overall methodological quality assessment for individual studies included in this meta-analysis was considered as moderate-to-high quality. Further analyses showed significant heterogeneity scores; thus, the pooled calculations were interpreted according to the random-effects model. For validity, meta-analysis results revealed that there was a raw score agreement between My Jump and the criterion measures, based on nonsignificant Hedge's g values as well as a high consistency of the within-group rankings, based on the pooled correlation result. For reliability, our meta-analysis showed near-perfect reliability for My Jump, based on the pooled ICC value. Additionally, sub-analyses suggested that the results were robust across different types of jumps, reference devices used, and types of reliability.

In fact, unlike the usual study designs that investigate the validity and the test–retest reliability of athletic performance measures, validity, and reliability analyses of the My Jump application can be completed without the need for re-testing14,45,46. As the participant performs a vertical jump once on a platform, the nature of which is accepted as the criterion (force plate, mat, photocell sensors, etc.), a video can be simultaneously recorded12,13,47. Thus, possible biases can be attributed to other factors not regarding participants. For example, because take-off and landing points are manually marked, minor variations are likely when a rater measures the same vertical jump performance consecutively. Or, in a video recording measuring a single jump performance, two raters may mark take-off and landing points differently. However, the My Jump provides a very functional method to minimize these errors, as it offers the possibility to pause the video and play it frame by frame12,13,47. Additionally, the formula it uses (h = t2 × 1.22625)48 is equivalent to most criterion devices. In this case, the major handicap seems to be small variations that can arise from manually determining the take-off and landing points.

While the My Jump application relies on flight time to calculate jump height, force platforms are often considered the gold standard due in part to their ability to calculate jump height based on the impulse-momentum theorem, which takes into account the total force applied during the jump and the duration of this force, providing a more holistic assessment15,16. However, it's worth noting that flight time-based calculations are also commonly used in force platforms. In fact, in the initial study introducing the My Jump, a force platform was used as a reference device, and it too employed a flight time-based methodology for the sake of comparison47. This methodological overlap offers some advantages. For instance, strategies that could artificially lengthen flight time are also applicable to force platform measurements based on flight time15. Therefore, in scenarios where the device rather than the method serves as the reference for jump performance, My Jump appears to be a viable alternative. The primary objective of our study is to test whether My Jump can serve as an alternative to more expensive and less portable devices. Our findings suggest that My Jump can be reliably used for practical applications such as ranking the jump performance of members within a group or tracking an athlete's jump performance over time, provided that the same methodology is consistently applied.

The importance of a high sampling rate is indeed critical for the My Jump app, which utilizes video recordings to calculate jump metrics3,19,47. In the study introducing this smartphone application, an iPhone 5 s was employed, featuring a 120 Hz high-speed camera at a quality of 720p, deemed adequate for such calculations47. Moreover, newer models of the device offer even more advanced capabilities3,49,50. When calculating jump height based on flight time, consider an average jump height of, for instance, 30 cm. The time an athlete would spend airborne for such a jump is approximately half a second. With a 120 Hz high-speed camera, this duration would translate into 60 frames (120 frames/second * 0.5 s = 60 frames). This high frame rate can allow for an accurate and reliable calculation of flight time, thus offering a valid measurement of jump height. What appears to be crucial is the adoption of a standardized procedure for selecting the frames where the jump starts and ends, ensuring consistency and reliability in measurements1,51,52.

Measurement errors related to instruments and raters in vertical jump assessments by smartphone applications can be caused to miss clinically crucial changes in performance1,2,8. However, while all of the studies3,1214,19,4547,49,5355 that evaluated the validity and reliability of the My Jump included designs in which vertical jump performance was measured with two devices at the same time, no study had a comprehensive design containing the vertical jump height differences between two scores by the same rater and between the scores of two different raters, from the same video recording. Although some studies reported the comparison results related to differences between successive jumps performed several minutes apart or the differences between two test days55,56, possible errors in these designs can be attributed to the participants. For an application such as My Jump, in which data can be collected simultaneously along with the criterion device, it is more critical to focus on errors between raters, between devices, and between the same participant’s scores, rather than participant-sourced factors. Presenting a pooled reliability score using all reported ICC scores of original studies since they did not consistently provide ICC reports can be considered as a limitation for the present systematic review and meta-analysis.

Twenty studies3,1214,19,4547,4955,5761 using force plate, contact mat, and photocell system to examine the validity and reliability of the My Jump reported high (ICC > 0.80) reliability scores. However, one study14 compared vertical jump heights obtained from Vertec with the heights obtained from the My Jump and found the ICC score for absolute agreement to be 0.665. Although studies are showing that Vertec offers valid and reliable results62,63, considering the results of the other studies using more valid criterion methods, it seems highly likely that the inconsistency between the scores from two methods is related to the linear position transducers method used by Yingling et al.14. In addition, the fact that individual studies comprising participants that represent a wide range of the population, such as healthy adults, athletes, both men and women, children, and the elderly, strengthen the competence of this smartphone application to produce valid and reliable outputs. Consequently, the present systematic review and meta-analysis showed that the My Jump presented high agreement and consistency scores with the force plate, contact mat, and photocell systems as reference methods, demonstrating a pooled nearly perfect reliability score. In addition to its low-cost and simplicity, the My Jump smartphone application could be considered a valid and reliable method of assessing vertical jump height in various populations (Supplementary Information).

Conclusions

This is the first investigation using meta-analytical methods to confirm the validity and reliability of the My Jump smartphone application to measure vertical jump heights. In terms of validity, meta-analysis results revealed that there was a raw score agreement between My Jump and the criterion measures, based on nonsignificant Hedge’s g values as well as a high consistency of the within-group rankings, based on the pooled correlation result. In terms of reliability, our present meta-analysis showed near-perfect reliability for My Jump, based on the pooled ICC value. Data from this systematic review and meta-analysis suggests that the My Jump can be used for assessing and monitoring vertical jump performance, which is a parameter included in global physical fitness test batteries and which provides information about the neuromuscular function and explosive power of the lower body. However, included studies mostly targeted on adults, only one study focused on children. More research need to be conducted on this population to precisely ensure the validity and reliability of My Jump smartphone application.

Practical application

The findings of this systematic review and meta-analysis offer several practical implications for sports scientists, strength and conditioning practitioners, and coaches. The My Jump smartphone application provides a cost-effective and portable alternative to traditional laboratory equipment, making it accessible and convenient for teams or organizations with budget constraints. Its high validity and reliability make it a trustworthy tool for assessing an athlete's neuromuscular function and explosive power. While the application has been primarily validated in adult populations, its potential applicability across different age groups, including children, suggests a broader utility, although more research is needed in this area. The applications mobile-based platform allows for real-time monitoring and immediate feedback during training sessions, facilitating data-driven adjustments to training programs. Overall, the My Jump serves as a reliable and valid tool for accurately assessing vertical jump performance, offering a cost-effective and accessible means for data collection in various settings.

Supplementary Information

Author contributions

S.U. and C.G. performed the statistical analysis. M.T., S.U., S.Ö., C.G., S.A., and B.Ç.S. performed the literature search, collected the data, wrote the manuscript, and made edits. CG, S.U. and M.T. were mainly responsible for the interpretation of the data and preparing the final version. İ.İ. created the figures. All authors provided critical feedback and contributed to the final manuscript. Correspondence and requests for materials should be addressed to S.U.

Data availability

The datasets used and/or analyzed during the current study available from the corresponding author on reasonable request.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

The online version contains supplementary material available at 10.1038/s41598-023-46935-x.

References

  • 1.Sharp AP, Cronin JB, Neville J. Using smartphones for jump diagnostics: A brief review of the validity and reliability of the my jump app. Strength Cond. J. 2019;41(5):96–107. doi: 10.1519/SSC.0000000000000472. [DOI] [Google Scholar]
  • 2.Silva R, et al. Validity and reliability of mobile applications for assessing strength, power, velocity, and change-of-direction: A systematic review. Sensors. 2021;21(8):2623. doi: 10.3390/s21082623. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Brooks ER, Benson AC, Bruce LM. Novel technologies found to be valid and reliable for the measurement of vertical jump height with jump-and-reach testing. J. Strength Cond. Res. 2018;32(10):2838–2845. doi: 10.1519/JSC.0000000000002790. [DOI] [PubMed] [Google Scholar]
  • 4.Ulupinar S, Özbay S. An easy-to-apply series of field test for physical education teachers in an educational setting: ALPHA test battery. J. Pedag. Res. 2020;4(3):262–271. [Google Scholar]
  • 5.Currell K, Jeukendrup AE. Validity, reliability and sensitivity of measures of sporting performance. Sports Med. 2008;38(4):297–316. doi: 10.2165/00007256-200838040-00003. [DOI] [PubMed] [Google Scholar]
  • 6.Mijnarends DM, et al. Validity and reliability of tools to measure muscle mass, strength, and physical performance in community-dwelling older people: A systematic review. J. Am. Med. Dir. Assoc. 2013;14(3):170–178. doi: 10.1016/j.jamda.2012.10.009. [DOI] [PubMed] [Google Scholar]
  • 7.Pecorelli N, et al. An app for patient education and self-audit within an enhanced recovery program for bowel surgery: A pilot study assessing validity and usability. Surg. Endosc. 2018;32(5):2263–2273. doi: 10.1007/s00464-017-5920-3. [DOI] [PubMed] [Google Scholar]
  • 8.Koo T, Li M. A guideline of selecting and reporting intraclass correlation coefficients for reliability research. J. Chiropr. Med. 2016. 2020;15(2):155–163. doi: 10.1016/j.jcm.2016.02.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Mundy PM, Clarke ND. Performance Assessment in Strength and Conditioning. Routledge; 2018. Reliability, validity and measurement error; pp. 23–33. [Google Scholar]
  • 10.Hopkins WG, Schabort EJ, Hawley JA. Reliability of power in physical performance tests. Sports Med. 2001;31(3):211–234. doi: 10.2165/00007256-200131030-00005. [DOI] [PubMed] [Google Scholar]
  • 11.Ulupinar S, Özbay S, Gençoğlu C. Counter movement jump and sport specific frequency speed of kick test to discriminate between elite and sub-elite kickboxers. Acta Gymnica. 2021;50(4):141–146. doi: 10.5507/ag.2020.019. [DOI] [Google Scholar]
  • 12.Haynes T, et al. The validity and reliability of the My Jump 2 app for measuring the reactive strength index and drop jump performance. J. Sports Med. Phys. Fit. 2019 doi: 10.23736/S0022-4707.18.08195-1. [DOI] [PubMed] [Google Scholar]
  • 13.Bogataj Š, et al. Concurrent validity and reliability of my jump 2 app for measuring vertical jump height in recreationally active adults. Appl. Sci. 2020;10(11):3805. doi: 10.3390/app10113805. [DOI] [Google Scholar]
  • 14.Yingling VR, et al. The reliability of vertical jump tests between the Vertec and My Jump phone application. PeerJ. 2018;6:e4669. doi: 10.7717/peerj.4669. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Attia A, et al. Measurement errors when estimating the vertical jump height with flight time using photocell devices: The example of Optojump. Biol. Sport. 2017;34(1):63–70. doi: 10.5114/biolsport.2017.63735. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Xu J, et al. A systematic review of the different calculation methods for measuring jump height during the countermovement and drop jump tests. Sports Med. 2023;53(5):1055–1072. doi: 10.1007/s40279-023-01828-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Moir GL. Three different methods of calculating vertical jump height from force platform data in men and women. Meas. Phys. Educ. Exerc. Sci. 2008;12(4):207–218. doi: 10.1080/10913670802349766. [DOI] [Google Scholar]
  • 18.Wade L, Lichtwark GA, Farris DJ. Comparisons of laboratory-based methods to calculate jump height and improvements to the field-based flight-time method. Scand. J. Med. Sci. Sports. 2020;30(1):31–37. doi: 10.1111/sms.13556. [DOI] [PubMed] [Google Scholar]
  • 19.Stanton R, Kean CO, Scanlan AT. My Jump for vertical jump assessment. Br. J. Sports Med. 2015;49(17):1157–1158. doi: 10.1136/bjsports-2015-094831. [DOI] [PubMed] [Google Scholar]
  • 20.Bui HT, et al. Comparison and analysis of three different methods to evaluate vertical jump height. Clin. Physiol. Funct. Imaging. 2015;35(3):203–209. doi: 10.1111/cpf.12148. [DOI] [PubMed] [Google Scholar]
  • 21.Morin J-B, et al. When jump height is not a good indicator of lower limb maximal power output: Theoretical demonstration, experimental evidence and practical solutions. Sports Med. 2019;49:999–1006. doi: 10.1007/s40279-019-01073-1. [DOI] [PubMed] [Google Scholar]
  • 22.Walsh MS, et al. The validation of a portable force plate for measuring force-time data during jumping and landing tasks. J. Strength Cond. Res. 2006;20(4):730. doi: 10.1519/R-18225.1. [DOI] [PubMed] [Google Scholar]
  • 23.Cabarkapa DV, et al. Impact of the anatomical accelerometer placement on vertical jump performance characteristics. Sports. 2023;11(4):92. doi: 10.3390/sports11040092. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Patiño-Palma BE, Wheeler-Botero CA, Ramos-Parrací CA. Validation and reliability of the wheeler jump sensor for the execution of the countermovement jump. Apunts. Educació Física i Esports. 2022;149:37–44. [Google Scholar]
  • 25.Šarabon N, Kozinc Ž, Marković G. Force–velocity profile during vertical jump cannot be assessed using only bodyweight jump and isometric maximal voluntary contraction tasks. Sci. Rep. 2020;10(1):1–12. doi: 10.1038/s41598-020-76262-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Eekhoff JD, Iannucci LE, Lake SP. Shear mechanosensing drives tendon adaptation. Nat. Biomed. Eng. 2021;5(12):1409–1410. doi: 10.1038/s41551-021-00724-x. [DOI] [PubMed] [Google Scholar]
  • 27.Özbay S, Ulupınar S. Strength-power tests are more effective when performed after exhaustive exercise in discrimination between top-elite and elite wrestlers. J. Strength Cond. Res. 2022;36(2):448–454. doi: 10.1519/JSC.0000000000003456. [DOI] [PubMed] [Google Scholar]
  • 28.Kemper HC, Van Mechelen W. Physical fitness testing of children: A European perspective. Pediatr. Exerc. Sci. 1996;8(3):201–214. doi: 10.1123/pes.8.3.201. [DOI] [Google Scholar]
  • 29.Ruiz JR, et al. Field-based fitness assessment in young people: The ALPHA health-related fitness test battery for children and adolescents. Br. J. Sports Med. 2011;45(6):518–524. doi: 10.1136/bjsm.2010.075341. [DOI] [PubMed] [Google Scholar]
  • 30.Morrow JR, Jr, Martin SB, Jackson AW. Reliability and validity of the FITNESSGRAM®: Quality of teacher-collected health-related fitness surveillance data. Res. Q. Exerc. Sport. 2010;81(sup3):S24–S30. doi: 10.1080/02701367.2010.10599691. [DOI] [PubMed] [Google Scholar]
  • 31.Higgins JP, et al. Cochrane Handbook for Systematic Reviews of Interventions. Wiley; 2019. [Google Scholar]
  • 32.Moher D, et al. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. PLoS Med. 2009;6(7):e1000097. doi: 10.1371/journal.pmed.1000097. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Downs SH, Black N. The feasibility of creating a checklist for the assessment of the methodological quality both of randomised and non-randomised studies of health care interventions. J. Epidemiol. Community Health. 1998;52(6):377–384. doi: 10.1136/jech.52.6.377. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Borenstein, M. et al. Comprehensive meta-analysis, version 2 biostat (2005).
  • 35.Higgins JP, et al. Measuring inconsistency in meta-analyses. BMJ. 2003;327(7414):557–560. doi: 10.1136/bmj.327.7414.557. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Cohen J. A power primer. Psychol. Bull. 1992;112(1):155. doi: 10.1037/0033-2909.112.1.155. [DOI] [PubMed] [Google Scholar]
  • 37.Barnett LM, et al. Correlates of gross motor competence in children and adolescents: A systematic review and meta-analysis. Sports Med. 2016;46(11):1663–1688. doi: 10.1007/s40279-016-0495-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Kramer MS, Feinstein AR. Clinical biostatistics: LIV. The biostatistics of concordance. Clin. Pharmacol. Ther. 1981;29(1):111–123. doi: 10.1038/clpt.1981.18. [DOI] [PubMed] [Google Scholar]
  • 39.Brown DJ, Fletcher D. Effects of psychological and psychosocial interventions on sport performance: A meta-analysis. Sports Med. 2017;47(1):77–99. doi: 10.1007/s40279-016-0552-7. [DOI] [PubMed] [Google Scholar]
  • 40.Cochran WG. The combination of estimates from different experiments. Biometrics. 1954;10(1):101–129. doi: 10.2307/3001666. [DOI] [Google Scholar]
  • 41.Ulupınar S, et al. Performance differences between Greco-roman and freestyle wrestlers: A systematic review and meta-analysis. J. Strength Cond. Res. 2021;35(11):3270–3279. doi: 10.1519/JSC.0000000000004129. [DOI] [PubMed] [Google Scholar]
  • 42.Borenstein M, et al. Basics of meta-analysis: I2 is not an absolute measure of heterogeneity. Res. Synth. Methods. 2017;8(1):5–18. doi: 10.1002/jrsm.1230. [DOI] [PubMed] [Google Scholar]
  • 43.Wang DX, et al. Muscle mass, strength, and physical performance predicting activities of daily living: A meta-analysis. J. Cachexia Sarcopenia Muscle. 2020;11(1):3–25. doi: 10.1002/jcsm.12502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Egger M, et al. Bias in meta-analysis detected by a simple, graphical test. BMJ. 1997;315(7109):629–634. doi: 10.1136/bmj.315.7109.629. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Carlos-Vivas J, et al. Validation of the iPhone app using the force platform to estimate vertical jump height. J. Sports Med. Phys. Fit. 2016;58(3):227–232. doi: 10.23736/S0022-4707.16.06664-0. [DOI] [PubMed] [Google Scholar]
  • 46.Cruvinel-Cabral RM, et al. The validity and reliability of the “My Jump App” for measuring jump height of the elderly. PeerJ. 2018;6:e5804. doi: 10.7717/peerj.5804. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Balsalobre-Fernández C, Glaister M, Lockey RA. The validity and reliability of an iPhone app for measuring vertical jump performance. J. Sports Sci. 2015;33(15):1574–1579. doi: 10.1080/02640414.2014.996184. [DOI] [PubMed] [Google Scholar]
  • 48.Bosco C, Luhtanen P, Komi PV. A simple method for measurement of mechanical power in jumping. Eur. J. Appl. Physiol. Occup. Physiol. 1983;50(2):273–282. doi: 10.1007/BF00422166. [DOI] [PubMed] [Google Scholar]
  • 49.Bogataj Š, et al. Validity, reliability, and usefulness of My Jump 2 App for measuring vertical jump in primary school children. Int. J. Environ. Res. Public Health. 2020;17(10):3708. doi: 10.3390/ijerph17103708. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Fatih G, Vedat A. My Jump 2 mobil uygulamasinin geçerlilik ve güvenilirlik analizi. Spor Bilimleri Araştırmaları Dergisi. 2023;8(1):127–135. doi: 10.25307/jssr.1192168. [DOI] [Google Scholar]
  • 51.Jimenez-Olmedo JM, et al. Reliability of My Jump 2 derived from crouching and standing observation heights. Int. J. Environ. Res. Public Health. 2022;19(16):9854. doi: 10.3390/ijerph19169854. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Soares D, et al. Validity and reliability of My Jump 2 app for jump performance in Judo players. Open Sports Sci. J. 2023 doi: 10.2174/1875399X-v16-e230714-2023-6. [DOI] [Google Scholar]
  • 53.Barbalho M, et al. Assessing interlimb jump asymmetry in young soccer players: The my jump 2 APP. Int. J. Sports Physiol. Perform. 2020;16(1):19–27. doi: 10.1123/ijspp.2019-0981. [DOI] [PubMed] [Google Scholar]
  • 54.Driller M, et al. Assessing a smartphone application to measure counter-movement jumps in recreational athletes. Int. J. Sports Sci. Coach. 2017;12(5):661–664. doi: 10.1177/1747954117727846. [DOI] [Google Scholar]
  • 55.Gallardo-Fuentes F, et al. Intersession and intrasession reliability and validity of the My Jump app for measuring different jump actions in trained male and female athletes. J. Strength Cond. Res. 2016;30(7):2049–2056. doi: 10.1519/JSC.0000000000001304. [DOI] [PubMed] [Google Scholar]
  • 56.Brooks GA, et al. Temperature, skeletal muscle mitochondrial functions, and oxygen debt. Am. J. Physiol. Leg. Content. 1971;220(4):1053–1059. doi: 10.1152/ajplegacy.1971.220.4.1053. [DOI] [PubMed] [Google Scholar]
  • 57.Alias MF, et al. Inter-rater reliability and intra-rater reliability testing of my jump 2 mobile application in measuring countermovement jump. J. Hum. Mov. Sports Sci. 2021;9(6):1319–1323. [Google Scholar]
  • 58.Chow GC-C, Kong Y-H, Pun W-Y. The concurrent validity and test-retest reliability of possible remote assessments for measuring countermovement jump: My jump 2, HomeCourt & Takei vertical jump meter. Appl. Sci. 2023;13(4):2142. doi: 10.3390/app13042142. [DOI] [Google Scholar]
  • 59.Bishop C, et al. Jump and change of direction speed asymmetry using smartphone apps: Between-session consistency and associations with physical performance. J. Strength Cond. Res. 2022;36(4):927–934. doi: 10.1519/JSC.0000000000003567. [DOI] [PubMed] [Google Scholar]
  • 60.Bishop C, et al. Validity and reliability of strategy metrics to assess countermovement jump performance using the newly developed smartphone application. J. Hum. Kinet. 2022;83(1):185–195. doi: 10.2478/hukin-2022-0098. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Plakoutsis G, et al. Validity and reliability of the portable Kforce plates system with the use of a smartphone application for measuring countermovement jump. Gait Posture. 2023;106:S167–S168. doi: 10.1016/j.gaitpost.2023.07.201. [DOI] [Google Scholar]
  • 62.Mahmoud I, et al. The reliability of a real time wearable sensing device to measure vertical jump. Procedia Eng. 2015;112:467–472. doi: 10.1016/j.proeng.2015.07.226. [DOI] [Google Scholar]
  • 63.Nuzzo JL, Anning JH, Scharfenberg JM. The reliability of three devices used for measuring vertical jump height. J. Strength Cond. Res. 2011;25(9):2580–2590. doi: 10.1519/JSC.0b013e3181fee650. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

The datasets used and/or analyzed during the current study available from the corresponding author on reasonable request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES