Ask Question
27 December, 08:45

The average estimated hours a person in the United States spent playing video games per year from 2002 to 2012 were 71, 80, 82, 78, 80, 91, 107, 121, 125, 131, and 142. Use the statistics calculator to find the variance and population standard deviation. Round answers to the nearest whole number.

+3
Answers (2)
  1. 27 December, 09:50
    0
    Variance: 636.4181818181823

    Standard Deviation: 25.227330057264926

    Solutions:

    - 71 + 78 + 80 + 80 + 82 + 91 + 107 + 121 + 125 + 131 + 142 = 1108

    - 1108 x 1108 = 1227664

    - 1227664 / 11 = 111605.81818181818

    - (71 x 71) + (78 x 78) + (80 x 80) + (80 x 80) + (82 x 82) + (91 x 91) + (107 x 107) + (121 x 121) + (125 x 125) + (131 x 131) + (142 x 142) = 117970

    - 117970 - 111605.81818181818 = 6364.1818181818235

    - 11 - 1 = 10

    - 6364.1818181818235 / 10 = 636.4181818181823 (Variance)

    - √ (636.4181818181823) = 25.227330057264926 (Standard Deviation)
  2. 27 December, 12:24
    0
    The variance of this data set is 579

    The population standard deviation for this data set is 24
Know the Answer?
Not Sure About the Answer?
Get an answer to your question ✅ “The average estimated hours a person in the United States spent playing video games per year from 2002 to 2012 were 71, 80, 82, 78, 80, 91, ...” in 📙 Mathematics if there is no answer or all answers are wrong, use a search bar and try to find the answer among similar questions.
Search for Other Answers