Friday 28 June 2013

It is claimed that two cesium clocks, if allowed to run for 100 years, free from any disturbance, may differ by only about 0.02 s. What does this imply for the accuracy of the standard cesium clock in measuring a time-interval of 1 s?

It is claimed that two cesium clocks, if allowed to run for 100 years, free from any disturbance, may differ by only about 0.02 s. What does this imply for the accuracy of the standard cesium clock in measuring a time-interval of 1 s?
Answer:
Difference in time of caesium clocks = 0.02 s
Time required for this difference = 100 years
= 100 × 365 × 24 × 60 × 60 = 3.15 × 10^9 s
In 3.15 × 109 s, the caesium clock shows a time difference of 0.02 s.
In 1s, the clock will show a time difference of .

Hence, the accuracy of a standard caesium clock in measuring a time interval of 1 s is

1 comment:

  1. Why did you reciprocal the time difference for accuracy?

    ReplyDelete

Please g plus the post