How is Accuracy Measured?

Discussion in 'General Clock Discussions' started by SpaceCowboy850, Nov 9, 2019.

1. SpaceCowboy850 Registered User

Oct 29, 2019
8
0
1
Male
Country Flag:
#1
I've been exploring mechanical and quartz designs a bit as of late, and stumbled upon this:

Burgess Clock B
Burgess Clock B, The World's Most Precise Pendulum Clock, Is Made To A 250-Year-Old Design By John Harrison, Longitude Prize Winner And Inventor Of The Marine Chronometer | Quill & Pad

A mechanical pendulum clock that loses a bit over half a second over the course of 100 days.

Then there is this
World's Most Accurate Clock will Lose One Second Every 14 Billion Years | Digital Trends

Which is an atomic clock that loses 1 second every 14 billion years.

So the question for both is...how do they know?

In the case of the mechanical clock, it was designed 250 years ago, and would have been the most precise time keeping measure at the time, so how could you know if it is losing half a second or not? Likewise with the 1 second in 14 billion years? Is it based on mathematics of what a "perfect" tick would look like and where the pendulum should be given a mathematically perfect tick? Is it based on calculations of where the sun/moon/stars should be, once again, given equations and seeing how that compares to what the respective clocks give?

2. Willie X Registered User

Feb 9, 2008
12,006
698
113
#2
I'm careful to never use the word "perfect" in any endeavor.

You can buy an Atomic slave for 20 bucks.

Willie X

3. shutterbug Moderator NAWCC Member

Oct 19, 2005
40,775
722
113
Male
Self employed interpreter/clock repairer
Iowa
Country Flag:
Region Flag:
#3
14 billion years is a pretty safe claim, since nothing mechanical can last that long. 100 years might be a better point for advertising