yeah. definitely see the utility of the au (for near-scale intra-solar system measurements, comparisons), sounds like its useful to astronomers, which makes sense, of course: lightyears, for their workflow, would require a conversion :
If, for example, we observe a star in January, and then look at it again in July, the Earth will have gone halfway around its orbit. We’re looking at the star from two locations around 200 million miles (~300 million km) apart. If the star is reasonably close, then – from one side of Earth’s orbit to the other – it will appear to move ever so slightly.
Add some trigonometry, and the parallax angle, combined with the size of Earth’s orbit, lets astronomers calculate the distance to the star.
These angles are miniscule. They’re too small for degrees to be a practical unit of measurement. That’s why parallax angles are typically measured in arcseconds – a unit of measurement equivalent to the width of an average human hair seen from 65 feet (20 meters) away – not degrees. There are 3,600 arcseconds in one degree.
And here’s how we arrive at parsecs as a unit of distance: one parsec is the distance to an object whose parallax angle is one arcsecond.
The term parsec is just over 100 years old. It first appeared in a 1913 paper by English astronomer Sir Frank Watson Dyson, and the term stuck. If you see a star with 1/2 arcsecond of parallax, it is two parsecs away. At 1/3 arcsecond, it is three parsecs away. And so on.
Basically, astronomers liked it because it made the math easier!
https://earthsky.org/space/what-is-a-parsec
reply
share