• 0 Posts
  • 8 Comments
Joined 2 years ago
cake
Cake day: August 4th, 2023

help-circle
  • There’s another factor - days where thr earth is orbiting faster, eg on the closer side of the ellipse - are a different length midday to midday from when we are on the far side of the ellipse.

    You can convince yourself of this when you consider that the area of the arc we traverse each day is the same (Kepler’s law). On the short side of our eliptical orbit, since the orbital distance is shorter, the arc must have a larger angle that we travel. That means the amount a point on the earth rotates to have the sun come back directly overhead must be different in different parts of the year.

    This difference, summed day over day, results in a +/- 20 min movement of actual midday to 12pm. The ‘mean’ in Greenwich Mean Time refers to averaging this difference over the whole orbit.






  • I’m guessing that exactly the same LLM model is used (somehow) on both sides - using different models or different weights would not work at all.

    An LLM is (at core) an algorithm that takes a bunch of text as input and produces an output of a list of word/probabilities such that the sum of all probabilities adds to 1.0. You could place a wrapper on this that creates a list of words by probability. A specific word can be identified by the index in the list, i.e. first word, tenth word etc.

    (Technically the system uses ‘tokens’ which represent either whole words or parts of words, but that’s not important here).

    A document can be compressed by feeding in each word in turn, creating the list in the LLM, and searching for the new word in the list. If the LLM is good, the output will be a stream of small integers. If the LLM is a perfect predictor, the next word will always be the top of the list, i.e. a 1. A bad prediction will be a relatively large number in the thousands or millions.

    Streams of small numbers are very well (even optimally) compressed using extant technology.