-
@McaleerStephen @drjwrae This is cool. Got me thinking that maybe LLM weights are “lossy ‘compression’” of training data in some ways like how 10,000 pages of the Journal of Shakespeare Studies is a lossy compression of Shakespeare’s 1300-page collected works.