@hjhornbeck That's a good qualifier!
To clarify my own comment, I think it's _possible_, but I don't know if it's actually been done successfully (by Swiss AI or anyone else). I feel like a model could be trained pretty well just using older material explicitly in the public domain (e.g. from Project Gutenberg), along with material explicitly made available for it to use. (E.g. our business welcomes models being trained using our reference manuals and support articles and MIT-licensed code.)
