Remix.run Logo
seizethecheese 11 hours ago

> Imagine you could interview thousands of educated individuals from 1913—readers of newspapers, novels, and political treatises—about their views on peace, progress, gender roles, or empire. Not just survey them with preset questions, but engage in open-ended dialogue, probe their assumptions, and explore the boundaries of thought in that moment.

He’ll yeah, sold, let’s go…

> We're developing a responsible access framework that makes models available to researchers for scholarly purposes while preventing misuse.

Oh. By “imagine you could interview…” they didn’t mean me.

pizzathyme 16 minutes ago | parent | next [-]

They did mean you, they just meant "imagine" very literally!

leoedin 5 hours ago | parent | prev | next [-]

It's a shame isn't it! The public must be protected from the backwards thoughts of history. In case they misuse it.

I guess what they're really saying is "we don't want you guys to cancel us".

danielbln 7 hours ago | parent | prev | next [-]

How would one even "misuse" a historical LLM, ask it how to cook up sarine gas in a trench?

DonHopkins 7 hours ago | parent [-]

Ask it to write a document called "Project 2025".

JKCalhoun 3 hours ago | parent | next [-]

"Project 1925". (We can edit the title in post.)

ilaksh 6 hours ago | parent | prev [-]

Well but that wouldn't be misuse, it would be perfect for that.

DGoettlich 3 hours ago | parent | prev | next [-]

understand your frustration. i trust you also understand the models have some dark corners that someone could use to misrepresent the goals of our project. if you have ideas on how we could make the models more broadly accessible while avoiding that risk, please do reach out @ history-llms@econ.uzh.ch

tombh 41 minutes ago | parent | next [-]

Of course, I have to assume that you have considered more outcomes than I have. Because, from my five minutes of reflection as a software geek, albeit with a passion for history, I find this the most surprising thing about the whole project.

I suspect restricting access could equally be a comment on modern LLMs in general, rather than the historical material specifically. For example, we must be constantly reminded not to give LLMs a level of credibility that their hallucinations would have us believe.

But I'm fascinated by the possibility that somehow resurrecting lost voices might give an unholy agency to minds and their supporting worldviews that are so anachronistic that hearing them speak again might stir long-banished evils. I'm being lyrical for dramatic affect!

I would make one serious point though, that do I have the credentials to express. The conversation may have died down, but there is still a huge question mark over, if not the legality, but certainly the ethics of restricting access to, and profiting from, public domain knowledge. I don't wish to suggest a side to take here, just to point out that the lack of conversation should not be taken to mean that the matter is settled.

2 hours ago | parent | prev | next [-]
[deleted]
naasking an hour ago | parent | prev | next [-]

What are the legal or other ramifications of people misrepresenting the goals of your project? What is it you're worried about exactly?

unethical_ban 24 minutes ago | parent | prev [-]

A disclaimer on the site that you are not bigoted or genocidal, and that worldviews from the 1913 era were much different than today and don't necessarily reflect your project.

Movie studios have done that for years with old movies. TCM still shows Birth of a Nation and Gone with the Wind.

Edit: I saw further down that you've already done this! What more is there to do?

BoredPositron 8 hours ago | parent | prev | next [-]

You would get pretty annoyed on how we went backwards in some regards.

speedgoose 8 hours ago | parent [-]

Such as?

JKCalhoun 3 hours ago | parent [-]

Touché.

ImHereToVote 8 hours ago | parent | prev [-]

I wonder how much GPU compute you would need to create a public domain version of this. This would be a really valuable for the general public.

wongarsu 5 hours ago | parent [-]

To get a single knowledge-cutoff they spent 16.5h wall-clock hours on a cluster of 128 NVIDIA GH200 GPUs (or 2100 GPU-hours), plus some minor amount of time for finetuning. The prerelease_notes.md in the repo is a great description on how one would achieve that

IanCal 5 hours ago | parent [-]

While I know there's going to be a lot of complications in this, given a quick search it seems like these GPUs are ~$2/hr, so $4000-4500 if you don't just have access to a cluster. I don't know how important the cluster is here, whether you need some minimal number of those for the training (and it would take more than 128x longer or not be possible on a single machine) or if a cluster of 128 GPUs is a bunch less efficient but faster. A 4B model feels like it'd be fine on one to two of those GPUs?

Also of course this is for one training run, if you need to experiment you'd need to do that more.