Remix.run Logo
drbig 3 days ago

Instruction pipelining and this is exactly why I wish we still have the time to go back to "it is exactly as it is", think the 6502 or any architecture that does not pretend/map/table/proxy/ringaway anything.

That, but a hell lot of it with fast interconnect!

... one can always dream.

ojbyrne 3 days ago | parent | next [-]

The article is essentially describing virtual memory (with enhancements) which predates the 6502 by a decade or so.

Delk 2 days ago | parent [-]

IMO it's not even quite right in its description. The first picture that describes virtual memory shows all processes as occupying the same "logical" address space with the page table just mapping pages in the "logical" address space to physical addresses one-to-one. In reality (at least in all VM systems I know of) each process has its own independent virtual address space.

drbig 2 days ago | parent | prev | next [-]

The point is that we should acknowledged those "cheats" came with their reasons and that they did improve performance etc. But, they also did come with a cost (Meltdown, Spectre anyone?) and fundamentally introduced _complexities_, which at today's level of manufacturing and end of Moore's law may not be the best tradeoffs.

I'm just expressing the general sentiment of distaste for piling stuff upon stuff and holding it with a duct-tape, without ever stepping back and looking at what we have, or at least should have, learnt and where we are today in the technological stack.

eru 2 days ago | parent | prev | next [-]

Do you want to throw out out-of-order-execution and pipelining while you are at it, too?

I'm semi-serious: there are actually modern processor designs that put this burden on the programmer (or rather their fancy compiler / code generator) in order to keep the silicon simple. See eg https://en.wikipedia.org/wiki/Groq#Language_Processing_Unit

taeric 3 days ago | parent | prev | next [-]

I'm curious how this dream is superior to where we are? Yes, things are more complex. But it isn't like this complexity didn't buy us anything. Quite the contrary.

harry8 3 days ago | parent [-]

> ...buy us anything.

Totally depends on who "us" and isn't. What problem is being solved etc. In the aggregate clearly the trade off has been beneficial to the most people. If what you want to do got traded, well you can still dream.

taeric 2 days ago | parent [-]

Right, but that was kind of my question? What is better about not having a lot of these things?

That is, phrasing it as a dream makes it sound like you imagine it would be better somehow. What would be better?

harry8 2 days ago | parent | next [-]

Think about using a modern x86-64 cpu core to run one process with no operating system. Know exactly what is in cache memory. Know exactly what deadlines you can meet and guarantee that.

It's quite a different thing to running a general purpose OS to multiplex each core with multiple processes and a hardware walked page table, TLB etc.

Obviously you know what you prefer for your laptop.

As we get more and more cores perhaps the system designs that have evolved may head back toward that simplicity somewhat? Anything above %x cpu usage gets its own isolated, un-interrupted core(s)? Uses low cost IPC? Hard to speculate with any real confidence.

taeric 2 days ago | parent [-]

I just don't know that I see it running any better for the vast majority of processes that I could imagine running on it. Was literally just transcoding some video, playing a podcast, and browsing the web. Would this be any better?

I think that is largely my qualm with the dream. The only way this really works is if we had never gone with preemptive multitasking, it seems? And that just doesn't seem like a win.

You do have me curious to know if things really do automatically pin to a cpu if it is above a threshold. I know that was talked of some, did we actually start doing that?

harry8 13 hours ago | parent [-]

> Was literally just transcoding some video, playing a podcast, and browsing the web.

Yeah that's the perfect use case for current system design. Nobody sane wants to turn that case into an embedded system running a single process with hard deadline guarantees. Your laptop may not be ideal for controlling a couple of tonnes of steel at high speed, for example. Start thinking about how you would design for that and you'll see the point (whether you want to agree or not).

layer8 2 days ago | parent | prev [-]

Things would be simpler, more predictable and tractable.

For example, real-time guarantees (hard time constraints on how long a particular type of event will take to process) would be easier to provide.

taeric 2 days ago | parent [-]

But why do we think that? The complexity would almost certainly still exist. Would just now be up a layer. With no guarantees that you could hit the same performance characteristics that we are able to hit today.

Put another way, if that would truly be a better place, what is stopping people from building it today?

layer8 2 days ago | parent [-]

Performance wouldn’t be the same, and that’s why nobody is manufacturing it. The industry prefers living with higher complexity when it yields better performance. That doesn’t mean that some people like in this thread wouldn’t prefer if things were more simple, even at the price of significantly lower performance.

> The complexity would almost certainly still exist.

That doesn’t follow. A lot of the complexity is purely to achieve the performance we have.

taeric 2 days ago | parent [-]

I'm used to people arguing for simpler setups because the belief is that they could make them more performant. This was specifically the push for RISC back in the day, no?

To that end, I was assuming the idea would be that we think we could have faster systems if we didn't have this stuff. If that is not the assumption, I'm curious what the appeal is?

layer8 2 days ago | parent [-]

That’s certainly not the assumption here. The appeal is, as I said, that the systems would be more predictable and tractable, instead of being a tarpit of complexity. It would be easier to reason about them, and about their runtime characteristics. Side-channel attacks wouldn’t be a thing, or at least not as much. Nowadays it’s rather difficult to reason about the runtime characteristics of code on modern CPUs, about what exactly will be going on behind the scenes. More often than not, you have to resort to testing how specific scenarios will behave, rather than being able to predict the general case.

taeric 2 days ago | parent [-]

I guess I don't know that I understand why you would dream of this, though? Just go out and program on some simpler systems? Retro computing makes the rounds a lot and is perfectly doable.

loeg 3 days ago | parent | prev [-]

But why?