this post was submitted on 02 Apr 2024
140 points (95.5% liked)
Videos
14915 readers
392 users here now
For sharing interesting videos from around the Web!
Rules
- Videos only
- Follow the global Mastodon.World rules and the Lemmy.World TOS while posting and commenting.
- Don't be a jerk
- No advertising
- No political videos, post those to [email protected] instead.
- Avoid clickbait titles. (Tip: Use dearrow)
- Link directly to the video source and not for example an embedded video in an article or tracked sharing link.
- Duplicate posts may be removed
Note: bans may apply to both [email protected] and [email protected]
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Moore's law predicts that compared to 1980, computers in 2040 would be a BILLION times faster.
Also that compared to 1994 computers, the ones rolling out now are a MILLION times faster.
A cheap Raspberry PI would easily be able to handle the computational workload of a room full of equipment in 1984.
What would have taken a million years to calculate in 1984 would theoretically take 131 hours today and 29 seconds in 2044...
Correction: Moore's law predicts that the number of transistors on an integrated circuit would double every two years. It doesn't make predictions about computers being "faster" or able to handle a certain "workload". The only thing it predicts is the growth in physical capacity of a single chip.
And we actually broke Moore's law and this capacity growth slowed a decade ago since manufacturing techniques started being the bottleneck.
Yes yes single threaded execution etc but now we just build a crap ton more and keep increasing the computational throughput per watt etc.
We've moved massive calculations into GPUs and thus in terms computational capabilities it holds up.
I mean check this out https://en.wikipedia.org/wiki/FLOPS
The geometric growth is real. Moore's law was just one way to explain it.