this post was submitted on 12 Sep 2024
54 points (87.5% liked)

PC Gaming

8556 readers
470 users here now

For PC gaming news and discussion. PCGamingWiki

Rules:

  1. Be Respectful.
  2. No Spam or Porn.
  3. No Advertising.
  4. No Memes.
  5. No Tech Support.
  6. No questions about buying/building computers.
  7. No game suggestions, friend requests, surveys, or begging.
  8. No Let's Plays, streams, highlight reels/montages, random videos or shorts.
  9. No off-topic posts/comments.
  10. Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] aluminium@lemmy.world 2 points 2 months ago* (last edited 2 months ago) (1 children)

been like this with the Apple A chips for years

[–] Ugurcan@lemmy.world 0 points 2 months ago* (last edited 2 months ago) (2 children)

I have to demonstrate to my friends every time how my MBP M2 blows my Ryzen 5950x desktop out of the water for my professional line of work.

I can’t catch quite the drift what x86/x64 chips are good for anymore, other than gaming, nostalgia and spec boasting.

[–] WolfLink@sh.itjust.works 4 points 2 months ago* (last edited 2 months ago) (1 children)

I have a 5950X computer and a Mac mini with some form of M2.

I render video on the M2 computer because I have that sweet indefinite Final Cut Pro license, but then I copy it to the 5950X computer and use ffmpeg to recompress it, which is like an order of magnitude faster than using the M2 computer to do the video compression.

I have some other tasks I’ve given both computers and when the 5950X actually gets to use all its cores, it blows the M2 out of the water.

[–] Ugurcan@lemmy.world 2 points 2 months ago (1 children)

Is it possible you’re using your desktop’s GPU for ffmpeg encoding, and not the CPU, by chance?

[–] WolfLink@sh.itjust.works 3 points 2 months ago

No, you need to manually specify that, and the options are more limited, so I usually do CPU encoding unless I’m prioritizing encoding speed over quality for some reason. (And yes, I have verified it’s using the CPU by looking at the CPU usage while it’s encoding).

[–] psvrh@lemmy.ca 2 points 2 months ago

I can’t catch quite the drift what x86/x64 chips are good for anymore, other than gaming, nostalgia and spec boasting.

Probably two things:

  • Cost- and power-no-object performance, which isn't necessarily a positive as it encourages bad behaviour.
  • The platform is much more open, courtesy of some quirks of how IBM spec'ed BIOS back before the dawn of time. Yes, you can get ARM and RISC-V licenses (openPOWER is kind of a non-entity these days) and design your own SBC, but every single ARM and RISC-V machine boots differently, while x86 and amd64 have a standard boot process.

All those fancy "CoPilot ready" Qualcomm machines? They're following the same path as ARM-based smartphones have, where every single machine is bespoke and you're looking for specific boot images on whatever the equivalent of xda-developers is, or (and this is more likely) just scrapping them when they're used up, which will probably happen a lot faster, given Qualcomm's history with support.

I'd love to see a replacement for x86/amd64 that isn't a power suck, but has an open interface to BIOS.