A prominent open-source dev publishes their findings as to what's going on with Starfield's performance, and it's pretty darn strange.
According to Hans-Kristian Arntzen, a prominent open-source developer working on Vkd3d, a DirectX 12 to Vulkan translation layer, Starfield is not interacting properly with graphics card drivers.
To attribute this most recent failure to an overabundance of hardware variety is a joke. This issue persists on all Nvidia and Intel cards. Why? Because it’s an oversight pertaining to one thing they all share in common: their shared interaction with DirectX.
Let me repeat myself for the people in the back. The number of items they had to account for with this failure is one. One API.
This sounds more like hardware manufacturers haven’t provided a good enough abstraction layer across their devices, or they did (vulkan) but everyone is just stuck on bad apis that don’t properly map to the abstractions for the hardware. Or even more likely the publishers cheaped out and pushed something to release when it wasn’t ready like they have been forever.
It’s also a lack of specialized talent. There’s lots of great “talent” at game devs and even middleware devs. There’s just not much great talent that deals with renderers and API development. The vast majority of devs just lean on the middleware developer to push out the renderer codebase. In a situation like Bethesda running their own studio engine, they just don’t have the right people for it. This plagued the 90’s when people were trying to code for Glide, OGL, DX5,6,7,8, and 9. Many studios folded because they couldn’t get their tech to work with hardware acceleration.
Pc gaming is and forever will be way better then games on consoles.
Why?
I’ve 3 letters for you.
R G B
( ͡° ͜ʖ ͡°)
tbf pc gaming was always a fight for performance, I never felt superior back in the day fighting with qemm, irqs for the soundblaster or glide3d, it’s always had been a shitshow. It was a super shitshow in the nineties, it was a bit better in the zero’s and nowadays it again became a tad better.
deleted by creator
To attribute this most recent failure to an overabundance of hardware variety is a joke. This issue persists on all Nvidia and Intel cards. Why? Because it’s an oversight pertaining to one thing they all share in common: their shared interaction with DirectX.
Let me repeat myself for the people in the back. The number of items they had to account for with this failure is one. One API.
This sounds more like hardware manufacturers haven’t provided a good enough abstraction layer across their devices, or they did (vulkan) but everyone is just stuck on bad apis that don’t properly map to the abstractions for the hardware. Or even more likely the publishers cheaped out and pushed something to release when it wasn’t ready like they have been forever.
It’s also a lack of specialized talent. There’s lots of great “talent” at game devs and even middleware devs. There’s just not much great talent that deals with renderers and API development. The vast majority of devs just lean on the middleware developer to push out the renderer codebase. In a situation like Bethesda running their own studio engine, they just don’t have the right people for it. This plagued the 90’s when people were trying to code for Glide, OGL, DX5,6,7,8, and 9. Many studios folded because they couldn’t get their tech to work with hardware acceleration.
*for current wage
Excellent point.
Lol
Pc gaming is and forever will be way better then games on consoles.
Why?
I’ve 3 letters for you.
R G B
( ͡° ͜ʖ ͡°)
tbf pc gaming was always a fight for performance, I never felt superior back in the day fighting with qemm, irqs for the soundblaster or glide3d, it’s always had been a shitshow. It was a super shitshow in the nineties, it was a bit better in the zero’s and nowadays it again became a tad better.
But somehow I enjoyed that shitshow. Still do.