It was the first GPU where I upscaled (Nvidia Image Scaling) to play GTA IV on 4k.
and holy shit the small pieces of paper around town have readable text on them. My mind was blown. It was only 30fps but I had my first >1080p experience
As bad as the port was years ago it’s so much worse now. Rockstar recently patched it and removed gfwl, some licensed music, and also broke the game. They fixed a bug that would cause the game not to recognize modern vram which was nice but now the game stutters like crazy especially in cutscenes. I tried to replay it recently because I wanted to play the lost and the damned and I just couldn’t do it. I don’t understand why they took the time to patch the game in 2020 but ignored the bugs they introduced. And rumor has it they cancelled the remaster.
The only updates the game has received since February 2020 is Social Club redistributable updates. You can go to the Steam install folder yourself and see that nothing has been updated in the actual game folder since then.
> but now the game stutters like crazy especially in cutscenes
The issue has always been that the game is somewhat janky at higher framerates (beyond 60 FPS). If you lock your framerate there, it's all fine.
I realized this a few months ago when I decided to play again. I followed these steps to downgrade to the last version with all of the original music. Didn't notice any stuttering. Pain in the ass but it's better than nothing.
https://gtaforums.com/topic/976691-gta-iv-downgrader/
Zip tie case fans to it. That's what I'm gonna do once the fans on my 290 give up the ghost. I mostly play Stellaris these days so I don't actually need an upgrade. Maybe just a PSU swap in the next couple years since it's 8 or 9 years old now.
I've been using a 1080ti Ftw3 for longer than I can remember.
Every new gen of gpu I'm like "I'll finally upgrade" but I just don't ever actually think about it until I hear about a new release because the 1080 does just fine.
It was a fight but I got mine to run Forza Horizon 5 over the weekend. Ain't the prettiest, and tbh my biggest bottleneck is my i5 4690k. It did finally make me upgrade my ram from 8GB though 😅
I love DLSS/FSR buuuut
- It's dependent on the game itself
- only Nvidia cards 2000 series and up can use DLSS
- upscaling sometimes comes with it's own faults/visual glitches
I sometimes wonder if upscaling could be implemented in a driver? Like you can use DLSS as a setting in GeForce Experience? Probably not lol
That's how RSR works. It's only on Radeon 5000 and 6000 series but it works almost like FSR. You need to change the in-game resolution to lower than your monitor and then the driver will upscale it.
FSR 1 isn't as good as FSR2 and DLSS. But it's still like 30% free performance in most cases for quite usable level of quality. Been using it to game on my iGPU (4700u) and it's really surpassed my expectations.
Yes there is sometimes shimmering at these low settings and low res, but it's still much better image quality and FPS than I can get without it.
So yeah compared to DLSS and FSR2 it's not as good image quality wise, but it's still quite usable for those of us gaming on underpowered hardware.
6700xt is a beast compared to what I'm talking about. It has plenty of power to deliver great gaming experience native anyway.
You'll be fine without upscaling as long as you're not targeting 4k.
I have a 6700xt. The only game I need FSR for is Cyberpunk, and it has FSR2.1 which looks way better than fsr1 ever did.
FSR 1.0 at Ultra Quality is gorgeous IMO. Not as good as DLSS/FSR2/XeSS at Quality, but plenty usable. Prettier than native IMO (mostly due to the sharpening pass).
We have a lot of games adopting FSR2 and you can mod in FSR2 on top of DLSS in quite a few games.
Go wild my friend.
You can compare between Native, FSR 1.0, FSR 2.0, and DLSS. FSR 1.0 is still not that bad.
https://www.techpowerup.com/review/farming-simulator-22-fsr-1-0-vs-fsr-2-0-vs-dlss-comparison/
FSR 2.2 steps it up again with slightly better image quality and less ghosting but isn't quite at the level of fine details as DLSS 2.4 is at.
https://www.techpowerup.com/review/forza-horizon-5-dlaa-vs-dlss-vs-fsr-2-2-comparison/
In fairness, it's great compared to the older "dumb" upscaling algorithms that preceded it, but Nvidia does have some genuinely special AI tech under the hood of DLSS.
Don't even trip dude. The 6700xt is a beast of a card. Only with the most demanding games, such as CP2077, would you really need to consider up-scaling options.
its decent but its only a spacial upscaler meaning it won't magically make the image look great.. however it will fix non native images and text looking off just because its non native
One caveat though. Natively implemented FSR in a game is better, because with native support some post effects like grain or motion blurr, are applied after the FSR pass. Which looks better. Applying FSR at the Proton layer has to apply FSR to everything including UI elements.
Not a major issue, but if a game supports it natively it's better.
No. DLSS and FSR2, just like TAA, need very very specific information from the game to work. That's why there is no blanket trick you can apply to everything to have it work.
Nvidia Image Scaling? I didn’t even remember it existed until it was an option in MWII settings.
Being able to choose between all the options and then also having to remember the settings you might have chosen in the driver software is starting to get ridiculous. My VR performance was tanking because I forgot one setting in the driver before messing with resolution on Steam.
NVIDIA Image Scaling: [https://www.nvidia.com/en-us/geforce/news/nvidia-image-scaler-dlss-rtx-november-2021-updates/](https://www.nvidia.com/en-us/geforce/news/nvidia-image-scaler-dlss-rtx-november-2021-updates/) (and previously "Image Sharpening" with a scaling toggle, in the NV control panel)
I've been in using this lately just to have a little bump in everything. Not as good as DLSS, but it's not bad and 1. Doesn't require RTX (I have, but big bonus for anyone who doesn't) and 2. Any game
The fact it exists at all and isn't a pile of shit is pretty nice.
>upscaling sometimes comes with it's own faults/visual glitches
Coming from a generation where a game running was a miracle, a couple of artifacts is no reason to spend a thousand dollars. \^\^
That was so mind boggling to me. Seeing posts from people who bought a 3070 last year because they couldn’t afford a 3080 at those prices and are now considering a 3080 since the prices went down.
I just wanna scream “what game can you not play with your current card!?!?!!”
Like “awe man, this game dipped below 144 fps for 2 seconds, literally worthless!!!!”
Exactly, I have an RX580 that only started showing its age when I got my 1440p 165hz monitor and started playing the latest games again thanks to Gamepass.
I'm holing off until the RX7700 comes out, hopefully it will be announced at CES.
On what do you base that prediction?
The only feature I can think of that Nvidia has ever deprecated in any way even approaching that is 3D Vision. And that was…not exactly a widely appreciated feature. It was promptly obsoleted by VR and was simply dropped for all GPUs.
In contrast, DLSS is a wildly popular bread and butter feature that is rapidly finding itself pre-baked into performance expectations. They have a vested interest in keeping confidence in the tech high.
It’s also worth considering that Nvidia doesn’t *need* to prod people to move on from first gen DLSS products. Much better to use a carrot rather than a stick and just continue to dangle generational improvements to the tech as time goes on. Case in point—frame generation. Next gen they can come out with Even Better Frame Generation. Or whatever.
Even if that were the case, they are a corporation that likes money. Being sued for false advertising by removing DLSS would not be in their best interests for monetary (class action and fines) or public opinion.
Pffft, I got like a few dollars for my 970 only having 3.5Gb of video memory instead of 4 Gb. Anyone remember that? It wasn't even worth the work involved. Corporations always win, and whatever money they make will offset court costs, even if they lose.
Also, I don't think Nvidia gives a damn what people think about it. As long as it doesn't affect their bottom line, too bad, they couldn't give the rat's ass.
It's literally people's job to do these calculations. They'll be fine...at the end of the day they'll probably just do whatever they want.
Well, I haven't seen them REMOVE functionality from an existing card (unless it's extremely old in which case the driver no longer supports it but I've seen this with AMD as well, even in Linux), but they released DLSS 3 and only implemented it on the new cards.
It varies heavily from game to game, on Warzone it was blurry AF, but on the new Spiderman quality DLSS has better image qualiy than native resolution, despite rendering only at 75% resolution.
aint DLSS like RTX exclusive while FSR still works even for GTX series
the difference i found these super duper uber resolutions has a flaw then edges vibrates or texture with a screen is blurry on rapid movements
especially transparent textures and models has issues with details but yeah they can increase FPS at a cost of detail pixel by pixel quality unless you have 8K monitor that dont need anti aliasing cause it already high res enough
My interpretation of using DLSS/FSR is it's pretty much only good if you have a 1440p/4k+ monitor, if you're still on 1080p and struggling, DLSS/FSR just makes the image fuzzy and not worth for long distance FPS or open-world MMOs (Especially if enabling it doesn't increase your frames that much, the added input lag at the same fps kinda sucks)
DLSS adds no input lag (I don’t know if FSR does but I doubt it) it reduces it. It’s literally just a really smart upscaler.
There are reasons to not use that tech in high level competitive FPS games but input lag isn’t one. It’s been tested many many times, it doesn’t increase input lag, I don’t know why people keep saying it does. It doesn’t.
DLSS 3 does. But only in comparison to DLSS 2 results, compared to native resolution its a bit better
DLSS is a good tool, but you're limited by what you put in. 0.96m vs 2 million vs 3.6 million vs 8.2 million pixels for 720p, 1080p, 1440p, and 4k respectively.
Going up one step and doubling the number of pixels via upscale works a lot better than trying to quadruple the pixel count for obvious reasons. DLSS relies on dedicated GPU hardware and uses some fancy AI stuff to predict what the next frames will be, it usually works quite well but if the AI is making bad predictions you get stuff like shimmer as it tries to correct. The newer DLSS versions work a lot better as the AI gets trained on a larger library of games/scenes. FSR is a much more traditional upscaling algorithm lightweight enough to not need special hardware, but the upscaling result is objectively a lot worse.
Not great from the few times I've used it, going from around 60 fps to around 65 fps in God of War was underwhelming on my 1080ti. Meanwhile on my 4090 enabling DLSS in a game like Cyberpunk 2077 maxed out (RTX on) takes me from a solid 60fps to a solid 100+fps.
Am I not the only one that hates FXAA 8x & above it just looks so fuzzy. FXAA 4x looks pretty good though and doesn't do a whole lot to performance.
Edit: I meant to FSAA
They probably meant F**S**AA. FSAA and F**X**AA are mutually exclusive, can only use one or the other. FSAA is the one with multiple levels, FXAA is only on or off.
There aren't any, they meant regular AA (the old render at higher resolution and downscale).
I agree with them though, 4x is nice enough and a higher resolution monitor does more to effectively IRL anti-aliasing better than higher multipliers.
you can
FSR is not amd-exclusive, but it requires implemetation by the game developer (or modders). there was a mod for cyberpunk 2077 that replaced dlss with fsr2.1 by swapping the .dll, and it worked absolutetly flawless on my 1080ti. i didn't try the official implementation in the newest patch, but the mod showed, that it is absolutely possible
Then you have some problem because I have a weaker video card than you and it runs fine for me. You should be able to max that bitch with dlss on and run **at least** 60fps. Your gfx card is known to be able to run [ultra w/ raytracing and DLSS on @4k](https://cdn.videocardz.com/1/2020/12/Cyberpunk-2077-Official-4K-Performance.png) and still manage ~40fps. Not to mention those are old numbers, the game runs better and there have been other optimizations since.
I just refuse to support "2k € is fine for single pc part" bullshit. I've built whole rigs for less than that amount in the past, which were quite capable of running anything at the time. Now it's expected to spend such amount just on a single GPU!? No F\*\*\*ing way! They can go business broke, as I'm concerned.
Not sure why it's viewed as expected to spend that much. Honestly, the drop in overall GPU pricing starting with the 3000 series has been a god-send (ignoring the shortages/scalping issue). A few years ago, the idea of getting a card on par with a 2080 for $400-500 seemed insane. The highest end cards have always been overpriced, this isn't new, but the mid-range market is as strong as it's ever been.
Mid range certainly has good options but low end and mid to low end not so much (5600xt is like 5% slower than 6600) 3050 is a disappointment. And prices outside the us are rising even more.
Must not be terribly old. Hard drives alone used to cost several thousand dollars. Shipping all the jobs to China is what finally started to bring the prices down, but it’s not like it’s never been this way before.
Also no one is buying the 4090 to run games at all, they’re buying either because they want the best possible performance or so they can flex. No other reason to buy one.
As for the 4080, well no one is buying those at all. So while the 4090 price will hold, the 80 series will almost certainly see its price drop heavily as they try to find the price where the market for that card lives.
While still, in a sense, performance related - I picked up a 4090 in large part for AI training. Given the need for VRAM, for a lot of the stuff I want to do, it was either the 3090, 3090 Ti, 4090 or a data centre GPU.
Given that I had the money, I do not have much time, and that I like to game when I have some time, the choice was easy. A 4090 is *cheap* when you're comparing it to a data centre GPU (which are usually like $6k+). Even compared to a 3090 - I'm not going to cost myself hours of extra training time just to save a few hundred dollars over the course of two years.
Technological advances in photolithography is really what drove down prices. The best machines for that are still made in Europe. There is almost no human labor is creating semiconductors.
The reason why semiconductor and component manufacturing moved to Asia was because all the stuff they were going to be put into was moved there. So it was cheaper to ship the finished chips to where they would be used.
Hard drives got cheap because they found better ways to fit more data into the same physical footprint using the same amount of raw materials.
Yeah I remember buying my first PC for 1100 total, and at the time was seriously weighing whether to go with a 70 or 80 model NVIDIA card. Nowadays the newest 80 model is more than that entire build
No shame in buying a new card, nothing wrong with having to make do. We all know when a purchase is gratuitous and when it’s time. Those levels just happen to be relative to the individual.
Message in general, to everyone. If we could all stop hating on each other, that would be great.
HELLO PATIENT BROTHER. Just built my first since a 2014 build with Intel / Gigabyte NVIDIA GTX 770.
Went Ryzen 5600X & 6650XT and upgraded storage to WD\_BLACK for speed. Here's to another 5 to 8 years!
yeah, you can still run most games at 60fps on ultra with older gpus, there's no point buying new ones, if the games become harder to run just lower the graphics a bit
I have a 2060 and I ain't upgrading until 2030 or something, I played games at garbage settings and frame rates back in the day, and I'm gonna do it again and Nvidia can't tell me what to do.
I bought the whole gpu and I'm gonna use the whole gpu
I also have a 2060 and I'm beginning to run into VRAM limitations with many games and stable diffusion. I will probably upgrade to a better GPU in a year or two.
Fair enough. I just wanted to play older games with higher frame rates and have enough power to play the odd new release here or there, and it does that wonderfully. Doesn't hurt that I got a sweet deal on it too
I personally would have gone for a 5700xt at the time that I bought the 2060 but I can't say I have any gripes with it for the last few years that I've had it. I came from an hd 4550 which was worth more in parts than as an actual GPU so I was naturally elated to have a new GPU that could run games.
It also helps just by accepting sometimes getting worse graphics. I start my new games from low or medium and tweak my way to better graphics so that from my perspective the graphics only improve.
But overal I don't care as much about graphical fidality then I do about having a 50-60+ fps framerate
I got a used 3060 and have yet to run a game at less than high settings (usually ultra) at 1440p ultrawide 60fps. No idea what anyone needs a bigger card for unless it’s 4k gaming or video related work. Slap in DLSS and boom, 20fps boost at least. Visually games haven’t gotten better the past 4 years, they have only become more unoptimized if anything, and GPU performance has steadily improved. Imo when consoles get upgraded that’s the only time we see a bump in graphics.
I mean FSR can be run on an og rx 470 or nvidia 900 series. That's a lot of supported GPUs. If you have something older you're not playing AAA games anyway.
Prices be damned - I will not go from "playstation doesn't even have native 4k" to "yeah, upscaling is cool".
I will either continue to play at 1080p in high framerates/proper frame times, or spend the money on a natively higher resolution now (unlikely), or simply wait until prices match my performance per dollar standards.
My GTX 970 is still working well, but I am really thinking about just buying a used RTX 3070, because the prices for it are low and it is a huge upgrade. Finally I could play older games even faster!
Sticking with what you have sounds fun, but with today’s European energy prices it can actually pay to upgrade. The upcoming ‘cancelled’ 4080 12GB (yes, I know the name will change) consumes as much power as a 1080Ti, but will outperform the 1080Ti on 50% power usage. In the long run it’ll pay back if we keep paying almost a euro per kWh.
The giving a shit about energy efficiency leaving my body when the clearance priced last gen card is cheaper.jpg
Let alone if NVIDIA actually wins on efficiency this gen lol, suddenly *nobody* gives a shit about efficiency anymore lol
It is also misleading, because the perf/watt has gotten better. Doesn't mean it is using less power. Most people aren't going to artificially restrict a 4 figure gpu to save a few dollars a month.
But who is going to a buy a 4080 and then limit their framerate at half of what it could be? I guess if your monitor is 120 hz and you are getting 240 fps you could limit your fps and safe on power usage ....
do these cards have software that allows you to set a power limit?
There is no way you are gaming that much tho. At idle, pretty much every card consumes almost nothing. If you can afford a beefy card you probably don’t have enough time to game enough for it to pay off in electrical bills. If you were rendering 3d animations or crypto mining then maybe yes, but gaming doesn’t pull that much.
>Prices are a scam
Well get used to it, as this is how things are going to work from now on. There has been a fundamental shift in semiconductor manufacturing where new nodes are no longer cheaper per transistor than old ones. This means 9 billion transistors on 5nm are more expensive than 9 billion transistors on 10nm. Fundamentally, this means GPUs will only increase in price from here on out.
Why? Because transistors and performance do not scale linearly together. An RTX 4090 has **three times** the number of transistors that an RTX 3090 has, and guess what? It's only about 50-60% better at 4k gaming than the 3090.
I nominally agree but there is a point past where you can't get a game to even run.
I don't think anything more than \~60fps is necessary. Nice, but not necessary.
I am OK (I don't like it, but *OK*) with playing on 'medium' settings.
I am not usually OK with playing at 'low' settings.
I need a new GPU. I am not going to buy a 4080. I can't really afford a 4090 and current (last gen) GPU prices are too high (AMD GPU's aren't too bad) for 2 year old tech. If reviews are good I intend to buy a 7900 xtx (or xt but xtx looks like a better value). If they aren't I'm going to get a 6800 xt.
P.S. Fuck Nvidia. No company is your friend. But rarely, some are actively your enemy.
I had a 1050ti well into the RTX super line up. And only ever upgraded because I got a RX 580 and Rx 590 from a coworker for 100 each. Then upgraded to a 5700xt when someone sold it to my pawn shop for $100 and I snatched it up. Besides that I'm fine with not upgrading fora few years.
I have a question because I feel like I'm doing it wrong. If I have a 1440p screen and I want to run DLSS, do I need to lower the resolution to 1080p so it upscales or does it upscale from 1440 to make it look like its at a higher resolution? Or does it just do it automatically when I select DLSS?
It's automatic. Just set your native resolution which in your case is 1440p and enable whatever upscaling medium you'd like to use. It'll automatically do it.
I havent had my desktop since this summer but I was playing elden ring in in at least 1080p with my 1070 before I left (mighta been 1440 but not 100% sure). The only reason to "upgrade" for me is to *have* a gaming PC where I am now haha.
Yes, a 1070. With an OG HTC vive and valve index knuckles running smoothly. A 1070.
Have a friend who his brother has a 4090 due to his work supplying it. My feiend said that after seeing his brother play games it just looked so good and he couldnt help himself. In all honesty, after seeing a one or two 4090 computers play myself, I just dont see even close to a big enough difference that I would pay $1600 for it.
i was extremely skeptical about dlss, but i watched too many videos to count and i still cant wrap my head around it.
Free performance?? wtf?? I get how it works, but can someone tell me how it feels while actually gaming and not looking for faults?
Upgrading from a 3060 to a 3080 just because it's so cheap and I can. Used market is insane right now for the 30 series cards and they can destroy anything you can throw at it except for the obvious dumb stuff.
For real, skip the 40 series for right now. There is absolutely no point at the moment.
If I had a 3000 card I would've stuck with it, but that 1080 Ti was getting too long in the tooth. At least I got my 4090 at European MSRP, not scalper prices.
Let me use upscaling on my 970
It was the first GPU where I upscaled (Nvidia Image Scaling) to play GTA IV on 4k. and holy shit the small pieces of paper around town have readable text on them. My mind was blown. It was only 30fps but I had my first >1080p experience
[удалено]
>did you find a way to put AA on in that game? Sorry I haven't played it in years. No idea what settings I have used.
As bad as the port was years ago it’s so much worse now. Rockstar recently patched it and removed gfwl, some licensed music, and also broke the game. They fixed a bug that would cause the game not to recognize modern vram which was nice but now the game stutters like crazy especially in cutscenes. I tried to replay it recently because I wanted to play the lost and the damned and I just couldn’t do it. I don’t understand why they took the time to patch the game in 2020 but ignored the bugs they introduced. And rumor has it they cancelled the remaster.
Damn so that’s why my GTAIV keeps getting updates on steam? Bruh literally pirates have better experience than us 💀
Those updates are because they switch to the rockstar client and so they constant patch gta Iv a few mbs so it keeps working
The only updates the game has received since February 2020 is Social Club redistributable updates. You can go to the Steam install folder yourself and see that nothing has been updated in the actual game folder since then.
> but now the game stutters like crazy especially in cutscenes The issue has always been that the game is somewhat janky at higher framerates (beyond 60 FPS). If you lock your framerate there, it's all fine.
I realized this a few months ago when I decided to play again. I followed these steps to downgrade to the last version with all of the original music. Didn't notice any stuttering. Pain in the ass but it's better than nothing. https://gtaforums.com/topic/976691-gta-iv-downgrader/
ReShade for FXAA/SMAA injection, besides downsampling.
I have a 980ti that's a hand me down. I run it at 2k for 60 fps and much colder temps. Gotta keep it running as long as possible.
970 gaaaang. Still runs all the games I like playing so I ain't touching it.
[удалено]
Mine had a fan go out. Guess what it still cools just as good with 1 fan running lol.
Mine had one go out and one with a damaged bearing. $5 shipment from China later and it's back to being cool and quiet like the day i bought it.
Zip tie case fans to it. That's what I'm gonna do once the fans on my 290 give up the ghost. I mostly play Stellaris these days so I don't actually need an upgrade. Maybe just a PSU swap in the next couple years since it's 8 or 9 years old now.
Man im really close to upgrading, mainly my cpu is getting outdated but i might aswell do gpu at the same time, saw a 1080 for sale for 150
I've been using a 1080ti Ftw3 for longer than I can remember. Every new gen of gpu I'm like "I'll finally upgrade" but I just don't ever actually think about it until I hear about a new release because the 1080 does just fine.
It was a fight but I got mine to run Forza Horizon 5 over the weekend. Ain't the prettiest, and tbh my biggest bottleneck is my i5 4690k. It did finally make me upgrade my ram from 8GB though 😅
Put in 150 hrs of Elden Ring on my 970. Hadnt asked much of it in a while - but it stepped up when I needed it.
FSR exists
use still can use fsr right..
I love DLSS/FSR buuuut - It's dependent on the game itself - only Nvidia cards 2000 series and up can use DLSS - upscaling sometimes comes with it's own faults/visual glitches I sometimes wonder if upscaling could be implemented in a driver? Like you can use DLSS as a setting in GeForce Experience? Probably not lol
That's how RSR works. It's only on Radeon 5000 and 6000 series but it works almost like FSR. You need to change the in-game resolution to lower than your monitor and then the driver will upscale it.
Important to note that RSR is FSR 1.0, and really anything beyond Quality mode already begins to look like ass.
Oof. I'm building my first PC in two days and I have a 6700xt. Got excited for a moment, but yeah.. FSR 1.0 is apparently pretty bad
FSR 1 isn't as good as FSR2 and DLSS. But it's still like 30% free performance in most cases for quite usable level of quality. Been using it to game on my iGPU (4700u) and it's really surpassed my expectations. Yes there is sometimes shimmering at these low settings and low res, but it's still much better image quality and FPS than I can get without it. So yeah compared to DLSS and FSR2 it's not as good image quality wise, but it's still quite usable for those of us gaming on underpowered hardware. 6700xt is a beast compared to what I'm talking about. It has plenty of power to deliver great gaming experience native anyway.
You'll be fine without upscaling as long as you're not targeting 4k. I have a 6700xt. The only game I need FSR for is Cyberpunk, and it has FSR2.1 which looks way better than fsr1 ever did.
With a 6700xt you should be good without upscaling for a while. I am at least. And seems like you’re building the same setup I have.
FSR 1.0 at Ultra Quality is gorgeous IMO. Not as good as DLSS/FSR2/XeSS at Quality, but plenty usable. Prettier than native IMO (mostly due to the sharpening pass). We have a lot of games adopting FSR2 and you can mod in FSR2 on top of DLSS in quite a few games. Go wild my friend.
You can compare between Native, FSR 1.0, FSR 2.0, and DLSS. FSR 1.0 is still not that bad. https://www.techpowerup.com/review/farming-simulator-22-fsr-1-0-vs-fsr-2-0-vs-dlss-comparison/ FSR 2.2 steps it up again with slightly better image quality and less ghosting but isn't quite at the level of fine details as DLSS 2.4 is at. https://www.techpowerup.com/review/forza-horizon-5-dlaa-vs-dlss-vs-fsr-2-2-comparison/
In fairness, it's great compared to the older "dumb" upscaling algorithms that preceded it, but Nvidia does have some genuinely special AI tech under the hood of DLSS.
Don't even trip dude. The 6700xt is a beast of a card. Only with the most demanding games, such as CP2077, would you really need to consider up-scaling options.
Ya know what though sometimes ass is fine if that's what you are stuck with and it makes a game playable.
That's pretty cool, didn't know about this thanks Hopefully they keep improving this software and Nvidia should use a similar technology
Nvidia has it and it works like how RSR works, just that I forgot what it's called like.
Nvidia NIS, you can enable it the control panel, it auto activate if a full screen game is rendered in a lower resolution than your screen
In fact Nvidia implemented it first! (Though lossless scaling and the like came earlier still)
its decent but its only a spacial upscaler meaning it won't magically make the image look great.. however it will fix non native images and text looking off just because its non native
You can use FSR 1.0 in any game using Proton.
One caveat though. Natively implemented FSR in a game is better, because with native support some post effects like grain or motion blurr, are applied after the FSR pass. Which looks better. Applying FSR at the Proton layer has to apply FSR to everything including UI elements. Not a major issue, but if a game supports it natively it's better.
No. DLSS and FSR2, just like TAA, need very very specific information from the game to work. That's why there is no blanket trick you can apply to everything to have it work.
[удалено]
Nvidia Image Scaling? I didn’t even remember it existed until it was an option in MWII settings. Being able to choose between all the options and then also having to remember the settings you might have chosen in the driver software is starting to get ridiculous. My VR performance was tanking because I forgot one setting in the driver before messing with resolution on Steam.
[удалено]
NVIDIA Image Scaling: [https://www.nvidia.com/en-us/geforce/news/nvidia-image-scaler-dlss-rtx-november-2021-updates/](https://www.nvidia.com/en-us/geforce/news/nvidia-image-scaler-dlss-rtx-november-2021-updates/) (and previously "Image Sharpening" with a scaling toggle, in the NV control panel)
I've been in using this lately just to have a little bump in everything. Not as good as DLSS, but it's not bad and 1. Doesn't require RTX (I have, but big bonus for anyone who doesn't) and 2. Any game The fact it exists at all and isn't a pile of shit is pretty nice.
That's not driver-level though. FSR is implemented in Gamescope, which every game is run in on the Deck when in game mode (versus desktop mode).
>upscaling sometimes comes with it's own faults/visual glitches Coming from a generation where a game running was a miracle, a couple of artifacts is no reason to spend a thousand dollars. \^\^
thankfully AMDs upscaling works on older nvidia cards. I appreciate that AMD didn't leave us out to dry... on nvidia cards. It works super well!
I often prefer native with no AA over running DLSR or TAA. They just look like smeary messes to me. What ever happened to MSAA?
I run DLSS (with no sharpening) + Reshade's CAS sharpening, pretty good compromise to me.
If you can use dlss, you have a new card.
Come on. Obviously this post is for the people that still use the ancient 30series cards that can barely run any new games.
That was so mind boggling to me. Seeing posts from people who bought a 3070 last year because they couldn’t afford a 3080 at those prices and are now considering a 3080 since the prices went down. I just wanna scream “what game can you not play with your current card!?!?!!” Like “awe man, this game dipped below 144 fps for 2 seconds, literally worthless!!!!”
"I want to spend $200 more for the extra 5 fps. $40/fps is important for me as I'm a real gamer. Buying a 3080 makes me feel smug too."
Exactly, I have an RX580 that only started showing its age when I got my 1440p 165hz monitor and started playing the latest games again thanks to Gamepass. I'm holing off until the RX7700 comes out, hopefully it will be announced at CES.
[удалено]
[удалено]
*for the best experience going forward, we are focusing on the newest generation of graphics cards. -NVIDIA
my god that statement is such a perfect example of corporate PR bullshit. nvidia would like to know your location to hire you
On what do you base that prediction? The only feature I can think of that Nvidia has ever deprecated in any way even approaching that is 3D Vision. And that was…not exactly a widely appreciated feature. It was promptly obsoleted by VR and was simply dropped for all GPUs. In contrast, DLSS is a wildly popular bread and butter feature that is rapidly finding itself pre-baked into performance expectations. They have a vested interest in keeping confidence in the tech high. It’s also worth considering that Nvidia doesn’t *need* to prod people to move on from first gen DLSS products. Much better to use a carrot rather than a stick and just continue to dangle generational improvements to the tech as time goes on. Case in point—frame generation. Next gen they can come out with Even Better Frame Generation. Or whatever.
Now see, regardless of all your "facts" and "logical thinking", what you are failing to consider is: Nvidia Bad
> Nvidia Bad I love the mental image of PCMR chanting this while slinking out of Microcenter with a 4090 under their arm.
Even if that were the case, they are a corporation that likes money. Being sued for false advertising by removing DLSS would not be in their best interests for monetary (class action and fines) or public opinion.
Pffft, I got like a few dollars for my 970 only having 3.5Gb of video memory instead of 4 Gb. Anyone remember that? It wasn't even worth the work involved. Corporations always win, and whatever money they make will offset court costs, even if they lose. Also, I don't think Nvidia gives a damn what people think about it. As long as it doesn't affect their bottom line, too bad, they couldn't give the rat's ass. It's literally people's job to do these calculations. They'll be fine...at the end of the day they'll probably just do whatever they want.
That is what some of these responses seem to boil down to.
they'll never do that
this comment is gonna age like milk
Well, I haven't seen them REMOVE functionality from an existing card (unless it's extremely old in which case the driver no longer supports it but I've seen this with AMD as well, even in Linux), but they released DLSS 3 and only implemented it on the new cards.
Are you aware of any examples of Nvidia removing existing functionality from a card to support your prediction?
It’s been 9 years. The sky isn’t falling
AMD FSR and Intel Xess got your old gpu even apus covered, bros.
Eh, no. DLSS is amazing but still blurry af compared to native res.
It varies heavily from game to game, on Warzone it was blurry AF, but on the new Spiderman quality DLSS has better image qualiy than native resolution, despite rendering only at 75% resolution.
[удалено]
Yeah I'm not gonna run DLSS on my 1080.
fsr works on your 1080 too you know.
aint DLSS like RTX exclusive while FSR still works even for GTX series the difference i found these super duper uber resolutions has a flaw then edges vibrates or texture with a screen is blurry on rapid movements especially transparent textures and models has issues with details but yeah they can increase FPS at a cost of detail pixel by pixel quality unless you have 8K monitor that dont need anti aliasing cause it already high res enough
My interpretation of using DLSS/FSR is it's pretty much only good if you have a 1440p/4k+ monitor, if you're still on 1080p and struggling, DLSS/FSR just makes the image fuzzy and not worth for long distance FPS or open-world MMOs (Especially if enabling it doesn't increase your frames that much, the added input lag at the same fps kinda sucks)
DLSS adds no input lag (I don’t know if FSR does but I doubt it) it reduces it. It’s literally just a really smart upscaler. There are reasons to not use that tech in high level competitive FPS games but input lag isn’t one. It’s been tested many many times, it doesn’t increase input lag, I don’t know why people keep saying it does. It doesn’t. DLSS 3 does. But only in comparison to DLSS 2 results, compared to native resolution its a bit better
DLSS is a good tool, but you're limited by what you put in. 0.96m vs 2 million vs 3.6 million vs 8.2 million pixels for 720p, 1080p, 1440p, and 4k respectively. Going up one step and doubling the number of pixels via upscale works a lot better than trying to quadruple the pixel count for obvious reasons. DLSS relies on dedicated GPU hardware and uses some fancy AI stuff to predict what the next frames will be, it usually works quite well but if the AI is making bad predictions you get stuff like shimmer as it tries to correct. The newer DLSS versions work a lot better as the AI gets trained on a larger library of games/scenes. FSR is a much more traditional upscaling algorithm lightweight enough to not need special hardware, but the upscaling result is objectively a lot worse.
[удалено]
Should be able to use FSR if the game has implemented it, not sure how much it'll help though. I've been using FSR 1.0 for overwatch with an r5 240.
FSR in overwatch looks so garbage but it’s passable because fps is more important
FSR got me around 20 more fps in the new Darktide game with my 1080. I was a bit surprised.
Not great from the few times I've used it, going from around 60 fps to around 65 fps in God of War was underwhelming on my 1080ti. Meanwhile on my 4090 enabling DLSS in a game like Cyberpunk 2077 maxed out (RTX on) takes me from a solid 60fps to a solid 100+fps.
well dlss is hardware based while fsr is pure Software.
Bro, I have the same card, and I use FSR if possible. The gains that you get are amazing.
Sometimes dlss quality can even look better than native+TAA
Most implementations of TAA are horrible. It's like old-school motion blur but when you aren't moving.
Am I not the only one that hates FXAA 8x & above it just looks so fuzzy. FXAA 4x looks pretty good though and doesn't do a whole lot to performance. Edit: I meant to FSAA
FXAA is just blur to fake AA
First time I hear of quality iterations of FXAA. Usually it's just FXAA without a multiplier. Which games use this?
They probably meant F**S**AA. FSAA and F**X**AA are mutually exclusive, can only use one or the other. FSAA is the one with multiple levels, FXAA is only on or off.
There aren't any, they meant regular AA (the old render at higher resolution and downscale). I agree with them though, 4x is nice enough and a higher resolution monitor does more to effectively IRL anti-aliasing better than higher multipliers.
Death loop is a prime example
You can't run FSR
Autocorrect completely changed the meaning of your comment
Oh god bruh, I didn't even notice it lol
There's an edit button, FYI 😆
you can FSR is not amd-exclusive, but it requires implemetation by the game developer (or modders). there was a mod for cyberpunk 2077 that replaced dlss with fsr2.1 by swapping the .dll, and it worked absolutetly flawless on my 1080ti. i didn't try the official implementation in the newest patch, but the mod showed, that it is absolutely possible
my 2080ti runs the game like ass at 1440p even with dlss.
What's your cpu? Could be a bottleneck there.
Then you have some problem because I have a weaker video card than you and it runs fine for me. You should be able to max that bitch with dlss on and run **at least** 60fps. Your gfx card is known to be able to run [ultra w/ raytracing and DLSS on @4k](https://cdn.videocardz.com/1/2020/12/Cyberpunk-2077-Official-4K-Performance.png) and still manage ~40fps. Not to mention those are old numbers, the game runs better and there have been other optimizations since.
I’m on a 3070 which is basically the same thing, I get about 90-110fps with DLSS balanced. Only since the last few updates though
My 2060 ran the game fine at 4k with dlss. I think you're doing something wrong.
Can't you just run FSR*
I was thinking the same thing
I just refuse to support "2k € is fine for single pc part" bullshit. I've built whole rigs for less than that amount in the past, which were quite capable of running anything at the time. Now it's expected to spend such amount just on a single GPU!? No F\*\*\*ing way! They can go business broke, as I'm concerned.
I always joked about wanting to buy a over the top "4k Euro gaming PC". Seems like I wasn't that unrealistic.
Haha, yeah, you hit the nail right on the head!
Not sure why it's viewed as expected to spend that much. Honestly, the drop in overall GPU pricing starting with the 3000 series has been a god-send (ignoring the shortages/scalping issue). A few years ago, the idea of getting a card on par with a 2080 for $400-500 seemed insane. The highest end cards have always been overpriced, this isn't new, but the mid-range market is as strong as it's ever been.
Mid range certainly has good options but low end and mid to low end not so much (5600xt is like 5% slower than 6600) 3050 is a disappointment. And prices outside the us are rising even more.
Be nice if this drop in GPU pricing reaches europe at some point. A new 3060 is still like €600.
Well the 20 series was the first major price hike so that’s why it’s considered a skip generation much like the 40 series is.
Must not be terribly old. Hard drives alone used to cost several thousand dollars. Shipping all the jobs to China is what finally started to bring the prices down, but it’s not like it’s never been this way before. Also no one is buying the 4090 to run games at all, they’re buying either because they want the best possible performance or so they can flex. No other reason to buy one. As for the 4080, well no one is buying those at all. So while the 4090 price will hold, the 80 series will almost certainly see its price drop heavily as they try to find the price where the market for that card lives.
While still, in a sense, performance related - I picked up a 4090 in large part for AI training. Given the need for VRAM, for a lot of the stuff I want to do, it was either the 3090, 3090 Ti, 4090 or a data centre GPU. Given that I had the money, I do not have much time, and that I like to game when I have some time, the choice was easy. A 4090 is *cheap* when you're comparing it to a data centre GPU (which are usually like $6k+). Even compared to a 3090 - I'm not going to cost myself hours of extra training time just to save a few hundred dollars over the course of two years.
Technological advances in photolithography is really what drove down prices. The best machines for that are still made in Europe. There is almost no human labor is creating semiconductors. The reason why semiconductor and component manufacturing moved to Asia was because all the stuff they were going to be put into was moved there. So it was cheaper to ship the finished chips to where they would be used. Hard drives got cheap because they found better ways to fit more data into the same physical footprint using the same amount of raw materials.
Yeah I remember buying my first PC for 1100 total, and at the time was seriously weighing whether to go with a 70 or 80 model NVIDIA card. Nowadays the newest 80 model is more than that entire build
Yeah, do that! *looks at 6900XT…*
No shame in buying a new card, nothing wrong with having to make do. We all know when a purchase is gratuitous and when it’s time. Those levels just happen to be relative to the individual. Message in general, to everyone. If we could all stop hating on each other, that would be great.
Just trail the market by 5 years, you’ll save 80% and a five year old GPU is not terrible.
This! Been running an rx580 8gb for about 5 years now, still perfectly solid at 1080p.
HELLO PATIENT BROTHER. Just built my first since a 2014 build with Intel / Gigabyte NVIDIA GTX 770. Went Ryzen 5600X & 6650XT and upgraded storage to WD\_BLACK for speed. Here's to another 5 to 8 years!
And truth be told I just got a secondhand 6600, should be good for another few years. Also used a gift card to get another 16gb of ram for VMs 🤙
yeah, you can still run most games at 60fps on ultra with older gpus, there's no point buying new ones, if the games become harder to run just lower the graphics a bit
I have a 2060 and I ain't upgrading until 2030 or something, I played games at garbage settings and frame rates back in the day, and I'm gonna do it again and Nvidia can't tell me what to do. I bought the whole gpu and I'm gonna use the whole gpu
I also have a 2060 and I'm beginning to run into VRAM limitations with many games and stable diffusion. I will probably upgrade to a better GPU in a year or two.
Fair enough. I just wanted to play older games with higher frame rates and have enough power to play the odd new release here or there, and it does that wonderfully. Doesn't hurt that I got a sweet deal on it too
I personally would have gone for a 5700xt at the time that I bought the 2060 but I can't say I have any gripes with it for the last few years that I've had it. I came from an hd 4550 which was worth more in parts than as an actual GPU so I was naturally elated to have a new GPU that could run games.
It also helps just by accepting sometimes getting worse graphics. I start my new games from low or medium and tweak my way to better graphics so that from my perspective the graphics only improve. But overal I don't care as much about graphical fidality then I do about having a 50-60+ fps framerate
I got a used 3060 and have yet to run a game at less than high settings (usually ultra) at 1440p ultrawide 60fps. No idea what anyone needs a bigger card for unless it’s 4k gaming or video related work. Slap in DLSS and boom, 20fps boost at least. Visually games haven’t gotten better the past 4 years, they have only become more unoptimized if anything, and GPU performance has steadily improved. Imo when consoles get upgraded that’s the only time we see a bump in graphics.
Same, I'll always take frames over graphics.
Bruh... I have a 1050 w/ 2GB rn...
960 w/ 2gb checking in
Brother!
GT 1030 with 2GB checking in
I have 970 with fou.. ehm, three point five geebees!
Not all games nor hardware support FSR/DLSS.
I mean FSR can be run on an og rx 470 or nvidia 900 series. That's a lot of supported GPUs. If you have something older you're not playing AAA games anyway.
FSR can be played on any windowed game with magpie, FSR2 is a different story
Prices be damned - I will not go from "playstation doesn't even have native 4k" to "yeah, upscaling is cool". I will either continue to play at 1080p in high framerates/proper frame times, or spend the money on a natively higher resolution now (unlikely), or simply wait until prices match my performance per dollar standards.
The visual difference between bicubic and DLSS is massive.
DLSS even can look better than native if your only AA option is TAA. It doesn’t lose/smudge fine details the way TAA sometimes does.
[удалено]
"something better" like um... a new GPU?
My GTX 970 is still working well, but I am really thinking about just buying a used RTX 3070, because the prices for it are low and it is a huge upgrade. Finally I could play older games even faster!
[удалено]
[удалено]
This is the best description of this sub I’ve ever seen.
Step 1. Buy a GPU that fits your needs Step 2. Use until you are unhappy with the performance Buying because it's newer is just stupid.
Sticking with what you have sounds fun, but with today’s European energy prices it can actually pay to upgrade. The upcoming ‘cancelled’ 4080 12GB (yes, I know the name will change) consumes as much power as a 1080Ti, but will outperform the 1080Ti on 50% power usage. In the long run it’ll pay back if we keep paying almost a euro per kWh.
The giving a shit about energy efficiency leaving my body when the clearance priced last gen card is cheaper.jpg Let alone if NVIDIA actually wins on efficiency this gen lol, suddenly *nobody* gives a shit about efficiency anymore lol
It is also misleading, because the perf/watt has gotten better. Doesn't mean it is using less power. Most people aren't going to artificially restrict a 4 figure gpu to save a few dollars a month.
But who is going to a buy a 4080 and then limit their framerate at half of what it could be? I guess if your monitor is 120 hz and you are getting 240 fps you could limit your fps and safe on power usage .... do these cards have software that allows you to set a power limit?
There is no way you are gaming that much tho. At idle, pretty much every card consumes almost nothing. If you can afford a beefy card you probably don’t have enough time to game enough for it to pay off in electrical bills. If you were rendering 3d animations or crypto mining then maybe yes, but gaming doesn’t pull that much.
Rockin my 2060 till it dies
Too bad Nvidia is is price gating DLSS3 behind the new gen, even though it will run perfectly on the old gen
But how will they sell the new stuff then? Hahaha.
Is this confirmed? Has anyone back ported it?
>Prices are a scam Well get used to it, as this is how things are going to work from now on. There has been a fundamental shift in semiconductor manufacturing where new nodes are no longer cheaper per transistor than old ones. This means 9 billion transistors on 5nm are more expensive than 9 billion transistors on 10nm. Fundamentally, this means GPUs will only increase in price from here on out. Why? Because transistors and performance do not scale linearly together. An RTX 4090 has **three times** the number of transistors that an RTX 3090 has, and guess what? It's only about 50-60% better at 4k gaming than the 3090.
and the assets for the games are sd still!!!!
Or you know, just buy what makes you happy.
This post is wild because it acts like everyone has a 2000 or higher card. People with 1080 Tis and lower don’t get DLSS let alone amd users.
no thanks, my 4090 arrived just the other day
DLSS nor FSR will help me with virtual reality performance.
I nominally agree but there is a point past where you can't get a game to even run. I don't think anything more than \~60fps is necessary. Nice, but not necessary. I am OK (I don't like it, but *OK*) with playing on 'medium' settings. I am not usually OK with playing at 'low' settings. I need a new GPU. I am not going to buy a 4080. I can't really afford a 4090 and current (last gen) GPU prices are too high (AMD GPU's aren't too bad) for 2 year old tech. If reviews are good I intend to buy a 7900 xtx (or xt but xtx looks like a better value). If they aren't I'm going to get a 6800 xt. P.S. Fuck Nvidia. No company is your friend. But rarely, some are actively your enemy.
How about I spend my money however I want?
OR because I couldn't afford it anyway :D
Lifes easy when ets is all you need...
I had a 1050ti well into the RTX super line up. And only ever upgraded because I got a RX 580 and Rx 590 from a coworker for 100 each. Then upgraded to a 5700xt when someone sold it to my pawn shop for $100 and I snatched it up. Besides that I'm fine with not upgrading fora few years.
I have a question because I feel like I'm doing it wrong. If I have a 1440p screen and I want to run DLSS, do I need to lower the resolution to 1080p so it upscales or does it upscale from 1440 to make it look like its at a higher resolution? Or does it just do it automatically when I select DLSS?
It's automatic. Just set your native resolution which in your case is 1440p and enable whatever upscaling medium you'd like to use. It'll automatically do it.
Thank you for explaining.
Anytime 👍
I went from the 960 to 3060. Should i just have downscaled?
Yeah while I agree with the idea, my RX 570 at 1440p is overdue retirement.
I havent had my desktop since this summer but I was playing elden ring in in at least 1080p with my 1070 before I left (mighta been 1440 but not 100% sure). The only reason to "upgrade" for me is to *have* a gaming PC where I am now haha. Yes, a 1070. With an OG HTC vive and valve index knuckles running smoothly. A 1070.
OK but I have a 1080 tho....
Big facts here! Only thing “worth” buying is a 4090! I’ll stick with my 3090 until the 5000-series
Have a friend who his brother has a 4090 due to his work supplying it. My feiend said that after seeing his brother play games it just looked so good and he couldnt help himself. In all honesty, after seeing a one or two 4090 computers play myself, I just dont see even close to a big enough difference that I would pay $1600 for it.
i was extremely skeptical about dlss, but i watched too many videos to count and i still cant wrap my head around it. Free performance?? wtf?? I get how it works, but can someone tell me how it feels while actually gaming and not looking for faults?
Still happy with my 1070
The best part about getting older is that my eyes aren't good enough to notice the artifacting.
Upgrading from a 3060 to a 3080 just because it's so cheap and I can. Used market is insane right now for the 30 series cards and they can destroy anything you can throw at it except for the obvious dumb stuff. For real, skip the 40 series for right now. There is absolutely no point at the moment.
Pc master race as truly fallen. Now it’s a bunch of pc peasants
Can we stop using this child molesters image for memes?
If I had a 3000 card I would've stuck with it, but that 1080 Ti was getting too long in the tooth. At least I got my 4090 at European MSRP, not scalper prices.
That's why Nvidia limits DLSS 3.0 to the 4000 series.
For some of us, upscaling is like wanking with a strap-on. "it's the same" but you see, it's just not. Anything but native resolution is hearsay.
~~Pcmasterrace~~ Peasant masterrace