If my 1080 gives up in the near future, I’ll probably just give up AAA gaming. BG3 is literally the only game in the last 5 years I have loved which would require more than a potato to run.
Honestly I have a gtx1080 and I can run BG3 at close to ultra settings for around 30-60fps. I think my actual issue is my CPU I have an i7 3770 which came out in like 2012. And I have like ddr 3 ram from 10 years ago as well.
Same here but with an i7 6700K. If I happen to find a decent deal on something like a 12th gen i5 or i7 and chuck in an extra 16gb ram then I think I’d be all set for another couple of years. Although I still don’t desperately need an upgrade, everything works well enough except for maybe that one nightclub map in Ready or Not with more npcs than my cpu can keep up with.
My SO upgraded to a Ryzen 5600 and 6600 GPU from a 1050TI and some integrated old CPU and saw massive improvements for BG3. I don’t know the integrated CPU but I think that made the difference for this game. Loading in went from crashes or naked people to no issue at all.
Only reason I recently moved from that very card was to play with stable diffusion, otherwise yeah the ‘need’ to boost things for games and whatnot just isn’t there. Might be just that I’'ve never been a major twitch type gamer though, more Civ/StarCraft style.
Edit: my dumb, was actually a 1650S I last switched from. Pretty sure the before that was a 760TI and probably mixed it up somewhere between them
Yeah, I only have a 1080p monitor so the only times my 3060 feels limiting are with stable diffusion and in VR
So I’ve only used those tools on Linux and found it works quite well. Perhaps its a problem related to the windows environment?
The setups I saw kept demanding 8GB of vram at the time, the card I had was only 4 so it would constantly give an out memory state.
Oh the 1080ti has 11GB of vram.
Im going to switch my 1060 only becuz of wayland
Did the same thing. 1080ti to 6700xt. No regrets whatsoever, loving the Wayland experience.
Literally the same thing I did, switched from 1060 6gb to rx 6600 just for wayland.
It’s not worth it tbh
Is it not possible to run Wayland?
Wayland doesn’t play well with proprietary Nvidia drivers.
Manjaro stuck with X11, and people complain about it on the forums alot. I run Nvidia, and X11 while dead works just fine for me.
It does run, but not as fine as it would on AMD or Intel
50% chance it will be ok. It worked perfectly on arch linux with hyprland, but the same set up on NixOS is having some troubles
1080ti is still a beast
My 980 still plays anything I want fine
I fucking love my 1080ti and I will never, ever get rid of it. It’s part of my first custom gaming PC and I have promised to frame it whenever the time comes I can’t use it anymore.
My 1070 is doing fine, thank you very much. I haven’t felt any need to replace it.
My 660ti is finally showing its age…
I have a 670. Beat you by three months!
In all seriousness, they perform about the same. But I can’t really play any recent titles bar Indie games.
Still got life in her yet
I knew a guy at NVidia and got a screaming deal on a 980ti hybrid. Just replaced it with a used 2070 this year.
I replaced my 980ti two years ago. It was heartbreaking…
You don’t need more for d2r and it’s the newest game anyone needs to be playing.
💀 what goshdarn movie is this from
It’s from a video by the brilliant James Lee: https://youtu.be/A7PJQa1RfM0?si=_AAwpY_rJC1o-j4P
Kay
…
…
I have never felt more called out.
Mine is a FE and still chugging along great. I did replace the godawful stock cooler with an aftermarket cooler with a GIANT heatsink. Temps never go over 60 anymore.
Similarly, I repasted and replaced thermal pads on my 1080ti SC2 and it is still performing beautifully at 1440p High settings on most things
A 1080TI still plays every release at medium or higher settings. /shrug
Unless you’re worried about 4k or VR, I wouldn’t upgrade anyway.
If you care about refresh rate it matters, not a lot of people can stand 30-40fps with hard drops to single digits just to be able to play a game.
That’s how I akways played games :(
I’m curious. What game do you think drops to single digits fps on medium settings with a 1080TI?
I was playing Darktide on a 1060 with minimum 30fps recently, and that game is optimized like absolute trash.
Starfield.
1080 is the minimum card, the TI is decently more powerful (30%), but you’ve got to make concessions on medium to get 30fps, and there’s drops.
Sounds like I’ll never want to even play it lol
That’s Bethesda’s fault. There is no fucking reason that game can’t run well on a 1080ti for how mediocre it looks.
To be fair though, that’s a VERY new game and they cared so little about optimizing it that they went out and said “you’re probably going to need a new computer to play this” …
I hear the 1080ti runs Doom just fine 😛
My 1070 handled Doom Eternal just fine with pretty high settings. I’m sure it helps that I only use a single 1080p monitor for games, but it was still pretty enough for me.
*Edit - I picked up on the sarcasm after posting this reply. Oh well.
Every release means every release, and the requirements aren’t going to get lower. It’s a great card and I know people hate losing it, but it’s on its last legs and likely won’t be able to play new releases at all next year.
Cod MW2/3 are total crapshoots with frame rates, even on a 3080 and set to performance it can still just turn to crap. It seems to run more stable and on higher settings on 2070 laptop. I don’t understand. (I tried to get as much hardware running DMZ as possible for friends and family, lots of machines)
I do 4k and VR on my 1080Ti with no issues, on the highest settings, too. That said, I don’t do a lot of AAA gaming, so take that as you will.
Have a 1060 for vr.actually works fine
I actually prefer the crisp edges without as much post-processing effects sometimes. Source engine games look great to me, just minimal crisp and clean geometry. I find a lot of modern graphics distracting, but it depends on the game. I do love really pushing graphics for a game like Skyrim.
Modern game engines don’t use the amazing SSAA (super sampling anti aliasing). Most have post processing anti aliasing like FXAA or TXAA which always makes edges look fuzzy. Source engine is one of those that still supports super sampling
Yeah that’s exactly it, MSAA isn’t too bad but FXAA makes edges look pretty blurry. Temporal anti aliasing is also really blurry looking sometimes but gives the impression that the edges could be crisp.
My 980ti is still a toss up between amazing or mediocre performance. The big issue is that I bought it for £600 which is a lot of money to me, and new GPUs are 3 times that, or more.
Depends on the games. My 980TI can still rock 3440x1440 in most of the games I play.
The fact that what I play is mostly metroidvania shouldn’t be an issue, right? 😅
My 980ti still holds up pretty well at 1920x1440 (high-end CRT monitors were beautiful things, restart production you cowards) for most 3d games I play on Linux, but it is starting to have performance issues in some games, and I’m getting real sick and tired of the dumb shit Nvidia keeps pulling with their Linux drivers. The current driver gives me horrible black flickering in a lot of games, and of course they arbitrarily lock me out maxing out my CRT monitor (which don’t have a fixed resolution, only a balance of resolution vs refresh rate, and it keeps blocking me from a whole range of refresh rate/resolution combinations). So I confess I am starting to eye the higher-end AMD 6xxx GPUs, and I would definitely try and grab one as cheaply as I could if I ever got a 3440x1440 ultrawide.
Incidentally, how are ultrawides for having two or three windows open side-by-side at the same time?
Incidentally, how are ultrawides for having two or three windows open side-by-side at the same time?
Awesome. For work (even if I am a Linux system engineer) I need to use W11 due to corporate policy. I have two 34" in landscape and a 27" in portrait. I split the screens with FancyZones.
Time for my bad drawing skills, lol.
In order:
- SSH
- SSH
- SSH
- Outlook
- Edge for work
- Teams
- Firefox with YouTube running. Firefox is the only browser that allows for in-window full screen.
I see
I’m debating getting a 3440x1440 monitor for coding and because I hear they work well with tiling window managers (hence the question), it’s just annoying that I have almost no chances to try them out for free, and also the cost is enough that I wouldn’t get one without serious consideration first. Although you have nudged me a bit closer to “maybe I could get one without testing them first, if it’s second hand and cheap(er)”.
Also I’d be replacing my existing 27 inch LCD with it, and keeping the 4:3, 21 inch CRT, for a highly cursed monitor setup, where everything gets letterboxed or pillarboxed. And then to make things worse, I could grab a 16:10 monitor to put in portrait besides one of the other two, for maximum “what is 16:9 and why do I have black bars on everything”.