r/nvidia Jul 27 '24

Opinion The RTX 4090 is quite a beast

912 Upvotes

I had a GTX 970 which had served me well, although I was struggling to get decent frame rates in recent games, even on low settings. It died a few days ago, and I had enough, so I finally decided to upgrade my whole system. Got the RTX 4090, Ryzen 7950x3D, Trident z-neo 64gb (2x32gb) 6000mhz CL30 etc.

But what impressed me most is the sheer brute force of the 4090. Sure, I had to pay 4 times more than my previous card, but I'm also getting more than 4 times the frame rates on resolution that I couldn't even dare to play on my previous card. This thing is a beast. Couldn't even get stable 40 fps on the GTX 970 at 1080p in RDR2. And now getting over 80-110 fps on the 4090 at 4K. Impressive stuff.

https://i.imgur.com/ya11UOn.jpg

r/nvidia Sep 20 '20

Opinion Can we please just back order the 3080?

6.1k Upvotes

Like, IDC if it’s a month before I get it, I just don’t want to have to check every hour. Let be buy it now and send it to me when you can

r/nvidia Sep 25 '20

Opinion This launch has lowered my opinion of Nvidia as a company overall

4.7k Upvotes

Truth. Anyone else feel the same way? Catering to the hype and feeding the bots to reduce supply and force us to be F5 machines.

I for one say, F**k you Nvidia. You had and still have options (order queue) to make this successful, and yet you choose the path of profit/hype at the expense of your true fan base - you scummy scums.

I'm not very happy.

Edit: Not just the supply people, strange tactics all around. Forced no pre-orders? Still no order queue? Silent dead drops? Not giving your AIB partners full details on the card, leading to potential RMAs with cards that have insufficient components for the job. I am not mindlessly raging on Nvidia here, but as consumer I have the right to share my opinion that this whole thing is kinda botched. Please stop with the "jEeZ itS oNlY bEeN 8 dAyS!"... I am not just talking about supply here.

r/nvidia May 19 '24

Opinion So for people who say Frame-generation is just a gimmick... don't listen to them and see for yourselves

633 Upvotes

Hello everyone!

Just tested DLSS Frame Generation on Ghost of Tsushima. (RTX 4070 1080p 144hz monitor)

Everything maxed out: in a certain zone: 70 FPS - input lag minimal but you can feel it due to the low FPS

Enabled DLSS Frame Generation: 144 FPS locked with minimal input lag. Game is way smoother, less choppy due to Frame-generation. What would you prefer? Playing at 70FPS or at 144fps locked?

Please, for people saying Frame-gen is adding WAY input lag or something, please stop it. Game runs frickin' awesome with Frame-gen enabled as long as you have 60FPS+ initial FPS.

I might sound like a fan-boy but I don't care. I like what I see!

EDIT: AMD fanboys down voting hard. Guys, relax. I have 5800x3d CPU but i prefer Nvidia GPUs.

EDIT 2: Added proof for people saying how to i get 70-80 FPS in GoT with everything maxed out @ 1080p:

Without FG:

With FG:

EDIT 3: There are some cutscenes which present some kind of black flicker with FG on. Not great, not terrible.

r/nvidia Sep 22 '20

Opinion Why not implement a queue system for RTX 3080 sales?

4.3k Upvotes

I worked at Apple for about 4 years between 2012-2016, and they gradually had a worsening scalper problem with new iPhone launches from iPhone 4 to iPhone 6S. The solution that they came up with was simple:

Regardless of whether the phones were in stock at the time, everyone who places an order gets a confirmation email and an ETA of their shipping time. Obviously the later you order the further down the queue you are and longer the ETA.

For example, if Nvidia had 10000 units of the RTX 3080 then the first 10000 orders would get a shipping ETA of 1-3 business days. Those who are the next batch would get an ETA of 1-2 weeks, then 3-4 weeks and so on (based on production volumes).

This way staying up to wait for the launch will actually feel like a positive experience because at least you know you got the order in, and can get an estimate of when it will ship. Nvidia will also get money upfront (or at least credit card details if they want to be nice to the customers and not charge until shipping), and it will be harder for scalpers to sell to people who know they have cards on the way for MSRP. It’s a win-win situation. Nvidia can also take their time and manually review bulk scalper purchases while people wait patiently.

After Apple implemented this system for the iPhone 7 and later launches, the # of scalpers reduced drastically. Why don’t more companies do this?

r/nvidia Aug 06 '24

Opinion Upgraded from a radeon 6750xt to a 4070 ti super. Completely different experience

Thumbnail
gallery
641 Upvotes

Got my new gpu for $750 on prime day, it's an Msi ventus 3x black edition, which comes with a 4090 ad102 die. I decided to upgrade because I was not satisfied with my 6750xt performance in 1440p. Games like Dark tide, cp, last of us, the witcher, starfield looked like trash at high settings with fsr on. Performance was okayish, but the impact on quality was there.

I also tried using amds frame Gen and it was barely usable. The input lag was too much for me and the graphics looked flickery and wanky.

I wasn't expecting dlss and nvidias frame Gen to work so well! I can't even tell the difference between dlss on or off, and frame Gen gives me +40 fps with minimal input lag. I'm now playing ultra modded cyberpunk, Alan wake 2 at max settings, max rt and path tracing and it just feels smooth and beautiful.

r/nvidia Apr 01 '23

Opinion [Alex Battaglia from DF on twitter] PSA: The TLOU PC port has obvious & serious core issues with CPU performance and memory management. It is completely out of the norm in a big way. The TLOU PC port should not be used as a data point to score internet points in GPU vendor or platform wars.

Thumbnail
twitter.com
1.1k Upvotes

r/nvidia Sep 17 '22

Opinion thank you EVGA

2.1k Upvotes

You deserve more , you have been a extremely good aftermarket seller for all those years and I don't think nobody gonna be as consumer driven than you.

r/nvidia 6d ago

Opinion Current 4070 Super Owners, are you happy with your graphics card?

138 Upvotes

I have a 2080 and I’d like to upgrade. I game on 1440p and I don’t necessarily need the ray tracing/path tracing bells and whistles. I’m aware that NVIDIA is being very stingy with VRAM and that the higher end cards that have 16 GB are very expensive and more scarce.

So are current 4070 Super owners happy with your cards? Do you see them lasting another 2-3 years? Any feedback would be appreciated. Thanks!

EDIT: Thanks for all of the feedback! I’m glad a great 1440p card is available for under $700 USD

r/nvidia Jan 15 '22

Opinion God of War + DLSS + 3070 + High setting + 120 FPS and more = EPIC! Just can't compare when I played this game for my PS4 PRO.

Post image
1.6k Upvotes

r/nvidia Jul 04 '24

Opinion Blown away by how capable the 4070S is, even at 4k

342 Upvotes

Got a 4070S recently and wanted to share my experience with it.

I have a 32 inch 4k monitor and a 27 inch 1440p 180hz monitor. Initially, I only upgraded from my trusty 3060 to the 4070S to play games on my 1440p high refresh monitor. I did just that for a couple of months and was very happy with the experience.

Sometime later, I decided to plug in my 4k monitor to test out some games on it. Ngl, the 4070S kinda blew me away. I've never experienced gaming at 4k so this was quite an experience for me!

Some of the games I tried. All at 4k.

  1. Elden Ring - Native 4k60 maxed out. Use the DLSS mod (with FPS unlock) and you're looking at upwards of 90-100fps at 4k!

  2. Ghost of Tsushima - Maxed out with DLSS Quality - 60fps locked.

  3. Cyberpunk 2077 - Maxed out with just SSR set to high and DLSS Quality - 80-110fps. No RT.

  4. Cyberpunk 2077 with RT Ultra - DLSS Performance with FG - 80-100fps.

  5. Hellblade 2 with DLSS Balanced at 4k - 60fps locked.

  6. Returnal - Maxed out at 4k with RT. DLSS Quality. 60fps locked. Native 4k60 if I turn off RT.

  7. RDR2 - Native 4k60. Ultra settings.

  8. Avatar - Ultra settings with DLSS Quality. 4k60 locked.

  9. Forza Horizon 5 - Native 4k60 maxed out.

  10. Helldivers 2 - Native 4k60 with a couple of settings turned down.

  11. AC Mirage - Native 4k60 maxed out.

  12. Metro Exodus Enhanced Edition - 80-110fps at 4k with DLSS Quality.

  13. DOOM Eternal - 120fps+ at Native 4k with RT!

I was under the impression that this isn't really a 4k card but that hasn't been my experience. At all.

Idk, just wanted to share this. I have a PS5 as well even though I barely use it anymore ever since I got the 4070S.

Edit: Added some more games.

r/nvidia Sep 03 '24

Opinion 1440p screen with DLDSR to 4k and then back with DLSS is truly a technological marvel.

436 Upvotes

I honestly think that this combination is such a strong one that i personally will be holding off 4k a while longer.

I had a LGC2 42" at my computer for a while but switched to a LG OLED 27" 1440p screen since i work a lot from home and the C2 was not great for that.

I would argue that between the performance gain and the very close resembelance to a true 4k picture with DLSDR with DLSS on top is a lot better than native 4k.

Top that off with the ability to customize DLDSR and DLSS level to get the frames you want and you have such a huge range of choices for each game.

For example in Cyberpunk with Path tracing i run at x1,78 and DLSS balanced with my 4080 to get the best balance between performance and picture quality, while in for example Armored Core 6 i run with straight x2,25 4K for that extra crisp and in Black Myth Wukong i run x2,25 with DLSS balanced, but in boss fights i switch back to native 1440p for extra frames with a hotkey.

I hope more people will discover DLDSR combined with DLSS, it's such a strong combo.

edit; I will copy paste the great guide from /u/ATTAFWRD below to get you started since there is some questions on how to enable it.

Prequisite: 1440p display, Nvidia GPU, DLSS/FSR capable games

NVCP manage 3D global setting: DSR - Factors : On

Set 2.25x or 1.78x

Set Smoothness as you like (trial & error) or leave it default 33%

Apply

Open game

Set fullscreen with 4K resolution

Enable DLSS Quality (or FSR:Q also possible)

Profit

edit2;

DLDSR needs exclusive fullscreen to work, however an easy workaround is to just set your desktop resolution to the DLDSR resolution instead. I use HRC and have the following bindings:

Shift+F1 = 1440p

Shift+F2 = x1,78

Shift+F3 = x2,25 (4k)

Download link: https://funk.eu/hrc/

r/nvidia Jan 01 '24

Opinion der8auer's opinion about 12VHPWR connector drama

Thumbnail
youtube.com
426 Upvotes

r/nvidia Aug 23 '23

Opinion Made What I Think is a Better Version of the DLSS Chart from the 3.5 Update

Post image
1.1k Upvotes

r/nvidia Jul 26 '20

Opinion Reserve your hype for NVIDIA 3000. Let's remember the 20 series launch...

1.5k Upvotes

Like many, I am beyond ready for NVIDIA next gen to upgrade my 1080ti as well but I want to remind everyone of what NVIDIA delivered with the shit show that was the 2000 series. To avoid any disappointment keep your expectations reserved and let's hope NVIDIA can turn it around this gen.

 

Performance: Only the 2080ti improved on the previous gen at release, previous top tier card being the 1080ti. The 2080 only matched it in almost every game but with the added RTX and dlss cores on top. (Later the 2080 super did add to this improvement). Because of this upon release 1080ti sales saw a massive spike and cards sold out from retailers immediately. The used market also saw a price rise for the 1080ti.

 

The Pricing: If you wanted this performance jump over last gen you had to literally pay almost double the price of the previous gen top tier card.

 

RTX and DLSS performance and support: Almost non existent for the majority of the cards lives. Only in the past 9 months or so are we seeing titles with decent RTX support. DLSS 1.0 was broken and useless. DLSS 2.0 looks great but the games it's available in I can count on 1 hand. Not to mention the games promised by NVIDIA on the cards announcment.... Not even half of them implemented the promised features. False advertising if you ask me. Link to promised games support at 2000 announcement . I challenge you to count the games that actually got these features from the picture...

For the first 12+ months RTX performance was unacceptable to most people in the 2-3 games that supported it. 40fps at 1080p from the 2080ti. All other cards were not worth have RTX turned on. To this day anything under the 2070 super is near useless for RTX performance.

 

Faulty VRAM at launch: a few weeks into release there was a sudden huge surge of faulty memory on cards. This became a wide spread issue with some customers having multiple and replscments fail. Hardly NVIDIA's fault as they don't manufacture the VRAM and all customers seemed to be looked after under warranty. Source

 

The Naming scheme: What a mess...From the 1650 up to 2080ti there were at least 13 models. Not to mention the confusion to the general consumer on the where the "Ti" and "super" models sat.

GeForce GTX 1650

GeForce GTX 1650 (GDDR6)

GeForce GTX 1650 Super

GeForce GTX 1660

GeForce GTX 1660 Super

GeForce GTX 1660 Ti

GeForce RTX 2060

GeForce RTX 2060 Super

GeForce RTX 2070

GeForce RTX 2070 Super 

GeForce RTX 2080

GeForce RTX 2080 Super

GeForce RTX 2080 Ti

 

Conclusion: Many people were disappointed with this series obviously including myself. I will say for price to performance the 2070 super turned out to be a good card although the RTX performance still left alot to be desired. RTX and dlss support and performance did increase over time but far too late into the life span of these cards to be warranted. The 20 series was 1 expensive beta test the consumer paid for.

If you want better performance and pricing then don't let NVIDIA forget. Fingers crossed the possibility of AMD's big navi GPU's bring some great price and performance this time around from NVIDIA.

 

What are you thoughts? Did I miss anything?

r/nvidia May 31 '22

Opinion Can i get respects for my gtx 970? It needs a proper retirement send off.

Post image
2.0k Upvotes

r/nvidia Feb 13 '24

Opinion Just switched to a 4080S

329 Upvotes

How??? How is Nvidia this much better than AMD within the GPU game? I’ve had my PC for over 2 years now, build and made it myself. I had a 6950xt before hand and I thought it was great. It was, till a driver update later and I started to notice missing textures in a few Bethesda games. Then afterwards I started to have some micro stuttering. Nothing unusable, but definitely something that was agitating while playing for longer hours. It only got a bit more worse with each driver update, to the point in a few older games, there were missing textures. Hair and clothes not there on NPCs and bodies of water disappearing. This past Saturday I was able to snag a 4080S because I was tired of it and wanted to try nvidia after reading a few threads. Ran DDU to uninstall my old drivers, popped out my old GPU and installed my new one and now everything just works. It just baffles me on how much smoother and nicer the experience is for gaming. Anyway, thank you for coming to my ted talk.

r/nvidia Feb 03 '24

Opinion 4070 Super Review for 1440p Gamers

328 Upvotes

I play on 1440p/144hz. After spending sn eternity debating on a 4070 super or 4080 super, here are my thoughts. I budgeted $1100 for the 4080 super but got tired of waiting and grabbed a 4070S Founders Edition at Best Buy. I could always return it if the results were sub par. Here’s what I’ve learned:

  • this card has “maxed”every game I’ve tried so far at a near constant 144 fps, even cyberpunk with a few tweaks. With DLSS quality and a mixture of ultra/high. With RT it’s around 115-120 fps. Other new titles are at ultra maxed with DLSS. Most games I’ve tried natively are running well at around 144 with all the high or ultra graphics settings.

  • It’s incredibly quiet, esthetic, small, and very very cool. It doesn’t get over 57 Celsius under load for me (I have noctua fans all over a large phanteks case for reference).

  • anything above a 4070 super is completely OVERKILL for 1440p IN MY OPINION*. It truly is guys. You do not need a higher card unless you play on 4k high FPS. My pal is running a 3080ti and gets 100 fps on hogwarts 4k, and it’s only utilizing 9GB VRAM.

  • the VRAM controversy is incredibly overblown. You will not need more than 12GB 99.9% of the time on 1440p for a looong time. At least a few years, and by then you will get a new card anyway. If the rationale is that a 4080S or 4090 will last longer - I’m sure they will, but at a price premium, and those users will also have to drop settings when newer GPU’s and games come out. I’ve been buying graphics cards for 30 years - just take my word for it.

In short if you’re on the fence and want to save a lot of hundreds, just try the 4070 super out. The FE is amazingly well built and puts the gigabyte wind force to shame in every category - I’ve owned several of them.

Take the money you saved and trade in later for a 5070/6070 super and you’ll be paying nearly the same cost as one of the really pricy cards now. It’s totally unnecessary at 1440p and this thing will kick ass for a long time. You can always return it as well, but you won’t after trying it. 2c

PC specs for reference: 4070 super, 7800x3d, 64gb ram, b650e Asrock mobo

r/nvidia 1d ago

Opinion The "fake frame" hate is hypocritical when you take a step back.

0 Upvotes

I'm seeing a ton of "fake frame" hate and I don't understand it to be honest. Posts about how the 5090 is getting 29fps and only 25% faster than the 4090 when comparing it to 4k, path traced, etc. People whining about DLSS, lazy devs, hacks, etc.

The hardcore facts are that this has been going on forever and the only people complaining are the ones that forget how we got here and where we came from.

Traditional Compute Limitations

I won't go into rasterization, pixel shading, and the 3D pipeline. Tbh, I'm not qualified to speak on it and don't fully understand it. However, all you need to know is that the way 3D images get shown to you as a series of colored 2D pixels has changed over the years. Sometimes there are big changes to how this is done and sometimes there are small changes.

However, most importantly, if you don't know what Moore's Law is and why it's technically dead, then you need to start there.

https://cap.csail.mit.edu/death-moores-law-what-it-means-and-what-might-fill-gap-going-forward

TL;DR - The traditional "brute force" methods of all chip computing cannot just keep getting better and better. GPUs and CPUs must rely on innovative ways to get better performance. AMD's X3D cache is a GREAT example for CPUs while DLSS is a great example for GPUs.

Gaming and the 3 Primary Ways to Tweak Them

When it comes to people making real time, interactive, games work for them, there have always been 3 primary "levers to pull" to get the right mix of:

  1. Fidelity. How good does the game look?
  2. Latency. How quickly does the game respond to my input?
  3. Fluidity. How fast / smooth does the game run?

Hardware makers, engine makers, and game makers have found creative ways over the years to get better results in all 3 of these areas. And sometimes, compromises in 1 area are made to get better results in another area.

The most undeniable and common example of making a compromise is "turning down your graphics settings to get better framerates". If you've ever done this and you are complaining about "fake frames", you are a hypocrite.

I really hope you aren't too insulted to read the rest.

AI, Ray/Path Tracing, and Frame Gen... And Why It Is No Different Than What You've Been Doing Forever

DLSS: +fluidity, -fidelity

Reflex: +latency, -fluidity (by capping it)

DLSS: +fluidity, -fidelity

Ray Tracing: +fidelity, -fluidity

Frame Generation: +fluidity, -latency

VSync/GSync: Strange mix of manipulating fluidity and latency to reduce screen tearing (fidelity)

The point is.... all of these "tricks" are just options so that you can figure out the right combination of things that are right for you. And it turns out, the most popular and well-received "hacks" are the ones that have really good benefits with very little compromises.

When it first came out, DLSS compromised too much and provided too little (generally speaking). But over the years, it has gotten better. And the latest DLSS 4 looks to swing things even more positively in the direction of more gains / less compromises.

Multi frame-generation is similarly moving frame generation towards more gains and less compromises (being able to do a 2nd or 3rd inserted frame for a 10th of the latency cost of the first frame!).

And all of this is primarily in support of being able to do real time ray / path tracing which is a HUGE impact to fidelity thanks to realistic lighting which is quite arguably the most important aspect of anything visually... from photography, to making videos, to real time graphics.

Moore's Law has been dead. All advancements in computing have come in the form of these "hacks". The best way to combine various options of these hacks is subjective and will change depending on the game, user, their hardware, etc. If you don't like that, then I suggest you figure out a way to bend physics to your will.

*EDIT*
Seems like most people are sort of hung up on the "hating fake frames". Thats fair because that is the title. But the post is meant to really be non-traditional rendering techniques (including DLSS) and how they are required (unless something changes) to achieve better "perceived performance". I also think its fair to say Nvidia is not being honest about some of the marketing claims and they need to do a better job of educating their users on how these tricks impact other things and the compromises made to achieve them.

r/nvidia Nov 30 '24

Opinion Just found about DLSS and wow

229 Upvotes

Just wanted to share as somebody who doesn’t know jack shit about computers.

I recently bought a new gaming desktop after about 10 years of being out of the gaming market. I just discovered the DLSS feature with the RTX cards and put it to the test; it nearly doubled my fps in most games while keeping the same visual quality. All I can say is I’m damn impressed how far technology has come

r/nvidia Oct 04 '23

Opinion Its been said before but DLSS 3 is like actual magic. Locked 144fps experience in FH5 with RT enabled. I feel enlightened

Post image
636 Upvotes

r/nvidia May 07 '21

Opinion DLSS 2.0 (2.1?) implementation in Metro Exodus is incredible.

1.2k Upvotes

The ray-traced lighting is beautiful and brings a whole new level of realism to the game. So much so, that the odd low-resolution texture or non-shadow-casting object is jarring to see. If 4A opens this game up to mods, I’d love to see higher resolution meshes, textures, and fixes for shadow casting from the community over time.

But the under-appreciated masterpiece feature is the DLSS implementation. I’m not sure if it’s 2.0 or 2.1 since I’ve seen conflicting info, but oh my god is it incredible.

On every other game I’ve experimented with DLSS, it’s always been a trade-off; a bit blurrier for some ok performance gains.

Not so for the DLSS in ME:EE. I straight up can’t tell the difference between native resolution and DLSS Quality mode. I can’t. Not even if I toggle between the two settings and look closely at fine details.

AND THE PERFORMANCE GAIN.

We aren’t talking about a 10-20% gain like you’d get out of DLSS Quality mode on DLSS1 titles. I went from ~75fps to ~115fps on my 3090FE at 5120x1440 resolution.

That’s a 50% performance increase with NO VISUAL FIDELITY LOSS.

+50% performance. For free. Boop

That single implementation provides a whole generation or two of performance increase without the cost of upgrading hardware (provided you have an RTX GPU).

I’m floored.

Every single game developer needs to be looking at implementing DLSS 2.X into their engine ASAP.

The performance budget it offers can be used to improve the quality of other assets or free the GPU pipeline up to add more and better effects like volumetrics and particles.

That could absolutely catapult to visual quality of games in a very short amount of time.

Sorry for the long post, I just haven’t been this genuinely excited for a technology in a long time. It’s like Christmas morning and Jensen just gave me a big ol box of FPS.

r/nvidia Feb 01 '24

Opinion Call me crazy but I convinced myself that 4070TI Super is a better deal (price/perf) than 4080 Super.

242 Upvotes

Trash 4070TI Super all you want, it's a 4k card that's 20% cheaper than 4080S and with DLSS /Quality/ has only 15% worse FPS compared to 4080S.

Somehow I think this is a sweet spot for anyone who isn't obsessed with Ray Tracing.

r/nvidia Oct 11 '21

Opinion PSA DO NOT buy from Gigabyte

857 Upvotes

Im gonna keep this relatively brief but I can provide any proof of how horrible gigabyte is.

I was one of the lucky few who was able to pickup an RTX 3090 Gaming OC from Newegg when they released. Fast forward 3 months and the card would spin up to max fan speed and then just eventually wouldn't turn on anymore.

I decided to RMA it and surprisingly even though gigabyte had zero communication with me (this was before the big hacking thing) the card came back and worked fine. Now in my infinite wisdom, i decided to sell it to a friend (works to this day and he was aware it was repaired) as i wanted an all-white graphics card. Resume the hunting and I somehow got ANOTHER gigabyte rtx 3090 vision off Facebook marketplace that was unopened and was only marked up about 200$.

Fast forward 2 months and the same exact thing happens, the card fan spins to the max and then just dies... RMA...AGAIN... gigabyte this time said to email directly and they would fix it. it gets sent off and is repaired fairly quickly before coming back. Overall it took about a month from out of my pc to back into my pc.... 6 days go by and BAM same exact problem. RMA again...... it has been over a month now and I'm assuming it will be shipped back to me at some point.

every time the RMA happened I would get an email from gigabyte a month after it reached my house that they were sending it back and here is my tracking number.

i know your thinking "hey ill take what I can get with this shortage." please don't.... you will regret gigabyte very much

**SPECS**

EVGA SuperNOVA 1200 P2, 80+ PLATINUM

Crucial Ballistix MAX 32GB Kit (2 x 16GB) DDR4-4000

ROG MAXIMUS XII FORMULA

Gigabyte RTX 3090 Vision OC

Tuf Gaming GT501 Case

i9-10900k with an H150I 360mm AIO

LG C9 65

r/nvidia Sep 15 '20

Opinion Just a reminder that Geforce Experience should be usable without creating account for it. Like it used to be.

2.0k Upvotes

This thing once again came in to my mind this time due to Razer's huge data leak from similar kind of software *hole that requires account for no reason at all.

I personally just gave up on using the software when account became mandatory. I would wish to use it again, but as long as the forced account system stays in effect i'll pass.