Nvidia: AMD's DirectX 11 lead doesn't matter

Status
Not open for further replies.
nVidia has more to worry about than mkaing a more powerful gpu, they need to address the eyefinity which for a gamer....is all the reason one needs to buy 3+ monitors...very cool and I would call it a killer app for the gamer.


now if monitor vendors will start making slim bezel (<3mm) 1080p 30" screens!
 
Nvidia just don't know what to say, if they can say any thing....
The truth is that AMD made a huge advance, even without DX11 is kicking Nvidia's top grafic cards... in perfomance and price, and don't forget about HD video and eyefinity...
And about what Nvidia says about AMD DX11, just read what Nvidia told us about dx10 back in 2007, when dx10 whas only ment to run on vista and nobody even whanted(whats) to hear about Windows Vista.. and of course not a single game whas ready for Vista and Dx10, when Nvidia released the DX10 grafic cards...
http://www.nvidia.com/object/IO_41448.html
 
Guest said:
nVidia has more to worry about than mkaing a more powerful gpu, they need to address the eyefinity which for a gamer....is all the reason one needs to buy 3+ monitors...very cool and I would call it a killer app for the gamer.

now if monitor vendors will start making slim bezel (<3mm) 1080p 30" screens!

we'll see them very soon. i think samsung said that they'll release something like that but i'm not sure.
 
Kibaruk said:
The one saying "I have a 4xxx and looks like postcard showoff, against the 285 that flies through games" according to the benchmarks at 8x MSAA resolution they run with like 5 fps difference, not 40 that is what it requires to look like postcard game.

First off I have 2 GTX280's (not GTX285)
Secondly...
http://www.bit-tech.net/hardware/graphics/2009/11/24/multi-gpu-round-up/7

As all the hoo-hah seems to be regarding the top-end cards then pay attention to benchmarks at 1920 and 2560.

If I run card "A" in game for 5 minutes, and during that time it runs at a MINIMUM framerate of 15 fps for 30 seconds and the rest of the time runs at 200 fps then the AVERAGE framerate is 181.5 fps.
If card "B" runs the at 60 fps for the entire 5 mins then it's average framerate is 60 fps.

By your logic card "A" gives you the better gameplay experience because the average would be three times higher.

I run two systems that are broadly comparable in performance, one nvidia and one ATI and while some games run better on one card/s or the other I have found that the GTX's work with a new game out of the box generally- not so much with ATI's offerings (Saboteur and Resident Evil 5 for example) and the minimum framerate is the telling factor for good gameplay.

Secondly...

Neither of my systems are primarily used for gaming. The bulk of their usage comes from video transcoding and graphics design. The new nVIDIA architecture (if and when it arrives) holds a great deal of promise as opposed to the HD3xxx > HD4xxx > HD5xxx incremental tweaking of an architecture that is heading towards a dead end.
ATI selling cards, GREAT, channel some of the funds into R&D ! because selling the next best thing based on the fact that it can push out 120 fps instead of a measly 100 fps seems shortsighted at best.
As for when the green teams offering is supposed to rear its head I thought it had been established that the timeframe was Q1 2010.
As for nVidia losing out to AMD...If the HD5870/5970 and GF100 were released at the same time they would be sharing revenue. As it stands AMD get nearly 100% of the enthusiast market for a few months, interest wanes after initial gpu-fever, nVidia launch GF100 with only AMD refreshes and AIB "special edition"/OC as competition and makes it's money in Q1 2010 instead of Q4 2009- assuming GF100 outperforms the HD5870!. This scenario is just a continuation of the HD3xxx > G80/92 > HD4xxx > GTX200 theme. When was the last time AMD/ATI and nVIDIA went head-to-head by releasing a whole new product line each at the same time?
 
I recently pulled my nvidia card and put in a new AMD card, and I have to say the AMD card definitely has much more kick than the nvidia that is comparable to it. I was quite surprised by the results.
 
Nvidia's comments are totally damage control. At this point they have dropped the ball and for this they will pay for the time being. Will this always be the case? Well, we shall wait and see upon the arrival of Fermi. Fermi may help or it may hurt them depending on the prices of the cards they will release in their next line-up. It sounds like what they (Nvidia) are doing could turn into something big if it succeeds, but if Fermi doesn't it could hurt them big time. Nvidia has been the leader in video cards for a long time, but AMD seems to have knocked them off their perch (or at least rattled their cage) with the 4XXX series of ATi cards.

I have to give ATi/AMD a hand for the release of their DirectX 11 compatible cards (although it's a little early in the game for them) simply because more DirectX 11 games can be tested with them, and thus be out on the market. Like I said earlier, it's way too early in the game to predict a winner between Nvidia's Fermi and ATi's DirectX 11 cards. I guess after the first half of 2010 we'll know for sure who's plan worked better.

***I have a 1 GB Radeon HD 5770 on the way and I'm curious to see what I can do with it.***
 
Say this
Nvidia wants fermi is one another G80/90 remarked chip
but they slammed into dx11
 
Status
Not open for further replies.
Back