Approvals 5/3 ▼
twdll (1) - 3898d ago Cancel
cl1983 (2) - 3898d ago Cancel
derkasan (2) - 3898d ago Cancel
120°

AMD Hawaii GPU to undercut Nvidia's GTX 780

If all the rumours surrounding AMD’s new line of graphics cards are to be believed, this Q4 is going to be a bloodbath in the graphics card sector.

Create Report !X

Add Report

Reports

+ Updates (1)- Updates (1)

Updates

Changed from Pending to Approved
Community3898d ago
mewhy323898d ago

AMD is on a role now. I mean they're suddenly getting millions upon millions of chips out the door with the PS4 and xbone abeit less power gpu for xbone. Now they're bustin' up nVidia. Run with it!!!!!

bicfitness3898d ago

Price has never been the issue with AMD. Their driver-support - or lack therof - is. I waited 6 months for my last AMD driver CF 7990, before just swapping for Nvidia cards. Doesn't matter how cheap your product is if it doesn't have any software support.

Perhaps a greater marketshare and console foothold will smarten them up a bit. I'm not a brand loyalist - consoles, PC parts or otherwise - so if they could straighten themselves out, I'd build my next rig with AMD GPUs for sure.

infamous-butcher3897d ago

why buy a dual gpu, when you know that there are issues with the drivers for them?

bicfitness3897d ago

For enthusiast level performance obviously. Why would you even ask that?

jeffgoldwin3897d ago

@mewhy32

I dunno if you've never watched pc hardware before, but its a very normal cycle and has been for over a decade+ where they leap frog each other back and forth.

Up next in the news, NVidia releases new gpu faster than blah, blah, blah.

steve30x3898d ago

If they can match the GTX780 but at a reasonable price then I will go back to AMD / ATI. I havent owned an AMD GPU since I owned the miserable HD4870 for a month. That thing was aweful

awi59513898d ago

MY 4870 was awesome i dont know what you are talking about. I had that card for like 4 years and it played all games on max. You must have bought a cheap one that ran too hot because those cards had high temps. I had that card untill BF3 came out at that point i was rocking 3 of them in my pc and it still ran BF3 at 30 fps on ultra. I upgraded just for direct x 11 effects.

sinjonezp3898d ago

As much as I praise amd for their price/performance solutions, there driver updates hold me back from switching from nvidia. If they were to update as well as nvidia they could take the market share swiftly.

Codeman4203898d ago

Agreed, the drivers for Nvidia cards are so much better than ATI's, thats the main reason i got rid of two 5970's and went to one 660ti.

jeffgoldwin3897d ago

Nvidia cards run cooler too.

GentlemenRUs3898d ago

As long as its cheaper then GeForce, I'll start leaning towards AMD again.

hollabox3898d ago

Is it me or are graphics too damn expensive? Bring back the days when top of the line graphics cards cost $300 rather than $600 plus. My personal limit for graphics cards is $400, never paid more than that, don't see the value in return with anything higher.

papashango3898d ago

$400 will buy you a 7970...

that's high end bro.

hollabox3897d ago

Sorry my card is still high end, using Nvidia GTX 670 overclocked to 1300 boost clock. I'm looking to go back to AMD after having so many driver issues with Nvidia after all my friends claimed they got their crap together. But like I said paying $600 for one of the newest high end video cards I don't see the return on investment since I upgrade about every 18 months.

papashango3897d ago

I actually own this sapphire 7970. got it for $400

I've yet to come across a driver issue. Probably the most solid GPU i've purchased.

I've owned

x1800xt
HD 3870
HD 4970
HD 5870
HD 6970
HD 7970

Only issue I had was with my 3870. I would freeze ingame often and had to hard boot my pc. This was back in 05ish. AMD drivers have been good to me since.

hollabox3897d ago (Edited 3897d ago )

I hated my ATi 3870 as well, no issues regarding crashes but it was so slow in DX10 games, DX9 it performed decent. I have owned the following graphics cards listed below, some I upgraded because of slow performance and missing features, some died, others either ran too hot or too loud with no or limited aftermarket cooling.

AMD/ATI
ATI 9000 Pro
ATI 9700 Non Pro
ATI AIW800
ATI X1800 XT
ATI 3870
ATI 4870
AMD 6970

Nvidia
Riva TNT 2
Geforce 2 MX
Geforce 3 TI
Geforce TI4400
Geforce 6800 GT
Geforce GTX 460
Geforce GTX 670

Voodoo 3

Out of all of my graphics cards I probably like the ATI 4870 the most. Great framerates, ran cool at 54C gaming, and not too expensive. In fact this card is still being used in a relative computer. AMD 9700 was my 2nd best or atleast compared to my failed TI4400. Third best is probably my GTX 670. When its not having random driver crashes, crash to desktop driver failed to respond its still pretty damn fast.

Worst cards Geforce GTX 6800 GT from Gigabyte. It was fast but this card only lasted 6 months before it died, thats pretty bad for a $400 graphics card that only lasted half a year. 2nd worst Geforce TI 4400, died on me in 8 months, shoddy performance with anisotropic filtering enabled. Third worst is probably my Voodoo 3 GC. This card lacked features, 256X256 max textures resolution, couldn't do true 32 bit color even though GLIDE was pretty good at faking it.

papashango3897d ago

A fellow voodoo owner and a 3870 owner.

I got hooked on PC gaming on the voodoo 2. I've gone through my fair share of nvidia cards but none as special as that voodoo.

Its rare to meet other 3870 owners. considering the reign of the 8800 series

Good times

+ Show (2) more repliesLast reply 3897d ago
awi59513896d ago

Nividia and ATI both suck i remember the days when you could crossfire and SLI 3 or 4 cheap cards and kick the highend cards asses. Well NIvida and ATI teamed up to put a stop to that crap so now cards only have one sli or crossfire link. Thats a bunch of bull crap we should have to option to link more than 2 cards. ATI and nivida are full of crap for that it was a total dick move they both agreed on.

Show all comments (20)
330°

Nvidia DLSS 3.7 drives a further nail in the coffin of native performance

Nvidia DLSS 3.7 is the latest update to the long-running AI upscaling technology, and it further shows native performance doesn't matter.

Create Report !X

Add Report

Reports

+ Updates (1)- Updates (1)

Updates

Changed from Pending to Approved
Community3d ago
DustMan3d ago

I think hardware development is at a point where they need to figure out how to draw less power, These beefy high end cards eat wattage, and I'm curious if using DLSS & AI in general will lower the power draw. It would seem like the days of just adding more VRAM & horsepower is over. Law of diminishing returns. Pretty soon DLSS/FSR will be incorporated into everything, and eventually the tech will be good enough to hardly notice a difference if at all. AI is the future and it would be foolish to turn around and not incorporate it at all. Reliance on AI is only going to pick up more & more.

Tapani3d ago (Edited 3d ago )

DLSS certainly lowers power consumption. Also, the numbers such as the 4090 at 450W does not tell you everything, most of the time the GPU stays between 200-350W in gameplay, which is not too different from the highest end GPU of 10 years ago. Plus, today you can undervolt + OC GPUs by a good margin to keep stock performance while utilizing 80% of the power limit.

You can make the 4090 extremely power efficient and keep 90% of its performance at 320W.

However, in today's world the chip manufacturing is limited by physics and we will have power increases in the next 5-10 years at the very least to keep the technology moving forward at a pace that satisfies both businesses and consumers.

Maybe in 10 years we have new tech coming to the markets which we are yet to invent or perhaps we can solve existing technologies problems with manufacturing or cost of production.

On the other hand, if we were to solve the energy problem on earth by utilizing fusion and solar etc. it would not matter how much these chips require. That being said, in the next 30-40 years that is a pipedream.

MrBaskerville3d ago

I don't think fusion is the way forward. It will mosy likely be too late when it's finally ready, meaning it will probably never be ready. Something else might arrive before though and then it becomes viable.

Firebird3602d ago

We need to stop the smear campaign on nuclear energy.
We could power everything forever if we wanted too.

Tacoboto3d ago

PS4 Pro had dedicated hardware in it for supporting checkerboard rendering that was used significantly in PS4 first party titles, so you don't need to look to PC or even modern PC gaming. The first RTX cards released nearly 6 years ago, so how many nails does this coffin need?

InUrFoxHole3d ago

Well... its a coffin man. So atleast 4?

Tacoboto3d ago

PSSR in the fall can assume that role.

anast3d ago

and those nails need to be replaced annually

Einhander19723d ago

I'm not sure what the point you're trying to make is, but PS4 Pro was before DLSS and FSR, and it still provides one of the highest performance uplifts while maintaining good image quality.

DLSS is it's own thing but checkerboarding om PS5 still is a rival to the likes of FSR2.

Tacoboto3d ago

Um. That is my point. That there have been so many nails in this "native performance" coffin and they've been getting hammered in for years, even on PS4 Pro before DLSS was even a thing.

RaidenBlack2d ago

Don't know what's OP's point is either but ... checkerboard rendering was good enough for its time but in terms of image quality its wayy behind what's DLSS 3 or FSR 3 is currently offering.
The main point of the article and what OP missed here is that DLSS 3.7 is soo good that its nearly undisguisable from native rendering and basically throws the "its still blurry and inferior to native rendering" debacle, (that's been going around in PC community since 2019), right out of the window.

Einhander19722d ago

RaidenBlack

DLSS is as i said a different thing from FSR and checkerboard.

But you're talking about FSR 3 which probably is better than checkerboard, but FSR 3 has only started to get games this year, so checkerboard which was the first hardware upscaling solution was and is still one of the best upscaling solutions.

Give credit where credit is due, PlayStation was first and they got it right from the get go, and PSSR will almost certainly be better than it will be given credit for, heck digital foundry is already spreading misinformation about the Pro.

Rhythmattic3d ago

Tacoboto
Yes... Its amazing how many talekd about KZ2 deferred rendering, pointing out the explosions were lower res than the frame itself..
And of course, Then the idea of checkerboard rendering, not being native....
For sure, maybe this tech makes it minimal while pixel counting, but alas, seems performance and close enough , and not native now matters.....
I want to see it run native without DLSS.. why not?

RonsonPL3d ago

Almost deaf person:
- lightweight portable 5$, speakers of 0,5cm diameter are the final nail in coffin of Hi-Fi audio!

Some people in 2010:
- smartphones are the final nain in the console gaming's coffin!

This is just the same.
AI upscalling is complete dogshit in terms of motion quality. The fact that someone is not aware of it (look at the deaf guy example) doesn't mean the flaws are not there. They are. And all it takes to see them, is to use a display that handles motion well, so either gets true 500fps at 500Hz LCD TN or OLED (or faster tech) or uses low persistence mode (check blurbusters.com if you don't know what it means) also known as Black Frame Insertion or backlight strobing.

Also, image ruined by any type of TAA is just as "native image" as chineese 0,5$ screwdriver is "high quality, heavy duty, for professional use". It's nowhere near it. But if you're an ignorant "journalist", you will publish crap like this article, just to flow with the current.

There's no coffin to native res quality and there never will be. Eventually, we'll have enough performance in rasterization to drive 500fps, which will be a game changer for motion quality while also adding other benefit - lower latency.
And at 500fps, the amount of time required for upscalling makes it completely useless.
This crap is only usable for cinematic stuff, like cutscenes and such. Not for gaming. Beware of ignorants on the internet. The TAA is not "native" and the shitty look of the modern games when you disable any TAA, is not "native" either as it's ruined by the developer's design choice - you can cheat by rendering every 4th pixel when you plan to put a smeary TAA pass on it later on. When you disable it, you will see a ruined image, horrible pixellation and other visual "glitches" but it is NOT what native would've looked like if you'd like to honestly compare the two.

Stay informed.

RaidenBlack2d ago

Main point of the article is how far DLSS has come with v3.7 since 2018. If this is what we're getting already, then imagine what we'll get within next ~3 years. Yes parity will obviously be there compared to the then native rendering tech but it'll slowly narrow down to the point it'll be indistinguishable.
Something similar is like the genAI Sora ... AI generative videos were turd back when they were introduced (the infamous Will Smith eating video) ... but now look at Sora, generating videos that just looks like real life.

2d ago
Yui_Suzumiya2d ago

How much VRAM is standard today? My laptop has a 1080p QLED display but only an Intel Iris Xe with 128MB of VRAM. I currently do all my gaming on it but certain titles do suffer because of it. I plan on getting a Steam Deck OLED soon to play the newer and more demanding titles.

purple1012d ago

Maybe better to get a budget gaming laptop and link a dualsense to it

= Portable console with far better graphics than a steam deck! + bigger screen and able to use it for work / etc

170°

Why I'm worried about the Nvidia RTX 50 series

Aleksha writes: "Nvidia has established itself as a dominant force in the world of AI, but I can't shake the worry of what this means for the RTX 50 series."

Create Report !X

Add Report

Reports

+ Updates (1)- Updates (1)

Updates

Changed from Pending to Approved
Community6d ago
Tal1696d ago

Echo sentiment here - I think the way GPUs are going, gaming could be secondary to deep learning. Wonder if the 40 series was the last true generation of GPUs?

Number1TailzFan6d ago

No.. Jensen believes GPUs should stay expensive. Those wanting a top end GPU will have to splash out for it, or play at just 1080p and 60fps or something if you can only afford a low end option.

On the other hand if you don't care about RT or AI performance then there's always AMD that are doing ok at the mid range.

Christopher6d ago

***or play at just 1080p and 60fps or something***

My over 2-year-old laptop GPU still runs fine. I think this is more a reason why GPUs are going to other things in priority, because the market reach for new users is shrinking as more PC gamers focus less on replacing older and still working parts that run RT/AI fine enough as it is. Not to say there aren't people who still do it, but I think the market is shrinking for having the latest and greatest like it has been the past two decades. Problem is we aren't growing things at a rate as we were, we're reaching the the flattening of that exponential curve in regards to advancement. We need another major technological advancement to restart that curve.

D0nkeyBoi6d ago

The irremoval ad makes it impossible to read article

Tzuno6d ago (Edited 6d ago )

I hope Intel takes some lead and do a big dent to nvidia sales

Jingsing6d ago

You also need to consider that NVIDIA are heavily invested in cloud gaming. So they are likely going to make moves to push you into yet another life subscription service.

Kayser815d ago

NVIDIA will never change their price point until AMD or intel makes a GPU that is comparable and cheaper than them .
it happend before in the days of gtx280 which they changed the price from 650$ to 450$ in a matter of 2 weeks because of rx4870 which is being sold at 380$.

Show all comments (8)
270°

AMD FSR 3.1 Announced at GDC 2024, FSR 3 Available and Upcoming in 40 Games

Last September, we unleashed AMD FidelityFX™ Super Resolution 3 (FSR 3)1 on the gaming world, delivering massive FPS improvements in supported games.

Read Full Story >>
community.amd.com
Create Report !X

Add Report

Reports

✔ Fixed
Bad Editing
When copying text from the source article make sure you read over the text and edit it if necessary. A good example is when a site writes "We" or "I" , as this should then be replaced with the site's or person's name, or you can quote the source.
Nyxus24d ago WhoDisagree(0)Agree(1)
+ Updates (1)- Updates (1)

Updates

Changed from Pending to Approved
Community23d ago
Eonjay24d ago (Edited 24d ago )

So to put 2 and 2 together... FSR 3.1 is releasing later this year and the launch game to support it is Rachet and Clank: Rift Apart. In Sony's DevNet documentation it shows Rachet and Clank: Rift Apart as the example for PSSR. PS5 Pro also launches later this year... but there is something else coming too: AMD RDNA 4 Cards (The very same technology thats in the Pro). So, PSSR is either FSR 3.1 or its a direct collaboration with AMD for that builds on FSR 3.1. Somehow they are related. I think PSSR is FSR 3.1 with the bonus of AI... now lets see if RDNA 4 cards also include an AI block.

More details:
FSR 3.1 fixes Frame Generation
If you have a 30 series RTX card you can now use DLSS3 with FSR Frame Generation (No 40 Series required!)
Its Available on all Cards (we assume it will come to console)
Fixes Temporal stability

MrDead23d ago

I've been using a mod that allows dlss frame gen on my 3080 it works on all rtx series. It'll be good not to rely on mods for the future.

darksky23d ago

The mods avaiable are actually using FSR3 frame gen but with DLSS or FSR2 upscaling.

Babadook723d ago (Edited 23d ago )

I think that the leaks about the 5 Pro would debunk the notion that the two (FSR 3.1 and PSSR) are the same technology. PSSR is a Sony technology.

MrDead24d ago (Edited 24d ago )

I wonder how much they fixed the ghosting in dark areas as Nvidia are leaving them in the dust with image quality. Still good that they are improving in big leaps, I'll have to see when the RTX5000 series is released who I go with... at the moment the RTX5000's are sounding like monsters.

just_looken23d ago

Did you see the dell leaks were they are trying to cool cards using over 1k watts of power.

We are going to need 220 lines for next gen pcs lol

MrDead23d ago

That's crazy! Sounds like heating my house won't be a problem next winter.

porkChop23d ago

As much as I hate supporting Nvidia, AMD just doesn't even try to compete. Their whole business model is to beat Nvidia purely on price. But I'd rather pay for better performance and better features. AMD also doesn't even try to innovate. They just follow Nvidia's lead and make their own version of whatever Nvidia is doing. But they're always 1 or 2 generations behind when it comes to those software/driver innovations, so Nvidia is always miles ahead in quality and performance.

MrDead23d ago

I do a lot of work on photoshop so an Intel Nvidia set up has been the got to because of performance edge, more expensive but far more stable too. Intel also have the edge over AMD processors with better load distribution on the cores, less spikes and jitters. When you're working large format you don't want lag or spikes when you're editing or drawing.

I do think AMD has improved massively though and whist I don't think they threaten Nvidia on the tech side they do make very well priced cards and processors for the power. I'm probably going with a 5080 or 5090 but AMD will get a little side look from me, which is a first in a long time... but like you said they are a generation or two behind at the moment.

Goosejuice22d ago

While I can't argue for amd gpu, they aren't bad but they aren't great either. The cpu for amd have great. I would argue the 7800x3d as one of the best cpu for gaming right now. Idk about editing so I take ur word for that but gaming amd cpu is a great option these days.

porkChop22d ago

@Goosejuice

I have a 7800X3D. It certainly is great for gaming. Though for video editing, rendering, etc, I think Intel have the advantage from what I remember. I just mean from a GPU standpoint I can't support them.