TweakTown NewsRefine News by Category:
AMD has launched a new website to promote GPUOpen, its initiative intended to assist graphics developers with open source tools, effects, and more.
As the site explains, the point of the tools is to foster more efficient game development. To that end, they are readily downloadable from GitHub and can be shared at will among fellow developers.
A couple of the technologies featured are HIP, a C++ interface that lets you create portable applications to run on any GPU, and AMD's LiquidVR, a Direct3D 11 interface that can empower applications with various GPU features, including multi-GPU support. Effects on display include AMD's TressFX for fancy hair and fur, to name one of many.
The last time we physically saw the dual-GPU version of the Fury X was at the launch event itself in Sydney, Australia - where we had our hands-on that beautiful PCB. But, the Radeon R9 Fury X2 has shown up again, this time at VRLA.
The VRLA expo was an event for all things virtual reality, held in LA last week. During the event, some of the HTC Vive demos were powered using the Radeon R9 Fury X2. Thanks to Facebook, we noticed Antal Tungler, PR Manager for AMD and all-round cool guy, posted on his Twitter account. He said: "Prototype Tiki from @FalconNW powering #htcvive with dual Fiji @AMDRadeon at the #vrla".
Someone asked Tungler "When you say "Dual Fiji" do you mean 2x Fiji cards, or 2x Fiji GPUs on 1 card? ;)", to which he replied with "One card". So we know that it wasn't 2 x R9 Fury X cards in the machine, but a single, dual-GPU beast. But with Polaris around the corner, I have to ask the question: where does the R9 Fury X2 fit in? It would only have 4GB of HBM1 per GPU, which really isn't enough VRAM considering it will be $1000+. VR headsets are pushing 90FPS, and a high-resolution to boot. I guess we'll see in the coming months, maybe AMD will launch the Fury X2 in between now and the release of Polaris in June/July.
But, that's with DX11 games. When DX12 rolls around, we should begin to see the 2 x 4GB lots of HBM1 turn into 8GB of HBM. This will be a huge deal for AMD, as they'll be able to market it as "8GB for DX12", right in time for DX12 games and VR.
The final specification for GDDR5X, the successor to GDDR5, has been decided, and though it doesn't allow for quite as much bandwidth as HBM or HBM2, though it's a technology that's a lot easier to implement than the latter, with fewer modifications needed to the GPU design to use.
GDDR5X allows for up to 14Gbps of total bandwidth and because it's based so heavily on its predecessor, it's pin compatible though highly internally revised in order to facilitate actual advancements in memory speed and bandwidth without making something entirely new. How JEDEC and Micron have done this is by increasing the prefetch by double, mandating the use of Phase Locked Loops and Delay Locked Loops as well as being able to transmit data at a rate that's quadruple the actual clock speed. In other words, it's fast. For comparison, GDDR5X running at the top-end 14Gpbs could potentially provide 448GBps of full bandwidth, which isn't too far off of the memory bandwidth of the R9 Fury X.
Micron, one of the leading manufacturers working on GDDR5X, estimates around a 10% power consumption decrease at the same VRAM size. VRAM sizes of 4Gb up to 16Gb can be used with the new specification. The reason for coming out with this new specification is to further address every segment of the market, especially those where HBM2 might not be economical, despite AMD's efforts to implement HBM in all segments of their GPUs. Now all GPUs can enjoy a healthy bandwidth increase for very little, if any, cost increase.
AMD might be looking to lower the price of the vanilla R9 Fury, or so a rumor seems to suggest. And this would be shortly after they lowered the price of the Nano in response to customer demand and the market.
KitGuru says that one of their sources that just happens to be part of the retail chain seems to be privy to the knowledge that there is indeed a price-cut that's going to happen on the R9 Fury. They don't know what that cut actually is, but just that it's going to happen in the coming weeks. They even suspect the price-cut might include the Fury X as well.
This is great news because it's a great card that can provide a good experience at 1440P and even at 4K if the visual quality is turned down some. A price reduction would make more competitive against the card, the 980, that it actually competes against. If this holds true, it'll be a good move by AMD.
Now that HBM2 is beginning to flow into the market, thanks to Samsung making 4GB HBM2-based DRAM, NVIDIA is getting confident with Pascal - with the latest rumor stating that the company will unveil its next-gen GPUs in the first half of this year, with availability to follow in 2H 2016.
We know this will happen, where my sources tease that both AMD and NVIDIA will have next-gen GPUs prepared for June/July, but I've got a feeling NVIDIA will introduce a next-gen enthusiast GPU at their GPU Technology Conference in early April. NVIDIA is reportedly already playing around with the 16nm-based Pascal GPUs internally, but we should expect GDDR5X- and HBM2-powered offerings, with a GeForce GTX Titan X successor to be unveiled at GTC 2016. We might see the new Titan X with 16GB of HBM2, and possibly a professional-grade Tesla/Quadro GPU with 32GB of HBM2 teased, too.
As for the GP100, I don't think NVIDIA will unveil the GeForce GTX 980 Ti successor just yet as the GTX 980 Ti is still one of the best video cards you can buy. We should see a Titan X successor unveiled - powered by HBM2, followed by a successor to the GTX 980 - powered by GDDR5X. The HBM2-powered offerings will be able to pack 32GB of HBM2, and offer up to 1TB/sec of memory bandwidth, up from the 334GB/sec on the GeForce GTX 980 Ti and its GDDR5.
GPU's are fantastic tools for completing scientific computational work. They're effective at calculating math that's highly parallel, doing it far faster than any CPU could do alone. And now the GPU is going open-source with Binghamton University's new Nyami architecture that researchers have developed.
Timothy Miller and his colleagues have finally been able to test their open-source GPU design, called Nyami. It's essentially a GPGPU focused design that's borrowed a lot from Intel's Larrabee (Xeon Phi) while still being, at it's heart, a GPU.
This is the first open-source modifiable and synthesizable GPU made by anyone. The architecture has a measure of modularity so that any aspiring researcher or scientist can modify it to their hearts content, provided they have the expertise to do so, though. But really this is revolutionary because now software and hardware can reach a nexus and be developed with the help of the open-hardware community, which is a well supported community. The problem has always been getting an architecture started, which is a highly technical engineering problem. But now the first part is solved, and we might see some great scientific progress that could even spill over to consumer GPU's one day. Just keep in mind this isn't something you'll be playing Assassin's Creed Redundancy on.
There's a rumor floating around the Internet that seems to suggest that NVIDIA is very close to releasing their top-end mobile GPU, the GeForce GTX 980MX as well as the little brother, the GTX 970MX. Keep in mind that these are strictly mobile parts, and not related to the full-fat GTX 980 that's being stuffed into laptops.
There's doesn't appear to be any actual source to confirm the imminent release, though they seem to be very adamant that NVIDIA is intent on releasing these high-end mobile parts soon. And these chips will be plenty fast and actually provide power efficiency that'll be necessary in thinner laptops.
The GTX 980MX is rumored to have 1664 CUDA cores, 104 texture units, 64 raster devices and a clock speed of up to 1048MHz on a 256-bit memory bus. This is slightly more CUDA cores than the slightly smaller part, the 980M. Oddly the TDP is only 25W less than the full-blow 980 laptop variant at 125W. That's still a lot of power, and you definitely wouldn't be gaming with a notebook powered by this monster without being tethered to the wall.
AMD's new Polaris chip might have just been caught on the Zauba import/export table, if we can truly believe what other sources have decoded while reading these manifests.
The price of per unit of this particualr "printed circuit board assembly for personal computer(video/ video card)" is such that it lines up with other AMD shipments in the past. It also indicate that the bigger Polaris die which we were able to see at CES, might be well into production. That's good news for us enthusiasts.
But then again, decoding the serial numbers and the entire manifest is very difficult, and even though some might claim that they know that these particular shipments are indeed for AMD and are a chip with a particular architecture, we don't actually know. But it is exciting that at least production appears to be marching on as you read this. Just don't forget the tablespoon of salt. And yeah, it's okay to be excited too. I know I am.
Whether you need to upgrade your outdated Radeon card or want to ensure your PC is ready for the incoming VR boom, we've found a duo of GPU deals that will fit the bill nicely. Today we have two different flavors of AMD's Radeon R9 390's discounted over at Newegg, both of which offer some impressive performances with 4K resolution support.
These sales are complimented by mail-in-rebates, which is par for the course for NewEgg sales. First up we have a PowerColor Radeon R9 390 for just $268 after a $20 mail-in-rebate (regular price $268), and the eggmen are tossing in a free $10 gift card to boot--but I don't think you can use the card for this purchase, only future purchases. The PowerColor R9 390 sports 1x HDMI, 1x Display Port and 2x DVI slots, and requires a 6-Pin / 8-Pin connectors with a 750W PSU.
Next up is an XFX Radeon R9 390 dropped down to $274 after a $30 MIR (original price $304). This card is similar to the PowerColor model, featuring a single HDMI and DisplayPort slot accompanied with two DVI ports. The power requirements are the same, with a minimum 750W PSU and the 6 and 8-pin PCIe power connectors.
During NVIDIA's GPU Technology Conference last year, NVIDIA unveiled its new NVLink interconnect that would find its BFF in their upcoming Pascal architecture.
At the time, we wrote that NVLink had 5x the bandwidth of PCIe 3.0, with NVLink opening up the possibilities for 8-way GPU setups - compared to the limit of 4-way SLI that we have now. Well, AMD is now talking about its upcoming next-gen coherent fabric, which will offer speeds of an insane 100GB/sec throughout multi-GPU setups. AMD has said that its new APUs will also be supported, with compute machines set to benefit greatly, too.
The big question for AMD is still in the air - but RTG boss Raja Koduri has said that he can't reveal if memory coherency and sharing between the GPUs and APUs will happen with the new interconnect. It would make sense to see it happen, but I'm sure AMD is rolling towards a big reveal in the near future with its Polaris architecture.