Video Cards & GPUs News - Page 330

All the latest graphics cards and GPU news, with everything related to Intel Arc, NVIDIA GeForce, AMD Radeon & plenty more - Page 330.

Follow TweakTown on Google News

NVIDIA's new $40 video card is '10x better than integrated graphics'

Derek Strickland | Jan 27, 2016 2:33 PM CST

NVIDIA has just launched its new GeForce GT 710 video card, a cheap $40 GPU that the company claims can deliver "up to 10x the performance of integrated graphics", and gaming up to "80% faster" than traditional iGPUs.

While the GeForce GT 710 gets beaten by Intel HD 530 integrated graphics found in Skylake chips, the card is an excellent option for budget gamers who have older rigs and want a cheap DirectX 12 card. As far as specs go, NVIDIA's inexpensive card uses DDR3 memory on a 64-bit bus with 14.4GB/s bandwidth, has a base clock of 954 MHz with 192 CUDA cores, and a memory clock speed of 1.8 Gbps.

The GeForce 710 supports a host of features including G-Sync adaptive sync, multi-monitor support for up to 3 displays, 3D Vision, NVIDIA's PhysX tech, and is OpenGL 4.5 and DirectX 12-ready out of the box. The card has a maximum resolution of 2560x1600 via HDMI and 2048x1536 on VGA, and features 1x Dual-Link DVI-D port, 1x HDMI and 1x VGA. It requires a minimum 300W power supply to function.

Continue reading: NVIDIA's new $40 video card is '10x better than integrated graphics' (full post)

NVIDIA to launch HBM2-powered GeForce GTX Titan X successor in April

Anthony Garreffa | Jan 26, 2016 6:47 PM CST

If you have been reading our GPU-related content, you should know that we are set for the biggest year in GPU history this year, from both sides: AMD and NVIDIA.

Well, at NVIDIA's GPU Technology Conference in April, we should see NVIDIA unveil the biggest GPU they've ever made - the successor to the GeForce GTX Titan X. The next-gen card could be called the GTX Titan X2, which would pull some of the wind out of AMD's sails with the dual-GPU Radeon R9 Fury X2, and we should see it featuring HBM2 - scaling up to 16-32GB with 1TB/sec of memory bandwidth. Insanity.

Back in September, we exclusively said that NVIDIA would release both a HBM2 and GDDR5X range of cards - something that will kick off with the HBM2-based Titan X successor. Towards June, we should see NVIDIA unveil a new GP104-based GeForce GTX 980 successor, based on GDDR5X - which offers 448GB/sec of memory bandwidth.

Continue reading: NVIDIA to launch HBM2-powered GeForce GTX Titan X successor in April (full post)

AMD's dual Fiji-powered GPU used with HTC Vive at VRLA event

Anthony Garreffa | Jan 24, 2016 4:46 PM CST

The last time we physically saw the dual-GPU version of the Fury X was at the launch event itself in Sydney, Australia - where we had our hands-on that beautiful PCB. But, the Radeon R9 Fury X2 has shown up again, this time at VRLA.

AMD's dual Fiji-powered GPU used with HTC Vive at VRLA event

The VRLA expo was an event for all things virtual reality, held in LA last week. During the event, some of the HTC Vive demos were powered using the Radeon R9 Fury X2. Thanks to Facebook, we noticed Antal Tungler, PR Manager for AMD and all-round cool guy, posted on his Twitter account. He said: "Prototype Tiki from @FalconNW powering #htcvive with dual Fiji @AMDRadeon at the #vrla".

Someone asked Tungler "When you say "Dual Fiji" do you mean 2x Fiji cards, or 2x Fiji GPUs on 1 card? ;)", to which he replied with "One card". So we know that it wasn't 2 x R9 Fury X cards in the machine, but a single, dual-GPU beast. But with Polaris around the corner, I have to ask the question: where does the R9 Fury X2 fit in? It would only have 4GB of HBM1 per GPU, which really isn't enough VRAM considering it will be $1000+. VR headsets are pushing 90FPS, and a high-resolution to boot. I guess we'll see in the coming months, maybe AMD will launch the Fury X2 in between now and the release of Polaris in June/July.

Continue reading: AMD's dual Fiji-powered GPU used with HTC Vive at VRLA event (full post)

JEDEC makes GDDR5X official, features 2x the bandwidth of GDDR5

Jeff Williams | Jan 22, 2016 8:03 PM CST

The final specification for GDDR5X, the successor to GDDR5, has been decided, and though it doesn't allow for quite as much bandwidth as HBM or HBM2, though it's a technology that's a lot easier to implement than the latter, with fewer modifications needed to the GPU design to use.

GDDR5X allows for up to 14Gbps of total bandwidth and because it's based so heavily on its predecessor, it's pin compatible though highly internally revised in order to facilitate actual advancements in memory speed and bandwidth without making something entirely new. How JEDEC and Micron have done this is by increasing the prefetch by double, mandating the use of Phase Locked Loops and Delay Locked Loops as well as being able to transmit data at a rate that's quadruple the actual clock speed. In other words, it's fast. For comparison, GDDR5X running at the top-end 14Gpbs could potentially provide 448GBps of full bandwidth, which isn't too far off of the memory bandwidth of the R9 Fury X.

Micron, one of the leading manufacturers working on GDDR5X, estimates around a 10% power consumption decrease at the same VRAM size. VRAM sizes of 4Gb up to 16Gb can be used with the new specification. The reason for coming out with this new specification is to further address every segment of the market, especially those where HBM2 might not be economical, despite AMD's efforts to implement HBM in all segments of their GPUs. Now all GPUs can enjoy a healthy bandwidth increase for very little, if any, cost increase.

Continue reading: JEDEC makes GDDR5X official, features 2x the bandwidth of GDDR5 (full post)

NVIDIA should launch its next-gen Pascal GPUs with HBM2 in 2H 2016

Anthony Garreffa | Jan 20, 2016 7:17 PM CST

Now that HBM2 is beginning to flow into the market, thanks to Samsung making 4GB HBM2-based DRAM, NVIDIA is getting confident with Pascal - with the latest rumor stating that the company will unveil its next-gen GPUs in the first half of this year, with availability to follow in 2H 2016.

We know this will happen, where my sources tease that both AMD and NVIDIA will have next-gen GPUs prepared for June/July, but I've got a feeling NVIDIA will introduce a next-gen enthusiast GPU at their GPU Technology Conference in early April. NVIDIA is reportedly already playing around with the 16nm-based Pascal GPUs internally, but we should expect GDDR5X- and HBM2-powered offerings, with a GeForce GTX Titan X successor to be unveiled at GTC 2016. We might see the new Titan X with 16GB of HBM2, and possibly a professional-grade Tesla/Quadro GPU with 32GB of HBM2 teased, too.

As for the GP100, I don't think NVIDIA will unveil the GeForce GTX 980 Ti successor just yet as the GTX 980 Ti is still one of the best video cards you can buy. We should see a Titan X successor unveiled - powered by HBM2, followed by a successor to the GTX 980 - powered by GDDR5X. The HBM2-powered offerings will be able to pack 32GB of HBM2, and offer up to 1TB/sec of memory bandwidth, up from the 334GB/sec on the GeForce GTX 980 Ti and its GDDR5.

Continue reading: NVIDIA should launch its next-gen Pascal GPUs with HBM2 in 2H 2016 (full post)

Open-source GPU from Binghamton University could shake up the industry

Jeff Williams | Jan 20, 2016 2:00 PM CST

GPU's are fantastic tools for completing scientific computational work. They're effective at calculating math that's highly parallel, doing it far faster than any CPU could do alone. And now the GPU is going open-source with Binghamton University's new Nyami architecture that researchers have developed.

Timothy Miller and his colleagues have finally been able to test their open-source GPU design, called Nyami. It's essentially a GPGPU focused design that's borrowed a lot from Intel's Larrabee (Xeon Phi) while still being, at it's heart, a GPU.

This is the first open-source modifiable and synthesizable GPU made by anyone. The architecture has a measure of modularity so that any aspiring researcher or scientist can modify it to their hearts content, provided they have the expertise to do so, though. But really this is revolutionary because now software and hardware can reach a nexus and be developed with the help of the open-hardware community, which is a well supported community. The problem has always been getting an architecture started, which is a highly technical engineering problem. But now the first part is solved, and we might see some great scientific progress that could even spill over to consumer GPU's one day. Just keep in mind this isn't something you'll be playing Assassin's Creed Redundancy on.

Continue reading: Open-source GPU from Binghamton University could shake up the industry (full post)

Rumor: GeForce GTX 980MX and 970MX mobile GPUs on the way?

Jeff Williams | Jan 19, 2016 6:08 PM CST

There's a rumor floating around the Internet that seems to suggest that NVIDIA is very close to releasing their top-end mobile GPU, the GeForce GTX 980MX as well as the little brother, the GTX 970MX. Keep in mind that these are strictly mobile parts, and not related to the full-fat GTX 980 that's being stuffed into laptops.

There doesn't appear to be any actual source to confirm the imminent release, though they seem to be very adamant that NVIDIA is intent on releasing these high-end mobile parts soon. And these chips will be plenty fast and actually provide power efficiency that'll be necessary in thinner laptops.

The GTX 980MX is rumored to have 1664 CUDA cores, 104 texture units, 64 raster devices and a clock speed of up to 1048MHz on a 256-bit memory bus. This is slightly more CUDA cores than the slightly smaller part, the 980M. Oddly the TDP is only 25W less than the full-blow 980 laptop variant at 125W. That's still a lot of power, and you definitely wouldn't be gaming with a notebook powered by this monster without being tethered to the wall.

Continue reading: Rumor: GeForce GTX 980MX and 970MX mobile GPUs on the way? (full post)

More AMD Polaris shipments potentially spotted, needs some salt

Jeff Williams | Jan 19, 2016 12:21 PM CST

AMD's new Polaris chip might have just been caught on the Zauba import/export table, if we can truly believe what other sources have decoded while reading these manifests.

The price of per unit of this particualr "printed circuit board assembly for personal computer(video/ video card)" is such that it lines up with other AMD shipments in the past. It also indicate that the bigger Polaris die which we were able to see at CES, might be well into production. That's good news for us enthusiasts.

But then again, decoding the serial numbers and the entire manifest is very difficult, and even though some might claim that they know that these particular shipments are indeed for AMD and are a chip with a particular architecture, we don't actually know. But it is exciting that at least production appears to be marching on as you read this. Just don't forget the tablespoon of salt. And yeah, it's okay to be excited too. I know I am.

Continue reading: More AMD Polaris shipments potentially spotted, needs some salt (full post)

GPU sales: PowerColor R9 390 for $268, XFX R9 390 for $274

Derek Strickland | Jan 19, 2016 11:26 AM CST

Whether you need to upgrade your outdated Radeon card or want to ensure your PC is ready for the incoming VR boom, we've found a duo of GPU deals that will fit the bill nicely. Today we have two different flavors of AMD's Radeon R9 390's discounted over at Newegg, both of which offer some impressive performances with 4K resolution support.

GPU sales: PowerColor R9 390 for $268, XFX R9 390 for $274

These sales are complimented by mail-in-rebates, which is par for the course for NewEgg sales. First up we have a PowerColor Radeon R9 390 for just $268 after a $20 mail-in-rebate (regular price $268), and the eggmen are tossing in a free $10 gift card to boot--but I don't think you can use the card for this purchase, only future purchases. The PowerColor R9 390 sports 1x HDMI, 1x Display Port and 2x DVI slots, and requires a 6-Pin / 8-Pin connectors with a 750W PSU.

Next up is an XFX Radeon R9 390 dropped down to $274 after a $30 MIR (original price $304). This card is similar to the PowerColor model, featuring a single HDMI and DisplayPort slot accompanied with two DVI ports. The power requirements are the same, with a minimum 750W PSU and the 6 and 8-pin PCIe power connectors.

Continue reading: GPU sales: PowerColor R9 390 for $268, XFX R9 390 for $274 (full post)

AMD's next-gen interconnect will offer 100GB/sec for multi-GPU setups

Anthony Garreffa | Jan 18, 2016 12:16 AM CST

During NVIDIA's GPU Technology Conference last year, NVIDIA unveiled its new NVLink interconnect that would find its BFF in their upcoming Pascal architecture.

At the time, we wrote that NVLink had 5x the bandwidth of PCIe 3.0, with NVLink opening up the possibilities for 8-way GPU setups - compared to the limit of 4-way SLI that we have now. Well, AMD is now talking about its upcoming next-gen coherent fabric, which will offer speeds of an insane 100GB/sec throughout multi-GPU setups. AMD has said that its new APUs will also be supported, with compute machines set to benefit greatly, too.

The big question for AMD is still in the air - but RTG boss Raja Koduri has said that he can't reveal if memory coherency and sharing between the GPUs and APUs will happen with the new interconnect. It would make sense to see it happen, but I'm sure AMD is rolling towards a big reveal in the near future with its Polaris architecture.

Continue reading: AMD's next-gen interconnect will offer 100GB/sec for multi-GPU setups (full post)