<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>VR World &#187; gt200</title>
	<atom:link href="http://www.vrworld.com/tag/gt200/feed/" rel="self" type="application/rss+xml" />
	<link>http://www.vrworld.com</link>
	<description></description>
	<lastBuildDate>Fri, 10 Apr 2015 07:54:22 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>http://wordpress.org/?v=4.1.1</generator>
	<item>
		<title>Nvidia&#8217;s discloses its DP performance limitations</title>
		<link>http://www.vrworld.com/2009/02/11/nvidias-discloses-its-dp-performance-limitations/</link>
		<comments>http://www.vrworld.com/2009/02/11/nvidias-discloses-its-dp-performance-limitations/#comments</comments>
		<pubDate>Wed, 11 Feb 2009 13:38:20 +0000</pubDate>
		<dc:creator><![CDATA[Theo Valich]]></dc:creator>
				<category><![CDATA[3D]]></category>
		<category><![CDATA[Business]]></category>
		<category><![CDATA[CPU]]></category>
		<category><![CDATA[Graphics]]></category>
		<category><![CDATA[Hardware]]></category>
		<category><![CDATA[double precision]]></category>
		<category><![CDATA[double preicision performance]]></category>
		<category><![CDATA[dp support]]></category>
		<category><![CDATA[gflops]]></category>
		<category><![CDATA[GPGPU]]></category>
		<category><![CDATA[GPU Computing]]></category>
		<category><![CDATA[gt200]]></category>
		<category><![CDATA[Nvidia]]></category>
		<category><![CDATA[nvidia vs ati 2009]]></category>
		<category><![CDATA[TFLOPS]]></category>

		<guid isPermaLink="false">http://theovalich.wordpress.com/?p=1058</guid>
		<description><![CDATA[<p>When Nvidia launched GT200 chip, the company claimed around 1TFLOPS of Single-Precision computing power, and roughly 150 GFLOPS of Dual-Precision performance. This discrepancy was mostly ...</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2009/02/11/nvidias-discloses-its-dp-performance-limitations/">Nvidia&#8217;s discloses its DP performance limitations</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p>When Nvidia launched GT200 chip, the company claimed around 1TFLOPS of Single-Precision computing power, and roughly 150 GFLOPS of Dual-Precision performance.<br />
This discrepancy was mostly due to the fact that Nvidia went with dedicated hardware for the DP support. Every eight-shader cluster had one dedicated dual-precision unit, costing millions of additional transistors and resulted in doubtful performance.</p>
<p>Fast forward to January 2009, and we have SP performance at 933 GFLOPS, while achievable DP performance dipped down to 78 GFLOPS. This figure is roughly half of what Nvidia boasted about at the time of launch, and sheer evidence that both manufacturers like to overstate the performance of actual parts. What makes things interesting is the fact that Tesla GPGPU boards aren&#8217;t even most powerful parts in the Nvidia line-up. That &#8220;honor&#8221; goes to newly introduced GTX285 and 295. In professional line-up, Quadro FX 5800 has more &#8220;oomph&#8221;, thanks to higher shader clock. but even FX5800 will remain below 100 GFLOPS in dual-precision operations&#8230; making this GPU &#8220;just&#8221; 2.5x faster than quad-core Xeon processor.</p>
<p>Then again, if you activate parallel execution, CPU will drop to sub-10 GFLOPS values, while the GPU will remain at 78 GFLOPS for DP and 933 GFLOPS in single precision. At the same time, ATI&#8217;s architectural concept of &#8220;emulating&#8221; the DP units by pairing more processing units in one cluster resulted in actual peak performance of 900 GFLOPS for the 4870 part (claimed performance: 1.2 TFLOPS) and 250 GFLOPS for the Dual-Precision formats. This is an impressive difference, showcasing ATI&#8217;s lead from the architectural standpoint. Extractable performance is a bit different, since some ISVs managed to extract that performance, such as ElcomSoft password cracker, while some hit different walls and could not get better performance.</p>
<p>The real dilemma now is to wait and see what kind of computing performance lies with upcoming 40nm GPUs.</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2009/02/11/nvidias-discloses-its-dp-performance-limitations/">Nvidia&#8217;s discloses its DP performance limitations</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://www.vrworld.com/2009/02/11/nvidias-discloses-its-dp-performance-limitations/feed/</wfw:commentRss>
		<slash:comments>1</slash:comments>
		</item>
		<item>
		<title>100th Story- ANALYSIS: Why will GDDR5 rule the world?</title>
		<link>http://www.vrworld.com/2008/11/22/100th-story-gddr5-analysis-or-why-gddr5-will-rule-the-world/</link>
		<comments>http://www.vrworld.com/2008/11/22/100th-story-gddr5-analysis-or-why-gddr5-will-rule-the-world/#comments</comments>
		<pubDate>Sat, 22 Nov 2008 21:00:46 +0000</pubDate>
		<dc:creator><![CDATA[Theo Valich]]></dc:creator>
				<category><![CDATA[3D]]></category>
		<category><![CDATA[AMD]]></category>
		<category><![CDATA[Business]]></category>
		<category><![CDATA[Graphics]]></category>
		<category><![CDATA[Hardware]]></category>
		<category><![CDATA[Intel]]></category>
		<category><![CDATA[Memory & Storage Space]]></category>
		<category><![CDATA[256 Bit]]></category>
		<category><![CDATA[40nm]]></category>
		<category><![CDATA[512-bit]]></category>
		<category><![CDATA[55nm]]></category>
		<category><![CDATA[ATI]]></category>
		<category><![CDATA[differential]]></category>
		<category><![CDATA[Differential GDDR5]]></category>
		<category><![CDATA[FirePro]]></category>
		<category><![CDATA[gddr3]]></category>
		<category><![CDATA[gddr4]]></category>
		<category><![CDATA[GDDR5]]></category>
		<category><![CDATA[GeForce]]></category>
		<category><![CDATA[gt200]]></category>
		<category><![CDATA[gt206]]></category>
		<category><![CDATA[gt212]]></category>
		<category><![CDATA[joe macri]]></category>
		<category><![CDATA[larrabee]]></category>
		<category><![CDATA[Nvidia]]></category>
		<category><![CDATA[PlayStation 4]]></category>
		<category><![CDATA[Quadro]]></category>
		<category><![CDATA[Radeon]]></category>
		<category><![CDATA[S.E. GDDR5]]></category>
		<category><![CDATA[single-ended]]></category>
		<category><![CDATA[xbox 720]]></category>

		<guid isPermaLink="false">http://theovalich.wordpress.com/?p=534</guid>
		<description><![CDATA[<p>As &#8220;Theo&#8217;s Bright Side of IT&#8221; turns a century (100 stories) after 5 weeks of existence, it would be right to write an article about ...</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2008/11/22/100th-story-gddr5-analysis-or-why-gddr5-will-rule-the-world/">100th Story- ANALYSIS: Why will GDDR5 rule the world?</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p>As &#8220;Theo&#8217;s Bright Side of IT&#8221; turns a century (100 stories) after 5 weeks of existence, it would be right to write an article about technology that is set to become an everyday word during the next couple of years: GDDR5.<br />
This memory standard will become a pervasive memory during next four years in much more fields than &#8220;just&#8221; graphics. Just like GDDR3 ended up in all three consoles, network switches, cellphones and even cars and planes, GDDR5 brings a lot of new features that are bound to win more customers from different markets.</p>
<p><strong>Background</strong><br />
The reason for development of radical ideas inside GDDR5 lies in the fact that ATI was looking at future GPU architectures, and concluded that the DRAM industry has to take a radical step in design and offer interface more flexible than any other memory standard. Then, ATI experienced huge issues with R600 and its huge monolithic die. After a lot of internal struggle, engineering teams came to agreement that a change of course is necessary for generations to come: R700/RV770, R800/RV870, R900, R1K… all of these engineering designs are reshaped and refocused. Current and future goal is to design a compact and affordable transistor design that would not play a game of Russian roulette with yields coming from <a title="MAD AMD or GloblaFoundries" href="http://www.tomshardware.com/news/amd-corporate-culture,5206.html" target="_blank">MAD AMD</a>, TSMC&#8217;s and UMC&#8217;s foundries.<br />
Development of this JEDEC certified standard happened under the lead of Joe Macri, Director of engineering at AMD and chairman of JEDEC&#8217;s Future DRAM Task Group JC42.3. Joe and his small ex-ATI/AMD GPGP team are mostly known for the development of the GDDR3 and GDDR4 memory standards, with former being probably the best thing ever to come out of the former ATI. ATI worked in solitude for a whole year before it sent initial specification to JEDEC in 2005. Then, Hynix, Qimonda and Samsung joined the effort to bring the new memory standard to life. When AMD acquired ATI in 2006, new management didn&#8217;t touch GDDR5 development and let the team to work in peace. Reason was simple: R&amp;D team warned the management that GDDR5 development is much more difficult from work done on GDDR3 and GDDR4.<br />
GDDR5 was seen as a path towards next-generation clients, that being consoles, desktop computing, networking equipment, HPC arena, handhelds&#8230; all of these roads start with one memory standard. At the time, engineers at ATI saw the path of success that GDDR3 took, and decided to create a spec that would outlive and outshine GDDR3.<br />
In May 2008, AMD finally announced the launch of GDDR5 memory standard. Soon after, the company revealed its Radeon 4800 series and cards equipped with GDDR5 memory. Given the performance of Radeon 4870 512MB, 4870 1GB and 4870X2 2GB, it is obvious that the future of graphics (and not just!) belongs to GDDR5 memory.<br />
At its very core, it is important to know that the main difference between LP-DDR (handhelds, PDAs), DDR (one fits all) and GDDR (Graphics) is the fact that capacity is not crucial, but performance is. Low-Power DDR and standard DDR are geared to enabling as much capacity as possible, while GDDR is usually referred to as the &#8220;Ferrari of the bunch&#8221;.</p>

<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_01_gpu-ram-roadmap1.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_01_gpu-ram-roadmap1-750x420.jpg" class="attachment-vw_medium" alt="Roadmap shows that DDR3 will replace DDR2 in low-end market, and GDDR5 will take over GDDR3" /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_03_gddr345-diferences.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_03_gddr345-diferences-750x420.jpg" class="attachment-vw_medium" alt="Description of differences between the standards..." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_04_gddr345-diferences.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_04_gddr345-diferences-750x420.jpg" class="attachment-vw_medium" alt="... and continuing with differences." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_05_ram-roadmap.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_05_ram-roadmap-750x420.jpg" class="attachment-vw_medium" alt="In 2010, we should see Differential GDDR5, and then the available bandwidth on GPUs will double over the night." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_06_gddr5_key-features.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_06_gddr5_key-features-750x420.jpg" class="attachment-vw_medium" alt="According to Qimonda, these are key features of GDDR5 standard." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_07_gddr5-lowmedhighfr.jpg' rel="lightbox[gallery-0]"><img width="750" height="372" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_07_gddr5-lowmedhighfr-750x372.jpg" class="attachment-vw_medium" alt="GDDR5 is divided into three different memory types, and clocks and voltage change according to specified role." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_08_gddr5-pcb-tracing_.jpg' rel="lightbox[gallery-0]"><img width="489" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_08_gddr5-pcb-tracing_-489x420.jpg" class="attachment-vw_medium" alt="Note the absence of &quot;combs&quot; on PCB using GDDR5 memory. This will enable cheaper PCBs and higher performance at the same time." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_09_gddr5-overclocking.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_09_gddr5-overclocking-750x420.jpg" class="attachment-vw_medium" alt="GDDR5 is also the first memory standard designed with overclocking in mind." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_10_gddr5-clockingandd.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_10_gddr5-clockingandd-750x420.jpg" class="attachment-vw_medium" alt="The way how clock works...four data transfers over a single clock." /></a>
<a href='http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_11_gddr5_x16-mode.jpg' rel="lightbox[gallery-0]"><img width="750" height="420" src="http://cdn.vrworld.com/wp-content/uploads/2008/11/gddr5_11_gddr5_x16-mode-750x420.jpg" class="attachment-vw_medium" alt="Clamshell mode - very important feature, will enable doubling the amount of memory in near future." /></a>

<p><strong><br />
DDR, DDR2, DDR3, GDDR3, GDDR4, GDDR5 … got it?</strong></p>
<p>If you can’t find your way through the jungle of different memory standards, don&#8217;t worry, you&#8217;re not alone. There is a lot of confusion in the world of DRAM memory, and sadly, there is no simple explanation. Most important thing to remember is that GDDR and DDR are not the same memory, and do not operate on same data sets.<br />
As you can see, GDDR memory transfers 32-bit data, while conventional DRAM transfers 64-bit data chunks. Previous generations of graphics memory (GDDR2, GDDR3) were remotely based on the DDR2-SDRAM memory standard, while GDDR5 is heading into a new direction.</p>
<p>In fact, GDDR5 standard actually splits into two different ways how DRAM operates: Single-Ended and Differential. This is a revolutionary step for GDDR memory, since it was widely expected that Single-Ended memory is the only way to go. In a way, you can say that ATI developed GDDR5 and GDDR &#8220;5.5&#8221; or &#8220;6&#8221; at the same time. Single-ended support is compatible with existing memory standards such as DDR1/2/3/GDDR3/4 and represents evolutional path for DRAM. First products to market will use single-ended chips, but as soon as Hynix, Qimonda and Samsung start manufacturing differential modules (2009-10), a new era will begin.<br />
Differential clock signaling is a method similar to interconnect buses such as HyperTransport, PCI Express, or Intel&#8217;s Quick Path Interface from Core i7. Differential introduces Reference clock, a clock that memory cell follows. Instead of using Ground wire as a passive driver, Differential mode enables precise communication and exactly this feature is the reason why available bandwidth is set for a dramatic change during lifetime of GDDR5.<br />
The sheer bandwidth gain from one GDDR generation to another is impressive. GDDR3 peaked at 2.4 Gbps, GDDR4 concluded at 3.2 Gbps. GDDR5 chips split into two: Single-Ended will offer between 3.4 and 6.4 Gbps of bandwidth, while differential chips will yield between 5.6 and 12.8 Gbps.</p>
<p>Besides Differential mode, GDDR5 also introduces an Error Correction Protocol based on a progressive algorithm that actually enables more aggressive overclocking. Major changes in internal chip design also include Quarter-Data Rate clock, continuous WRITE clock, CDR based READ (no reading clock/strobe information), DRAM Interface training, Internal and External VREF and x16 mode.<strong></strong></p>
<p><strong>Power Saving</strong></p>
<p>One of very important things with GDDR5 is power reduction. If you take GDDR3 and GDDR5 modules, clocked at 1.0 GHz each, GDDR3 will have to operate at 2.0V, while GDDR5 needs only 1.5V. This results in 30% reduction of power consumption, while raising available per-pin bandwidth by almost 100%.</p>
<p>GDDR5 is designed to operate at low, medium and high frequencies. Low frequency (0.2-1.5 Gbps) calls for low-voltage (0.8-1.0V), while medium (1.0-3.0 Gbps) and high (2.5-5.0 Gbps) frequencies call for higher voltage, in a range between 1.4-1.6V.<br />
High frequency is the only one that utilizes CDR (Command Data Rate) circuitry, while medium and low frequencies call for conventional mode (RDQS with Preamble).<br />
Seeing the drop in power below levels of FB-DIMM DDR2-800 only makes us wonder what would happen if CPU manufacturers would implement Differential GDDR5 as system memory. Would we really need Gigabytes of system memory if we would have system memory with higher bandwidth than L2 and L3 cache? Intel is looking in similar direction, considers <a href="http://www.tomshardware.com/news/Intel-DRAM-CPU,5697.html" target="_blank">replacing SRAM cache with DRAM technology</a>.</p>
<p>Sadly, the changes that would be required in memory controller are such that only place where GDDR5 will see the light of day as system memory are closed designs, such as consoles, set top boxes and so on. There is hope that some future AMD&#8217;s Fusion designs might implement GDDR support, but it is too early to tell.</p>
<p><strong>How to lower the cost of manufacturing?</strong></p>
<p><strong><br />
</strong>During design stages of GDDR5 memory, one of main concerns was how to simplify tracing on the PCB (Printed Circuit Board). On current GDDR3 and GDDR4 graphics boards, synchronization issues are solved by using traces of the same length from every pin on DRAM chip to the GPU. This causes quite a messy design, with traces going everywhere.</p>
<p>IF you&#8217;re PCB designer, there is one thing that you don&#8217;t want: complex routing of traces. This eventually leads to more PCB layers, higher cost and most importantly &#8211; more ways for *something* to go wrong. Every trace has increased isolation from electromagnetic interferences (EMI), while Asymmetrical Interface compensates for differences in length. In order to keep the signal integrity, several optimizations were made.<br />
As you could see on picture above, GDDR5 PCB route is much cleaner than GDDR3, and you can see that if you compare Radeon 4850 to Radeon 4870, for instance. This was paid by additional resistors around memory chips, but second generation of GDDR5 graphics cards should feature cleaner design.</p>
<p><strong>Memory designed for overclocking?</strong></p>
<p><strong><br />
</strong>With power saving and performance-related tweaks, it is obvious that this memory is designed for overclocking. This was confirmed to us just by looking at slides from AMD and Qimonda.</p>
<p>The GDDR5 specification delivers a combination of three technologies: Adaptive Training and CDR, Error Detection and an on-die thermal sensor. Adaptive Training is combined with the Error Detection algorithm and enables the memory controller of the GPU to keep thermals on a tight leash. If you want to overclock the memory, it will go up until the error correction algorithm hits a thermal wall.</p>
<p>Error Detection works with both read and write instructions, offering real time repeat and resend operations. Thanks to asynchronous clocks, memory controller can control flow of data and resend bits of information that fail to arrive in time (or arrive corrupted). Error Detection algorithm will try to avoid a crash until the number of errors passes 1 Error/sec.<br />
In order to maintain the signal stability, additional resistors were placed inside and outside the memory chip (take a look at the back of 4870 and compare it to 4850). AMD also addressed the issue spotted on GDDR4. Overclocking of GDDR4 memory was limited because DRAM timing loop would run out of power. GDDR5 changed the way how clock is generated and kept, so memory chip should never starve for power. No timing loop issue = no memory freeze. According to our sources, GDDR5 memory clocking in the end depends on the manufacturing process (used by the chip manufacturer) and the amount of voltage provided to the chip.<br />
But main difference in clocking of GDDR3 and GDDR5 is the fact that PVT (Power, Voltage, and Temperature) is no longer the unbreakable barrier. Now, it is GPU&#8217;s memory controller that will keep (or fail to keep) the flow of data.</p>
<p><strong>Coalition between the GPU and the RAM</strong></p>
<p>Unlike previous memory standards, in order to extract the best possible performance memory controller has to support ALL of the GDDR5 features. This especially goes to Asymmetrical interface, since WRITE and READ clocks are programmed by the GPU. Advanced Clock Training calibrates GPU-RAM signals &#8211; without this feature, you cannot count on high clocks or overclocking capabilities. With four bits of data being sent per clock (instead of two), memory controller is exposed to a lot of stress, and has to be able to do error checking on the fly. Any misses on GPU side will lead to cycle losses &#8211; leading to instability.<br />
Good example is memory controller tucked inside the Radeon 4800. This 256-bit controller supports DDR2, DDR3, GDDR3, GDDR4 and GDDR5 memory standards. The memory controller is tuned up to the point where bandwidth and clock limitation are on the side of the SGRAM chips: If the fastest GDDR5 memory chips were available today, you could build a 4800-series card with them. This also opens up revenue opportunities for Hynix, Samsung and Qimonda. All three manufacturers could earn a small fortune by selling gold sampled memory chips to premium graphics card manufacturers.<br />
When it comes to Nvidia, answer to the question why the company went with GDDR3 for GTX 200 series of cards is not a simple one: according to our sources, GT200 chip supports GDDR3 and GDDR4, while engineers ran out of time to adjust memory controller to asymmetrical interface (advanced interface training), key feature for stable operation. But, if Nvidia sticks with 512-bit memory controller for NV70 generation (GT300?), we should see Nvidia GPUs featuring bandwidth in excess of 300 GB/s, more than twice that is available today. There is also a question what will Nvidia do with its two refreshes, 55nm GT206 and 40nm GT212 chips.<br />
Intel is not giving out any details on Larrabee&#8217;s architecture, but we know for sure that the 1024-bit internal/512-bit external memory controller will support GDDR5 and its advanced features. Given the late 2009 release, support for differential mode should be a given. When it comes to christening, Larrabee with GDDR5 memory will debut during this winter, with <a href="http://www.tomshardware.com/news/intel-larrabee-graphics,5847.html" target="_blank">first graphics cards delivered to Dreamworks</a>.</p>
<p><strong>Capacity – just how big can we go?<br />
</strong>Now that you&#8217;ve seen all of the performance elements, time to write about capacity. While Joe told us that GDDR should be considered as &#8220;The Ferrari of DDR world&#8221;, GDDR5 introduces x16 mode. This mode has nothing to do with PCI Express x16 (to kill any potential confusion).</p>
<p>As you can see on the slide above, Clamshell mode is introduced to enable two memory chips sitting on a single x32 node. If we take ATI Radeon 4800 series, GPU features eight x32 I/O controllers. In theory, this should top at 16 memory chips per GPU, or 1GB of onboard memory using conventional 512Mbit chips. With x16 mode, card designer can put up to 32 chips (good luck with finding available space), or 2GB memory with 512Mbit (64MB) chips. With 1Gbit (128MB) chips, this number grows to 4GB. Qimonda is expected to ship 2Gbit (256MB) chips during 2009, enabling 8GB of on-board memory.</p>
<p>This number is increasingly important for GPGPU market, which wants as much on-board memory as possible. Bear in mind that Tesla 10-Series features 4GB of GDDR3 memory, and some contacts we&#8217;ve talked with &#8211; claim they would fill even more.</p>
<p>Eight GB of video memory may sound too much for consumer space, but if world is to usher into the era of <a href="http://www.tomshardware.com/news/Larrabee-Ray-Tracing,5769.html" target="_blank">Ray-tracing</a>, we have to get enough space for gigabytes of data. Jules Urbach from JulesWorld explained that he is working with datasets bigger than 300 GB, and has to resort using AMD&#8217;s CAL (Compression Algorithm) to fit all the data inside 1GB per GPU (Jules uses R700 boards).</p>
<p><strong>Conclusion</strong></p>
<p><strong></strong>GDDR5 ramped up during 2008 and we expect the technology becoming a standard for GPU add-in-boards in 2009. ATI will migrate to GDDR5, so will Nvidia. With Intel joining the pack with Larrabee, volumes should be ready to drive the cost of GDDR5 into budget for next generation of game consoles, starting in the 2010-11 timeframe.<br />
This is by far the most developed and well-thought memory standard that lacks childhood sicknesses like DDR2 and DDR3. GDDR5 is coming to market as a complete product, and offers solid future roadmap, with Differential GDDR5 even surpassing XDR2 DRAM in quest for highest possible per-pin bandwidth.<br />
By that time, Differential GDDR5 should be cheaper than GDDR3 is today.</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2008/11/22/100th-story-gddr5-analysis-or-why-gddr5-will-rule-the-world/">100th Story- ANALYSIS: Why will GDDR5 rule the world?</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://www.vrworld.com/2008/11/22/100th-story-gddr5-analysis-or-why-gddr5-will-rule-the-world/feed/</wfw:commentRss>
		<slash:comments>8</slash:comments>
		</item>
		<item>
		<title>TSMC introduces 40nm volume production, advances in front of Intel</title>
		<link>http://www.vrworld.com/2008/11/18/tsmc-introduces-40nm-volume-production-advances-in-front-of-intel/</link>
		<comments>http://www.vrworld.com/2008/11/18/tsmc-introduces-40nm-volume-production-advances-in-front-of-intel/#comments</comments>
		<pubDate>Tue, 18 Nov 2008 09:56:21 +0000</pubDate>
		<dc:creator><![CDATA[Theo Valich]]></dc:creator>
				<category><![CDATA[3D]]></category>
		<category><![CDATA[AMD]]></category>
		<category><![CDATA[Business]]></category>
		<category><![CDATA[Graphics]]></category>
		<category><![CDATA[Hardware]]></category>
		<category><![CDATA[Intel]]></category>
		<category><![CDATA[2009]]></category>
		<category><![CDATA[2010]]></category>
		<category><![CDATA[2011]]></category>
		<category><![CDATA[22nm]]></category>
		<category><![CDATA[28nm]]></category>
		<category><![CDATA[32nm]]></category>
		<category><![CDATA[40nm]]></category>
		<category><![CDATA[chip]]></category>
		<category><![CDATA[CPU]]></category>
		<category><![CDATA[GPU]]></category>
		<category><![CDATA[gt200]]></category>
		<category><![CDATA[gt206]]></category>
		<category><![CDATA[gt212]]></category>
		<category><![CDATA[larabee]]></category>
		<category><![CDATA[manufacturing]]></category>
		<category><![CDATA[TSMC]]></category>

		<guid isPermaLink="false">http://theovalich.wordpress.com/?p=470</guid>
		<description><![CDATA[<p>A while ago, I spoke with my sources at TSMC, who were quite decisive to make it to the front on the field of chip ...</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2008/11/18/tsmc-introduces-40nm-volume-production-advances-in-front-of-intel/">TSMC introduces 40nm volume production, advances in front of Intel</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p>A while ago, I spoke with my sources at TSMC, who were quite decisive to make it to the front on the field of chip manufacturing. Heads of this Taiwanese giant decided to invest more than 10 billion USD in order to become world&#8217;s most advanced manufacturer, and their roadmap is more aggressive than anyone in the industry.</p>
<p>The results of that investment are slowly coming to life, and as of today, TSMC has more advanced manufacturing process than any other competitor in the manufacturing business. Intel will argue its (very important, though) Hafnium or High-K material, but ever since I became a journalist, Intel touted its manufacturing capabilities and ability to go small &#8220;sooner than anyone else&#8221;. Well, that is about to change.<br />
For instance, Intel will introduce 32nm process in late 2009, and mass production in 2010. Due to separation between AMD and &#8220;MAD AMD&#8221; (The Foundry Company) will introduce 32nm (bulkPG, not for CPUs) only at the end of 2009, with 2010 being the year of mass production. If all things go well, that is.<br />
During that same time, TSMC will introduce 32nm (Q4&#8217;2009), 28nm (Q2&#8217;2010) and 22nm will debut in first half of 2011. This is very, very aggressive roadmap that will give Nvidia and ATI leverage in development of graphics parts.</p>
<p>This also does not sound good for Intel&#8217;s own Larrabee, which will rely on Intel&#8217;s own manufacturing capabilities. While this was viewed as a huge strength in the previous years, TSMC may actually give AMD and Nvidia more than a fighting chance &#8211; a winning cost/die ratio.</p>
<p>As a case of demonstration, my source gave me a comparison while using Nvidia&#8217;s GT200 chip. This estimated comparison gave me shivers, because in 28nm (available in a bit more than a year), die size for 1.4 billion transistors would drop to incredible 160mm2. Of course, don&#8217;t expect that ATI or Nvidia will stand still. They will keep making big GPUs and put more and more core logic inside.</p>
<p><strong>GT200 die through different TSMC manufacturing processes (&#8220;wild&#8221; estimate):</strong><br />
65nm: 576 mm2 (GT200)<br />
55nm: 470 mm2 (GT206)<br />
40nm: 320 mm2 (GT212)<br />
32nm: 220 mm2 (die-shrink estimate)<br />
28nm: 150 mm2 (die-shrink estimate)</p>
<p>Given this table, we can see that if Nvidia would want to keep the 500mm2 die size, it could manufacture a chip with 500 processors in 40nm, 700 processors in 32nm or massive 1200 shader processors using 28nm process. But don&#8217;t expect that either ATI or Nvidia will go linear with their GPUs.<br />
What I personally expect is 512-bit bus, GDDR5 memory controller for both companies (regardless of what ATI is saying now), and increasing the capabilities of shaders. Currently, ATI is supporting FP64 DP format through their 80 shader lines (e.g. in RV780, you have 80 shader pipelines with 10 units in each &#8211; you can either output one FP64 Dual Precision or ten FP32 Single Precision number formats). Nvidia features one FP64 DualPrecision unit on eight of their regular shader cores.<br />
With 32nm available in 2009 and 28nm available a year later, it is easy to predict that we will see a tremendous increase in processing power not through the sheer number of shaders, but rather increasing existing shader capabilities.<br />
My $0.02 is that we will see 4-10TFLOPS parts coming in next 24 months, essentially increasing the computational power by anywhere between four and ten times. All thanks to massive effort put in by TSMC.<br />
For now, Nvidia can announce the mass production of Tegra mobile SoC chips and its notebook lineup, while ATI can launch their notebook line-up. 40nm High-Performance arrives in Q1&#8217;2009, and you can expect GT212 and RV870 coming &#8220;your way in May&#8221;.</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2008/11/18/tsmc-introduces-40nm-volume-production-advances-in-front-of-intel/">TSMC introduces 40nm volume production, advances in front of Intel</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://www.vrworld.com/2008/11/18/tsmc-introduces-40nm-volume-production-advances-in-front-of-intel/feed/</wfw:commentRss>
		<slash:comments>0</slash:comments>
		</item>
		<item>
		<title>UPDATE: 55nm GT206 GPU powers both GTX290 and Quadro FX 5800</title>
		<link>http://www.vrworld.com/2008/11/11/55nm-gt206-gpu-powers-both-gtx290-and-quadro-fx-5800/</link>
		<comments>http://www.vrworld.com/2008/11/11/55nm-gt206-gpu-powers-both-gtx290-and-quadro-fx-5800/#comments</comments>
		<pubDate>Tue, 11 Nov 2008 02:42:16 +0000</pubDate>
		<dc:creator><![CDATA[Theo Valich]]></dc:creator>
				<category><![CDATA[3D]]></category>
		<category><![CDATA[AMD]]></category>
		<category><![CDATA[Business]]></category>
		<category><![CDATA[Graphics]]></category>
		<category><![CDATA[Hardware]]></category>
		<category><![CDATA[240 shaders]]></category>
		<category><![CDATA[4800]]></category>
		<category><![CDATA[4890]]></category>
		<category><![CDATA[55nm]]></category>
		<category><![CDATA[a3]]></category>
		<category><![CDATA[ATI]]></category>
		<category><![CDATA[christmas]]></category>
		<category><![CDATA[g200]]></category>
		<category><![CDATA[g200-200]]></category>
		<category><![CDATA[g200-202]]></category>
		<category><![CDATA[g200-300]]></category>
		<category><![CDATA[g200-302]]></category>
		<category><![CDATA[GPU]]></category>
		<category><![CDATA[gt200]]></category>
		<category><![CDATA[gt206]]></category>
		<category><![CDATA[gt212]]></category>
		<category><![CDATA[gtx 270]]></category>
		<category><![CDATA[gtx 290]]></category>
		<category><![CDATA[GTX270]]></category>
		<category><![CDATA[gtx290]]></category>
		<category><![CDATA[higher clock]]></category>
		<category><![CDATA[mod]]></category>
		<category><![CDATA[Radeon]]></category>
		<category><![CDATA[rv780]]></category>
		<category><![CDATA[rv790]]></category>
		<category><![CDATA[shopping]]></category>
		<category><![CDATA[Thanksgiving]]></category>

		<guid isPermaLink="false">http://theovalich.wordpress.com/?p=365</guid>
		<description><![CDATA[<p>The honor of being the first product powered by 55nm G200-302 chip (a.k.a. GT206/212) went to Quadro FX 4800/5800, products that launched with a lot ...</p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2008/11/11/55nm-gt206-gpu-powers-both-gtx290-and-quadro-fx-5800/">UPDATE: 55nm GT206 GPU powers both GTX290 and Quadro FX 5800</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p>The honor of being the first product powered by 55nm G200-302 chip (a.k.a. GT206/212) went to Quadro FX 4800/5800, products that launched with <a href="http://theovalich.wordpress.com/2008/11/10/nvidia-officially-unveils-civil-cx-and-fx5800-monster/" target="_blank">a lot of fanfare earlier today</a>.</p>
<p>Besides Quadro FX 4800 and 5800, the new 55nm GPU will also power GeForce GTX 270 and 290. Essentially, we&#8217;re talking about the same parts. Quadro FX 4800 is nothing more but GTX270 with double the amount of video memory, while Quadro FX 5800 is equal to GTX290, but with four times the video memory. ATI is not sleeping, as the company is preparing an RV790 part , beefed-up version of already existing RV770 chip.</p>
<p>G200-302 Rev A2 begun manufacturing back in September, and the first parts are now finding their way to mass manufacturing. The chip features a die size of 470mm2, 107mm2 less than the original G200 chip. This just goes to show the vast difference between 65nm and 55nm &#8211; if Nvidia had the balls to go with 55nm chip back in May, the prices of GTX260/280 parts could have been way cheaper and offer much more flexibility, but we can&#8217;t cry over spilt milk. 55nm part is here now, and it will consume much less power than is the case with the 65nm one.</p>
<p>The 55nm GPU consumes roughly 50% less power than it was the case with 65nm one, and this difference is more than massive. When I did quick power checks, the GTX280 at 650/1500/2200 would eat around 266W, while the default clocked GTX280 (600/1300/2200) was specc&#8217;ed at 238W.</p>
<p>Well, the 55nm GPU will eat around 170W at 650/1500/2200, meaning that GTX290 just got 100W of power to play with. If you&#8217;re into overclocking, you can now start dreaming about clocking those 240 shaders to 1.7-1.8 GHz range (perhaps even 2.0 if water-cooling setup is powerful enough), and achieve massive performance gains, all happening while you&#8217;re consuming <strong>less </strong>power than a stock clocked GTX280.</p>
<p>As far as the naming convention goes, Nvidia calls their chips NVxx (we&#8217;re at NV60 right now) or Gxx/Gxxx internally, and partners get the GT200-XXX name. But at the end of the day, the number that matters is the one on the chip.<br />
GTX 260 and 280 both came with G200-200 and G200-300 chips, while GTX270 and 290 will feature G206-202 and G206-302 chips. Essentially, there is no difference between the two, sans the hardwired part that decides how many shaders a certain chip has. If you&#8217;re brave enough, you&#8217;ll pop the massive HIS and play around with resistors. Who knows, perhaps you can enable 240 shaders on GTX260/270… or maybe not.<br />
In any case, we can&#8217;t wait for these new babies to show up. FX4800, FX5800, GTX270 and 290 are all coming to market very, very soon.<br />
My personal take is that Nvidia will try to steal the limelight of official Core i7 launch on 11/17 and ship the GTX270/290 to reviewers, trying to tell them that they&#8217;re still on top. All hopes with ATI lie in the form of upcoming 4890. But still, Nvidia does not offer a compelling $199 experience and this is where ATI will take them to the cleaners.</p>
<p>Of course&#8230; unless you see a GTX260-216 at a completely new price point, and GTX270 costing just $50 more, dropping to $199 for Christmas. Crazy scenario, but competition brings the best for us, consumers.</p>
<p><strong>UPDATE: </strong>Picture that accompanied the story did not feature GT206 chip, thus I removed it. The rest of the info is pretty valid <img src="http://cdn.vrworld.com/wp-includes/images/smilies/icon_smile.gif" alt=":-)" class="wp-smiley" /></p>
<p>The post <a rel="nofollow" href="http://www.vrworld.com/2008/11/11/55nm-gt206-gpu-powers-both-gtx290-and-quadro-fx-5800/">UPDATE: 55nm GT206 GPU powers both GTX290 and Quadro FX 5800</a> appeared first on <a rel="nofollow" href="http://www.vrworld.com">VR World</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://www.vrworld.com/2008/11/11/55nm-gt206-gpu-powers-both-gtx290-and-quadro-fx-5800/feed/</wfw:commentRss>
		<slash:comments>15</slash:comments>
		</item>
	</channel>
</rss>

<!-- Performance optimized by W3 Total Cache. Learn more: http://www.w3-edge.com/wordpress-plugins/

Content Delivery Network via Amazon Web Services: CloudFront: cdn.vrworld.com

 Served from: www.vrworld.com @ 2015-04-10 17:35:59 by W3 Total Cache -->