Days of 3d accelerators numbered?

Hellbinder

New member
Odd topic I know,

With the ever increasing focus on complete programability throughout the entire 3D rendering process.... Its looking more and more like the old software days. Comments from various developers have gone from "wow we can do all this cool stuff now" to "we are being held back by limitations in progamability".

The most excitement about any new product recently seems to be focused on the P10, which is nearly like a second complete CPU. I am wondering if what we really need is actually a new powerful 3D programing language, and the focus placed on more powerful CPU's. I dont see why companies like ATi, Nvidia and others simply develope a specialized CPU with the instruction set tailored to 3D. GPU's are already starting to go that route. Surely it would be faster to have a 2Ghz C-GPU than the setup we have today. Of course there are many other factors involved to make such a design really fly Like memmory access, bandwidth, cache, Its own specialized bus etc etc...

Anyone else have any imput on the idea?
 
I'm not sure I agree with that. I think there will always be some form of independant graphics processor (even if that means it is integrated into the core of the CPU).

In 5-10 years I could mebbie see what you're taling about. When we're talking cpus in the 100s of GHz. At which point processing power will be such that there will be power to spare, and room in them on small processes to integrate functions to help accelerate some of the really complicated math functions required for graphics which really aren't needed in day-to-day computing.
 
That would be nice but I agree with Ichneumon. That thing will need alot more power than what's available today to get everything done. I don't see why it wouldn't be possible in the future though:)
 
Re: Days of 3d accelerators numbered?

Hellbinder said:
I dont see why companies like ATi, Nvidia and others simply develope a specialized CPU with the instruction set tailored to 3D. GPU's are already starting to go that route.

Well, this is a general thing in computing, in automation and - generally speaking - in decision making, that is de-centralizing.

You already have several processors in your pc, apart from the cpu, which is 'general purpose', there is the gpu, the 'chipset' - which makes a lot 'alone', without assistance from the cpu - the ide controller (well, it's become part of the chipset... years ago the i8272, VLSI, was just controlling a floppy - or two) even the keyboard has a controller - i.e. dedicated processor of itself. The purpose is to reduce execution time by using 'specialists' for each function, and to simplify comunication by reducing message length between them.

A good team of specialists with good communication.
 
I'm an idiot so I have a dumb question.

Why is it that GPU's have such a lower clock speed compared to todays CPU's when the die size is even smaller and transistor count is higher in most cases?.

Would higher clocked GPU's be an advantage or is there a bottle neck somewhere that would make it unnecessary?
 
BeardedClem said:
I'm an idiot so I have a dumb question.

Why is it that GPU's have such a lower clock speed compared to todays CPU's when the die size is even smaller and transistor count is higher in most cases?.

Would higher clocked GPU's be an advantage or is there a bottle neck somewhere that would make it unnecessary?

i think the high transistor count of the GPU's is what holds the speed down for reasons of heat and others that i'm not quite sure...maybe someone else can answer that..

later
 
wandernACE64 said:


i think the high transistor count of the GPU's is what holds the speed down for reasons of heat and others that i'm not quite sure...maybe someone else can answer that..

later

More than likely. When a company is producing a chip with a lifespan of 12-18 months they are probably not going to put much effort into efficiency. It seems to be more of a "Do what you need to do to hit this target" approach that they are using, and if they need 80 million transistors to do it, but can only reach 250-300 MHz then so be it. It would probably take a lot more time and effort to make a smarter chip with fewer transistors and a higher clock rate than to just bash something through that works.

Chris.
 
Re: Days of 3d accelerators numbered?

Hellbinder said:
Odd topic I know,

--->8---
I dont see why companies like ATi, Nvidia and others simply develope a specialized CPU with the instruction set tailored to 3D. GPU's are already starting to go that route. Surely it would be faster to have a 2Ghz C-GPU than the setup we have today.
---8<---
Anyone else have any imput on the idea?

Actually, that was what they used to do long, long ago. Way back in the day Number Nine (and a few others) made a really good series of cards based on a Texas Instruments DSP. The upshot was that if they wanted to add or fix Windows accelerator functions, it was just a driver update rather than new firmware or a new chip. Problem was that fixed function accelerators were a lot cheaper to design and build, so they took over. That was when S3 took over the market for the 2D accelerators with their fixed function 801/805 series chips. Funny how things come around again if you wait long enough.

Chris.
 
The term DSP is a little ambiguous - it stands for Digital Signal Processor, and there are a lot of chips marketed under this acronzm. However, they are general purpose DSPs, that means processors with built in support for some functions most widely used in digital signal processing, especially for digital filters and Fourier transforms.

A GPU is - in a way - a DSP (it does digital signal processing), only it needs more specialized functions than - say - digital audio boxes, Dolby 5.1 etc - or the EMU10k1.

The point in using dedicated chips is that the GPU will not waste time in fetching and processing code (like 'get x', 'get y', 'calculate x2+y2', 'next x', etc.) to simulate those functions, but will accept comands like 'draw a circle', the rest of the 'get x' stuff being performed implicitely by the whole bunch of transistors.

Yes, it's cheaper to build those chips than to buy some 200$ TI DSP (check the DSP prices, you'll be surprised), and the current DSPs run also in the 100-300 MHz range...

The slowness of GPU's is, IMO, in the fact that they are not built from scratch, like CPUs, but are in fact some sort of ready made configurable chips (PGA, ASIC and the sort), which are slower because of the structure overhead implied by configurability. That is, you buy the chip and 'configure' it to be a GPU (or anything else), but the internal structures that allow you to do that remain within the chip and consume some bandwith, power etc.

As far as development is concerned, it's easyer to do DSP, not that easy with ASIC - and really not that easy with 'transistor by transistor' chip design. In the end, it's easiest to leave it all tu the CPU.

Ten years ago, Intel made 50 MHz chips - I think. The consummer graphics industry will come of age, eventually.
 
merry said:
Yes, it's cheaper to build those chips than to buy some 200$ TI DSP (check the DSP prices, you'll be surprised), and the current DSPs run also in the 100-300 MHz range...

...

As far as development is concerned, it's easyer to do DSP, not that easy with ASIC - and really not that easy with 'transistor by transistor' chip design. In the end, it's easiest to leave it all tu the CPU.

Well, it might be easier to build the actual DSP chip, but then you have a whole new department to develop the internal code for it. It's the engineering that's expensive, not the chip. The reason that DSPs are expensive is the toolkits they provide to help you actually do something with it. The upshot is that you can tell the DSP exactly what you want rather than having to munge something out with the grocery list of supported functions like an ASIC, which may or may not do exactly what you need.

As far as keeping everything in the CPU goes, who here got a chance to use a Commodore Amiga? They had a graphics processor, a sound processor, and the CPU as separate hardware, which is why they could do motion video or play games with great colour and digital audio when people on IBM PCs were playing Commander Keen in EGA with an AdLib card.

Chris.
 
Think about the sound cards. About 4 years ago they often came with advanced signal processor for voice recognition e.t.c. They often had a pretty advanced midi wavwtable synthesizer. Look at the sound cards of today. The wavetable synthesizer is often gone and the card only handles big number of streams instead. The wavetable synthesizer is moved from hardware to software (just listen Yamaha software synth, better than almost all hardware synths) and the software synth doesn't make that much impact on performance today.

One interesting aspect of the future CPU:s are that they will come with multiple cores on one chip. One question is if that will make the GPU:s more redudant since they almost are an extra CPU albeit a bit more specialized.

However, there are still very much to do in the visual domain. Scientists are saying that real-time raytracing will take over the world one day. Raytracing would make great benefits of hardware since it can be done in parallell to great extent. I am really wondering why the big GF companies don't put more effort in this area since raytracing gives much more for "free" (reflections, refractions, easy portals).

For those who are interested in real-time raytracing go to www.openrt.de and read some interesting papers.
 
Last edited:
Intergrating features onto chips is an on going process, it reduces overall costs, improves dependability and probably a number of other aspects. I remember XT motherboards that would make the mother boards of today look like unpopulated toys bit yet today motherboards are eons more advance and more capable. I don't think we are even remotely close in intergrating a R300 or NV30 onto a cpu or having a cpu capable enough to reproduce even the basic features we take for granted. 5 to 10 years will probably be way different but I think dedicated chips are here to stay. I would like to see the wire traces dissappear and light traces or fiber optics take over though. :)
 
Integration might work for low-cost systems, but for high-bandwidth/high-end applications it still makes sense for a processor to be where the work is. Why share the memory and I/O bandwidth of the CPU if you can have an external unit take care of it by itself with a completely separate memory bus? The only reason to do differently is to make it cheap (like those sound cards).

Chris.
 
cbsboyer said:


Well, it might be easier to build the actual DSP chip, but then you have a whole new department to develop the internal code for it. It's the engineering that's expensive, not the chip

I guess this is the point where my english starts failing me. This is precisely what I wanted to say. Intel has the lot of bucks to support chip design.

The point with expensive DSP chips is that, well, to do a good job you need an expensive chip, more expensive than a middle of the road graphics card - mine was 85$, at this money I couldn't buy a reasonably performant DSP (I tried to find a DSP56307 or 311 evaluation module, I should have to pay some 300$, board, sw and all - and it's not even suitable for graphics).

A chip that you develop yourself will be cheaper than that, design costs incuded.

I don't know much about designing w ASIC, I mentioned it because I've heard this it what they do for GPUs, ATI at least.

And it can't be as fast as a CPU, because of the internal complexity required for the engineering of the chip, but useless for the final activity of the chip.

Uf, I hope this makes any sense. I actually totaly agree with you:
Integration might work for low-cost systems, but for high-bandwidth/high-end applications it still makes sense for a processor to be where the work is. Why share the memory and I/O bandwidth of the CPU if you can have an external unit take care of it by itself with a completely separate memory bus?

merry
 
Two years later, and 3D acceleration is still alive and kicking.. with no sign of degrading.

It's just getting faster and faster and bigger and has bigger heatsinks/fans now.



My Voodoo 3 did not have a fan. It had just a heatsink, and that alone was probably not nessecary.
 
I think an interesting point here is that Tim Sweeney (of Unreal fame) seems to be thinkinf along similar lines - He seems to be convinced that at some stage in the future, the CPU and GPU will converge into a single, all-purpose module.

I can't say that I agree with him personally, but it's an interesting theory.
 
Hanners said:
I think an interesting point here is that Tim Sweeney (of Unreal fame) seems to be thinkinf along similar lines - He seems to be convinced that at some stage in the future, the CPU and GPU will converge into a single, all-purpose module.

I can't say that I agree with him personally, but it's an interesting theory.

That would be cool. I'm sick and tired of hearing "cpu limited" :rolleyes: Why is it that you can go out and buy a $500.00 video card and your 2.8 P4 is holding it back? :nuts:
 
Vengeance said:
That would be cool. I'm sick and tired of hearing "cpu limited" :rolleyes: Why is it that you can go out and buy a $500.00 video card and your 2.8 P4 is holding it back? :nuts:
I think is because the CPU is not as fast in doing things like AI or physics. The GPU ends up waiting on the CPU for things like that. It would be nice if either a) CPU manufacturers can speed that kind of processing up or b) GPU manufacturers use the GPU to accelerate AI/physics/etc. calculations. Either way we will be CPU limited for sometime. I do not know much about dual core processors and SMP, but it would be nice if during a gaming session, one core handled the necessary processing tasks and some game specific calculations, while the other core did game specific calculations exclusively.
 
I think once PCI Express has matured and become the standard, and games start being coded around its inherent benefits, then we may well see a lot of things like physics shifted onto the GPU.

We're already seeing things like geometry instancing taking some of the strain away from the CPU, and I imagine that trend will only continue.
 
Isn't there a shift towards parallel processing in the computing world such as dual-core processors or dual CPU systems.

I think this shift is due to the logical assumption that if two tasks are unreleated they can be done concurrently and save the time to do each seperately.

With this in mind I think the tasks and methods for completing them are unreleated with respect to the CPU and GPU. The tasks a CPU performs alot of are unreleated tasks that require well for lack of a better way of saying it step by step decision making. Tasks like physics and even more so AI are very suited for being placed on the CPU. The GPU handles batches or very similiar tasks that don't require alot of decision making (ignoring dynamic branching in shaders for a moment). As someone previously noted where a CPU would plot a circle point by point a GPU would simply draw a circle as it is optimized to do.

Since general tasks and graphic processing are for the most part unrelated tasks I don't think the call for parallelism would be defeated by a merger of the CPU and GPU. Making a single core that was efficiently at doing both tasks concureently seems a very difficult task. I feel it is likely however that a dual core processor may however replace the CPU and GPU down the road. That is to say a GPU core will be added to a traditional CPU core on a single chip. This would give the GPU core the much needed speed boost it needs without crippling the CPU core with tasks it is not optimized to perform and since the GPU is up to speed there would be no real need to labor the CPU core with geometry at all allowing for it to be solely dedicated to AI, physics, and system tasks. This of course means you need a faster bus and more bandwith but that already happening.

Who knows, multicore design may be even more explicit than that where a core will be dedicated to graphics, AI, physics, and general system tasks respectively. Sony certainly seems unafraid to add cores to the Cell chip. (maybe they should be...)

Well that's my 2 cents. I think multi-"specialized"-core processors will win out over a complicated single core attempt to do all tasks efficiently. I'm not as as smart as you guys so I apologize if I'm aggregiously off base.
 
Last edited:
Back
Top