LinuxQuestions.org
Help answer threads with 0 replies.
Home Forums Tutorials Articles Register
Go Back   LinuxQuestions.org > Forums > Linux Forums > Linux - General
User Name
Password
Linux - General This Linux forum is for general Linux questions and discussion.
If it is Linux Related and doesn't seem to fit in any other forum then this is the place.

Notices


Reply
  Search this Thread
Old 08-23-2008, 05:16 AM   #46
resetreset
Senior Member
 
Registered: Mar 2008
Location: Cyberspace
Distribution: Dynebolic, Ubuntu 10.10
Posts: 1,340

Rep: Reputation: 62

Quote:
Originally Posted by maxreason View Post
I have no time for that - BUT - I'd do it anyway IF I have the complete detailed specifications for the latest nvidia cards. I've been wanting to write directly to the cards "forever", so I can completely drop OpenGL, D3D, etc - and program "pedal to the metal". Fact is, that would let me do the other thing I wanted to do - merge the vertex, pixel, geometry, CUDA/OpenGL into one unified language, so object/geometry/texture/physics/general computations can be performed in a perfectly unified, integrated way.

Yeah, I know, I should get a job at nvidia. But I hate cities, and silicon valley is a long commute from phobos.

Yeah! THAT's the spirit!
The folks already DOING the drivers may have reverse engineered some or quite a bit of it by now, so you won't be starting from scratch.
What do you say?
 
Old 08-23-2008, 05:21 AM   #47
resetreset
Senior Member
 
Registered: Mar 2008
Location: Cyberspace
Distribution: Dynebolic, Ubuntu 10.10
Posts: 1,340

Rep: Reputation: 62
Just to add, I've been wanting this "pedal to the metal" info for 3D cards almost as soon as they started coming out, but no one has been able to help me - they all went either the Open GL or D3D way. But that's not the demoscener spirit, I want the *basic* info - such as, the card has memory for holding textures, right? Can that memory be used for anything then? I've already heard of people programming spreadsheets and whatever, to run on GPUs.....?
What does the phrase "Hardware transformation and lighting" exactly MEAN?

Last edited by resetreset; 08-23-2008 at 05:40 AM.
 
Old 08-23-2008, 08:20 PM   #48
maxreason
Member
 
Registered: Dec 2007
Location: phobos, mars
Distribution: 64-bit linux mint v20
Posts: 259

Original Poster
Rep: Reputation: 17
Well, the phrase "hardware transformation and lighting" means very little since the advent of programmable "shaders" in OpenGL and D3D, because you [can] do all transformation and lighting in your own [shader] programs. However - and this IS important in a few contexts - the 3D graphics cards still contain some very real ("hard") hardware aspects to perform certain simple, low-level highly repetitious processes specific to graphics rendering. Still, those aspects can be bypassed to perform non-rendering processes, which is increasingly what GPUs do.

When you program with "shaders" (or GPUPU languages like OpenCL/CUDA) the memory on the graphics card can serve any purpose you [shader] programs wish.

Frankly, it is STUPID that OpenCL and CUDA are being developed - those capabilities should be part of the already-existing [shader] languages. But then again, nobody ever asks me until it is too late [cuz I don't work at nvidia].

One of the advantages of my engine [for demo scene purposes and others] is that it provides a very fast/compact/efficient way to issue [graphics and other] commands !!! without knowing or caring what is the interface to the hardware !!!. Therefore, my engine may perform its graphics [and GPUPU] functions on OpenGL [and OpenCL/CUDA] initially, but if and when I/someone figures out how to drive the video cards directly, the engine would simply become faster (but continue to accept the same commands, and function identially). This kind of interface never breaks, and never needs an overhaul - because the commands deal with fundamental/enduring concepts, not implementation-specific architecture quirks.

Sorry, but the work I am doing is far too important to abandon [even temporarily] unless I had complete graphics card specs to work with. I have reverse engineered a couple things in my day, and though it can be satisfying in the end, it is brutally difficult and time-consuming. I don't have that kind of time, which my personal messages should explain - hopefully!

Last edited by maxreason; 08-23-2008 at 08:26 PM.
 
  


Reply



Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off



Similar Threads
Thread Thread Starter Forum Replies Last Post
Eclipse : installing CDT vmelkon Ubuntu 1 08-02-2007 10:01 PM
configuring cdt on eclipse mohtasham1983 Programming 1 01-25-2007 10:22 PM
eclipse & cdt vshenoy Programming 1 04-27-2006 09:36 AM
Eclipse CDT for Amd64??? ssobeht Programming 4 02-07-2005 06:33 AM
autoindent in eclipse/CDT spuzzzzzzz Linux - Software 0 07-23-2004 06:27 PM

LinuxQuestions.org > Forums > Linux Forums > Linux - General

All times are GMT -5. The time now is 06:47 AM.

Main Menu
Advertisement
My LQ
Write for LQ
LinuxQuestions.org is looking for people interested in writing Editorials, Articles, Reviews, and more. If you'd like to contribute content, let us know.
Main Menu
Syndicate
RSS1  Latest Threads
RSS1  LQ News
Twitter: @linuxquestions
Open Source Consulting | Domain Registration