Author Topic: [MOVED] Hacking NVidia Cards into their Professional Counterparts  (Read 1650513 times)

0 Members and 6 Guests are viewing this topic.

Offline FireDragon

  • Regular Contributor
  • *
  • Posts: 62
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #700 on: November 09, 2013, 05:29:06 pm »
I would also like to get 12 7Ghz GDDR5 chips and add them to the back of the board if it has the chip spots are available so that I can get 6GB. That might work with or without modding. I don't have the 780 Ti yet -- waiting on the EGVA ACX OC version so that I can get binned parts, but hoping for the next month or so. Some people might overclock to 1100 and others manage over 1300. Don't really care about the ACX cooler, that will eventually be replaced with water cooling.

For the amount that would cost you, you might as well get a Titan to begin with. In fact, for the number of cards you'll destroy soldering on the BGA chips manually, you might as well just get a K6000 outright.

Binning these days does next to nothing. Silicon manufacturing has gotten to the point where all chips will do the same speeds to within a few %, and those last few % are down to luck and generally not worth bothering with. Haven't you noticed that in the past decade if the clock range of a particular Intel CPU series was, say, 2.4GHz for the slowest model and 3.33 GHz for the fastest model, they will all do about 3.4GHz regardless of what they were sold as? Granted, Intel silicon is better than most, but it's not THAT much better.

Thing is, I would be willing to pay for both of the features. But nobody is even talking about releasing either a 6GB card or unlocking the DP. I don't game, but I can use the DP for computation. I also want the 780 Ti because I will upping to the ASUS 39" UHD monitor once they release it. Both computation and the UHD resolution make 3GB a little iffy.

Sounds like what you really should be getting is a K6000. Full shader count of the 780Ti, full DP performance of the Titan, and 12GB of RAM. Yours for a mere £4K on ebay. Given I've not been able to get either my 690 or the Titan to work virtualized, I'm tempted to just trade them in for a pair of genuine K5000 cards, seen as they are now going for around £600 on ebay.

You may be right about the cost of the chips. That will figure in my decision. I was assuming that the cost would only be slightly unreasonable. Like, perhaps 1/4 the cost of the 780 Ti itself. These are not cutting edge chips (anymore).

As far as BGA soldering is concerned, I haven't done that. However, wasn't there someone here who has BGA equipment? Perhaps I could make a deal with him (or someone else) to actually do the soldering. The amount of memory is usually picked up by the amount that is physically present. And, of course, the BIOS can be tweaked to be consistent so I assume that a memory mod would work.

You're right about binning -- until  you overclock. Then the differences in the chips show up. I am currently running a 3770K on stock air at 4.4Ghz. It was not stable at 4.6, but almost. I possibly could do 4.5 but haven't pushed it. Some people have gotten as high as 4.7 and been stable. My understanding is that EVGA bins the ACX OC parts so that at each voltage level the chip does just a little bit better in speed and power consumption. If you want to OC it, that can make a difference. Even if you don't you can run a bit cooler.

I'm willing to pay for the cost of the card -- but not a factor of six! You would be looking at no more than 1/14 increase over the Titan and probably would NOT clock anywhere near as high. So, if the Titan is 1000, then 1200 or so is reasonable. Perhaps 1500 for high speed board.

It might be a completely moot point. Changing the ID might not allow the DP to be unlocked. We don't know if the switch to turn it on is is done by fuses or simply made not available unless it is the Titan. If, as seems likely from previous posts, the ID change can be done via soft strapping (since probably only nibble four is involved and the only board changes appear to be in the power distribution area) then it will be a cheap test.

All of the other attempts to turn on DP have not involved changing to a Titan which has a software switch, but rather something else where it is always turned on.

Can't do anything though until people start getting cards.
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #701 on: November 09, 2013, 06:13:07 pm »
I would also like to get 12 7Ghz GDDR5 chips and add them to the back of the board if it has the chip spots are available so that I can get 6GB. That might work with or without modding. I don't have the 780 Ti yet -- waiting on the EGVA ACX OC version so that I can get binned parts, but hoping for the next month or so. Some people might overclock to 1100 and others manage over 1300. Don't really care about the ACX cooler, that will eventually be replaced with water cooling.

For the amount that would cost you, you might as well get a Titan to begin with. In fact, for the number of cards you'll destroy soldering on the BGA chips manually, you might as well just get a K6000 outright.

Binning these days does next to nothing. Silicon manufacturing has gotten to the point where all chips will do the same speeds to within a few %, and those last few % are down to luck and generally not worth bothering with. Haven't you noticed that in the past decade if the clock range of a particular Intel CPU series was, say, 2.4GHz for the slowest model and 3.33 GHz for the fastest model, they will all do about 3.4GHz regardless of what they were sold as? Granted, Intel silicon is better than most, but it's not THAT much better.

Thing is, I would be willing to pay for both of the features. But nobody is even talking about releasing either a 6GB card or unlocking the DP. I don't game, but I can use the DP for computation. I also want the 780 Ti because I will upping to the ASUS 39" UHD monitor once they release it. Both computation and the UHD resolution make 3GB a little iffy.

Sounds like what you really should be getting is a K6000. Full shader count of the 780Ti, full DP performance of the Titan, and 12GB of RAM. Yours for a mere £4K on ebay. Given I've not been able to get either my 690 or the Titan to work virtualized, I'm tempted to just trade them in for a pair of genuine K5000 cards, seen as they are now going for around £600 on ebay.

You may be right about the cost of the chips. That will figure in my decision. I was assuming that the cost would only be slightly unreasonable. Like, perhaps 1/4 the cost of the 780 Ti itself. These are not cutting edge chips (anymore).

As far as BGA soldering is concerned, I haven't done that. However, wasn't there someone here who has BGA equipment? Perhaps I could make a deal with him (or someone else) to actually do the soldering. The amount of memory is usually picked up by the amount that is physically present. And, of course, the BIOS can be tweaked to be consistent so I assume that a memory mod would work.

If you are doing this for fun as the primary motivation, that's great, but otherwise you need to factor in the cost of the man-hours that is going to go into all of this. Combine that with the outcome being far from certain, and the economics of it start to look very questionable.

You're right about binning -- until  you overclock. Then the differences in the chips show up. I am currently running a 3770K on stock air at 4.4Ghz. It was not stable at 4.6, but almost. I possibly could do 4.5 but haven't pushed it. Some people have gotten as high as 4.7 and been stable. My understanding is that EVGA bins the ACX OC parts so that at each voltage level the chip does just a little bit better in speed and power consumption. If you want to OC it, that can make a difference. Even if you don't you can run a bit cooler.

I'm willing to pay for the cost of the card -- but not a factor of six! You would be looking at no more than 1/14 increase over the Titan and probably would NOT clock anywhere near as high. So, if the Titan is 1000, then 1200 or so is reasonable. Perhaps 1500 for high speed board.

My view of OC is that there are a lot of "fishermen's tales" and that what most people consider "stable" isn't actually all that stable. If you can run each of the OCCT tests and the multi-threaded tmpfs file hash tests on linux for 24 hours each without any errors, then I may be a little more convinced. Most of the time when people have claimed stability I have been able to shake the machine loose in under 10 minutes.

It might be a completely moot point. Changing the ID might not allow the DP to be unlocked. We don't know if the switch to turn it on is is done by fuses or simply made not available unless it is the Titan. If, as seems likely from previous posts, the ID change can be done via soft strapping (since probably only nibble four is involved and the only board changes appear to be in the power distribution area) then it will be a cheap test.

All of the other attempts to turn on DP have not involved changing to a Titan which has a software switch, but rather something else where it is always turned on.

Can't do anything though until people start getting cards.

The whole notion of DP boost by changing the 780's ID to Titan stems from the part in Tom's Hardware review that suggests that the DP shader clock speed is driver controlled.
 

Offline FireDragon

  • Regular Contributor
  • *
  • Posts: 62
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #702 on: November 09, 2013, 06:51:53 pm »
If you are doing this for fun as the primary motivation, that's great, but otherwise you need to factor in the cost of the man-hours that is going to go into all of this. Combine that with the outcome being far from certain, and the economics of it start to look very questionable.

My view of OC is that there are a lot of "fishermen's tales" and that what most people consider "stable" isn't actually all that stable. If you can run each of the OCCT tests and the multi-threaded tmpfs file hash tests on linux for 24 hours each without any errors, then I may be a little more convinced. Most of the time when people have claimed stability I have been able to shake the machine loose in under 10 minutes.

The whole notion of DP boost by changing the 780's ID to Titan stems from the part in Tom's Hardware review that suggests that the DP shader clock speed is driver controlled.

Nothing about my current build is cost effective. If include the cost of man hours, then the total cost is astronomical. But, it has also been an excellent project for learning things. Extreme - yes. Just the disk storage system is nearly 5K (2K in SSDs, another 1.5K in raid controller / expanders and the rest in hard drives and optical drives). The water cooling is already over 2k and going up. And none of that includes the tools that I have purchased to do the build. So, while it will be a system that I actually use, and is intended for long term use and easy changes, it is also intended to be the system I want without really counting the cost (doesn't mean that I don't consider tradeoffs or have budget limitations - funding is via overtime).

For 4.6Ghz I was able to run Prime95 for several hours before it failed. I was going for a full day.

I understand where the DP boost idea comes from. But, for a Titan, it can clearly be changed in software using the control panel. No one (that I know about) has tracked that all the way down to the hardware / bios, but clearly on the Titan the block is not via fuses. It is far from certain, but if the boost option is simply turned off if it is not a Titan, it is possible that no fuses are involved on the 780 Ti parts either. They appear to be using the same chips. It is certainly worth a try, especially if it has zero cost to try.
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #703 on: November 09, 2013, 07:12:53 pm »
If you are doing this for fun as the primary motivation, that's great, but otherwise you need to factor in the cost of the man-hours that is going to go into all of this. Combine that with the outcome being far from certain, and the economics of it start to look very questionable.

My view of OC is that there are a lot of "fishermen's tales" and that what most people consider "stable" isn't actually all that stable. If you can run each of the OCCT tests and the multi-threaded tmpfs file hash tests on linux for 24 hours each without any errors, then I may be a little more convinced. Most of the time when people have claimed stability I have been able to shake the machine loose in under 10 minutes.

The whole notion of DP boost by changing the 780's ID to Titan stems from the part in Tom's Hardware review that suggests that the DP shader clock speed is driver controlled.

Nothing about my current build is cost effective. If include the cost of man hours, then the total cost is astronomical. But, it has also been an excellent project for learning things. Extreme - yes. Just the disk storage system is nearly 5K (2K in SSDs, another 1.5K in raid controller / expanders and the rest in hard drives and optical drives). The water cooling is already over 2k and going up. And none of that includes the tools that I have purchased to do the build. So, while it will be a system that I actually use, and is intended for long term use and easy changes, it is also intended to be the system I want without really counting the cost (doesn't mean that I don't consider tradeoffs or have budget limitations - funding is via overtime).

So just save up for a K6000 and save yourself the disappointment.

For 4.6Ghz I was able to run Prime95 for several hours before it failed. I was going for a full day.

Prime95 is a completely useless test of stability. I have seen Prime95 and various Pi calculators run for days without error only for OCCT or hash calculations to return an error in under 30 seconds. If anything, what you said confirms my view of fishermen's tales.

I understand where the DP boost idea comes from. But, for a Titan, it can clearly be changed in software using the control panel. No one (that I know about) has tracked that all the way down to the hardware / bios, but clearly on the Titan the block is not via fuses. It is far from certain, but if the boost option is simply turned off if it is not a Titan, it is possible that no fuses are involved on the 780 Ti parts either. They appear to be using the same chips. It is certainly worth a try, especially if it has zero cost to try.

As long as you are happy with the assumed outcome of no DP improvement, that's fine. There are several things to try, including changing the device ID and flashing the Titan BIOS onto the card. But that doesn't mean it'll succeed. I have a 4GB GTX680 running a K5000 BIOS, and there is no obvious performance benefit in any test over a standard GTX680 - the only advantage is in persuading the drivers to allow VGA passthrough operation - and even that doesn't seem to work on models more recent than the GTX680.
« Last Edit: November 10, 2013, 09:48:06 am by gordan »
 

Offline FireDragon

  • Regular Contributor
  • *
  • Posts: 62
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #704 on: November 10, 2013, 04:58:35 am »

So just save up for a K6000 and save yourself the disappointment.

Prime95 is a completely useless test of stability. I have seen Prime95 and various Pi calculators run for days without error only for OCCT or hash calculations to return an error in under 30 seconds. If anything, what you said confirms my view of fisherments' tales.

As long as you are happy with the assumed outcome of no DP improvement, that's fine. There are several things to try, including changing the device ID and flashing the Titan BIOS onto the card. But that doesn't mean it'll succeed. I have a 4GB GTX680 running a K5000 BIOS, and there is no obvious performance benefit in any test over a standard GTX680 - the only advantage is in persuading the drivers to allow VGA passthrough operation - and even that doesn't seem to work on models more recent than the GTX680.

A K6000 isn't what I want.

I am interested in alternate ways of determining stability, Prime95 appears to be what most overclockers used to validate a clock. I want to increase the clock rate, but not at the expense of stability or correctness. I do not run Linux, I am currently using Windows 7 64. I would appreciate any suggestions you have for validating a clock as stable.

I am not a dedicated overclocker, but once I am on water cooling I would like to see 5Ghz on my 3770K on a 24/7 basis. Moderate overclocking is not really overclocking at all because Intel drastically underspecifies the performance of the cpus. It isn't until you push past about 4.5Ghz that you are really in the overclocking range.

I will accept no DP improvement. I am just hoping and this thread has exposed an interesting possibility. Unlike others in this thread, I am not trying to turn a consumer product into a commercial product, nor am I interested in the card sharing in VMs so my use of the information is for different purposes. While I can use the card for compute purposes, I would not be doing so professionally. And, if I really needed the compute boost, then I would get a K6000 or whatever would be best.

Why do you appear to be so against the idea? I would think it that it would be of interest to at least some of the participants of this forum and certainly is harmless. It would also resolve an open question about the Titan vs 680 / 680 Ti.
 

Offline aardvark0

  • Newbie
  • Posts: 1
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #705 on: November 10, 2013, 07:10:00 am »
I have a 4GB GTX680 running a K5000 BIOS, and there is no obvious performance benefit in any test over a standard GTX680 - the only advantage is in persuading the drivers to allow VGA passthrough operation - and even that doesn't seem to work on models more recent than the GTX680.

Does the K5000 modified GTX680 allow 30bit color on the displayport output like a real K5000?
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #706 on: November 10, 2013, 08:19:24 am »

So just save up for a K6000 and save yourself the disappointment.

Prime95 is a completely useless test of stability. I have seen Prime95 and various Pi calculators run for days without error only for OCCT or hash calculations to return an error in under 30 seconds. If anything, what you said confirms my view of fishermen's tales.

As long as you are happy with the assumed outcome of no DP improvement, that's fine. There are several things to try, including changing the device ID and flashing the Titan BIOS onto the card. But that doesn't mean it'll succeed. I have a 4GB GTX680 running a K5000 BIOS, and there is no obvious performance benefit in any test over a standard GTX680 - the only advantage is in persuading the drivers to allow VGA passthrough operation - and even that doesn't seem to work on models more recent than the GTX680.

I am interested in alternate ways of determining stability, Prime95 appears to be what most overclockers used to validate a clock.

Because it helps perpetuate their fishermen's tales while giving them an appearance of validity. It is one of the reasons why projects such as F@H nowdays have to send each work unit out multiple times in hope to get two matching results back for cross-check, resulting in a huge waste of computational resources.

I want to increase the clock rate, but not at the expense of stability or correctness. I do not run Linux, I am currently using Windows 7 64. I would appreciate any suggestions you have for validating a clock as stable.

OCCT is a good place to start (small test for CPU/cache testing, large test for MCH/RAM testing, then re-check with linpack test for good measure as I have seen it throw errors where the small and large tests passed).

I am not a dedicated overclocker, but once I am on water cooling I would like to see 5Ghz on my 3770K on a 24/7 basis.

Dream on.

Moderate overclocking is not really overclocking at all because Intel drastically underspecifies the performance of the cpus. It isn't until you push past about 4.5Ghz that you are really in the overclocking range.

They don't overspecify them by much per se. You'll find there's maybe 10% of headroom on the top of the range model, and you probably want to keep most of that 10% margin for error. The smart way to approach the matter is to buy the bottom of the range CPU and crank it up to the top of the range clock speed as that will by and large "just work". Above that things rapidly start to get uneconomical in terms of cost and effort.

Why do you appear to be so against the idea? I would think it that it would be of interest to at least some of the participants of this forum and certainly is harmless. It would also resolve an open question about the Titan vs 680 / 680 Ti.

I am not against it, I was just pointing out that your approach involving getting a 3GB card and soldering another 3GB of RAM chips onto it instead of getting a 6GB card to begin with was at best misguided.

I have a 4GB GTX680 running a K5000 BIOS, and there is no obvious performance benefit in any test over a standard GTX680 - the only advantage is in persuading the drivers to allow VGA passthrough operation - and even that doesn't seem to work on models more recent than the GTX680.

Does the K5000 modified GTX680 allow 30bit color on the displayport output like a real K5000?

I don't know - I don't have any displayport monitors.
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #707 on: November 10, 2013, 09:28:23 am »
oguz286, any chance of a howto write up on the subject of memory size modding?
I got my hands on a K6000 BIOS from here:
http://forum.techinferno.com/nvidia-video-cards/5022-%5Breq%5D-raise-maximum-power-target-quadro-k6000-bios-attached.html
and I want to try it on my Titan to see if VGA passthrough starts working, but I need to halve the memory size first.

Also, do you have a download link for a CUDA binary you want me to test on my Titan with a K20c BIOS? Or is that no longer of interest?

Also regarding your article on the website about PCIe 2.0 vs. 3.0, are you saying that flashing a K20c BIOS onto the 780 makes it go in to PCIe 2.0 mode? How can you tell? If you are judging this by what GPU-Z says, I wouldn't trust it too much - my Titan shows up as PCIe 1.1 on a 2.0 motherboard, while my GTX690 shows up as PCIe 3.0 even though the motherboard most definitely cannot do more than PCIe 2.0.
 

Offline KamranB

  • Newbie
  • Posts: 2
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #708 on: November 11, 2013, 06:10:03 pm »
Nice! If it works out I'm getting a Titan too :) Then I can concentrate on 'Teslafying' the GTX780 completely and running CUDA on it without getting unknown errors when calling cudaMemcpy.

I must say I am really curious if you will be able to figure out where and how the memory configuration is stored. If you look a back on this thread on page 38, you will find this post:
https://www.eevblog.com/forum/chat/hacking-nvidia-cards-into-their-professional-counterparts/msg292021/#msg292021
containing the hex diff between 1.5GB and 3GB variants of a GTX580 BIOS of the same version number. Unless I made a huge mistake somewhere (or the BIOSes are mislabeled on TPU, I no longer have a GTX580 I could flash with those BIOSes to test), the memory difference should be encoded somewhere in those 10 lines.

Yeah I already checked the diffs of many BIOSes but the actual size of the memory is not stored literally in the BIOS. The type of memory, the configuration, the clocks etc. are stored as a table in the BIOS and according to these variables you can calculate what the memory size is.

Can you elaborate on this? What  byte offset locations in the GeForce BIOS contains the number of chips and their size?

Back in the GeForce 2 days, you could turn certain models into a Quadro 2, though in those cases it wasn't just a straight performance unlock. It was a tradeoff. Something like far better CAD and wireframe performance, but games weren't so well optimized anymore. It wasn't something a gamer would do to get a few extra FPS.

Not really the case any more. On GTS450 -> Quadro 2000 (GF106) there is a marginal improvement in some SPEC components (e.g. Maya gets a 40% boost, the rest remains the same), and I hadn't noticed any gaming degradation. On GF100 (GTX470/GTX480) and later there is no performance difference in the SPEC benchmarks, but there is a memory I/O boost (potentially up to double) from enabling the bidirectional async DMA. From the GTX580 (Q7000) onward there is no difference in any aspect of performance that I have been able to observe. I have a GTX680 running a K5000 BIOS and there is no obvious performance difference in either SPEC or gaming benchmarks.

We faced the exact problem that you have mentioned in the forum and changed the resistors in a way to get a Quadro Graphic card but it did not work for us. By the way I see that there are little differences in our board with the image that you have shared in the forum.
1) In the upper column that you showed there is a 25K resistor that should be removed and a 20K resistor should be mount below that one.Ok and we did it. But in our board the second right side column is different. There is a resistor on top of this row
which is not on your board and reversely there is a resistor under that on your board which is not present in our board.

2) We plugged in the board and there were one long beep and three short beeps on windows startup. And it did not work.

Did you change just the 3rd nibble resistor pair or did you change the 4th one as well? I suggest you put the 4th nibble (lower pair in the photo) back as it was and soft-mod that part instead. For the 3rd nibble resistor, you can either leave it off and stabilize wiith the soft-mod on the lowest bit of the nibble, or put in a resistor. With 25K or more, the 3rd nibble will go to 0xB.

1)   Sorry I don’t know exactly which resistors determine which nibbles but I’m sure that I did the same thing that was proposed in the image as highlighted by the red rectangles. Removed the 25K on the right top. Added a 20K below the removed one. And changed the forth column top resistor from 5K to 15K. Could you please show in the image what is your suggestion??
 
2)   Also could you please tell me more about soft-mod?? Is that some sort of firmware update for the graphic Card? What do you mean by stabilize with soft-mod??

Thanks again
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #709 on: November 11, 2013, 07:13:22 pm »
1)   Sorry I don’t know exactly which resistors determine which nibbles but I’m sure that I did the same thing that was proposed in the image as highlighted by the red rectangles. Removed the 25K on the right top. Added a 20K below the removed one. And changed the forth column top resistor from 5K to 15K. Could you please show in the image what is your suggestion??

My first suggestion is that you understand what the mod does and why it works before taking a soldering iron to an expensive piece of equipment.

The top pair of resistors controls the 3rd nibble of the device ID. Nibble is half a byte. So if your card's PCI device ID is, e.g. 0x1180 (for a GTX680), the 3rd nibble is 8, and the 4th nibble s 0. The resistors controlling the 3rd nibble are at the upper pair in the picture, and for the 4th nibble are the lower ones.

The values for the 4th nibble are fully configurable in both hardware or software (see first page for details of resistor values for the hard-mod). The 3rd nibble is partially controllable in software (the 5th bit is controllable via a soft-strap in the BIOS, the other 3 are not). Thus, to convert to a Tesla K10 no hardware modification is required - it can be done in software. See my post on page 41 of this thread for an explanation of soft-modding on UEFI BIOS cards.

For modifying to a K5000 or K2 you have to hard-mod at least the 3rd nibble resistor pair.

2)   Also could you please tell me more about soft-mod?? Is that some sort of firmware update for the graphic Card? What do you mean by stabilize with soft-mod??

Yes. See page 41.
What I have found is that you can remove the 3rd nibble resistor without replacing it. This is a bad idea, but it does work for most people, at least on the GTX680. The result is that the value of the 3rd nibble will flap between the highest settable value and one blow it (in the case of GK104 based cards that means 0xA or 0xB, randomly after reboots). This is only a difference in the least significant bit of the 3rd nibble, so you can "stabilize" it by setting that bit in the soft strap.

If you are not getting even a POST, then something else has been damaged/modified on the card, e.g. if you shorted something by too much solder, static shock damage, or similar. Or you just didn't push the card into the slot properly. Even without any of the hard-strap resistors the card should still POST, but it's device ID may be unstable (for a lot of people, myself included, no resistors in any of the 4 locations results in reasonably stable 0x11BF (K2) device ID on the GTX680. Note that leaving pins disconnected on an analogue circuit is a bad idea as it can cause inpredictable behaviour, so don't do this unless you have no choice.
 

Offline C_Coffie

  • Newbie
  • Posts: 3
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #710 on: November 12, 2013, 02:46:28 am »
So I read almost halfway through the thread and then searched the rest and I still have some questions. So I current have two GTX 660 (non-ti) that was hoping to convert to k4000's. I simply would like to do this for the virtualization benefits. I currently have 32gb ram in my desktop and it's killing not being able to use all of it. I was thinking of implementing a Xen setup but I think I would need the gtx 660 profession counterpart which I believe is the k4000. I know early on in this post- https://www.eevblog.com/forum/chat/hacking-nvidia-cards-into-their-professional-counterparts/msg203239/#msg203239 They posted a potential position of the resistors but I didn't see anything come of this. So my question is if anyone has successfully modified a gtx 660 to be a k4000? or would you guys be able to point me towards how I would figure this out myself? Thanks for the help. Also you guys are awesome.
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #711 on: November 12, 2013, 10:36:47 am »
So I read almost halfway through the thread and then searched the rest and I still have some questions. So I current have two GTX 660 (non-ti) that was hoping to convert to k4000's. I simply would like to do this for the virtualization benefits. I currently have 32gb ram in my desktop and it's killing not being able to use all of it. I was thinking of implementing a Xen setup but I think I would need the gtx 660 profession counterpart which I believe is the k4000. I know early on in this post- https://www.eevblog.com/forum/chat/hacking-nvidia-cards-into-their-professional-counterparts/msg203239/#msg203239 They posted a potential position of the resistors but I didn't see anything come of this. So my question is if anyone has successfully modified a gtx 660 to be a k4000? or would you guys be able to point me towards how I would figure this out myself?

You might want to look at oguz286's web site (link under his username). He produced a diagram of how the ID strapping seems to hang together circuit-wise on the GTX780. You could arm yourself with a multimeter and try to find if there is a similar arrangement on the 660. As an added bonus, if there is, you might be able to get away with a similar trick to avoid manually soldering 0402 components.

ION: In Linux if you run the nouveau driver, you can cat the BIOS from the GPU via /sys/kernel/debug/dri/*/vbios.rom. The reason this is useful is because what you will get out of it is the BIOS stripped of the UEFI headers, which takes guessing out of where to start stripping the footer out. It can also be educational to see the length of the UEFI header for those that haven't figured it out by eyeballing it (0x400 bytes (1KiB) on GK104, 0x600 (1.5KiB) bytes on GK110).
« Last Edit: November 12, 2013, 12:16:07 pm by gordan »
 

Offline verybigbadboy

  • Contributor
  • Posts: 38
  • Country: ru
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #712 on: November 13, 2013, 07:36:38 am »
hello, may be it is offtopic

I found that my modded gtx650 card to grid k1, and gtx680 -> grid k2 have WGL_NV_gpu_affinity
But I have issue, These card are in different PCs.

Did someone try to use this extension on modded cards?
6'7''
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #713 on: November 13, 2013, 08:03:44 am »
What is WGL_NB_gpu_affinity supposed to do?
Did you mod just the device ID, or did you flash the Grid BIOS onto the card?
 

Offline verybigbadboy

  • Contributor
  • Posts: 38
  • Country: ru
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #714 on: November 13, 2013, 08:16:39 am »
What is WGL_NB_gpu_affinity supposed to do?
Did you mod just the device ID, or did you flash the Grid BIOS onto the card?

It need for multigpu render to select right render gpu.
I modded device ID only.

I not able to test it because I have only one videocard in each pc :(
6'7''
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #715 on: November 13, 2013, 09:16:11 am »
What is WGL_NB_gpu_affinity supposed to do?
Did you mod just the device ID, or did you flash the Grid BIOS onto the card?

It need for multigpu render to select right render gpu.
I modded device ID only.

I not able to test it because I have only one videocard in each pc :(

My 690/K2 doesn't show that capability.
My 680/K5000 does.

I can also confirm that the 680 also has this capability with it's original BIOS (as opposed to the K5000 BIOS that I normally run it with).

Tested using:
http://www.equalizergraphics.com/downloads/affinityCheck.exe and
http://www.geeks3d.com/20131113/gpu-caps-viewer-1-19-0-videocard-information-utility-opengl-opencl-geforce-quadro-radeon-gpu/

Another thing I can also confirm is that in the control panel I get the ECC option now even with the stock BIOS, as had been previously reported. It would appear that once the Quadro/Grid BIOS has been flashed onto the card, something else ends up changed (FPGA microcode push?) and flashing the BIOS back doesn't undo it.

Quite strange. I wonder what is going on there.
« Last Edit: November 13, 2013, 09:43:32 pm by gordan »
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #716 on: November 13, 2013, 11:26:15 pm »
I have made progress on adjusting the VRAM size in the BIOS. What happens is that BIOS has "init scripts", which is basically binary init code that gets run as-is on the GPU to initialize it.

I worked on comparing two similar BIOSes (between Gainward Phantom GTX680 2GB and 4GB models):
http://www.techpowerup.com/vgabios/130996/Gainward.GTX680.2048.120710.rom
http://www.techpowerup.com/vgabios/131820/Gainward.GTX680.4096.120709.rom

I first adjusted the clocks and voltages on the 2GB BIOS to match the 4GB BIOS to reduce the differences.

Then I compared the BIOSes and found that there were just 9 blocks that differ between them. With a bit of binary search experimentation, I got there in the end.

The first two blocks caused screen corruption and lack of card initialization (VGA 4-bit color only in Windows, some corruption in text mode) when applied to the 4GB card. The 3rd block proved to be the one containing the memory init payload. Here is the complete diff block:

Code: [Select]
2529,2532c2529,2532
< 0009e00: 0000 90cf 0000 90cf 0000 90cf 0000 96cf  ................
< 0009e10: 0200 96cf 0200 9055 0200 9055 0200 9044  .......U...U...D
< 0009e20: 0200 9044 0200 a055 0200 a055 0200 a055  ...D...U...U...U
< 0009e30: 0200 a055 026e 0402 1100 ffff ffff 0000  ...U.n..........
---
> 0009e00: 0000 90cf 0000 90cf 0000 90cf 0000 968f  ................
> 0009e10: 0200 908f 0200 9055 0200 9055 0200 9044  .......U...U...D
> 0009e20: 0200 9044 0200 a055 0200 a055 0200 9055  ...D...U...U...U
> 0009e30: 0200 9055 026e 0402 1100 ffff ffff 0000  ...U.n..........

And again with the deltas highlighted in red:
2529,2532c2529,2532
< 0009e00: 0000 90cf 0000 90cf 0000 90cf 0000 96cf
< 0009e10: 0200 96cf 0200 9055 0200 9055 0200 9044
< 0009e20: 0200 9044 0200 a055 0200 a055 0200 a055
< 0009e30: 0200 a055 026e 0402 1100 ffff ffff 0000
---
> 0009e00: 0000 90cf 0000 90cf 0000 90cf 0000 968f
> 0009e10: 0200 908f 0200 9055 0200 9055 0200 9044
> 0009e20: 0200 9044 0200 a055 0200 a055 0200 9055
> 0009e30: 0200 9055 026e 0402 1100 ffff ffff 0000


The next part of the effort will involve dissasembling this block of init code to figure out what it actually does. In absence of more info on what he did on the 780/K20c mod from oguz286's, with a bit of luck, this will yield enough clues on how to modify the K6000 BIOS for the Titan.

Stay tuned.
 

Offline tgtweak

  • Newbie
  • Posts: 1
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #717 on: November 14, 2013, 05:51:21 pm »
Anybody try NVENC functionality on their modded 680-> K2?  This is a very big part of the grid/quadro cards featureset. 
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #718 on: November 14, 2013, 06:59:16 pm »
Anybody try NVENC functionality on their modded 680-> K2?  This is a very big part of the grid/quadro cards featureset.

Explain in detail how to test it (including a testing binary if appropriate) and I'll let you know.
 

Offline Rutgerb

  • Newbie
  • Posts: 1
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #719 on: November 14, 2013, 07:06:11 pm »
Can someone help me? I am trying to make a k5000 from my MSI gtx 660 ti, but I don't know where I can find the resistors and what I need to change them with.

Thanks!
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #720 on: November 14, 2013, 07:38:43 pm »
On the GPU side of the card, I can see two 8-pin chips that look like they could be the EEPROM. You'll need to check the writing on them and google it to establish which is the EEPROM. Once you have established that, take a look at this diagram:

http://www.guztech.nl/wordpress/index.php/2013/11/researching-nvidia-gpus-geforce-gtx780-and-gtx-titan-to-tesla-k20-and-tesla-k20x/

See if you can find a resistor with an alternate position that goes to SCLK pin on the EEPROM. That should be the 3rd nibble. That's the one you'll need to change, most likely to 40K. That should get you the device ID 0x11B3 (from the default 0x1183). From there, edit the BIOS at 0x1C as discussed a few pages pack to set bits 0-4 accordingly for the 4th nibble.
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #721 on: November 14, 2013, 11:39:47 pm »
I've been doing a bit of digging to try to figure out why my GTX690 is, infuriatingly, not playing ball in a VM, and I just noticed this in a new tool that I decided to try after seeing it in verybigbadboy's recent post.

Spot the device ID it reports: 0x11A8.
This must be coming from the driver since Windows Device Manager and Linux lspci both report it as device ID 0x11BF.

The crucial point is that I only part hard-modded the card. I removed the 3rd nibble resistor, and fixed the flapping by setting the soft-strap bit-5 high to compensate. I soft modded the low 4 bits.

What's the bet that the driver also keys off the same place where GPU Caps viewer gets the device ID, and pretends to work fine while in fact not doing it's Quadro thing?

* gordan will take a soldering iron to the GTX690 tomorrow put some 40K resistors on and see if that fixes the problem.

The interesting thing is that this seems to be new to the Kepler GPUs. It isn't the case with the Fermi GPUs - on the Fermis the soft-mod sticks completely.
« Last Edit: November 14, 2013, 11:47:28 pm by gordan »
 

Offline vacaloca

  • Contributor
  • Posts: 14
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #722 on: November 15, 2013, 06:08:03 pm »
oguz286, any chance of a howto write up on the subject of memory size modding?
I got my hands on a K6000 BIOS from here:
http://forum.techinferno.com/nvidia-video-cards/5022-%5Breq%5D-raise-maximum-power-target-quadro-k6000-bios-attached.html
and I want to try it on my Titan to see if VGA passthrough starts working, but I need to halve the memory size first.

Also, do you have a download link for a CUDA binary you want me to test on my Titan with a K20c BIOS? Or is that no longer of interest?

Also regarding your article on the website about PCIe 2.0 vs. 3.0, are you saying that flashing a K20c BIOS onto the 780 makes it go in to PCIe 2.0 mode? How can you tell? If you are judging this by what GPU-Z says, I wouldn't trust it too much - my Titan shows up as PCIe 1.1 on a 2.0 motherboard, while my GTX690 shows up as PCIe 3.0 even though the motherboard most definitely cannot do more than PCIe 2.0.

Bhahaha, I posted that Quadro K6000 BIOS a bit ago there... which reminds me, I'm going to sent a message to svl7 to see if he can do more tweaks to remove the power target throttling past 225W I've experienced. I'm also curious on the PCIe 2.0 vs 3.0 speeds on a GPU modded to K20c BIOS. If PCIe 3.0 is possible, the only thing missing would be the dual-DMA engines.
 

Offline gordan

  • Frequent Contributor
  • **
  • Posts: 277
  • Country: 00
    • Tech Articles
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #723 on: November 15, 2013, 07:38:37 pm »
oguz286, any chance of a howto write up on the subject of memory size modding?
I got my hands on a K6000 BIOS from here:
http://forum.techinferno.com/nvidia-video-cards/5022-%5Breq%5D-raise-maximum-power-target-quadro-k6000-bios-attached.html
and I want to try it on my Titan to see if VGA passthrough starts working, but I need to halve the memory size first.

Also, do you have a download link for a CUDA binary you want me to test on my Titan with a K20c BIOS? Or is that no longer of interest?

Also regarding your article on the website about PCIe 2.0 vs. 3.0, are you saying that flashing a K20c BIOS onto the 780 makes it go in to PCIe 2.0 mode? How can you tell? If you are judging this by what GPU-Z says, I wouldn't trust it too much - my Titan shows up as PCIe 1.1 on a 2.0 motherboard, while my GTX690 shows up as PCIe 3.0 even though the motherboard most definitely cannot do more than PCIe 2.0.

Bhahaha, I posted that Quadro K6000 BIOS a bit ago there... which reminds me, I'm going to sent a message to svl7 to see if he can do more tweaks to remove the power target throttling past 225W I've experienced. I'm also curious on the PCIe 2.0 vs 3.0 speeds on a GPU modded to K20c BIOS. If PCIe 3.0 is possible, the only thing missing would be the dual-DMA engines.

PCIe 3.0 definitely is possible if the card can do it as a 780, it's just a matter of finding the right bit in the binary init script in the BIOS payload that configures the PCIe speed. But that's not easy, unless you are fluent in talking binary to GPU registers.

Dual DMA engines are most likely not going to work - from the GTX580 onward I'm reasonably certain they were laser cut from the silicon. GTX480 was the last card where it was still present on GeForce cards. But who knows, maybe they left it in some of the GK110 GeForce cards. It certainly doesn't look like it's available on GK104s.
 

Offline takeawaydave

  • Newbie
  • Posts: 1
Re: [MOVED] Hacking NVidia Cards into their Professional Counterparts
« Reply #724 on: November 16, 2013, 10:47:24 am »
Can someone help me? I am trying to make a k5000 from my MSI gtx 660 ti, but I don't know where I can find the resistors and what I need to change them with.

Thanks!

Hello - I was taking a look at the photos that you posted however they aren't all that clear. Try to get some better lighting and try and get the tiny details, lettering etc.. all in focus - this will really help in getting identified what needs to be identified :)
 


Share me

Digg  Facebook  SlashDot  Delicious  Technorati  Twitter  Google  Yahoo
Smf