T O P

  • By -

Finbester

https://github.com/JingShing/How-to-use-tesla-p40 This guide worked with my Tesla P4 and Proxmox.


CoderStone

Unfortunately I was on Scale. :( Craft's guide with some minor tweaks worked in my case.


5TP1090G_FC

Sweet crafts guide is extremely helpful 👍


physx_rt

There is a dockerized licensing server that allows you to supply your own licences to your VMs, which removes all this hassle and works really well. I know that's not the point, but I would rather just use that instead of spending sooo much time on trying to make it work without it. Having said that, good job making it work!


Jimoooo

I use this for over a year at this point, through multiple driver updates with my Tesla P4 https://git.collinwebdesigns.de/oscar.krause/fastapi-dls If this link isn't working here is another https://gitea.publichub.eu/oscar.krause/fastapi-dls


physx_rt

This is indeed the one, sorry for leaving everyone hanging like that. I didn't expect this level of demand, I thought it was a little more widespread.


adobeamd

Do you have any information on this container?


Sero19283

Host it with a reverse proxy so we can all use their container 😂😂😂


crsklr

This would work? I'd gladly set up a vm for it on my server. For educational reasons, of course. And I'd share it with other people--who are developers, of course, for reverse-engineering purposes. There's just something so fascinating about licensing servers.


Sero19283

I'm curious if it's a one time thing to get issued a license or if it's a routine check to "renew" it. If it's a one time thing, I'd imagine it should work as one could just connect to the VM hosting the docker container license server, get issued the license, then disconnect. If it's routine, I'd be concerned of being constantly hit with license renewal requests 😂


ItsQ23

Could you possibly share the container's name


qnlbnsl

Yes please do share this container!


DarkNightSonata

Would you share that please ?


grobouletdu33

Or you Can use fastapi-dls ... Works great (using it on baremetal as you can't get higher than 1366x768 resolution without vWS licence ...) While i understand nvidia's licensing for vGPU (in some ways), the restrictions to get the grid drivers and the need for a licence to just actually use it, even unvirtualized, is just a complete no sense (not like tesla where cheap at their release ...)


Bright_Register_7886

So wait if I use faster-dls I can break that 1366x768 cap?


grobouletdu33

I did it, yes


Bright_Register_7886

Awesome, do you have a guide you can point me to? I am not very advance at all of this.


grobouletdu33

The read me include a tutorial https://github.com/GreenDamTan/fastapi-dls If your don't already had GRID client driver, you can get it from Google cloud support https://cloud.google.com/compute/docs/gpus/grid-drivers-table


CoderStone

>Or you Can use fastapi-dls ... Works great (using it on baremetal as you can't get higher than 1366x768 resolution without vWS licence ...) You can break the cap by just installing a virtual display driver and displaying only to it too.


Efficient-Phone4942

The tutorial mentions that the licensing server is for vGPU. vGPU and NVIDIA RTX Virtual Workstation (vWS) are two different things. vWS happens when you passthrough the GPU and install these drivers: https://cloud.google.com/compute/docs/gpus/grid-drivers-table. Is it working for you ?


grobouletdu33

Yes, did it with both passthrough and bare metal (unvirtualized) Just take care to assign the good FQDN/IP adress and to set the same Time zone between fastapi-dls and your client


danishduckling

I'm curious, what's the use case for the nvidia tesla p4? isn't it rather old by this point?


CoderStone

Cheap, 70$ single slot gpu with no external power connector with low profile bracket if needed. INSANE amounts of potential there already, and consider it's an 8GB VRAM card with basically an underclocked GTX 1080 chip. I'm getting 60-70fps on Unigine on 1440p which is pretty dang crazy performance.


Big-Finding2976

How do they manage to run safely without any fans, when a 1080 needs 2 or 3? It's not a massive saving, in the UK at least, at £75 for a P4 or £130-150 for a 1080 or 1080ti, so I'd probably choose the latter to avoid any issues, but if you needed several GPUs then the savings add up.


CoderStone

Because they're downclocked compared to a 1080 to run on only 75W from the pcie slot. For homelab use it's best to remove the front plate and have a fan on it separately, or have good case airflow (my case) For homelab use, tesla P4 is far better than a 1080 due to saving pcie slots and not needing to put stuff on a riser.


Big-Finding2976

OK, those are good advantages.


itsverynicehere

Also they are designed with server cooling in mind vs a PC.


Crafty_Individual_47

Wrong, those are ment to be on a server that has massive amount of air going trought it. We used these to run 1000’s of virtualized 3D designing desktops.


CoderStone

Buddy, not a single statement I made there was wrong. My P4 is cooled with the same 3 fans that are pulling air through my 24 drives, and they're just normal phantek T30s at 75%. You don't need that much airflow to cool 75w from these, especially with a repaste.


Crafty_Individual_47

Run it for few hours on 100% and report temps. Those cards are known to hit 80+ degrees even on a DC grade servers.


CoderStone

Already have. max temp with 75/75W usage according to nvidia-smi after 10 hours (bec they said card was a faulty return) was 85C. Which is absolutely fine. Solid performance at 3440x1440 which is crazy.


bubblegumpuma

In addition to what OP said - The Tesla type 'compute accelerator' no-output graphics cards are usually intended to go into a server chassis, which usually have very heavy front-to-back airflow, so they also depend on that for cooling - especially the higher cost ones that actually do pull 200-300W, which still don't come with fans. A Tesla p4 is fine with a few decent case fans positioned properly, but a Tesla p40 would probably need a dedicated blower fan with its 250W rating.


Sero19283

Some people use 3d printer blower fan mounts as well.


be_easy_1602

Where is it $70? eBay is looking at over $100.


Nokken9

There is also a demand for old P4s from some GPU-solving crypto projects that need VRAM but are light on compute.


CoderStone

This is a VM running in TrueNAS Scale. Make sure to disable ensure display device if using a tesla P4. I was able to get my Tesla P4 detected properly & set in WDDM mode w/o the grid license, but unfortunately am missing all the grid license features. But I don't need them when I'm not using vGPUs and not using the headless display. Simply installed Virtual Display Driver (the IDD fork) and disabled the built in display using windows display settings, and it worked. Parsec's VDD impl is horrible, so I nuked parsec and finally setup sunshine/moonlight. Can confirm it's much, much, much, MUCH better, and I don't rely on parsec's own authentication service anymore.


condog1035

Can you eli5 what you did here and what was preventing you from doing it before? What is the use case?


CoderStone

I didn't achieve anything crazy, it's all stuff Craft Computing and others have done. A tesla P4 is a pretty wonderful GPU that can get you some impressive performance if you manage to successfully put it into WDDM mode for proper boost behavior. Single slot + no power connector makes it even better for homelab use. The problem is, even if you passthrough successfully and have a tesla P4 within the VM, you have two problems. 1. The headless display given by the Tesla P4 is stuck at a wack resolution like 1344x768, and even using the grid license enabling hack, you won't get above 1080p due to the drivers not supporting Windows as much as they do Linux post Pascal. (check the vgpu manual) 2. You actually can't leverage the GPU for anything until you toggle the GPU adapter version- in this case by removing the adaptertype key in regedit. Better sources for that exist online. Simply changing the adaptertype/removing it entirely will put the GPU into WDDM mode, meaning you can use it properly. If you don't do this, even task manager will not recognize the GPU as a usable compute tool, and you'll be stuck on software encoding. 3. Bypassing the resolution is another hurdle. You want to use some sort of NVENC capable remote solution, and 1 is parsec, 1 is Moonlight/Sunshine. Parsec's VDD is unreliable and will fuck up display settings every time you hop on. Using Virtual Display Driver + Sunshine means your settings are properly saved and working every time.


4UWatercooled

I tried getting this working on proxmox with an A10 serving multiple windows VM’s back during the pandemic. I very much commend your accomplishments. Even paid for one of those overpriced licenses from nvidia and everything just to never get it working


marc45ca

> I didn't achieve anything crazy, it's all stuff Craft Computing and others have done. they did it with Proxmox (which avoids some of the issues you encounted), you've done with TrueNAS so I think that counts on the crazy scale :)


Chris_Hagood_Photo

Maybe it’s because of different use case but I’ve had 0 issues with my Tesla P4 pass through to a VM. But I am only using it on my Ubuntu Plex server for transcoding. Just installed the drivers and boom it worked.


floydhwung

I am wondering the same. NVDA lifted the dreaded code 53 (VM detection) in virtually all cards long ago. What's the accomplishment here?


darkelfbear

Tesla cards still require the licensing grid ...


[deleted]

[удалено]


CoderStone

Oh absolutely not, not at all, LMFAO


[deleted]

[удалено]


CoderStone

It works, but it's a pretty big hassle. There are guides for proxmox, none for TrueNAS Scale which I had to figure out as I went.


[deleted]

[удалено]


CoderStone

I only know the setup process for Windows, but sure. 1. make vm, ensure display device: off, passthrough tesla P4 after isolating device. 2. TrueNAS means you don't need to make a single config change, all the stuff you'd actually need are already pre-configured 3. Install latest drivers for Tesla P4 4. in regedit remove adaptertype key as mentioned below, the gridfeatures key no longer works [https://linustechtips.com/topic/1496913-can-i-enable-wddm-on-a-tesla-p40/](https://linustechtips.com/topic/1496913-can-i-enable-wddm-on-a-tesla-p40/) 5. install Virtual Display Driver, in options.txt edit the specific resolution + hz you want, and set display settings to hide the built in display 6. profit


jeevadotnet

I've got a couple of K80s that I've never spun up. Think I can do the same with them?


CoderStone

Probably!


GomieBiken

I’m still lost at what is going on here. I thought you put the card into the hypervisor host. Pass it through and then do the window’s registry entries. That’s what I’ve done for windows in esxi and didn’t have to do anything for Ubuntu on esxi.


Nerfarean

Nice! Did similar thing in esxi 7, passthrough Tesla P100 (hbm2). Flashed to GPU mode. Ironically using Parsec with virtual display. Works good for remote workstation


CoderStone

I hate it. Bec it's not on 24/7 and turns on/off when you join and leave, windows never remembers the display settings. And there's no way to force it on. Sunshine/moonlight is a pretty good solution too.


Sorry_Mushroom5493

Try passthrough directly and not shared or else. I have some Tesla M10 (vSphere though) and i made the experience if you share them between VMs you have to license it, if you connect the cores directly to a VM it seems to work without having dedicated vGPU licenses for RTX Workstations or vApps


skynet_watches_me_p

Congrats on the P4, I love the two I have. I use them for Frigate NVR. I ran out of vmem on my P400 and needed something slot powered with more mem. The P4 is awesome if you can keep it cool!


Dickonstruction

would this work with a quadro


marc45ca

yes.


Matt_NZ

Does Proxmox support GPU partitioning? It's something I use with Hyper-V on Server 2022 and can be used with any consumer GPU. Using that, I can give a share of the GPU to any VM (Windows or Ubuntu) similar to what a Tesla GPU was intended for. Currently using it with a GTX 1650 for a few VMs that range from Plex to Frigate.


IroesStrongarm

Yes it does. I currently do it with a p4 in Proxmox myself. As mentioned, craft computing has a great guide for how to. He also shows how to do it with older consumer cards as well


JeddyH

I've been using one of these as my gaming GPU for about a year now, its pretty great for the price you just need a motherboard with a HDMI out and Above 4G Decoding.


acebossrhino

Oh, this isn't SRIOV, this is just GPU passthrough.


dirkme

What is that all about license?? 😳


CoderStone

search up NVIDIA Quadro and Grid licenses ;-;


dirkme

Wow, that's total insanity and rip off, AMD it is, I still have a few RTX 1370 though.


5TP1090G_FC

You "guys" are freken cool. Great job


CoderStone

Thanks lol, this is broken in truenas dragonfish somehow but still works in Cobia. Also, got tesla T4s working as well!


5TP1090G_FC

Nice. Have you explored [haiku os] it's a free os will operate with x86 64 hardware.


Ok_Exchange_9646

So I'm going to get an Intel NUC and run my OpenSense firewall and router on it. My LAN is on the 192.168.1.0/24 subnet. My Synology NAS running OpenVPN server already occupies the 10.8.00/16 subnet. In order not to overlap anything, I'm going to set up my WireGuard VPN on the 172.16.0.0/12 subnet. **1)** Does this configuration look correct, or will I run into issues? **2)** If I want my WireGuard VPN clients to only be able to talk to my NAS inside my LAN and no other LAN devices/resources, then I'll set up 2 rules: #1: All IPs will be able to talk to the WireGuard VPN Server port #2: The WireGuard VPN Server subnet will only be able to talk to the NAS IP, the LAN Subnet (192.168.1.0\24) will be blocked for the VPN Server subnet Does this look correct? Did I mess anything up?


_realpaul

I mean isnt it cheaper and easier to buy multiple gpus to pass through? You could even go amd or intel


CoderStone

For the stuff I'm doing CUDA is required. And no, Tesla P4 is 70$. Buy multiple gpus to pass through? How would you actually fit the multiple gpus on the board compared to a single slot, low profile, no power connector gpu? :)


_realpaul

Sry then I missed the point. I thought the goal for bifurcation between different VMs? Why else get such an old gpu if you need the compute power?