-
Compressor/Qmaster error: shared storage client timed out while subscribing to…
First, thanks in advance to anybody who can shed some light on what’s going on! I’ve been practically tearing my hair out over this for days. Here’s my Qmaster setup:
computer 1: CONTROLLER, no nodes
– 8TB RAID hooked up via Fiber
– connected to the GigE network switch via a 6-port bond
– cluster storage set to a path on the RAIDcomputers 2, 3, 4, 5: RENDER NODES
* each computer has a 2-port bonded connection with the GigE switch
computer 6: Client, with FCS2 installed.
* connected with a single GigE link
I have set up this cluster primarily for command-line renders, and it works great. I submit command-line renders from the client computer, which get distributed and executed on each node. The command line renders specify a source file on the RAID, and a destination path on the RAID. Everything works great.
I run into trouble when trying to use Compressor with this same setup. The files are on the RAID, and all my computers have an NFS automount that puts it in the /Volumes folder on each computer.
I set up my Compressor job and submit it to the cluster. It submits sucessfully, and distributes the work. After a few seconds, each node gives me a timeout error:
“Shared storage client timed out while subscribing to https://computer1.local/path to cluster storage”
Is this a bandwidth issue? Command line renders work fine, I can render 16 simultaneous Quicktimes to the RAID over NFS. I don’t see much network activity on any of the computers when it’s trying to start the Compressor render, it’s as if it’s not even trying to connect.
If I submit the SAME compressor job to a cluster with nodes ONLY on the controller computer, it renders fine. Clearly the networked nodes are having trouble connecting to the share for some reason.
Does anybody have any ideas? I have tried almost everything to get this to work. Hooking up each node locally to the RAID is NOT an option unfortunately.