AWS Thinkbox Discussion Forums

Houdini Redshift Render won't start - DEBUG (new user)

I have an issue where a submitted task won’t start rendering and i am not sure how i can debug this…looking for clues.

It’s a scene on my local machine(same repository/worker) that i am rendering using Redshift. I have set up the Deadline ROP under my Redshift ROP, everything is under a local $JOB folder structure, everything is mostly cached out in terms of geo and sim caches, apart from a few attributes being modified after caches for shading and common stuff. I can submit the task successfully, everything looks fine but the render won’t start making progress in the monitor. One weird thing is that although Deadline Worker is firing up and seems to respond to the submitted task, all it does is chug on CPU and RAM usage although this is a Redshift render…it’s not even firing up the GPU.

As for basic checks :

  • The render works when done inside Houdini.

  • I made a test scene with a sphere and 1 light, submitted to Deadline and it chewed it in an instant.(So it means my Deadline installation and basic functionalities are covered)

  • The scene i wish to render is far from being heavy geo wise, i’m using only RS proxies and very little geo(it’s motion graphics). I optimized pretty much everything that could be optimized.(The frames takes about 1:30 each on a 1080ti)

I’ve attached a log of the problematic scenedeadlineslave-DESKTOP-5U1FHGC-2021-01-25-0000.zip (32.0 KB)

Looking at the attached log file, most of it is full with Deadline Worker complaining that it cannot find a license.

There is a single line early on that indicates 2-Workers License Free (TRIAL) mode:

2021-01-25 10:09:34:  The Worker is running in TRIAL MODE because a license could not be found, and there are two or less Workers in the Deadline Repository.

However, this is the only occurrence. It might have helped you render the simpler sphere+light test file?

The rest of the log is basically the Worker reporting no license could be found, which usually means that the Workers list had 3 or more Workers listed at the time. Usually, the Worker should turn yellow on the Workers List when this happens…

2021-01-25 10:42:44:  0: STDOUT: Detected Houdini version: (18, 0, 499)
2021-01-25 10:42:44:  0: STDOUT: ['C:\\ProgramData\\Thinkbox\\Deadline10\\workers\\DESKTOP-5U1FHGC\\plugins\\600ee6dd59b0db14706d6103\\hrender_dl.py', '-f', '1156', '1300', '1', '-o', '$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr', '-g', '-d', '/out/Redshift_ROP_render', '-tempdir', 'C:\\ProgramData\\Thinkbox\\Deadline10\\workers\\DESKTOP-5U1FHGC\\jobsData\\600ee6dd59b0db14706d6103\\0_tempIBOMX0', '-arnoldAbortOnLicenseFail', '1', 'C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/600ee6dd59b0db14706d6103/popcorn_balls_wide_render_v001.hiplc']
2021-01-25 10:42:44:  0: STDOUT: Start: 1156
2021-01-25 10:42:44:  0: STDOUT: End: 1300
2021-01-25 10:42:44:  0: STDOUT: Increment: 1
2021-01-25 10:42:44:  0: STDOUT: Ignore Inputs: True
2021-01-25 10:42:44:  0: STDOUT: Output: $JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr
2021-01-25 10:42:44:  0: STDOUT: Driver: /out/Redshift_ROP_render
2021-01-25 10:42:44:  0: STDOUT: Input File: C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/600ee6dd59b0db14706d6103/popcorn_balls_wide_render_v001.hiplc
2021-01-25 10:44:26:  Could not obtain license because:
2021-01-25 10:44:26:  Cannot find license file.
2021-01-25 10:44:26:   The license files (or license server system network addresses) attempted are 
2021-01-25 10:44:26:  listed below.  Use LM_LICENSE_FILE to use a different license file,
2021-01-25 10:44:26:   or contact your software provider for a license file.
2021-01-25 10:44:26:  Feature:       deadline
2021-01-25 10:44:26:  Filename:      C:\flexlm\license.dat
2021-01-25 10:44:26:  License path:  C:\flexlm\license.dat;
2021-01-25 10:44:26:  FlexNet Licensing error:-1,359.  System Error: 2 "No such file or directory"
2021-01-25 10:44:26:  For further information, refer to the FlexNet Licensing documentation,
2021-01-25 10:44:26:  available at "www.flexerasoftware.com".
2021-01-25 10:44:26:  Failed to checkout a license using current configuration, checking for auto configuration...
2021-01-25 10:44:26:  Auto Configuration: Picking configuration based on: DESKTOP-5U1FHGC / 192.168.0.191
2021-01-25 10:44:26:  Auto Configuration: No auto configuration could be detected, using local configuration

If there are some Offline Workers on the list, you can safely delete them. Just make sure that you have only 1 or 2 Workers listed, or email thinkbox-sales@amazon.com and ask for a 30 days evaluation license for whatever number of machines you want to test…

Of course, if you DO have a Thinkbox license server, be sure to configure your Workers (or set up an Auto configuration rule) to point at it.

Hi Bobo.

Thanks for your help.

-There is indeed the test scene in the log file. It’s located between 10:10:25 and 10:11:40.

-I only have one worker in total, everything i am doing is tested on my local machine with no local network connected. There are no offline workers on my list inside Monitor.

Can you right-click the Deadline Launcher on the Worker, select Change License Server, and switch to “No License”? I would like to know what happens then (from the log it appears that you don’t have a license server configured anyway).

Hi!

I followed up on your advice, but still no luck.

Here’s a screenshot of my worker information as per the licensing option you asked me to configure within the launcher.

As well as a log from my latest attempt.

Blockquote
2021-01-27 11:28:06: Triggering house cleaning events…
2021-01-27 11:28:13: Skipping pending job scan because it is not required at this time
2021-01-27 11:28:13: Skipping repository repair because it is not required at this time
2021-01-27 11:28:13: Skipping house cleaning because it is not required at this time
2021-01-27 11:28:20: Skipping pending job scan because it is not required at this time
2021-01-27 11:28:20: Skipping repository repair because it is not required at this time
2021-01-27 11:28:20: Skipping house cleaning because it is not required at this time
2021-01-27 11:28:27: Skipping pending job scan because it is not required at this time
2021-01-27 11:28:27: Skipping repository repair because it is not required at this time
2021-01-27 11:28:27: Skipping house cleaning because it is not required at this time
2021-01-27 11:28:33: Skipping pending job scan because it is not required at this time
2021-01-27 11:28:33: Skipping repository repair because it is not required at this time
2021-01-27 11:28:33: Skipping house cleaning because it is not required at this time
2021-01-27 11:28:41: Skipping pending job scan because it is not required at this time
2021-01-27 11:28:41: Skipping repository repair because it is not required at this time
2021-01-27 11:28:41: Skipping house cleaning because it is not required at this time
2021-01-27 11:28:48: Skipping pending job scan because it is not required at this time
2021-01-27 11:28:48: Skipping repository repair because it is not required at this time
2021-01-27 11:28:48: Skipping house cleaning because it is not required at this time
2021-01-27 11:28:48: Scheduler Thread - Job’s Limit Groups:
2021-01-27 11:28:48: Scheduler Thread - Error mapping D: to D:/: The network name cannot be found.
2021-01-27 11:28:48: 0: Loading Job’s Plugin timeout is Disabled
2021-01-27 11:28:48: 0: SandboxedPlugin: Render Job As User disabled, running as current user ‘julie’
2021-01-27 11:28:51: Synchronization time for job files: 43.385 ms
2021-01-27 11:28:51: Synchronizing Plugin Houdini from C:\DeadlineRepository10\plugins\Houdini took: 0 seconds
2021-01-27 11:28:51: 0: nogui was not added to the CommandLineParser.
2021-01-27 11:28:51: 0: Executing plugin command of type ‘Initialize Plugin’
2021-01-27 11:28:51: 0: INFO: Executing plugin script ‘C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601194b9e15609552c7f70b8\Houdini.py’
2021-01-27 11:28:51: 0: INFO: About: Houdini Plugin for Deadline
2021-01-27 11:28:51: 0: INFO: The job’s environment will be merged with the current environment before rendering
2021-01-27 11:28:51: 0: Done executing plugin command of type ‘Initialize Plugin’
2021-01-27 11:28:52: 0: Start Job timeout is disabled.
2021-01-27 11:28:52: 0: Task timeout is disabled.
2021-01-27 11:28:52: 0: Loaded job: popcorn_balls_wide_render_v001 - /out/Redshift_ROP_render (601194b9e15609552c7f70b8)
2021-01-27 11:28:52: 0: Error mapping D: to D:/: The network name cannot be found.
2021-01-27 11:28:52: 0: Executing plugin command of type ‘Start Job’
2021-01-27 11:28:52: 0: DEBUG: S3BackedCache Client is not installed.
2021-01-27 11:28:52: 0: INFO: Executing global asset transfer preload script ‘C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601194b9e15609552c7f70b8\GlobalAssetTransferPreLoad.py’
2021-01-27 11:28:52: 0: INFO: Looking for legacy (pre-10.0.26) AWS Portal File Transfer…
2021-01-27 11:28:52: 0: INFO: Looking for legacy (pre-10.0.26) File Transfer controller in C:/Program Files/Thinkbox/S3BackedCache/bin/task.py…
2021-01-27 11:28:52: 0: INFO: Could not find legacy (pre-10.0.26) AWS Portal File Transfer.
2021-01-27 11:28:52: 0: INFO: Legacy (pre-10.0.26) AWS Portal File Transfer is not installed on the system.
2021-01-27 11:28:52: 0: Done executing plugin command of type ‘Start Job’
2021-01-27 11:28:52: 0: Plugin rendering frame(s): 1001-1300
2021-01-27 11:28:52: 0: Executing plugin command of type ‘Render Task’
2021-01-27 11:28:52: 0: INFO: Starting Houdini Job
2021-01-27 11:28:52: 0: INFO: Stdout Redirection Enabled: True
2021-01-27 11:28:52: 0: INFO: Stdout Handling Enabled: True
2021-01-27 11:28:52: 0: INFO: Popup Handling Enabled: True
2021-01-27 11:28:52: 0: INFO: QT Popup Handling Enabled: False
2021-01-27 11:28:52: 0: INFO: WindowsForms10.Window.8.app.* Popup Handling Enabled: False
2021-01-27 11:28:52: 0: INFO: Using Process Tree: True
2021-01-27 11:28:52: 0: INFO: Hiding DOS Window: True
2021-01-27 11:28:52: 0: INFO: Creating New Console: False
2021-01-27 11:28:52: 0: INFO: Running as user: julie
2021-01-27 11:28:52: 0: INFO: Executable: “C:\Program Files\Side Effects Software\Houdini 18.0.499\bin\Hython.exe”
2021-01-27 11:28:52: 0: INFO: Argument: “C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601194b9e15609552c7f70b8\hrender_dl.py” -f 1001 1300 1 -o “$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr” -g -d /out/Redshift_ROP_render -tempdir “C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\jobsData\601194b9e15609552c7f70b8\0_tempZnBDe0” -arnoldAbortOnLicenseFail 1 “C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601194b9e15609552c7f70b8/popcorn_balls_wide_render_v001.hiplc”
2021-01-27 11:28:52: 0: INFO: Full Command: “C:\Program Files\Side Effects Software\Houdini 18.0.499\bin\Hython.exe” “C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601194b9e15609552c7f70b8\hrender_dl.py” -f 1001 1300 1 -o “$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr” -g -d /out/Redshift_ROP_render -tempdir “C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\jobsData\601194b9e15609552c7f70b8\0_tempZnBDe0” -arnoldAbortOnLicenseFail 1 “C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601194b9e15609552c7f70b8/popcorn_balls_wide_render_v001.hiplc”
2021-01-27 11:28:52: 0: INFO: Startup Directory: “C:\Program Files\Side Effects Software\Houdini 18.0.499\bin”
2021-01-27 11:28:52: 0: INFO: Process Priority: BelowNormal
2021-01-27 11:28:52: 0: INFO: Process Affinity: default
2021-01-27 11:28:52: 0: INFO: Process is now running
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Redshift for Houdini plugin version 3.0.30 (Sep 26 2020 14:49:18)
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Plugin compile time HDK version: 18.0.499
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Houdini host version: 18.0.499
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Plugin dso/dll and config path: C:/ProgramData/Redshift/Plugins/Houdini/18.0.499/dso
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Core data path: C:\ProgramData\Redshift
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Local data path: C:\ProgramData\Redshift
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Procedurals path: C:\ProgramData\Redshift\Procedurals
2021-01-27 11:29:01: 0: STDOUT: [Redshift] Preferences file path: C:\ProgramData\Redshift\preferences.xml
2021-01-27 11:29:01: 0: STDOUT: [Redshift] License path: C:\ProgramData\Redshift
2021-01-27 11:29:04: 0: STDOUT: Detected Houdini version: (18, 0, 499)
2021-01-27 11:29:04: 0: STDOUT: [‘C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601194b9e15609552c7f70b8\hrender_dl.py’, ‘-f’, ‘1001’, ‘1300’, ‘1’, ‘-o’, ‘$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr’, ‘-g’, ‘-d’, ‘/out/Redshift_ROP_render’, ‘-tempdir’, ‘C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\jobsData\601194b9e15609552c7f70b8\0_tempZnBDe0’, ‘-arnoldAbortOnLicenseFail’, ‘1’, ‘C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601194b9e15609552c7f70b8/popcorn_balls_wide_render_v001.hiplc’]
2021-01-27 11:29:04: 0: STDOUT: Start: 1001
2021-01-27 11:29:04: 0: STDOUT: End: 1300
2021-01-27 11:29:04: 0: STDOUT: Increment: 1
2021-01-27 11:29:04: 0: STDOUT: Ignore Inputs: True
2021-01-27 11:29:04: 0: STDOUT: Output: $JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr
2021-01-27 11:29:04: 0: STDOUT: Driver: /out/Redshift_ROP_render
2021-01-27 11:29:04: 0: STDOUT: Input File: C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601194b9e15609552c7f70b8/popcorn_balls_wide_render_v001.hiplc
2021-01-27 11:39:48: Scheduler Thread - Task “0_1001-1300” has been requeued, because DeletedJob. Cancelling task…
2021-01-27 11:39:48: 0: RenderThread CancelCurrentTask called, will transition from state None to None
2021-01-27 11:39:48: 0: In the process of canceling current task: ignoring exception thrown by PluginLoader
2021-01-27 11:39:48: 0: Executing plugin command of type ‘Cancel Task’
2021-01-27 11:39:48: 0: Done executing plugin command of type ‘Cancel Task’
2021-01-27 11:39:48: 0: Done executing plugin command of type ‘Render Task’
2021-01-27 11:39:48: 0: Error in EndJob: Process was canceled externally.
2021-01-27 11:39:48: at Deadline.Plugins.PluginWrapper.RenderTasks(Task task, String& outMessage, AbortLevel& abortLevel)
2021-01-27 11:39:51: Scheduler Thread - In the process of canceling current tasks: ignoring exception thrown by render thread 0
2021-01-27 11:39:53: Performing pending job scan
2021-01-27 11:39:53: Pending Job Scan - Loading pending and active jobs
2021-01-27 11:39:53: Pending Job Scan - Loaded 0 pending and active jobs in 596.500 μs
2021-01-27 11:39:53: Pending Job Scan - Done.
2021-01-27 11:39:53: Processing Pending Job Events
2021-01-27 11:39:53: Pending Job Events - Checking for pending job events…
2021-01-27 11:39:53: Pending Job Events - Processing 0 job events
2021-01-27 11:39:53: Pending Job Events - No more job events to process
2021-01-27 11:39:53: Pending Job Events - Done.
2021-01-27 11:39:53: Performing repository repair
2021-01-27 11:39:53: Performing Orphaned Task Scan…
2021-01-27 11:39:53: Orphaned Task Scan - Loading rendering jobs
2021-01-27 11:39:53: Orphaned Task Scan - Loaded 0 rendering jobs in 537.800 μs
2021-01-27 11:39:53: Orphaned Task Scan - Done.
2021-01-27 11:39:53: Performing Orphaned Limit Stub Scan…
2021-01-27 11:39:53: Orphaned Limit Stub Scan - Loading limits
2021-01-27 11:39:53: Orphaned Limit Stub Scan - Loaded 0 limits in 627.500 μs
2021-01-27 11:39:53: Orphaned Limit Stub Scan - Done.
2021-01-27 11:39:53: Checking Available Database Connections
2021-01-27 11:39:53: Available Database Connections - Skipping because there are no Low Database Connection notification email addresses set in the Repository Options
2021-01-27 11:39:53: Performing Stalled Worker Scan…
2021-01-27 11:39:53: Stalled Worker Scan - Loading Worker states
2021-01-27 11:39:53: Stalled Worker Scan - Loaded 1 Worker states in 383.100 μs
2021-01-27 11:39:53: Stalled Worker Scan - Scanning Worker states
2021-01-27 11:39:53: Stalled Worker Scan - Cleaned up 0 stalled Worker in 120.700 μs
2021-01-27 11:39:53: Stalled Worker Scan - Done.
2021-01-27 11:39:53: Performing Stalled Pulse Scan…
2021-01-27 11:39:53: Stalled Pulse Scan - Loading pulse states
2021-01-27 11:39:53: Stalled Pulse Scan - Loaded 0 pulse states in 284.700 μs
2021-01-27 11:39:53: Stalled Pulse Scan - Done.
2021-01-27 11:39:53: Performing Stalled Balancer Scan…
2021-01-27 11:39:53: Stalled Balancer Scan - Loading balancer states
2021-01-27 11:39:53: Stalled Balancer Scan - Loaded 0 balancer states in 306.800 μs
2021-01-27 11:39:53: Stalled Balancer Scan - Done.
2021-01-27 11:39:53: Performing Stalled License Forwarder Scan…
2021-01-27 11:39:53: Stalled License Forwarder Scan - Loading forwarder states
2021-01-27 11:39:53: Stalled License Forwarder Scan - Loaded 0 forwarder states in 283.000 μs
2021-01-27 11:39:53: Stalled License Forwarder Scan - Done.
2021-01-27 11:39:53: Performing Stalled Connection Server Scan…
2021-01-27 11:39:53: Stalled Connection Server Scan - Loading Connection Server states
2021-01-27 11:39:53: Stalled Connection Server Scan - Loaded 0 Connection Server states in 276.700 μs
2021-01-27 11:39:53: Stalled Connection Server Scan - Done.
2021-01-27 11:39:53: Triggering repository repair events…
2021-01-27 11:39:53: Performing house cleaning
2021-01-27 11:39:53: Performing Job Cleanup Scan…
2021-01-27 11:39:53: Job Cleanup Scan - Loading completed jobs
2021-01-27 11:39:53: Job Cleanup Scan - Loaded 0 completed and 0 active/pending jobs in 2.255 ms
2021-01-27 11:39:53: Job Cleanup Scan - Done.
2021-01-27 11:39:53: Purging Unsubmitted Jobs
2021-01-27 11:39:53: Unsubmitted Job Scan - Loading unsubmitted jobs
2021-01-27 11:39:53: Unsubmitted Job Scan - Loaded 0 unsubmitted jobs in 519.800 μs
2021-01-27 11:39:53: Unsubmitted Job Scan - Done.
2021-01-27 11:39:53: Purging Deleted Jobs
2021-01-27 11:39:53: Deleted Job Scan - Loading deleted jobs
2021-01-27 11:39:53: Deleted Job Scan - Loaded 2 deleted jobs in 1.494 ms
2021-01-27 11:39:53: Deleted Job Scan - Purged 0 deleted jobs in 2.100 μs
2021-01-27 11:39:53: Deleted Job Scan - Done.
2021-01-27 11:39:53: Purging Obsolete Workers
2021-01-27 11:39:53: Obsolete Worker Scan - Skipping because it is disabled in the Repository Options
2021-01-27 11:39:53: Purging Obsolete AWS Portal Workers
2021-01-27 11:39:53: Obsolete AWS Portal Worker Scan - Loading Worker states
2021-01-27 11:39:53: Obsolete AWS Portal Worker Scan - Loading 1 Worker states in 650.100 μs
2021-01-27 11:39:53: Obsolete AWS Portal Worker Scan - Purging offline/stalled Workers older than 5 days, 0 hours
2021-01-27 11:39:53: Obsolete AWS Portal Worker Scan - Purged 0 obsolete AWS Portal Workers in 1.200 μs
2021-01-27 11:39:53: Obsolete AWS Portal Worker Scan - Done.
2021-01-27 11:39:53: Purging Old Worker Reports
2021-01-27 11:39:53: Worker Report Scan - Loading Worker report collections
2021-01-27 11:39:53: Worker Report Scan - Found 1 report collections and 1 job state objects in 1.495 ms
2021-01-27 11:39:53: Worker Report Scan - Loading Worker IDs
2021-01-27 11:39:53: Worker Report Scan - Loaded 1 Worker IDs in 511.100 μs
2021-01-27 11:39:53: Worker Report Scan - Purged 0 report collections, 0 job state objects, and 0 licensed Worker stubs in 0
2021-01-27 11:39:53: Worker Report Scan - Done.
2021-01-27 11:39:53: Purging Timed Out Workers in Throttle Queue
2021-01-27 11:39:53: Purging Old Statistics
2021-01-27 11:39:53: Old Statistics - Skipping job statistics because the option to purge them is disabled in the Repository Options
2021-01-27 11:39:53: Old Statistics - Purging Worker statistics that are older than Sep 29/20 11:39:53
2021-01-27 11:39:53: Old Statistics - Purged old Worker statistics in 1.096 ms
2021-01-27 11:39:53: Old Statistics - Purging repository statistics that are older than Sep 29/20 11:39:53
2021-01-27 11:39:53: Old Statistics - Purged old repository statistics in 533.900 μs
2021-01-27 11:39:53: Purging expired Remote Command Tokens
2021-01-27 11:39:53: Purging Deleted Document Stubs From Database
2021-01-27 11:39:53: Deleted Document Stubs - Deleting stubs that are older than 3 days
2021-01-27 11:39:53: Deleted Document Stubs - Deleted 0 stubs in 775.900 μs
2021-01-27 11:39:53: Triggering house cleaning events…
2021-01-27 11:39:54: WARNING: Limit ‘601194b9e15609552c7f70b8’ with holder named ‘desktop-5u1fhgc’ could not be found.
2021-01-27 11:40:02: Skipping pending job scan because it is not required at this time
2021-01-27 11:40:02: Skipping repository repair because it is not required at this time
2021-01-27 11:40:02: Skipping house cleaning because it is not required at this time

Sorry for the log formatting…how do you encapsulate a log in the box format like you did?

Here’s a zipped version of it if you prefer.

deadline_log_2021-01-27_114000.zip (3.2 KB)

Use the Preformatted Text instead of Blockquote (or just prefix a line by 4 spaces like below:

2021-01-27 11:39:48: Scheduler Thread - Task “0_1001-1300” has been requeued, because DeletedJob. Cancelling task…

This line claims that the Job was deleted (if you can trust the grammar of that sentence :slight_smile: ), so the Task was cancelled due to that. There are no license errors reported anymore.

Yes i indeed cancelled it after waiting for a good amount of time (10 mins i think?). Is it possible to pinpoint the cause of that stalling?

Oh, right, I see the 10 minutes wait there.

What our support team would typically ask you is “can you run the command line that Deadline is running, but outside of Deadline?”

The full command seems to be

2021-01-27 11:28:52: 0: INFO: Full Command: 

“C:\Program Files\Side Effects Software\Houdini 18.0.499\bin\Hython.exe” “C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601194b9e15609552c7f70b8\hrender_dl.py” -f 1001 1300 1 -o “$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr” -g -d /out/Redshift_ROP_render -tempdir “C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\jobsData\601194b9e15609552c7f70b8\0_tempZnBDe0” -arnoldAbortOnLicenseFail 1 “C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601194b9e15609552c7f70b8/popcorn_balls_wide_render_v001.hiplc”

Most of the arguments are temp. files that are copied over to the local temp folder when the Worker runs, so they are ephemeral. You can try to run that command line right after your Workers has tried to (and stalled).

Now looking under the Repository folder /plugins/Houdini/, the hrender_dl.py is responsible for the prints in the log like

Detected Houdini version: (18, 0, 499)

It then prints the full command line arguments, the Start, End and Increment values, etc.

The last print is of the Input File: value, and then it hangs.

Looking further at the code, around line 418 the script tries to perform

try:
    hou.hipFile.load( inputFile )
except hou.LoadWarning as e:
    print(e)

There is nothing being printed about an error in this call, but I would highly recommend adding some debug prints to the plugin script yourself, like

try:
    print ("About to call hou.hipFile.load()")
    hou.hipFile.load( inputFile )
    print ("hou.hipFile.load() finished without an error.")
except hou.LoadWarning as e:
    print(e)

None of the following prints related to ROP types etc. are ever written to your log, so I have to assume the function to load the input file never finishes. Thus it would be great to know if you ever get “hou.hipFile.load() finished without an error.” printed to the Deadline Worker log once you have added it to the script…

At first sight it is Houdini that is hanging in the loading input file stage, but I could be wrong.

Hi again. So i did as you asked, modified hrender_dl.py to add those extra prints. Here’s the new log:

27 22:05:11:  Scheduler Thread - Job's Limit Groups: 
2021-01-27 22:05:11:  0: Loading Job's Plugin timeout is Disabled
2021-01-27 22:05:11:  0: SandboxedPlugin: Render Job As User disabled, running as current user 'julie'
2021-01-27 22:05:15:  Synchronization time for job files: 74.239 ms
2021-01-27 22:05:15:  0: nogui was not added to the CommandLineParser.
2021-01-27 22:05:15:  Synchronizing Plugin Houdini from C:\DeadlineRepository10\plugins\Houdini took: 0 seconds
2021-01-27 22:05:15:  0: Executing plugin command of type 'Initialize Plugin'
2021-01-27 22:05:15:  0: INFO: Executing plugin script 'C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601229e13c6aeae5741965ad\Houdini.py'
2021-01-27 22:05:15:  0: INFO: About: Houdini Plugin for Deadline
2021-01-27 22:05:15:  0: INFO: The job's environment will be merged with the current environment before rendering
2021-01-27 22:05:15:  0: Done executing plugin command of type 'Initialize Plugin'
2021-01-27 22:05:15:  0: Start Job timeout is disabled.
2021-01-27 22:05:15:  0: Task timeout is disabled.
2021-01-27 22:05:15:  0: Loaded job: popcorn_balls_wide_render_v001 - /out/Redshift_ROP_render (601229e13c6aeae5741965ad)
2021-01-27 22:05:15:  0: Executing plugin command of type 'Start Job'
2021-01-27 22:05:15:  0: DEBUG: S3BackedCache Client is not installed.
2021-01-27 22:05:15:  0: INFO: Executing global asset transfer preload script 'C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601229e13c6aeae5741965ad\GlobalAssetTransferPreLoad.py'
2021-01-27 22:05:15:  0: INFO: Looking for legacy (pre-10.0.26) AWS Portal File Transfer...
2021-01-27 22:05:15:  0: INFO: Looking for legacy (pre-10.0.26) File Transfer controller in C:/Program Files/Thinkbox/S3BackedCache/bin/task.py...
2021-01-27 22:05:15:  0: INFO: Could not find legacy (pre-10.0.26) AWS Portal File Transfer.
2021-01-27 22:05:15:  0: INFO: Legacy (pre-10.0.26) AWS Portal File Transfer is not installed on the system.
2021-01-27 22:05:15:  0: Done executing plugin command of type 'Start Job'
2021-01-27 22:05:15:  0: Plugin rendering frame(s): 1001-1300
2021-01-27 22:05:16:  0: Executing plugin command of type 'Render Task'
2021-01-27 22:05:16:  0: INFO: Starting Houdini Job
2021-01-27 22:05:16:  0: INFO: Stdout Redirection Enabled: True
2021-01-27 22:05:16:  0: INFO: Stdout Handling Enabled: True
2021-01-27 22:05:16:  0: INFO: Popup Handling Enabled: True
2021-01-27 22:05:16:  0: INFO: QT Popup Handling Enabled: False
2021-01-27 22:05:16:  0: INFO: WindowsForms10.Window.8.app.* Popup Handling Enabled: False
2021-01-27 22:05:16:  0: INFO: Using Process Tree: True
2021-01-27 22:05:16:  0: INFO: Hiding DOS Window: True
2021-01-27 22:05:16:  0: INFO: Creating New Console: False
2021-01-27 22:05:16:  0: INFO: Running as user: julie
2021-01-27 22:05:16:  0: INFO: Executable: "C:\Program Files\Side Effects Software\Houdini 18.0.499\bin\Hython.exe"
2021-01-27 22:05:16:  0: INFO: Argument: "C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601229e13c6aeae5741965ad\hrender_dl.py" -f 1001 1300 1 -o "$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr" -g -d /out/Redshift_ROP_render -tempdir "C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\jobsData\601229e13c6aeae5741965ad\0_tempterwD0" -arnoldAbortOnLicenseFail 1 "C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601229e13c6aeae5741965ad/popcorn_balls_wide_render_v001.hiplc"
2021-01-27 22:05:16:  0: INFO: Full Command: "C:\Program Files\Side Effects Software\Houdini 18.0.499\bin\Hython.exe" "C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\plugins\601229e13c6aeae5741965ad\hrender_dl.py" -f 1001 1300 1 -o "$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr" -g -d /out/Redshift_ROP_render -tempdir "C:\ProgramData\Thinkbox\Deadline10\workers\DESKTOP-5U1FHGC\jobsData\601229e13c6aeae5741965ad\0_tempterwD0" -arnoldAbortOnLicenseFail 1 "C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601229e13c6aeae5741965ad/popcorn_balls_wide_render_v001.hiplc"
2021-01-27 22:05:16:  0: INFO: Startup Directory: "C:\Program Files\Side Effects Software\Houdini 18.0.499\bin"
2021-01-27 22:05:16:  0: INFO: Process Priority: BelowNormal
2021-01-27 22:05:16:  0: INFO: Process Affinity: default
2021-01-27 22:05:16:  0: INFO: Process is now running
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Redshift for Houdini plugin version 3.0.30 (Sep 26 2020 14:49:18)
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Plugin compile time HDK version: 18.0.499
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Houdini host version: 18.0.499
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Plugin dso/dll and config path: C:/ProgramData/Redshift/Plugins/Houdini/18.0.499/dso
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Core data path: C:\ProgramData\Redshift
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Local data path: C:\ProgramData\Redshift
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Procedurals path: C:\ProgramData\Redshift\Procedurals
2021-01-27 22:05:25:  0: STDOUT: [Redshift] Preferences file path: C:\ProgramData\Redshift\preferences.xml
2021-01-27 22:05:25:  0: STDOUT: [Redshift] License path: C:\ProgramData\Redshift
2021-01-27 22:05:29:  0: STDOUT: Detected Houdini version: (18, 0, 499)
2021-01-27 22:05:29:  0: STDOUT: ['C:\\ProgramData\\Thinkbox\\Deadline10\\workers\\DESKTOP-5U1FHGC\\plugins\\601229e13c6aeae5741965ad\\hrender_dl.py', '-f', '1001', '1300', '1', '-o', '$JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr', '-g', '-d', '/out/Redshift_ROP_render', '-tempdir', 'C:\\ProgramData\\Thinkbox\\Deadline10\\workers\\DESKTOP-5U1FHGC\\jobsData\\601229e13c6aeae5741965ad\\0_tempterwD0', '-arnoldAbortOnLicenseFail', '1', 'C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601229e13c6aeae5741965ad/popcorn_balls_wide_render_v001.hiplc']
2021-01-27 22:05:29:  0: STDOUT: Start: 1001
2021-01-27 22:05:29:  0: STDOUT: End: 1300
2021-01-27 22:05:29:  0: STDOUT: Increment: 1
2021-01-27 22:05:29:  0: STDOUT: Ignore Inputs: True
2021-01-27 22:05:29:  0: STDOUT: Output: $JOB/render/popcorn/popcorn_wide/popcorn_wide.$F4.exr
2021-01-27 22:05:29:  0: STDOUT: Driver: /out/Redshift_ROP_render
2021-01-27 22:05:29:  0: STDOUT: Input File: C:/ProgramData/Thinkbox/Deadline10/workers/DESKTOP-5U1FHGC/jobsData/601229e13c6aeae5741965ad/popcorn_balls_wide_render_v001.hiplc
2021-01-27 22:05:29:  0: STDOUT: About to call hou.hipFile.load()
2021-01-27 22:10:51:  Scheduler Thread - Task "0_1001-1300" has been requeued, because DeletedJob. Cancelling task...
2021-01-27 22:10:51:  0: RenderThread CancelCurrentTask called, will transition from state None to None
2021-01-27 22:10:51:  0: In the process of canceling current task: ignoring exception thrown by PluginLoader
2021-01-27 22:10:51:  0: Executing plugin command of type 'Cancel Task'
2021-01-27 22:10:51:  0: Done executing plugin command of type 'Cancel Task'
2021-01-27 22:10:51:  0: Done executing plugin command of type 'Render Task'
2021-01-27 22:10:51:  0: Executing plugin command of type 'End Job'
2021-01-27 22:10:51:  0: Done executing plugin command of type 'End Job'
2021-01-27 22:10:51:  0: Error in EndJob: Process was canceled externally.
2021-01-27 22:10:51:     at Deadline.Plugins.PluginWrapper.RenderTasks(Task task, String& outMessage, AbortLevel& abortLevel)
2021-01-27 22:10:53:  Scheduler Thread - In the process of canceling current tasks: ignoring exception thrown by render thread 0
2021-01-27 22:10:56:  Performing pending job scan
2021-01-27 22:10:56:      Pending Job Scan - Loading pending and active jobs
2021-01-27 22:10:56:      Pending Job Scan - Loaded 0 pending and active jobs in 995.100 μs
2021-01-27 22:10:56:      Pending Job Scan - Done.
2021-01-27 22:10:56:  Processing Pending Job Events
2021-01-27 22:10:56:      Pending Job Events - Checking for pending job events...
2021-01-27 22:10:56:      Pending Job Events - Processing 0 job events
2021-01-27 22:10:56:      Pending Job Events - No more job events to process
2021-01-27 22:10:56:      Pending Job Events - Done.
2021-01-27 22:10:56:  Performing repository repair

Can that help you narrrow down the diagnostic? I don’t have the IT skills to investigate further, and this is my first experience with Deadline. Any help is greatly appreciated.

You mentionned temp files being copied…why would that be necessary if all i am doing is on the same local machine?(worker + repository)

Ok, this confirms that Houdini never finishes loading the scene.

When the Worker dequeues a job, it checks the Plugin= key value and copies the Plugin’s script file over to the local temp folder for the Job.
Then, if the SceneFile was sent as an auxiliary file to the Repository during submission (there is an option in most submitters whether to send the Scene with the Job or not), the scene file is also copied from the Job’s folder in the Repository to the local temp folder.
If the option to send the scene with the Job was disabled, then the Job metadata would contain a key SceneFile= with a value of the network path where the scene to load resides, and no copy to the local temp. folder will be made.

From looking at the log, I cannot tell why it is not starting. I would highly recommend emailing support@thinkboxsoftware.com to create an actual support ticket with our Support team. Please include a link to this forum thread so they can see what you have tried so far. This forum is not the official support channel, it is mostly for peer-to-peer communication, which is good in case someone else has seen this problem and wants to chime in…

I am neither a support engineer, nor a Houdini guru, so I have a limited ability to help beyond this point.

Ok! Well so far you’ve helped me a lot.

So if i understand correctly, as long as i am on the same repository/worker(single machine) i have no reason to check “Submit scene”? If that is the case i could easily workaround this issue for now, if the problem resides in the copying part, by unchecking “Submit scene”?

If there’s a problem with copying that would be strange since everything is on the same machine and the scene barely weights a couple hundreds MBs.

In the event of using a NAS in the future, i would have to find the underlying cause to this problem? Therefore, i will contact support as you suggested.

Not quite.

If the scene is located on a network path (e.g. a NAS drive) and both the submission machine and the render node running the Worker (which could be the same, or different computers) can access it either through a UNC path or a Windows Mapped Drive, there is no reason to submit the scene as auxiliary file with the job.

The main reasons for having a “submit scene with the job” option is when the scene file is stored on the local C: drive of the submitting machine and the Repository/Workers are elsewhere and cannot access that C: drive, or when you want to include the scene file in the job data for archiving purposes (e.g. if you Archive the Job which creates a ZIP file you can de-archive years later, and you want the scene to be part of that as the network copy could be gone or modified by then).

I don’t think the copying of the file to the temp. folder has anything to do with your problem. But you are welcome to try submitting the job without including the scene file to see if it changes anything. I suspect that there is an issue with the file running in this context, and I don’t think Deadline is in the core of the problem.

Can you try creating a different, SUPER SIMPLE scene with one Sphere on a Plane, one Light, one Camera, no surface shaders, no animation. Set it to render in Redshift as usual, and submit. Does it still fail? At this point we are debugging the scene itself, not your Deadline setup. Knowing whether the issue is scene-specific or general would be a good next step…

Yes i did try with a simple test scene and everything was fine.

Is there a guide or a step-by-step process that could help me debug the scene, since the deadline logs seems to be out of the option for this…? Once again, the scene renders fine by itself inside Houdini.

I’ll try deactivating “Submit scene” to see if i’m lucky even though it isn’t really logical as you say! :slight_smile:

The best way to debug this would be to try to do a command line rendering with Houdini OUTSIDE of Deadline. Does it render properly outside of an interactive Houdini session?

Then start removing components of the scene - e.g. if there is a particle system, delete it, resubmit, see if anything changes. Still failing? Delete objects one by one. Or better, delete HALF of the content - if that changes things, you know that one of the deleted components in the deleted half is the culprit. Go back to the original scene and delete 1/4 of the scene (half of the half you deleted before). Keep on subdividing until you zero in on the offending content.

Since the issue could be Redshift-related, you can also try modifying the scene to render in Mantra to see if it is really something about the Renderer, or a more general problem.

Normally I would expect the loading of a scene with a problem to lead to some kind of error message, or even hard crash. Since this is just hanging, I am not sure what else we can do to pinpoint but tweak the content until it works, and then reason from there what the issue was…

Hi Bobo!

I made some progress trying to figure out my issues.

First, i’m not exactly sure why, but reducing the frame per task to 12 rather than the total amount of the job helped a lot. I can now get the rendering to start.

Altough it is technically working, i noticed another strange issue, which is probably why it seemed to “freeze” in the first place, forcing me to suspend and delete jobs.

When Deadline works on a frame or completes one for a current task, it keeps showing 0% progress in the task progress bar. I guess that’s why when i submitted a 300+ frames per task job, i thought Deadline wasn’t doing anything at all. Even though it was actually the case until i started lowering the frames per task, i can start to get a clue why. Is this normal behavior?

I know for a fact that when i had a high number of frames per task the render wasn’t starting at all first by checking the logs with you but also because once a frame begins rendering, Redshift creates a “.lock” file of the current frame file. (i.e.: <render.####.exr.lock>). When i was submitting lots of frames per task it would never gets to this point. Now it does but as a frame progresses or completes, the progress bar still shows 0%.

2021-01-31 09:45:26:  0: STDOUT: [Redshift] Redshift for Houdini plugin version 3.0.30 (Sep 26 2020 14:49:18)
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Plugin compile time HDK version: 18.0.499
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Houdini host version: 18.0.499
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Plugin dso/dll and config path: C:/ProgramData/Redshift/Plugins/Houdini/18.0.499/dso
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Core data path: C:\ProgramData\Redshift
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Local data path: C:\ProgramData\Redshift
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Procedurals path: C:\ProgramData\Redshift\Procedurals
2021-01-31 09:45:26:  0: STDOUT: [Redshift] Preferences file path: C:\ProgramData\Redshift\preferences.xml
2021-01-31 09:45:26:  0: STDOUT: [Redshift] License path: C:\ProgramData\Redshift
2021-01-31 09:45:28:  0: STDOUT: Detected Houdini version: (18, 0, 499)
2021-01-31 09:45:28:  0: STDOUT: ['C:\\ProgramData\\Thinkbox\\Deadline10\\workers\\DESKTOP-5U1FHGC\\plugins\\6016bdafb1bd1194507526a4\\hrender_dl.py', '-f', '888', '899', '1', '-o', '$JOB/render/seq_1/seq_1.$F4.exr', '-g', '-d', '/out/Redshift_ROP_render', '-tempdir', 'C:\\ProgramData\\Thinkbox\\Deadline10\\workers\\DESKTOP-5U1FHGC\\jobsData\\6016bdafb1bd1194507526a4\\0_tempSzNZs0', '-arnoldAbortOnLicenseFail', '1', 'D:/projects/Invicare/hip/render/seq_1_render_v001.hiplc']
2021-01-31 09:45:28:  0: STDOUT: Start: 888
2021-01-31 09:45:28:  0: STDOUT: End: 899
2021-01-31 09:45:28:  0: STDOUT: Increment: 1
2021-01-31 09:45:28:  0: STDOUT: Ignore Inputs: True
2021-01-31 09:45:28:  0: STDOUT: Output: $JOB/render/seq_1/seq_1.$F4.exr
2021-01-31 09:45:28:  0: STDOUT: Driver: /out/Redshift_ROP_render
2021-01-31 09:45:28:  0: STDOUT: Input File: D:/projects/Invicare/hip/render/seq_1_render_v001.hiplc
2021-01-31 09:45:28:  0: STDOUT: About to call hou.hipFile.load()
2021-01-31 09:45:32:  0: STDOUT: hou.hipFile.load() finished without an error.
2021-01-31 09:45:32:  0: STDOUT: Begin Path Mapping
2021-01-31 09:45:32:  0: STDOUT: End Path Mapping
2021-01-31 09:45:32:  0: STDOUT: ROP type: Redshift_ROP
2021-01-31 09:45:32:  0: STDOUT: Rendering frame 888 to 899
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Scene 'seq_1_render_v001', ROP node 'Redshift_ROP_render' render started for 12 frames. Time from 36.958333 to 37.416667
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Houdini/FX or Indie license found, no render restrictions
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Closing the RS scene
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Initializing the render session
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Textures cache size (MB): 32768
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Textures cache path: %LOCALAPPDATA%\Redshift\Cache
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Devices enabled: 0:GeForce GTX 1080 Ti,
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Redshift Initialized
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Version: 3.0.30, Sep 26 2020
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Windows Platform (Windows 10 Pro)
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Release Build
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Number of CPU HW threads: 12
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	CPU speed: 3.21 GHz
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Total system memory: 23.99 GB
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	TDR delay: 8s
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Driver version: 457.30
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Current working dir: C:\Program Files\Side Effects Software\Houdini 18.0.499\bin
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Creating CUDA contexts
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	CUDA init ok
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Ordinals: { 0 }
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Initializing GPUComputing module (CUDA). Ordinal 0
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	CUDA Driver Version: 11010
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	CUDA API Version: 11000
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Device 1/1 : GeForce GTX 1080 Ti 
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Compute capability: 6.1
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Num multiprocessors: 28
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	PCI busID: 4, deviceID: 0, domainID: 0
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Theoretical memory bandwidth: 484.440002 GB/Sec
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Measured PCIe bandwidth (pinned CPU->GPU): 5.552693 GB/s
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Measured PCIe bandwidth (pinned GPU->CPU): 6.205368 GB/s
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Measured PCIe bandwidth (paged CPU->GPU): 2.885061 GB/s
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Measured PCIe bandwidth (paged GPU->CPU): 3.044155 GB/s
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Estimated GPU->CPU latency (0): 0.043044 ms
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Estimated GPU->CPU latency (1): 0.042438 ms
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Estimated GPU->CPU latency (2): 0.044547 ms
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Estimated GPU->CPU latency (3): 0.043237 ms
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	New CUDA context created
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Available memory: 10096.2500 MB out of 11264.0000 MB
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Loading Redshift procedural extensions...
2021-01-31 09:56:19:  0: STDOUT: [Redshift] 	Done!
2021-01-31 09:56:19:  0: STDOUT: [Redshift] RS render object initialized successfully
2021-01-31 09:56:19:  0: STDOUT: [Redshift] ROP node 'Redshift_ROP_render' rendering frame 888 (time: 36.958333)
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Loading RS rendering options
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Initializing the AOVs system
2021-01-31 09:56:19:  0: STDOUT: [Redshift] AOVs initialized sucessfully
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Loading the RS scene
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Loading the camera data
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Loading the scene OBJ mesh nodes data
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Processing the objects to be extracted list
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Building the additional objects extraction list from the instancer objects
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Disabled objects to be extracted and used as base instance objects: 0
2021-01-31 09:56:19:  0: STDOUT: [Redshift] =================================================================================================
2021-01-31 09:56:19:  0: STDOUT: [Redshift] Rendering frame 888...

I made some progress again.

I realized some of the extra long hanging times were due to a dependency in my scene network. I don’t know if you are familiar with Houdini, but within the scene you have file caches nodes. Well, even it those have an option to “Load from Disk”, therefore bypassing any inputs, my guess would be that Deadline doesn’t recognize this “bypass” feature and it was most likely trying to recook the whole setup above the file cache node. Disconnecting the file cache input (which had no use in my case except for a potential recook of the assets) or completly deleting all the nodes upstream of the final render caches solved the hanging times problems.

My previous reply is still valid though, Deadline doesn’t show the progress per task nor per frame in the monitor. But anyone using Houdini and Redshift : be aware, Deadline WILL read everything above your File Caches even if the “Load from disk” option is enabled. :slight_smile:

Have a good day!

The main thing you need to keep in mind is that Deadline does not render the job, Houdini does. With all rendering applications, Deadline simply tells the application what to do, and then sits around and watches for output (STDOUT from the process, or a returns value when running a command line). So it is a bit unfair to blame Deadline for most of the things you wrote, as it tells Houdini to load the scene and render whatever frames you requested. It then has Output Handlers that perform RegEx on the STDOUT to figure out if a frame progress was reported, watch out for known and unknown Error messages in the STDOUT, and react to them.

By default, the Frames Per Task should be 1. The Task progress will only increase if the renderer is actually outputting progress information to STDOUT.

For a non-Houdini example, when V-Ray renders in Maya, you see something like this:

2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:07] V-Ray: Additional progressive sampling filter buffers memory usage: 176.81 MB
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m 56.9s): 1 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray: Noise threshold goal lowered to 0.142857
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m 18.6s): 4 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray: Noise threshold goal lowered to 0.090909
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray: Noise threshold goal lowered to 0.052632
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m 17.3s): 5 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray: Noise threshold goal lowered to 0.028571
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m 15.1s): 6 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray: Noise threshold goal lowered to 0.014925
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m 11.1s): 8 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray: Noise threshold goal set: 0.010000
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  5.6s): 16 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  1.5s): 44 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  1.1s): 53 % completed
2019-12-03 16:00:08:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.8s): 62 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.7s): 67 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.6s): 70 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.5s): 74 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.4s): 80 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.3s): 84 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.2s): 86 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:08] V-Ray:  (Estimated time remaining:  0h  0m  0.2s): 88 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.1s): 91 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.1s): 93 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.1s): 95 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.1s): 96 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.0s): 97 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.0s): 98 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.0s): 99 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray:  (Estimated time remaining:  0h  0m  0.0s): 100 % completed
2019-12-03 16:00:09:  0: STDOUT: [2019/Dec/3|16:00:09] V-Ray: Progressive rendering finished: specified noise level reached

In the MayaBatch.py integration script, there are several STDOUT Handlers that are looking for the % display to update the progress, catch errors etc.

    self.AddStdoutHandlerCallback( "V-Ray error: .*" ).HandleCallback += self.HandleFatalError
    self.AddStdoutHandlerCallback( "V-Ray: Building light cache*" ).HandleCallback += self.HandleVrayMessage
    self.AddStdoutHandlerCallback( "V-Ray: Prepass ([0-9]+) of ([0-9]+)*" ).HandleCallback += self.HandleVrayMessage
    self.AddStdoutHandlerCallback( "V-Ray: Rendering image*" ).HandleCallback += self.HandleVrayMessage
    self.AddStdoutHandlerCallback( "V-Ray: +([0-9]+)%" ).HandleCallback += self.HandleVrayProgress
    self.AddStdoutHandlerCallback( "V-Ray: +([0-9]+) %" ).HandleCallback += self.HandleVrayProgress
    self.AddStdoutHandlerCallback( "([0-9]+) % completed" ).HandleCallback += self.HandleVrayProgress
    self.AddStdoutHandlerCallback( "V-Ray: Total frame time" ).HandleCallback += self.HandleVrayFrameComplete
    
    self.AddStdoutHandlerCallback( "V-Ray: Updating frame at time ([0-9]+)" ).HandleCallback += self.HandleVrayExportProgress
    self.AddStdoutHandlerCallback( "V-Ray: Render complete" ).HandleCallback += self.HandleVrayExportComplete
    self.AddStdoutHandlerCallback( ".*V-Ray warning.* The file path for .* cannot be created.*" ).HandleCallback += self.HandleFatalError

We have similar ones registered for Redshift there:

    if self.Renderer == "redshift":
        self.AddStdoutHandlerCallback( r"Frame rendering aborted." ).HandleCallback += self.HandleFatalError
        self.AddStdoutHandlerCallback( r"Rendering was internally aborted" ).HandleCallback += self.HandleFatalError
        self.AddStdoutHandlerCallback( r'Cannot find procedure "rsPreference"' ).HandleCallback += self.HandleFatalError
        self.AddStdoutHandlerCallback( "Rendering frame \\d+ \\((\\d+)/(\\d+)\\)" ).HandleCallback += self.HandleRedshiftNewFrameProgress
        self.AddStdoutHandlerCallback( "Block (\\d+)/(\\d+) .+ rendered" ).HandleCallback += self.HandleRedshiftBlockRendered

When Redshift renders in Maya, it outputs something like this:

2019-11-13 18:30:40:  0: STDOUT: Rendering blocks... (resolution: 960x540, block size: 128, unified minmax: [4,16])
2019-11-13 18:30:40:  0: STDOUT: 	Allocating VRAM for device 0 (Tesla M60)
2019-11-13 18:30:40:  0: STDOUT: 		Redshift can use up to 6399 MB
2019-11-13 18:30:40:  0: STDOUT: 		Fixed: 0 MB
2019-11-13 18:30:40:  0: STDOUT: 		Geo: 402 MB, Tex: 39 MB, Rays: 1840 MB, NRPR: 1142240
2019-11-13 18:30:40:  0: STDOUT: 		Done! (  6ms). CUDA reported free mem: 728 MB
2019-11-13 18:30:40:  0: STDOUT: 	Block 1/40 (3,1) rendered by GPU 0 in 551ms
2019-11-13 18:30:40:  0: STDOUT: 	Block 2/40 (4,1) rendered by GPU 0 in 286ms
2019-11-13 18:30:41:  0: STDOUT: 	Block 3/40 (4,2) rendered by GPU 0 in 401ms
2019-11-13 18:30:41:  0: STDOUT: 	Block 4/40 (3,2) rendered by GPU 0 in 380ms
2019-11-13 18:30:42:  0: STDOUT: 	Block 5/40 (2,3) rendered by GPU 0 in 306ms
2019-11-13 18:30:42:  0: STDOUT: 	Block 6/40 (2,2) rendered by GPU 0 in 299ms
2019-11-13 18:30:42:  0: STDOUT: 	Block 7/40 (2,1) rendered by GPU 0 in 264ms
2019-11-13 18:30:42:  0: STDOUT: 	Block 8/40 (2,0) rendered by GPU 0 in 250ms
2019-11-13 18:30:43:  0: STDOUT: 	Block 9/40 (3,0) rendered by GPU 0 in 263ms
2019-11-13 18:30:43:  0: STDOUT: 	Block 10/40 (4,0) rendered by GPU 0 in 169ms
2019-11-13 18:30:43:  0: STDOUT: 	Block 11/40 (5,0) rendered by GPU 0 in 156ms
2019-11-13 18:30:43:  0: STDOUT: 	Block 12/40 (5,1) rendered by GPU 0 in 236ms
2019-11-13 18:30:44:  0: STDOUT: 	Block 13/40 (5,2) rendered by GPU 0 in 354ms
2019-11-13 18:30:44:  0: STDOUT: 	Block 14/40 (5,3) rendered by GPU 0 in 398ms
2019-11-13 18:30:44:  0: STDOUT: 	Block 15/40 (4,3) rendered by GPU 0 in 396ms
2019-11-13 18:30:45:  0: STDOUT: 	Block 16/40 (3,3) rendered by GPU 0 in 459ms
2019-11-13 18:30:45:  0: STDOUT: 	Block 17/40 (1,4) rendered by GPU 0 in 83ms
2019-11-13 18:30:45:  0: STDOUT: 	Block 18/40 (1,3) rendered by GPU 0 in 241ms
2019-11-13 18:30:45:  0: STDOUT: 	Block 19/40 (1,2) rendered by GPU 0 in 378ms
2019-11-13 18:30:46:  0: STDOUT: 	Block 20/40 (1,1) rendered by GPU 0 in 284ms
2019-11-13 18:30:46:  0: STDOUT: 	Block 21/40 (1,0) rendered by GPU 0 in 263ms
2019-11-13 18:30:46:  0: STDOUT: 	Block 22/40 (6,0) rendered by GPU 0 in 222ms
2019-11-13 18:30:47:  0: STDOUT: 	Block 23/40 (6,1) rendered by GPU 0 in 244ms
2019-11-13 18:30:47:  0: STDOUT: 	Block 24/40 (6,2) rendered by GPU 0 in 374ms
2019-11-13 18:30:47:  0: STDOUT: 	Block 25/40 (6,3) rendered by GPU 0 in 322ms
2019-11-13 18:30:47:  0: STDOUT: 	Block 26/40 (6,4) rendered by GPU 0 in 102ms
2019-11-13 18:30:47:  0: STDOUT: 	Block 27/40 (5,4) rendered by GPU 0 in 118ms
2019-11-13 18:30:48:  0: STDOUT: 	Block 28/40 (4,4) rendered by GPU 0 in 135ms
2019-11-13 18:30:48:  0: STDOUT: 	Block 29/40 (3,4) rendered by GPU 0 in 171ms
2019-11-13 18:30:48:  0: STDOUT: 	Block 30/40 (2,4) rendered by GPU 0 in 102ms
2019-11-13 18:30:48:  0: STDOUT: 	Block 31/40 (0,4) rendered by GPU 0 in 110ms
2019-11-13 18:30:48:  0: STDOUT: 	Block 32/40 (0,3) rendered by GPU 0 in 322ms
2019-11-13 18:30:49:  0: STDOUT: 	Block 33/40 (0,2) rendered by GPU 0 in 350ms
2019-11-13 18:30:49:  0: STDOUT: 	Block 34/40 (0,1) rendered by GPU 0 in 264ms
2019-11-13 18:30:49:  0: STDOUT: 	Block 35/40 (0,0) rendered by GPU 0 in 200ms
2019-11-13 18:30:49:  0: STDOUT: 	Block 36/40 (7,0) rendered by GPU 0 in 132ms
2019-11-13 18:30:49:  0: STDOUT: 	Block 37/40 (7,1) rendered by GPU 0 in 138ms
2019-11-13 18:30:50:  0: STDOUT: 	Block 38/40 (7,2) rendered by GPU 0 in 210ms
2019-11-13 18:30:50:  0: STDOUT: 	Block 39/40 (7,3) rendered by GPU 0 in 183ms
2019-11-13 18:30:50:  0: STDOUT: 	Block 40/40 (7,4) rendered by GPU 0 in 67ms
2019-11-13 18:30:50:  0: STDOUT: 	Processing blocks...
2019-11-13 18:30:50:  0: STDOUT: 	Time to render 40 blocks: 10.3s

In the case of Houdini, it is possible that Redshift does not write anything usable to the log to handle as progress. Here are the Houdini STDOUT Handlers found in Repo/plugins/Houdini/Houdini.py:

    self.AddStdoutHandlerCallback( "(Couldn't find renderer.*)" ).HandleCallback += self.HandleStdoutRenderer
    self.AddStdoutHandlerCallback( "(Error: Unknown option:.*)" ).HandleCallback += self.HandleStdoutUnknown
    self.AddStdoutHandlerCallback( "(Error: .*)" ).HandleCallback += self.HandleStdoutError
    self.AddStdoutHandlerCallback(r"(ERROR\s*\|.*)").HandleCallback += self.HandleStdoutError #Arnold errors
    self.AddStdoutHandlerCallback(r"\[Error\].*").HandleCallback += self.HandleStdoutError
    self.AddStdoutHandlerCallback( ".*(Redshift cannot operate with less than 128MB of free VRAM).*" ).HandleCallback += self.HandleStdoutError
    self.AddStdoutHandlerCallback( ".*(No licenses could be found to run this application).*" ).HandleCallback += self.HandleStdoutLicense
    self.AddStdoutHandlerCallback( ".*ALF_PROGRESS ([0-9]+)%.*" ).HandleCallback += self.HandleStdoutFrameProgress
    self.AddStdoutHandlerCallback( ".*Render Time:.*" ).HandleCallback += self.HandleStdoutFrameComplete
    self.AddStdoutHandlerCallback( ".*Finished Rendering.*" ).HandleCallback += self.HandleStdoutDoneRender
    self.AddStdoutHandlerCallback( ".*ROP type: (.*)" ).HandleCallback += self.SetRopType
    self.AddStdoutHandlerCallback( ".*?(\d+)% done.*" ).HandleCallback += self.HandleStdoutFrameProgress
    self.AddStdoutHandlerCallback( "\[render progress\] ---[ ]+(\d+) percent" ).HandleCallback += self.HandleStdoutFrameProgress
    self.AddStdoutHandlerCallback( "(License error: No license found)").HandleCallback += self.HandleStdoutLicense

The two handlers for FrameProgress are

    self.AddStdoutHandlerCallback( ".*?(\d+)% done.*" ).HandleCallback += self.HandleStdoutFrameProgress
    self.AddStdoutHandlerCallback( "\[render progress\] ---[ ]+(\d+) percent" ).HandleCallback += self.HandleStdoutFrameProgress

Your log does not show anything like that, so the Tasks would not show any progress.
However, if you have 100 frames and 1 Frame Per Task, you will get a Job Progress advancing from 0 to 100% as each Task finishes.

It only makes sense to combine multiple frames in a single Task if there is an overhead in relaunching the application between Tasks. For example, AfterEffects notoriously performs better if your Frames Per Task value is high (10 or so). But for applications that keep the scene in memory and just switch between frames, it is probably ok to keep the value at 1.

I cannot comment on why the caches were being re-cooked - I wonder if the file cache paths were invalid in the context of the render node, and forced Houdini to recompute?

1 Like

I didn’t know that! I thought Deadline was re-lauching Houdini/Redshift for each new tasks. Thanks for the heads-up.

The file paths were valid, simply disconnecting their inputs worked fine, their paths remain the same.

Here’s an example of how it looks like inside Houdini if you’re curious :

Sorry for the confusion, the “keep in memory” option applies to applications that launch a Managed Process (e.g. MayaBatch and 3ds Max). We don’t do that with Houdini as far as I can tell.

Privacy | Site terms | Cookie preferences