Applets and Entry Points
About Applets and Entry Points
Applets
Applets are executable data objects. Like other data objects, applets exist inside projects and can be cloned between projects. By default, applets also have VIEW permissions into the projects from which they are run. Applets can be used to create private, customized scripts for specialized needs, or for testing and developing more general apps that may be of interest to the community at large.
Applets must be created with a run specification so that the system knows how to run it, but they can be created with or without input and output specifications. Providing I/O specifications is encouraged because it allows the system to validate the input arguments when an applet is launched, and validate the produced output when an applet is finished. Also, when an applet has an I/O specification, the DNAnexus website can automatically render a configuration form for users who want to launch the applet (since the system is aware of the names and types of inputs that the applet requires). Finally, other developers who want to invoke the applet from their applets can look at the I/O specification for help on how to launch the applet programmatically (i.e., what to give as inputs, or what to expect from the outputs).
If I/O specifications are not provided, users can launch the applet with any input (which the system will not validate, except for DNAnexus links and job-based object references), and the applet can produce any output. Developers of such applets are responsible for documenting what their applet expects in the input and what outputs their applet can produce. Likewise, they are responsible for providing a user interface for someone to configure the applet’s input on the DNAnexus’ website. This allows for building powerful applets whose input or output might be hard to formally describe, perhaps because it is variable or polymorphic, or would otherwise be constrained by what DNAnexus allows in its input/output definition grammar.
Running an applet: Project context
Running an applet is slightly different depending on whether the applet is launched by a user from outside of the DNAnexus platform, or by another running job.
Launching an applet outside of the platform requires associating a project with it. As mentioned earlier, this project context is important for the following reasons:
Any charges related to the execution of the applet are associated with that project.
Jobs (such as the job created by launching the applet, as well as any other jobs created by the applet itself while running) will be given VIEW access to that project.
Any objects output by the applet will be placed into that project.
When launching an applet from another running job, this parent job is already associated with a project. This project is carried forward to the launched master job; more specifically:
Any charges related to the execution of the master job are associated with that project.
Jobs (such as the job created by launching the master job, as well as any other descendant jobs of the master job) will be given VIEW access to that project.
Any objects output by the master job will be placed into the workspace of the parent job.
When running an applet, the applet object does not need to be inside the project context. Whoever runs the applet needs to have VIEW access to the applet object and CONTRIBUTE access to the project context. However, the generated job will have VIEW access to the project context and CONTRIBUTE access to the workspace only, and the system "acts as that job" for the purposes of accessing any references to file objects mandated by the applet spec, (such as assets mentioned under bundledDepends
, etc.); therefore, these objects need to be accessible by the job, and this can only be done if they happen to be in the project context or in a public project. Thus, if an applet object’s dependencies are not in the project context, then running it will eventually fail unless it is in a public project, because the system will not be able to fetch associated files. As such, when an applet is cloned, all objects linked to in the bundledDepends
of its run specification will also be cloned with it.
Running Entry Points
Any job can issue the /job/new API call to create a subjob that will run a particular entry point. Running an entry point is like calling a function in your code that will then execute on its own worker in the cloud. Because each job runs on a separate worker, any communication between jobs under the same master job must be via their input, output, and/or stored data objects in their shared temporary workspace. When an applet is first run, it automatically runs the entry point called "main". See Code Interpreters for more information on how to create entry points in your code given the interpreter of your choice.
Requesting Instance Types
Different instance types and other system resources can be requested for different entry points of an applet using the systemRequirements
field. This can be specified in an applet's run specification and partially (or fully) overridden at runtime using the systemRequirementsByExecutable
or systemRequirements
runtime arguments. The systemRequirements
argument used at runtime overrides uses the same syntax as in the run specification for applets and apps. systemRequirementsByExecutable
argument syntax is described below.
When computing the effective value of systemRequirements
, the keys are retrieved in the following order (using the first one found):
The
systemRequirementsByExecutable
values provided at launch of the job's ancestors and the job itself. Note that child job'ssystemRequirementsByExecutable
are merged into parent'ssystemRequirementsByExecutable
without overriding parental values.The
systemRequirements
value requested for the job's master job.If the job was run via /app-xxxx/run or /applet-xxxx/run, then this is the value of
systemRequirements
in the API call.If the job was run as a stage in a workflow, then the effective value is a combination of fields provided in /workflow-xxxx/run and of any default values stored in the workflow itself.
The
systemRequirements
field provided to the /job/new method.The
runSpec.systemRequirements
field in the applet or app.If none of these values are present (or contain a relevant entry point), system-wide defaults are used.
Using the systemRequirements Argument
The "*" entry point in systemRequirements
argument refers to all other entry points that are not already named in systemRequirements
. Note that if an applet has a run specification that includes an instance type X for entry point "main", and the applet is run with systemRequirements
specifying that "*" should have a different instance type Y, the "main" entry point will be run on instance type Y because the "main" key has been implicitly overridden by the "*" entry point specified at runtime.
Using the systemRequirementsByExecutable Argument
The systemRequirementsByExecutable
argument to /app-xxxx/run, /applet-xxxx/run, /workflow-xxxx/run, /globalworkflow-xxxx/run and /job/new allows users to specify instanceType
, fpgaDriver
, nvidiaDriver
and clusterSpec.initialInstanceCount
fields for all jobs in the resulting execution tree, configurable by executable ID and then by entry point. If present, it includes at least one of the following key-value pairs:
key executable id (
applet-xxxx
orapp-xxxx
)or"*"
to indicate all executables not explicitly specified in this mapping.value mapping System requirement for the corresponding executable. It includes at least one of the following key-value pairs:
key Entry point name or
"*"
to indicate all entry points not explicitly specified in this mapping.value mapping Requested resources for the entry point:
instanceType
string (optional): A string specifying the instance type used to execute jobs running the specified entry point of the specified executable. See Instance Types for a list of possible values.clusterSpec
mapping (optional). If specified, must containinitialInstanceCount
positive integer: The number of nodes in the cluster including the driver node. Value of 1 indicates a cluster with no worker nodes.
fpgaDriver
string (optional): Specifies the FPGA driver that will be installed on the FPGA-enabled cloud host instance prior to app's code execution. Accepted values are "edico-1.4.2" (installed on FPGA-enabled instances by default) , "edico-1.4.5", and "edico-1.4.7".nvidiaDriver
string (optional): Specifies the NVIDIA driver to install on the GPU-enabled cloud host instance prior to app's code execution. Accepted values are"R470"
(470.256.02, default value) and"R535"
(535.183.01).
The systemRequirementsByExecutable
argument applies to the entire resulting job execution tree and merges with all downstream runtime inputs, with explicit upstream inputs taking precedence.
Second-level systemRequirementsByExecutable
keys (instanceType
, fpgaDriver
, nvidiaDriver,
clusterSpec
) are resolved independently from each other.
For example, calling
forces mem2_ssd1_v2_x2
for all jobs running all entry points of applet-1
, but allows overrides of fpgaDriver
, nvidiaDriver
and clusterSpec.initialInstanceCount
for applet-1
via downstream specification of systemRequirementsByExecutable
, systemRequirements
/ stageSystemRequirements
specified at runtime, system requirements embedded in workflow stages, and systemRequirements
supplied to /executable/new
.
systemRequirementsByExecutable
specified at the root level can not be overridden by later specifications within the execution tree. Children's systemRequirementsByExecutable
are merged into parent's systemRequirementsByExecutable
without overriding parental values. Specification of "*" at a higher level (both for executable key and entry point key) precludes overrides at lower levels of execution subtree. "*" at either executable or entry point level refers to "everything else" and does not take precedence over any sibling fields in the systemRequirementsByExecutable
object that mention specific executable IDs or entry points, but precluding overrides at lower level:
If a parent has "*", the child's specification of any key in that hash is ignored because otherwise the child's key (either * or some other specific key) would override the parental specification.
If a parent does not have "*", the child's specification of "*" will apply to all keys not specified at the parent level or by an explicit non-* key inside the child. The child's non-* keys that are already mentioned in the parent are ignored, while non-* keys that are not mentioned in any of the parents will take effect.
Invoking /executable-xxxx/run with "detach": true
argument will cause the new detached execution tree to disregard detachedFrom
's systemRequirementsByExecutable
. In other words, a detached execution tree is treated like a new root execution: detached execution tree does not inherit detachedFrom
's systemRequirementsByExecutable
setting, but honors an optional systemRequirementsByExecutable
S argument if it is supplied to /executable-xxxx/run(detached=true, systemRequirementsByExecutable=S)
call that created the detached execution tree.
Examples:
Force all jobs in the execution tree to use
mem2_ssd1_v2_x2
. Can not be overridden at either executables or entry point level of any downstream job in the resulting execution tree by eithersystemRequirementsByExecutable
specifications or bysystemRequirements
specifications.Force all jobs in the execution tree executing
applet-1
to run onmem2_ssd1_v2_x2
. Downstream jobs in the resulting execution tree will be able to overrideinstanceType
for executables other thanapplet-1
, but will not be able to overrideapplet-1
's instance types for a specific entry point because of the"*"
entry point in the below specification.Force all jobs in the execution tree that are executing the
applet-1
main
entry point to run onmem2_ssd1_v2_x2
. Downstream jobs will be able to overrideinstanceType
for executables other thanapplet-1
and for entry points ofapplet-1
other thanmain
.Force all jobs in the execution tree that are executing the
applet-1
main
entry point to run onmem2_ssd1_v2_x2
and jobs executing all other entry points ofapplet-1
to run onmem2_ssd1_v2_x4
. Also force all jobs in the execution tree executing themain
entry point ofapplet-2
to run onmem2_ssd1_v2_x8
. Downstream jobs will be able to overrideinstanceType
for executables other thanapplet-1
, except for themain
entry point ofapplet-2
.
For more examples of how system requirements are resolved, see
.
Specifying Job Timeouts
Job timeout policies (configurable by executable ID and then by entry point) may be specified to request timeouts for jobs running specific executables at specific entry points. When a job times out, the job is either terminated or restarted (depending on the job restart policy). Job timeout policies may be specified upon creation of the executable (app or applet; configurable by entry point) in the executable's timeoutPolicy
field of the run specification and may be overridden (in part or in full) at runtime (via an input to /app-xxxx/run, /applet-xxxx/run, /workflow-xxxx/run , /globalworkflow-xxx/run, or /job/new; configurable by executable ID and then by entry point). The effective timeout of a job running a specific executable at a specific entry point is determined from the following sources (in decreasing priority):
The runtime input
timeoutPolicyByExecutable.<executable_id>.<entry_point>
field. Note that this field overrides both user-specified and system default timeout policies, and it propagates down the entire resulting job execution tree and merges with all downstream runtime inputs, with explicit upstream inputs taking precedence.The
timeoutPolicy.<entry_point>
field provided in the executable's run specification upon executable creation. This field serves as the user-specified default timeout policy and will override the system default timeout policy.The system default of 30 days. This limit is enforced for all jobs billed to orgs that do not have the
allowJobsWithoutTimeout
license.
The '*'
entry point refers to all other entry points that are not named in timeoutPolicyByExecutable.<executable_id>
or the timeoutPolicy
field of the executable's run specification.
Note that setting the timeout of a specific executable at a specific entry point to 0 is the same as not setting a timeout for that executable at that entry point at all (i.e. there exists no runtime or default run specification entry for that executable at that entry point).
Example: Suppose that an applet with ID <applet_id>
has a run specification that includes a timeout policy for entry point '*'
that specifies a timeout of 5 hours, and suppose that that applet is run with a runtime policy for <applet_id>
for entry point 'main'
that specifies a timeout of 2 hours. The resulting job will have a timeout of 2 hours. Now, suppose that the resulting job runs a subjob with no runtime policies at entry point 'foo'
. The resulting subjob will have a timeout of 5 hours.
Applet API Method Specifications
API Method: /applet/new
/applet/new
Specification
Creates a new applet object with the given applet specification.
Links specified in "bundledDepends" contribute to the "links" array returned by a describe call and are always cloned together with the applet regardless of their visibility.
Note that an applet that is not written in a supported interpreted language can still be run if a separate file object is created containing the compiled code. It is suggested that the user mark this file object as "hidden", and create an applet object with a link to it in the "bundledDepends" and code which specifically runs the compiled binary. Note that the file object does not need to be hidden for it to be bundled with the applet object, but for this purpose, it may be more useful to hide the file object if it is unlikely to be used in isolation. See the Execution Environment Reference for more details on how the contents of "bundledDepends" are handled.
The applet object does not receive special permissions for any referenced data objects (such as "id" in "bundledDepends"); these entries are accessed every time the applet is run, with the same permissions as the job. This means that if some referenced file is later deleted or is not present in the project context, the applet will not be able to run. However, the system will not automatically "invalidate" the applet object for any broken links; if the referenced file is reinstated from a copy existing in another project, the applet can then be run.
Inputs
project
string ID of the project or container to which the applet should belong (e.g. the string "project-xxxx")name
string (optional, default is the new ID) The name of the objecttitle
string (optional, default "") Title of the applet, e.g. "Micro Map"summary
string (optional, default "") A short description of the appletdescription
string (optional, default "") A longer description about the appletdeveloperNotes
string (optional, default "") More detailed notes about the applettags
array of strings (optional) Tags to associate with the objecttypes
array of strings (optional) Types to associate with the objecthidden
boolean (optional, default false) Whether the object should be hiddenproperties
mapping (optional) Properties to associate with the objectkey Property name
value string Property value
details
mapping or array (optional, default { }) JSON object or array that is to be associated with the object; see the Object Details section for details on valid inputfolder
string (optional, default "/") Full path of the folder that is to contain the new objectparents
boolean (optional, default false) Whether all folders in the path provided infolder
should be created if they do not existignoreReuse
boolean (optional, default false) If true then no job reuse will occur by default for this applet.inputSpec
array of mappings (optional) An input specification as described in the Input Specification sectionoutputSpec
array of mappings (optional) An output specification as described in the Output Specification sectionrunSpec
mapping A run specification as described in the run specification sectiondxapi
string The version of the API that the applet was developed with, for example, "1.0.0"access
mapping (optional) Access requirements as described in the Access Requirements sectionhttpsApp
mapping (optional) HTTPS app configurationports
array of integers Array of ports open for inbound access. Allowed ports are443
,8080
, and8081
shared_access
string HTTPS access restriction for jobs run from this executable. Allowed values are "VIEW", "CONTRIBUTE", "ADMINISTER", "NONE". VIEW, CONTRIBUTE, ADMINISTER require the specified permission level or greater for the project in which the job executes. Most restrictive NONE setting limits access to only the user who launched the job.dns
mapping DNS configuration for the jobhostname
string (optional) if specified, the URL to access this job in the browser will be "https://hostname.dnanexus.cloud" instead of the default "https://job-xxxx.dnanexus.cloud". The hostname must consist of lower case alphanumeric characters and a hyphen (-
) character and must match/^[a-z][a-z0-9]{2,}-[a-z][a-z0-9-]{2,}[a-z0-9]$/
regular expression. A user may run an app with custom hostname subject to these additional restrictions:The effective
billTo
of the running app must be an organization.The user must be an admin of that organization.
If the id of the
billTo
organization isorg-some_name
then the hostname must resemblehostprefix-some-name
, i.e. it must end in thehandle
portion of the org id with_
and.
replaced by hyphen (-
) and start with an alphanumeric "prefix" (in this casehostprefix
).
For an org with id of
org-some_name
, valid hostnames includemyprefix-some-name
andab5-some-name
; invalid hostnames includea-some-name
(prefix too short),a-some_name
(underscore present),4abc-some-name
(starts with a numerical character),myprefix-someotherorgname
(does not end with-some-name
).When two jobs attempt to use the same URL, the newer job will take over the hostname from an already running job. We recommend having a single job running with any given URL to avoid the risk of having the URL re-assigned to another job with same URL if the other job restarts due to its restart policy.
nonce
string (optional) Unique identifier for this request. Ensures that even if multiple requests fail and are retried, only a single applet is created. For more information, see Nonces.treeTurnaroundTimeThreshold
integer (optional, default: N/A or not available.) The turnaround time threshold (in seconds) for trees, specifically, root executions, that run this executable. See Job Notifications for more information about turnaround time and managing job notifications.
A license is required to use the Job Notifications feature. Contact DNAnexus Sales for more information.
Outputs
id
string ID of the created applet object (i.e. a string in the form "applet-xxxx")
Errors
InvalidInput
A reserved linking string ("$dnanexus_link") appears as a key in a mapping in
details
but is not the only key in the hash.A reserved linking string ("$dnanexus_link") appears as the only key in a hash in
details
but has value other than a string.The spec is invalid.
All specified bundled dependencies must be in the same region as the specified project.
A
nonce
was reused in a request but some of the other inputs had changed signifying a new and different requestA
nonce
may not exceed 128 bytestreeTurnaroundTimeThreshold
must be a non-negative integer less than 2592000timeoutPolicy for all entry points should not exceed 30 days
PermissionDenied
UPLOAD access is required to the specified project.
The
billTo
of the project is not licensed to usejobNotifications
. Contact sales@dnanexus.com to enablejobNotifications
.
ResourceNotFound
The route in
folder
does not exist whileparents
is false.
API Method: /applet-xxxx/describe
/applet-xxxx/describe
Specification
Describes an applet object.
Alternatively, you can use the /system/describeDataObjects method to describe a large number of data objects at once.
Inputs
project
string (optional) Project or container ID to be used as a hint for finding the object in an accessible projectdefaultFields
boolean (optional, default false iffields
is supplied, true otherwise) whether to include the default set of fields in the output (the default fields are described in the "Outputs" section below). The selections are overridden by any fields explicitly named infields
.fields
mapping (optional) include or exclude the specified fields from the output. These selections override the settings indefaultFields
.key Desired output field; see the "Outputs" section below for valid values here
value boolean whether to include the field
The following options are deprecated (and will not be respected if fields
is present):
properties
boolean (optional, default false) Whether the properties should be returneddetails
boolean (optional, default false) Whether the details should also be returned
Outputs
id
string The object ID (i.e. the string "applet-xxxx")
The following fields are included by default (but can be disabled using fields
or defaultFields
):
project
string ID of the project or container in which the object was foundclass
string The value "applet"types
array of strings Types associated with the objectcreated
timestamp Time at which this object was createdstate
string Either "open" or "closed"hidden
boolean Whether the object is hidden or notlinks
array of strings The object IDs that are pointed to from this object, including links found in both thedetails
and inbundledDepends
(if it exists) of the appletname
string The name of the objectfolder
string The full path to the folder containing the objectsponsored
boolean Whether the object is sponsored by DNAnexustags
array of strings Tags associated with the objectmodified
timestamp Time at which the user-provided metadata of the object was last modifiedcreatedBy
mapping How the object was createduser
string ID of the user who created the object or launched an execution which created the objectjob
string present if a job created the object ID of the job that created the objectexecutable
string present if a job created the object ID of the app or applet that the job was running
runSpec
mapping The run specification of the applet but without thecode
field (use the /applet-xxxx/get method for obtaining the source code)dxapi
string The version of the API usedaccess
mappings The access requirements of the applettitle
string The title of the appletsummary
string The summary of the appletdescription
string The description of the appletdeveloperNotes
string The developer notes of the appletignoreReuse
boolean Whether job reuse is disabled for this applethttpsApp
mapping HTTPS app configurationshared_access
string HTTPS access restriction for this jobports
array of integers Ports that are open for inbound accessdns
mapping DNS configuration for the jobhostname
string if specified, the URL to access this job in the browser will be "https://hostname.dnanexus.cloud" instead of the default "https://job-xxxx.dnanexus.cloud"
treeTurnaroundTimeThreshold
integer or null The turnaround time threshold (in seconds) for trees (specifically, root executions) that run this executable. See Job Notifications for more information about turnaround time and managing job notifications.
A license is required to use the jobNotifications
feature. Contact DNAnexus Sales to enable jobNotifications
.
The following field (included by default) is available if an input specification is specified for the applet:
inputSpec
array of mappings The input specification of the applet
The following field (included by default) is available if an output specification is specified for the applet:
outputSpec
array of mappings The output specification of the applet
The following field (included by default) is available if the object is sponsored by a third party:
sponsoredUntil
timestamp Indicates the expiration time of data sponsorship (this field is only set if the object is currently sponsored, and if set, the specified time is always in the future)
The following fields are only returned if the corresponding field in the fields
input is set to true
:
properties
mapping Properties associated with the objectkey Property name
value string Property value
details
mapping or array Contents of the object’s details
Errors
ResourceNotFound
the specified object does not exist or the specified project does not exist
InvalidInput
the input is not a hash,
project
(if supplied) is not a string, or the value ofproperties
(if supplied) is not a boolean
PermissionDenied
VIEW access required for the
project
provided (if any), and VIEW access required for some project containing the specified object (not necessarily the same as the hint provided)
API Method: /applet-xxxx/get
/applet-xxxx/get
Specification
Returns the full specification of the applet, i.e. the same output as /applet-xxxx/describe but with the runSpec
field left intact.
Inputs
None
Outputs
project
string ID of the project or container in which theobject was found
id
string The object ID (i.e. the string "applet-xxxx")class
string The value "applet"types
array of strings Types associated with the objectcreated
timestamp Time at which this object was createdstate
string Either "open" or "closed"hidden
boolean Whether the object is hidden or notlinks
array of strings The object IDs that are pointed to from this object, including links found in both thedetails
and inbundledDepends
(if it exists) of the appletname
string The name of the objectfolder
string The full path to the folder containing the objectsponsored
boolean Whether the object is sponsored by DNAnexustags
array of strings Tags associated with the objectmodified
timestamp Time at which the user-provided metadata of the object was last modifiedcreatedBy
mapping How the object was createduser
string ID of the user who created the object or launched an execution which created the objectjob
string present if a job created the object ID of the job that created the objectexecutable
string present if a job created the object ID of the app or applet that the job was running
runSpec
mapping The run specification of the appletdxapi
string The version of the API usedaccess
mappings The access requirements of the applettitle
string The title of the appletsummary
string The summary of the appletdescription
string The description of the appletdeveloperNotes
string The developer notes of the applet
If an input specification is specified for the applet:
inputSpec
array of mappings The input specification of the applet
If an output specification is specified for the applet:
outputSpec
array of mappings The output specification of the applet
Errors
ResourceNotFound
the specified object does not exist
PermissionDenied
VIEW access required
API Method: /applet-xxxx/run
/applet-xxxx/run
Specification
Creates a new job which will execute the code of this applet. The default entry point for the applet’s interpreter (given in the runSpec.interpreter field of the applet spec) will be called.
Interpreter
Entry point
bash
main() in top level scope with no args, if it exists. Also, $1 is set to "main"
python3
Any function decorated with @dxpy.app_entry(func="main") with no args
If constraints on inputs are specified in the applet spec, and the given inputs do not satisfy those constraints at the time the API call is performed, or if the names of inputs given do not exactly match the inputs listed in the applet object, or if an input is omitted and no default is listed in the applet object, an InvalidInput error will result. For inputs given as job-based object reference, an equivalent error may result at job dispatch time, in which case the job will fail.
The job might fail for the following reasons (this list is non-exhaustive):
A reference such as one mentioned in "bundledDepends" could not be accessed using the job’s credentials (VIEW access to project context, CONTRIBUTE access to a workspace, VIEW access to public projects)
A job-based object reference did not resolve successfully (invalid job, job ID not found, job not in that project, job is in failed state, field does not exist, field does not contain a valid object link).
An input object does not exist.
Permission denied accessing an input object.
An input object is not a data object (things like users, projects, or jobs are not data objects)
An input object does not satisfy the class constraints.
An input object does not satisfy the type constraints.
An input object is not in the "closed" state.
Insufficient credits.
The user has too many jobs that are in a nonterminal state.
Inputs
name
name (optional, default is the applet's title if set and otherwise the applet's name) Name for the resulting jobinput
mapping Input that the applet is launched withkey Input field name. If the applet has an input specification, it must be one of the names of the inputs; otherwise, it can be any valid input field name.
value Input field value
dependsOn
array of strings (optional) List of job, analysis and/or data object IDs; the applet will not begin running any of its entry points until all jobs listed have transitioned to the "done" state, and all data objects listed are in the "closed" stateproject
string (required if invoked by a user; optional if invoked from a job withdetach: true
option; prohibited when invoked from a job withdetach: false
) The ID of the project in which this applet will be run (i.e., the project context). If invoked with thedetach: true
option, then the detached job will run under the providedproject
(if provided), otherwise project context is inherited from that of the invoking job. If invoked by a user or run as detached, all output objects are cloned into the project context; otherwise, all output objects will be cloned into the temporary workspace of the invoking job. See this documentation for more information on project context on the DNAnexus Platform.folder
string (optional, default "/") The folder into which objects output by the job will be placed. If the folder does not exist when the job completes, the folder (and any parent folders necessary) will be created. The folder structure that output objects reside in is replicated within the target folder, e.g. iffolder
is set to "/myJobOutput" and the job outputs an object which is in the folder "/mappings/mouse" in the workspace, the object is placed into "/myJobOutput/mappings/mouse".tags
array of strings (optional) Tags to associate with the resulting jobproperties
mapping (optional) Properties to associate with the resulting jobkey Property name
value string Property value
details
mapping or array (optional, default { }) JSON object or array that is to be associated with the jobsystemRequirements
mapping (optional) Request specific resources for each of the executable's entry points; see the Requesting Instance Types section above for more detailssystemRequirementsByExecutable
mapping (optional) Request system requirements for all jobs in the resulting execution tree, configurable by executable and by entry point, described in more detail in the Requesting Instance Types section.timeoutPolicyByExecutable
mapping (optional) The timeout policies for jobs in the resulting job execution tree, configurable by executable and the entry point within that executable. If unspecified, it indicates that all jobs in the resulting job execution tree will have the default timeout policies present in the run specifications of their executables. Note thattimeoutPolicyByExecutable
(keyed first by app or applet ID and then entry point name) will propagate down the entire job execution tree, and that explicitly specified upstream policies always take precedence. If present, includes at least one of the following key-value pairs:key app or applet ID. If an executable is not explicitly specified in
timeoutPolicyByExecutable
, then any job in the resulting job execution tree that runs that executable will have the default timeout policy present in the run specification of that executable.value mapping Timeout policy for the corresponding executable. Includes at least one of the following key-value pairs:
key Entry point name or
"*"
to indicate all entry points not explicitly specified in this mapping. If an entry point name is not explicitly specified and"*"
is not present, then any job in the resulting job execution tree that runs the corresponding executable at that entry point will have the default timeout policy present in the run specification of the corresponding executable.value mapping Timeout for a job running the corresponding executable at the corresponding entry point. Includes at least one of the following key-value pairs:
key Unit of time; one of "days", "hours", or "minutes".
value number Amount of time for the corresponding time unit; must be non-negative. The effective timeout is the sum of the units of time represented in this mapping.
executionPolicy
mapping (optional) A collection of options that govern automatic job restart upon certain types of failures. The format of this field is identical to that of theexecutionPolicy
field in the run specification supplied to /applet/new and can override part or all of theexecutionPolicy
found in the applet's run specification (if present).delayWorkspaceDestruction
boolean (optional) If not given, the value defaults to false for root executions (launched by a user or detached from another job), or to the parent'sdelayWorkspaceDestruction
setting. If set to true, the temporary workspace created for the resulting job will be preserved for 3 days after the job either succeeds or fails.allowSSH
array of strings (optional, default [ ]) Array of IP addresses or CIDR blocks (up to /16) from which SSH access will be allowed to the user by the worker running this job. Array may also include '*' which is interpreted as the IP address of the client issuing this API call as seen by the API server. See Connecting to Jobs for more information.debug
mapping (optional, default { }) Specify debugging options for running the executable; this field is only accepted when this call is made by a user (and not a job)debugOn
array of strings (optional, default [ ]) Array of job errors after which the job's worker should be kept running for debugging purposes, offering a chance to SSH into the worker before worker termination (assuming SSH has been enabled). This option applies to all jobs in the execution tree. Jobs in this state for longer than 2 days will be automatically terminated but can be terminated earlier. Allowed entries include "ExecutionError", "AppError", and "AppInternalError".
singleContext
boolean (optional) If true then the resulting job and all of its descendants will only be allowed to use the authentication token given to it at the onset. Use of any other authentication token will result in an error. This option offers extra security to ensure data cannot leak out of your given context. In restricted projects user-specified value is ignored, andsingleContext: true
setting is used instead.ignoreReuse
boolean (optional) If true then no job reuse will occur for this execution. Takes precedence over value supplied forapplet-xxxx/new
.nonce
string (optional) Unique identifier for this request. Ensures that even if multiple requests fail and are retried, only a single job is created. For more information, see Nonces.detach
boolean (optional) This option has no impact when the API is invoked by a user. If invoked from a job with detach set to true, the new job will be detached from the creator job and will appear as a typical root execution. A failure in the detached job will not cause a termination of the job from which it was created and vice versa. Detached job will inherit neither the access to the workspace of its creator job nor the creator job's priority. Detached job's access permissions will be the intersection (most restricted) of access permissions of the creator job and the permissions requested by the detached job's executable. To launch the detached job, creator job must have CONTRIBUTE or higher access to the project in which the detached job is launched. Additionally, the billTo of the project in which the creator job is running must be licensed to launch detached executions.
For more information on a license that supports launching detached executions, contact DNAnexus Sales.
rank
integer (optional) An integer between -1024 and 1023, inclusive. The rank indicates the priority in which the executions generated from this executable will be processed. The higher the rank, the more prioritized it will be. If no rank is provided, the executions default to a rank of zero. If the execution is not a root execution, it will inherit its parent's rank.
A license is required to use the Job Ranking feature. Contact DNAnexus Sales for more information.
costLimit
float (optional) The limit of the cost that this execution tree should accrue before termination. This field will be ignored if this is not a root execution.headJobOnDemand
boolean (optional) If true, then the resulting master job will be allocated to an on-demand instance, regardless of its scheduling priority. All of its descendent jobs (if any) inherit its scheduling priority, and their instance allocations are independent from this option. This option overrides the settings in the app’sheadJobOnDemand
(if any).preserveJobOutputs
mapping (optional, default is null). Preserves all cloneable outputs of every completed, non-jobReused job in the execution tree launched by this api call in the root execution project, even if root execution ends up failing. Preserving the job outputs in the project trades off higher costs of storage for the possibility of subsequent job reuse.When a non-jobReused job in the root execution tree launched with non-null
preserveJobOutputs
enters the "done" state, all cloneable objects (e.g. files, records, applets, closed workflows, but not databases) referenced by the$dnanexus_link
in the job'soutput
field will be cloned to the project folder described bypreserveJobOutputs.folder
, unless the output objects already appear elsewhere in the project. If the folder specified bypreserveJobOutputs.folder
does not exist in the project, the system will create the folder and its parents. As the root job or root analysis' stages complete, the regular outputs of the root execution will be moved frompreserveJobOutputs.folder
to the regular output folder(s) of the root execution. So if you [1] run your root execution without thepreserveJobOutputs
option to completion, some root execution outputs will appear in the project in the root execution's output folder(s). If you had run the same execution withpreserveJobOutputs.folder
set to"/pjo_folder"
, the same set of outputs would appear in the same set of root execution folders as in [1] at completion of the root execution, while some additional job outputs that are not outputs of the root execution would appear in"/pjo_folder"
.preserveJobOutputs
argument can be specified only when starting a root execution or a detached job.preserveJobOutputs
value, if not null, should be a mapping that may contain the following:key
"folder"
string (optional)value
path_to_folder
string (required if "folder" key is specified). Specifies a folder in the root execution project where the outputs of jobs that are a part of the launched execution will be stored.path_to_folder
starting with "/
" is interpreted as absolute folder path in the project the job is running in.path_to_folder
not starting with "/
" is interpreted as a path relative to the root execution'sfolder
field.path_to_folder
value of""
(i.e. empty string) will preserve job outputs in the folder described by root execution'sfolder
field. If thepreserveJobOutputs
mapping does not have afolder
key, the system will use the default folder value of"intermediateJobOutputs"
(i.e."preserveJobOutputs": {}
is equivalent to"preserveJobOutputs": {"folder":"intermediateJobOutputs"}
).
It is recommended to place preserveJobOutputs outputs for different root executions into different folders so as not to create a single folder with a very large (>450K) number of files.
A license is required to use preserveJobOutputs. Contact DNAnexus Sales for more information.
detailedJobMetrics
boolean Requests detailed metrics collection for jobs if set to true. The default value for this flag is projectbillTo
'sdetailedJobMetricsCollectDefault
policy setting or false if org default is not set. This flag can be specified for root executions and will apply to all jobs in the root execution. The list of detailed metrics collected every 60 seconds and viewable for 15 days from the start of a job is here.
A license is required to use detailedJobMetrics
. Contact DNAnexus Sales for more information.
Outputs
id
string ID of the created job (i.e. a string in the form "job-xxxx")
Errors
ResourceNotFound
The specified applet object or project context does not exist.
One of the IDs listed in
dependsOn
does not exist.
PermissionDenied
The requesting user must have VIEW access to all objects listed in
dependsOn
, and to all project contexts of all jobs listed independsOn
.The requesting user must have VIEW access to the applet object.
If invoked by a user, then the requesting user must have CONTRIBUTE access to the project context.
The requesting user must be able to describe all jobs used in a job-based object reference -- see /job-xxxx/describe.
The requesting user has too many (65536, by default) nonterminal (e.g., running, runnable) jobs and must wait for some to finish before creating more.
The
billTo
of the job's project must be licensed to start detached executions when invoked from the job withdetach: true
argument.If rank is provided and the billTo does not have license feature executionRankEnabled set to true.
if
preserveJobOutputs
is not null andbillTo
of the project where execution is attempted does not have preserveJobOutputs license.detailedJobMetrics
setting of true requires project'sbillTo
to havedetailedJobMetrics
license feature set to true.
InvalidInput
input
does not satisfy the input specification of this applet; an additional field is provided in the error JSON for this error that looks likeThe possible error reasons are described in Input Specification Errors.
If invoked by a user, then
project
must be specified.If invoked by a job, then
project
must not be specified.The project context must be in the same region as this applet.
All data object inputs that are specified directly must be in the same region as this applet.
All inputs that are job-based object references must refer to a job that was run in the same region as this applet.
allowSSH
accepts only IP addresses or CIDR blocks up to /16.A
nonce
was reused in a request but some of the other inputs had changed signifying a new and different request.A
nonce
may not exceed 128 bytes.preserveJobOutputs.folder
value is a syntactically invalid path to a folder.preserveJobOutputs
is specified when launching a non-detached execution from a job.detailedJobMetrics
can not be specified when launching a non-detached execution from a job.timeoutPolicyByExecutable
for all executables should not benull
timeoutPolicyByExecutable
for all entry points of all executables should not benull
timeoutPolicyByExecutable
for all entry points of all executables should not exceed 30 daysExpected key "
timeoutPolicyByExecutable
.*" of input to match "/^(app|applet)-[0-9A-Za-z]{24}$/"
InvalidState
Some specified input is not in the "closed" state.
Some job in
dependsOn
has failed or has been terminated.
Input Specification Errors
The following list describes the possible error reasons and what the fields mean:
"class": the specified "field" was expected to have class "expected". If the input spec required an array but it was not an array, the value for "expected" will be "array". If the input spec required an array but an element was of the wrong class, then the value for "expected" will be the actual class the entry was expected to be, e.g. "record".
"type": the specified "field" either needs to have the type "expected" or does not satisfy the or-condition in "expected"
"missing": the specified "field" was not provided but is required in the input specification
"unrecognized": the given "field" is not present in the input specification
"malformedLink": incorrect syntax was given either for a job-based object reference or for a link to a data object. Possible values for "expected" include:
"key "field"": the key "field" was missing in a job-based object reference
"only two keys": exactly two keys were expected in the hash for the job-based object reference
"key "$dnanexus_link"": the key "$dnanexus_link" was missing in a link for specifying a data object
"choices": the specified "field" must be one of the values in "expected"
API Method: /applet-xxxx/validateBatch
/applet-xxxx/validateBatch
Specification
This API call verifies that a set of input values for a particular applet can be used to launch a batch of jobs in parallel. The applet must have input specification defined.
Batch and common inputs:
batchInput
: mapping of inputs corresponding to batches. The nth value of each array corresponds to nth execution of the applet. Including a null
value in an array at a given position means that the corresponding applet input field is optional and the default value, if defined, should be used. E.g.:
commonInput
: mapping of non-batch, constant inputs common to all batch jobs, e.g.:
File references:
files
: list of files (passed as $dnanexus_link references), must be a superset of files included in batchInput
and/or commonInput
e.g.:
Output: list of mappings, each mapping corresponds to an expanded batch call. Nth mapping contains the input values with which the nth execution of the applet will be run, e.g.:
It performs the following validation:
the input types match the expected applet input field types,
provided inputs are sufficient to run the applet,
null values are only among values for inputs that are optional or have no specified default values,
all arrays of
batchInput
are of equal size,every file referred to in
batchInputs
exists infiles
input.
Inputs
batchInput
mapping Input that the applet is launched withkey Input field name. It must be one of the names of the inputs defined in the applet input specification.
value Input field values. It must be an array of fields.
commonInput
mapping (optional) Input that the applet is launched withkey Input field name. It must be one of the names of the inputs defined in the applet input specification.
value Input field values. It must be an array of fields.
files
list (optional) Files that are needed to run the batch jobs, they must be provided as$dnanexus_links
. They must correspond to all the files included incommonInput
orbatchInput
.Outputs
expandedBatch
list of mappings Each mapping contains the input values for one execution of the applet in batch mode.
Errors
InvalidInput
inputSpec
must be specified for the appletExpected
batchInput
to be a JSON objectExpected
commonInput
to be a JSON objectExpected
files
to be an array of$dnanexus_link
references to filesThe
batchInput
field is required but empty array was providedExpected the value of
batchInput
for an applet input field to be an arrayExpected the length of all arrays in
batchInput
to be equalThe applet input field value must be specified in
batchInput
The applet input field is not defined in the input specification of the applet
All the values of a specific
batchInput
field must be provided (cannot benull
) since the field is required and has no default valueExpected all the files in
batchInput
andcommonInput
to be referenced in thefiles
input array
Job API Method Specifications
API Method: /job/new
/job/new
Specification
This API call may only be made from within an executing job. This call creates a new job which will execute a particular function (from the same applet as the one the current job is running) with a particular input. The input will be checked for links, and any job-based object reference will be honored. However, the input is not checked against the applet spec. Since this is done from inside another job, the new job will inherit the same workspace and project context -- no objects will be cloned, and no other modification will take place in the workspace.
The entry point for the job’s execution will be determined as follows, where f is the string given in the function parameter in the input:
Interpreter
Entry point
bash
f() in top level scope with no args, if it exists. Also, $1 is set to f
python3
Any function decorated with @dxpy.entry_point("f")
See Execution Environment Reference for more info.
This call will fail if the specified OAuth2 token does not internally represent a currently executing job.
The system takes note of the job that this new job is created from; this information is available in the "parent" field when describing the new job object. Moreover, the parent-child relationship is tracked by the system, and used to advance the job state of the parent. Specifically, a job which has finished executing remains in the "waiting_on_output" state until all of its child jobs have proceeded to the "done" state. See Job Lifecycle for more information.
The new job may fail for at least the following reasons:
A reference such as one mentioned in "bundledDepends" could not be accessed using the job’s credentials (VIEW access to project context, CONTRIBUTE access to workspace, VIEW access to public projects)
A job-based object reference did not resolve successfully (invalid job, job ID not found, job not in that project, job is in failed state, field does not exist, field does not contain a valid object link).
Insufficient credits.
Inputs
name
string (optional, default "<parent job's name>:<function
>") Name for the resulting jobinput
mapping Input that the job is launched with; no syntax checking occurs, but the mapping will be checked for links and create dependencies on any open data objects or unfinished jobs accordinglykey Input field name
value Input field value
dependsOn
array of strings (optional) List of job, analysis and/or data object IDs; the newly created job will not run until all executions listed independsOn
have transitioned to the "done" state, and all data objects listed are in the "closed" statefunction
string The name of the entry point or function of the applet’s code that will be executedtags
array of strings (optional) Tags to associate with the resulting jobproperties
mapping (optional) Properties to associate with the resulting jobkey Property name
value string Property value
details
mapping or array (optional, default { }) JSON object or array that is to be associated with the jobsystemRequirements
mapping (optional) Request specific resources for each of the executable's entry points; see the Requesting Instance Types section above for more detailssystemRequirementsByExecutable
mapping (optional) Request system requirements for all jobs in the resulting execution subtree, configurable by executable and by entry point, described in more detail in the Requesting Instance Types section.timeoutPolicyByExecutable
mapping (optional) Similar to thetimeoutPolicyByExecutable
field supplied to /applet-xxxx/runignoreReuse
boolean (optional) If true then no job reuse will occur for this execution. Takes precedence over value supplied forapplet-xxxx/new
.singleContext
boolean (optional) If true then the resulting job and all of its descendants will only be allowed to use the authentication token given to it at the onset. Use of any other authentication token will result in an error. This option offers extra security to ensure data cannot leak out of your given context. In restricted projects user-specified value is ignored, andsingleContext: true
setting is used instead.nonce
string (optional) Unique identifier for this request. Ensures that even if multiple requests fail and are retried, only a single job is created. For more information, see Nonces.headJobOnDemand
boolean (optional) If true, then the resulting root job will be allocated to an on-demand instance, regardless of its scheduling priority. All of its descendent jobs (if any) inherit its scheduling priority, and their instance allocations are independent from this option. This option overrides the settings in the app’sheadJobOnDemand
(if any).
Outputs
id
string ID of the created job (i.e. a string in the form "job-xxxx").
Errors
InvalidAuthentication (the usual reasons InvalidAuthentication is thrown, or the auth token used is not a token issued to a job)
ResourceNotFound (one of the IDs listed in
dependsOn
does not exist)PermissionDenied (VIEW access required for any objects listed in
dependsOn
and for the project contexts of any jobs listed independsOn
; ability to describe any job used in a job-based object referenceInvalidInput
The input is not a hash
input
is missing or is not a hashan invalid link syntax appears in the
input
dependsOn
, if given, is not an array of strings"details", if given, is not a conformant JSON object
For each property key-value pair, the size, encoded in UTF-8, of the property key may not exceed 100 bytes and the property value may not exceed 700 bytes
A
nonce
was reused in a request but some of the other inputs had changed signifying a new and different requestA
nonce
may not exceed 128 bytes
InvalidState
some job in
dependsOn
has already failed or been terminated
API Method: /job-xxxx/describe
/job-xxxx/describe
Specification
Describes a job object. Each job is the result of either running an applet by calling "run" on an applet or app object, or the result of launching a job from an existing job by calling the "new" job class method. In all cases, the job is associated with either an applet or an app, and a project context reflected in the "project" field. Moreover, the ID of the user who launched the execution that this job is a part of is reflected in the "launchedBy" field (and is propagated to all child jobs as well, including any executables launched from those jobs). A job is always in a particular state; for information about job states, see the Job Lifecycle section.
Users with reorg apps that rely on describing the currently running job may want to check the output field "dependsOn" before the full analysis description becomes available using dx describe analysis-xxx --json | jq -r .dependsOn
or equivalent dxpy
bindings. The output of the command will be an empty array "[]" if it no longer depends on anything (e.g. status "done") which is the signal to proceed, or some job / subanalysis IDs if it is not ready, and the reorg script should wait.
Inputs
defaultFields
boolean (optional, defaultfalse
if fields is supplied,true
otherwise) Whether to include the default set of fields in the output (default fields described in the "Outputs" section below). The selections are overridden by any fields explicitly named in fields.fields
mapping (optional) Include or exclude the specified fields from the output. These selections override the settings indefaultFields
.key Desired output field; see the "Outputs" section below for valid values here
value boolean Whether to include the field
try
non-negative integer (optional). Specifies a particular try of a restarted job. Value of 0 refers to the first try. Defaults to the latest try (i.e. the try with the largesttry
attribute) for the specified job ID. See Restartable Jobs section for details.
The following option is deprecated (and will not be respected if fields
is present):
io
boolean (optional, default true) Whether the input and output fields (runInput
,originalInput
,input
, andoutput
) should be returned
Outputs
id
string The job ID (i.e. the string "job-xxxx")
The following fields are included by default (but can be disabled using fields
):
try
non-negative integer or null. Returns the try for this job, with 0 corresponding to the first try, 1 corresponding to the second try for restarted jobs and so on.null
is returned for jobs belonging to root executions launched before July 12, 2023 00:13 UTC and information for the latest job try is returned.class
string The value "job"name
string The name of the job tryexecutableName
string The name of the executable (applet or app) that the job was created to runcreated
timestamp Time at which this job was created. All tries of this job have the samecreated
value corresponding to creation time of the first job try.tryCreated
timestamp or null Time at which this job'stry
was created.null
is returned for jobs belonging to root executions launched before July 12, 2023 00:13 UTC. For job try 0, this field has the same value as thecreated
field.modified
timestamp Time at which this job try was last updatedstartedRunning
timestamp (present once the transition has occurred) Time at which this job try transitioned into the "running" state (see Job Lifecycle)stoppedRunning
timestamp (present once the transition has occurred) Time at which this job try transitioned out of the "running" state (see Job Lifecycle)egressReport
mapping or undefined A mapping detailing the total bytes of egress for a particular job try.regionLocalEgress
int Amount in bytes of data transfer between IP in the same cloud region.internetEgress
int Amount in bytes of data transfer to IP outside of the cloud provider.interRegionEgress
int Amount in bytes of data transfer to IP in other regions of the cloud provider.
billTo
string ID of the account to which any costs associated with this job will be billedproject
string The project context associated with this jobfolder
string The output folder in which the outputs of this job’s master job will be placedrootExecution
string ID of the top-level job or analysis in the execution treeparentJob
string or null ID of the job which created this job, or null if this job is an origin jobparentJobTry
non-negative integer or null.null
is returned if the job try being described had no parent, or if the parent itself had a nulltry
attribute. Otherwise, thisjob-xxxx
with the try attributes specified in the method's input was launched from theparentJobTry
try of theparentJob
.originJob
string The closest ancestor job whoseparentJob
is null, either because it was run by a user directly or was run as a stage in an analysisdetachedFrom
string or null The ID of the job this job was detached from via thedetach
option, otherwise nulldetachedFromTry
non-negative integer or null. If this job was detached from a job,detachedFrom
anddetachedFromTry
describe the specific try of the job this job was detached from.null
is returned if this job was not detached from another job or if thedetachedFrom
had anull
try
attribute.parentAnalysis
string or null ID of the analysis, present for an origin job that is run as a stage in an analysis, otherwise nullanalysis
string or null ID of the nearest ancestor analysis in the execution tree, if one exists, otherwise nullstage
string or null Null if this job was not run as part of a stage in an analysis; otherwise, the ID of the stage this job is part ofstate
string The job state: one of "idle", "waiting_on_input", "runnable", "running", "waiting_on_output", "done", "debug_hold", "restartable", "failed", "terminating", and "terminated"); see Job Lifecycle for more details on job statesstateTransitions
array of mappings Each element in the list indicates a time at which the state of the job try changed; the initial state of a job try is always "idle" when it is created and is not included in the list. Each hash has the key/values:newState
string The new state, e.g. "runnable"setAt
timestamp Time at which the new state was set for the job try
workspace
string (present once the workspace has been allocated) ID of the temporary workspace assigned to the job try (e.g. "container-xxxx")launchedBy
string ID of the user who launched the original jobfunction
string Name of the function, or entry point, that this job is runningtags
array of strings Tags associated with the job tryproperties
mapping Properties associated with the job trykey Property name
value string Property value
finalPriority
string The final priority this job try was run at. If the job try was run on an on-demand instance,finalPriority
will be set to "high" regardless the original priroity setting.rank
integer The rank assigned to the job and all of its child executions. The range is from [-1024 to 1023].
A license is required to use the Job Ranking feature. Contact DNAnexus Sales for more information.
details
mapping or array JSON details that were stored with this jobsystemRequirements
mapping Resolved resources requested for each of the executable's entry points based onmergedSystemRequirementsByExecutable
,runStageSystemRequirements
,runSystemRequirements
, system requirements embedded in workflow stages, andsystemRequirements
supplied to/executable/new
. See the Requesting Instance Types section above for more details.executionPolicy
mapping Options specified for this job to perform automatic restart upon certain types of failures. The format of this field is identical to that of theexecutionPolicy
field in the run specification supplied to /applet/new. This hash is computed from the following sources, in order of precedence:Values given to /app-xxxx/run, /applet-xxxx/run, or /workflow-xxxx/run
Values given to /workflow-xxxx/addStage
Values given to /applet/new in the run specification
timeout
int The effective timeout, in milliseconds, for this job. This number will be smaller or equal to 30 days (2592000000 milliseconds) for most jobs. Most jobs launched prior to December 19, 2024 or jobs billed to orgs with theallowJobsWithoutTimeout
license may have this field set to "undefined," indicating that no timeout policy was specified for the executable being run by this job, or that the effective timeout resolved to 0ms.instanceType
string Instance type for this job try, computed from thesystemRequirements
specification or the system-wide default if no instance type was requested.networkAccess
array of strings The computed network access list that is available to the job; this may be a subset of the requested list in the Access Requirements of the executable if an ancestor job had access to a more restricted list of domains.delayWorkspaceDestruction
boolean (present for origin and master jobs only) Whether the job's temporary workspace will be kept around for 3 days after the job either succeeds or failsdependsOn
array of strings If the job is in a waiting state ("waiting_on_input" or "waiting_on_output"), an array of job IDs and data object IDs that must transition to the done or closed state for the job to transition out of the waiting state to either "runnable" or "done"failureReason
string (present if the job try is in a failing, failed or restarted state) A short string describing where the error occurred, e.g. "AppError" for errors thrown in the execution of the applet or appfailureMessage
string (present if the job try is in a failing, failed or restarted state) A more detailed message describing why the error occurredfailureFrom
mapping (present if the job is try in a failing, failed or restarted state) Metadata describing the job try which caused the failure of this job try (which may be the same job and job try as the one being described):id
string ID of the failed jobtry
non-negative integer or undefined.try
of the failure-causing job for thisjob-xxxx
’stry
. Present only if the failure-causing job had the try attribute.name
string name of the jobexecutable
string ID (of the form "applet-xxxx" or "app-xxxx") of the executable the job was runningexecutableName
string Name of the executable the job was runningfunction
string name of the function, or entry point, the job was runningfailureReason
stringfailureReason
of the failed jobfailureMessage
stringfailureMessage
of the failed job
failureReports
array of mappings (present if this job failure was reported through a support mechanism) Each item in the list has the following fields:to
string Email address the failure was reported toby
string ID of the user who reported the failureat
timestamp Time at which the report was made
failureCounts
mapping A mapping from failure types (e.g. "AppError") to the number of times that type occurred and caused the job to be restarted prior to the job try being described.runInput
mapping Theinput
field that was provided when launching this joboriginalInput
mapping The same asrunInput
but with default values filled in for any optional inputs that were not providedinput
mapping The same asoriginalInput
, except that if any job-based references have since been resolved, they are replaced with the resulting object IDs; once the job's state has transitioned to "runnable", this represents exactly the input that will be given to the joboutput
mapping or null If the job is in the "done" state, this is the output that this job has generated as output. Otherwise, this value is null. The output may contain unresolved job-based object references.region
string The region (e.g., "aws:us-east-1") in which the job is running.singleContext
boolean Whether the job was specified, at run time, to be locked down in order to only issue requests from its own job token.ignoreReuse
boolean Whether job reuse was disabled for this jobhttpsApp
mapping HTTPS app configurationenabled
boolean Whether HTTPS app configuration is enabled for this jobshared_access
string HTTPS access restriction for this jobports
array of integers Ifenabled
is true, which ports are open for inbound accessdns
mapping DNS configuration for the joburl
string Defaults to "https://job-xxxx.dnanexus.cloud" unless overridden byhostname
in the executable's dxapp.json.
preserveJobOutputs
null or a mapping withpreserveJobOutputs.folder
expanded to start with"/"
.detailedJobMetrics
boolean Set to true only if the detailed job metrics collection was enabled for this jobclusterSpec
mapping A copy of the clusterSpec (if present) in the executable used to launch this jobclusterID
string (present only for jobs with a clusterSpec) Unique ID used to identify the cluster of workers running this job trycostLimit
float If the job is a root execution, and has the root execution cost limit, this is the cost limit for the root execution.
If this job is a root execution, the following fields are included by default (but can be disabled using fields
):
selectedTreeTurnaroundTimeThreshold
integer or null The selected turnaround time threshold (in seconds) for this root execution. WhentreeTurnaroundTime
reachesselectedTreeTurnaroundTimeThreshold
, the system sends an email about this root execution to thelaunchedBy
user and thebillTo
profile.selectedTreeTurnaroundTimeThresholdFrom
string or null WhereselectedTreeTurnaroundTimeThreshold
is from.executable
means thatselectedTreeTurnaroundTimeThreshold
is from this root execution's executable'streeTurnaroundTimeThreshold
.system
means thatselectedTreeTurnaroundTimeThreshold
is from the system's default threshold.treeTurnaroundTime
integer The turnaround time (in seconds) of this root execution, which is the time between its creation time and its terminal state time (or the current time if it is not in a terminal state. Terminal states for an execution include done, terminated, and failed. See Job Lifecycle for information on them). If this root execution can be retried, the turnaround time begins at the creation time of the root execution's first try, so it includes the turnaround times of all tries.
A license is required to use the jobNotifications
feature. Contact sales@dnanexus.com to enable jobNotifications
.
The following fields (included by default) are only available if the requesting user has permissions to view debug options for the job (either the user launched the job, or the user has ADMINISTER permissions in the job's project context; see Connecting to Jobs for more information):
debugOn
array of strings Array of error types (e.g. AppError) upon which the job will be held for debugging
The following fields (included by default) are only available if the requesting user has permissions to view the pricing model of the billTo
of the job, the job is the last try of an origin job, and the job's price has been computed:
isFree
boolean Whether this job will be charged to thebillTo
(usually set to true if the job has failed for failure reasons which are usually indicative of some system error rather than user error)currency
mapping Information about currency settings, such as dxCode, code, symbol, symbolPosition, decimalSymbol and groupingSymbol.totalPrice
number Price (incurrency
) for how much this job (along with all its subjobs) costs (or would cost ifisFree
is true)priceComputedAt
timestamp Time at whichtotalPrice
was computed. For billing purposes, the cost of the job accrues to the invoice of the month that containspriceComputedAt
(in UTC).totalEgress
mapping Egress (inByte
) for how much data amount this job (along with all its subjobs) has egressed.regionLocalEgress
int Amount in bytes of data transfer between IP in the same cloud region.internetEgress
int Amount in bytes of data transfer to IP outside of the cloud provider.interRegionEgress
int Amount in bytes of data transfer to IP in other regions of the cloud provider.
egressComputedAt
timestamp Time at whichtotalEgress
was computed. For billing purposes, the cost of the analysis accrues to the invoice of the month that contains egressComputedAt (in UTC).
The following fields (included by default) are only available if the requesting user has permissions to view worker information for the job (either the user launched the job, or the user has CONTRIBUTE permissions in the job's project context project):
allowSSH
array of strings Array of IP addresses or CIDR blocks from which SSH access will be allowed to the user by the worker running this job try.
The following fields (included by default) are only available if the requesting user has permissions to view worker information for the job (either the user launched the job, or the user has CONTRIBUTE permissions in the job's project context project) and a worker has started running the job try:
sshHostKey
string The worker's SSH host keyhost
string The worker's public FQDN or IP addresssshPort
string TCP port that can be used to connect to the SSH daemon for monitoring or debugging the jobclusterSlaves
array of mappings For jobs running on a cluster of instances, an array describing the slave nodes with keys:host
string the public hostname of the worker, can be used to SSH to the node if allowSSH was enabled for the jobsshPort
string TCP port to use when connecting via SSH to this nodeinternalIp
string The private IP address for this node, only accessible from other hosts in this cluster
The following fields (included by default) are only available if this job try is running an applet:
applet
string ID of the applet from which the job was run
The following fields (included by default) are only available if this job try is running an app:
app
string ID of the app from which the job was runresources
string ID of the app’s resources containerprojectCache
string ID of the project cache
The following field is only returned if the corresponding field in the fields
input is set to true
:
headJobOnDemand
boolean The value ofheadJobOnDemand
that the job was started withrunSystemRequirements
mapping or null A mapping with thesystemRequirements
values that were passed explicitly to/executable-xxxx/run
or/job/new
when the job was created, ornull
if thesystemRequirements
input was not supplied the API call that created the job.runSystemRequirementsByExecutable
mapping or null Similar torunSystemRequirements
but forsystemRequirementsByExecutable
.mergedSystemRequirementsByExecutable
mapping or null A mapping with values ofsystemRequirementsByExecutable
supplied to all the ancestors of this job and the value supplied to create this job, merged as described in the Requesting Instance Types section. If neither the ancestors of this job nor this job were created with thesystemRequirementsByExecutable
input,mergedSystemRequirementsByExecutable
value ofnull
is returned.
The following field is only returned if the corresponding field in the fields
input is set to true, and
the caller is an ADMIN of the org that the project of job-xxxx
is currently billed to (in addition to the caller having access to the project, as already required by /job-xxxx/describe
) and the billTo
of the project in which the job ran (at the time of the /job-xxxx/describe
call) is licensed to collect and view job's internet usage IPs. If any of these conditions are not met, /job-xxxx/describe
will omit internetUsageIPs
from its output, while returning other valid requested output fields.
A license is required to use the Internet Usage IPs
feature. Contact DNAnexus Sales for more information.
Note that this information should only be used by org admins for forensic investigation and fraud prevention purposes.
internetUsageIPs
array of strings with unique string-encoded IP addresses that the job code communicated with, in no specific order, subject to the following conditionshttpsApp and ssh-to-worker connections are included
IP traffic involving Application Execution Environment (AEE)
IP traffic passing through DNAnexus proxies running on the worker is excluded
All other IP traffic involving AEE is included
internetUsageIPs
includes IPs that were communicated with over various IP protocols, not just TCP (e.g. UDP, ICMP)IP addresses accessed by a restarted job will be rolled up into internetUsageIPs field of the restarted job's closest visible ancestor job for jobs whose root execution was created before July 12, 2023 00:13 UTC.
internetUsageIPs for
restarted jobs in root executions created after July 12, 2023 00:13 UTC can be described using thetry
input argument to this API method.internetUsageIPs
for a cluster job will include IP addresses that were communicated with from the cluster worker nodes as well as the main node, including from restarted cluster nodes.
The following field is only returned if the corresponding field in the fields
input is set to true
, the requesting user has permissions to view the pricing model of the billTo
of the job, and the job is the last try of a root execution:
subtotalPriceInfo
mapping Information about the current costs associated with all jobs in the tree rooted at this jobsubtotalPrice
number Current cost (incurrency
) of the job tree rooted at this jobpriceComputedAt
timestamp Time at whichsubtotalPrice
was computed
subtotalEgressInfo
mapping Information about the aggregated egress amount in bytes associated with all jobs in the tree rooted at this analysissubtotalRegionLocalEgress
int Amount in bytes of data transfer between IP in the same cloud region.subtotalInternetEgress
int Amount in bytes of data transfer to IP outside of the cloud provider.subtotalInterRegionEgress
int Amount in bytes of data transfer to IP in other regions of the cloud provider.egressComputedAt
timestamp Time at which subtotalEgress was computed
The following field is included only if explicitly requested in the fields input, by a user with VIEW access to a job that is billed to an org with the job logs forwarding feature enabled:
jobLogsForwardingStatus
mapping Information on the status of job logs for the job; ornull
, ifjobLogsForwarding
has not been configured for this job, or ifjobLogsForwardingStatus
has not been updated yet. It has the following keys:linesDropped
int The number of job log lines whose delivery to Splunk failed from the start of this job's try.
A license is required to use the Forwarding Job Logs to Customer's Splunk feature. Contact DNAnexus Sales for more information
Errors
ResourceNotFound
The specified object does not exist
try
inputT
is specified, but there is no tryT
forjob-xxxx.
Also returned iftry
input was specified for jobs in root executions created before July 12, 2023 00:13 UTC.
PermissionDenied
Require either VIEW access to the job's temporary workspace, or VIEW access to the parent job's temporary workspace
InvalidInput
Input is not a hash, or
fields
if present, is not a hash or has a non-boolean keytry
input should be a non-negative integer.
API Method: /job-xxxx/update
/job-xxxx/update
Specification
Update a job. Note that most runtime options for jobs are immutable for reproducibility reasons.
A license is required to use the Job Ranking feature. Contact DNAnexus Sales for more information.
If a rank field is present, a valid rank must be provided. The organization associated with this job must have the license feature executionRankEnabled
active in order to update rank. In addition, the user must be the original launcher of the analysis or is an administrator of the organization.
When supplying rank, the job/analysis being updated must be a rootExecution
, and the job/analysis being updated must be in a state that is capable of creating more jobs. Rank cannot be supplied for terminal states like "terminated", "done", "failed", "debug_hold".
Inputs
allowSSH
array of strings (optional, default [ ]) Array of IP addresses or CIDR blocks (up to /16) from which SSH access will be allowed to the user by the worker running this job. Array may also include '*' which is interpreted as the IP address of the client issuing this API call as seen by the API server. See Connecting to Jobs for more information. Changing this value after a job has started running may not be taken into account for the running job, but it may be passed on to new child jobs created after the update.rank
integer (optional) The rank to set the job and all of its children executions to.
Outputs
id
string ID of the updated job (i.e., the string "job-xxxx")
Errors
InvalidInput
Input is not a hash
Expected key "rank" of input to be an integer
Expected key "rank" of input to be in range [-1024, 1023]
Not a root execution
allowSSH accepts only IP addresses or CIDR blocks up to /16
ResourceNotFound
The specified job does not exist
PermissionDenied
billTo does not have license feature executionRankEnabled
Not permitted to change rank
The supplied token does not belong to the user who started the job or who has ADMINISTER access to the job's parent project.
API Method: /job-xxxx/addTags
/job-xxxx/addTags
Specification
Adds the specified tags to the specified job. If any of the tags are already present, no action is taken for those tags.
Inputs
tags
array of strings Tags to be addedtry
non-negative integer (optional) Specifies a particular try of a restarted job. Value of 0 refers to the first try. Defaults to the latest try (i.e. the try with the largesttry
attribute) for the specified job ID. See Restartable Jobs section for details.
Outputs
id
string ID of the manipulated job
Errors
InvalidInput
The input is not a hash, the key
tags
is missing, or its value is not an array, or the array contains at least one invalid (not a string of nonzero length) tag.try
input should be a non-negative integer
ResourceNotFound
The specified job does not exist
try
inputT
is specified, but there is no tryT
forjob-xxxx.
Also returned iftry
input was specified for jobs in root executions created before July 12, 2023 00:13 UTC.
PermissionDenied
CONTRIBUTE access is required for the job's project context; otherwise, the request can also be made by jobs sharing the same workspace as the specified job or the same workspace as the parent job of the specified job.
API Method: /job-xxxx/removeTags
/job-xxxx/removeTags
Specification
Removes the specified tags from the specified job. Ensures that the specified tags are not part of the job -- if any of the tags are already missing, no action is taken for those tags.
Inputs
tags
array of strings Tags to be removedtry
non-negative integer (optional). Specifies a particular try of a restarted job. Value of 0 refers to the first try. Defaults to the latest try (i.e. the try with the largesttry
attribute) for the specified job ID. See Restartable Jobs section for details.
Outputs
id
string ID of the manipulated job
Errors
InvalidInput
The input is not a hash, or the key
tags
is missing, or its value is not an array, or the array contains at least one invalid (not a string of nonzero length) tagtry
input should be a non-negative integer
ResourceNotFound
The specified job does not exist
try
inputT
is specified, but there is no tryT
forjob-xxxx.
Also returned iftry
input was specified for jobs in root executions created before July 12, 2023 00:13 UTC
PermissionDenied
CONTRIBUTE access is required for the job's project context; otherwise, the request can also be made by jobs sharing the same workspace as the specified job or the same workspace as the parent job of the specified job
API Method: /job-xxxx/setProperties
/job-xxxx/setProperties
Specification
Sets properties on the specified job. To remove a property altogether, its value needs to be set to the JSON null (instead of a string). This call updates the properties of the job by merging any old (previously existing) ones with what is provided in the input, the newer ones taking precedence when the same key appears in the old.
Best practices: to completely "reset" properties (i.e. remove all existing key/value pairs and replace them with some new), issue a describe call to get the names of all properties, then issue a setProperties request to set the values of those properties to null.
Inputs
properties
mapping Properties to modifykey Name of property to modify
value string or null Either a new string value for the
property, or null to unset the property
try
non-negative integer (optional). Specifies a particular try of a restarted job. Value of 0 refers to the first try. Defaults to the latest try (i.e. the try with the largesttry
attribute) for the specified job ID. See Restartable Jobs section for details.
Outputs
id
string ID of the manipulated job
Errors
InvalidInput
The input is not a hash,
properties
is missing or is not a hash, or there exists at least one value inproperties
which is neither a string nor the JSON null.try
input should be a non-negative integer.
ResourceNotFound
The specified job does not exist
try
inputT
is specified, but there is no tryT
forjob-xxxx.
Also returned iftry
input was specified for jobs in root executions created before July 12, 2023 00:13 UTC.
PermissionDenied
CONTRIBUTE access is required for the job's project context; otherwise, the request can also be made by jobs sharing the same workspace as the specified job or the same workspace as the parent job of the specified job.
API Method: /job-xxxx/terminate
/job-xxxx/terminate
Specification
Terminates a job and its job tree. If the job is already in a terminal state such as "terminated", "failed", or "done", no action will be taken. Otherwise, all jobs left in the job tree that have not reached a terminal state will eventually be put into the "terminated" state with failure reason "Terminated". Any authentication tokens generated for this execution will be invalidated, and any running jobs will be stopped. See Job Lifecycle for more details on job states.
Jobs can only be terminated by the user who launched the job or by any user with ADMINISTER access to the project context.
Inputs
None
Outputs
id
string ID of the terminated job (i.e., the string "job-xxxx")
Errors
ResourceNotFound
The specified object does not exist
PermissionDenied
Either (1) the user must match the "launchedBy" entry of the job object, and CONTRIBUTE access is required to the project context of the job, or (2) ADMINISTER access is required to the project context of the job
API Method: /job-xxxx/getIdentityToken
/job-xxxx/getIdentityToken
Specification
Get a signed DNAnexus JSON Web Token (JWT) that establishes a security-hardened and verifiable identity linked to the specific DNAnexus jobs. This job identity token can be provided to a 3rd party service (such as AWS), which will validate the job identity token and exchange it for a temporary access credentials (e.g. AWS token) allowing job code to securely access specific 3rd party resources, such as external storage buckets, external lambda functions, external databases, or external secrets vaults. See Job Identity Tokens for Access to Clouds and 3rd-Party Services for details.
This API method must be called from a DNAnexus job with the DNAnexus job token corresponding to the job.
Internally DNAnexus uses the RSASSA_PSS_SHA_256
algorithm to sign the retrieved JWT. This algorithm follows the following spec:
Algorithm
Algorithm description
RSASSA_PSS_SHA_256
PKCS #1 v2.2, Section 8.2, RSA signature with PKCS #1v1.5 Padding and SHA-256
Inputs
audience
string of between 1 and 255 characters containing only alphanumeric characters and.
(period),_
(underscore) and-
(dash) characters. User can specify the intended audience claim from the token. This value corresponds to the audience IdP setting in the 3rd party service such as AWS.subject_claims
non-empty array of strings An array of unique, valid DNAnexus claims that will be joined together to overwrite the default "sub" claim. The default value is["launched_by", "job_worker_ipv4"]
Valid claims:'job_id'
'root_execution_id'
'root_executable_id'
'root_executable_name'
'root_executable_version'
'executable_id'
'app_name'
'app_version'
'project_id'
'bill_to'
'launched_by'
'region'
'job_worker_ipv4'
'job_try'
Outputs
This method returns a JSON object with a Token
field containing a signed JSON Web Token (JWT), with standard and custom claims, represented as a JSON string.
Errors
InvalidInput
job-xxxx/getIdentityToken
is missing parameteraudience
audience input expected to be a non empty string between 1 and 255 characters containing only alphanumeric or
'.', '-', '_'
charactersExpected
subject_claims
to be a nonempty array of stringsThe
subject_claims
array contains at least one invalid claimClaim
'<input_claim>'
is expected to only occur at most once insubject_claims
PermissionDenied
job-xxxx/getIdentityToken can only be called with a job token
This error is returned if this API method is called with a non-job DNAnexus token
InvalidAuthentication
The token could not be found
This error is returned if this API method is called with an invalid DNAnexus token
The token cannot be used from this IP address
This error is returned if this API method is called with a DNAnexus job token corresponding to a different DNAnexus job.
Last updated