Problem
While setting up a new job cluster, you indicate the default policy, Job Compute, and Single user under Access mode. You notice a tooltip in the UI saying you can select “single-node” from the compute mode while your indicated number of workers is set to 1 but cannot see how to select the single node cluster in the UI.
The following screenshot shows the UI with the tooltip after selecting 1 worker.
Cause
The Job Compute default policy is designed to create clusters with multiple workers.
Solution
Create a custom job policy that instructs Databricks to create a single-node cluster. Use the following configuration as a minimum. For details on creating a custom job policy, refer to the “Use policy families to create custom policies” section of the Default policies and policy families (AWS | Azure | GCP) documentation.
{
  "cluster_type": {
    "type": "fixed",
    "value": "job"
  },
  "num_workers": {
    "type": "fixed",
    "value": 0
  },
  "spark_conf.spark.databricks.cluster.profile": {
    "type": "fixed",
    "value": "singleNode"
  },
  "spark_conf.spark.master": {
    "type": "fixed",
    "value": "local[*,4]"
  },
  "custom_tags.ResourceClass": {
    "type": "fixed",
    "value": "SingleNode"
  }
}
Then, when you return to the job cluster setup UI view, you can indicate your custom policy, and see Single node selected as well.