库 CLI(旧版)

重要

本文档已过时,将来可能不会更新。

此信息适用于旧版 Databricks CLI 0.18 及更低版本。 Databricks 建议改用较新的 Databricks CLI 0.205 或更高版本。 请参阅什么是 Databricks CLI?。 若要查找你的 Databricks CLI 的版本,请运行 databricks -v

要从 Databricks CLI 0.18 或更低版本迁移到 Databricks CLI 0.205 或更高版本,请参阅 Databricks CLI 迁移

可以通过将 Databricks 库 CLI 子命令追加到 databricks libraries 之后来运行这些命令。 这些子命令调用库 API

databricks libraries -h
Usage: databricks libraries [OPTIONS] COMMAND [ARGS]...

  Utility to interact with libraries.

Options:
  -v, --version  [VERSION]
  -h, --help     Show this message and exit.

Commands:
  all-cluster-statuses  Get the status of all libraries.
  cluster-status        Get the status of all libraries for a cluster.
    Options:
      --cluster-id CLUSTER_ID   Can be found in the URL at https://<databricks-instance>/?o=<16-digit-number>#/setting/clusters/$CLUSTER_ID/configuration.
  install               Install a library on a cluster.
    Options:
      --cluster-id CLUSTER_ID   Can be found in the URL at https://<databricks-instance>/?o=<16-digit-number>#/setting/clusters/$CLUSTER_ID/configuration.
      --jar TEXT                JAR on DBFS or WASB.
      --egg TEXT                Egg on DBFS or WASB.
      --whl TEXT                Wheel or zipped wheelhouse on DBFS or WASB. Supported in CLI 0.8.2 and above.
      --maven-coordinates TEXT  Maven coordinates in the form of GroupId:ArtifactId:Version (i.e.org.jsoup:jsoup:1.7.2).
      --maven-repo TEXT         Maven repository to install the Maven package from. If omitted, both Maven Repository and Spark Packages are searched.
      --maven-exclusion TEXT    List of dependences to exclude. For example: --maven-exclusion "slf4j:slf4j" --maven-exclusion "*:hadoop-client".
      --pypi-package TEXT       The name of the PyPI package to install. An optional exact version specification is also supported. Examples "simplejson" and "simplejson==3.8.0".
      --pypi-repo TEXT          The repository where the package can be found. If not specified, the default pip index is used.
      --cran-package TEXT       The name of the CRAN package to install.
      --cran-repo TEXT          The repository where the package can be found. If not specified, the default CRAN repo is used.
  list                  Shortcut to `all-cluster-statuses` or `cluster-status`.
    Options:
      --cluster-id CLUSTER_ID   Can be found in the URL at https://<databricks-instance>/?o=<16-digit-number>#/setting/clusters/$CLUSTER_ID/configuration.
  uninstall             Uninstall a library on a cluster.
    Options:
      --cluster-id CLUSTER_ID   Can be found in the URL at https://<databricks-instance>/?o=<16-digit-number>#/setting/clusters/$CLUSTER_ID/configuration. [required]
      --all                     Uninstall all libraries.
      --jar TEXT                JAR on DBFS or WASB.
      --egg TEXT                Egg on DBFS or WASB.
      --whl TEXT                Wheel or zipped wheelhouse on DBFS or WASB. Supported in CLI 0.8.2 and above.
      --maven-coordinates TEXT  Maven coordinates in the form of GroupId:ArtifactId:Version (i.e.org.jsoup:jsoup:1.7.2).
      --maven-repo TEXT         Maven repository to install the Maven package from. If omitted, both Maven Repository and Spark Packages are searched.
      --maven-exclusion TEXT    List of dependences to exclude. For example: --maven-exclusion "slf4j:slf4j" --maven-exclusion "*:hadoop-client".
      --pypi-package TEXT       The name of the PyPI package to install. An optional exact version specification is also supported. Examples "simplejson" and "simplejson==3.8.0".
      --pypi-repo TEXT          The repository where the package can be found. If not specified, the default pip index is used.
      --cran-package TEXT       The name of the CRAN package to install.
      --cran-repo TEXT          The repository where the package can be found. If not specified, the default CRAN repo is used.

列出所有群集上所有库的状态。

若要显示使用情况文档,请运行 databricks libraries all-cluster-statuses --help

databricks libraries all-cluster-statuses
{
  "statuses": [
    {
      "cluster_id": "1234-567890-lest123",
      "library_statuses": [
        {
          "library": {
            "jar": "dbfs:/FileStore/jars/bbf81650_a62b_4b7a_b47e_7bdd9505792a-SparkJDBC42.jar"
          },
          "status": "INSTALLED",
          "is_library_for_all_clusters": true
        },
        ...
      ]
    },
    ...
  ]
}

列出某个群集上所有库的状态。

若要显示使用情况文档,请运行 databricks libraries cluster-status --helpdatabricks libraries list --help

databricks libraries cluster-status --cluster-id 1234-567890-lest123

或:

databricks libraries list --cluster-id 1234-567890-lest123
{
  "cluster_id": "1234-567890-lest123",
  "library_statuses": [
    {
      "library": {
        "jar": "dbfs:/FileStore/jars/bbf81650_a62b_4b7a_b47e_7bdd9505792a-SparkJDBC42.jar"
      },
      "status": "INSTALLED",
      "is_library_for_all_clusters": false
    },
    ...
  ]
}

在群集上安装库

若要显示使用情况文档,请运行 databricks libraries install --help

databricks libraries install --cluster-id 1234-567890-lest123 --jar dbfs:/test-dir/test.jar

如果成功,则不显示任何输出。

从群集中卸载库

若要显示使用情况文档,请运行 databricks libraries uninstall --help

databricks libraries uninstall --cluster-id 1234-567890-lest123 --jar dbfs:/test-dir/test.jar
WARNING: Uninstalling libraries requires a cluster restart.
databricks clusters restart --cluster-id 1234-567890-lest123