Skip to content

salesforcecli/plugin-agent

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

plugin-agent - PREVIEW

THIS PLUGIN IS A PREVIEW VERSION AND IS NOT MEANT FOR PRODUCTION USAGE UNTIL ANNOUNCED.

NPM Downloads/week License

Install

sf plugins install @salesforce/[email protected]

Contributing

  1. Please read our Code of Conduct
  2. Create a new issue before starting your project so that we can keep track of what you are trying to add/fix. That way, we can also offer suggestions or let you know if there is already an effort in progress.
  3. Fork this repository.
  4. Build the plugin locally
  5. Create a topic branch in your fork. Note, this step is recommended but technically not required if contributing using a fork.
  6. Edit the code in your fork.
  7. Write appropriate tests for your changes. Try to achieve at least 95% code coverage on any new code. No pull request will be accepted without unit tests.
  8. Sign CLA (see CLA below).
  9. Send us a pull request when you are done. We'll review your code, suggest any needed changes, and merge it in.

CLA

External contributors will be required to sign a Contributor's License Agreement. You can do so by going to https://cla.salesforce.com/sign-cla.

Build

To build the plugin locally, make sure to have yarn installed and run the following commands:

# Clone the repository
git clone [email protected]:salesforcecli/plugin-agent

# Install the dependencies and compile
yarn && yarn build

To use your plugin, run using the local ./bin/dev.js or ./bin/dev.cmd file.

# Run using local run file.
./bin/dev.js agent

There should be no differences when running via the Salesforce CLI or using the local run file. However, it can be useful to link the plugin to do some additional testing or run your commands from anywhere on your machine.

# Link your plugin to the sf cli
sf plugins link .
# To verify
sf plugins

Commands

sf agent create

Create an agent in your org from a local agent spec file.

USAGE
  $ sf agent create -o <value> -f <value> -n <value> [--json] [--flags-dir <value>] [--api-version <value>]

FLAGS
  -f, --spec=<value>         (required) Path to an agent spec file.
  -n, --name=<value>         (required) API name of the new agent.
  -o, --target-org=<value>   (required) Username or alias of the target org. Not required if the `target-org`
                             configuration variable is already set.
      --api-version=<value>  Override the api version used for api requests made by this command

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.
  --json               Format output as json.

DESCRIPTION
  Create an agent in your org from a local agent spec file.

  To generate an agent spec file, run the "agent generate spec" CLI command, which outputs a JSON file with the list of
  jobs and descriptions that the new agent can perform. Then specify this generated spec file to the --spec flag of this
  command, along with the name of the new agent.

  When this command finishes, your org contains the new agent, which you can then edit in the Agent Builder UI. The new
  agent already has a list of topics and actions that were automatically created from the list of jobs in the provided
  agent spec file. This command also retrieves all the metadata files associated with the new agent to your local DX
  project.

  To open the new agent in your org's Agent Builder UI, run this command: "sf org open agent --name
  <api-name-of-your-agent>".

EXAMPLES
  Create an agent called "CustomerSupportAgent" in an org with alias "my-org" using the specified agent spec file:

    $ sf agent create --name CustomerSupportAgent --spec ./config/agentSpec.json --target-org my-org

See code: src/commands/agent/create.ts

sf agent generate definition

Interactively generate a new AiEvaluationDefinition.

USAGE
  $ sf agent generate definition [--flags-dir <value>]

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.

DESCRIPTION
  Interactively generate a new AiEvaluationDefinition.

  This command will prompt you for the necessary information to create a new AiEvaluationDefinition. The definition will
  be saved to the `aiEvaluationDefinitions` directory in the project.

  You must have the `Bots` and `AiEvaluationTestSets` metadata types present in your project to use this command.

EXAMPLES
  $ sf agent generate definition

See code: src/commands/agent/generate/definition.ts

sf agent generate spec

Generate an agent spec, which is the list of jobs that the agent performs.

USAGE
  $ sf agent generate spec -o <value> [--json] [--flags-dir <value>] [--api-version <value>] [-t customer|internal]
    [--role <value>] [--company-name <value>] [--company-description <value>] [--company-website <value>] [-d <value>]
    [-f <value>]

FLAGS
  -d, --output-dir=<value>           [default: config] Directory where the agent spec file is written; can be an
                                     absolute or relative path.
  -f, --file-name=<value>            [default: agentSpec.json] Name of the generated agent spec file.
  -o, --target-org=<value>           (required) Username or alias of the target org. Not required if the `target-org`
                                     configuration variable is already set.
  -t, --type=<option>                Type of agent to create.
                                     <options: customer|internal>
      --api-version=<value>          Override the api version used for api requests made by this command
      --company-description=<value>  Description of your company.
      --company-name=<value>         Name of your company.
      --company-website=<value>      Website URL of your company.
      --role=<value>                 Role of the agent.

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.
  --json               Format output as json.

DESCRIPTION
  Generate an agent spec, which is the list of jobs that the agent performs.

  When using Salesforce CLI to create an agent in your org, the first step is to generate the local JSON-formatted agent
  spec file with this command.

  An agent spec is a list of jobs and descriptions that capture what the agent can do. Use flags such as --role and
  --company-description to provide details about your company and the role that the agent plays in your company; you can
  also enter the information interactively if you prefer. When you then execute this command, the large language model
  (LLM) associated with your org uses the information to generate the list of jobs that the agent most likely performs.
  We recommend that you provide good details for --role, --company-description, etc, so that the LLM can generate the
  best and most relevant list of jobs and descriptions. Once generated, you can edit the spec file; for example, you can
  remove jobs that don't apply to your agent.

  When your agent spec is ready, you then create the agent in your org by specifying the agent spec file to the
  --job-spec flag of the "agent create" CLI command.

EXAMPLES
  Create an agent spec for your default org in the default location and use flags to specify the agent's role and your
  company details:

    $ sf agent generate spec --type customer --role "Assist users in navigating and managing bookings" \
      --company-name "Coral Cloud" --company-description "Resort that manages guests and their reservations and \
      experiences"

  Create an agent spec by being prompted for role and company details interactively; write the generated file to the
  "specs" directory and use the org with alias "my-org":

    $ sf agent generate spec --output-dir specs --target-org my-org

See code: src/commands/agent/generate/spec.ts

sf agent generate testset

Interactively generate an AiEvaluationTestSet.

USAGE
  $ sf agent generate testset [--flags-dir <value>]

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.

DESCRIPTION
  Interactively generate an AiEvaluationTestSet.

  Answer the prompts to generate an AiEvaluationTestSet that will be written to a file. You can then run "sf agent
  generate definition" to generate the AiEvaluationDefinition that can be used to evaluate the test set.

EXAMPLES
  $ sf agent generate testset

See code: src/commands/agent/generate/testset.ts

sf agent preview

Interact with an active agent, as a user would, to preview responses

USAGE
  $ sf agent preview -o <value> -n <value> [--flags-dir <value>] [--api-version <value>]

FLAGS
  -n, --name=<value>         (required) The name of the agent you want to preview
  -o, --target-org=<value>   (required) Username or alias of the target org. Not required if the `target-org`
                             configuration variable is already set.
      --api-version=<value>  Override the api version used for api requests made by this command

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.

DESCRIPTION
  Interact with an active agent, as a user would, to preview responses

  XXX

EXAMPLES
  $ sf agent preview --agent HelpDeskAgent

  $ sf agent preview --agent ConciergeAgent --target-org production

FLAG DESCRIPTIONS
  -n, --name=<value>  The name of the agent you want to preview

    the API name of the agent? (TBD based on agents library)

See code: src/commands/agent/preview.ts

sf agent test cancel

Cancel an agent test that's currently running in your org.

USAGE
  $ sf agent test cancel -o <value> [--json] [--flags-dir <value>] [--api-version <value>] [-i <value>] [-r]

FLAGS
  -i, --job-id=<value>       Job ID of the running agent test that you want to cancel.
  -o, --target-org=<value>   (required) Username or alias of the target org. Not required if the `target-org`
                             configuration variable is already set.
  -r, --use-most-recent      Use the job ID of the most recently-run agent test.
      --api-version=<value>  Override the api version used for api requests made by this command

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.
  --json               Format output as json.

DESCRIPTION
  Cancel an agent test that's currently running in your org.

  This command requires a job ID, which the original "agent test run" command displays when it completes. You can also
  use the --use-most-recent flag to see results for the most recently run agent test.

EXAMPLES
  Cancel an agent test currently running in your default org using a job ID:

    $ sf agent test cancel --job-id 4KBfake0000003F4AQ

  Cancel the most recently run agent test in the org with alias "my-org":

    $ sf agent test cancel --job-id 4KBfake0000003F4AQ --target-org my-org

See code: src/commands/agent/test/cancel.ts

sf agent test results

Get the results of a completed agent test run.

USAGE
  $ sf agent test results -o <value> -i <value> [--json] [--flags-dir <value>] [--api-version <value>] [--result-format
    json|human|junit|tap] [-f <value>]

FLAGS
  -f, --output-dir=<value>      Directory to write the agent test results into.
  -i, --job-id=<value>          (required) Job ID of the completed agent test run.
  -o, --target-org=<value>      (required) Username or alias of the target org. Not required if the `target-org`
                                configuration variable is already set.
      --api-version=<value>     Override the api version used for api requests made by this command
      --result-format=<option>  [default: human] Format of the agent test run results.
                                <options: json|human|junit|tap>

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.
  --json               Format output as json.

DESCRIPTION
  Get the results of a completed agent test run.

  This command requires a job ID, which the original "agent test run" command displays when it completes. You can also
  use the --use-most-recent flag to see results for the most recently run agent test.

  By default, this command outputs test results in human-readable tables for each test case. The tables show whether the
  test case passed, the expected and actual values, the test score, how long the test took, and more. Use the
  --result-format to display the test results in JSON or Junit format. Use the --output-dir flag to write the results to
  a file rather than to the terminal.

EXAMPLES
  Get the results of an agent test run in your default org using its job ID:

    $ sf agent test results --job-id 4KBfake0000003F4AQ

  Get the results of the most recently run agent test in an org with alias "my-org":

    $ sf agent test results --use-most-recent --target-org my-org

  Get the results of the most recently run agent test in your default org, and write the JSON-formatted results into a
  directory called "test-results":

    $ sf agent test results --use-most-recent --output-dir ./test-results --result-format json

FLAG DESCRIPTIONS
  -f, --output-dir=<value>  Directory to write the agent test results into.

    If the agent test run completes, write the results to the specified directory. If the test is still running, the
    test results aren't written.

See code: src/commands/agent/test/results.ts

sf agent test resume

Resume an agent test that you previously started in your org so you can view the test results.

USAGE
  $ sf agent test resume -o <value> [--json] [--flags-dir <value>] [--api-version <value>] [-i <value>] [-r] [-w
    <value>] [--result-format json|human|junit|tap] [-f <value>]

FLAGS
  -f, --output-dir=<value>      Directory to write the agent test results into.
  -i, --job-id=<value>          Job ID of the original agent test run.
  -o, --target-org=<value>      (required) Username or alias of the target org. Not required if the `target-org`
                                configuration variable is already set.
  -r, --use-most-recent         Use the job ID of the most recent agent test run.
  -w, --wait=<value>            [default: 5 minutes] Number of minutes to wait for the command to complete and display
                                results to the terminal window.
      --api-version=<value>     Override the api version used for api requests made by this command
      --result-format=<option>  [default: human] Format of the agent test run results.
                                <options: json|human|junit|tap>

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.
  --json               Format output as json.

DESCRIPTION
  Resume an agent test that you previously started in your org so you can view the test results.

  This command requires a job ID, which the original "agent test run" command displays when it completes. You can also
  use the --use-most-recent flag to see results for the most recently run agent test.

  Use the --wait flag to specify the number of minutes for this command to wait for the agent test to complete; if the
  test completes by the end of the wait time, the command displays the test results. If not, the CLI returns control of
  the terminal to you, and you must run "agent test resume" again.

  By default, this command outputs test results in human-readable tables for each test case. The tables show whether the
  test case passed, the expected and actual values, the test score, how long the test took, and more. Use the
  --result-format to display the test results in JSON or Junit format. Use the --output-dir flag to write the results to
  a file rather than to the terminal.

EXAMPLES
  Resume an agent test in your default org using a job ID:

    $ sf agent test resume --job-id 4KBfake0000003F4AQ

  Resume the most recently-run agent test in an org with alias "my-org" org; wait 10 minutes for the tests to finish:

    $ sf agent test resume --use-most-recent --wait 10 --target-org my-org

  Resume the most recent agent test in your default org, and write the JSON-formatted results into a directory called
  "test-results":

    $ sf agent test resume --use-most-recent --output-dir ./test-results --result-format json

FLAG DESCRIPTIONS
  -f, --output-dir=<value>  Directory to write the agent test results into.

    If the agent test run completes, write the results to the specified directory. If the test is still running, the
    test results aren't written.

See code: src/commands/agent/test/resume.ts

sf agent test run

Start an agent test in your org.

USAGE
  $ sf agent test run -o <value> -n <value> [--json] [--flags-dir <value>] [--api-version <value>] [-w <value>]
    [--result-format json|human|junit|tap] [-f <value>]

FLAGS
  -f, --output-dir=<value>      Directory to write the agent test results into.
  -n, --name=<value>            (required) Name of the agent test to start.
  -o, --target-org=<value>      (required) Username or alias of the target org. Not required if the `target-org`
                                configuration variable is already set.
  -w, --wait=<value>            Number of minutes to wait for the command to complete and display results to the
                                terminal window.
      --api-version=<value>     Override the api version used for api requests made by this command
      --result-format=<option>  [default: human] Format of the agent test run results.
                                <options: json|human|junit|tap>

GLOBAL FLAGS
  --flags-dir=<value>  Import flag values from a directory.
  --json               Format output as json.

DESCRIPTION
  Start an agent test in your org.

  Use the --name flag to specify the name of the agent test you want to run; find the agent test's name in the Testing
  Center page in the Setup UI of your org.

  This command starts the agent test in your org, but doesn't by default wait for it to finish. Instead, it displays the
  "agent test resume" command, with a job ID, that you execute to see the results of the test run, and then returns
  control of the terminal window to you. Use the --wait flag to specify the number of minutes for the command to wait
  for the agent test to complete; if the test completes by the end of the wait time, the command displays the test
  results. If not, run "agent test resume".

  By default, this command outputs test results in human-readable tables for each test case, if the test completes in
  time. The tables show whether the test case passed, the expected and actual values, the test score, how long the test
  took, and more. Use the --result-format to display the test results in JSON or Junit format. Use the --output-dir flag
  to write the results to a file rather than to the terminal.

EXAMPLES
  Start a test called MyAgentTest for an agent in your default org, don't wait for the test to finish:

    $ sf agent test run --name MyAgentTest

  Start a test for an agent in an org with alias "my-org" and wait for 10 minutes for the test to finish:

    $ sf agent test run --name MyAgentTest --wait 10 --target-org my-org

  Start a test and write the JSON-formatted results into a directory called "test-results":

    $ sf agent test run --name MyAgentTest --wait 10 --output-dir ./test-results --result-format json

FLAG DESCRIPTIONS
  -f, --output-dir=<value>  Directory to write the agent test results into.

    If the agent test run completes, write the results to the specified directory. If the test is still running, the
    test results aren't written.

See code: src/commands/agent/test/run.ts