Salto for
Salesforce
Articles
SHARE
Julian Joseph
February 25, 2025
5
min read
I started this experiment thinking, "How can this replace me?" or "How can I set this up so I never have to do these tasks again?" Instead, I came away with something completely different—an approach to understanding confusing parts of our product by watching AI attempt tasks and struggle.
I asked ChatGPT Operator to perform a Salesforce deployment using Salto. Here’s the exact prompt I used:
"Using Salto, create a deployment between Developer and Partner orgs. I don’t care what metadata is deployed."
With some guidance, it navigated Salto’s UI, set up the deployment, and even reported any errors back to me. But the real insight came from watching where it struggled.
✅ It technically worked, which is exciting.
✅ Operator generally knew when to stop and ask for important decisions.
✅ It understood Salesforce concepts at a high level and seemed to know where to look.
❌ It was very slow and didn't replace a human doing the same tasks.
❌ It could still make mistakes if the UI isn’t clear enough.
What stood out the most was not just how Operator interacted with Salto but what that revealed about our own product design. Watching an AI struggle through tasks provided an unexpected form of user research:
Rather than asking, "Can AI replace humans for these tasks?" a more valuable question might be:
This experiment didn’t convince me that AI is ready to automate deployments fully, but it did make me want to use Operator as a user research tool—a way to see Salto through fresh eyes and spot areas that could be improved.
By using AI not just for automation but as a diagnostic tool, we can make our products more intuitive for both humans and AI alike.
Salto for
Salesforce
Salesforce
SHARE
Julian Joseph
February 25, 2025
5
min read
I started this experiment thinking, "How can this replace me?" or "How can I set this up so I never have to do these tasks again?" Instead, I came away with something completely different—an approach to understanding confusing parts of our product by watching AI attempt tasks and struggle.
I asked ChatGPT Operator to perform a Salesforce deployment using Salto. Here’s the exact prompt I used:
"Using Salto, create a deployment between Developer and Partner orgs. I don’t care what metadata is deployed."
With some guidance, it navigated Salto’s UI, set up the deployment, and even reported any errors back to me. But the real insight came from watching where it struggled.
✅ It technically worked, which is exciting.
✅ Operator generally knew when to stop and ask for important decisions.
✅ It understood Salesforce concepts at a high level and seemed to know where to look.
❌ It was very slow and didn't replace a human doing the same tasks.
❌ It could still make mistakes if the UI isn’t clear enough.
What stood out the most was not just how Operator interacted with Salto but what that revealed about our own product design. Watching an AI struggle through tasks provided an unexpected form of user research:
Rather than asking, "Can AI replace humans for these tasks?" a more valuable question might be:
This experiment didn’t convince me that AI is ready to automate deployments fully, but it did make me want to use Operator as a user research tool—a way to see Salto through fresh eyes and spot areas that could be improved.
By using AI not just for automation but as a diagnostic tool, we can make our products more intuitive for both humans and AI alike.