Currently, when a step in the API test fails, the whole test is stopped.
Please add the continue on failure option for steps in API project.
The Keyboard (KeyPress) test step is missing the option to change the ScrollToVisibleType in the properties pane. Although the step actually performs the scroll to visible action and I see the code, when I select Edit in Code.
Please add the ScrollToVisibleType option.
It would be nice if there was a way to avoid simulating real typing for a search box. There is no option to enable/disable it, but it is clearly using this behavior. I've tried a workaround of entering text directly in the input element, but it doesn't seem to register it when this technique is used. I don't see how a textbox can be made to work without this behavior but a search box cannot.
Simulating real typing tends to be the most fragile part of our tests, and all we really need is to enter text and then search. It also slows down the tests quite a bit vs. just setting the text directly.
The ScrollToVisible option is not available for some of the actions that use translators. One of them is the RadSearchBox, where this property is not available.
Currently the workaround is to add an Enter text step against the <input> element. To do that, you need to use the highlighter and record any verification step. Then add the Enter text step from the step builder, enable SimulateRealTyping and change the ScrollToVisible option.
It will be useful to improve the output data in API test results. Currently the generated results files from an API test run do not contain data for the actual sent and received requests.
The available workaround is to use Fiddler and capture the generated traffic during the API test run.
Tests that have Not Completed tests, shouldn't list the overall test as Pass or Fail....rather it should set the overall run as "not complete"
In addition - when loading not completed tests within test runs, they are just shown as fail with no background and white text. The run is listed as pass if there are no fails but existing "not completes".
Recommendation to add Not Complete as overall result in addition to individual tests that are no in a passed or failed state
Currently Test Studio updates only the necessary tests and resources to the storage server. When the test execution starts on the execution server, Test Studio will download everything available in the storage server.
Please check how this can be optimized to ensure that there are all the necessary resources for the test execution.
Conditional statements (if, while, and such) should be able to use an extracted values as a condition.
Also multiple conditions will be very useful in some cases.
I think it would be neat to be able to view / edit markdown files within a project within the Test Studio application. Being able to see a markdown file in the project list, and have them open in a editor tab so you can review / make edits.
For example I create markdown files that have high level summary's of what each test completes, or potential helpful debugging information on a particular test.
Really excited to have a dashboard to present to the team now
Is there anyway to configure the number of Tests that display from within a result? Right now it only looks like 5 is what is the default, that requires quite a big of pagination.