Currently, when a step in the API test fails, the whole test is stopped.
Please add the continue on failure option for steps in API project.
The ScrollToVisible option is not available for some of the actions that use translators. One of them is the RadSearchBox, where this property is not available.
Currently the workaround is to add an Enter text step against the <input> element. To do that, you need to use the highlighter and record any verification step. Then add the Enter text step from the step builder, enable SimulateRealTyping and change the ScrollToVisible option.
The Keyboard (KeyPress) test step is missing the option to change the ScrollToVisibleType in the properties pane. Although the step actually performs the scroll to visible action and I see the code, when I select Edit in Code.
Please add the ScrollToVisibleType option.
It will be useful to improve the output data in API test results. Currently the generated results files from an API test run do not contain data for the actual sent and received requests.
The available workaround is to use Fiddler and capture the generated traffic during the API test run.
It would be nice if there was a way to avoid simulating real typing for a search box. There is no option to enable/disable it, but it is clearly using this behavior. I've tried a workaround of entering text directly in the input element, but it doesn't seem to register it when this technique is used. I don't see how a textbox can be made to work without this behavior but a search box cannot.
Simulating real typing tends to be the most fragile part of our tests, and all we really need is to enter text and then search. It also slows down the tests quite a bit vs. just setting the text directly.
Currently when a test step is set to SimulateRealClick or SimulateRealTyping, the target element is unconditionally scrolled to the top of the browser window, even in cases where the element is already visible. It would be more efficient to test if the target element is already fully visible and only scroll when needed.
Tests that have Not Completed tests, shouldn't list the overall test as Pass or Fail....rather it should set the overall run as "not complete"
In addition - when loading not completed tests within test runs, they are just shown as fail with no background and white text. The run is listed as pass if there are no fails but existing "not completes".
Recommendation to add Not Complete as overall result in addition to individual tests that are no in a passed or failed state
Currently Test Studio updates only the necessary tests and resources to the storage server. When the test execution starts on the execution server, Test Studio will download everything available in the storage server.
Please check how this can be optimized to ensure that there are all the necessary resources for the test execution.
Really excited to have a dashboard to present to the team now
Is there anyway to configure the number of Tests that display from within a result? Right now it only looks like 5 is what is the default, that requires quite a big of pagination.
Conditional statements (if, while, and such) should be able to use an extracted values as a condition.
Also multiple conditions will be very useful in some cases.
I think it would be neat to be able to view / edit markdown files within a project within the Test Studio application. Being able to see a markdown file in the project list, and have them open in a editor tab so you can review / make edits.
For example I create markdown files that have high level summary's of what each test completes, or potential helpful debugging information on a particular test.
1. Open the attached aiiresult file Expected: To load the results in the Results viewer Actual: The results viewer crashes Resources shared internally.
In two attached test projects TestProjectForTelerik_BigReport and TestProjectForTelerik_SmallReport the same test actions are executed (tests set is the same too but there are some differences in test configurations). 1. Load www.telerik.com page 2. Come back to the main page 3. Click on a link to the products. The link is found by content (product's name) from data source (TelerikProducts.xls file, Products sheet) But different ways are used to run tests in the projects. In TestProjectForTelerik_SmallReport we start from SmallReportTest (use SmallReport test list from the project) and bind ClickProduct test to the data (the data source) directly. In TestProjectForTelerik_BigReport we start from BigReportTest (use BigReport test list from the project) and bind IntermediateStep test to the data directly. Data binding is removed from ClickProduct test but this one must inherit the parent data source. In IntermediateStep test we choose which test will be run next according to data from data source. In IntermediateStep we always run ClickProduct test. Please see comments in the IntermediateStep test for more information. This way is very needed for my real project. So, these two projects have very different report size for the small report it is 249674 bytes, for another 989670 bytes. The difference is about 4 times. If steps count in ClickProduct test or row count in data source increases, size difference will be more. As I can see in the report from TestProjectForTelerik_BigReport project, there are a lot of duplicated information in the report. The information from StepResults section of ClickProduct test is duplicated in DataIterationResults section of IntermediateStep test. Please check this. Maybe there is a way to decrease the report size. Resources shared internally