A picture is worth a thousand words...
Test Studio Welcome screen crashes in specific customer environment.
When starting a project from the Settings.aiis file, it loads correctly, but triggering the Welcome screen results in unexpected error.
Allow the ArtOfTest.Runner to publish the results to the Executive Dashboard. There they can be seen by members of the team, who do not have Test Studio installed.
Make it possible for the ArtOfTest.Runner to use multiple execution machines. Distribute the test list execution to other execution servers.
Currently the exported result contains extended details only for the failed steps. If there is a warning for a step - like the warning that the element was found only by image, this can be only seen in the Test Studio result file.
Extending the HTML exported result to show the step warning details will be very useful for anyone who review this type of result (attached in an email after a scheduled run, for example).
We have a large PDF (60+ pages) and try to verify certain content on specific pages. During the recording process some elements are captured with TagName and TextContent and others with TagIndex. In this scenario TagIndex is not reliably, because the document is large and the elements change dynamicaly.
One solution is to update the element's find logic manually on all elements that use TagIndex, but this is very time consuming.
Another option is to use coded step to find the target element by TextContent and verify it.
When the Scheduling server is not added during installation, the Configure button in Test Studio will be greyed out (disabled). It will be helpful to add some kind of tool tip or note about this.
The solution is to modify the existing installation of Test Studio and add the Scheduling Server feature to the installation.
Allow users to add custom arguments when executing tests against any of the supported browsers.
This includes the Chrome Headless execution, which always defaults to en-EN language argument for example and you can't change it.
When generating videos for the test runs from a test list, the output video files uses the name of the test only. There is no indication which is the test list from which this test was executed and when having multiple runs and videos it is difficult to relate these with the generated results.
It will be useful to generate the names of the videos from a test list in a way to correspond to the test list name and particular run.
We need to be able to verify the text in dialogs in different browsers.
The current solution is in code, but is not stable due to browser structure changes -> https://docs.telerik.com/teststudio/advanced-topics/coded-samples/html/verify-dialog-text-chrome
It will be helpful to have an element, which is defined based on another element. This is similar to chained find expression, but instead of defining the parent element with its find expression, add the ability to reference an already recorded element in the project as a parent one. Then add find expression for the child element.
As a result it seems that the translators for Kendo Angular Input and TextBox components are, currently, not backwards compatible with the previous versions of these controls.
It will be great to fix these so that the latest Test Studio can be used for testing web pages built with previous versions of Kendo Angular controls.
A specific coded steps test cause Chrome to be closed when trying to perform a partial run using Run->To Here.
The coded step before the navigate one is starting a proxy to log the traffic.
Sample project is provided internally.
Http requests' responses with content-encoding type 'br' cannot be decoded in Test Studio load testing. This results in the inability of using these responses to generate dynamic targets to cover the proper load test run.
Workaround: Modify the traffic for a load user profile by removing the 'br' encoding type. A third party Chrome extension can be used for this modification.
Test Config Scenario (sample proj attached):
Test A (Maintest) calls Test B
Test B (filled with local Data) calls Test C
Test C (inheritParentData, should get it's data from Test A) calls Test D
Test D (inheritParentData, should get it's data from Test A as well)
Expected: Upon execution Test B executes all its iterations correctly for each iteration of Test A.
Actual: Test B executes all its iterations, but uses only the first row value for each of the iterations.
Note: If Test C is excluded from the test execution, all iterations from test B are performed correctly.
If Test C and Test D are using their own data instead of inheriting the parent source, the execution is as well as expected.
When one need to reconfigure a very high volume of Custom Dynamic Targets regularly in different user profiles, the GUI approach is very slow and inconvenient.
Would it be possible to implement a feature wherein a file can be uploaded to a User Profile in order to automatically configure Custom Dynamic Targets that are used frequently? Or somehow transfer the desired set of targets from one profile to another test or profile?