I hope my email finds you well and safe.
I successfully managed to setup executive server.
This ticket is not an issue but a suggestion from a team member of my company.
On the executive dashboard page it would be nice to have a description to show what this test list do. Think that the description it will be something that i will write it somewhere and show it up on the executive dashboard.
Thank you for your support
Kyriakos, Prevention at Sea
The problem we're currently having is being able to leverage Test Studio in the same capacity of distributed testing through Azure DevOps which our IT organization uses for Continuous Integration.
Although the AOR-CLI does provide some benefits, it significantly is missing distributed testing which creates a large bottleneck in the turn-around times of our tests.
A potential way to solve this, as I've seen other testing platforms, is create a Test Studio Add-on for Azure DevOps in which you can handle calling your tests in the same fashion that you can from the Test Studio platform
Nested data driven tests generates quite large results, even when there is no data in the bound data table (built-in or external). The result is almost twice the size of the same tests, if these are not using the data tables. The tests may be using InheritParentDataSource, but the behavior is similar if these are not using it as well.
For real user scenarios, the large results may exceed the limitation of the file size (16MB), which can be transferred through the storage service.
Investigate if it is possible to optimize the data driven test results, at least for the case, when the data source is empty.
It would be nice to have the story board capture subtest screenshots instead of the "Test Step" placeholder. It would be fine maybe to watermark the screenshots with "Test Step," but still show the screens for reference.
Not a big deal, or anything requiring further support, but an observation worth noting nonetheless:
If a dynamic target is set between 2 steps and one such step is removed from HTTP traffic, rather than being able to edit the target to reflect a new destination step (or rather than the target information simply passing to step taking the place of the deleted step in the HTTP order), the target destination defaults to zero and any attempt to edit that traffic or the dynamic target further results in the application crashing.
Also of note, the reason that this has been an issue, is that deleted traffic seems to be reappearing on application close-->reopen, even when the traffic is saved (with or without the crashing behavior).
Hope this helps! Thanks.
ResultsViewer crashes when closing the Step Failure Details window with the OK button.
The version I am using is not listed below: 2020.1.403.0
I am noticing an inconsistency when using the Replace Element feature i.e. the attributes selected are not respecting the priority set in the Settings - Find Logic (Html) screen.
Please refer to screenshots attached.
See the screen shot.
There are two annotation markers. For some reason, a "MouseClick:LeftClick" will appear, as it should, but then it sticks there, even as 10 or 12 other steps execute and their annotations come and go. Sometimes another MouseClick:LeftClick will appear and the old one will go away and then the new one will stick. But this does not happen every times. Sometimes a new MouseClick:LeftClick annotation comes in and goes away and the old one is still hanging around.
This happens will almost all of my scripts.
Is it because some of the mouse clicks are from recorded steps and some are from code steps? This is just a guess.
On a data driven web test, setting the DataRange property to the example "SingleRow" causes Visual Studio 2017 to crash. I have three rows of data in a data table that are bound to specific steps of my test. I was attempting to limit execution to only one row temporarily. Using SingleRow or 'SingleRow' causes Visual Studio 2017 to crash with the following stack trace as found in Windows Event log. I was able to successfully use '1:1' or ':1' as a work around.Application: devenv.exe
Often in our web app a lot of logs, reports or forms are downloaded through Step Builder "Handle 'Download' dialog".
We would like to compare the downloaded file (xls, xlsx, pdf are mostly used) to a baseline file so that we can verify the downloaded file has the same content as expected result.
This definitely can be achieved through "Coded Step" but it will be great if there is a built in function to compare and verify file, similar to how to verify image.
Is it possible to have the Test List result in Excel format to be as detailed as HTML format?
Is there any particular reason for the results in Excel format to be so simplified?
Management level would prefer to be able to filter results conveniently in Excel meanwhile the results reporting in Excel is overly summarized.
It would be great if we are able to view which steps fail with details in Excel format too.
We are using telerik testing framework for Automation testing. Currently we are executing our scripts in a client VM using TFS builds. Now we would like transform to Docker Execution. We couldn't get the direct solution from any forum for Telerik Testing framework and Docker integration. So please help us on identifying the docker image for Telerik testing framework, so that we will make use of it and we will create a container for the same to run our test scripts in docker vm. Any websites/links related to this topic would be helpful for us.