I noticed that in the recurring scheduled test the settings do not persist if we select " Only on error..." .
It defaults back to "Always". I would expect it to remember the original setting.
Currently WPF tests always require an application to be configured and launched upon test start. So, one need to always use a dummy app when using the 'Connect to app' and 'Launch app' steps.
Enable the WPF test configuration with the option to not launch an application - similar to the configuration experience for Desktop test.
Additional details shared internally!!!
A workaround is to manually refresh the DOM tree so that it gets updated to the current state of the page.
During a recording session the refresh DOM option is available in the Advanced Recording Tools window -> DOM Explorer tab -> Toolbar options -> Refresh button.
During test execution the refresh DOM option is available in a coded step - Manager.ActiveBrowser.RefreshDomTree();
While researching the case we also noticed that after some delay the DOM tree is refreshed automatically, so a delay step is also an option to be tried.
when I open TS -> Results ->
lets say it's Wednesday, a noon, and I have daily runs. I see some already executed - marked as green or red, and some yellow tiles, from the future. When I try to check what is the PC node name where a single task (list) will be executed, I don't see that info on a tile directly. Moreover, when I go and edit it, I see settings for a frequency and time, in next step I see what is the name of list selected, but on the very next step, where all nodes are listed, nothing is selected. That looks like a bug. So, if I have 20 things scheduled, I need to rely on my own notes to be sure I edit correct task. The name of the PC is available only for already executed task. No idea why it's not visible for future runs.
1. Record a test with elements that weren't in the repository
2. Move the element via Drag & Drop into an if-else case
3. Rename element in the TreeView on the left hand side
Normally the element name gets refreshed in the project and shows the correct name.
In an if-else case the name won't be refreshed.
Screenshot:
1. start recording
2. enable hover over highlighting
3. switch window with [ALT]+[TAB]
After switching from WPF-Window into Telerik Teststudio the hover over highlighting is still active
I need to export the contents of our test lists to a CSV, TXT or Excel file. There is no option other than exporting the generated results from a test list run.
However I need to be able to export the tests in a test list before I get to the point of executing these.
There is a dialogbox in WPF app to which the recorder toolbar gets attached. When the highlighting is enabled for this window no of the elements get highlighted.
The workaround is to use the DOM explorer tab in the Advanced Recording Tools to locate the desired element and build the desired step that way.
Details for the dialogbox and app shared internally!
There is no option in Test Studio recording capabilities to add a step which sets value for the WPF RadSlider control. It will be useful to have such similar to the WPF slider control.
The workaround is to set this in a coded step like this:
// Accepts values from 0 to 1
Applications.SliderTestexe.MainWindow.Item0Radslider.Value = 0.25;
Although the actual image files are with good resolution and are smooth, when displayed in the Failure Details Images tab they are blurry.
It will be great to display the images better.
Steps to reproduce:
Expected: The references are added with relative path starting under the project root.
Actual: The project root is included in the relative path.
Hello Progress Support,
I wonder if we can have "Bind data" applied on the "String Length" in "Generate random string" step?
Thank you for your help.
Regards,
Lisa
Hello,
Currently when I create a Git repo and connect Test Studio it creates default branch called "master".
Would it be possible to make it consistent with good practice and Git standard and rename it to "main" in the next version of Test Studio?
Can this be added as feature request please.
Thank you,
Max
When generating videos for the test runs from a test list, the output video files uses the name of the test only. There is no indication which is the test list from which this test was executed and when having multiple runs and videos it is difficult to relate these with the generated results.
It will be useful to generate the names of the videos from a test list in a way to correspond to the test list name and particular run.
Currently the exported result contains extended details only for the failed steps. If there is a warning for a step - like the warning that the element was found only by image, this can be only seen in the Test Studio result file.
Extending the HTML exported result to show the step warning details will be very useful for anyone who review this type of result (attached in an email after a scheduled run, for example).