Steps to reproduce: 1. Add a data source to a test project which contains more than one row. 2. Bind a test to the data source. 3. Sort the rows by clicking the column header. 4. Enable Filtering and choose the second row (2-2) for example. Actual: The second row before sorting is selected. Expected: The second row after sorting should be selected.
Sample date picker used here.
Steps to reproduce:
Expected: To record a step to 'Enter Text'
Actual: A 'Click' step gets recorded and no date entered is registered in the test.
Workaround:
Steps to reproduce: 1. A demo application is provided 2. Record a test against it The application crashes after the available button is clicked.
It would be useful to have a method where you could specify global/environmental variables/parameters that is separate and independent from the data driven feature. These global parameters would be avalaible to all tests using the Data Driven interface of binding values to test steps, but would not be mingled in the data source itself. This way you can create a data driven test that runs iterations but keep the environmental static data e.g. userid's to login as, separate from the data itself.
The test list events are fired in case of running on the scheduling server, however the problem occurs in case of running on the execution client setup (out of the server). Steps to reproduce: 1. Create your own execution extension. 2. Override OnBeforeTestlistStarted and OnAfterTestListCompleted. 3. Create a test list and run it remotely. Expected: OnBeforeTestListStarted and OnAfterTestListCompleted are called. Actual: OnBeforeTestListStarted and OnAfterTestListCompleted aren't called.
In two attached test projects TestProjectForTelerik_BigReport and TestProjectForTelerik_SmallReport the same test actions are executed (tests set is the same too but there are some differences in test configurations). 1. Load www.telerik.com page 2. Come back to the main page 3. Click on a link to the products. The link is found by content (product's name) from data source (TelerikProducts.xls file, Products sheet) But different ways are used to run tests in the projects. In TestProjectForTelerik_SmallReport we start from SmallReportTest (use SmallReport test list from the project) and bind ClickProduct test to the data (the data source) directly. In TestProjectForTelerik_BigReport we start from BigReportTest (use BigReport test list from the project) and bind IntermediateStep test to the data directly. Data binding is removed from ClickProduct test but this one must inherit the parent data source. In IntermediateStep test we choose which test will be run next according to data from data source. In IntermediateStep we always run ClickProduct test. Please see comments in the IntermediateStep test for more information. This way is very needed for my real project. So, these two projects have very different report size for the small report it is 249674 bytes, for another 989670 bytes. The difference is about 4 times. If steps count in ClickProduct test or row count in data source increases, size difference will be more. As I can see in the report from TestProjectForTelerik_BigReport project, there are a lot of duplicated information in the report. The information from StepResults section of ClickProduct test is duplicated in DataIterationResults section of IntermediateStep test. Please check this. Maybe there is a way to decrease the report size. Resources shared internally
When DBMigrator is in progress all you see is "Please wait...". There's no indication how far it's gotten nor how much farther there is to go. The user can't tell if he needs to wait another 5 minutes or another 5 days, or even if anything is actually happening. Some sort of feedback needs to be shown the user indicating progress is actually being made by the migration process. In addition, when the process is completed, it would be useful to show something like "N records migrated" to indicate success or failure of the migration process. Instead currently when the process is finished the "Please wait..." is removed. The user can only assume the process was successful and how much data was migrated.
Displayu the full test list name/Execured browser in the tooltip in the same way as it is in the calendar view on results tab. Refer to the attachment provided internally
Test Studio removes elevated trust privileges from an OOB application when connected to the application for recording. Please refer to the screenshot attached from the local repro. The sample project is attached. The ArtOfTest.SLExtension.dll is included in the SL application.
Capture the URL of the current browser page when creating a feedback item. Include the full URL with query & fragments.
The Send dialog should close automatically after submitting a bug or mailing feedback. Generally people won't be using more than one action, so let's eliminate the extra step of making them close another dialog.
Some customers display reports using the XAML RadPdfViewer control. They would like to be able to parse the content displayed in it and verify the application generated reports are correct. Currently this can't be done because there seems to be no way to extract the various text fields that do get displayed. Every single letter is a separate Glyph element and you can't even get at which letter the Glyph is representing.
Add Test Studio tests to Visual Studio 2012 Update 2 test playlists.
The test freezes when trying to perform multiple actions in parallel. Here's a sample code: Parallel.Invoke ( () => {Actions.SetText(Pages.ClientLogIn.GeneralBoxContentUcClientRegistrationScetbLastNameTbSanitizedText, "test");}, () => {Actions.SetText(Pages.ClientLogIn.GeneralBoxContentUcClientRegistrationScetbFirstNameTbSanitizedText, "test");} ); If we comment the first action, the second will run just fine. To reproduce: Sample project is attached to the internal description.
Include a way to mark a Test Studio test with a Visual Studio 2012 test trait to allow test discovery in the VS Test Explorer.
TTS opens and closes browser even though command line test does not require any browser interaction. It increases test execution time. Add n option not to launch browser would be helpful to speed up the test execution and save processing power.
Having the ability to move an automated test over to a manual one. Would also be nice to export the automated tests to excel.
We have no need ever for the Team Pulse docked window. It would be nice to be able to keep this optional docking window from displaying at all.