How to change snapshots size in LoadRunner - snapshot

I'm writing a HTTP/HTML script on LoadRunner VuGen 11.50 and I need to inspect the results of a web_custom_request, but the snapshot created is chunked to 100 Kb. It says that "Snapshot Step data exceeds maximum size and cannot be fully displayed..."
Is there a way to modify this maximum size so I can read the complete response of the server?

Load Runner 11.5 VuGen Menu: Tools-> Options -> Scripting "Tab"
On the left pane click "Snapshot". The right pane will display the options available. Uncheck all the "Do not load snapshot larger than" .... options.

Change your log level in the run time settings to extended log, with appropriate sub-options. Then examine the log which has been generated when you run your virtual user inside of VUGEN.

Related

The window terminated unexpectedly [duplicate]

I am trying to open a folder that I opened before, but it crashed.
I can open other projects, and restarting the computer didn't help.
Maybe it's because I had a big file opened (400mb) in this folder, but I cant close this file because the vscode crashing every time when I tried open the workspace..
https://github.com/microsoft/vscode/issues/126127
https://github.com/microsoft/vscode/issues/130375
I had the same problem.
Just delete the folders under %appdata%/code/backups/ and restart VS.
I solved the problem by restarting my computer and then opening VS code from a different folder.
To open large JSON files, I use Dadroit JSON Viewer, thus preventing the problem from repeating itself.
Not worry much about that!, same issue occurred in my pc but there is easy steps to fix this error as following.
solution: Enable "paging files" option OR increase virtual memory.
steps:
1). Right click on "ThisPC" or "MyPC" and go to properties. then go to "Advanced System setting". [ you can also find that using SEARCH button ]
2). then click on settings of Performance block. its open another pop-up window in that ,go to the "advanced" option .
3). you can see the Virtual memory section ,in that click on change button.
4). now first disable the checkbox of "Automatically manage paging file size for all drive".
-> and select the "custom size:"
-> in Initial size(MB): { you need to add the same value as recommended, which is also displayed downside section }
-> in Maximum size(MB): you need to enter the value that you get multiply by 3 of your allocated RAM size.
for example:
your pc's ram is 4 gb then you can enter 4098*3= "12288mb".
then set it , click ok, ok, ok, and then just RESTART your pc. 🥹
hope god will help you.
It's easy than you think, you just need to update your laptop. and try to use Vs code again you will see that it works.

Unable to run all values from csv file while doing a performance test

Im using VS Web Performance Test Tool.
My Application has a login and few functionalities but doesnt have a logout . I one closes the browser and opens , it asks for login .
Now, i have 5 rows in excel , but only the first is run and it stops.
Please let me know if anyone knows how to make all rows run.
Thanks,
SK
By default, running a Web Performance Test (WPT) only does one execution of the test. A WPT can be called from a Visual Studio Load Test to run it many times by many virtual (i.e. simulated) users (VUs).
A data driven WPT can easily be run more than once by either of two methods.
After the test has been run once, click the "Edit run settings" on the Web Test results viewer. The modified run settings only apply to runs started by clicking the "Click here to run again" and are lost when the Web Test results viewer is closed. See here for more details.
The default can be changed in a .testsettings file. If no such file is already in the solution, then use the context (right click) menu of the solution in Solution Explorer and select Add => New item then follow the prompts to add a Test settings file. The Web test section of the file controls how WPTs are executed. See here for more details.

Visual Studio Load Test View Raw Results

I am running a Coded UI test using Visual Studio load testing capabilities.
My test runs successfully, but I am trying to get a better understanding of results. My results are stored as "All Individual Details" and are stored in a local database. However, I'd like to extract raw results for each individual transaction. All I can really see in the results is the number of times each transaction ran, the average response time and the min and max etc. I'd like to get hold of the raw results for each individual transaction however. Is this possible?
I'm using VS2012.
Visual Studio stores load test results in an SQL database, provided the "run settings" specify so. The contents of the database for Visual Studio 2010 is defined here. A few changes have been made for later versions but much remains the same.
Much of the contents of the database is derived from Web Performance tests but the question asks about load tests running Coded UI tests. Transactions and pages are features of Web Performance tests, whilst the author of a Coded UI test may think of transactions and pages, Coded UI has no inbuilt items that provide that data. Hence various parts of the database will be empty.
To find the connection string for the SQL load test database, open any load test file then from toolbar (or context menu) select "Manage Test Controllers". The connection string will be shown within the window.
to get the raw data do the following,
1.open the results
2. select the graph tab
3. Select the drop down, e.g. transaction response time
4. In the top right corner, select Export graph data to excel

Query runs fast but report renders slow: how to debug this?

I'm running SSRS 08. My query has four parameters (all varchars). I can run the query through management studio and it runs in under 10sec. I get the same performance if I run the query in Visual Studio's query designer.
However...if I try to preview the report or actually run the report once it's deployed the actual report sometimes takes more than 5 minutes. I ran a trace and the query seems to be in/out of the SQL server side quickly. How can I find where/why I'm loosing this performance? I've tried/checked so far the following:
No images are referenced in the report
No grouping/sorting outside of the stored procedure
Since it's a straight forward report I rebuilt it with same result (to check if it's not an issue of a corrupted report).
ran SQL trace when I executed the report to make sure the query did not have issues
Based on some suggestions before I rewrote the query to use variables instead of parameters like so
-- ...
-- Note: #Parameter is a varchar(40)
-- ...
declare #Var as varchar(40)
set #var=#parameter
select * from table where fieldvalue=#var
Ran the trace and it's not an issue on the query side.
I had the report html output trouble on report retrieving 32000 lines. In my case I had to activate “Interactive Paging” to allow user see first page and able to generate Excel file. The pro is that first page appears fast and user can generate export to Excel or PDF, the cons is that user can scroll only current page. If user wants to see more content he\she must use navigation buttons above the grid. In my case user accepted this behavior because the export to Excel was more important.
To activate “Interactive Paging” you must click on the free area in the report pane and change property “InteractiveSize”\ “Height” on the report level in Properties pane. Set this property to different from 0. I set to 8.5 inches in my case. Also ensure that you unchecked “Keep together on one page if possible” property on the Tablix level (right click on the Tablix, then “Tablix Properties”, then “General”\ “Page Break Options”).
I was troubleshooting a similar issue a while ago and it turned out to be related to the browser I was using to view the report. If the report returns a lot of data to the screen consider splitting it into multiple pages if it is not already configured that way.
Why Does Business Intelligence Studio Load Reports Faster than the Report Server
I think I may have found my answer on a social.msdn.microsoft.com thread
http://social.msdn.microsoft.com/Forums/en-US/sqlreportingservices/thread/1e34dc76-4d78-4376-89b0-7c381ed82726/
I unchecked the "Keep together" setting and it now actually previews coming from the server and I can export the report. Much improvement

Firefox error console buffer size is too small

How to change the default Firefox error console buffer size? The problem is that there isn't enough time to read messages: old ones get overridden by the newer messages.
I know this is an old one, but this came up as my top search and the setting can now be found in the about:config settings.
Type about:config into the address bar and select ok to the dragons.
type this into the filter
devtools.hud.loglimit.console
AMO :: DevPrefs
Can increase buffer size from the options panel of that addon:
(source: mozilla.net)
EDIT: There are two parts to this answer. The Console Service has one limit, which applies when the Error Console is closed. The Error Console has another limit, which applies when it is open. Normally these limits are both set to 250.
It is not useful to have the Console Service limit is higher than the Error Console limit, since you need to open the Error Console in order to view the errors, but in theory you can do this by debugging the Firefox executable and setting a breakpoint on xul!nsConsoleService::nsConsoleSerice and set mBufferSize to a larger value. Symbols to allow you to do this are available on the Mozilla symbol server.
You can increase the Error Console limit by editing the consoleBindings.xml file. This file is stored in one of the .jar files provided with Firefox (omni.jar in Firefox 4). Note that editing a .jar file will stop partial updates from working, instead the Firefox updater will reinstall the full updated version.

Resources