I executed the jmx file from ubuntu environment on amazon server. It got executed and report files are also generated. But i want to transfer/move these reports to my windows local. I dont want to use winscp/filezilla, as my data is very huge for huge threads(i tried this, but its taking hell lot of time to push).
Pls suggest a way to do it.
You can use some form of JMeter Plugin which will upload your test results into Cloud Analysis services. Amazon instances usually have good bandwidth so it shouldn't take long. If you need results locally you can download them from services, usually they provide possibility to fetch compressed results files. As a bonus you'll get advanced analysis toolkit.
Currently I'm aware of the following plugins:
Loadosophia.org Uploader
BlazeMeter's Plug-in for JMeter
Both are free and easy to install/use.
Related
Azure DevOps offers us the possibility of running Load Tests in the Cloud. Thus, we can use multiple servers to hit the web app under test from different locations.
The Azure DevOps UI allows us to upload a JMeter test file, plus some supporting files, like CSV files that will be used by the test.
When we develop the test, we'll most certainly be running JMeter against a locally running application, to make sure our requests are properly formatted and are hitting the application as desired. Thus, we'll be running JMeter locally against localhost:.
When we upload the test plan file to Azure DevOps, we'll expect the test to run against the application that is deployed to Azure App Services (for example). Hard-coding the URL in the test plan is quite inconvenient. Isn't there a way to make Azure DevOps pass this a parameter to JMeter before the load test runs?
JMeter accepts variables to be defined in the local environment, outside of the test plan, but the Load Test UI in Azure DevOps doesn't seem to support this.
Looking into Azure DevOps documentation it is possible to provide "Supporting Files"
So you can put your URL(s) into i.e. CSV file and load it inside your JMeter test using one of the following approaches:
CSV Data Set Config
__FileToString()
__CSVRead()
I am trying to do the MSI web deployment with chef. I have about 400 web servers with same configuration. We will do deployment in two slots with 200 servers each.
I will follow below steps for new release,
1) Increase the cookbook version.
2) Upload the cookbook to server.
3) Update the cookbook version to role and run list.
I will do lot of steps from cookbook like install 7 msi, update IIS settings, update web.configure file and add registry entry. Once deployment is done we need to update testing team, so that they can start the testing. My question is how could I ensure deployment is done in all the machines successfully? How could I find if one MSI is not installed in one machine or one web.config file is not updated properly?
My understanding is chef client will run every 30 Mins default, so I have wait for next 30 mins to complete the deployment. Is there any other way with push (I can’t use push job, since chef is removed push job support from chef High Availability servers) like knife chef client from workstation?
It would be fine, If anyone share their experience who is using chef in large scale windows deployment.
Thanks in advance.
I personnaly use rundeck to trigger on demand chef runs.
According to your description, I would use 2 prod env, one for each group where you'll bump the cookbook version limitation for each group separately.
For the reporting, at this scale consider buying a license to get chef-manage and chef-reporting so you'll have a complete overview, next option is to use a handler to report the run status and send a mail if there was an error during the run.
Nothing in here is specific to Windows, so more you are asking how to use Chef in a high-churn environment. I would highly recommend checking out the new Policyfile workflow, we've had a lot of success with it though it has some sharp limitations. I've got a guide up at https://yolover.poise.io/. Another solution on the cookbook/data release side is to move a lot of your tunables (eg. versions of things to deploy) out of the cookbook and in to a little web service somewhere, than have your recipe code read from that to get their tuning data. As for the push vs. pull question, most people end up with a hybrid. As #Tensibai mentioned, RunDeck is a popular push-based option. Usually you still leave background interval runs on a longer cycle time (maybe 1 or 2 hours) to catch config drift and use the push system for more specific deploy tasks. Beyond RunDeck you can also check out Fabric, Capistrano, MCollective, and SaltStack (you can use its remote execution layer without the CM stuffs). Chef also has its own Push Jobs project but I think I can safely say you should avoid it at this point, it never got enough community momentum to really go anywhere.
I wasn't able to find solid information on this and I wanted to ask developers who use Parse Dashboard:
What are the pros/cons of Parse Dashboard local installation vs deployment?
I currently run the Parse Dashboard on local installation, but I know that deployment to Heroku is also an option (my app is deployed on Heroku). I wanted to gather some information before deploying/not deploying.
Thank you!
I also have it running locally and I think for security reasons it's best to do so. If you setup the dashboard on the same server on which Parse is running, then you will have to take security measure to protect access to the dashboard and the config file which includes your masterkey and all that. This definitely outweighs the arguments to host it locally, which in my opinion only is that it's easier to access the dashboard.
If you really want to setup a dashboard on a server at least do it on a separate server.
Just installed and started using VSSonarExtension, but already have a couple of questions that I am unable to answer myself:
Why does the extension need to connect to the server if it scans local code with local tools, such as FxCop and StyleCop?
After looking through roughly 15 files, I get an error saying that I need a license to scan more than 15 files in one session - isn't the extension open source?
I want to make sure that the rules specified on the server match the rules locally, so that I get the same output with regards to technical debt - how can I achieve it?
Seems like the extension does not use FxCop, according to the logs; moreover, sometimes logs indicate that some code has been skipped because some rules were not present or defined in StyleCop. What does that mean?
Looking forward to your replies, thanks for help.
answers
so it can sync the quality profile in server, you dont want your users to complain the local analysis is different from what they have in server
local files analysis, is provided by external plugin: please contact support their for details
its the default behaviour. therefore ansering question 1
please open a ticket in github with those issues. it can be that you have some differences between server profile and what is installed locally.
I am doing a seminar on JMeter and I need to demo on some software. My professor say we need an open source software. So that we can set it up on local host and testing against it
The problem is I don't know which open source software should be used. I am new to JMeter and saw some examples with testing a website only
So can anyone tell me some open source software that can be put under test with JMeter?
Thanks in advance.
Yes, start with this GitHub project and then, since the project includes Groovy, and requires Maven to run, then you have 2 additional open source projects involved. All you have to do is run JMeter using the included .bat script and then create a TestPlan with a JSR223-Groovy sampler in it. (The project I mention above will launch a certain version of Jmeter 2.11)
You could also go one step further and install PostgresSQL (also open source) and then create a JDBC sampler that connects to it using Jmeter (because the Jmeter project I mentioned above also includes the postgres jdbc driver by default).
If you need a REST api to test against, then install Jenkins on your local computer and you can get Jmeter to perform rest queries against it.