Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 9 years ago.
Improve this question
I'm unable to decide which is a better benchmarking tool-Is it apachebench tool or httperf?
Please help me to decide which one would be the most appropriate tool to be chosen for benchmarking the web application.
I suppose it depends on what kind of testing you want to do.
ApacheBench basically tests a general load, but httperf gives more a surgical precision and can simulate real user activity.
is “ab” or “httperf” better for checking performance of a website?
How do you test the performance of a website?
Most effective and realistic free web-app load tester?
Tools to benchmark web-services
They're both good for doing different types of testing. httperf will just send a barage of data to your server, whether or not it's answered. That gives you the breaking point of your app, and shows you how it fails out. Use this to find bottlenecks.
apachebench is a bit different in that it sends exactly what you tell it, and is more typically used to measure response times (for example, if you optimize). Once the bottlenecks are found and fixed, check the speed with this.
Please Note: That's what I do, but I'm not a master
I'm unable to decide which is a better
benchmarking tool-Is it apachebench
tool or httperf?
There is an extensive discussion about the benefit of each tool here:
http://dsec.com/en_apachebench_httperf.html
The guy went as far as to write a wrapper (public domain C code) around Apache Benchmark and HTTPerf to illustrate the pro and cons of each solution.
He also explains how to make sure system settings will not make the benchmark fail, and how to get the most of a server machine (OS, virtualization, etc.).
Hope this helps!
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I want to conduct online coding competitions. For that I am developing a website. I have a webpage to take the code from user. Once the user has finished his coding, on submitting I want to send that code to some compiler to check the code. And I can check with the test cases if its compiled without errors. Is there any online compilers available for that. I mean, Is there any site to send my_code, test_cases, programming_languages to that and get back the results? Or should I use system commands to do this process?
Compiling requires a fair amount of computation. You will have to consider how well your system will scale to user demand. Most online services would eventually charge a fee to cover their infrastructure costs (+ profit).
See ideone.com/sphere-engine
Depending on the scale, you may want to use system commands. I would suggest doing so until your project reach a certain maturity. You could also have a setup like Jenkins, were compilation work is dispatched to build nodes which can be added for increased demand.
Existing software like PC2 ask the hosts of the competition to provide the compilation server. However, the setup is local.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 7 years ago.
Improve this question
We are looking for a software that will help to decide which MOM solution should be used in a given case, taking into account throughput and latency figures it measures.
Not to reinvent the wheel we'd like to investigate what solutions already exist on the market.
For example, there are:
JMeter with its Point-to-Point Test Plan and Topic Test Plan
QPID performance and latency test utils
ActiveMQ JMeter performance test
HornetQ performance test
RabbitMQ performance test
etc.
It seems there are no tools to test performance and latency of these MOMs somehow in a general way.
So could you please recommend tools which worth looking at or some ideas how to implement such a tool in a cross-platform, cross-language way.
While someone may know a tool to test them all I'd recommend that you build a testing scenario that resembles your intended use (message sized, integration patterns (e.g. do you need CBR and if so what), required latencies, message rates, target environment etc.)
Synthetic benchmarks can be misleading and will most likely not represent your usage scenario
As a side note I'd say that AMQP has the advantage that it also defines the wire protocol so if you choose one you can change the backend without affecting the client
What about SPEC JMS 2007 ??? I think it would be quite usefull for your case :)
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 3 years ago.
Improve this question
I'm currently trying to write up some software specs for a large-ish project that I'd like to bid out on oDesk (maybe 3k-6k). I have a classical education in computer science, but minimal work experience and I've never worked in a corporate environment.
I've written up 15 pages of project specs for an API that I want to run on Google App Engine. I'm finding the subtleties difficult with issues such as: where to draw the line on what to specify and what not to; how to break the project up into pieces so that if one part goes wrong, other parts can still be used; where and how to set milestones; and how to screen for a candidate who has the skills to properly implement the specs.
I asked a friend what issues I should watch out for and he said:
Make sure specs do not have contradictory information
Make sure specs are not unnecessarily specific because this will turn off more skilled developers
What other pitfalls and common mistakes should I look out for while drawing up project specs?
While it is important to focus on the functionality of what you want to achieve, also keep in mind the non-functional requirements as well. Think about how you want it to perform, security considerations, etc.
One to add to the list
Make sure the specs are unambiguous. (Can only be interpreted one way...Hopefully)
I think these two links will help:
On Reqs And Specs: http://www.pragmaticmarketing.com/publications/topics/02/0204sj
Painless Functional Specifications - Part 2: What's a Spec?: http://www.joelonsoftware.com/articles/fog0000000035.html
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 5 years ago.
Improve this question
There are lots of texts on how to plan software projects (with user stories, etc), but they usually assume you have a large budget, liberal timeframes and/or a real dev team available. While they sound fantastic, they never seem to account for solo devs working on a short deadline.
There is also a lot of talk about test-based methodologies where you write a test case for every method before you implement it, but I feel that these are difficult to impossible to apply if your software is GUI-focused (e.g. (server-side) web programming or Flash/ActionScript).
Although I try to make heavy use of refactoring to improve my code whenever I have finished a section of it, last minute hacks and additions tend to make this incredibly frustrating and I often feel that there should be a way I can utilise at least some of the planning theory that's apparently meant to help large dev teams and developers of software libraries first and foremost.
What is The Right Way to go about writing small-ish applications as a solo dev and how do you prevent last minute changes from making your code worse?
There is no "Right Way" unfortunately, however, there are a lot of better ways. I think there is no real distinction between small and large projects - the same kind of things need to happen, it's just the depth of those things that changes.
In your situation - working to a short deadline, preventing last minute problems - the same old tried and true methods are going to work:
Use source control effectively, develop last minute changes in a separate branch so you can drop them easily if required.
Test, test, test. If your tests cover the intent of what you were trying to achieve properly then last-minute changes can be measured.
I suspect you need to look seriously at some different types of testing and tools - there are plenty around that will help you manage these issues.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 5 years ago.
Improve this question
We have been asked to provide a data reporting solution. The followng are the requirements:
i. The client has a lot of data which is generated everyday as an outcome of the tests they run. These tests are run at several sites and they get automatically backed up into a central server.
ii. They already have perl scripts which post process them and generates excel based reports.
iii. They need a web based interface for comparing those reports and they need to mark and track issues which might be present in those data.
I am confused if we should build our own tool for this or we should go for already exiting tool(any suggestions?). Can you please provide supportive arguments for the decision that you would suggest?
You need to narrow down your requirements (what kind of data needs to be compared, and in which format?). Then check if there is already a software available (commercial or free) that fulfills your needs. Based on that, decide if its better (i.e. cheaper) to implement the functionality yourself, or use the other software.
Don't reinvent the wheel.
There are quite a few tools out there that specialise in this sort of thing, my gut feeling is that you can find something ready made that does what you need.
As a side note, that tool may also be a better solution for creating those excel reports than the perl scripts.