which is best for the client performance [closed] - performance

Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 8 years ago.
Improve this question
Which is better, small data size and more server requests, or large data size and fewer requests? Larger data size means longer processing time on ther server. How does this scale out thouh as the number of users and their activity increases?

It is always ideal to get all that you need in as many less calls as you can. If you can get all that you need in one call, definitely do it.
Also it will be very helpful if you can give some insight of what kind of application are we talking about here.

Related

How to improve performances in UIPath? [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
How to improve performances in UIPath? What are the activities can we use in automation to increase performance speed? List out the activities that should be avoid.
First, it is always a good idea to program your RPA with visible actions to make sure that you can achieve the result that you want while seeing the workflow in action, then you can increase performance switching between the input methods (Default, SimulateType, SendWindowMessages) so you can make your RPA work in the background.
https://www.youtube.com/watch?v=fuyRjiXLaRc
As of the output methods, avoid using OCR activities as much as you can since they will slow down your process comparing to (Full text, native).
https://studio.uipath.com/docs/output-or-screen-scraping-methods
If you are working in a citrix environment and you want to interact with an element that doesn't show up immediately avoid using the delay activity since it could not show up in the amount of time that you have programmed, use the wait image activity instead.
https://www.youtube.com/watch?v=EmaQ_BLwgaw

Render large point cloud set (More than 1 GB) in web browser [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
Now I am struggling in rendering point cloud in WebGL. Once the file size get larger than 200 MB, the browser will crash. I am now using Points in Three.js to render my data.
Is there any way to optimise it or good library I can use to do this case?
This is very broad topic where you can do many ridiculous optimizations.
Take a look on this http://www.potree.org/

Searching through an list [closed]

Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 6 years ago.
Improve this question
I'm reading about AI and in the notes it is mentioned
A lookup table in chess would have roughly 35^100 entries.
But what does this mean? Is there any way we could find out how long it would take the computer to search through and find it's entry? Would we assume thereis some order or that there is no order?
The number of atoms in the known universe is estimated to be around 10^80 which is much less than 35^100. With current technology, at least a few thousand atoms are required to store a single bit. I assume that each entry of your table would have multiple bits. You would need some really advanced technology to implement the memory of your computer.
So the answer is: With current technology it is not a matter of time, it is simply impossible.

Are applications such as image processing naturally slow in Scheme for lacking a random-access data-structure? [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
Scheme lists are slow for random access, which is a common operation in many applications such as image processing. Does this make it naturally handicapped for that kind of application?
If performance is a concern, then you should definitely consider using fixed-access-time structures. Fortunately, Scheme has lots of these, too. The "vector" is the simplest one; it's a close match to what most languages call an "array".

Find plagiarism in bulk articles [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 9 years ago.
Improve this question
I have a 20,000 collection of master articles and I will get about 400,000 articles of one or two pages everyday. Now, I am trying to see if each one of this 400k articles are a copy or modified version of my collection of master articles (a threshold of above 60% plagiarism is fine with me)
What are the algorithms and technologies I should use to tackle the problem in a very efficient and timely manner.
Thanks
Fingerprint the articles (i.e. intelligently hash them based on the word frequency) and then look for statistical connection between the fingerprints. Then if there is a hunch on some of the data set, do a brute force search for matching strings on those.

Resources