It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 9 years ago.
While preparing for an exam, I came across this doubt. Does a CPU process instructions as data?
The short answer is yes. There's a Wikipedia entry on it. The gist of it is, if you have the address of the instructions you would like to read, then you can load those into data registers, test the instructions, and even write back new instructions. Many modern operating systems will probably not allow you write the instructions back however. It is a safety precaution to prevent malicious code from modifying your system, or trust applications.
Depends on the computer:
In von Neumann architecture, yes.
In Harvard architecture, no.
Related
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 10 years ago.
I was wondering if folks use VHDL/FPGAs in scientific computing.
An example scenario that I was thinking off was say:
Construct an arbitrary precision floating point adder
Configure an FPGA board to then add such numbers
So I was looking for references (example code) where VHDL/FPGAs have been used in scientific computing.
Thanks in advance.
There are several vendors who build heterogeneous computing systems using FPGAs. I doubt you'll find complete source code for such systems.
SRC Computing
Convey Computer
Mitrionics. A reseller of other systems.
Novo-G. An academic project.
Look into radio astronomy. With arrays such as the VLA and ALMA, the massively parallel correlator is the part that could be considered most important. These typically use FPGAs but could use custom-designed chips for extreme performance at higher cost.
Some fine reading:
https://science.nrao.edu/facilities/cdl/digital-signal-processing
http://web.njit.edu/~gary/728/Lecture8.html
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
My multicore machine is receiving packets and distributes them evenly (sort-of round robin) among its cores. Each core should decide whether to let the packet pass or drop it, depending solely on its origin address.
I need to find lock-less algorithm and data structure to allow this. Can you help?
If you are happy to use Java, or look at the design of Java source, you could chose a random key and then retrieve a queue from http://docs.oracle.com/javase/1.5.0/docs/api/java/util/concurrent/ConcurrentHashMap.html. Given the queue, you could add the packet to it without blocking if it was a http://docs.oracle.com/javase/1.5.0/docs/api/java/util/concurrent/ConcurrentLinkedQueue.html
Much of java.util.concurrent is due to Doug Lea, who has information on it at http://gee.cs.oswego.edu/dl/concurrency-interest/index.html.
Possibly overkill for your particular problem, but might satisfy a general request for info on data structures in this area.
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
Has anyone tried to implement Numenta's most recent cortical learning theory (http://www.numenta.com)? I'm working on it and would like to share experience.
I think the ideas from Numeta are very promising. But as with any company that wants to make money they are not verbose enough about the technology so that one should be able to re-implement their system (at least from what I have seen so far). It is probably not in their interest that one could just rebuild the system from which they plan to make money.
Also their system is very general and complex. So unless you have a lot of experience with other kinds of neural networks and learning algorithms, I would not recommend experimenting with their ideas. First try to do Backpropagation and maybe some of the less advanced temporal learning until you are really familiar with that.
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
I want to know if is there any specific algorithm that can be followed to understand the meaning of a word/sentence/paragraph. Basically, I want to write a program that will take text/paragraph as input and try to find out what its meaning is. And thereby highlight the emotions within the text.
Also, if there is an algorithm to understand things, can the same algorithm be applied to itself? It reduces the quest further to a point where we become interested in knowing meaning of meaning OR rather definition of definition.
You want Natural Language Processing and Semantic Technology. This is still a flourishing area in computer science. Look at things such as a Semantic Reasoner. You can start with Jena. There are also other things you can look at such as Academic Thesis papers.
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
I came across NUMA-aware allocation in relation to improving the performance of the application. It is something related to Multicore programming. I know NUMA means non-uniform memory access but I am not aware of NUMA-aware allocation.
Do anyone know about this?
NUMA aware allocation is basically the ability to allocate resources in a way such that the NUMA-properties of the system does not result in unnecessarily long commnication paths. I.e memory should be allocated in the RAM connected directly to the CPU where the process needing it runs, etc