Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
Where to get data sets for random or test data generation, such as names/surnames with distribution, address data, university/school names, company names etc.?
I've found the list of English names and surnames, with the count of them (unfortunately I haven't noted from where I got that). I got address database from Poland. However these data sets from other countries would also be very useful for me. So with university and school names.
What data do you need as source for such information? Could you provide links to such data? (of course, only those who are free publicly available)
I think you will find answer to your question in following topics:
Sample database for exercise
https://stackoverflow.com/questions/202092/where-can-i-find-free-and-open-data
There are many open source and commericial test data generators on internet. Below 2 are good ones
http://www.sqledit.com/dg/
http://www.generatedata.com/#about
for random numbers/strings: http://www.random.org/
Amazon has made several public data sets available for free download:
http://aws.amazon.com/publicdatasets/
Try http://www.mockaroo.com
You can generate up to 100,000 rows of data in CSV, tab-delimited and SQL formats, save & reuse schemas, and automate test data generation using curl.
There is a free API at http://randomprofile.com/api-for-developers/ for generating test user profiles which include name, surname, address, bank info, CC number, blood type etc. Not sure about the schools, but useful if you're looking after data on Asian users.
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 9 months ago.
Improve this question
Hail all:
Though I am 52, and active with computers since 1986 (those were more oversized calculators, to be correct), i have no idea what is meant with Form Data.
I know, I know, this is insane, but i just never came across this before.
Might be, I DO know it, but by a different name, maybe Dutch.
Still I am quite stuck, to be quite honest.
Tried to google it, got heaps on information as to how to clear it, restore it, save it, unsave it, ... , but nowhere an in-depth explanation of what is meant with form Data.
Now, I am building a Batch File, for speed-cleaning of certain data, when my Firefox starts to get slow ... .
Came across the "erase Form Data" command, but nae clue as to what Form Data is.
Thank you.
Ben
That rundll32 command only applies to Internet Explorer and will not touch Firefox.
Anyway, form data is data entered into html forms like your name, address and telephone etc. These form fields are edit boxes, check/radio buttons and drop-down selectors.
Firefox has its own setting to clear form data when you close the browser. Form data is very low on the list of what slows down your browser...
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 8 years ago.
Improve this question
I am new to ABAP programming. To prepare myself for my new job, I am reading ABAP books. During reading, I learned that ABAP has several legacy elements to keep it backwards compatible with older SAP releases.
Regarding GUIs, I am reading about SAP-UI (PARAMETERS, etc.) Dynpros and WebDynpros. Now, I am unsure about on what to focus my learning efforts on.
Are the common rules like "You should know a little about basic SAP-UI, but mainly focus on WebDypros."
Background information: My new employee does SAP customizing for small and medium sized enterprises.
I'm not a consultant, but I work for a medium (~120 employees) sized company myself. If you were to work for us you would mostly create custom abap reports, maybe sometimes program a user exit. Small companies usually don't spend the money needed for big SAP driven portals, so they probably don't use Netweaver AS Java at all. That means abap dynpro and abap lists as your main UI elements. Sometimes it is good to also know your way around other ways of creating reports, for instance SAP Query.
If I were you I would start with basic abap. You won't have any fun working with dynpros if you haven't gotten your head around the basic stuff first. Learn to work with internal tables, work areas, field symbols. Have a look at some basic ABAP Objects stuff (for instance the ALV grid, very useful for displaying all sorts of tables). You should also understand the ABAP Dictionary, the place where structures, tables, data elements, data domains ans search helps are defined.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I am trying to query Windows Event log for events and am a bit stuck at which approach to use. Windows Developer documentation lists two examples (and approaches) of querying log data.
Querying for Events
http://msdn.microsoft.com/en-us/library/aa385650%28v=vs.85%29.aspx
using EvtQuery function
Querying for Event Information
http://msdn.microsoft.com/en-us/library/windows/desktop/bb427356%28v=vs.85%29.aspx
If anyone has experience querying for event data, are these two approaches equal? Or is any of them outdated or not recommended for actual use? I am new to Windows programming and not really found any recommendations regarding any of these approaches on MSDN.
http://msdn.microsoft.com/en-us/library/aa385650%28v=vs.85%29.aspx using EvtQuery function
The difference is in the use of header. This approach is simpler but you will need to design either an XML or structured XML query.
http://msdn.microsoft.com/en-us/library/windows/desktop/bb427356%28v=vs.85%29.aspx
If you need more flexibility in the format of the returned events (like time stamp format) use this approach. This is harder but gives more flexibility if you need control over the format of the extracted data.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
I need a large data (more than 10GB) to run Hadoop demo. Anybody known where I can download it. Please let me know.
I would suggest you downloading million songs Dataset from the following website:
http://labrosa.ee.columbia.edu/millionsong/
The best thing with Millions Songs Dataset is that you can download 1GB (about 10000 songs), 10GB, 50GB or about 300GB dataset to your Hadoop cluster and do whatever test you would want. I love using it and learn a lot using this data set.
To start with you can download dataset start with any one letter from A-Z, which will be range from 1GB to 20GB.. you can also use Infochimp site:
http://www.infochimps.com/collections/million-songs
In one of my following blog I showed how to download 1GB dataset and run Pig scripts:
http://blogs.msdn.com/b/avkashchauhan/archive/2012/04/12/processing-million-songs-dataset-with-pig-scripts-on-apache-hadoop-on-windows-azure.aspx
Tom White mentioned about a sample weather data set in his Book(Hadoop: the definitive guide).
http://hadoopbook.com/code.html
Data is available for more than 100 years.
I used wget in linux to pull the data. For the year 2007 itself the data size is 27 GB.
It is hosted as an FTP link. So, you can download with any FTP utility.
ftp://ftp.ncdc.noaa.gov/pub/data/noaa/
For complete details please check my blog:
http://myjourneythroughhadoop.blogspot.in/2013/07/how-to-download-weather-data-for-your.html
There are public datasets availbale on Amazon:
http://aws.amazon.com/publicdatasets/
I would suggest to consider running demo cluster there - and thus to save downloading.
There is also good dataset of the crowled web from Common Crawl, which is also available on amazon s3. http://commoncrawl.org/
An article that might be of interest to you, "Using Hadoop to analyze the full Wikipedia dump files using WikiHadoop".
If you are after Wikipedia page view statistics, then this might help. You can download pagecount files from 2007 up until current date. Just to give an idea of the size of the files, 1.9 GB for a single day (here I chose 2012-05-01) spread across 24 files.
Currently, 31 countries have sites which make available public data in various formats, http://www.data.gov/opendatasites. In addition, the World Bank makes available data at http://data.worldbank.org/data-catalog
What about "Internet Census 2012", data gathered by a distributed scan over the whole Internet:
Announcement: http://seclists.org/fulldisclosure/2013/Mar/166
Data: http://internetcensus2012.bitbucket.org/
The whole data is 7TB, (obviously) only available by torrent.
If you are interested in countries indicators, the best source I found was worldbank.org. The data they offer can be exported as CSV which makes it very easy to work with in Hadoop. If you are using .NET, I wrote a blogpost http://ryanlovessoftware.blogspot.ro/2014/02/creating-hadoop-framework-for-analysing.html where you can see how the data looks, and if you download the code from gidhub https://github.com/ryan-popa/Hadoop-Analysis, you already have the string parsing methods.
It might be faster to generate the data than it is to download it and put it up. This has the advantage of giving you control of the problem domain and letting your demo mean something to the people who are watching.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 5 years ago.
Improve this question
I'm having trouble with project management & am looking for a good tool that will be a good match for the way my brain works (very associatively). I'd like a bug-tracker but one that I can group tasks into topics and associate the topics to each other in a graph (see the Wikipedia entry on Topic Maps ) so that I can find & visualize easily the "big picture". I've tried using AbstractSpoon's ToDoList and it works well but it's hierarchical and after about 30 or 40 entries I get lost in a maze of things to do.
any suggestions?
edit: I've now tried Freemind, Conzilla, XMind, and VUE. Freemind and Conzilla were a little flaky. XMind seems to be the most polished of the four; they have a "pro" version which is non-free (pay by the month >:( which is weird) but an open-source base version which is free. You can't export the data directly from the program with the free version, but the storage format is just a .jar-style (ZIP file w/ extension .xmind) file that contains a "contents.xml" that is easily parsed if I needed to.
#codeslave:
but how important is the visual
representation any way
Visualization is everything! I've got information overload and I need to be able to navigate a mess of information. I don't want it to be super-Powerpoint-polished, but I need to be able to use the associations that I create to remind myself how to find what I'm looking for. In an ideal world you could just full-text search everything, but that only works if you can remember the search phrase. Often I'll file something under "algorithm" and when I go to look for it I look under "programming" instead, or vice-versa. Associativity solves that problem by allowing me to visually browse my "mental model" of the information I've stored.
You can always get an CVS export from your "favourite" tool and create Topic Maps maps you can view with the Omnigator or the xSiteable tool. I used to have a few XSLT files to do such a job dealing with JIRA data. If the interest is high enough, maybe a ressurection is needed?
I've developed a small utility that will import MindMaps into Project plans. Let me know if something like this is helpful and I will develop it further.
Right now I just use it one-way from MindMap -> Project file. I generally use this for brainstorming and scope management, then export to Project when we like the scope of what we are working with for more formal project management.
What about using good old FogBugz? You can associate cases together pretty easily. You don't get the pretty graph of the topic space/mind map (feature idea Joel) but how important is the visual representation any way.