Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed last month.
Improve this question
According to developpers.google.com, we can store the placeIDs and refresh them to get the details. I need to get the place nearby in order by distance over 500 meters. I don't know how to do it without the location storing.
Can I store the place locations with the placeIds according to developpers.google.com?
Thank you for you replies.
SyLaDe
In my opinion, Google Maps place_id is just the key to request basic information about that place (and it's free) or you can request detail information such as contact and atmosphere (with additional charge).
If you wanna get the place nearby in order by the distance over 500 meters, you can simply use Nearby Search requests under Place API.
for example:
https://maps.googleapis.com/maps/api/place/nearbysearch/json?location=-6.321516009,106.86227379999998&radius=500&key=yourkey
and you will get json result that you can store and organize.
To get longitude and latitude, you can manually choose from google maps, or you can get it from GIS tool with coordinates acquisition.
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about programming within the scope defined in the help center.
Closed 6 years ago.
Improve this question
I'm working on a book for Manning and want to use Alchemy News API as part of one of the examples. I have a free license which says it allows for 1,000 transactions per day. Does that mean 1,000 queries or something else? I hit the limit today way earlier than I expected to, at significantly less than 1,000 queries.
Each type of call has different amounts of transactions it uses. The text analysis uses just 1 transaction but the image analysis uses more. I believe it used about 5 transactions per image but it's been awhile since I've used the image recognition.
The number of transactions used is given in the response from AlchemyAPI. It also gives you more details in the documentation.
Query Cost + Result Cost = Total Cost
This is something I learnt today. Therefore you'll run out of transactions easily with loads of data. Keep your queries to a minimum.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about programming within the scope defined in the help center.
Closed 5 years ago.
Improve this question
is there a specific reason why Ryanair is missing from the QPX Express API results while it appears in a normal search on Google Flight?
Thanks
Max
I noticed that the QPX API and Ita Software Matrix (by Google) gives the same results, not including a lot of companies like Ryanair.
I asked Google Support about it.
The answer is:
"QPX Express API and Matrix only include fares that are filed with ATPCO. There are a number of small carriers and low cost carriers that do not file."
Asking Google if they consider to add these flights to their APIs the answer was:
"Airlines determine which distribution channel they want to participate in. If any of these airlines were to file with ATPCO, then we would certainly be interested."
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about programming within the scope defined in the help center.
Closed 8 years ago.
Improve this question
I should to prepare myself for upcoming task which consist of a lot of graphs.
I need some data (available in free domain) to train myself.
Bigger - is better...
could you suggest some open data resource?
I'll appreciate this.
You can visit http://snap.stanford.edu/data/ . It contains many different kind of network or graph data.
Here is an answer for your could you suggest some open data resource? and not for which consist of a lot of graphs. So, plz, keep it in mind.
Here (data.gov.au) you can find a huge datasets (864!) of a different types in a different formats (txt, csv, xml, ). You will find a Finance, Industry, Geography, etc. datasets.
In other case, if you want some special (and meaningful data, for example, global population density) you can see this (a bit outdated, but usefull) source from readwriteweb.com.
And one more source: "Open Governmental Datasets" - it's worth to see it indeed.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 9 years ago.
Improve this question
I want to generate some fake people for a piece of software I'm writing. These people also have ethnicities, and I'd prefer not to have names that don't look totally out of place when compared to those ethnicities.
My first idea was to base it on data. There is a table of first and last names from the 1990 US census with attached frequencies, but that says nothing about ethnicity. There is also a table of last names from the 2000 US census which is broken down by ethnicity, but it says nothing about first names.
So I need some way of generating first names based on ethnicity. Any ideas?
Use behindthename.com. They have very extensive lists of names by usage, including lists of popular names.
The site http://www.babynamefacts.com/ contains lists of most popular baby names per country. This may be a good starting point. For example, this page shows the most popular baby names for Serbia in 2009: http://www.babynamefacts.com/popularnames/countries.php?country=SRB .
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
i would like to know what algorithm and what programming language wikipedia is using to randomly choose an article to display.
i would also like to know how does it work so fast?
Here's information on that.
Every article is assigned a random number between 0 and 1 when it is created (these are indexed in SQL, which is what makes selection fast). When you click random article it generates a target random number and then returns the article whose recorded random number is closest to this target.
If you are interested you can read the actual code here.
Something along this lines:
"SELECT cur_id,cur_title
FROM cur USE INDEX (cur_random)
WHERE cur_namespace=0 AND cur_is_redirect=0
AND cur_random>RAND()
ORDER BY cur_random
LIMIT 1"
From MediaWiki.org:
MediaWiki is a free software wiki
package written in PHP, originally
for use on Wikipedia. It is now used
by several other projects of the
non-profit Wikimedia Foundation and by
many other wikis, including this
website, the home of MediaWiki.
MediaWiki is open source, so you can download the code and inspect it, to see how they have implemented this feature.
If you look at the source, they use PHP/MySQL a sort and filter pages by pregenerated random values (page_random column) that have an index on them.