How can you implement gis server using laravel? - laravel

I want to start a project where my web application would suggest best path available to reach from one destination to another.
https://github.com/eleven-lab/laravel-geo
I found above url but I really don't know how to use that above mentioned thing properly. Can someone please help me understand it?

There are many applications to do routing in GIS.If you have ESRI, they have a network builder module. In open source you can start with OSM https://wiki.openstreetmap.org/wiki/Routing. Also check out OSM routing with SpatiaLite. https://www.gaia-gis.it/fossil/libspatialite/wiki?name=misc-docs ( http://www.gaia-gis.it/gaia-sins/Using-Routing.pdf).

Related

Unable to access H2O Flow using H2O_connection_url

I am using both H2O and Sparkling Water on Amazon Clusters. I have been using Qubole and have been able to access the Flow UI on that platform. I am currently testing Databricks and Sagemaker, but I am unable to access the Flow UI using either platform (using port 54321). I am using H2O_cluster_version: 3.32.1.3. Do I need to use another port?
Getting the right Flow URL can be tricky because of the changes in the base URL at DBC. There were some improvements in more recent releases of SW that give the proper URL within Databricks, so make sure you try the latest version.
You should get it from your print/output, when you create an H2OContext. The port would be 9009. If you want to change it, you can use spark.ext.h2o.client.web.port.
You can also find the link in "Spark UI" -> "Sparkling Water" tab
The format would be something like: https://your-dbc-domain/driver-proxy/o/xxxxxxxx/yyyyyyy/9009/flow/index.html
From the docs for reference:
Flow is accessible via the URL printed out after H2OContext is
started. Internally we use open port 9009. If you have an environment
where a different port is open on your Azure Databricks cluster, you
can configure it via spark.ext.h2o.client.web.port or corresponding
setter on H2OConf.

Internal Server Error- http://localhost:1080/webTours/home.html

Application: http://localhost:1080/webTours/home.html [I intend to use it for Jmeter testing]
Below error is displayed when I am trying to click on "sign up now" link. Kindly help to fix it. Pls note, I have strawberry-perl-5.30.0.1-64bit.msi also installed and path is set to the environment.
Internal Server Error The server encountered an internal error or
misconfiguration and was unable to complete your request.
Please contact the server administrator, admin#localhost and inform
them of the time the error occurred, and anything you might have done
that may have caused the error.
More information about this error may be available in the server error
log.
Thanks.
Probably, you need to do the correlation for the dynamic parameters like sessionID etc.
Record two scripts using tools like fiddler or LR or any other sniffing tool and compare two scripts to find out dynamic variables required. Then correlate them. For correlation, identify the first occurrence of the variable in the response then extract them using post processor and pass them further below/to all the request that requires it.
Hope this helps.
It appears that your Web Tours Sample Application installation is broken, looking into MICRO FOCUS SUPPORTED stanza in the application details you should contact Micro Focus Support or ask the question in Micro Focus Community Forums.
In the meantime you can use http://blazedemo.com/ page and Getting Started with JMeter - A Basic Tutorial as the alternative test application for your JMeter training
If you are not getting internal server error while sign up (or) accessing administration link then follow the below,
a. close the server from the cmd
b. In the Web Tours 1.0 parent folder, find the file .msi file: strawberry-perl-5.10.1.0
c. Double click on strawberry-perl-5.10.1.0 msi file to install
d. Once it is installed, Now re-start the server (Double click on StartServer.bat file) under WebTours sub-folder
e. Now, Go to link http://localhost:1080/webtours/home.html
f. Click on Sign up (or) administration link
g. Hope you are able to access the Web Tours app in local server now. Thanks! :-)

how-to configure a bgp-ls peer

I have brought up an Opendaylight instance in order to establish a BGP-LS session with our network and get topology information.
So, the objective is to configure 1 peering with a router and get node, link,network information.
I have concluded that the documentation is more accurate over here: https://github.com/opendaylight/docs/blob/master/docs/user-guide/bgpcep-guide/bgp/bgp-user-guide-linkstate-family.rst
I have done this step https://github.com/opendaylight/docs/blob/master/docs/user-guide/bgpcep-guide/bgp/bgp-user-guide-running-bgp.rst and works.
Unfortunately, there are no such files like it is described here https://github.com/opendaylight/docs/blob/stable/lithium/manuals/user-guide/src/main/asciidoc/bgpcep/odl-bgpcep-bgp-all-user.adoc
Meaning, I cannot locate 31-bgp.xml and 41-bgp-example.xml
I would like to ask if there are or can be derived somehow .xml files that describe fully all the parameters available, so that I can configure OpenDaylight by using these .xml files.
The examples mentioned in the first link are targeting a few specific parameters, and it is not about a nearly complete example.
Could you please advise how to:
get the necessary bgp-ls configuration .xml files that fully
configure ODL as BGP speaker and a peering ?
Many Thanks.
These files has been removed as a part of task BGPCEP-685.
You can still find them in previous release like here or here
There is also wiki about configuring BGP peer.

H2O Steam deploy can't connect to Prediction Service Builder

I am trying to use h2o steam (running on localhost) to deploy a model. After importing the model from h2o flow, clicking the "deploy model" option in the "models" section of the project, filling out the resulting dialog box, and clicking the "deploy" button, the following messages are displayed:
At first I thought that it was because maybe I needed to start up the service builder on my own, so I started it up following the docs here, but still got the same error. Any suggestions would be appreciated. Thanks :)
Just make sure jetty HTTP server is running locally by executing the following in your shell:
java -jar var/master/assets/jetty-runner.jar var/master/assets/ROOT.war
Looking here, it seems like I would need to "override" some kind of default browser restriction for accessing localhost:8080 (which is what I assume steam is trying to do to launch the service builder (I don't know much about networking related stuff)). I got around this by launching steam with the command:
$ ./steam serve master --prediction-service-host=localhost --prediction-service-port-range=12345:22345
where the ports are some arbitrary range between (1025, 65535) which I got by word-searching the a page of the steam source code (line 182 as of the date of this posting).
Doing this lets me deploy the models through the steam dialog without any error messages. Again, I don't know much about networking related stuff, so if anyone has a better way to solve this problem (ie. allow access of localhost:8080) please post or comment. Thanks.

Mule Connect to remote flat files

I am new to Mule and I have been struggling with a simple issue for a while now. I am trying to connect to flat files (.MDB, .DBF) located on a remote desktop through my Mule application using the generic database connector of Mule. I have tried different things here:
I am using StelsDBF and StelsMDB drivers for the JDBC connectivity. I tried connecting directly using jdbc URL - jdbc:jstels:mdb:host/path
I have also tried to access through FTP by using FileZilla server on remote desktop and using jdbc URL in my app - jdbc:jstels:dbf:ftp://user:password#host:21/path
None of these seem to be working as I am always getting Connection exceptions. If anyone has tried this before, what is the best way to go about it? Connecting a remote flat file with Mule? Your response on this will be greatly appreciated!
If you want to load the contents of the file inside a Mule flow you should use the file or FTP connector, i don't know for sure about your JDBC option.
With the File connector you can access local files (files on the server where mule is running), you could try to mount the folders as a share.
Or run an FTP server like you already tried, that should work.
There is probably an error in your syntax / connection.
Please paste the complete XML of your Mule flow so we can see what you are trying to do.
Your usecase is still not really clear to me, are you really planning to use http to trigger the DB everytime? Anyway did you try putting the file on a local path and use that path in your database url. Here is someone that says he had it working, he created a separate bean.
http://forums.mulesoft.com/questions/6422/setting_property_dynamically_on_jdbcdatasource.html
I think a local path is maybe possible and it's better to test that first.
Also take note of how to refer to a file path, look at the examples for the file connector: https://docs.mulesoft.com/mule-user-guide/v/3.7/file-transport-reference#namespace-and-syntax
If you manage to get it working and you can use the path directly in the JDBC url, you should have a look at the poll scope.
https://docs.mulesoft.com/mule-user-guide/v/3.7/poll-reference
You can use your DB connector as an inbound endpoint when wrapped in a poll scope.
I experienced the same issue when connect to Microsoft Access Database (*.mdb, *.accdb) using Mule Database Connector. After further investigation, it's solved by installing Microsoft Access Database Engine
Another issue, I couldn't pass parameter to construct a query as same as I do for other databases. e.g.: SELECT * FROM emplcopy WHERE id = #[payload.id]
To solve this issue:
I changed the Query type from Parameterized into Dynamic.
I generated the query inside Set Payload transformer (generate the query in form of String, e.g.: SELECT * FROM emplcopy WHERE id = '1').
Finally, put it into the Dynamic query area: #[payload]

Resources