Detecting Hololens in the browser - user-agent

How do you detect if a user is visiting your website using a Hololens?
The user-agent of the hololens seems to be equal to the one used by Edge.

Take a look the image attached showing a Hololens with a user-agent, the Edge build number is higher than the publicly released version of Edge for Windows 10 desktop.
Ideally feature detection would be preferred, or detect on navigator.prototype.platform == "Windows.Holographic" would be best, but this seems to be the best way to detect for it.

Related

DJI-Mobile-SDK/position control

I am useing DJI Mobile SDK to crtate an APP by Android Studio now.I want to know how to use the GPS signal of the aircraft and the phone to realize position control. Is there any API in the DJI Mobile SDK I can use?
You may follow this sample to run simple higher-level GPS position control https://developer.dji.com/mobile-sdk/documentation/ios-tutorials/GSDemo.html
If you stop at some waypoint, it can automatically hold the position. It is a simple recreation of the DJI pilot app waypoint planning.
For low-level GPS position control requires a higher understanding of the system. This usually allows interesting applications such as allow drone to follow some person or precision landing to some mark or circle around some tower. There is not much open-source implementation available on the internet. You have to search in the MSDK for the API for some basic control and you also need to have deep understanding in the field that you are trying to achieve e.g real-time object detection, low-level control framework, Visual-Inertial SLAM etc

Installing Microsft Edge in Windows Containers

Context
I'm trying to create a selenium-standalone windows container so we can run e2e-tests on Microsoft Edge. To do this I need to install Microsoft Edge in the container. Afaik, Microsoft does not publish Edge in any form. They only pre-package Edge with Windows.
Unfortunately Edge isn't pre-packaged in any Windows Container. I verified this by adding: RUN powershell (Get-AppxPackage Microsoft.MicrosoftEdge).Version to my .Dockerfile which always came up empty regardless of the tag and image.
Specific question
What I would like to know is whether someone knows where I can find an installer for Edge that's atleast semi-official. I had no luck finding one so far neither online nor in any package-manager.
General question
If you can't answer the question above but do know a different method of installing Microsoft Edge in a Windows Container, please share.
There actually is now a Docker Image with Edge and Edgedriver, the link is at the bottom of this page: https://learn.microsoft.com/en-us/microsoft-edge/webdriver-chromium/?tabs=c-sharp
https://hub.docker.com/_/microsoft-msedge-msedgedriver?tab=description
"This experimental docker container contains Microsoft Edge (Chromium) and Microsoft Edge Driver."
As far as I know, there have Docker images for Chrome and Firefox, but not images for IE or Microsoft Edge. You could add a feedback from here.
At present, as a provisional workaround, you could try to use Chrome and Firefox to do the selenium test. Please check the README at https://github.com/SeleniumHQ/docker-selenium
I will also try to submit the feedback regarding this issue. If I get any solution or a work around than I will try to provide you in future. Thanks for your understanding.

Do Cell Phone Repeaters have CID and LAC?

I want to detect offline mobile phones in a specific area, they are offline no GPS and no internet. I have an application and it should take an action when it detects that it is in the area.
the area may be a floor on a building or even a room.
Can I set a repeater device and wait for its signal to show up then to take an action based on the Cell Id?
knowing that I don't need the mobile phone to connect the repeater.
is there a better approach?
I think you mean that you want an app on the phone to detect it is in the room or building, even when the app is not connected to the internet and not using GPS.
An alternative approach to using a specialist device as you outline, and which might be worth testing, is to leverage the WiFi scanning that your device is probably already doing, even when not connected to the internet.
In Android, for example, you can register a broadcast listener for 'SCAN_RESULTS_AVAILABLE_ACTION' which will provide you a list of all the nearby wifi access points including their BSID's and SSID's ():https://developer.android.com/reference/android/net/wifi/ScanResult
There are permission restrictions which you need to be aware of but they are quite common ones for users - more details here:
https://developer.android.com/guide/topics/connectivity/wifi-scan
Many location services use this type of approach, often to supplement the GPS info and improve accuracy.

Getting detected features from Google Tango Motion Tracking API

I would like to know how to get the current feature points used in motion tracking and the ones that are present in the learned area (detected or not).
There is an older, related post without an useful answer:
How is it possible to get tracked features from tango APIs used for motion tracking. I'm using the tango to not do SLAM and IMU-integration on my own.
What do I need to do, to visualize the tracked features like they did in some of the presentation videos. https://www.youtube.com/watch?v=2y7NX-HUlMc (0:35 - 0:55)
What I want in general is some kind of measure or visual guidance on how good the devices learned the current environment. I know, there is is the Inspector App but I need this information on the fly.
Thanks for your Help ;)
If you want to check if an area is present in your learned area model and which is not, you can use the Tango Debug Overlay App. It has a field 'Tracking Success' that only counts up if the device sees learned feature points (ADF on) or finds new feature points (ADF off) (http://grauonline.de/alexwww/tmp/tango_debug_overlay_app.jpg). Additionally, you can request that debug information like Tango Debug Overlay App does (as a simple text) via UDP port 29361 in your App and parse the returned debug text (although this is not recommended at all for a real app as this interface is not documented)
PS: In Tango Core 01-19-2017 this counter does not seem to work anymore.

Is it possible to control DJI Phantom 2 Vision (plus) with smartphone?

I'm checking at DJI specification for Phantom 2 Vision and Phantom 2 Vision plus and also API reference, but I'm not sure wether is it possible to pilot drone without remote controller? How is the communication via SDK done - is smartphone directly communicating with drone, or is going via remote controller?
DJI SDK Level 2 does allow direct flight control of the DJI Phantom 2 Vision+. It communicates the commands to the unit using wifi. That being said, the wifi extender that your phone is connected to is attached to the controller, so you still need the controller to be turned on and in the vicinity in order to operate, but the device is still what is sending the commands to the Phantom. I would recommend that someone always have the controller at the ready incase the wifi network has interference / phone dies etc. I am using the SDK in this way for my side project www.followmephantom.com.
Hope that helps.
FYI - to use the level SDK you first have to submit a proposal for your project to DJI and get their approval..... and pay a fee in most cases.
Here is a helpful link that shows the difference in features for the SDK level 1 and 2.
https://dev.dji.com/en/products/sdk/mobile-sdk/features/level-compare
From 3 weeks of playing: the smart phone controls the gimbal and the camera in general, all the flight controls are strictly from the remote control unit albeit waypoints from the smart phone could be argued as control by the phone. I havent flown it without the smart phone but I am fairly sure I could, without blades on the sitting room floor it seemed possible. I cant imagine why you would want controls from your phone, when things go pear-shaped you would be pretty much screwed with just a touch screen.

Resources