At the very end of the DJI Simulator Tutorial it says this:
This demo is a simple demonstration of using DJISimulator, to have a better user experience, you can create a 3D simulated environment using 3D game engine like Unity3Dto show the simulated data and aircraft flight behavious inside your mobile application (Like the Flight Simulator in DJI Go app)!
I have been looking for a way to integrate a custom 3D model/environment objects into the simulation environment, but there does not seem to be a way. I can´t find any forum posts in regard to this either.
Does anyone have experience doing something like this or can point me in the direction?
Thanks!
It looks like there is not any sample integrating a custom 3D mode.
But there are some states such as 'roll','yaw','pitch', etc in DJISimulatotState.
It is helpful to build another customized simulator using 3D game engine.
https://developer.dji.com/api-reference/android-api/Components/Simulator/DJISimulator_DJISimulatorState.html#djisimulator_djisimulatorstate
Related
I am useing DJI Mobile SDK to crtate an APP by Android Studio now.I want to know how to use the GPS signal of the aircraft and the phone to realize position control. Is there any API in the DJI Mobile SDK I can use?
You may follow this sample to run simple higher-level GPS position control https://developer.dji.com/mobile-sdk/documentation/ios-tutorials/GSDemo.html
If you stop at some waypoint, it can automatically hold the position. It is a simple recreation of the DJI pilot app waypoint planning.
For low-level GPS position control requires a higher understanding of the system. This usually allows interesting applications such as allow drone to follow some person or precision landing to some mark or circle around some tower. There is not much open-source implementation available on the internet. You have to search in the MSDK for the API for some basic control and you also need to have deep understanding in the field that you are trying to achieve e.g real-time object detection, low-level control framework, Visual-Inertial SLAM etc
I am new to the drone community. I am trying to use a drone to follow a driving vehicle and record a video, this would be part of my graduate school project.
I know DJI provides an "active track" function but it seems like the maximum speed for the tracking is around 20-25 mph, and it cannot track the vehicle at a top-down angle (drone looks straight down at the vehicle).
I have an idea to send the GPS info of the ego vehicle to my mobile device and use a customized app to read the GPS location and set it as a target so the drone can follow it.
Is it doable? Is it in general worth the effort? Or is it better off to just use the active track and work with the best angle that I can get?
I've done that.
Send gps updates from a tracker phone to the phone connected to rc.
You have to use virtualstick mode in the MSDK. Tha max speed is 15m/s
Virtual stick functions in the DJI Mobile SDK simulate the remote controller's joysticks, and therefore an aircraft can be automated to fly in any way a human can manually fly it. Compared to missions, this is a more complicated, but flexible way to automate flight.
Works very well.
An example:
https://youtu.be/i3axYfIOHTY?t=55
It's a little shaky, but thats fixed now :-)
Hello ~ I am a Korean student who is studying how to make 'DJI SPARK' app using Android Studio. The app I want to create is to allow the drones to shoot a stationary object with a certain pattern(For example, a pattern that rotates about an object constantly). However, I could not find information on the homepage.
I uploaded the DJI Sample Code to Android Studio, entered an app key on the Android Manifest.xml file. Then I think I need to put coding into MissionManager, but I do not know how to make the code to implement the pattern. I need help from several developers. Thank you.
I can help but most of what you ask sounds specific to your project and I don't want to give you the answer.
I will gladly point you into the correct directions and help you locale the specific areas of the documentation in order to compete for project.
You will want to use DJI's APIs under Mission Classes - specifically the HotpointMissionOperator and WaypointMissionOperator
The Hotpoint Mission will allow your drone to fly circles around a specified point. If you want to fly a different pattern other than a circle then you can plot out a series of points with Waypoint Missions in whichever pattern you designate.
I would like to know how to get the current feature points used in motion tracking and the ones that are present in the learned area (detected or not).
There is an older, related post without an useful answer:
How is it possible to get tracked features from tango APIs used for motion tracking. I'm using the tango to not do SLAM and IMU-integration on my own.
What do I need to do, to visualize the tracked features like they did in some of the presentation videos. https://www.youtube.com/watch?v=2y7NX-HUlMc (0:35 - 0:55)
What I want in general is some kind of measure or visual guidance on how good the devices learned the current environment. I know, there is is the Inspector App but I need this information on the fly.
Thanks for your Help ;)
If you want to check if an area is present in your learned area model and which is not, you can use the Tango Debug Overlay App. It has a field 'Tracking Success' that only counts up if the device sees learned feature points (ADF on) or finds new feature points (ADF off) (http://grauonline.de/alexwww/tmp/tango_debug_overlay_app.jpg). Additionally, you can request that debug information like Tango Debug Overlay App does (as a simple text) via UDP port 29361 in your App and parse the returned debug text (although this is not recommended at all for a real app as this interface is not documented)
PS: In Tango Core 01-19-2017 this counter does not seem to work anymore.
What I understand from technical specs of Google glass is that it displays a 2D plane on one of the eye's projector. Android sdk in addition with GDK provides tools for writing apps for the device with features that can sense eye and voice actions. But, it does not provide 3D stereoscopic vision as this would require projector on both eyes.
On the other hand Epson Moverio promises true 3D augmented reality experience Having used Moverio, I can see two projector for both eyes that is able to project steroscopic images.
Perhaps I should have done a more extensive research regarding the spectrum of products/toolkit available, still I have some Questions/Doubts of which until now I could not find any information.
Q1. Does google provide any 2-eye-projector kind of glasses product?
ANS: No
Q2. Does google glasses development kit (the api) provides features for generating left & right views for a 3D object for EPSON Moverio? I have seen that Wikitude and Metaio comes with these kind of features. Did google provide any support in gdk?
ANS: NO. Not from google.
Q3. Does Epson plan to roll out any developer's tool for easily create 3D markers and plot them in the projected space?
ANS: Not announced yet from EPSON.
There is no current support in Google Glass for stereoscopic views.