This is likely to be a lengthy one and I am really hoping someone has the time and patience to assist me before the 14th June 2017. I have jumped on an IoT opportunity at work for a customer that needs a solution built using Azure IoT suite. For those of you reading this that are keen on a pretty sweet project, feel free to follow the walkthrough to see if you end up on the same snag. The walkthrough is pretty detailed, see below:
1) azure.microsoft.com/en-us/resources/samples/iot-hub-c-raspberrypi-getstartedkit/#section2.8
So what I have done is I have setup an Rpi3 with Raspian Lite (headless). The walkthrough suggests using an Adafruit BME280 multi-sensor which I haven't connected yet, but I will be buying one this week to do that. I didn't think it would be necessary to have it connected upfront as the majority of the steps in the walkthrough are about getting the Azure IoT Suite remote monitoring solution setup. The pi setup is straight forward and for me to connect to it I used raspi-config to enable SSH and are able to connect to the device without a glitch.
Everything in the walkthrough goes as expected on the pi up until the point where I get to step 1.6. In this step there is a line that suggests running the following code:
~/iot-hub-c-raspberrypi-getstartedkit/samples/build.sh
See the following image:
Image 1: This is where build.sh code initially fails
Image 2: This is where build.sh continues to fail and then quits
So, considering I am a sales guy with very limited technical expertise, and consider myself more of a tinkerer than a seasoned coder, the C code in this build.sh, seems to hit a snag and my problem is that I cannot seem to understand what the failure reason is trying to tell me. I'd really appreciate anyone's assistance here as I'm scrambling for time to have this up and running for my customer meeting before the 14th June 2017. If anyone has the capacity to assist me, I'd really appreciate it more than you know :-)
Related
Hopefully this is the best place to ask this... We have a Google Mini that now gets stuck on the Enterprise Network service step and won't go past that. Is there anything we can do? We have probably been out of support for around 8 years, and everyone who set it up has left our organization. I found this post about it https://groups.google.com/forum/#!topic/Google-Search-Appliance-Help/91AhqGXc4HM, but the link at the end doesn't work anymore... Hopefully someone has a solution to that problem, that doesn't involve buying a new Search Appliance...
You cannot directly edit the configuration of the Google Mini unless you can boot it up and get to the configuration services. The hard drives are encrypted.
Since the Mini is a licensed piece of hardware and will have expired by now you won't be able to have Support log in to resolve the issue.
You should look at replacement technology at this point.
Maybe I have't looked hard enough, but I spent yesterday googling for a bit and found no relevant projects on hacking the DJI Phantom Drone in order to create new coordinating apps. This is besides the app for coordination DJI currently uses for their drone. I'm trying to see if there's a way to communicate with the Drone with a specific protocol in order to accept a set of procedures.
Any help would be awesome,
Thanks.
Great News for you and all us Droneys! DJI has launched their SDK since you asked this question. They released it last November and you can now apply for a license and write your own apps for the Phantom2 Vision+ using their SDK.
Check it out at https://developer.dji.com/
I am already building a project using the SDK - you can follow my progress on my blog / product site. I will also try to update it with good DJI related development links and tips.
This post is old but I think it is good to leave a foot print for others :)
There is this new company called NVdrones, which created a peace of hardware that you can attach to any drone (you need physical access to the flight controller), and once you do that you can use their SDK (Arduino, Java, Android and Javascript) to write your app without the need of hacking, soldering or anything else. It is just plug and play.
Another benefit is that you are not locked with a specific drone (DJI SDK or 3DRobotics SDK), you can use the board on anything you want. Which gives lots of flexibility.
The developer site is http://developers.NVdrones.com
Hope this helps.
This is a great topic!
You could check how to hack your copter here: https://github.com/flyver/Flyver-SDK/wiki/-2.2--How-To:-Flyver-Hack-a-Copter
By opening the drone, taking out the original controller, soldering a few wires and sticking an Android phone to it, you will have the ability to program your Phantom in a modern manner with an open source SDK and application based development. This means that you could add computer vision to it, automation or additional hardware. You could also use smartphones, web and other interactive devices for remote controlling the copter instead of using the standard remote controls.
The Phantom, however, is offcenter balanced due to the fact that most people use gimbal with it. Without the gimbal is a lot less stable from my experiments so you will have to put some extra work in center balancing it.
I'm aware this isn't exactly a programming question, but it directly impacts our developers and the code we're assigned to write. If there's another SO-like forum where this could be better posted, please let me know and I'll take the question down from here & post it there.
Our work environment is a couple of developers creating (20-30%) and maintaining (lion's share) legacy software for factory production floor and test workers to use to calibrate or test the equipment the company sells. We've implemented a very simple Google form based bug reporting page, but we're already running into problems of scale (approx 40:1 them:us and lots of old-old buggy software that we didn't write). The company has tried using Bugzilla before my arrival with little success, the factory folks were apparently intimidated by it and wouldn't use it. However, they seem to like the simple Google form and the wizard-like steps to file a bug or request a feature. We're currently manually cutting & pasting their bug/feature requests from the Google form spreadsheet into Trac, and manually tracking the bugs/feature requests on a white board with magnetic bug cards. We're only a few weeks into this system and it's already showing it fragility and lack of scalability.
Ideally we'd have a Windows >= XP web or desktop client that would provide:
Simplified bug reporting, a Wizard like approach seems to work well
Customizable for our software packages (like drop downs for each)
Bugzilla or Trac integration
Standard bug tracking features developers and management can use
I've found the winners of the "Make Bugzilla Pretty" contest, but coming from a pure software house where we just used straight Bugzilla out of the box, I'm unclear on how to configure and install these skins. Obviously I can figure this out but don't want to go down that path if it's not going to solve our basic problem which is non-technical people reporting bugs.
TaskCompiler, found on the Bugzilla wiki site seemed like a candidate because it talks to both Bugzilla & Trac, but their sales page is offline and the site hasn't been updated since 2012 and I'm unsure as to their viability.
I'm certain we're not the first production facility to run into problems like this, I'm looking for recommendations to help solve both our scalability as well as-ease-of-use problem.
Another thought that occurs to me is a GAS script to push our current Google forms based bug reports into Trac or Bugzilla.
Edit: The decision between Bugzilla/Trac seems to have been made for us. I'm exploring options for using Trac here if you want to follow along.
Okay so I am not sure if a lot of you have started to work on Microsoft Kinect for Windows that has been released in February 2012. I am trying to start developing and I wanted to know if there are any tutorials I can find on how to use the SDK or if someone can guide me How the RGB stream can be captured using the Kinect?
There are many tutorials. Some can be found at Channel 9's Kinect Quick Start Series , Channel 9 also has many articles on Kinect. All of the classes and variables found in the SDK can be found at MSDN, on Rob Relyea's Blog there are many tutorials. And if you ever are struggling, you can visit the Kinect Development Chatroom (assuming you have 20 rep).
Hope this helps!
Personally, I wouldn't start with Channel 9, or any tutorials for that matter. The most enjoyable way to jump into the Kinect and start messing around with stuff is to install the Developer Toolkit. It was update 3 days ago to include some really cool 3D point cloud stuff. Download/install the toolkit, run the Kinect Studio application it comes with, and spend some time checking out what the Kinect is capable of. If you see something of interest, install it to your computer and open it in Visual Studio. If you don't have Visual Studio, you can download the C# Express version for free. The source code is all very well commented and I find that's the best way to learn by example. You don't have to sit through Channel 9's sometimes painful videos or spend time reading a blog, you can just jump in and have fun with it. If you get stuck, then refer back to Channel9 or come back to Stack Overflow.
The best place to start learning is MSDN, and where you got the driver for kinect. They offer many tutorials and videos that explain most concepts for the kinect.
You can refer Kinect 1.0 for kinect for Windows SDk 1.0
I've been trying to find a guide on how to get Team Foundation Server to turn on a lava lamp or traffic light to indicate the status of the build. I want to set up something that's visible right across the office so there's some peer pressure to encourage developers not to break the build; but I also want it to be fun.
There's a lot of examples for CruiseControl that use X.10 devices which seems like a good way to go. But I can't find anything similar for TFS. I'm sure that somebody must be doing this somewhere...?
Using X.10 has one problem in that it requires a serial port - but our TFS is completely virtualised in a data centre somewhere. Maybe there is some way to trigger the traffic light via an email?
Any advice appreciated. Thanks.
The TFS have got a nice API for getting the status of recent builds. You can use the API and design your own fun system.
Also take a look at:
TFS Build Monitor
TFS Build Light
At some point I had stumbled upon this youtube-video, where Martin Woodward presents Brian, the funky TFS-build bunny. Might be worth checking. It might also be worth checking this SO post.
The TFS API's are terrible they're a pain to do yourself. You could start with this open source project on Google Code: http://code.google.com/p/siren-of-shame/. That project is designed to work with a couple of different build servers, but everything is broken out, so you could start with the TFS 2010 project (TfsServices.csproj). Or if you don't want to do it all yourself that project is designed to work with a USB Siren that they sell (see http://www.sirenofshame.com/).