|Subject:||Re: [Paparazzi-devel] FPV app on AppStore for H.264 wifi video links|
|Date:||Wed, 20 Mar 2013 17:04:28 +0100|
I'd like to announce something not immediately related to paparazzi *now*, although this may change in the future…
I've been working on an affordable, accessible and low-latency (HD) video downlink (for consumers and hobbyists) for a year or so
using my own resources. Software wise things have been going smoothly, but hardware wise there are still issues to resolve.
There's a forum post where a lot of research so far is documented on fpvlab.com:
The ground station software that decodes and shows the transmitted video is now stable enough that I released this as
an app on the AppStore. It's called "FPV" (FIrst Person View) as it targets mostly hobbyists, but newer features that I have
in the pipeline may make this very interesting for attempts to integrating this with some paparazzi ground station modules or
the autopilot itself. The FPV app decodes H.264 video over wifi and uses some Long Range wifi modules to get all the
data across. (actually, in theory, paparazzi could already use a wifi link for telemetry using the w5100 module that I created, but that still
needs some work to remove the wait loops and other cleanups. ). There has been some other work done with Gumstix I believe.
The link to the App in the AppStore is here:
A video showing the app with a video stream from a simulator is here:
( the video stream is created by running the FlightGear simulator on Linux, recording the desktop in realtime and transferring this over
my private network. The telemetry is taken from FlightGear as well and used to paint the OSD ). It's also interesting to note that the OSD
at the top shows the "technical" info that you'd typically find on an OSD designed by an engineer. The "bottom" view is more of a
view designed after Cognitive Work Analysis.
The manual, protocol description and more information on the hardware that I used for field testing is here:
These new features mentioned are all about showing virtual cues in the streamed image. Think "virtual tunnel" to fly a specific trajectory manually,
points of interest and how these pois could be generated by touching a display showing the same streamed image.
The idea behind this is that spectators who're watching along with an FPV flight might want to request flying over
a specific area or point and how they'd communicate that exact point with the pilot. In more serious applications
you'd probably have a vehicle with an AP. It is then necessary to communicate effectively between team members on the ground.
One person is scanning a camera image for POIs and at some point needs to communicate his perceptions with the
flight operator. How do you get this POI seen from a camera image quickly on a 2D map so that navigation changes
can be made effectively?
Paparazzi-devel mailing list
|[Prev in Thread]||Current Thread||[Next in Thread]|