Loading AI tools
From Wikipedia, the free encyclopedia
A synthetic vision system (SVS) is a computer-mediated reality system for aerial vehicles, that uses 3D to provide pilots with clear and intuitive means of understanding their flying environment.
Synthetic vision provides situational awareness to the operators by using terrain, obstacle, geo-political, hydrological and other databases. A typical SVS application uses a set of databases stored on board the aircraft, an image generator computer, and a display. Navigation solution is obtained through the use of GPS and inertial reference systems.
Highway In The Sky (HITS), or Path-In-The-Sky, is often used to depict the projected path of the aircraft in perspective view. Pilots acquire instantaneous understanding of the current as well as the future state of the aircraft with respect to the terrain, towers, buildings and other environment features.
A forerunner to such systems existed in the 1960s, with the debut into U.S. Navy service of the Grumman A-6 Intruder carrier-based medium-attack aircraft. Designed with a side-by-side seating arrangement for the crew, the Intruder featured an advanced navigation/attack system, called the Digital Integrated Attack and Navigation Equipment (DIANE), which linked the aircraft's radar, navigation and air data systems to a digital computer known as the AN/ASQ-61. Information from DIANE was displayed to both the Pilot and Bombardier/Navigator (BN) through cathode ray tube display screens. In particular, one of those screens, the AN/AVA-1 Vertical Display Indicator (VDI), showed the pilot a synthetic view of the world in front of the aircraft and, in Search Radar Terrain Clearance mode (SRTC), depicted the terrain detected by the radar, which was then displayed as coded lines that represented preset range increments. Called 'Contact Analog', this technology allowed the A-6 to be flown at night, in all weather conditions, at low altitude, and through rugged or mountainous terrain without the need for any visual references.[1]
Synthetic vision was developed by NASA and the U.S. Air Force in the late 1970s[2] and 1980s in support of advanced cockpit research, and in 1990s as part of the Aviation Safety Program. Development of the High Speed Civil Transport fueled NASA research in the 1980s and 1990s. In the early 1980s, the USAF recognized the need to improve cockpit situation awareness to support piloting ever more complex aircraft, and pursued SVS (also called pictorial format avionics) as an integrating technology for both crewed and remotely piloted systems.[3]
In 1979, the FS1 Flight Simulator by Bruce Artwick for the Apple II microcomputer introduced recreational uses of synthetic vision.[4]
NASA used synthetic vision for remotely piloted vehicles (RPVs), such as the High Maneuverability Aerial Testbed or HiMAT.[5] According to the report by NASA, the aircraft was flown by a pilot in a remote cockpit, and control signals up-linked from the flight controls in the remote cockpit on the ground to the aircraft, and aircraft telemetry downlinked to the remote cockpit displays (see photo). The remote cockpit could be configured with either nose camera video or with a 3D synthetic vision display. SV was also used for simulations of the HiMAT. Sarrafian reports that the test pilots found the visual display to be comparable to output of camera on board the RPV.[5]
The 1986 RC Aerochopper simulation by Ambrosia Microcomputer Products, Inc. used synthetic vision to aid aspiring RC aircraft pilots in learning to fly. The system included joystick flight controls which would connect to an Amiga computer and display.[6] The software included a three-dimensional terrain database for the ground as well as some man-made objects. This database was basic, representing the terrain with relatively small numbers of polygons by today's standards. The program simulated the dynamic three-dimensional position and attitude of the aircraft using the terrain database to create a projected 3D perspective display. The realism of this RPV pilot training display was enhanced by allowing the user to adjust the simulated control system delays and other parameters.
Similar research continued in the U.S. military services, and at Universities around the world. In 1995-1996, North Carolina State University flew a 17.5% scale F-18 RPV using Microsoft Flight Simulator to create the three-dimensional projected terrain environment.[7]
In 2005 a synthetic vision system was installed on a Gulfstream V test aircraft as part of NASA's "Turning Goals Into Reality" program.[8] Much of the experience gained during that program led directly to the introduction of certified SVS on future aircraft. NASA initiated industry involvement in early 2000 with major avionics manufacturers.
Eric Theunissen, a researcher at Delft University of Technology in the Netherlands, contributed to the development of SVS technology.[9]
At the end of 2007 and early 2008, the FAA certified the Gulfstream Synthetic Vision-Primary flight display (SV-PFD) system for the G350/G450 and G500/G550 business jet aircraft, displaying 3D color terrain images from the Honeywell EGPWS data overlaid with the PFD symbology.[10] It replaces the traditional blue-over-brown artificial horizon.
In 2017, Avidyne Corporation certified Synthetic Vision capability for its air navigation avionics.[11] Other glass cockpit systems such as the Garmin G1000 and the Rockwell Collins Pro Line Fusion offer synthetic terrain.
Lower-cost, non-certified avionics offer synthetic vision like apps available for Android or iPad tablet computers from ForeFlight,[12] Garmin,[13] Air Navigation Pro,[14] or Hilton Software[15]
Seamless Wikipedia browsing. On steroids.
Every time you click a link to Wikipedia, Wiktionary or Wikiquote in your browser's search results, it will show the modern Wikiwand interface.
Wikiwand extension is a five stars, simple, with minimum permission required to keep your browsing private, safe and transparent.