Curiosity, NASA’s newest Mars rover, is continuing its journey to its first major science destination, Glenelg. Covering nearly 40 yards per day, it scans the surface for interesting rocks, using its mast-mounted camera. In order to improve this ability, NASA is planning to deploy new software that will make it simpler to target and capture images while moving.
The new software, dubbed the Autonomous Exploration for Gathering Increased Science (AEGIS), will allow the rover to look for rocks of certain shapes, sizes, or brightnesses. “The idea is, when it’s doing a long drive, a scientist can say, ‘Oh, if you see this type of rock in the middle of this drive or at the end of the drive, go ahead and take some high-quality, high-resolution images of it before the rover moves past,’ said Tara Estlin, a senior member of NASA Jet Propulsion Laboratory’s Artificial Intelligence Group and AEGIS project leader. Earlier this week, Curiosity parked in front of a football-size rock, which it photographed. NASA scientists plan to use the rover’s spectrometer to determine the rock’s composition.”
“AEGIS, which has been used on the Mars Exploration Rover Opportunity since 2009, will be installed on Curiosity in the next nine to 12 months, Estlin said in an interview with InformationWeek. The AEGIS software, developed by JPL, was named NASA’s ‘software of the year’ in 2011,” Information Week writes.
“AEGIS will let the rover automatically key in on geological features of interest to the Curiosity project team as the rover moves along on its research mission. An automated image-capture process is important because ‘we can’t stop at every point and wait a day or two for ground to be in the loop,’ Estlin said.”
“JPL developed AEGIS on Linux-based systems, then tested the software on research rovers. Opportunity uses the software to take a wide-angle image with a low-resolution camera, then picks out rocks in the image to see if there’s something of interest. If so, it takes a high-resolution image using an on-board science camera that’s capable of zooming in on the subject. The software has potential beyond picture taking. Its see-and-react code could be adapted to other instruments.”
“We have versions of the software that could stop the rover and call home if it sees something really interesting,” Estlin said. “You could [then] drive closer to the object of interest.”
The software is actually being considered for use in future space missions, when flying by asteroids or planets. “There are all sorts of surface missions that often don’t have a lot of time to determine a response,” Estlin said. “The AEGIS capability was developed as part of a framework called the Onboard Autonomous Science Investigation System (OASIS), which was designed to let a rover to identify and react to ‘serendipitous science opportunities,’ according to NASA.”