NASA has pioneered the development of all kinds of robots and robotic systems. Beyond its0 orbiters and satellites, which have been exploring the planets and bodies of the Solar System for decades, there’s also the growing army of landers and rovers that have been exploring planetary surfaces. Aboard the ISS, they even have floating robots (like CIMON) and humanoid robot helpers – a la Robonaut and Robonaut 2.
Looking to the future, NASA hopes to build robots that can do even more. While the current generation of rovers can drive across the plains and craters of Mars, what if they could explore cliffs, polar ice caps and other hard-to-reach places? That is the purpose behind the Limbed Excursion Mechanical Utility Robot (LEMUR) that is currently being developed by engineers at NASA’s Jet Propulsion Laboratory.
If something called “Project METERON” sounds to you like a sinister project involving astronauts, robots, the International Space Station, and artificial intelligence, I don’t blame you. Because that’s what it is (except for the sinister part.) In fact, the Meteron Project (Multi-Purpose End-to-End Robotic Operation Network) is not sinister at all, but a friendly collaboration between the European Space Agency (ESA) and the German Aerospace Center (DLR.)
The idea behind the project is to place an artificially intelligent robot here on Earth under the direct control of an astronaut 400 km above the Earth, and to get the two to work together.
“Artificial intelligence allows the robot to perform many tasks independently, making us less susceptible to communication delays that would make continuous control more difficult at such a great distance.” – Neil Lii, DLR Project Manager.
On March 2nd, engineers at the DLR Institute of Robotics and Mechatronics set up the robot called Justin in a simulated Martian environment. Justin was given a simulated task to carry out, with as few instructions as necessary. The maintenance of solar panels was the chosen task, since they’re common on landers and rovers, and since Mars can get kind of dusty.
The first test of the METERON Project was done in August. But this latest test was more demanding for both the robot and the astronaut issuing the commands. The pair had worked together before, but since then, Justin was programmed with more abstract commands that the operator could choose from.
American astronaut Scott Tingle issued commands to Justin from a tablet aboard the ISS, and the same tablet also displayed what Justin was seeing. The human-robot team had practiced together before, but this test was designed to push the pair into more challenging tasks. Tingle had no advance knowledge of the tasks in the test, and he also had no advance knowledge of Justin’s new capabilities. On-board the ISS, Tingle quickly realized that the panels in the simulation down here were dusty. They were also not pointed in the optimal direction.
This was a new situation for Tingle and for Justin, and Tingle had to choose from a range of commands on the tablet. The team on the ground monitored his choices. The level of complexity meant that Justin couldn’t just perform the task and report it completed, it meant that Tingle and the robot also had to estimate how clean the panels were after being cleaned.
“Our team closely observed how the astronaut accomplished these tasks, without being aware of these problems in advance and without any knowledge of the robot’s new capabilities,” says DLR engineer Daniel Leidner.
The next test will take place in Summer 2018 and will push the system even further. Justin will have an even more complex task before him, in this case selecting a component on behalf of the astronaut and installing it on the solar panels. The German ESA astronaut Alexander Gerst will be the operator.
If the whole point of this is not immediately clear to you, think Mars exploration. We have rovers and landers working on the surface of Mars to study the planet in increasing detail. And one day, humans will visit the planet. But right now, we’re restricted to surface craft being controlled from Earth.
What METERON and other endeavours like it are doing, is developing robots that can do our work for us. But they’ll be smart robots that don’t need to be told every little thing. They are just given a task and they go about doing it. And the humans issuing the commands could be in orbit around Mars, rather than being exposed to all the risks on the surface.
“Artificial intelligence allows the robot to perform many tasks independently, making us less susceptible to communication delays that would make continuous control more difficult at such a great distance,” explained Neil Lii, DLR Project Manager. “And we also reduce the workload of the astronaut, who can transfer tasks to the robot.” To do this, however, astronauts and robots must cooperate seamlessly and also complement one another.
That’s why these tests are important. Getting the astronaut and the robot to perform well together is critical.
“This is a significant step closer to a manned planetary mission with robotic support,” says Alin Albu-Schäffer, head of the DLR Institute of Robotics and Mechatronics. It’s expensive and risky to maintain a human presence on the surface of Mars. Why risk human life to perform tasks like cleaning solar panels?
“The astronaut would therefore not be exposed to the risk of landing, and we could use more robotic assistants to build and maintain infrastructure, for example, with limited human resources.” In this scenario, the robot would no longer simply be the extended arm of the astronaut: “It would be more like a partner on the ground.”
For over a decade, robots have been exploring Mars in advance of the crewed missions that are being planned for the coming decades. And when it comes time for astronauts to set foot on the Red Planet, they will be looking for robots to help them with some of the legwork. After all, exploring Mars is tough, laborious, and dangerous work, so some robotic assistance will probably be necessary.
These efforts are being led Professor Russ Tedrake, an electrical engineer and computer programmer who helped program the Atlas robot to take part in the 2015 DARPA Robotics Challenge. Together with members of an advanced independent research group – known as the Super Undergraduate Research Opportunities Program (SuperUROP) – he is getting this R5 robot ready for NASA’s Space Robotics Challenge.
As part of NASA’s Centennial Challenges Program, and with a prize purse of $1 million, this competition aims to push the boundaries of what robots are capable of in the realm of space exploration. In addition to MIT, Northeastern University and the University of Edinburgh have been tasked with programming an R5 to complete tasks normally handled by astronauts.
Ultimately, the robots will be tested in a simulated environment and judged based on their ability to complete three tasks. These include aligning a communications array, repairing a broken solar array, and identifying and repairing a habitat leak. There will also be a qualifying round where teams will be tasked with demonstrating autonomous tracking abilities (which will have to be completed in order to move towards the main round).
Naturally, this presents quite a few challenges. NASA designed the R5 robot to be capable of performing human tasks and move like a human being as much as possible, which necessitated a body with 28 torque-controlled joints. However, getting those joints to work together to perform mission-related work and operate independently is a bit of a challenge.
In short, the robot is not like other robotic missions – such as the Opportunity or Curiosity rovers. Instead of having a human being pushing levers to get them to move about and collect samples, the R5 will be tasked with things like opening airlock hatches, attaching and removing power cables, repairing equipment, and retrieving samples all on its own. And of course, if it takes a spill and falls down, it will have to be able to get up on its own.
With the help of the special algorithms being generated by Tedrake and his colleagues – as well as other teams competing in this challenge – robots could play an important role in future missions. This could involve robots selecting landing sites for astronaut crews, setting up habitats in advance of crews arriving, and even conducting preliminary research on celestial bodies.
In addition, robots could take the place of crews on long-distance missions (such as Europa). Instead of sending a crew that would require months of food and supplies, a robot crew could be dispatched to the Jovian moon to collect ice samples, explore the surface, and interface with drones being sent to explore the interior ocean. And if the mission failed, there would be no grieving families (just grieving robotics teams).
And now to address the elephant in the room. The idea of sending robot explorers on space missions to help astronauts (or even replace them) is sure to make some people out there nervous. But for those who fear that this might bring one step closer to a robot revolution, rest assured that the machines are nowhere near where they’d need to be to go all “Judgement Day” on us just yet.
Long before they can launch nuclear weapons, pick up laser guns and stalk us through a post-apocalyptic landscape, or start upgrading themselves to look (and feel) human, robots will first need to master the simple tasks of walking upright and holding a screwdriver.
Still, if any of the robots end up having creepy red visor eyes (or saying things like “by your command”), we might want to consider including the Three Laws of Robotics in their programming. It’s never too soon to make sure they can’t turn on humanity!
Registration for the Space Robotics Challenge opened in August, 2016. The qualifying round, which began in mid-October, will run until mid-December. Finalists of that round will be announced in January, with the final virtual competition taking place in June 2017. The winning team will be awarded $500,000 over a two year period from NASA’s Space Technology Mission Directive.
NASA’s Jet Propulsion Laboratory recently announced that it is developing a small drone helicopter to scout the way for future Mars rovers. Why would Mars rovers need such a robotic guide? The answer is that driving on Mars is really hard.
Here on Earth, robots exploring volcanic rims, or assisting rescuers, can be driven by remote control, with a joystick. This is because radio signals reach the robot from its control center almost instantly. Driving on the moon isn’t much harder. Radio signals traveling at the speed of light take about two and half seconds to make the round trip to the moon and back. This delay isn’t long enough to seriously interfere with remote control driving. In the 1970’s Soviet controllers drove the Lunokhod moon rovers this way, successfully exploring more than 40 km of lunar terrain.
Driving on Mars is much harder, because it is so much further away. Depending on its position with respect to Earth, signals can take between 8 and 42 minutes for the round trip. Pre-programmed instructions must be sent to the rover, which it then executes on its own. Each Martian drive takes hours of careful planning. Stereo images taken by the rover’s navigation cameras are carefully scrutinized by engineers. Images from spacecraft orbiting Mars sometimes provide additional information.
A rover can be programmed either to simply execute a list of driving commands sent from Earth, or it can use images taken by its navigation cameras and processed by its on-board computers to measure speed and detect obstacles or hazards by itself. It can even plot its own safe path to a specified goal. Drives based on instructions from the ground are the fastest.
The Mars Exploration Rovers Spirit and Opportunity could drive up to 124 meters in an hour this way. This corresponds to about the length of an American football field. But this mode was also the least safe.
When the rover actively guides itself with its cameras, progress is safer, but much slower because of all the image processing needed. It may progress by as little as 10 meters an hour, which is about the distance from the goal line to the 10 yard line on an American football field. This method must be used whenever the rover doesn’t have a clear view of the route ahead, which is often the case due to rough and hilly terrain.
As of early 2015, the farthest Curiosity has driven in a single day is 144 meters. Opportunity’s longest daily drive was 224 meters, a distance the length of two American football fields.
If ground controllers could get a better view of the path ahead, they could devise instructions allowing a future rover to safely drive much further in a day.
That’s where the idea of a drone helicopter comes in. The helicopter could fly out ahead of the rover every day. Images made from its aerial vantage point would be invaluable to ground controllers for identifying points of scientific interest, and planning driving routes to get there.
Flying a helicopter on Mars poses special challenges. One advantage is that Martian gravity is only 38% as strong as that of Earth, so that the helicopter wouldn’t need to generate as much lift as one of the same mass on Earth. A helicopter’s propeller blades generate lift by pushing air downward. This is harder to do on Mars than on Earth, because the Martian atmosphere is on hundred times thinner. To displace enough air, the propeller blades would need to spin very quickly, or to be very large.
The copter must be capable of flying on its own, using prior instructions, maintaining stable flight along a pre-specified route. It must land and take off repeatedly in rocky Martian terrain. Finally it must be capable of surviving the harsh conditions of Mars, where the temperature plummets to 100 degrees Fahrenheit or lower every night.
The JPL engineers designed a copter with a mass of 1 kilogram; a tiny fraction of the 900 kg mass of the Curiosity rover. Its propeller blades span 1.1 meters from blade tip to blade tip, and are capable of spinning at 3400 rotations per minute. The body is about the size of a tissue box.
The copter is solar powered, with a disk of solar cells gathering enough power every day to power a flight of two to three minutes and to heat the vehicle at night. It can fly about half a kilometer in that time, gathering images for transmission to ground control as it goes. Engineers expect that the reconnaissance that the drone copter gathers will be invaluable in planning a rover’s drives, tripling the distance that can be traveled in a day.
References and further reading:
Thanks to Mark Maimone of NASA Jet Propulsion Laboratory for information about the daily driving distances of Curiosity and Opportunity.
With robotic spacecraft, we have explored, discovered and expanded our understanding of the Solar System and the Universe at large. Our five senses have long since reached their limits and cannot reveal the presence of new objects or properties without the assistance of extraordinary sensors and optics. Data is returned and is transformed into a format that humans can interpret.
Humans remain confined to low-Earth orbit and forty-three years have passed since humans last escaped the bonds of Earth’s gravity. NASA’s budget is divided between human endeavors and robotic and each year there is a struggle to find balance between development of software and hardware to launch humans or carry robotic surrogates. Year after year, humans continue to advance robotic capabilities and artificial intelligence (A.I.), and with each passing year, it becomes less clear how we will fit ourselves into the future exploration of the Solar System and beyond.
Is it a race in which we are unwittingly partaking that places us against our inventions? And like the aftermath of the Kasparov versus Deep Blue chess match, are we destined to accept a segregation as necessary? Allow robotics, with or without A.I., to do what they do best – explore space and other worlds?
Should we continue to find new ways and better ways to plug ourselves into our surrogates and appreciate with greater detail what they sense and touch? Consider how naturally our children engross themselves in games and virtual reality and how difficult it is to separate them from the technology. Or is this just a prelude and are we all antecedents of future Captain Kirks and Jean Luc Picards?
Approximately 55% of the NASA budget is in the realm of human spaceflight (HSF). This includes specific funds for Orion and SLS and half measures of supporting segments of the NASA agency, such as Cross-Agency Support, Construction and Maintenance. In contrast, appropriations for robotic missions – project development, operations, R&D – represent 39% of the budget.
The appropriation of funds has always favored human spaceflight, primarily because HSF requires costlier, heavier and more complex systems to maintain humans in the hostile environment of space. And while NASA budgets are not nearly weighted 2-to-1 in favor of human spaceflight, few would contest that the return on investment (ROI) is over 2-to-1 in favor of robotic driven exploration of space. And many would scoff at this ratio and counter that 3-to-1 or 4-to-1 is closer to the advantage robots have over humans.
Politics play a significantly bigger role in the choice of appropriations to HSF compared to robotic missions. The latter is distributed among smaller budget projects and operations and HSF has always involved large expensive programs lasting decades. The big programs attract the interest of public officials wanting to bring capital and jobs to their districts or states.
NASA appropriations are complicated further by a rift between the White House and Capitol Hill along party lines. The Democrat-controlled White House has favored robotics and the use of private enterprise to advance NASA while Republicans on the Hill have supported the big human spaceflight projects; further complications are due to political divisions over the issue of Climate Change. How the two parties treat NASA is the opposite to, at least, how the public perceives the party platforms – smaller government or more social programs, less spending and supporting private enterprise. This tug of war is clearly seen in the NASA budget pie chart.
The House reduced the White House request for NASA Space Technology by 15% while increasing the funds for Orion and SLS by 16%. Space Technology represents funds that NASA would use to develop the Asteroid Redirect Mission (ARM), which the Obama administration favors as a foundation for the first use of SLS as part of a human mission to an asteroid. In contrast, the House appropriated $100 million to the Europa mission concept. Due to the delays of Orion and SLS development and anemic funding of ARM, the first use of SLS could be to send a probe to Europa.
While HSF appropriations for Space Ops & Exploration (effectively HSF) increased ~6% – $300 million, NASA Science gained ~2% – $100 million over the 2014 appropriations; ultimately set by Capitol Hill legislators. The Planetary Society, which is the Science Mission Directorate’s (SMD) staunchest supporter, has expressed satisfaction that the Planetary Science budget has nearly reached their recommended $1.5 billion. However, the increase is delivered with the requirement that $100 million shall be used for Europa concept development and is also in contrast to cutbacks in other segments of the SMD budget.
Note also that NASA Education and Public Outreach (EPO) received a significant boost from Republican controlled Capital Hill. In addition to the specific funding – a 2% increase over 2014 and 34% over the White House request, there is $42 million given specifically to the Science Mission Directorate (SMD) for EPO. The Obama Adminstration has attempted to reduce NASA EPO in favor of a consolidated government approach to improve effectiveness and reduce government.
The drive to explore beyond Earth’s orbit and set foot on new worlds is not just a question of finances. In retrospect, it was not finances at all and our remaining shackles to Earth was a choice of vision. Today, politicians and administrators cannot proclaim ‘Let’s do it again! Let’s make a better Shuttle or a better Space Station.’ There is no choice but to go beyond Earth orbit, but where?
While the International Space Station program, led by NASA, now maintains a continued human presence in outer space, more people ask the question, ‘why aren’t we there yet?’ Why haven’t we stepped upon Mars or the Moon again, or anything other than Earth or floating in the void of low-Earth orbit. The answer now resides in museums and in the habitat orbiting the Earth every 90 minutes.
The retired Space Shuttle program and the International Space Station represent the funds expended on human spaceflight over the last 40 years, which is equivalent to the funds and the time necessary to send humans to Mars. Some would argue that the funds and time expended could have meant multiple human missions to Mars and maybe even a permanent presence. But the American human spaceflight program chose a less costly path, one more achievable – staying close to home.
Ultimately, the goal is Mars. Administrators at NASA and others have become comfortable with this proclamation. However, some would say that it is treated more as a resignation. Presidents have been defining the objectives of human spaceflight and then redefining them. The Moon, Lagrangian Points or asteroids as waypoints to eventually land humans on Mars. Partial plans and roadmaps have been constructed by NASA and now politicians have mandated a roadmap. And politicians forced continuation of development of a big rocket; one which needs a clear path to justify its cost to taxpayers. One does need a big rocket to get anywhere beyond low-Earth orbit. However, a cancellation of the Constellation program – to build the replacement for the Shuttle and a new human-rated spacecraft – has meant delays and even more cost overruns.
During the ten years that have transpired to replace the Space Shuttle, with at least five more years remaining, events beyond the control of NASA and the federal government have taken place. Private enterprise is developing several new approaches to lofting payloads to Earth orbit and beyond. More countries have taken on the challenge. Spearheading this activity, independent of NASA or Washington plans, has been Space Exploration Technologies Corporation (SpaceX).
SpaceX’s Falcon 9 and soon to be Falcon Heavy represent alternatives to what was originally envisioned in the Constellation program with Ares I and Ares V. Falcon Heavy will not have the capability of an Ares V but at roughly $100 million per flight versus $600 million per flight for what Ares V has become – the Space Launch System (SLS) – there are those that would argue that ‘time is up.’ NASA has taken too long and the cost of SLS is not justifiable now that private enterprise has developed something cheaper and done so faster. Is Falcon Nine and Heavy “better”, as in NASA administrator Dan Golden’s proclamation – ‘Faster, Better, Cheaper’? Is it better than SLS technology? Is it better simply because its cheaper for lifting each pound of payload? Is it better because it is arriving ready-to-use sooner than SLS?
Humans will always depend on robotic launch vehicles, capsules and habitats laden with technological wonders to make our spaceflight possible. However, once we step out beyond Earth orbit and onto other worlds, what shall we do? From Carl Sagan to Steve Squyres, NASA scientists have stated that a trained astronaut could do in just weeks what the Mars rovers have required years to accomplish. How long will this hold up and is it really true?
Since Chess Champion Garry Kasparov was defeated by IBM’s Deep Blue, there have been 8 two-year periods representing the doubling of transistors in integrated circuits. This is a factor of 256. Arguably, computers have grown 100 times more powerful in the 17 years. However, robotics is not just electronics. It is the confluence of several technologies that have steadily developed over the 40 years that Shuttle technology stood still and at least 20 years that Space Station designs were locked into technological choices. Advances in material science, nano-technology, electro-optics, and software development are equally important.
While human decision making has been capable of spinning its wheels and then making poor choices and logistical errors, the development of robotics altogether is a juggernaut. While appropriations for human spaceflight have always surpassed robotics, advances in robotics have been driven by government investments across numerous agencies and by private enterprise. The noted futurist and inventor Ray Kurzweil who predicts the arrival of the Singularity by around 2045 (his arrival date is not exact) has emphasized that the surpassing of human intellect by machines is inevitable due to the “The Law of Accelerating Returns”. Technological development is a juggernaut.
In the same year that NASA was founded, 1958, the term Singularity was first used by mathematician John von Neumann to describe the arrival of artificial intelligence that surpasses humans.
Unknowingly, this is the foot race that NASA has been in since its creation. The mechanisms and electronics that facilitated landing men on the surface of the Moon never stopped advancing. And in that time span, human decisions and plans for NASA never stopped vacillating or stop locking existing technology into designs; suffering delays and cost overruns before launching humans to space.
So are we destined to arrive on Mars and roam its surface like retired geologists and biologists wandering in the desert with a poking stick or rock hammer? Have we wasted too much time and has the window passed in which human exploration can make discoveries that robotics cannot accomplish faster, better and cheaper? Will Mars just become an art colony where humans can experience new sunrises and setting moons? Or will we segregate ourselves from our robotic surrogates and appreciate our limited skills and go forth into the Universe? Or will we mind meld with robotics and master our own biology just moments after taking our first feeble steps beyond the Earth?
Here’s a new DARPA-inspired, NASA-built robot, complete with a glowing NASA Meatball in its chest, reminiscent of ET’s heart light. The robot’s name is Valkyrie and she was created by a team at the Johnson Space Center as part of the DARPA Robotics Challenge, a contest designed to find the life-saving robot of the future. While NASA’s current robot — Robonaut 2 – is just now getting a pair of legs, “Val” (officially named “R5″ by NASA) is a 1.9 meter tall, 125 kilogram, (6-foot 2-inch, 275-pound) rescue robot that can walk over multiple kinds of terrain, climb a ladder, use tools, and even drive.
According to an extensive article about the new robot in IEEE Spectrum, “This means that Valkyrie has to be capable of operating in the same spaces that a person would operate in, under the control of humans who have only minimal training with robots, which is why the robot’s design is based on a human form.”
Why is NASA building more robots? The thinking is that NASA could send human-like robots to Mars before they send humans. Right now, Valkyrie is not space-rated, but the team at JSC is just getting started.
She’s loaded with cameras, LIDAR, SONAR, is strong and powerful, and is just a great-looking robot.
“We really wanted to design the appearance of this robot to be one that was, when you saw it you’d say, wow, that’s awesome.” Nicolaus Radford, Project and Group Lead at the Dexterous Robotics Lab and JSC.
This is both wonderful and terrifying. A DARPA-funded four-legged robot named WildCat is being developed by a company called Boston Dynamics (tagline of “Changing Your Idea of What Robots Can Do”). They’ve previously developed a humanoid capable of walking across multiple terrains called Atlas, and the scarily-fast Cheetah which set a new land-speed record for legged robots. But the WildCat is a brand new robot created to run fast on all types of terrain, and so far its top speed has been about 16 mph on flat terrain using both bounding and galloping gaits.
The video, released yesterday, shows WildCat’s best performance so far. Don’t let the sound fool you — yes, it does sound like a weed-whacker. But as soon as it raises up off its haunches, you know you’re doomed.
I’ve been trying to figure out what sci-fi equivalent might describe it best: the Terminator’s pet? A lethal, non-fuzzy Daggit from Battlestar Galactica? An AT-AT Walker on speed?
The talking robot launched to the International Space Station in August has sent its first audio/visual message to Earth. Kirobo, the mini Japanese robot — which appears to have the bravado of Buzz Lightyear and the cuteness of WALL-E — is just .34 meters (13.4-inches) long. Kirobo is designed to be able to have conversations with its astronaut crewmates and to study how robot-human interactions can help the astronauts in the space environment. In its first message, Kirobo wished Earth a “good morning” and mentioned (and motioned) its giant step in getting to space.
Kirobo is part of a research project sponsored by the University of Tokoyo and Toyota, and the robot will be working closely with Koichi Wakata, slated to be the first Japanese commander of the ISS for Expedition 39, who will launch this November as part of the Expedition 38/39 crew. An identical robot named Mirata remains on Earth for additional testing.
Kirobo is designed to navigate in zero-gravity, have facial recognition of its fellow crewmates, and will assist Wakata in various experiments. No word on whether it will have access to opening or closing the various hatches on the space station.
Building a flying vehicle for Mars would have significant advantages for exploration of the surface. However, to date, all of our surface exploring vehicles and robotic units on Mars have been terrestrial rovers. The problem with flying on Mars is that the Red Planet doesn’t have much atmosphere to speak of. It is only 1.6% of Earth air density at sea level, give or take. This means conventional aircraft would have to fly very quickly on Mars to stay aloft. Your average Cessna would be in trouble.
But nature may provide an alternative way of looking at this problem.
The fluid regime of any flying (or swimming) animal, machine, etc. can be summarized by something called the Reynolds Number (Re). The Re is equal to the characteristic length x velocity x fluid density, divided by the dynamic viscosity. It is a measure of the ratio of inertial forces to viscous ones. Your average airplane flies at a high Re: lots of inertia relative to air stickiness. Because the Mars air density is low, the only way to get that inertia is to go really fast. However, not all flyers operate at high Re: most flying animals fly at much lower Re. Insects, in particular, operate at quite small Reynolds numbers (relatively speaking). In fact, some insects are so small that they swim through the air, rather than fly. So, if we scale up a bug-like critter or small bird just a bit, we might get something that can move in the Martian atmosphere without having to go insanely fast.
We need a system of equations to constrain our little bot. Turns out that’s not too tough. As a rough approximation, we can use Colin Pennycuick’s average flapping frequency equation. Based on the flapping frequency expectations from Pennycuick (2008), flapping frequency varies roughly as body mass to the 3/8 power, gravitational acceleration to the 1/2 power, span to the -23/24 power, wing area to the -1/3 power, and fluid density to the -3/8 power. That’s handy, because we can adjust to match Martian gravity and air density. But we need to know if we are shedding vortices from the wings in a reasonable way. Thankfully, there is a known relationship, there, as well: the Strouhal number. Str (in this case) is flapping amplitude x flapping frequency divided by velocity. In cruising flight, it turns out to be pretty constrained.
Our bot should, therefore, end up with a Str between 0.2 and 0.4, while matching the Pennycuick equation. And then, finally, we need to get a Reynolds number in the range for a large living flying insect (tiny insects fly in a strange regime where much of propulsion is drag-based, so we will ignore them for now). Hawkmoths are well studied, so we have their Re range for a variety of speeds. Depending on speed, it ranges from about 3,500 to about 15,000. So somewhere in that ballpark will do.
There are a few ways of solving the system. The elegant way is to generate the curves and look for the intersection points, but a fast and easy method is to punch it into a matrix program and solve iteratively. I won’t give all the possible options, but here’s one that worked out pretty well to give an idea:
Mass: 500 grams
Span: 1 meter
Wing Aspect Ratio: 8.0
This gives an Str of 0.31 (right on the money) and Re of 13,900 (decent) at a lift coefficient of 0.5 (which is reasonable for cruising). To give an idea, this bot would have roughly bird-like proportions (similar to a duck), albeit a bit on the light side (not tough with good synthetic materials). It would, however, flap through a greater arc at higher frequency than a bird here on Earth, so it would look a bit like a giant moth at distance to our Earth-trained eyes. As an added bonus, because this bot is flying in a moth-ish Reynolds Regime, it is plausible that it might be able to jump to the very high lift coefficients of insects for brief periods using unsteady dynamics. At a CL of 4.0 (which has been measured for small bats and flycatchers, as well as some large bees), the stall speed is only 19.24 m/s. Max CL is most useful for landing and launching. So: can we launch our bot at 19.24 m/s?
For fun, let’s assume our bird/bug bot also launches like an animal. Animals don’t take off like airplanes; they use a ballistic initiation by pushing from the substrate. Now, insects and birds use walking limbs for this, but bats (and probably pterosaurs) use the wings to double as pushing systems. If we made our bots wings push-worthy, then we can use the same motor to launch as to fly, and it turns out that not much push is required. Thanks to the low Mars gravity, even a little leap goes a long way, and the wings can already beat near 19.24 m/s as it is. So just a little hop will do it. If we’re feeling fancy, we can put a bit more punch on it, and that’ll get out of craters, etc. Either way, our bot only needs to be about 4% as efficient a leaper as good biological jumpers to make it up to speed.
These numbers, of course, are just a rough illustration. There are many reasons that space programs have not yet launched robots of this type. Problems with deployment, power supply, and maintenance would make these systems very challenging to use effectively, but it may not be altogether impossible. Perhaps someday our rovers will deploy duck-sized moth bots for better reconnaissance on other worlds.
Mission planners really hate it when space robots land off course. We’re certainly improving the odds of success these days (remember Mars Curiosity’s seven minutes of terror?), but one space agency has a fancy simulator up its sleeve that could make landings even more precise.
Shown above, this software and hardware (tested at the European Space Agency) so impressed French aerospace center ONERA that officials recently gave the lead researcher an award for the work.
“If I’m a tourist in Paris, I might look for directions to famous landmarks such as the Eiffel Tower, the Arc de Triomphe or Notre Dame cathedral to help find my position on a map,” stated Jeff Delaune, the Ph.D. student performing the research.
“If the same process is repeated from space with enough surface landmarks seen by a camera, the eye of the spacecraft, it can then pretty accurately identify where it is by automatically comparing the visual information to maps we have onboard in the computer.”
Because landmarks close-up can look really different from far away, this system has a method to try and get around that problem.
The so-called ‘Landing with Inertial and Optical Navigation’ (LION) system takes the real-time images generated by the spacecraft’s camera and compares it to maps from previous missions, as well as 3-D digital models of the surface.
LION can take into account the relative size of every point it sees, whether it’s a huge crater or a tiny boulder.
At ESA’s control hardware laboratory in Noordwijk, the Netherlands, officials tested the system with a high-res map of the moon.
Though this is just a test and there is still a ways to go before this system is space-ready, ESA said simulated positional accuracy was better than 164 feet at 1.86 miles in altitude (or 50 meters at three kilometers in altitude.)
Oh, and while it’s only been tested with simulated moon terrain so far, it’s possible the same system could help a robot land on an asteroid, or Mars, ESA adds.
No word on when the system will first hitch an interplanetary ride, but Delaune is working to apply the research to terrestrial matters such as unmanned aerial vehicles.