{"project": {
"benefits": "Like cockroaches, robots with this brain model could easily adapt to different environments. In addition to providing benefits to robotics in exploration and search and rescue, this work will provide hypotheses for biologists to test, increasing the performance and accuracy of the model. These robots would reduce the need for human input to in their activity, by requiring intervention in navigation decisions far less often and thus could be employed for efficient exploration of a large area, since little human intervention would be required.",
"programDirectors": {"programDirector": "Therese Griebel"},
"coInvestigators": {"coInvestigator": "Brian Tietz"},
"responsibleProgram": "Space Technology Research Grants",
"workLocations": {"workLocation": "Ohio"},
"website": "https://www.nasa.gov/directorates/spacetech/home/index.html",
"endDate": "Aug 2015",
"primaryTas": {"technologyAreas": [
{
"code": 4,
"name": "Robotics and Autonomous Systems",
"id": 3239
},
{
"code": 4.5,
"name": "System-Level Autonomy",
"id": 3278
}
]},
"programManagers": {"programManager": "Kurt Sacksteder"},
"description": "Often, robots are designed for an environment where most of the constraints are known by the programmer, therefore allowing for task-specific algorithms to be used. However in a new environment, randomly moving robots are prone to getting lost, while more deterministic robots can get stuck in an unexpected situation. Animals, unlike robots, are capable of quickly adapting to many different environments. Given their adaptive abilities one would expect animals to employ some method of navigation towards their goals that is simple and robust, if not efficient. Comparing cockroaches to robots, cockroaches have already found a successful balance between sensory input, mapping, and randomness through evolution. Their methods, once determined, could provide inspiration for more robust robotic navigation for applications such as search and rescue or planetary exploration. I propose to mimic cockroach sensory integration and decision making using a software model of the cockroaches' central complex. Using data from current studies in the Ritzmann biology lab, I can generate a model of the brain to synthesize tactile and visual inputs into navigation decisions. My approach to developing this controller will take place in three phases. First, I will implement an algorithm based on cockroach goal seeking behavior in a robot for proof of concept. Then I will look into the neural bases of these behavioral decisions, and generate a neural network that makes these decisions in simulation. This brain model could provide outputs as motor speeds and then be implemented on the robot from task one. My third and final task will be to work with a gait controller being developed a CWRU to determine descending commands for behaviors such as turning and obstacle avoidance, and integrating these commands with the brain model. Like cockroaches, robots with this brain model could easily adapt to different environments. In addition to providing benefits to robotics in exploration and search and rescue, this work will provide hypotheses for biologists to test, increasing the performance and accuracy of the model. These robots would reduce the need for human input to in their activity, by requiring intervention in navigation decisions far less often and thus could be employed for efficient exploration of a large area, since little human intervention would be required.",
"technologyMaturityCurrent": 3,
"title": "Sensory Integration and Decision Making Based on Insect Brain Model",
"technologyMaturityEnd": 3,
"additionalTas": {"technologyAreas": [
{
"code": 4,
"name": "Robotics and Autonomous Systems",
"id": 3239
},
{
"code": 4.5,
"name": "System-Level Autonomy",
"id": 3278
},
{
"code": "4.5.8",
"name": "Automated Data Analysis for Decision Making",
"id": 3671
}
]},
"principalInvestigators": {"principalInvestigator": "Roger Quinn"},
"lastUpdated": "2017-09-16",
"supportingOrganizations": {"organization": {
"name": "Case Western Reserve University",
"type": "Academic"
}},
"library": {"libraryItem": {
"description": "Project Image Sensory Integration and Decision Making Based on Insect Brain Model",
"files": {"file": {
"size": 194342,
"id": 1822,
"url": "https://techport.nasa.gov/file/1822"
}},
"id": 941,
"title": "4327-1363264966606.jpg",
"type": "Image"
}},
"technologyMaturityStart": 2,
"responsibleMissionDirectorateOrOffice": "Space Technology Mission Directorate",
"id": 4327,
"startDate": "Aug 2011",
"status": "Completed"
}}