Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

That's very cool. Does it use vision or just tactile feeling of where it puts its feed down?


I wouldn't be surprised if it uses neither. My guess would be it has encoders to know the state of the joint angles (so, a type of proprioception) but probably not for foot contact.

If someone has a reference to some technical details, though, I'd be very interested.


Posted: http://news.ycombinator.com/item?id=731300

It uses an IMU, optical vision, lidar, encoders, force sensors, and ground contact sensors.

Why do you think joint angles alone would be enough?


Thanks for the link! I just figured it would use the simplest solution that could get the job done. I don't know how important measuring foot slip/ground contact is for quadrupedal locomotion or how well it can be approximated by measuring joint angles and computing forward kinematics. I guess load cells are cheap so why not use them?

Again, thanks for digging that up. It's heaps better than idle speculation.


Getting the job done is important! A simple solution is no good if it doesn't work. Presumably they need to know where the ground is / how firm the ground is, before putting all it's weight down. There is a good bit in the video of it climbing over rubble.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: