>
Video: Spacious bubble-sub lets you tour the sea floor in first class
NASA just hacked a 1977 computer on a spacecraft way out past Pluto
First-ever autonomous motor race streams live this weekend
Kanye West plans to launch Yeezy PORN studio with Stormy Daniels' ex in latest shock move...
Blazing bits transmitted 4.5 million times faster than broadband
Scientists Close To Controlling All Genetic Material On Earth
Doodle to reality: World's 1st nuclear fusion-powered electric propulsion drive
Phase-change concrete melts snow and ice without salt or shovels
You Won't Want To Miss THIS During The Total Solar Eclipse (3D Eclipse Timeline And Viewing Tips
China Room Temperature Superconductor Researcher Had Experiments to Refute Critics
5 video games we wanna smell, now that it's kinda possible with GameScent
Unpowered cargo gliders on tow ropes promise 65% cheaper air freight
Wyoming A Finalist For Factory To Build Portable Micro-Nuclear Plants
Recently, veterinarians have developed a protocol for estimating the pain a sheep is in from its facial expressions, but humans apply it inconsistently, and manual ratings are time-consuming. Computer scientists at the University of Cambridge in the United Kingdom have stepped in to automate the task. They started by listing several "facial action units" (AUs) associated with different levels of pain, drawing on the Sheep Pain Facial Expression Scale. They manually labeled these AUs—nostril deformation, rotation of each ear, and narrowing of each eye—in 480 photos of sheep. Then they trained a machine-learning algorithm by feeding it 90% of the photos and their labels, and tested the algorithm on the remaining 10%. The program's average accuracy at identifying the AUs was 67%, about as accurate as the average human, the researchers will report today at the IEEE International Conference on Automatic Face and Gesture Recognition in Washington, D.C. Ears were the most telling cue.