>
They've Been Feeding You Poison (And Calling It Food)
Tattoo ink may cause prolonged changes to the immune system
Travel gadget promises to dry and iron your clothes – totally hands-free
Duckweed: A sustainable, protein-packed food source smeared by Big Ag
Perfect Aircrete, Kitchen Ingredients.
Futuristic pixel-raising display lets you feel what's onscreen
Cutting-Edge Facility Generates Pure Water and Hydrogen Fuel from Seawater for Mere Pennies
This tiny dev board is packed with features for ambitious makers
Scientists Discover Gel to Regrow Tooth Enamel
Vitamin C and Dandelion Root Killing Cancer Cells -- as Former CDC Director Calls for COVID-19...
Galactic Brain: US firm plans space-based data centers, power grid to challenge China
A microbial cleanup for glyphosate just earned a patent. Here's why that matters
Japan Breaks Internet Speed Record with 5 Million Times Faster Data Transfer

Google has curated a set of YouTube clips to help machines learn how humans exist in the world. The AVAs, or "atomic visual actions," are three-second clips of people doing everyday things like drinking water, taking a photo, playing an instrument, hugging, standing or cooking.
Each clip labels the person the AI should focus on, along with a description of their pose and whether they're interacting with an object or another human.
"Despite exciting breakthroughs made over the past years in classifying and finding objects in images, recognizing human actions still remains a big challenge," Google wrote in a recent blog post describing the new dataset. "This is due to the fact that actions are, by nature, less well-defined than objects in videos."
The catalog of 57,600 clips only highlights 80 actions but labels more than 96,000 humans. Google pulled clips from popular movies, emphasizing that they drew from a "variety of genres and countries of origin."