Tag Archives: create

#438014 Meet Blueswarm, a Smart School of ...

Anyone who’s seen an undersea nature documentary has marveled at the complex choreography that schooling fish display, a darting, synchronized ballet with a cast of thousands.

Those instinctive movements have inspired researchers at the Harvard John A. Paulson School of Engineering and Applied Sciences (SEAS), and the Wyss Institute for Biologically Inspired Engineering. The results could improve the performance and dependability of not just underwater robots, but other vehicles that require decentralized locomotion and organization, such as self-driving cars and robotic space exploration.

The fish collective called Blueswarm was created by a team led by Radhika Nagpal, whose lab is a pioneer in self-organizing systems. The oddly adorable robots can sync their movements like biological fish, taking cues from their plastic-bodied neighbors with no external controls required. Nagpal told IEEE Spectrum that this marks a milestone, demonstrating complex 3D behaviors with implicit coordination in underwater robots.

“Insights from this research will help us develop future miniature underwater swarms that can perform environmental monitoring and search in visually-rich but fragile environments like coral reefs,” Nagpal said. “This research also paves a way to better understand fish schools, by synthetically recreating their behavior.”

The research is published in Science Robotics, with Florian Berlinger as first author. Berlinger said the “Bluedot” robots integrate a trio of blue LED lights, a lithium-polymer battery, a pair of cameras, a Raspberry Pi computer and four controllable fins within a 3D-printed hull. The fish-lens cameras detect LED’s of their fellow swimmers, and apply a custom algorithm to calculate distance, direction and heading.

Based on that simple production and detection of LED light, the team proved that Blueswarm could self-organize behaviors, including aggregation, dispersal and circle formation—basically, swimming in a clockwise synchronization. Researchers also simulated a successful search mission, an autonomous Finding Nemo. Using their dispersion algorithm, the robot school spread out until one could detect a red light in the tank. Its blue LEDs then flashed, triggering the aggregation algorithm to gather the school around it. Such a robot swarm might prove valuable in search-and-rescue missions at sea, covering miles of open water and reporting back to its mates.

“Each Bluebot implicitly reacts to its neighbors’ positions,” Berlinger said. The fish—RoboCod, perhaps?—also integrate a Wifi module to allow uploading new behaviors remotely. The lab’s previous efforts include a 1,000-strong army of “Kilobots,” and a robotic construction crew inspired by termites. Both projects operated in two-dimensional space. But a 3D environment like air or water posed a tougher challenge for sensing and movement.

In nature, Berlinger notes, there’s no scaly CEO to direct the school’s movements. Nor do fish communicate their intentions. Instead, so-called “implicit coordination” guides the school’s collective behavior, with individual members executing high-speed moves based on what they see their neighbors doing. That decentralized, autonomous organization has long fascinated scientists, including in robotics.

“In these situations, it really benefits you to have a highly autonomous robot swarm that is self-sufficient. By using implicit rules and 3D visual perception, we were able to create a system with a high degree of autonomy and flexibility underwater where things like GPS and WiFi are not accessible.”

Berlinger adds the research could one day translate to anything that requires decentralized robots, from self-driving cars and Amazon warehouse vehicles to exploration of faraway planets, where poor latency makes it impossible to transmit commands quickly. Today’s semi-autonomous cars face their own technical hurdles in reliably sensing and responding to their complex environments, including when foul weather obscures onboard sensors or road markers, or when they can’t fix position via GPS. An entire subset of autonomous-car research involves vehicle-to-vehicle (V2V) communications that could give cars a hive mind to guide individual or collective decisions— avoiding snarled traffic, driving safely in tight convoys, or taking group evasive action during a crash that’s beyond their sensory range.

“Once we have millions of cars on the road, there can’t be one computer orchestrating all the traffic, making decisions that work for all the cars,” Berlinger said.

The miniature robots could also work long hours in places that are inaccessible to humans and divers, or even large tethered robots. Nagpal said the synthetic swimmers could monitor and collect data on reefs or underwater infrastructure 24/7, and work into tiny places without disturbing fragile equipment or ecosystems.

“If we could be as good as fish in that environment, we could collect information and be non-invasive, in cluttered environments where everything is an obstacle,” Nagpal said. Continue reading

Posted in Human Robots

#438001 How an Israeli Startup Is Using AI to ...

The first baby conceived using in-vitro fertilization (IVF) was born in the UK in 1978. Over 40 years later, the technique has become commonplace, but its success rate is still fairly low at around 22 to 30 percent. A female-founded Israeli startup called Embryonics is setting out to change this by using artificial intelligence to screen embryos.

IVF consists of fertilizing a woman’s egg with her partner’s or a donor’s sperm outside of her body, creating an embryo that’s then implanted in the uterus. It’s not an easy process in any sense of the word—physically, emotionally, or financially. Insurance rarely covers IVF, and the costs run anywhere from $12,000 to $25,000 per cycle (a cycle takes about a month and includes stimulating a woman’s ovaries to produce eggs, extracting the eggs, inseminating them outside the body, and implanting an embryo).

Women have to give themselves daily hormone shots to stimulate egg production, and these can cause uncomfortable side effects. After so much stress and expense, it’s disheartening to think that the odds of a successful pregnancy are, at best, one in three.

A crucial factor in whether or not an IVF cycle works—that is, whether the embryo implants in the uterus and begins to develop into a healthy fetus—is the quality of the embryo. Doctors examine embryos through a microscope to determine how many cells they contain and whether they appear healthy, and choose the one that looks most viable.

But the human eye can only see so much, even with the help of a microscope; despite embryologists’ efforts to select the “best” embryo, success rates are still relatively low. “Many decisions are based on gut feeling or personal experience,” said Embryonics founder and CEO Yael Gold-Zamir. “Even if you go to the same IVF center, two experts can give you different opinions on the same embryo.”

This is where Embryonics’ technology comes in. They used 8,789 time-lapse videos of developing embryos to train an algorithm that predicts the likelihood of successful embryo implantation. A little less than half of the embryos from the dataset were graded by embryologists, and implantation data was integrated when it was available (as a binary “successful” or “failed” metric).

The algorithm uses geometric deep learning, a technique that takes a traditional convolutional neural network—which filters input data to create maps of its features, and is most commonly used for image recognition—and applies it to more complex data like 3D objects and graphs. Within days after fertilization, the embryo is still at the blastocyst stage, essentially a microscopic clump of just 200-300 cells; the algorithm uses this deep learning technique to spot and identify patterns in embryo development that human embryologists either wouldn’t see at all, or would require massive collation of data to validate.

On top of the embryo videos, Embryonics’ team incorporated patient data and environmental data from the lab into its algorithm, with encouraging results: the company reports that using its algorithm resulted in a 12 percent increase in positive predictive value (identifying embryos that would lead to implantation and healthy pregnancy) and a 29 percent increase in negative predictive value (identifying embyros that would not result in successful pregnancy) when compared to an external panel of embryologists.

TechCrunch reported last week that in a pilot of 11 women who used Embryonics’ algorithm to select their embryos, 6 are enjoying successful pregnancies, while 5 are still awaiting results.

Embryonics wasn’t the first group to think of using AI to screen embryos; a similar algorithm developed in 2019 by researchers at Weill Cornell Medicine was able to classify the quality of a set of embryo images with 97 percent accuracy. But Embryonics will be one of the first to bring this sort of technology to market. The company is waiting to receive approval from European regulatory bodies to be able to sell the software to fertility clinics in Europe.

Its timing is ripe: as more and more women delay having kids due to lifestyle and career-related factors, demand for IVF is growing, and will likely accelerate in coming years.

The company ultimately hopes to bring its product to the US, as well as to expand its work to include using data to improve hormonal stimulation.

Image Credit: Gerd Altmann from Pixabay Continue reading

Posted in Human Robots

#437992 This Week’s Awesome Tech Stories From ...

ARTIFICIAL INTELLIGENCE
This Chinese Lab Is Aiming for Big AI Breakthroughs
Will Knight | Wired
“China produces as many artificial intelligence researchers as the US, but it lags in key fields like machine learning. The government hopes to make up ground. …It set AI researchers the goal of making ‘fundamental breakthroughs by 2025’ and called for the country to be ‘the world’s primary innovation center by 2030.’ BAAI opened a year later, in Zhongguancun, a neighborhood of Beijing designed to replicate US innovation hubs such as Boston and Silicon Valley.”

ENVIRONMENT
What Elon Musk’s $100 Million Carbon Capture Prize Could Mean
James Temple | MIT Technology Review
“[Elon Musk] announced on Twitter that he plans to give away $100 million of [his $180 billion net worth] as a prize for the ‘best carbon capture technology.’ …Another $100 million could certainly help whatever venture, or ventures, clinch Musk’s prize. But it’s a tiny fraction of his wealth and will also only go so far. …Money aside, however, one thing Musk has a particular knack for is generating attention. And this is a space in need of it.”

HEALTH
Synthetic Cornea Helped a Legally Blind Man Regain His Sight
Steve Dent | Engadget
“While the implant doesn’t contain any electronics, it could help more people than any robotic eye. ‘After years of hard work, seeing a colleague implant the CorNeat KPro with ease and witnessing a fellow human being regain his sight the following day was electrifying and emotionally moving, there were a lot of tears in the room,’ said CorNeat Vision co-founder Dr. Gilad Litvin.”

BIOTECH
MIT Develops Method for Lab-Grown Plants That May Eventually Lead to Alternatives to Forestry and Farming
Darrell Etherington | TechCrunch
“If the work of these researchers can eventually be used to create a way to produce lab-grown wood for use in construction and fabrication in a way that’s scalable and efficient, then there’s tremendous potential in terms of reducing the impact on forestry globally. Eventually, the team even theorizes you could coax the growth of plant-based materials into specific target shapes, so you could also do some of the manufacturing in the lab, by growing a wood table directly for instance.”

AUTOMATION
FAA Approves First Fully Automated Commercial Drone Flights
Andy Pasztor and Katy Stech Ferek | The Wall Street Journal
“US aviation regulators have approved the first fully automated commercial drone flights, granting a small Massachusetts-based company permission to operate drones without hands-on piloting or direct observation by human controllers or observers. …The company’s Scout drones operate under predetermined flight programs and use acoustic technology to detect and avoid drones, birds, and other obstacles.”

SPACE
China’s Surging Private Space Industry Is Out to Challenge the US
Neel V. Patel | MIT Technology Review
“[The Ceres-1] was a commercial rocket—only the second from a Chinese company ever to go into space. And the launch happened less than three years after the company was founded. The achievement is a milestone for China’s fledgling—but rapidly growing—private space industry, an increasingly critical part of the country’s quest to dethrone the US as the world’s preeminent space power.”

CRYPTOCURRENCY
Janet Yellen Will Consider Limiting Use of Cryptocurrency
Timothy B. Lee | Ars Technica
“Cryptocurrencies could come under renewed regulatory scrutiny over the next four years if Janet Yellen, Joe Biden’s pick to lead the Treasury Department, gets her way. During Yellen’s Tuesday confirmation hearing before the Senate Finance Committee, Sen. Maggie Hassan (D-N.H.) asked Yellen about the use of cryptocurrency by terrorists and other criminals. ‘Cryptocurrencies are a particular concern,’ Yellen responded. ‘I think many are used—at least in a transactions sense—mainly for illicit financing.’i”

SCIENCE
Secret Ingredient Found to Power Supernovas
Thomas Lewton | Quanta
“…Only in the last few years, with the growth of supercomputers, have theorists had enough computing power to model massive stars with the complexity needed to achieve explosions. …These new simulations are giving researchers a better understanding of exactly how supernovas have shaped the universe we see today.”

Image Credit: Ricardo Gomez Angel / Unsplash Continue reading

Posted in Human Robots

#437971 Video Friday: Teleport Yourself Into ...

Video Friday is your weekly selection of awesome robotics videos, collected by your Automaton bloggers. We’ll also be posting a weekly calendar of upcoming robotics events for the next few months; here's what we have so far (send us your events!):

HRI 2021 – March 8-11, 2021 – [Online]
RoboSoft 2021 – April 12-16, 2021 – [Online]
Let us know if you have suggestions for next week, and enjoy today's videos.

Samsung announced some new prototype robots at CES this week. It's a fancy video, but my guess is that the actual autonomy here is minimal at best.

[ Samsung ]

Some very impressive reactive agility from Ghost Robotics' little quadruped.

[ Ghost Robotics ]

Toyota Research Institute (TRI) is researching how to bring together the instinctive reflexes of professional drivers and automated driving technology that uses the calculated foresight of a supercomputer. Using a Toyota GR Supra, TRI will learn from some of the most skilled drivers in the world to develop sophisticated vehicle control algorithms. The project’s goal is to design a new level of active safety technology for the Toyota Guardian™ approach of amplifying human driving abilities and helping keep people safe.

[ TRI ]

The end of this video features one of the most satisfying-sounding drone outtakes I've ever heard,

[ ASL ]

Reachy can now run the first humanoid VR teleoperation app available on the market. This app allows you to place yourself in the body of a humanoid robot, in VR, wherever you are in the world, to remotely operate it and carry out complex tasks. With this new functionality, Reachy is able to learn from the demonstration of the humans who control it, which makes application development even easier.

[ Pollen Robotics ]

Thanks Elsa!

Boston Dynamics has inspired some dancing robot videos recently, including this from Marco Tempest.

[ Marco Tempest ]

MOFLIN is an AI Pet created from a totally new concept. It possesses emotional capabilities that evolve like living animals. With its warm soft fur, cute sounds, and adorable movement, you’d want to love it forever. We took a nature inspired approach and developed a unique algorithm that allows MOFLIN to learn and grow by constantly using its interactions to determine patterns and evaluate its surroundings from its sensors. MOFLIN will choose from an infinite number of mobile and sound pattern combinations to respond and express its feelings. To put it in simple terms, it’s like you’re interacting with a living pet.

I like the minimalist approach. I dislike the “it’s like you’re interacting with a living pet” bit.

[ Kickstarter ]

There's a short gif of these warehouse robots going around, but here's the full video.

[ BionicHIVE ]

Vstone's Robovie-Z proves that you don't need fancy hardware for effective teleworking.

[ Vstone ]

All dual-arm robots are required, at some point, to play pool.

[ ABB ]

Volkswagen Group Components gives us a first glimpse of the real prototypes. This is one of the visionary charging concepts that Volkswagen hopes will expand the charging infrastructure over the next few years. Its task: fully autonomous charging of vehicles in restricted parking areas, like underground car parks.

To charge several vehicles at the same time, the mobile robot moves a trailer, essentially a mobile energy storage unit, to the vehicle, connects it up and then uses this energy storage unit to charge the battery of the electric vehicle. The energy storage unit stays with the vehicle during the charging process. In the meantime, the robot charges other electric vehicles.

[ Volkswagen ]

I've got a lot of questions about Moley Robotics' kitchen. But I would immediately point out that the system appears to do no prep work, which (at least for me) is the time-consuming and stressful part of cooking.

[ Moley Robotics ]

Blueswarm is a collective of fish-inspired miniature underwater robots that can achieve a wide variety of 3D collective behaviors – synchrony, aggregation/dispersion, milling, search – using only implicit communication mediated through the production and sensing of blue light. We envision this platform for investigating collective AI, underwater coordination, and fish-inspired locomotion and sensing.

[ Science Robotics ]

A team of Malaysian researchers are transforming pineapple leaves into strong materials that can be used to build frames for unmanned aircraft or drones.

[ Reuters ]

The future of facility disinfecting is here, protect your customers, and create peace of mind. Our drone sanitization spraying technology is up to 100% more efficient and effective than conventional manual spray sterilization processes.

[ Draganfly ]

Robots are no long a future technology, as small robots can be purchased today to be utilized for educational purposes. See what goes into making a modern robot come to life.

[ Huggbees ]

How does a robot dog learn how to dance? Adam and the Tested team examine and dive into Boston Dynamics' Choreographer software that was behind Spot's recent viral dancing video.

[ Tested ]

For years, engineers have had to deal with “the tyranny of the fairing,” that anything you want to send into space has to fit into the protective nosecone on top of the rocket. A field of advanced design has been looking for new ways to improve our engineering, using the centuries-old artform to dream bigger.

[ JPL ] Continue reading

Posted in Human Robots

#437964 How Explainable Artificial Intelligence ...

The field of artificial intelligence has created computers that can drive cars, synthesize chemical compounds, fold proteins, and detect high-energy particles at a superhuman level.

However, these AI algorithms cannot explain the thought processes behind their decisions. A computer that masters protein folding and also tells researchers more about the rules of biology is much more useful than a computer that folds proteins without explanation.

Therefore, AI researchers like me are now turning our efforts toward developing AI algorithms that can explain themselves in a manner that humans can understand. If we can do this, I believe that AI will be able to uncover and teach people new facts about the world that have not yet been discovered, leading to new innovations.

Learning From Experience
One field of AI, called reinforcement learning, studies how computers can learn from their own experiences. In reinforcement learning, an AI explores the world, receiving positive or negative feedback based on its actions.

This approach has led to algorithms that have independently learned to play chess at a superhuman level and prove mathematical theorems without any human guidance. In my work as an AI researcher, I use reinforcement learning to create AI algorithms that learn how to solve puzzles such as the Rubik’s Cube.

Through reinforcement learning, AIs are independently learning to solve problems that even humans struggle to figure out. This has got me and many other researchers thinking less about what AI can learn and more about what humans can learn from AI. A computer that can solve the Rubik’s Cube should be able to teach people how to solve it, too.

Peering Into the Black Box
Unfortunately, the minds of superhuman AIs are currently out of reach to us humans. AIs make terrible teachers and are what we in the computer science world call “black boxes.”

AI simply spits out solutions without giving reasons for its solutions. Computer scientists have been trying for decades to open this black box, and recent research has shown that many AI algorithms actually do think in ways that are similar to humans. For example, a computer trained to recognize animals will learn about different types of eyes and ears and will put this information together to correctly identify the animal.

The effort to open up the black box is called explainable AI. My research group at the AI Institute at the University of South Carolina is interested in developing explainable AI. To accomplish this, we work heavily with the Rubik’s Cube.

The Rubik’s Cube is basically a pathfinding problem: Find a path from point A—a scrambled Rubik’s Cube—to point B—a solved Rubik’s Cube. Other pathfinding problems include navigation, theorem proving and chemical synthesis.

My lab has set up a website where anyone can see how our AI algorithm solves the Rubik’s Cube; however, a person would be hard-pressed to learn how to solve the cube from this website. This is because the computer cannot tell you the logic behind its solutions.

Solutions to the Rubik’s Cube can be broken down into a few generalized steps—the first step, for example, could be to form a cross while the second step could be to put the corner pieces in place. While the Rubik’s Cube itself has over 10 to the 19th power possible combinations, a generalized step-by-step guide is very easy to remember and is applicable in many different scenarios.

Approaching a problem by breaking it down into steps is often the default manner in which people explain things to one another. The Rubik’s Cube naturally fits into this step-by-step framework, which gives us the opportunity to open the black box of our algorithm more easily. Creating AI algorithms that have this ability could allow people to collaborate with AI and break down a wide variety of complex problems into easy-to-understand steps.

A step-by-step refinement approach can make it easier for humans to understand why AIs do the things they do. Forest Agostinelli, CC BY-ND

Collaboration Leads to Innovation
Our process starts with using one’s own intuition to define a step-by-step plan thought to potentially solve a complex problem. The algorithm then looks at each individual step and gives feedback about which steps are possible, which are impossible and ways the plan could be improved. The human then refines the initial plan using the advice from the AI, and the process repeats until the problem is solved. The hope is that the person and the AI will eventually converge to a kind of mutual understanding.

Currently, our algorithm is able to consider a human plan for solving the Rubik’s Cube, suggest improvements to the plan, recognize plans that do not work and find alternatives that do. In doing so, it gives feedback that leads to a step-by-step plan for solving the Rubik’s Cube that a person can understand. Our team’s next step is to build an intuitive interface that will allow our algorithm to teach people how to solve the Rubik’s Cube. Our hope is to generalize this approach to a wide range of pathfinding problems.

People are intuitive in a way unmatched by any AI, but machines are far better in their computational power and algorithmic rigor. This back and forth between man and machine utilizes the strengths from both. I believe this type of collaboration will shed light on previously unsolved problems in everything from chemistry to mathematics, leading to new solutions, intuitions and innovations that may have, otherwise, been out of reach.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Image Credit: Serg Antonov / Unsplash Continue reading

Posted in Human Robots