​Forget self-driving cars: What about self-flying drones?

Tina Amirtha for Benelux:  In 2014, three software engineers decided to create a drone company in Wavre, Belgium, just outside Brussels. All were licensed pilots and trained in NATO security techniques. But rather than build drones themselves, they decided they would upgrade existing radio-controlled civilian drones with an ultra-secure software layer to allow the devices to fly autonomously. Their company, EagleEye Systems, would manufacture the onboard computer and design the software, while existing manufacturers would provide the drone body and sensors. Fast-forward to the end of March this year, when the company received a Section 333 exemption from the US Federal Aviation Administration to operate and sell its brand of autonomous drones in the US. The decision came amid expectations that the FAA will loosen its restrictions on legal drone operations and issue new rules to allow drones to fly above crowds.   Cont'd...

SUNSPRING by 32 Tesla K80 GPUs

From Ross Goodwin on Medium: To call the film above surreal would be a dramatic understatement. Watching it for the first time, I almost couldn’t believe what I was seeing—actors taking something without any objective meaning, and breathing semantic life into it with their emotion, inflection, and movement.  After further consideration, I realized that actors do this all the time. Take any obscure line of Shakespearean dialogue and consider that 99.5% of the audience who hears that line in 2016 would not understand its meaning if they read it in on paper. However, in a play, they do understand it based on its context and the actor’s delivery.  As Modern English speakers, when we watch Shakespeare, we rely on actors to imbue the dialogue with meaning. And that’s exactly what happened inSunspring, because the script itself has no objective meaning. On watching the film, many of my friends did not realize that the action descriptions as well as the dialogue were computer generated. After examining the output from the computer, the production team made an effort to choose only action descriptions that realistically could be filmed, although the sequences themselves remained bizarre and surreal... (medium article with technical details) Here is the stage direction that led to Middleditch’s character vomiting an eyeball early in the film: C (smiles) I don’t know anything about any of this. H (to Hauk, taking his eyes from his mouth) Then what? H2 There’s no answer.

Employing Drones to Solve Business' Most Complex Issues

A true system doesnt only take the technology into account, but also the processes and human aspects.

5D Robotics + Aerial MOB = Autonomy and Reliability

The marrying of Aerial MOBs robust operational experience and IP portfolio with 5Ds robust autonomy and behavioral technology really bridges many of the gaps for delivering valuable products to many industrial type clients, such as those in oil and gas, utilities, and construction among others.

Scientists develop bee model that will impact the development of aerial robotics

Phys.org:  Scientists have built a computer model that shows how bees use vision to detect the movement of the world around them and avoid crashing. This research, published in PLOS Computational Biology, is an important step in understanding how the bee brain processes the visual world and will aid the development of robotics. The study led by Alexander Cope and his coauthors at the University of Sheffield shows how bees estimate the speed of motion, or optic flow, of the visual world around them and use this to control their flight. The model is based on Honeybees as they are excellent navigators and explorers, and use vision extensively in these tasks, despite having a brain of only one million neurons (in comparison to the human brain's 100 billion). The model shows how bees are capable of navigating complex environments by using a simple extension to the known neural circuits, within the environment of a virtual world. The model then reproduces the detailed behaviour of real bees by using optic flow to fly down a corridor, and also matches up with how their neurons respond.   Cont'd...

Billions Are Being Invested in a Robot That Americans Don't Want

Keith Naughton for Bloomberg Technology:  Brian Lesko and Dan Sherman hate the idea of driverless cars, but for very different reasons.  Lesko, 46, a business-development executive in Atlanta, doesn’t trust a robot to keep him out of harm’s way. “It scares the bejeebers out of me,” he says. Sherman, 21, a mechanical-engineering student at the University of Minnesota, Twin Cities, trusts the technology and sees these vehicles eventually taking over the road. But he dreads the change because his passion is working on cars to make them faster. “It’s something I’ve loved to do my entire life and it’s kind of on its way out,” he says. “That’s the sad truth.” The driverless revolution is racing forward, as inventors overcome technical challenges such as navigating at night and regulators craft new rules. Yet the rush to robot cars faces a big roadblock: People aren’t ready to give up the wheel. Recent surveys by J.D. Power, consulting company EY, the Texas A&M Transportation Institute, Canadian Automobile Association, researcher Kelley Blue Book and auto supplier Robert Bosch LLC all show that half to three-quarters of respondents don’t want anything to do with these models.   Cont'd...

Zero Zero Hover Camera drone uses face tracking tech to follow you

Lee Mathews for Geek:  Camera-toting drones that can follow a subject while they’re recording aren’t a new thing, but a company called Zero Zero is putting a very different spin on them. It’s all about how they track what’s being filmed. Zero Zero’s new Hover Camera doesn’t require you to wear a special wristband like AirDog. There’s no “pod” to stuff in your pocket like the one that comes with Lily, and it doesn’t rely on GPS either. Instead, the Hover Camera uses its “eyes” to follow along. Unlike some drones that use visual sensors to lock on to a moving subject, the Hover Camera uses them in conjunction with face and body recognition algorithms to ensure that it’s actually following the person you want it to follow. For now, it can only track the person you initially select. By the time the Hover Camera goes up for sale, however, Zero Zero says it will be able to scan the entire surrounding area for faces.   Cont'd...

Crowdfunding Projects For April

Here are a few projects we think are worth looking into. Be careful... it is crowdfunding.

Bring 3D printed robots to life with 'Ziro' hand-controlled robotics kit

Benedict for 3Ders.org:  Tech startup ZeroUI, based in San Jose, California, has launched an Indiegogo campaign for Ziro, the “world’s first hand-controlled robotics kit”. The modular kit has been designed to bring 3D printed creations to life, and has already surpassed its $30,000 campaign goal. It would be fair to say that the phenomenon of gesture recognition, throughout the wide variety of consumer electronics to which it has been introduced, has been a mixed success. The huge popularity of the Nintendo Wii showed that—for the right product—users were happy to use their hands and bodies as controllers, but for every Wii, there are a million useless webcam or smartphone functions, lying dormant, unused, and destined for the technology recycle bin.   Full Article:  

US to sail submarine drones in South China Sea

Geoff Dyer for CNBC:  As it watches China build up its presence in the South China Sea, one reclaimed island at a time, the US military is betting on a new technology to help retain its edge — submarine drones. During the past six months, the Pentagon has started to talk publicly about a once-secret program to develop unmanned undersea vehicles, the term given to the drone subs that are becoming part of its plan to deter China from trying to dominate the region. Ashton Carter, US defense secretary, made special mention of drone subs in a speech about military strategy in Asia and hinted at their potential use in the South China Sea, which has large areas of shallower water. The Pentagon's investment in subs "includes new undersea drones in multiple sizes and diverse payloads that can, importantly, operate in shallow water, where manned submarines cannot", said Mr Carter, who visited a US warship in the South China Sea on Friday.   Cont'd...

Interested in UAVs? Visit AUVSI Xponential 2016

This show has all of the tools to enable UAV developers, along with innovative software to manage your entire commercial UAS operation. It's a show you don't want to miss.

Upcoming Tradeshow, Conference & Exhibition Summary - May, June, July 2016

Here is a summary of what Tradeshows, Conferences & Exhibitions to look forward to in the coming months.

3D Robotics Undergoing Layoffs And Google Puts Boston Dynamics Up For Sale

3DR will now be focusing more narrowly on enterprise customers that are interested in using drones for such projects as utility line and pipeline inspections, and construction site inspections.

Over 1,000 Student-Led Robotics Teams Converge At VEX Worlds

VEX Worlds 2016 kicks off this week! Presented by the Robotics Education and Competition (REC) Foundation and the Northrop Grumman Foundation, this culminating event brings together the top 1,000 teams from around the world in one city and under one roof for one incredible celebration of robotics engineering, featuring the world's largest and fastest growing international robotics programs - the VEX IQ Challenge, the VEX Robotics Competition and VEX U. On April 20-23, at the Kentucky Exposition Center in Louisville, Ky., over 16,000 participants from 37 nations will come together to put their engineering expertise to the test as they seek to be crowned the Champions of VEX Worlds.   Follow the competition here:

AUVSI XPONENTIAL 2016 - What to Expect from Micromo!

MICROMO was been part of XPONENTIAL/AUVSI since 2010. Our team of Application Engineers will be at the show.

Records 841 to 855 of 1079

First | Previous | Next | Last

Featured Product

BitFlow Introduces 6th Generation Camera Link Frame Grabber: The Axion

BitFlow Introduces 6th Generation Camera Link Frame Grabber: The Axion

BitFlow has offered a Camera Link frame grabbers for almost 15 years. This latest offering, our 6th generation combines the power of CoaXPress with the requirements of Camera Link 2.0. Enabling a single or two camera system to operate at up to 850 MB/S per camera, the Axion-CL family is the best choice for CL frame grabber. Like the Cyton-CXP frame grabber, the Axion-CL leverages features such as the new StreamSync system, a highly optimized DMA engine, and expanded I/O capabilities that provide unprecedented flexibility in routing. There are two options available; Axion 1xE & Axion 2xE. The Axion 1xE is compatible with one base, medium, full or 80-bit camera offering PoCL, Power over Camera Link, on both connectors. The Axion 2xE is compatible with two base, medium, full or 80-bit cameras offering PoCL on both connectors for both cameras. The Axion-CL is a culmination of the continuous improvements and updates BitFlow has made to Camera Link frame grabbers.