http://WWW.ROBOTICS4U.IN
ROBOTICS 5ab356e70402d10aa42866f8 False 36 2
OK
background image not found
Found Update results for
'mit professor russ tedrake'
9
By 2023, India wants an advanced robotic soldier protecting its borders. This next-generation soldier should be intelligent enough to automatically recognize threats and take action. It should also be sophisticated enough to distinguish between threats and non-threats. If India achieves its objective, that will have a huge impact on two fronts at least. First, the robotic soldier would give India the ability to redefine geopolitics, regionally and globally. India could join a very small yet special club of countries (such as Russia and Israel) that are using robots to defend their borders. India may use its robotic soldier as a strategic weapon, like a nuclear bomb, to command attention and respect. From a nation that is currently a secondary partner to the U.S., Russia, or China, a robotic soldier would give India the capability to have a strategic agenda of its own. India will not just be a coalition partner. It will create its own coalition. The next U.N. peacekeeping mission might involve robotic soldiers imported from India or under the command of an Indian general experienced in commanding a robotic army. Second, building an army of robotic soldiers would affect the Indian economy. During the next financial year (2016-’17), India plans to spend nearly $40 billion on defense. This expenditure has quadrupled in the past 15 years. The expenditure was $11.8 billion in 2001. By 2022, India may be spending $620 billion on defense. It’s no wonder then that the Stockholm International Peace Research Institute (SIPRI) found India topping the list of nations importing weapons. According to SIPRI, India bought 14% of all weapons sold globally between 2011 and 2015. The defense budget not only accounts for 17.2 percent of the total planned government expenditure for the next fiscal year, but there is also an off-books number — pensions of defense personnel — that is rising rapidly. It will be around $10 billion in the next financial year. When one in five rupees is going toward defense operations, the economy takes a hit. While the robotic soldiers will not fix the problem by themselves or dramatically change the budget, they are likely to offer relief. Every rupee saved from defense will go toward development. What strategy will India adopt? Will it increase its imports of weapons and acquire the robotic soldiers from overseas, or will India create its robotic soldiers under the “Make in India” program? Or, just as Russia surprised the world with its intervention in the Syrian civil war, India could also enter and exit hot zones or create them in pursuit of its national interests. The robotic soldier would change the border dynamics with China, Bangladesh, and Pakistan, for sure. Information gathered by - Bangalore BTM Robotics training center, Bannerghatta Robotics training center.
This MIT Gadget Can Google Search Your Thoughts. Last night's 60 Minutes took a look inside MIT's Media Lab. Calling it “the Future Factory, " the show got a look at some of the remarkable projects being built, including devices that can Google your thoughts and capture the brain's creativity while going to sleep. Ever since its creation in 1988, books like Steward Brand's The Media Lab: Inventing the Future at M.I.T looked at its early work in studying how to create holograms, recreate human motion, create flat screens, and a host of other technical achievements of the past. As for the present, 60 Minute's Scott Pelley gets a firsthand look at a variety of projects from the Media Lab, with the most outlandish being the brain-to-Google device. Called AlterEgo, the system detects neuromuscular signals in the jaw and face that stem from internal verbalizations, the brain's thought process. After detecting questions in the mind, AlterEgo then transmits the answers through bone vibrations in the ear canal. This allows for the user's natural listening process to continue unabated. Other projects include growing plants without dirt, building better prosthetic limbs, and robots that record thoughts at the moment between consciousness and sleep, attempting to capture the brain at its most creative. Content gathered by BTM robotics training centre, robotics in Bangalore, stem education in Bangalore, stem education in Bannerghatta road, stem education in JP Nagar, robotics training centres in Bannerghatta road, robotics training centres in JP Nagar, robotics training for kids, robotics training for beginners, best robotics in Bangalore.
Firefighting Robot Snake Flies on Jets of Water. Using steerable jets of water like rockets, this robot snake can fly into burning buildings to extinguish fires. Fires have an unfortunate habit of happening in places that aren’t necessarily easy to reach. Whether the source of the fire is somewhere deep within a building, or up more than a floor or two, or both, firefighters have few good options for tackling them. They can either pour water into windows (which doesn’t always work that well), or they can try and get into the building, which seems like it’s probably super dangerous. At the International Conference on Robotics and Automation last month, researchers from Tohoku University and National Institute of Technology, Hachinohe College, in Japan, presented a new kind of snake-like robot with the body of a fire house. Like other snake robots, this one has the potential to be able to wiggle its way into windows or other gaps in a structure, with the benefit of carrying and directing water as it goes. What’s so cool about this particular design, though, is how it powers itself: By firing high pressure jets of water downwards like rocket engines, it can lift itself off of the ground and fly. What’s happening here might be complex to implement in practice, but in principle, it’s not too complicated: There are sets of steerable nozzle modules distributed along the length of the hose. These modules siphon water out of the high pressure stream inside of the hose, and spray it downwards. As the water exits downwards at high velocity, it pushes the hose upwards, and with enough of these modules squirting out high pressure water, the entire hose can be lifted into the air. Just like a rocket, it’s not dependent on ground proximity to work, so as long as you keep on giving it more hose and water at a high enough pressure, it’ll go as high as you want. Since the nozzles are steerable, each module can direct itself independently, letting the hose weave itself through small gaps deep into a structure in order to find the source of a fire. And the “head” module comes with a few extra degrees of freedom to allow the water stream to be directed more precisely. And of course, while the head nozzle is fighting the source of the fire, a byproduct of the body of the house keeping itself airborne is that it’s drenching everything that it’s passing over, while also keeping itself cool. The 2-meter long prototype in the video above is intended to be a single segment in a robot that can be extended to an arbitrary length by just adding on more segments. A gas engine powered a compressor that provided water at 0.7 MPa. It worked reasonably well, as prototypes go, but it’s really more of a proof of concept in hardware than anything else, and obviously there’s a lot to do before a system like this could be real-world useful. The researchers readily admit that their current control algorithms are “not sophisticated, ” and that they’ll need to put some work into making it more stable, more controllable, and able to handle more modules. They’re actively working on it, though, and we’re looking forward to this tech being adapted to garden hoses as well. Content gathered by BTM robotics training center, robotics in Bangalore, stem education in Bangalore, stem education in Bannerghatta road, stem education in JP nagar, robotics training centers in Bannerghatta road, robotics training centers in JP nagar, robotics training for kids, robotics training for beginners, best robotics in Bangalore,
Nvidia 's AI makes a breakthrough in ray tracing(computer graphics): Nvidia' AI that started out in 2017 as something that was not expected to do as much as it has already done like enriching graphics, transforming huge amounts of medical data into life saving breakthroughs, identifying diseases with a simple drop of blood, finding new ways to bring cures to the market faster, helping crops to flourish with optimal materials, customer assistance, self driving vehicles, analysis of various things in search for solutions etc. This AI became popular due to the Isaac robot simulator program. This got rid of programming and let the AI or robot think and learn on its own. Of course initially it was inefficient but after they found out that the robot could successfully learn something like hockey they wanted to make it efficient. This is where their idea of a virtual world comes in. They crated a program that takes the brain of a robot and puts it in a virtual world where it is allowed to try as much as it likes to achieve a certain task. This virtual world follows the laws of our world except for time. In this world a robot can practice its goal in very less time therefore becoming efficient. This AI has now made a breakthrough in computer graphics. ray tracing which is a method used for non real time instances due to its low computing speed has now been turned into a more faster computing one to handle real time gaming systems. Ray tracing is normally used to enhance effects bye understanding how our eye works, but was unable to handle on going instances like gaming and that was why it was limited to only movies.But now thanks to the AI and the quadro GV100 it is now possible to use get high quality graphics by using ray tracing for games as well. This also cuts the cost to 1/5 the original and takes 1/7 the original time taken. Issued by BTM layout robotic center
Nvidia is training robots to learn new skills by observing humans. Initial experiments with the process have seen a Baxter robot learn to pick up and move colored boxes and a toy car in a lab environment. The researchers hope the development of the new deep-learning based system will go some way to train robots to work alongside humans in both manufacturing and home settings. “In the manufacturing environment, robots are really good at repeatedly executing the same trajectory over and over again, but they don’t adapt to changes in the environment, and they don’t learn their tasks, ” Nvidia principal research scientist Stan Birchfield told VentureBeat. “So to repurpose a robot to execute a new task, you have to bring in an expert to reprogram the robot at a fairly low level, and it’s an expensive operation. What we’re interested in doing is making it easier for a non-expert user to teach a robot a new task by simply showing it what to do.” The researchers trained a sequence of neural networks to perform duties associated with perception, program generation, and program execution. The result was that the robot was able to learn a new task from a single demonstration in the real world. Once the robot witnesses the task, it generates a human-readable description of the states required to complete the task. A human can then correct the steps if necessary before execution on the real robot. “There’s sort of a paradigm shift happening in the robotics community now, ” Birchfield said. “We’re at the point now where we can use GPUs to generate essentially a limitless amount of pre-labeled data essentially for free to develop and test algorithms. And this is potentially going to allow us to develop these robotics systems that need to learn how to interact with the world around them in ways that scale better and are safer.” In a video released by the researchers, human operator shows a pair of stacks of cubes to the robot. The system then understands an appropriate program and correctly places the cubes in the correct order. Information gathered by - Robotics for u. Bangalore Robotics, BTM Robotics training center, Robotics spares, Bannerghatta Robotics training center, best robotics training in bangalore,
Controlling robots with brainwaves and hand gestures Computer Science and Artificial Intelligence Laboratory system enable people to correct robot mistakes on multiple-choice tasks. Getting robots to do things isn’t easy, usually, scientists have to either explicitly program them or get them to understand how humans communicate via language. But what if we could control robots more intuitively, using just hand gestures and brainwaves? A new system spearheaded by researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) aims to do exactly that, allowing users to instantly correct robot mistakes with nothing more than brain signals and the flick of a finger. Building off the team’s past work focused on simple binary-choice activities, the new work expands the scope to multiple-choice tasks, opening up new possibilities for how human workers could manage teams of robots. By monitoring brain activity, the system can detect in real-time if a person notices an error as a robot does a task. Using an interface that measures muscle activity, the person can then make hand gestures to scroll through and select the correct option for the robot to execute. Content gathered by BTM robotics training center, robotics in Bangalore, stem education in Bangalore, stem education in Bannerghatta road, stem education in JP nagar, robotics training centers in Bannerghatta road, robotics training centers in JP nagar, robotics training for kids, robotics training for beginners, best robotics in Bangalore,
Wireless 'RoboFly' Looks like an Insect, Gets Its Power from Lasers. You might remember RoboBee, an insect-sized robot that flies by flapping its wings. Unfortunately, though it has to be hard-wired to a power source. Well, one of RoboBee's creators has now helped develop RoboFly, which flies without a tether. Slightly heavier than a toothpick, RoboFly was designed by a team at the University of Washington – one member of that team, assistant professor Sawyer Fuller, was also part of the Harvard University team that first created RoboBee. That flying robot receives its power via a wire attached to an external power source, as an onboard battery would simply be too heavy to allow the tiny craft to fly. Instead of a wire or a battery, RoboFly is powered by a laser. That laser shines on a photovoltaic cell, which is mounted on top of the robot. On its own, that cell converts the laser light to just seven volts of electricity, so a built-in circuit boosts that to the 240 volts needed to flap the wings. That circuit also contains a microcontroller, which tells the robot when and how to flap its wings – on RoboBee, that sort of "thinking" is handled via a tether-linked external controller. Content gathered by BTM robotics training centre, robotics in Bangalore, stem education in Bangalore, stem education in Bannerghatta road, stem education in JP Nagar, robotics training centres in Bannerghatta road, robotics training centres in JP Nagar, robotics training for kids, robotics training for beginners, best robotics in Bangalore,
Former NASA Engineers Building Real-Life Underwater Transformer. In its ROV mode, Aquanaut has two arms for doing work. A transformer designed to do grunt work for the oil industry and the military is coming, and it’s admittedly kind of fun to look at. Houston Mechatronics, a small company founded and led by a team of former NASA robot engineers, May 1 some major strides toward building a transforming submersible the company calls "Aquanaut." The 2, 315-pound (1, 050 kilograms) unmanned underwater vehicle (UUV) will transform itself in order to operate in two modes, according to the company: a sleek, submarine-shaped autonomous underwater vehicle (AUV) mode and an unfolded, two-armed remotely operated vehicle (ROV) mode for work. Aquanaut will swim through the water in its sleek AUV mode. When Aquanaut moves through the water, we want as little drag as possible to extend the maximum range of what the vehicle can do on battery power, " Houston Mechatronics spokesperson Sean Halpin said. "By enclosing the limbs, we're able to operate the vehicle over great distances, up to 200 kilometers [124 miles] Content gathered by BTM robotics training centre, robotics in Bangalore, stem education in Bangalore, stem education in Bannerghatta road, stem education in JP Nagar, robotics training centres in Bannerghatta road, robotics training centres in JP Nagar, robotics training for kids, robotics training for beginners, best robotics in Bangalore,
Experimental drone uses AI to spot violence in crowds. Whether or not it works well in practice is another story. Drone-based surveillance still makes many people uncomfortable, but that isn't stopping research into more effective airborne watchdogs. Scientists have developed an experimental drone system that uses AI to detect violent actions in crowds. The team trained their machine learning algorithm to recognize a handful of typical violent motions (punching, kicking, shooting and stabbing) and flag them when they appear in a drone's camera view. The technology could theoretically detect a brawl that on-the-ground officers might miss, or pinpoint the source of a gunshot. As The Verge warned, the technology definitely isn't ready for real-world use. The researchers used volunteers in relatively ideal conditions (open ground, generous spacing and dramatic movements). The AI is 94 percent effective at its best, but that drops down to an unacceptable 79 percent when there are ten people in the scene. As-is, this system might struggle to find an assailant on a jam-packed street -- what if it mistakes an innocent gesture for an attack? The creators expect to fly their drone system over two festivals in India as a test, but it's not something you'd want to rely on just yet. There's a larger problem surrounding the ethical implications. There are questions about abuses of power and reliability for facial recognition systems. Governments may be tempted to use this as an excuse to record aerial footage of people in public spaces, and could track the gestures of political dissidents (say, people holding protest signs or flashing peace symbols). It could easily combine with other surveillance methods to create a complete picture of a person's movements. This might only find acceptance in limited scenarios where organizations both make it clear that people are on camera and with reassurances that a handshake won't lead to police at their door. Content gathered by BTM robotics training center, robotics in Bangalore, stem education in Bangalore, stem education in Bannerghatta road, stem education in JP nagar, robotics training centers in Bannerghatta road, robotics training centers in JP nagar, robotics training for kids, robotics training for beginners, best robotics in Bangalore.
1
false