Tag: autonomous

  • Autonomous Maneuver through Square Targets

    Autonomous Maneuver through Square Targets

    Students: Shakeeb Ahmad, Greg Brunson

    The main idea behind the project is to develop a fully autonomous system, free of external sensings like motion capture and GPS. At the same time, it should be capable of sensing its environment for various tasks. NVIDIA Jetson TK1 is used as the main processor on-board while a forward-facing ZED stereo camera is used to get visual odometry and to detect objects in the environment. The test prototype is then used to implement autonomous navigation through a set of square targets. The stereo camera is hence used to detect the square targets and their center points and finally, a path is planned through the center points. The algorithm is implemented in C++ using the Robot Operating System (ROS) framework.

    Papers:
    [1] S. Ahmad, “High-Performance Testbed for Vision-Aided Autonomous Navigation for Quadrotor UAVs in Cluttered Environments”, The University of New Mexico (Digital Repository), 2018

  • F1/10 Racing

    F1/10 Racing

    Students (from left to right): Carolina Gomez, Rebecca Kreitinger, Greg Brunson, Jonathan West 

    Support: The University of New Mexico Department of Electrical & Computer Engineering, Sandia National Labs

    On the web: The Official Home of F1/10

    The UNM MARHES Lab racing team, LoboRacers won 2nd place in the first-ever F1/10 International Autonomous Racing Competition at Carnegie Mellon University on October 1-2, 2016. The F1/10 Competition involves assembling, programming, and racing an autonomous car of the one-tenth scale of an actual Formula One race car. The objective of this competition is to motivate students to develop advanced algorithms that could be applied to self-driving cars in the future. The cars are programmed utilizing Python and Robot Operating System.

    At the moment, the LoboRacers’ car runs a simple wall following an algorithm that uses distance measurements from five different angles collected with the LiDAR sensor. This way the car can determine how close it is to obstacles and at what rate is that distance from the obstacle changing, such that it may adjust its speed and/or turn accordingly to avoid collisions. The LoboRacers team is currently working on improving the car’s localization and mapping algorithms in preparation for the next race in 2017. The localization and mapping driving approach involves first creating a map of the desired path or “racetrack” by manually driving the car through remote control. Afterward, the driving algorithm is applied, and through the car’s odometry, it will be able to know where it is on the map and expect any upcoming turns, again adjusting its speed to avoid colliding with the edges of the track. This method should allow the car to drive faster and more efficiently, as it will no longer need to “sense” obstacles in real-time and can instead anticipate them and be prepared to execute its speed and direction altering commands.

    Related Articles:
    UNM Electrical Engineering Team Places 2nd in Racing Contest, UNM Newsroom

    UNM Electrical Engineering Students to Compete in National Racing Competition, UNM Newsroom

    Related Videos:
    F1/10 Autonomous Racing