Understanding Self-Driving Cars and How to Profit From Them

    Date:

    If you’re a longtime reader of these issues, you’ll know that being at the forefront of innovation is central to my core investment approach. As such, I keep a close eye on several corners of the tech world, watching for the latest and most intriguing developments. And for the past few years, one such corner has held my attention – self-driving cars. 

    Indeed, I’ve been bullish on autonomous vehicles for a while now. But though the industry’s developments have been promising, self-driving cars have remained ‘five years away from being five years away’…

    Until now, that is. 

    Thanks to the rapid expansion of autonomous ride-hailing services in places like Phoenix and San Francisco, the rollout of autonomous trucking in Texas and Arizona, and the upcoming launch of Elon Musk’s Robotaxi on Oct. 10, I believe the stage is set for self-driving cars to begin transforming the $11 trillion transportation services industry. 

    Now, that’s all great information to know. But it doesn’t mean much if we don’t grasp how these vehicles actually work. 

    After all, understanding a burgeoning megatrend is key to finding the best stocks to buy to profit from it. 

    Therefore, to potentially turn the Age of AVs into a massive payday in the long run, we must first understand how a self-driving car works. 

    A Tech Trifecta

    At its core, a self-driving car is operated by a combination of sensors – the ‘hardware stack’ – and AI-powered software – unsurprisingly called the ‘software stack.’ 

    In short, the car’s sensors gather information about its surroundings. Then the AI software processes that data to determine whether the car accelerates, brakes, changes lanes, turns, etc. And this all happens on an instantaneous basis. 

    Usually, the ‘hardware stack’ comprises three sensors: cameras, radar, and lidar. A typical self-driving car uses all three sensors as each has strengths and weaknesses that complement the others nicely. 

    Cameras are used to collect visual data. They capture high-resolution images of the vehicle’s environment, similar to that of a human driver’s eye. These cameras help to recognize various signs, lane markings, and traffic lights and can distinguish between different objects, like pedestrians, cyclists, and vehicles. They are very good at providing detailed visual information, which helps the car understand the context of its surroundings. But they tend to perform poorly in bad visual environments, like when there’s low light or inclement weather. 

    An AV’s radar sensors emit radio waves that bounce off objects and return to the sensor, providing information about the distance, speed, and movement of obstacles in the car’s vicinity. These sensors work well in all weather conditions (complementing cameras nicely), but they provide limited resolution and detail (where cameras excel). 

    Lidar – which stands for light detection and ranging – is essentially radar powered by lasers. These sensors emit laser pulses that also bounce off surrounding objects and return to the sensor. By measuring the time it takes for the light to return, lidar can create a high-resolution 3D map of the vehicle’s environment. This provides accurate depth perception, enabling the car to understand the exact shape, size, and distance of surrounding objects. However, lidar doesn’t capture color or texture information (like cameras do). 

    In other words, cameras are used to see things. Radar is used to sense how fast those things are going. And lidar helps to calculate the exact position of those things.

    In this sense, it is easy to see how these three sensors work together within a self-driving car. 

    Chart

    SignUp For Breaking Alerts

    New Graphic

    We respect your email privacy

    Share post:

    Popular

    More like this
    Related

    Long End Yields Jump to 2-Month Highs on Blockbuster Jobs: Oct. 4, 2024

    Yields are surging on this Jobs Friday as the...

    How to Get a Job in a High-Frequency Trading Firm – Part III

    Read about the Impact of High-Frequency Trading in Part I...

    Navigating Cybersecurity in the Age of AI

    In this episode, Andrew Wilkinson speaks with Alexander Gunz,...

    The Fiscal Impact Of Mass Immigration On U.S. And European Taxpayers

    A year ago, I shared with you my thoughts on...