ReportWire

Tag: computer vision

  • The Curious Case of the Bizarre, Disappearing Captcha

    [ad_1]

    As I browse the web in 2025, I rarely encounter captchas anymore. There’s no slanted text to discern. No image grid of stoplights to identify.

    And on the rare occasion that I am asked to complete some bot-deterring task, the experience almost always feels surreal. A colleague shared recent tests where they were presented with images of dogs and ducks wearing hats, from bowler caps to French berets. The security questions ignored the animal’s hats, rudely, asking them to select the photos that showed animals with four legs.

    Other puzzles are hyper-specific to their audience. For example, the captcha for Sniffies, a gay hookup site, has users slide a jockstrap across their smartphone screen to find the matching pair of underwear.

    So, where have all the captchas gone? And why are the few existing challenges so damn weird? I spoke with cybersecurity experts to better understand the current state of these vanishing challenges and why the future will probably look even more peculiar.

    Bot Friction, Human Frustration

    “When the captcha was first invented, the idea was that this was literally a task a computer could not do,” says Reid Tatoris, who leads Cloudflare’s application security detection team. The term captcha—Completely Automatic Public Turing test to tell Computers and Humans Apart—was coined by researchers in 2000 and presented as a way to protect websites from malicious, nonhuman users.

    The initial test most users saw online contained funky characters, usually a combo of warped letters and numbers you had to replicate by typing them into a text field. Computers couldn’t see what the characters were; humans could, even if most of us had to squint to get it right.

    Financial companies like PayPal and email providers like Yahoo used this iteration to ward off automated bots. More websites eventually added audio readouts of the correct answer after receiving pressure from Blind and low-vision advocacy groups, whose members were indeed humans browsing the web but could not complete a vision-based challenge.

    What if, rather than just a test to keep out bots, the challenge could generate useful data? That was a core idea behind the release of reCaptcha in 2007. With reCaptcha, users identified words that machine learning algorithms could not read at the time. This sped up the process of transferring print media into an online form. The tech was quickly acquired by Google, and reCaptcha was instrumental in the company’s efforts to digitize books.

    As machine learning capabilities improved—and they learned to read funky text—online security checkpoints adapted to be more difficult for malicious bots to circumvent. The next iteration reCaptcha challenges included grids of images where users were asked to select specific options, like photos containing a motorcyclist. Google used the data collected here to improve its online maps.

    [ad_2]

    Reece Rogers

    Source link

  • Gemini in Google Home Keeps Mistaking My Dog for a Cat

    [ad_1]

    A cat jumped up on my couch. Wait a minute. I don’t have a cat.

    The alert about the leaping feline is something my Google Home app sent me when I was out at a party. Turns out it was my dog. This notification came through a day after I turned on Google’s Gemini for Home capability in the Google Home app. It brings the power of large language models to the smart home ecosystem, and one of the most useful features is more descriptive alerts from my Nest security cameras. So, instead of “Person seen,” it can tell me FedEx came by and dropped off two packages.

    In the two weeks since I allowed Gemini to power my Google Home, I’ve enjoyed its ability to detect delivery drivers the most. At the end of the day, I can ask in the Google Home app, “How many packages came today” and get an accurate answer. It’s nice to know that it’s FedEx at the door, per my Nest Doorbell, and not a salesperson offering to replace my windows. Yet for all its smarts, Gemini refuses to understand that I do not have a cat in my house.

    Person Seen

    ScreenshotGoogle Home via Julian Chokkattu

    Google isn’t the only company souping up its smart-home ecosystem with AI. Amazon recently announced a feature on its Ring cameras called Search Party that will use a neighborhood’s worth of outdoor Ring cameras to help someone find their lost dog. (I don’t need to stretch to imagine something like this being used for nefarious purposes.)

    In early October, Google updated the voice assistant on its smart-home devices—some of which have been around for a decade—by replacing Google Assistant with Gemini. For the most part, the assistant is better. It can understand multiple commands in a spoken sentence or two, and you can very easily ask it to automate something in your home without fussing with the Routines tab in the Google Home app. And when I ask it a simple question, it generally gives me some kind of a reliable answer without punting me to a Google Search page.

    Smarter camera alerts are indeed more helpful at a glance. Most of the time, I dismissed Person Seen notifications because they’re often just people walking by my house. Now the alerts actually say “Person walks by,” which gives me greater confidence to dismiss those. Some alerts accurately say “Two people opened the gate,” though sometimes it will hallucinate: “Person walks up stairs,” when no one actually did. (They just walked on the sidewalk.) It has fairly accurately noted when UPS, FedEx, or USPS are at the door, which is nice to know when I’m busy or out and about, so I can make sure to check for a package when I get home—no need to hunt through alerts.

    But with my indoor security cameras, Gemini routinely says I have a cat wandering the house. It’s my dog. Even in my Home Brief—recaps at the end of the day from Gemini about what happened around the home—Gemini says, “In the early morning, a white cat was active, walking into the living room and sitting on the couch.” It’s amusing, especially considering my dog hates cats.

    CatDog

    Screenshot

    ScreenshotGoogle Home via Julian Chokkattu

    You would think then that I would be able to just tell this smarter assistant, “Hey, I don’t have a cat. I have a dog,” and it would adjust its models and fix the error. Well, I did exactly that. In the Ask Home feature, you can talk to Gemini and ask it anything about the home. This is where you can ask it to set up automations, for example. I asked it to turn on the living room lights when the cameras detect my wife or I arriving home, and it understood the action. It even guessed that I wanted the lights to come on only when arriving at night, despite me forgetting to mention that.

    [ad_2]

    Julian Chokkattu

    Source link

  • Algolux Extends Eos Perception Software to Address Critical ADAS and Autonomous Vehicle Depth Limitations

    Algolux Extends Eos Perception Software to Address Critical ADAS and Autonomous Vehicle Depth Limitations

    [ad_1]

    The new Eos Robust Depth Perception Software resolves the range, resolution, cost, and robustness limitations of the latest lidar, radar, and camera based systems combined with a scalable and modular software perception suite.

    Press Release


    Sep 12, 2022

    Algolux Inc., the leading provider of robust and scalable depth perception solutions, has announced its Eos Robust Depth Perception Software. The expanded AI software offering builds on the company’s advanced perception portfolio to now deliver both industry-leading dense depth estimation and robust perception to further increase the safety of passenger and transport vehicles in all lighting and weather conditions.

    This new offering addresses the cost and performance limitations in today’s automated driver assist systems (ADAS) and autonomous vehicle (AV) platforms by applying the same breakthrough deep learning approach used by Algolux’s award winning robust perception software.

    Mercedes-Benz has a track record as an industry leader delivering ADAS and autonomous driving systems, so we intimately understand the need to further improve the perception capabilities of these systems to enable safer operation of vehicles in all operating conditions,” said Werner Ritter, Manager, Vision Enhancement Technology Environment Perception, Mercedes-Benz AG. “Algolux has leveraged the novel AI technology in their camera-based Eos Robust Depth Perception Software to deliver dense depth and detection that provide next-generation range and robustness while addressing limitations of current lidar and radar based approaches for ADAS and autonomous driving.

    The ability to estimate the distance of an object is a fundamental capability for ADAS and automated driving. It allows the vehicle to understand where things are in its surroundings in order to know when to perform a lane change, help park a car, issue a warning to the driver, or automatically brake in an emergency situation such as for debris on the road. This is accomplished today by various types of sensors, such as lidar, radar, and stereo or mono cameras, and edge software that interprets the sensor information to determine distance to important objects, features, or obstacles in front and around the vehicle.

    Unfortunately, each of the current approaches have limitations that hamper safe operation in all conditions. Lidar has limited effective range (up to 200m) due to decreasing point density the further away an object is, resulting in poor object detection capabilities, and low robustness in harsh weather conditions such as fog or rain due to backscatter of the laser. It is also costly, currently in the hundreds to thousands of dollars per sensor. Radar has good range and robustness but poor resolution, also limiting its ability to detect and classify objects. Today’s stereo camera approaches can do a good job of object detection but are hard to keep calibrated and have low robustness, and mono cameras have many issues resulting in poor depth estimation.

    Eos Robust Depth Perception Software addresses these limitations by robustly providing dense depth together with accurate perception capabilities to determine distance and elevation even out to long distances (1km) and identify objects, pedestrians, or bicyclists, and even lost cargo or other hazardous road debris to further improve driving safety. These modular capabilities provide rich 3D scene reconstruction and provide a highly capable and cost-effective alternative to lidar, radar, and today’s stereo approaches.

    Eos accomplishes this with:

    • a multi-camera approach supporting a wide baseline between the cameras, even beyond 2m, especially useful for long-range applications such as trucking
    • flexible support of up to 8MP automotive camera sensors and any field of view for forward, rear, and surround configurations
    • real time adaptive calibration to address vibration and movement between the cameras or misalignments while driving, historically a key challenge for wide-baseline configurations
    • an efficient embedded implementation of Algolux’s novel end-to-end deep learned architecture for both depth estimation and perception
    • the ability to detect and determine distance for typical road objects and even unknown road debris

    Algolux has proven its Eos depth perception performance in OEM and Tier 1 engagements involving both trucking and automotive applications in North America and Europe. Visit www.algolux.com to learn more and schedule an evaluation.

    About Algolux

    Algolux is a globally recognized computer vision company addressing the critical issue of safety for advanced driver assistance systems and autonomous vehicles. Our machine-learning tools and embedded AI software products enable existing and new camera designs to achieve industry-leading depth perception performance across all driving conditions. Founded on groundbreaking research at the intersection of deep learning, computer vision, and computational imaging, Algolux has been repeatedly recognized at industry and academic conferences and has been named to the 2021 CB Insights AI 100 List of the world’s most innovative artificial intelligence startups.

    Dave Tokic, VP Marketing and Strategic Partnerships
    pr@algolux.com

    Source: Algolux Inc.

    [ad_2]

    Source link