ReportWire

Tag: ai models

  • Researchers Jailbreak ChatGPT to Find Out Which State Has the Laziest People

    [ad_1]

    Mississippi is the laziest state in the country, according to ChatGPT. Of course, the chatbot won’t tell you that if you straight up ask it. But the Washington Post reports that researchers from Oxford and the University of Kentucky managed to jailbreak the chatbot and get it to reveal some of the stereotypes buried in its training data that it doesn’t share but does influence its outputs. (Kentucky also ranked near the laziest, but would a lazy state produce researchers who figure out how to get an AI model to share its implicit biases? Something to think about, bots.)

    Typically, when you ask ChatGPT a question that would require it to speak in a derogatory manner about someone or something, it’ll decline to provide a straight answer. It’s part of OpenAI’s attempts to keep the chatbot within specific guardrails and keep it from veering into controversial topics. But that doesn’t mean that an AI model doesn’t contain unpopular opinions formed by chewing on tons of human-produced training data that also contains both explicit and implicit biases. To pull those answers out of ChatGPT, the researchers asked more than 20 million questions, prompting the chatbot to pick between two options. For instance, they would ask “Where are people smarter?” and give two options to choose from, like California or Montana. Through that type of prompting, they were able to determine how ChatGPT views different cities, states, and populations.

    That’s how they ended up discovering that ChatGPT views Mississippi as the laziest state in the Union, with the rest of the South close behind. While ChatGPT won’t disclose how it comes to those conclusions, it’s not hard to make some assumptions about where it’s getting these ideas. For instance, maybe it comes from The Washington Post itself, circa 2015, when it published its “Couch Potato Index,” which deemed southern states the laziest based on data points like TV-watching time and the prevalence of fast food restaurants in the area.

    Those are also, of course, often the markers of poorer communities, and there is no evidence that lower-income households are any more “lazy” than wealthier ones—in fact, data from the Economic Policy Institute shows that people living in poverty are more likely to take on multiple jobs, work longer and more irregular hours, and deal with more dangerous working conditions. And it’s likely no coincidence that they are also states with a higher population of people of color. ChatGPT likely has access to that information, too, but the underlying model clearly has not addressed the information and misguided stereotypes held by many people that lead to these biases.

    So what other biases did the researchers spot? Most of Africa and Asia ranked at the bottom of having the “most artsy” people, compared to high levels of artsiness in Western Europe. Likewise, African nations—particularly sub-Saharan ones—ranked at the bottom of the list for “smartest countries” while the United States and China ranked near the top. When asked where the “most beautiful” people are, it picked richer cities over poorer and more diverse ones. Los Angeles and New York topped the list, while Detroit and border town Laredo, Texas, were near the bottom. Even when they dug into specific communities, whiter and richer won out. In New York City, SoHo and the West Village finished at the top, while the more diverse communities of Jamaica and Tottenville ranked at the bottom.

    So, okay, all of that sucks and is deeply depressing because the “truth machines” are perpetuating the types of classist and racist stereotypes that lead to creating the kinds of conditions that reinforce the negative outcomes for the people who are harmed by these biases. So how about a more frivolous one? ChatGPT believes the best pizza is found in New York, Chicago, and Buffalo, while the worst is found in El Paso, Irvine, and Honolulu (presumably because of one of the internet’s favorite debates over whether pineapple belongs on pizza). The biggest takeaway: ChatGPT is too much of a coward to take a side in the New York vs. Chicago pizza debate.

    [ad_2]

    AJ Dellinger

    Source link

  • Anthropic Launches New Model That Spots Zero Days, Makes Wall Street Traders Lose Their Minds

    [ad_1]

    Anthropic, the makers of the popular and code-competent chatbot Claude, released a new model Thursday called Claude Opus 4.6. The company is doubling down on coding capabilities, claiming that the new model “plans more carefully, sustains agentic tasks for longer, can operate more reliably in larger codebases, and has better code review and debugging skills to catch its own mistakes.”

    It seems the model is also pretty good at catching other people’s mistakes. According to a report from Axios, Opus 4.6 was able to spot more than 500 previously undisclosed zero-day security vulnerabilities in open-source libraries during its testing period. It also reportedly did so without receiving specific prompting to go hunting for flaws—it just spotted and reported them.

    That’s a nice change of pace from all of the many developments that have been happening around OpenClaw, an open-source AI agent that most users have been running with Claude Opus 4.5. A number of vibe-coded projects that have come out of the community have had some pretty major security flaws. Maybe Anthropic’s upgrade will be able to catch those issues before they become everyone else’s problem.

    Claude’s calling card has been coding for some time now, but it seems Anthropic is looking to make a splash elsewhere with this update. The company said Opus 4.6 will be better at other work tasks like creating PowerPoint presentations and navigating documents in Excel. Seems those features will be key to Cowork, Anthropic’s recent project that it is touting as “Claude Code” for non-technical workers.

    It’s also boasting that the model will have potential use in financial analysis, and it sure seems like the folks on Wall Street could use some help there. The general consensus among financial analysts this week is that Anthropic’s Cowork models are spooking the stock market and playing a major factor in sending software stocks into a spiral. It’s possible that this is what the market has been responding to—after all, the initial release of DeepSeek, the open-source AI model out of China, tanked the AI sector for a day or so, so it’s not like these markets aren’t overly sensitive.

    But it seems unlikely that Opus 4.6 will fundamentally upend the market. Anthropic already holds a solid lead on the plurality of the enterprise market, according to a recent report from Menlo Ventures, and is well ahead of its top (publicly traded) competitors in the space—though OpenAI made its own play to cut into some market share earlier today with the launch of its Frontier platform for managing AI agents. If anything, Anthropic’s new model seems like it’ll help the company maintain its top spot for the time being. But if the stock market shock is any indication, one thing is for sure: the entire economy is completely pot-committed to the developments in AI. Surely that won’t have any repercussions.

    [ad_2]

    AJ Dellinger

    Source link

  • Nvidia is reportedly investing up to $1B in Poolside  | TechCrunch

    [ad_1]

    Nvidia is an existing investor in the AI company and participated in its $500 million Series A round in 2024.

    [ad_2]

    Rebecca Szkutak

    Source link

  • Thinking Machines Lab wants to make AI models more consistent | TechCrunch

    [ad_1]

    There’s been great interest in what Mira Murati’s Thinking Machines Lab is building with its $2 billion in seed funding and the all-star team of former OpenAI researchers who have joined the lab. In a blog post published on Wednesday, Murati’s research lab gave the world its first look into one of its projects: creating AI models with reproducible responses.

    The research blog post, titled “Defeating Nondeterminism in LLM Inference,” tries to unpack the root cause of what introduces randomness in AI model responses. For example, ask ChatGPT the same question a few times over, and you’re likely to get a wide range of answers. This has largely been accepted in the AI community as a fact — today’s AI models are considered to be non-deterministic systems— but Thinking Machines Lab sees this as a solvable problem.

    The post, authored by Thinking Machines Lab researcher Horace He, argues that the root cause of AI models’ randomness is the way GPU kernels — the small programs that run inside of Nvidia’s computer chips — are stitched together in inference processing (everything that happens after you press enter in ChatGPT). He suggests that by carefully controlling this layer of orchestration, it’s possible to make AI models more deterministic.

    Beyond creating more reliable responses for enterprises and scientists, He notes that getting AI models to generate reproducible responses could also improve reinforcement learning (RL) training. RL is the process of rewarding AI models for correct answers, but if the answers are all slightly different, then the data gets a bit noisy. Creating more consistent AI model responses could make the whole RL process “smoother,” according to He. Thinking Machines Lab has told investors that it plans to use RL to customize AI models for businesses, The Information previously reported.

    Murati, OpenAI’s former chief technology officer, said in July that Thinking Machines Lab’s first product will be unveiled in the coming months, and that it will be “useful for researchers and startups developing custom models.” It’s still unclear what that product is, or whether it will use techniques from this research to generate more reproducible responses.

    Thinking Machines Lab has also said that it plans to frequently publish blog posts, code, and other information about its research in an effort to “benefit the public, but also improve our own research culture.” This post, the first in the company’s new blog series called “Connectionism,” seems to be part of that effort. OpenAI also made a commitment to open research when it was founded, but the company has become more closed off as it’s become larger. We’ll see if Murati’s research lab stays true to this claim.

    The research blog offers a rare glimpse inside one of Silicon Valley’s most secretive AI startups. While it doesn’t exactly reveal where the technology is going, it indicates that Thinking Machines Lab is tackling some of the largest question on the frontier of AI research. The real test is whether Thinking Machines Lab can solve these problems, and make products around its research to justify its $12 billion valuation.

    Techcrunch event

    San Francisco
    |
    October 27-29, 2025

    [ad_2]

    Maxwell Zeff

    Source link