ReportWire

Tag: DeepMind

  • Google’s AI Weather Model Nailed Its First Major Storm Forecast

    [ad_1]

    While generative AI tools that primarily amount to slop generators grab most of the attention in the artificial intelligence space, there are occasionally some actually useful applications of the technology, like Google DeepMind’s use of AI weather models to predict cyclones. The experimental tool, launched earlier this year, successfully managed to provide accurate modeling of Hurricane Erin as it started gaining steam in the Atlantic Ocean earlier this month.

    As Ars Technica first reported, Hurricane Erin—which reached Category 5 status and caused some damage to the island of Bermuda, parts of the Caribbean, and the East Coast of the United States—provided Google DeepMind’s Weather Lab with the first real test of its capabilities.

    According to James Franklin, former chief of the hurricane specialist unit at the National Hurricane Center, it did quite well, outperforming the National Hurricane Center’s official model and topping several other physics-based models during the first 72 hours of modeling. It did ultimately fall off a bit the longer the prediction effort ran, but it still topped the consensus model through the five-day forecast.

    While Google’s model was impressively accurate in the first days of modeling, it’s the latter ones that are most important to experts, per Ars Technica, as days three through five of the model are the ones that officials count on to make decisions on calls for evacuation and other preparatory efforts. Still, it seems like there may be some promise in the possibility of AI-powered weather modeling—though the sample size here is pretty small.

    Most of the current gold standard modeling techniques used for storm prediction use physics-based prediction engines, which essentially try to recreate the conditions of the atmosphere by factoring in things like humidity, air pressure, and temperature changes to simulate how a storm might behave. Google’s model instead pulls from a massive amount of data that it was trained on, including a “reanalysis dataset that reconstructs past weather over the entire Earth from millions of observations, and a specialized database containing key information about the track, intensity, size and wind radii of nearly 5,000 observed cyclones from the past 45 years.”

    According to Google, it tested its model on storms from 2023 and 2024, and found that its five-day prediction managed to predict the path of a storm with more accuracy than most other models, coming about 140km or 90 miles closer to the ultimate location of the cyclone than the European Centre for Medium-Range Weather Forecasts’ ensemble model, which is considered the most accurate model available. Now it can point to a storm that it tracked in real-time as proof of concept, though there is no reason to think AI tools like this will completely displace the other approaches at this stage.

    [ad_2]

    AJ Dellinger

    Source link

  • Andreessen Horowitz Founders Notice A.I. Models Are Hitting a Ceiling

    [ad_1]

    The investment firm was founded by Ben Horowitz and Marc Andreessen in 2009. Photos by Phillip Faraone/Getty Images for WIRED and Paul Chinn/The San Francisco Chronicle via Getty Images

    Despite continuing to bet big on A.I. startups and chip programs, the founders of the venture capital firm Andreessen Horowitz say they’ve noticed a drop off in A.I. model capability improvements in recent years. Two years ago, OpenAI’s GPT-3.5 model was “way ahead of everybody else’s,” said Marc Andreessen, who co-founded Andreessen Horowitz alongside Ben Horowitz in 2009, on a podcast released yesterday (Nov. 5). “Sitting here today, there’s six that are on par with that. They’re sort of hitting the same ceiling on capabilities,” he added.

    That’s not to say the investment firm doesn’t have faith in the new technology. One of the most aggressive investors in the A.I. space, Andreessen Horowitz earlier this year earmarked $2.25 billion in funding for A.I.-focused applications and infrastructure and has led investments in notable companies including Mistral AI, a French startup founded by former DeepMind and Meta (META) researchers, and Air Space Intelligence, an aerospace company using A.I. to enhance air travel.

    Despite their embrace of the new technology, Andreessen and Horowitz concede there are growth limitations. In the case of OpenAI’s models, the difference in capability growth between its GPT-2.0, GPT-3 and GPT-3.5 models compared to the difference between GPT-3.5 and GPT-4 show that “we’ve really slowed down in terms of the amount of improvement,” said Horowitz.

    One of the primary challenges for A.I. developers has been a global shortage of graphics processing units (GPUs), the chips that power A.I. models. OpenAI CEO Sam Altman last week cited needs to allocate compute as causing the company to “face a lot of limitations and hard decisions” about what projects they focus on. Nvidia, the leading GPU maker, has previously described the shortage as making clients “tense” and “emotional.”

    In response to this demand, Andreessen Horowitz recently established a chip-lending program that provides GPUs to its portfolio companies in exchange for equity. The firm reportedly has been working on building a stockpile chip cluster of 20,000 GPUs, including Nvidia’s. However, chips aren’t the only aspect of compute that is of concern, according to Horowitz, who pointed to the need for more powering and cooling across the data centers housing GPUs. “Once they get chips we’re not going to have enough power, and once we have the power we’re not going to have enough cooling,” he said on yesterday’s podcast.

    But compute needs might not actually be the largest barrier when it comes to improving A.I. model capabilities, according to the venture capital firm. It’s the availability of training data needed to teach A.I. models how to behave that is increasingly becoming a problem. “The big models are trained by scraping the internet and pulling in all human-generated training data, all-human generated text and increasingly video and audio and everything else, and there’s just literally only so much of that,” said Andreessen.

    Between April of 2024 and 2023, 5 percent of all data and 25 percent of data from the highest quality sources was restricted by websites cracking down on the use of their text, images and videos in training A.I., according to a recent study from the Data Provenance Initiative.

    The issue has become so large that major A.I. labs are “hiring thousands of programmers and doctors and lawyers to actually handwrite answers to questions for the purpose of being able to train their A.I.’s—it’s at that level of constraint,” added Andreessen. OpenAI, for example, has a “Human Data Team” that works with A.I. trainers on gathering specialized data to train and evaluate models. And numerous A.I. companies have begun working with startups like Scale AI and Invisible Tech that hire human experts with specialized knowledge across medicine, law and other areas to help fine-tune A.I. model answers.

    Such practices fly in the face of fears relating to A.I.-driven unemployment, according to Andreessen, who noted that the dwindling supply of data has led to an unexpected A.I. hiring boom to help train models. “There’s an irony to this.”

    Andreessen Horowitz Founders Notice A.I. Models Are Hitting a Ceiling

    [ad_2]

    Alexandra Tremayne-Pengelly

    Source link

  • OpenAI Warns Users Could Become Emotionally Hooked on Its Voice Mode

    OpenAI Warns Users Could Become Emotionally Hooked on Its Voice Mode

    [ad_1]

    In late July, OpenAI began rolling out an eerily humanlike voice interface for ChatGPT. In a safety analysis released today, the company acknowledges that this anthropomorphic voice may lure some users into becoming emotionally attached to their chatbot.

    The warnings are included in a “system card” for GPT-4o, a technical document that lays out what the company believes are the risks associated with the model, plus details surrounding safety testing and the mitigation efforts the company’s taking to reduce potential risk.

    OpenAI has faced scrutiny in recent months after a number of employees working on AI’s long-term risks quit the company. Some subsequently accused OpenAI of taking unnecessary chances and muzzling dissenters in its race to commercialize AI. Revealing more details of OpenAI’s safety regime may help mitigate the criticism and reassure the public that the company takes the issue seriously.

    The risks explored in the new system card are wide-ranging, and include the potential for GPT-4o to amplify societal biases, spread disinformation, and aid in the development of chemical or biological weapons. It also discloses details of testing designed to ensure that AI models won’t try to break free of their controls, deceive people, or scheme catastrophic plans.

    Some outside experts commend OpenAI for its transparency but say it could go further.

    Lucie-Aimée Kaffee, an applied policy researcher at Hugging Face, a company that hosts AI tools, notes that OpenAI’s system card for GPT-4o does not include extensive details on the model’s training data or who owns that data. “The question of consent in creating such a large dataset spanning multiple modalities, including text, image, and speech, needs to be addressed,” Kaffee says.

    Others note that risks could change as tools are used in the wild. “Their internal review should only be the first piece of ensuring AI safety,” says Neil Thompson, a professor at MIT who studies AI risk assessments. “Many risks only manifest when AI is used in the real world. It is important that these other risks are cataloged and evaluated as new models emerge.”

    The new system card highlights how rapidly AI risks are evolving with the development of powerful new features such as OpenAI’s voice interface. In May, when the company unveiled its voice mode, which can respond swiftly and handle interruptions in a natural back and forth, many users noticed it appeared overly flirtatious in demos. The company later faced criticism from the actress Scarlett Johansson, who accused it of copying her style of speech.

    A section of the system card titled “Anthropomorphization and Emotional Reliance” explores problems that arise when users perceive AI in human terms, something apparently exacerbated by the humanlike voice mode. During the red teaming, or stress testing, of GPT-4o, for instance, OpenAI researchers noticed instances of speech from users that conveyed a sense of emotional connection with the model. For example, people used language such as “This is our last day together.”

    Anthropomorphism might cause users to place more trust in the output of a model when it “hallucinates” incorrect information, OpenAI says. Over time, it might even affect users’ relationships with other people. “Users might form social relationships with the AI, reducing their need for human interaction—potentially benefiting lonely individuals but possibly affecting healthy relationships,” the document says.

    Joaquin Quiñonero Candela, head of preparedness at OpenAI, says that voice mode could evolve into a uniquely powerful interface. He also notes that the kind of emotional effects seen with GPT-4o can be positive—say, by helping those who are lonely or who need to practice social interactions. He adds that the company will study anthropomorphism and the emotional connections closely, including by monitoring how beta testers interact with ChatGPT. “We don’t have results to share at the moment, but it’s on our list of concerns,” he says.

    [ad_2]

    Will Knight, Reece Rogers

    Source link

  • Google DeepMind’s Game-Playing AI Tackles a Chatbot Blind Spot

    Google DeepMind’s Game-Playing AI Tackles a Chatbot Blind Spot

    [ad_1]

    Several years before ChatGPT began jibber-jabbering away, Google developed a very different kind of artificial intelligence program called AlphaGo that learned to play the board game Go with superhuman skill through tireless practice.

    Researchers at the company have now published research that combines the abilities of a large language model (the AI behind today’s chatbots) with those of AlphaZero, a successor to AlphaGo also capable of playing chess, to solve very tricky mathematical proofs.

    Their new Frankensteinian creation, dubbed AlphaProof, has demonstrated its prowess by tackling several problems from the 2024 International Math Olympiad (IMO), a prestigious competition for high school students.

    AlphaProof uses the Gemini large language model to convert naturally phrased math questions into a programming language called Lean. This provides the training fodder for a second algorithm to learn, through trial and error, how to find proofs that can be confirmed as correct.

    Earlier this year, Google DeepMind revealed another math algorithm called AlphaGeometry that also combines a language model with a different AI approach. AlphaGeometry uses Gemini to convert geometry problems into a form that can be manipulated and tested by a program that handles geometric elements. Google today also announced a new and improved version of AlphaGeometry.

    The researchers found that their two math programs could provide proofs for IMO puzzles as well as a silver medalist could. Out of six problems total, AlphaProof solved two algebra problems and a number theory one, while AlphaGeometry solved a geometry problem. The programs got one problem in minutes but took up to several days to figure out others. Google DeepMind has not disclosed how much computer power it threw at the problems.

    Google DeepMind calls the approach used for both AlphaProof and AlphaGeometry “neuro-symbolic” because they combine the pure machine learning of an artificial neural network, the technology that underpins most progress in AI of late, with the language of conventional programming.

    “What we’ve seen here is that you can combine the approach that was so successful, and things like AlphaGo, with large language models and produce something that is extremely capable,” says David Silver, the Google DeepMind researcher who led work on AlphaZero. Silver says the techniques demonstrated with AlphaProof should, in theory, extend to other areas of mathematics.

    Indeed, the research raises the prospect of addressing the worst tendencies of large language models by applying logic and reasoning in a more grounded fashion. As miraculous as large language models can be, they often struggle to grasp even basic math or to reason through problems logically.

    In the future, the neural-symbolic method could provide a means for AI systems to turn questions or tasks into a form that can be reasoned over in a way that produces reliable results. OpenAI is also rumored to be working on such a system, codenamed “Strawberry.”

    There is, however, a key limitation with the systems revealed today, as Silver acknowledges. Math solutions are either correct or incorrect, allowing AlphaProof and AlphaGeometry to work their way toward the right answer. Many real-world problems—coming up with the ideal itinerary for a trip, for instance—have many possible solutions, and which one is ideal may be unclear. Silver says the solution for more ambiguous questions may be for a language model to try to determine what constitutes a “right” answer during training. “There’s a spectrum of different things that can be tried,” he says.

    Silver is also careful to note that Google DeepMind won’t be putting human mathematicians out of jobs. “We are aiming to provide a system that can prove anything, but that’s not the end of what mathematicians do,” he says. “A big part of mathematics is to pose problems and find what are the interesting questions to ask. You might think of this as another tool along the lines of a slide rule or calculator or computational tools.”

    Updated 7/25/24 1:25 pm ET: This story has been updated to clarify how many problems AlphaProof and AlphaGeometry solved, and of what type.

    [ad_2]

    Will Knight

    Source link

  • Google DeepMind’s Chatbot-Powered Robot Is Part of a Bigger Revolution

    Google DeepMind’s Chatbot-Powered Robot Is Part of a Bigger Revolution

    [ad_1]

    In a cluttered open-plan office in Mountain View, California, a tall and slender wheeled robot has been busy playing tour guide and informal office helper—thanks to a large language model upgrade, Google DeepMind revealed today. The robot uses the latest version of Google’s Gemini large language model to both parse commands and find its way around.

    When told by a human “Find me somewhere to write,” for instance, the robot dutifully trundles off, leading the person to a pristine whiteboard located somewhere in the building.

    Gemini’s ability to handle video and text—in addition to its capacity to ingest large amounts of information in the form of previously recorded video tours of the office—allows the “Google helper” robot to make sense of its environment and navigate correctly when given commands that require some commonsense reasoning. The robot combines Gemini with an algorithm that generates specific actions for the robot to take, such as turning, in response to commands and what it sees in front of it.

    When Gemini was introduced in December, Demis Hassabis, CEO of Google DeepMind, told WIRED that its multimodal capabilities would likely unlock new robot abilities. He added that the company’s researchers were hard at work testing the robotic potential of the model.

    In a new paper outlining the project, the researchers behind the work say that their robot proved to be up to 90 percent reliable at navigating, even when given tricky commands such as “Where did I leave my coaster?” DeepMind’s system “has significantly improved the naturalness of human-robot interaction, and greatly increased the robot usability,” the team writes.

    Courtesy of Google DeepMind

    A photo of a Google DeepMind employee interacting with an AI robot.

    Photograph: Muinat Abdul; Google DeepMind

    The demo neatly illustrates the potential for large language models to reach into the physical world and do useful work. Gemini and other chatbots mostly operate within the confines of a web browser or app, although they are increasingly able to handle visual and auditory input, as both Google and OpenAI have demonstrated recently. In May, Hassabis showed off an upgraded version of Gemini capable of making sense of an office layout as seen through a smartphone camera.

    Academic and industry research labs are racing to see how language models might be used to enhance robots’ abilities. The May program for the International Conference on Robotics and Automation, a popular event for robotics researchers, lists almost two dozen papers that involve use of vision language models.

    Investors are pouring money into startups aiming to apply advances in AI to robotics. Several of the researchers involved with the Google project have since left the company to found a startup called Physical Intelligence, which received an initial $70 million in funding; it is working to combine large language models with real-world training to give robots general problem-solving abilities. Skild AI, founded by roboticists at Carnegie Mellon University, has a similar goal. This month it announced $300 million in funding.

    Just a few years ago, a robot would need a map of its environment and carefully chosen commands to navigate successfully. Large language models contain useful information about the physical world, and newer versions that are trained on images and video as well as text, known as vision language models, can answer questions that require perception. Gemini allows Google’s robot to parse visual instructions as well as spoken ones, following a sketch on a whiteboard that shows a route to a new destination.

    In their paper, the researchers say they plan to test the system on different kinds of robots. They add that Gemini should be able to make sense of more complex questions, such as “Do they have my favorite drink today?” from a user with a lot of empty Coke cans on their desk.

    [ad_2]

    Will Knight

    Source link

  • Forget Chatbots. AI Agents Are the Future

    Forget Chatbots. AI Agents Are the Future

    [ad_1]

    This week a startup called Cognition AI caused a bit of a stir by releasing a demo showing an artificial intelligence program called Devin performing work usually done by well-paid software engineers. Chatbots like ChatGPT and Gemini can generate code, but Devin went further, planning how to solve a problem, writing the code, and then testing and implementing it.

    Devin’s creators brand it as an “AI software developer.” When asked to test how Meta’s open source language model Llama 2 performed when accessed via different companies hosting it, Devin generated a step-by-step plan for the project, generated code needed to access the APIs and run benchmarking tests, and created a website summarizing the results.

    It’s always hard to judge staged demos, but Cognition has shown Devin handling a wide range of impressive tasks. It wowed investors and engineers on X, receiving plenty of endorsements, and even inspired a few memes—including some predicting Devin will soon be responsible for a wave of tech industry layoffs.

    Devin is just the latest, most polished example of a trend I’ve been tracking for a while—the emergence of AI agents that instead of just providing answers or advice about a problem presented by a human can take action to solve it. A few months back I test drove Auto-GPT, an open source program that attempts to do useful chores by taking actions on a person’s computer and on the web. Recently I tested another program called vimGPT to see how the visual skills of new AI models can help these agents browse the web more efficiently.

    I was impressed by my experiments with those agents. Yet for now, just like the language models that power them, they make quite a few errors. And when a piece of software is taking actions, not just generating text, one mistake can mean total failure—and potentially costly or dangerous consequences. Narrowing the range of tasks an agent can do to, say, a specific set of software engineering chores seems like a clever way to reduce the error rate, but there are still many potential ways to fail.

    Not only startups are building AI agents. Earlier this week I wrote about an agent called SIMA, developed by Google DeepMind, which plays video games including the truly bonkers title Goat Simulator 3. SIMA learned from watching human players how to do more than 600 fairly complicated tasks such as chopping down a tree or shooting an asteroid. Most significantly, it can do many of these actions successfully even in an unfamiliar game. Google DeepMind calls it a “generalist.”

    I suspect that Google has hopes that these agents will eventually go to work outside of video games, perhaps helping use the web on a user’s behalf or operate software for them. But video games make a good sandbox for developing and testing agents, by providing complex environments in which they can be tested and improved. “Making them more precise is something that we’re actively working on,” Tim Harley, a research scientist at Google DeepMind, told me. “We’ve got various ideas.”

    You can expect a lot more news about AI agents in the coming months. Demis Hassabis, the CEO of Google DeepMind, recently told me that he plans to combine large language models with the work his company has previously done training AI programs to play video games to develop more capable and reliable agents. “This definitely is a huge area. We’re investing heavily in that direction, and I imagine others are as well.” Hassabis said. “It will be a step change in capabilities of these types of systems—when they start becoming more agent-like.”

    [ad_2]

    Will Knight

    Source link

  • Google’s Chess Experiments Reveal How to Boost the Power of AI

    Google’s Chess Experiments Reveal How to Boost the Power of AI

    [ad_1]

    His group decided to find out. They built the new, diversified version of AlphaZero, which includes multiple AI systems that trained independently and on a variety of situations. The algorithm that governs the overall system acts as a kind of virtual matchmaker, Zahavy said: one designed to identify which agent has the best chance of succeeding when it’s time to make a move. He and his colleagues also coded in a “diversity bonus”—a reward for the system whenever it pulled strategies from a large selection of choices.

    When the new system was set loose to play its own games, the team observed a lot of variety. The diversified AI player experimented with new, effective openings and novel—but sound—decisions about specific strategies, such as when and where to castle. In most matches, it defeated the original AlphaZero. The team also found that the diversified version could solve twice as many challenge puzzles as the original and could solve more than half of the total catalog of Penrose puzzles.

    “The idea is that instead of finding one solution, or one single policy, that would beat any player, here [it uses] the idea of creative diversity,” Cully said.

    With access to more and different played games, Zahavy said, the diversified AlphaZero had more options for sticky situations when they arose. “If you can control the kind of games that it sees, you basically control how it will generalize,” he said. Those weird intrinsic rewards (and their associated moves) could become strengths for diverse behaviors. Then the system could learn to assess and value the disparate approaches and see when they were most successful. “We found that this group of agents can actually come to an agreement on these positions.”

    And, crucially, the implications extend beyond chess.

    Real-Life Creativity

    Cully said a diversified approach can help any AI system, not just those based on reinforcement learning. He has long used diversity to train physical systems, including a six-legged robot that was allowed to explore various kinds of movement, before he intentionally “injured” it, allowing it to continue moving using some of the techniques it had developed before. “We were just trying to find solutions that were different from all previous solutions we have found so far.” Recently, he has also been collaborating with researchers to use diversity to identify promising new drug candidates and develop effective stock-trading strategies.

    “The goal is to generate a large collection of potentially thousands of different solutions, where every solution is very different from the next,” Cully said. So—just as the diversified chess player learned to do—for every type of problem, the overall system could choose the best possible solution. Zahavy’s AI system, he said, clearly shows how “searching for diverse strategies helps to think outside the box and find solutions.”

    Zahavy suspects that in order for AI systems to think creatively, researchers simply have to get them to consider more options. That hypothesis suggests a curious connection between humans and machines: Maybe intelligence is just a matter of computational power. For an AI system, maybe creativity boils down to the ability to consider and select from a large enough buffet of options. As the system gains rewards for selecting a variety of optimal strategies, this kind of creative problem-solving gets reinforced and strengthened. Ultimately, in theory, it could emulate any kind of problem-solving strategy recognized as a creative one in humans. Creativity would become a computational problem.

    Liemhetcharat noted that a diversified AI system is unlikely to completely resolve the broader generalization problem in machine learning. But it’s a step in the right direction. “It’s mitigating one of the shortcomings,” she said.

    More practically, Zahavy’s results resonate with recent efforts that show how cooperation can lead to better performance on hard tasks among humans. Most of the hits on the Billboard 100 list were written by teams of songwriters, for example, not individuals. And there’s still room for improvement. The diverse approach is currently computationally expensive, since it must consider so many more possibilities than a typical system. Zahavy is also not convinced that even the diversified AlphaZero captures the entire spectrum of possibilities.

    “I still [think] there is room to find different solutions,” he said. “It’s not clear to me that given all the data in the world, there is [only] one answer to every question.”


    Original story reprinted with permission from Quanta Magazine, an editorially independent publication of the Simons Foundation whose mission is to enhance public understanding of science by covering research developments and trends in mathematics and the physical and life sciences.

    [ad_2]

    Stephen Ornes

    Source link

  • A new breed of companies expand in San Francisco’s prime areas | TechCrunch

    A new breed of companies expand in San Francisco’s prime areas | TechCrunch

    [ad_1]

    Ten years ago, Pear VC, then a tiny new venture firm, operated out of a nondescript office in Palo Alto that was enlivened by bright, computer-themed art. Last week, the outfit — which closed its largest fund to date in May — quietly inked a deal to sublease 30,000 square feet of “Class A” office space in San Francisco’s Mission Bay neighborhood from the file-storage giant Dropbox.

    It’s among a number of fast-growing outfits taking up more space in San Francisco as an earlier generation of companies shrinks its physical footprint.

    As the San Francisco Chronicle first reported last week, ChatGPT creator OpenAI just subleased two buildings totaling a collective 486,600 square feet from Uber. The ride-share giant, which originally leased a grouping of four buildings down the street from Dropbox and will continue to occupy two of these, told the paper it is “right-sizing.”

    A rival to OpenAI — Anthropic — also just reportedly closed a sizable subleasing deal. Its plan: to take over the entire 250,000-square-foot building in downtown San Francisco that was previously Slack’s headquarters.

    Salesforce, which acquired Slack in 2021, is an investor in Anthropic. Meanwhile, Pear VC co-founder Pejman Nozad wrote one of the first small checks to Dropbox when he was still relatively new to the U.S. from Iran and selling Persian rugs to Silicon Valley bigwigs.

    Such subleases don’t necessarily begin with hand-shake deals, however. Asked if Nozad zeroed in on Pear’s new space owing to his connection to Dropbox, he scoffs. The office — which has room for more than 200 desks, features more than 20 conference and call rooms, and has dedicated event space to host talks — “was a business deal for them,” says Nozad. “The founders were not involved. As you know, I sold rugs for 17 years, so I have some skills in negotiation,” he adds with a laugh.

    Certainly, it’s a good time to strike a subleasing deal if you’re a well-funded company on the rise. According to Colin Yasukochi, an executive director at the commercial real estate services firm CBRE, subleases in prime areas like Mission Bay and the city’s Financial District currently range from $60 to $80 per square foot. The higher the floor and the more plentiful the amenities, the higher the price. For startups willing to sublease space with less than five years left on the lessee’s contract, the better the terms (as they’ll need to lease again somewhere else in the not-too-distant future). In comparison, office lease rates passed the $75 per square foot mark in September 2019 before the pandemic turned the city upside down.

    There’s no shortage of options right now. San Francisco’s commercial buildings are currently 35% vacant, and there are still more tenants flowing out the door than entering them.

    Dropbox originally leased the entire 750,000-square-foot space in the building it currently occupies, but it never filled it up entirely and after COVID struck, it began more aggressively whittling down its use. It paid $32 million in late 2021 to terminate part of its 15-year lease; before newly subleasing space to Pear VC, it separately subleased roughly 200,000 square feet to two different life sciences companies: Vir Biotechnology and BridgeBio. It’s still less than half full.

    This week, Adobe listed half its leased footprint in San Francisco’s Showplace Square neighborhood and is now looking to sublease 156,000 square feet across three floors of one of the buildings it used to occupy.

    But a tipping point is seemingly in sight. There was “negative net absorption” of 1.85 million square feet in San Francisco in the third quarter of this year, according to CBRE data; at the same time, market demand reached 5.2 million square feet, which is the highest increase since the first quarter of 2020.

    Much of that shift can be traced to companies like OpenAI, suggests Yasukochi, who says that a new spate of outfits is starting to set up shop, enticed by the opportunity to rent sleeker space for the same or better prices than was possible several years ago for less finished locations, and in more central areas of the city. “It’s a huge opportunity for companies that are trying to bring back their employees,” says Yasukochi. (OpenAI CEO Sam Altman has long said he thinks companies are more effective when employees convene in person.)

    Indeed, Yasukochi anticipates that if the economy improves in the second half of new year and interest rates come down, tech outfits in particular will be positioned to recover faster — and pull the city along with them. “Many tech companies were quick to cut excess employees, along with real estate and other costs,” says Yasukochi. He also says that while tech outfits are typically “early to cut back, they’re also early to grow. I don’t see any other industry that generates the volume of growth that tech can.”

    Worth noting: Yasukochi does not think those tech companies will necessarily be growing in San Francisco’s Hayes Valley. Though the small shop-studded neighborhood has led a resurgence of interest in San Francisco this year and eagerly embraced the moniker “Cerebral Valley,” owing to its concentration of AI communities, most of those teams, he observes, are “meeting in restaurants and bars and working out of their apartments.”

    The reality, Yasukochi continues, is “there isn’t a lot of office space there.”

    Pictured above: 1800 Owens Street in San Francisco, which is the site of Dropbox’s headquarters and now, Pear VC’s San Francisco office, too.

    [ad_2]

    Connie Loizos

    Source link

  • Google DeepMind’s robotics head on general purpose robots, generative AI and office WiFi | TechCrunch

    Google DeepMind’s robotics head on general purpose robots, generative AI and office WiFi | TechCrunch

    [ad_1]

    [A version of this piece first appeared in TechCrunch’s robotics newsletter, Actuator. Subscribe here.]

    Earlier this month, Google’s DeepMind team debuted Open X-Embodiment, a database of robotics functionality created in collaboration with 33 research institutes. The researchers involved compared the system to ImageNet, the landmark database founded in 2009 that is now home to more than 14 million images.

    “Just as ImageNet propelled computer vision research, we believe Open X-Embodiment can do the same to advance robotics,” researchers Quan Vuong and Pannag Sanketi noted at the time. “Building a dataset of diverse robot demonstrations is the key step to training a generalist model that can control many different types of robots, follow diverse instructions, perform basic reasoning about complex tasks and generalize effectively.”

    At the time of its announcement, Open X-Embodiment contained 500+ skills and 150,000 tasks gathered from 22 robot embodiments. Not quite ImageNet numbers, but it’s a good start. DeepMind then trained its RT-1-X model on the data and used it to train robots in other labs, reporting a 50% success rate compared to the in-house methods the teams had developed.

    I’ve probably repeated this dozens of times in these pages, but it truly is an exciting time for robotic learning. I’ve talked to so many teams approaching the problem from different angles with ever-increasing efficacy. The reign of the bespoke robot is far from over, but it certainly feels as though we’re catching glimpses of a world where the general-purpose robot is a distinct possibility.

    Simulation will undoubtedly be a big part of the equation, along with AI (including the generative variety). It still feels like some firms have put the horse before the cart here when it comes to building hardware for general tasks, but a few years down the road, who knows?

    Vincent Vanhoucke is someone I’ve been trying to pin down for a bit. If I was available, he wasn’t. Ships in the night and all that. Thankfully, we were finally able to make it work toward the end of last week.

    Vanhoucke is new to the role of Google DeepMind’s head of robotics, having stepped into the role back in May. He has, however, been kicking around the company for more than 16 years, most recently serving as a distinguished scientist for Google AI Robotics. All told, he may well be the best possible person to talk to about Google’s robotic ambitions and how it got here.

    Image Credits: Google

    At what point in DeepMind’s history did the robotics team develop?

    I was originally not on the DeepMind side of the fence. I was part of Google Research. We recently merged with the DeepMind efforts. So, in some sense, my involvement with DeepMind is extremely recent. But there is a longer history of robotics research happening at Google DeepMind. It started from the increasing view that perception technology was becoming really, really good.

    A lot of the computer vision, audio processing, and all that stuff was really turning the corner and becoming almost human level. We starting to ask ourselves, “Okay, assuming that this continues over the next few years, what are the consequences of that?” One of clear consequence was that suddenly having robotics in a real-world environment was going to be a real possibility. Being able to actually evolve and perform tasks in an everyday environment was entirely predicated on having really, really strong perception. I was initially working on general AI and computer vision. I also worked on speech recognition in the past. I saw the writing on the wall and decided to pivot toward using robotics as the next stage of our research.

    My understanding is that a lot of the Everyday Robots team ended up on this team. Google’s history with robotics dates back significantly farther. It’s been 10 yeas since Alphabet made all of those acquisitions [Boston Dynamics, etc.]. It seems like a lot of people from those companies have populated Google’s existing robotics team.

    There’s a significant fraction of the team that came through those acquisitions. It was before my time — I was really involved in computer vision and speech recognition, but we still have a lot of those folks. More and more, we came to the conclusion that the entire robotics problem was subsumed by the general AI problem. Really solving the intelligence part was the key enabler of any meaningful process in real-world robotics. We shifted a lot of our efforts toward solving that perception, understanding and controlling in the context of general AI was going to be the meaty problem to solve.

    It seemed like a lot of the work that Everyday Robots was doing touched on general AI or generative AI. Is the work that team was doing being carried over to the DeepMind robotics team?

    We had been collaborating with Everyday Robots for, I want to say, seven years already. Even though we were two separate teams, we have very, very deep connections. In fact, one of the things that prompted us to really start looking into robotics at the time was a collaboration that was a bit of a skunkworks project with the Everyday Robots team, where they happened to have a number of robot arms lying around that had been discontinued. They were one generation of arms that had led to a new generation, and they were just lying around, doing nothing.

    We decided it would be fun to pick up those arms, put them all in a room and have them practice and learn how to grasp objects. The very notion of learning a grasping problem was not in the zeitgeist at the time. The idea of using machine learning and perception as the way to control robotic grasping was not something that had been explored. When the arms succeeded, we gave them a reward, and when they failed, we give them a thumbs-down.

    For the first time, we used machine learning and essentially solved this problem of generalized grasping, using machine learning and AI. That was a lightbulb moment at the time. There really was something new there. That triggered both the investigations with Everyday Robots around focusing on machine learning as a way to control those robots. And also, on the research side, pushing a lot more robotics as an interesting problem to apply all of the deep learning AI techniques that we’ve been able to work so well into other areas.

    DeepMind embodied AI

    Image Credits: DeepMind

    Was Everyday Robots absorbed by your team?

    A fraction of the team was absorbed by my team. We inherited their robots and still use them. To date, we’re continuing to develop the technology that they really pioneered and were working on. The entire impetus lives on with a slightly different focus than what was originally envisioned by the team. We’re really focusing on the intelligence piece a lot more than the robot building.

    You mentioned that the team moved into the Alphabet X offices. Is there something deeper there, as far as cross-team collaboration and sharing resources?

    It’s a very pragmatic decision. They have good Wi-Fi, good power, lots of space.

    I would hope all the Google buildings would have good Wi-Fi.

    You’d hope so, right? But it was a very pedestrian decision of us moving in here. I have to say, a lot of the decision was they have a good café here. Our previous office had not so good food, and people were starting to complain. There is no hidden agenda there. We like working closely with the rest of X. I think there’s a lot of synergies there. They have really talented roboticists working on a number of projects. We have collaborations with Intrinsic that we like to nurture. It makes a lot of sense for us to be here, and it’s a beautiful building.

    There’s a bit of overlap with Intrinsic, in terms of what they’re doing with their platform — things like no-code robotics and robotics learning. They overlap with general and generative AI.

    It’s interesting how robotics has evolved from every corner being very bespoke and taking on a very different set of expertise and skills. To a large extent, the journey we’re on is to try and make general-purpose robotics happen, whether it’s applied to an industrial setting or more of a home setting. The principles behind it, driven by a very strong AI core, are very similar. We’re really pushing the envelope in trying to explore how we can support as broad an application space as possible. That’s new and exciting. It’s very greenfield. There’s lots to explore in the space.

    I like to ask people how far off they think we are from something we can reasonably call general-purpose robotics.

    There is a slight nuance with the definition of general-purpose robotics. We’re really focused on general-purpose methods. Some methods can be applied to both industrial or home robots or sidewalk robots, with all of those different embodiments and form factors. We’re not predicated on there being a general-purpose embodiment that does everything for you, more than if you have an embodiment that is very bespoke for your problem. It’s fine. We can quickly fine-tune it into solving the problem that you have, specifically. So this is a big question: Will general-purpose robots happen? That’s something a lot of people are tossing around hypotheses about, if and when it will happen.

    Thus far there’s been more success with bespoke robots. I think, to some extent, the technology has not been there to enable more general-purpose robots to happen. Whether that’s where the business mode will take us is a very good question. I don’t think that question can be answered until we have more confidence in the technology behind it. That’s what we’re driving right now. We’re seeing more signs of life — that very general approaches that don’t depend on a specific embodiment are plausible. The latest thing we’ve done is this RTX project. We went around to a number of academic labs — I think we have 30 different partners now — and asked to look at their task and the data they’ve collected. Let’s pull that into a common repository of data, and let’s train a large model on top of it and see what happens.

    DeepMind RoboCat

    Image Credits: DeepMind

    What role will generative AI play in robotics?

    I think it’s going to be very central. There was this large language model revolution. Everybody started asking whether we can use a lot of language models for robots, and I think it could have been very superficial. You know, “Let’s just pick up the fad of the day and figure out what we can do with it,” but it’s turned out to be extremely deep. The reason for that is, if you think about it, language models are not really about language. They’re about common sense reasoning and understanding of the everyday world. So, if a large language model knows you’re looking for a cup of coffee, you can probably find it in a cupboard in a kitchen or on a table.

    Putting a coffee cup on a table makes sense. Putting a table on top of a coffee cup is nonsensical. It’s simple facts like that you don’t really think about, because they’re completely obvious to you. It’s always been really hard to communicate that to an embodied system. The knowledge is really, really hard to encode, while those large language models have that knowledge and encode it in a way that’s very accessible and we can use. So we’ve been able to take this common-sense reasoning and apply it to robot planning. We’ve been able to apply it to robot interactions, manipulations, human-robot interactions, and having an agent that has this common sense and can reason about things in a simulated environment, alongside with perception is really central to the robotics problem.

    DeepMind Gato

    The various tasks that Gato learned to complete.

    Simulation is probably a big part of collecting data for analysis.

    Yeah. It’s one ingredient to this. The challenge with simulation is that then you need to bridge the simulation-to-reality gap. Simulations are an approximation of reality. It can be very difficult to make very precise and very reflective of reality. The physics of a simulator have to be good. The visual rendering of the reality in that simulation has to be very good. This is actually another area where generative AI is starting to make its mark. You can imagine instead of actually having to run a physics simulator, you just generate using image generation or a generative model of some kind.

    Tye Brady recently told me Amazon is using simulation to generate packages.

    That makes a lot of sense. And going forward, I think beyond just generating assets, you can imagine generating futures. Imagine what would happen if the robot did an action? And verifying that it’s actually doing the thing you wanted it to and using that as a way of planning for the future. It’s sort of like the robot dreaming, using generative models, as opposed to having to do it in the real world.

    [ad_2]

    Brian Heater

    Source link