Xorte logo

News Markets Groups

USA | Europe | Asia | World| Stocks | Commodities



Add a new RSS channel

 
 


Keywords

2025-04-22 10:00:00| Fast Company

Anthropic, Menlo Ventures, and other AI industry players are betting $50 million on a company called Goodfire, which aims to understand how AI models think and steer them toward better, safer answers. Even as AI becomes more embedded in business systems and personal lives, researchers still lack a clear understanding of how AI models generate their output. So far, the go-to method for improving AI behavior has focused on shaping training data and refining prompting methods, rather than addressing the models internal thought processes. Goodfire is tackling the latterand showing real promise. The company boasts a kind of dream team of mechanistic interpretability pioneers. Cofounder Tom McGrath helped create the interpretability team at DeepMind. Cofounder Lee Sharkey pioneered the use of sparse autoencoders in language models. Nick Cammarata started the interpretability team at OpenAI alongside Chris Olah, who later cofounded Anthropic. Collectively, these researchers have delivered some of the fields biggest breakthroughs. Goodfire founder and CEO Eric Ho, who left a successful AI app company in 2022 to focus on interpretability, tells Fast Company that the new funding will be used to expand the research team and enhance its Ember interpretability platform. In addition to its core research efforts, Goodfire also generates revenue by deploying field teams to help client organizations understand and control the outputs of their AI models. Goodfire is developing the knowledge and tools needed to perform brain surgery on AI models. Its researchers have found ways to isolate modules within neural networks to reveal the AIs thoughts. Using a technique they call neural programming, they can intervene and redirect a models cognition toward higher-quality, more aligned outputs. We envision a future where you can bring a little bit of the engineering back to neural networks, Ho says. The company has also been collaborating with other AI labs to solve interpretability challenges. For example, Goodfire has helped the Arc Institute interpret the inner workings of its Evo 2 DNA foundation model, which analyzes nucleotide sequences and predicts what comes next. By understanding how the model makes its predictions, researchers have uncovered unique biological conceptspotentially valuable for new scientific discoveries. Anthropic, too, may benefit from Goodfires insights. “Our investment in Goodfire reflects our belief that mechanistic interpretability is among the best bets to help us transform black-box neural networks into understandable, steerable systemsa critical foundation for the responsible development of powerful AI,” Anthropic CEO Dario Amodei said in a statement. According to Ho, Goodfire has also been fielding requests from Fortune 500 companies that want to better understand how the large language models they use for business are thinkingand how to change faulty reasoning into sound decision-making. He notes that many within businesses still see AI models as another kind of software, something that can be reprogrammed when it produces incorrect outputs. But AI works differently: It generates responses based on probabilities and a degree of randomness. Improving those outputs requires intervention within the models cognitive processes, steering them in more productive directions. This kind of intervention is still a new and imprecise science. It remains crude and at a high level and not precise, Ho says. Still, Goodfire offers an initial tool kit that gives enterprises a level of control more familiar from traditional deterministic software. As companies increasingly rely on AI for decisions that affect real lives, Ho believes the ability to understand and redirect AI models will become essential. For instance, if a developer equips a model with ethical or safety guardrails, an organization should be able to locate the layer or parameter in the neural network where the model chose to bypass the rulesor tried to appear compliant while it wasnt. This would mean turning the AI black box into a glass box, with tools to reach inside and make necessary adjustments. Ho is optimistic that interpretability research can rise to the challenge. This is a solvable, tractable, technical problem, but it’s going to take our smartest researchers and engineers to solve the really hard problem of understanding and aligning models to human goals and morals. As AI systems begin to surpass human intelligence, concerns are growing about their alignment with human values and interests. A major part of the challenge lies in simply understanding whats happening inside AI models, which often think in alien, opaque ways. Whether the big AI labs are investing enough in interpretability remains an open questionone with serious implications for our readiness for an AI-driven future. Thats why its encouraging to see major industry players putting real funding behind an interpretability research lab. Lightspeed Venture Partners, B Capital, Work-Bench, Wing, and South Park Commons also participated in the funding round. Menlo Ventures partner Deedy Das will join Goodfires board of directors. While most of the tech world now rushes ahead with the development and application of generative AI models, concerns about the inscrutable nature of the models often get brushed aside as afterthoughts. But that wasnt always the case. Google hesitated to put generative models into production because it feared being sued over unexpected and unexplainable model outputs. In some industries, however, such concerns remain very relevant, Das points out. There are extremely sensitive use cases in law, finance, and so on, where trying to deploy AI models as we know them today is just not feasible because you’re relying on a black box to make decisions that you don’t understand why it’s making those decisions, Das says. A good part of [Goodfires] mission is just to be able to do that.


Category: E-Commerce

 

Latest from this category

13.12How Taylor Swift is turning the NFLs mass-media machine into a a pipeline for new male fans
13.12AI advertising slop is on the rise. The cure? The STFU brand strategy
13.12CNBC replaces its peacock with . . . a triangle
13.12The 3 key financial lessons of Its a Wonderful Life
13.1290 housing markets cross critical inventory thresholdtilting power toward buyers
13.12Try these 4 Android battery tips to keep your Google Pixel running longer than ever
13.12Kara Swisher dishes on OpenAI, Meta, Googleand the bidding war for Warner Bros. Discovery
13.12Three hacks to improve your odds of success
E-Commerce »

All news

13.12A new AAA Alien game is reportedly in the works
13.12Half-Life 3 is rumored to be a Steam Machine launch title and could arrive in spring 2026
13.12iOS 26.2 is here with another Liquid Glass tweak, new Podcasts features and more
13.12CRKD's Nitro Deck 2 works for both the Nintendo Switch and Switch 2
13.12Planes, trains, automobiles: RDAs strategic plan sets 20-year goals
13.12Dalal Street Week Ahead: Nifty seen consolidating further before next directional move
13.12Richton Park tree lot that was the setting for a Hallmark Christmas movie shuts down after 40 years
13.12How Taylor Swift is turning the NFLs mass-media machine into a a pipeline for new male fans
More »
Privacy policy . Copyright . Contact form .