Pragmatic idealist. Worked on Ubuntu Phone. Inkscape co-founder. Probably human.
1515 stories
·
12 followers

How I Stopped Worrying and Learned to Love the Bots

1 Share

The decisions by major AI companies—xAI in July, Meta in August, and OpenAI last month—to open their chatbots to erotica have supercharged debate around humans forming romantic relationships with AI. Critics argue that this is the end of human connection.

I founded and run one of the largest romantic chat companies in the world, janitorAI. And yes, I chat with the bots myself—mafiosa Nova Marino is a personal favorite. When I launched the site in 2023, OpenAI sent me a cease-and-desist order because our users were using our platform together with OpenAI’s model to generate romantic content. Our website couldn’t access OpenAI’s application programming interface, and the company disabled some of our users’ OpenAI accounts for violating its terms of service. A few months later, that ban quietly disappeared.

Read the whole story
tedgould
2 hours ago
reply
Texas, USA
Share this story
Delete

Researchers isolate memorization from reasoning in AI neural networks

1 Share

When engineers build AI language models like GPT-5 from training data, at least two major processing features emerge: memorization (reciting exact text they’ve seen before, like famous quotes or passages from books) and reasoning (solving new problems using general principles). New research from AI startup Goodfire.ai provides the first potentially clear evidence that these different functions actually work through completely separate neural pathways in the model’s architecture.

The researchers discovered that this separation proves remarkably clean. In a preprint paper released in late October, they described that when they removed the memorization pathways, models lost 97 percent of their ability to recite training data verbatim but kept nearly all their “logical reasoning” ability intact.

For example, at layer 22 in Allen Institute for AI’s OLMo-7B language model, the bottom 50 percent of weight components showed 23 percent higher activation on memorized data, while the top 10 percent showed 26 percent higher activation on general, non-memorized text. This mechanistic split enabled the researchers to surgically remove memorization while preserving other capabilities.

Perhaps most surprisingly, the researchers found that arithmetic operations seem to share the same neural pathways as memorization rather than logical reasoning. When they removed memorization circuits, mathematical performance plummeted to 66 percent while logical tasks remained nearly untouched. This discovery may explain why AI language models notoriously struggle with math without the use of external tools. They’re attempting to recall arithmetic from a limited memorization table rather than computing it, like a student who memorized times tables but never learned how multiplication works. The finding suggests that at current scales, language models treat “2+2=4” more like a memorized fact than a logical operation.

It’s worth noting that “reasoning” in AI research covers a spectrum of abilities that don’t necessarily match what we might call reasoning in humans. The logical reasoning that survived memory removal in this latest research includes tasks like evaluating true/false statements and following if-then rules, which are essentially applying learned patterns to new inputs. This also differs from the deeper “mathematical reasoning” required for proofs or novel problem-solving, which current AI models struggle with even when their pattern-matching abilities remain intact.

Looking ahead, if the information removal techniques receive further development in the future, AI companies could potentially one day remove, say, copyrighted content, private information, or harmful memorized text from a neural network without destroying the model’s ability to perform transformative tasks. However, since neural networks store information in distributed ways that are still not completely understood, for the time being, the researchers say their method “cannot guarantee complete elimination of sensitive information.” These are early steps in a new research direction for AI.

Traveling the neural landscape

To understand how researchers from Goodfire distinguished memorization from reasoning in these neural networks, it helps to know about a concept in AI called the “loss landscape.” The “loss landscape” is a way of visualizing how wrong or right an AI model’s predictions are as you adjust its internal settings (which are called “weights”).

Imagine you’re tuning a complex machine with millions of dials. The “loss” measures the number of mistakes the machine makes. High loss means many errors, low loss means few errors. The “landscape” is what you’d see if you could map out the error rate for every possible combination of dial settings.

During training, AI models essentially “roll downhill” in this landscape (gradient descent), adjusting their weights to find the valleys where they make the fewest mistakes. This process provides AI model outputs, like answers to questions.

Figure 1: Overview of our approach. We collect activations and gradients from a sample of training data (a), which allows us to approximate loss curvature w.r.t. a weight matrix using K-FAC (b). We decompose these weight matrices into components (each the same size as the matrix), ordered from high to low curvature. In language models, we show that data from different tasks interacts with parts of the spectrum of components differently (c). Figure 1 from the paper “From Memorization to Reasoning in the Spectrum of Loss Curvature.” Credit: Merullo et al.

The researchers analyzed the “curvature” of the loss landscapes of particular AI language models, measuring how sensitive the model’s performance is to small changes in different neural network weights. Sharp peaks and valleys represent high curvature (where tiny changes cause big effects), while flat plains represent low curvature (where changes have minimal impact).

Using a technique called K-FAC (Kronecker-Factored Approximate Curvature), they found that individual memorized facts create sharp spikes in this landscape, but because each memorized item spikes in a different direction, when averaged together they create a flat profile. Meanwhile, reasoning abilities that many different inputs rely on maintain consistent moderate curves across the landscape, like rolling hills that remain roughly the same shape regardless of the direction from which you approach them.

“Directions that implement shared mechanisms used by many inputs add coherently and remain high-curvature on average,” the researchers write, describing reasoning pathways. In contrast, memorization uses “idiosyncratic sharp directions associated with specific examples” that appear flat when averaged across data.

Different tasks reveal a spectrum of mechanisms

The researchers tested their technique on multiple AI systems to verify the findings held across different architectures. They primarily used Allen Institute’s OLMo-2 family of open language models, specifically the 7 billion- and 1 billion-parameter versions, chosen because their training data is openly accessible. For vision models, they trained custom 86 million-parameter Vision Transformers (ViT-Base models) on ImageNet with intentionally mislabeled data to create controlled memorization. They also validated their findings against existing memorization removal methods like BalancedSubnet to establish performance benchmarks.

The team tested their discovery by selectively removing low-curvature weight components from these trained models. Memorized content dropped to 3.4 percent recall from nearly 100 percent. Meanwhile, logical reasoning tasks maintained 95 to 106 percent of baseline performance.

These logical tasks included Boolean expression evaluation, logical deduction puzzles where solvers must track relationships like “if A is taller than B,” object tracking through multiple swaps, and benchmarks like BoolQ for yes/no reasoning, Winogrande for common sense inference, and OpenBookQA for science questions requiring reasoning from provided facts. Some tasks fell between these extremes, revealing a spectrum of mechanisms.

Mathematical operations and closed-book fact retrieval shared pathways with memorization, dropping to 66 to 86 percent performance after editing. The researchers found arithmetic particularly brittle. Even when models generated identical reasoning chains, they failed at the calculation step after low-curvature components were removed.

Figure 3: Sensitivity of different kinds of tasks to ablation of flatter eigenvectors. Parametric knowledge retrieval, arithmetic, and memorization are brittle, but openbook fact retrieval and logical reasoning is robust and maintain around 100% of original performance. Figure 3 from the paper “From Memorization to Reasoning in the Spectrum of Loss Curvature.” Credit: Merullo et al.

“Arithmetic problems themselves are memorized at the 7B scale, or because they require narrowly used directions to do precise calculations,” the team explains. Open-book question answering, which relies on provided context rather than internal knowledge, proved most robust to the editing procedure, maintaining nearly full performance.

Curiously, the mechanism separation varied by information type. Common facts like country capitals barely changed after editing, while rare facts like company CEOs dropped 78 percent. This suggests models allocate distinct neural resources based on how frequently information appears in training.

The K-FAC technique outperformed existing memorization removal methods without needing training examples of memorized content. On unseen historical quotes, K-FAC achieved 16.1 percent memorization versus 60 percent for the previous best method, BalancedSubnet.

Vision transformers showed similar patterns. When trained with intentionally mislabeled images, the models developed distinct pathways for memorizing wrong labels versus learning correct patterns. Removing memorization pathways restored 66.5 percent accuracy on previously mislabeled images.

Limits of memory removal

However, the researchers acknowledged that their technique isn’t perfect. Once-removed memories might return if the model receives more training, as other research has shown that current unlearning methods only suppress information rather than completely erasing it from the neural network’s weights. That means the “forgotten” content can be reactivated with just a few training steps targeting those suppressed areas.

The researchers also can’t fully explain why some abilities, like math, break so easily when memorization is removed. It’s unclear whether the model actually memorized all its arithmetic or whether math just happens to use similar neural circuits as memorization. Additionally, some sophisticated capabilities might look like memorization to their detection method, even when they’re actually complex reasoning patterns. Finally, the mathematical tools they use to measure the model’s “landscape” can become unreliable at the extremes, though this doesn’t affect the actual editing process.

Read full article

Comments



Read the whole story
tedgould
2 hours ago
reply
Texas, USA
Share this story
Delete

“So much more menacing”: Formula E’s new Gen4 car breaks cover

1 Share

Formula E officially revealed its next electric racing car today. At first glance, the Gen4 machine looks similar to machinery of seasons past, but looks are deceiving—it’s “so much more menacing,” according to Formula E CEO Jeff Dodds. The new car is not only longer and wider, it’s far more powerful. The wings and bodywork now generate meaningful aerodynamic downforce. There will be a new tire supplier as Bridgestone returns to single-seat racing. The car is even completely recyclable.

I’m not sure that everyone who attended a Formula E race in its first season would have bet on the sport’s continued existence more than a decade down the line. When the cars took their green flag for the first time in Beijing in 2014, as many people derided it for being too slow or for the mid-race car swaps as praised it for trying something new in the world of motorsport.

Despite that, the racing was mostly entertaining, and it got better with the introduction of the Gen2 car, which made car swapping a thing of the past. Gen3 added more power, then temporary all-wheel drive with the advent of the Gen3 Evo days. That car will continue to race in season 12, which kicks off in Brazil on December 6 and ends in mid-August in London. When season 13 picks up in late 2026, we might see a pretty different kind of Formula E racing.

A Partially disassembled Formula E chassis on a stand The HALO head protection will be more necessary than ever, given the higher speeds of the new car. Credit: Formula E

“It feels like a real moment for us,” said Dodds. The new car will generate 603 hp in race mode, a 50 percent jump compared to the Gen3 Evo. That goes up to 804 hp (600 kW) in attack mode. For context, next year’s F1 cars will generate more power, but only when their batteries are fully charged; if the battery is depleted, that leaves just a 536 hp (400 kW) V6.

Acceleration should be extremely violent thanks to permanent AWD—the first for any single seater in FIA competition, at least for the last few decades. Top speed will be close to double that of the original race car, topping out at 210 mph (337 km/h). Now you can see why the sport decided that aerodynamic grip would be a useful addition.

In fact, there will be two different bodywork configurations, one for high downforce and the other with less. But that doesn’t mean Formula E teams will run out and build wind tunnels, like their F1 counterparts. “There’s significant gains that can be made out of software improvements, efficiency improvements, powertrain developments,” said Dodds, so there’s no incentive to spend lots of money on aero development that would only add fractions of a second.

The biggest opportunity for finding performance improvements may be with traction control and antilock braking systems. Formula E wants its technology to be road-relevant, so such driver aids will be unlimited in the Gen4 era. But efficiency will remain of utmost importance; the cars will still have to regenerate 40 percent of the energy they need to finish the race, as the 55 kWh battery is not sufficient to go flat-out to the end. Happily for the drivers, the new car can regen up to 700 kW of energy under braking.

A grainy shot of a Formula E gen4 car testing The Gen4 car in testing. Credit: Formula E

Finally, the car’s end of life has been considered. The entire race car is entirely recyclable, Formula E says, and it already contains 20 percent recycled content.

So far, the Gen4 car has been put through its paces for more than 5,000 miles (8,000 km), which is more than the mileage of an entire Formula E season, including testing. Now the teams have started to receive their chassis and have started the work of getting to know them and preparing to race them in season 13, all while getting ready to start season 12 next month.

What we won’t know until season 13 gets underway is how the Gen4 era will change the races. With bigger, faster cars, not every Formula E circuit will still be suitable, like London’s very tight Excel Arena track, but with a continued focus on making efficiency count, it’s quite likely we’ll continue to see the same close pack racing as before.

Read full article

Comments



Read the whole story
tedgould
3 days ago
reply
Texas, USA
Share this story
Delete

Would Elon Musk Work Harder for $1 Trillion Than $1 Billion?

1 Share
Economists and psychologists say that compensation may not provide as powerful an incentive as is often assumed.

Read the whole story
tedgould
3 days ago
reply
Texas, USA
Share this story
Delete

Google plans secret AI military outpost on tiny island overrun by crabs

1 Share

On Wednesday, Reuters reported that Google is planning to build a large AI data center on Christmas Island, a 52-square-mile Australian territory in the Indian Ocean, following a cloud computing deal with Australia’s military. The previously undisclosed project will reportedly position advanced AI infrastructure a mere 220 miles south of Indonesia at a location military strategists consider critical for monitoring Chinese naval activity.

Aside from its strategic military position, the island is famous for its massive annual crab migration, where over 100 million of red crabs make their way across the island to spawn in the ocean. That’s notable because the tech giant has applied for environmental approvals to build a subsea cable connecting the 135-square-kilometer island to Darwin, where US Marines are stationed for six months each year.

The project follows a three-year cloud agreement Google signed with Australia’s military in July 2025, but many details about the new facility’s size, cost, and specific capabilities remain “secret,” according to Reuters. Both Google and Australia’s Department of Defense declined to comment when contacted by the news agency.

Sir David Attenborough examines the great Christmas Island red crab migration.

Bryan Clark, a former US Navy strategist who ran recent war games featuring Christmas Island, told Reuters that the planned facility would enable AI-powered military command and control. Recent military exercises involving Australian, US, and Japanese forces show Christmas Island’s value as a forward defense position for launching uncrewed weapons systems. The island’s location allows the monitoring of traffic through the Sunda, Lombok, and Malacca straits, which are key waterways for global shipping and submarine movements.

Christmas Island has reportedly struggled with poor telecommunications and limited economic opportunities in the past, but some of the island’s 1,600 human residents are cautiously optimistic about the project.

Christmas Island Shire President Steve Pereira told Reuters that the council is examining community impacts before approving construction. “There is support for it, providing this data center actually does put back into the community with infrastructure, employment, and adding economic value to the island,” Pereira said.

That’s great, but what about the crabs?

Christmas Island’s annual crab migration is a natural phenomenon that Sir David Attenborough reportedly once described as one of his greatest TV moments when he visited the site in 1990.

Every year, millions of crabs emerge from the forest and swarm across roads, streams, rocks, and beaches to reach the ocean, where each female can produce up to 100,000 eggs. The tiny baby crabs that survive take about nine days to march back inland to the safety of the plateau.

While Google is seeking environmental approvals for its subsea cables, the timing could prove delicate for Christmas Island’s most famous residents. According to Parks Australia, the island’s annual red crab migration has already begun for 2025, with a major spawning event expected in just a few weeks, around November 15–16.

During peak migration times, sections of roads close at short notice as crabs move between forest and sea, and the island has built special crab bridges over roads to protect the migrating masses.

Parks Australia notes that while the migration happens annually, few baby crabs survive the journey from sea to forest most years, as they’re often eaten by fish, manta rays, and whale sharks. The successful migrations that occur only once or twice per decade (when large numbers of babies actually survive) are critical for maintaining the island’s red crab population.

How Google’s facility might coexist with 100 million marching crustaceans remains to be seen. But judging by the size of the event, it seems clear that it’s the crab’s world, and we’re just living in it.

Read full article

Comments



Read the whole story
tedgould
4 days ago
reply
Texas, USA
Share this story
Delete

Nick Fuentes’s Rise Puts MAGA Movement in a ‘Time of Choosing’

1 Share
After Mr. Fuentes’s interview with Tucker Carlson, Republicans are considering just how far his views are from the nationalism embraced by President Trump’s followers.

Read the whole story
tedgould
5 days ago
reply
Texas, USA
Share this story
Delete
Next Page of Stories