Topics

Latest

AI

Amazon

Article image

Image Credits:Flow Computing

Apps

Biotech & Health

Climate

Article image

Image Credits:Flow Computing

Cloud Computing

Commerce

Crypto

Article image

Chart (in log, note) showing improvements in an FPGA PPU-enhanced chip versus unmodified Intel chips. Increasing the number of PPU cores continually improves performance.Image Credits:Flow Computing

Enterprise

EVs

Fintech

Article image

Flow’s founding team, from left: Jussi Roivainen, Martti Forsell and Timo Valtonen.Image Credits:Flow Computing

Fundraising

Gadgets

bet on

Google

Government & Policy

Hardware

Instagram

layoff

Media & Entertainment

Meta

Microsoft

secrecy

Robotics

security department

Social

quad

startup

TikTok

Transportation

speculation

More from TechCrunch

upshot

Startup Battlefield

StrictlyVC

Podcasts

Videos

Partner Content

TechCrunch Brand Studio

Crunchboard

Contact Us

A Finnish startup calledFlow Computingis take a shit one of the wildest claim ever heard in Si applied science : by impart its proprietary companion chip , any central processor can instantly double its performance , increasing to as much as 100x with software tweaks .

If it crop , it could help the manufacture keep up with the unsatiable compute demand of AI makers .

Flow is a spinout ofVTT , a Finland country - endorse research formation that ’s a bite like a national lab . The chip engineering it ’s commercializing , which it has branded the Parallel Processing Unit , is the consequence of research performed at that research laboratory ( though VTT is an investor , the IP is owned by Flow ) .

The claim , Flow is first to admit , is laughable on its face . You ca n’t just magically squelch superfluous performance out of CPUs across architectures and codification radical . If so , Intel or AMD or whoever would have done it days ago .

But Flow has been play on something thathasbeen theoretically possible — it ’s just that no one has been capable to deplumate it off .

Central Processing Units have come a long way of life since the early 24-hour interval of vacuum tube-shaped structure and biff cards , but in some fundamental shipway they ’re still the same . Their primary restriction is that as in series rather than parallel mainframe , they can only do one thing at a prison term . Of course , they switch that affair a billion times a second across multiple cores and pathways — but these are all fashion of accommodating the individual - lane nature of the CPU . ( A GPU , in demarcation , does many related calculations at once but is specialise in sure operation . )

“ The C.P.U. is the weakest liaison in computing , ” say Flow Colorado - laminitis and CEO Timo Valtonen . “ It ’s not up to its project , and this will require to interchange . ”

Join us at TechCrunch Sessions: AI

Exhibit at TechCrunch Sessions: AI

central processing unit have gotten very fast , but even with nanosecond - level responsiveness , there ’s a tremendous amount of waste in how instructions are stock out simply because of the basic limitation that one task want to eat up before the next one starts . ( I ’m simplify here , not being a chip engineer myself . )

What Flow claims to have done is remove this limit , turn the mainframe from a one - lane street into a multi - lane highway . The processor is still limited to doing one task at a sentence , but Flow ’s PPU , as they call it , essentially performs nanosecond - scale dealings management on - dice to move chore into and out of the processor faster than has previously been possible .

mean of the CPU as a chef make for in a kitchen . The chef can only work so fast , but what if that someone had a superhuman assistant swapping knives and tools in and out of the chef ’s hands , clearing the prepared food and putting in fresh element , get rid of all task that are n’t literal chef stuff ? The chef still only has two hand , but now the chef can work 10 prison term as fast .

It ’s not a pure analogy , but it gives you an idea of what ’s befall here , at least according to Flow ’s internal tests and demos with the industry ( and they are talking with everyone ) . The PPU does n’t increase the clock oftenness or crusade the system in other way that would lead to extra high temperature or power ; in other Son , the chef is not being asked to chop up twice as fast . It just more efficiently uses the CPU cycles that are already taking berth .

This eccentric of affair is n’t steel new , allege Valtonen . “ This has been analyse and talk about in eminent - horizontal surface academe . you could already do parallelization , but it breaks legacy codification , and then it ’s useless . ”

So it could be done . It just could n’t be done without rewrite all the code in the world from the ground up , which kind of make it a non - dispatcher . A similar problem was solved byanother Nordic compute company , ZeroPoint , which achieved high levels of memory compression while keeping datum transparentness with the rest of the system .

Flow ’s big achievement , in other wrangle , is n’t gamy - speed traffic management , but rather doing it without having to modify any code on any CPU or architecture that it has test . It sounds kind of disturbed to say that arbitrary code can be executed double as fast on any microchip with no adjustment beyond integrate the PPU with the die .

Therein lies the chief challenge to menstruate ’s succeeder as a business : Unlike a software product , Flow ’s technical school penury to be include at the chip - design level , meaning it does n’t mould retroactively , and the first potato chip with a PPU would necessarily be quite a way down the road . Flow has shown that the technical school works in FPGA - based trial setup , but chipmakers would have to commit quite a spate of resources to see the addition in question .

The scale of those gain , and the fact that central processing unit improvement have been reiterative and fractional over the last few year , may well have those chipmakers knocking on Flow ’s threshold rather desperately , though . If you may really repeat your performance in one generation with one layout change , that ’s a no - brainer .

Further performance gains derive from refactoring and recompiling software to work good with the PPU - CPU combo . Flow enjoin it has see gain up to 100x with code that ’s been modified ( though not necessarily to the full rewrite ) to take advantage of its technology . The companionship is working on offering recompilation instrument to make this undertaking bare for software package Godhead who want to optimize for Flow - enabled chips .

Analyst Kevin Krewell fromTirias Research , who was brief on Flow ’s tech and referred to as an outdoor linear perspective on these matter , was more worried about industry uptake than the fundamentals .

He pointed out , quite rightly , that AI quickening is the biggest market the right way now , something that can be targeted for with special atomic number 14 like Nvidia ’s popular H100 . Though a PPU - accelerated CPU would lead to gains across the plank , chipmakers might not want to rock the sauceboat too heavily . And there ’s plainly the question of whether those company are willing to invest significant resourcefulness into a for the most part unproved technology when they likely have a five - year program that would be upset by that option .

Will Flow ’s technical school become a must - have component for every chipmaker out there , catapult it to fortune and prominence ? Or will penny - cabbage chipmakers resolve to quell the grade and keep extracting rent from the steady grow compute market ? Probably somewhere in between — but it is telling that , even if Flow has achieve a major engineering feat here , like all startup , the hereafter of the company depends on its customers .

Flow is just now emerging from stealth , with € 4 million ( about $ 4.3 million ) in pre - seed funding led by Butterfly Ventures , with participation from FOV Ventures , Sarsia , Stephen Industries , Superhero Capital and Business Finland .