Kazu Gomi has a giant view of the know-how world from his perch in Silicon Valley. And as president and CEO of NTT Research, a division of the large Japanese telecommunications agency NTT, Gomi can management the R&D price range for a large chunk of the essential analysis that’s accomplished in Silicon Valley.
And maybe it’s no shock that Gomi is pouring some huge cash into AI for the enterprise to find new alternatives to benefit from the AI explosion. Final week, Gomi unveiled a brand new analysis effort to deal with the physics of AI and nicely as a chip design for an AI inference chip that may course of 4K video quicker. This comes on the heels of analysis tasks introduced final yr that might pave the way in which for higher AI and extra vitality environment friendly information facilities.
I spoke with Gomi about this effort within the context of different issues huge firms like Nvidia are doing. Bodily AI has turn into a giant deal in 2025, with Nvidia main the cost to create artificial information to pretest self-driving automobiles and humanoid robotics to allow them to get to market quicker.
And constructing on a narrative that I first did in my first tech reporting job, Gomi mentioned the corporate is doing analysis on photonic computing as a method to make AI computing much more vitality environment friendly.

A long time in the past, I toured Bell Labs and listened to the ambitions of Alan Huang as he sought to make an optical pc. Gomi’s workforce is making an attempt to do one thing related a long time later. If they’ll pull it off, it may make information facilities function on so much much less energy, as gentle doesn’t collide with different particles or generate friction the way in which {that electrical} alerts do.
In the course of the occasion final week, I loved speaking to slightly desk robotic known as Jibo that swiveled and “danced” and instructed me my very important indicators, like my coronary heart price, blood oxygen stage, blood stress, and even my ldl cholesterol — all by scanning my pores and skin to see the tiny palpitations and colour change because the blood moved by means of my cheeks. It additionally held a dialog with me by way of its AI chat functionality.
NTT has greater than 330,000 staff and $97 billion in annual income. NTT Analysis is a part of NTT, a worldwide know-how and enterprise options supplier with an annual R&D price range of $3.6 billion. About six years in the past it created an R&D division in Silicon Valley.
Right here’s an edited transcript of our interview.

VentureBeat: Do you’re feeling like there’s a theme, a prevailing theme this yr for what you’re speaking about in comparison with final yr?
Kazu Gomi: There’s no secret. We’re extra AI-heavy. AI is entrance and middle. We talked about AI final yr as nicely, nevertheless it’s extra vivid at present.
VentureBeat: I wished to listen to your opinion on what I absorbed out of CES, when Jensen Huang gave his keynote speech. He talked so much about artificial information and the way this was going to speed up bodily AI. As a result of you may check your self-driving automobiles with artificial information, or check humanoid robots, a lot extra testing will be accomplished reliably within the digital area. They get to market a lot quicker. Do you’re feeling like this is smart, that artificial information can result in this acceleration?
Gomi: For the robots, sure, 100%. The robots and all of the bodily issues, it makes a ton of sense. AI is influencing so many different issues as nicely. In all probability not all the pieces. Artificial information can’t change all the pieces. However AI is impacting the way in which firms run themselves. The authorized division is likely to be changed by AI. The HR division is changed by AI. These sorts of issues. In these situations, I’m unsure how artificial information makes a distinction. It’s not making as huge an affect as it might for issues like self-driving automobiles.
VentureBeat: It made me assume that issues are going to return so quick, issues like humanoid robots and self-driving automobiles, that we now have to resolve whether or not we actually need them, and what we would like them for.
Gomi: That’s a giant query. How do you cope with them? We’ve positively began speaking about it. How do you’re employed with them?

VentureBeat: How do you employ them to enhance human staff, but additionally–I feel certainly one of your individuals talked about elevating the usual of residing [for humans, not for robots].
Gomi: Proper. If you happen to do it proper, completely. There are lots of good methods to work with them. There are definitely unhealthy situations which can be attainable as nicely.
VentureBeat: If we noticed this a lot acceleration within the final yr or so, and we are able to count on artificial information will speed up it much more, what do you count on to occur two years from now?
Gomi: Not a lot on the artificial information per se, however at present, one of many press releases my workforce launched is about our new analysis group, known as Physics of AI. I’m trying ahead to the outcomes coming from this workforce, in so many alternative methods. One of many fascinating ones is that–this humanoid factor comes close to to it. However proper now we don’t know–we take AI as a black field. We don’t know precisely what’s happening contained in the field. That’s an issue. This workforce is trying contained in the black field.
There are lots of potential advantages, however one of many intuitive ones is that if AI begins saying one thing incorrect, one thing biased, clearly it’s worthwhile to make corrections. Proper now we don’t have an excellent, efficient method to appropriate it, besides to only preserve saying, “That is incorrect, you need to say this as an alternative of that.” There’s analysis saying that information alone gained’t save us.
VentureBeat: Does it really feel such as you’re making an attempt to show a child one thing?
Gomi: Yeah, precisely. The fascinating superb situation–with this Physics of AI, successfully what we are able to do, there’s a mapping of information. In the long run AI is a pc program. It’s made up of neural connections, billions of neurons related collectively. If there’s bias, it’s coming from a specific connection between neurons. If we are able to discover that, we are able to in the end scale back bias by slicing these connections. That’s the best-case situation. Everyone knows that issues aren’t that simple. However the workforce could possibly inform that when you reduce these neurons, you would possibly be capable to scale back bias 80% of the time, or 60%. I hope that this workforce can attain one thing like that. Even 10% continues to be good.
VentureBeat: There was the AI inference chip. Are you making an attempt to outdo Nvidia? It looks like that will be very arduous to do.

Gomi: With that specific mission, no, that’s not what we’re doing. And sure, it’s very arduous to do. Evaluating that chip to Nvidia, it’s apples and oranges. Nvidia’s GPU is extra of a general-purpose AI chip. It could actually energy chat bots or autonomous automobiles. You are able to do all types of AI with it. This one which we launched yesterday is barely good for video and pictures, object detection and so forth. You’re not going to create a chat bot with it.
VentureBeat: Did it appear to be there was a chance to go after? Was one thing not likely working in that space?
Gomi: The quick reply is sure. Once more, this chip is unquestionably personalized for video and picture processing. The secret is that with out decreasing the decision of the bottom picture, we are able to do inference. Excessive decision, 4K photos, you should use that for inference. The profit is that–take the case of a surveillance digital camera. Possibly it’s 500 meters away from the thing you wish to take a look at. With 4K video you may see that object fairly nicely. However with typical know-how, due to processing energy, you must scale back the decision. Possibly you could possibly inform this was a bottle, however you couldn’t learn something on it. Possibly you could possibly zoom in, however then you definately lose different data from the realm round it. You are able to do extra with that surveillance digital camera utilizing this know-how. Greater decision is the profit.

VentureBeat: This is likely to be unrelated, however I used to be occupied with Nvidia’s graphics chips, the place they had been utilizing DLSS, utilizing AI to foretell the subsequent pixel it’s worthwhile to draw. That prediction works so nicely that it received eight occasions quicker on this technology. The general efficiency is now one thing like–out of 30 frames, AI would possibly precisely predict 29 of them. Are you doing one thing related right here?
Gomi: One thing associated to that–the explanation we’re engaged on this, we had a mission that’s the precursor to this know-how. We spent a variety of vitality and assets prior to now on video codec applied sciences. We bought an early MPEG decoder for professionals, for TV station-grade cameras and issues like that. We had that base know-how. Inside this base know-how, one thing just like what you’re speaking about–there’s a little bit of object recognition happening within the present MPEG. Between the frames, it predicts that an object is transferring from one body to the subsequent by a lot. That’s a part of the codec know-how. Object recognition makes that occur, these predictions. That algorithm, to some extent, is used on this inference chip.
VentureBeat: One thing else Jensen was saying that was fascinating–we had an structure for computing, retrieval-based computing, the place you go right into a database, fetch a solution, and are available again. Whereas with AI we now have the chance for reason-based computing. AI figures out the reply with out having to look by means of all this information. It could actually say, “I do know what the reply is,” as an alternative of retrieving the reply. It could possibly be a special type of computing than what we’re used to. Do you assume that will probably be a giant change?
Gomi: I feel so. Plenty of AI analysis is occurring. What you mentioned is feasible as a result of AI has “information.” As a result of you might have that information, you don’t must go retrieve information.

VentureBeat: As a result of I do know one thing, I don’t must go to the library and look it up in a guide.
Gomi: Precisely. I do know that such and such occasion occurred in 1868, as a result of I memorized that. You would look it up in a guide or a database, but when you realize that, you might have that information. It’s an fascinating a part of AI. Because it turns into extra clever and acquires extra information, it doesn’t have to return to the database every time.
VentureBeat: Do you might have any specific favourite tasks happening proper now?
Gomi: A pair. One factor I wish to spotlight, maybe, if I may choose one–you’re trying intently at Nvidia and people gamers. We’re placing a variety of deal with photonics know-how. We’re occupied with photonics in a few alternative ways. If you take a look at AI infrastructure–you realize all of the tales. We’ve created so many GPU clusters. They’re all interconnected. The platform is big. It requires a lot vitality. We’re operating out of electrical energy. We’re overheating the planet. This isn’t good.
We wish to tackle this difficulty with some totally different methods. Considered one of them is utilizing photonics know-how. There are a few alternative ways. First off, the place is the bottleneck within the present AI platform? In the course of the panel at present, one of many panelists talked about this. If you take a look at GPUs, on common, 50% of the time a GPU is idle. There’s a lot information transport taking place between processors and reminiscence. The reminiscence and that communication line is a bottleneck. The GPU is ready for the info to be fetched and ready to put in writing outcomes to reminiscence. This occurs so many occasions.
One thought is utilizing optics to make these communication strains a lot quicker. That’s one factor. Through the use of optics, making it quicker is one profit. One other profit is that in relation to quicker clock speeds, optics is rather more energy-efficient. Third, this entails a variety of engineering element, however with optics you may go additional. You may go this far, and even a few ft away. Rack configuration generally is a lot extra versatile and fewer dense. The cooling necessities are eased.
VentureBeat: Proper now you’re extra like information middle to information middle. Right here, are we speaking about processor to reminiscence?

Gomi: Yeah, precisely. That is the evolution. Proper now it’s between information facilities. The subsequent part is between the racks, between the servers. After that’s inside the server, between the boards. After which inside the board, between the chips. Ultimately inside the chip, between a few totally different processing items within the core, the reminiscence cache. That’s the evolution. Nvidia has additionally launched some packaging that’s alongside the strains of this phased method.
VentureBeat: I began protecting know-how round 1988, out in Dallas. I went to go to Bell Labs. On the time they had been doing photonic computing analysis. They made a variety of progress, nevertheless it’s nonetheless not fairly right here, even now. It’s spanned my entire profession protecting know-how. What’s the problem, or the issue?
Gomi: The situation I simply talked about hasn’t touched the processing unit itself, or the reminiscence itself. Solely the connection between the 2 parts, making that quicker. Clearly the subsequent step is we now have to do one thing with the processing unit and the reminiscence itself.
VentureBeat: Extra like an optical pc?
Gomi: Sure, an actual optical pc. We’re making an attempt to do this. The factor is–it sounds such as you’ve adopted this matter for some time. However right here’s a little bit of the evolution, so to talk. Again within the day, when Bell Labs or whoever tried to create an optical-based pc, it was mainly changing the silicon-based pc one to at least one, precisely. All of the logic circuits and all the pieces would run on optics. That’s arduous, and it continues to be arduous. I don’t assume we are able to get there. Silicon photonics gained’t tackle the problem both.
The fascinating piece is, once more, AI. For AI you don’t want very fancy computations. AI computation, the core of it’s comparatively easy. Every part is a factor known as matrix-vector multiplication. Data is available in, there’s a end result, and it comes out. That’s all you do. However you must do it a billion occasions. That’s why it will get difficult and requires a variety of vitality and so forth. Now, the great thing about photonics is that it may possibly do that matrix-vector multiplication by its nature.
VentureBeat: Does it contain a variety of mirrors and redirection?

Gomi: Yeah, mirroring after which interference and all that stuff. To make it occur extra effectively and all the pieces–in my researchers’ opinion, silicon photonics could possibly do it, nevertheless it’s arduous. You need to contain totally different supplies. That’s one thing we’re engaged on. I don’t know when you’ve heard of this, nevertheless it’s lithium niobate. We use lithium niobate as an alternative of silicon. There’s a know-how to make it into a skinny movie. You are able to do these computations and multiplications on the chip. It doesn’t require any digital parts. It’s just about all accomplished by analog. It’s tremendous quick, tremendous energy-efficient. To some extent it mimics what’s happening contained in the human mind.
These {hardware} researchers, their purpose–a human mind works with possibly round 20 watts. ChatGPT requires 30 or 40 megawatts. We will use photonics know-how to have the ability to drastically upend the present AI infrastructure, if we are able to get all the way in which there to an optical pc.
VentureBeat: How are you doing with the digital twin of the human coronary heart?
Gomi: We’ve made fairly good progress over the past yr. We created a system known as the autonomous closed-loop intervention system, ACIS. Assume you might have a affected person with coronary heart failure. With this method utilized–it’s like autonomous driving. Theoretically, with out human intervention, you may prescribe the precise medication and remedy to this coronary heart and convey it again to a standard state. It sounds a bit fanciful, however there’s a bio-digital twin behind it. The bio-digital twin can exactly predict the state of the center and what an injection of a given drug would possibly do to it. It could actually rapidly predict trigger and impact, resolve on a remedy, and transfer ahead. Simulation-wise, the system works. We’ve some good proof that it’s going to work.

VentureBeat: Jibo, the robotic within the well being sales space, how shut is that to being correct? I feel it received my ldl cholesterol incorrect, nevertheless it received all the pieces else proper. Ldl cholesterol appears to be a tough one. They had been saying that was a brand new a part of what they had been doing, whereas all the pieces else was extra established. If you may get that to excessive accuracy, it could possibly be transformative for the way typically individuals must see a health care provider.
Gomi: I don’t know an excessive amount of about that specific topic. The standard means of testing that, after all, they’ve to attract blood and analyze it. I’m positive somebody is engaged on it. It’s a matter of what sort of sensor you may create. With non-invasive units we are able to already learn issues like glucose ranges. That’s fascinating know-how. If somebody did it for one thing like ldl cholesterol, we may convey it into Jibo and go from there.
