We just sent you a verification email. Please verify your account to gain access to
theCUBE + NYSE Wired: Mixture of Experts Series. If you don’t think you received an email check your
spam folder.
Sign in to theCUBE + NYSE Wired: Mixture of Experts Series.
In order to sign in, enter the email address you used to registered for the event. Once completed, you will receive an email with a verification link. Open this link to automatically sign into the site.
Register For theCUBE + NYSE Wired: Mixture of Experts Series
Please fill out the information below. You will recieve an email with a verification link confirming your registration. Click the link to automatically sign into the site.
You’re almost there!
We just sent you a verification email. Please click the verification button in the email. Once your email address is verified, you will have full access to all event content for theCUBE + NYSE Wired: Mixture of Experts Series.
I want my badge and interests to be visible to all attendees.
Checking this box will display your presense on the attendees list, view your profile and allow other attendees to contact you via 1-1 chat. Read the Privacy Policy. At any time, you can choose to disable this preference.
Select your Interests!
add
Upload your photo
Uploading..
OR
Connect via Twitter
Connect via Linkedin
EDIT PASSWORD
Share
Forgot Password
Almost there!
We just sent you a verification email. Please verify your account to gain access to
theCUBE + NYSE Wired: Mixture of Experts Series. If you don’t think you received an email check your
spam folder.
Sign in to theCUBE + NYSE Wired: Mixture of Experts Series.
In order to sign in, enter the email address you used to registered for the event. Once completed, you will receive an email with a verification link. Open this link to automatically sign into the site.
Sign in to gain access to theCUBE + NYSE Wired: Mixture of Experts Series
Please sign in with LinkedIn to continue to theCUBE + NYSE Wired: Mixture of Experts Series. Signing in with LinkedIn ensures a professional environment.
Are you sure you want to remove access rights for this user?
Details
Manage Access
email address
Community Invitation
Erik Bernhardsson, Modal
In this theCUBE + NYSE Wired: Mixture of Experts segment, theCUBE’s Dave Vellante sits down with Jim McNiel, Chief Growth Officer at TAE Technologies, to demystify fusion vs. fission and explore how proton–boron fusion could reshape energy economics for enterprise and Wall Street alike. McNiel explains why TAE targets abundant, low-cost boron fuel and how its approach avoids long-lived radioactive waste, requires only light shielding and eliminates meltdown risk. He breaks down siting and regulation – fusion treated more like medical isotopes than fission – and outlines first-gen levelized energy costs in the 7–9¢ range with a path to sub-5¢ as the technology matures. The conversation ties these fundamentals to market dynamics: dispatchable, carbon-free baseload power for data centers, safer urban siting and a financing narrative that aligns with investor expectations and hyperscaler demand.
Listeners also get a clear milestone roadmap: Copernicus (commissioned to operate in 2028) targeting net energy out; Da Vinci as a 50-MW commercial prototype; and TAE Fusion 1 designed for 350 MW—scalable units that could colocate with gigawatt-scale AI facilities. McNiel details how AI already governs plasma stability via TAE’s “Optometrist Algorithm” developed with Google and notes strategic investors (e.g., Chevron, Sumitomo) plus near-term revenue from TAE Power Solutions and TAE Life Sciences. The discussion frames emerging trends in enterprise strategy – from energy as a core input to AI-driven productivity gains – and why the go-to-market has shifted from utility-first to hyperscaler-led demand for dispatchable, clean power.
>> Hey everybody, welcome back to the New York Stock Exchange. We're here in the Buttonwood podium overlooking the options exchange. My name is Dave Vellante and you're watching the NYSE Wired and theCUBE's Mixture of Experts series. Erik Bernhardsson is here. He's the CEO and co-founder of Modal, a company that does serverless compute and really cloud innovator, startup. Super psych to have you here, Erik.
Erik Bernhardsson
>> Yeah, it's great to be here. Thanks a lot for having me.
Dave Vellante
>> You were very welcome. Why did you start the company? What's Modal all about?
Erik Bernhardsson
>> Yeah, it's a long story, but I think the shorter version is I have built a lot of AI products in my career so far. In particular, I was at Spotify for many years and built the music recommendation system and realized at some point that a lot of traditional infrastructure just isn't built for AI and for a lot of these newer types of applications. And with infrastructure, I mean this sort of software layer sort of sitting under the applications, you have the applications and this is what the machine learning engineers are building, and these could be things like training models, or deploying models, or building vibe coding platforms, or computational biotech, or other things. They only need infrastructure, right? So that infrastructure layer today is often built for very different types of applications. So we realized at some point we should just build a whole new layer that is more geared towards these types of new applications and rethink a lot of the assumptions and focus on the needs of these new applications in terms of scaling up and down, or working with large models, working with GPU capacity all over the world, et cetera. So that's why we started Modal.
Dave Vellante
>> So interesting. I mean, I never studied Spotify deeply, although as a consumer and a user of the service, I always felt like there was some really smart people behind the scenes. You were obviously one of them building what I refer to as data products. I mean, essentially, you had data sets and you would serve up services to consumers like myself that were really well-thought-out and the infrastructure underneath was something that you probably didn't get off the shelf as you had to build this, right? Is that what you're implying? And then now what you're doing is bringing that to the masses? Is that-
Erik Bernhardsson
>> No, that's exactly right. So a big part of what Spotify focused on was, how do we make it easy for people to know what to play? You have 10 million tracks, where do you start? And so my team specifically focused a lot on music recommendations and built a lot of features around that. And my experience building those products was because you didn't have the right infrastructure, you ended up spending way more time building the infrastructure than actually spending time building the applications. Then I realized, "Someone should just build this layer." And I'm sure Spotify isn't the only company, so I'm sure there's 10,000 other companies out there who also need this infrastructure layer. And that's what turned out to be quite true. We realized so far there's a lot of companies out there who need better infrastructure, especially now that we're switching to GPUs and very large models and LMs and all these things. You especially need more infrastructure.
Dave Vellante
>> So there are a lot of recommendation engines out there and a lot of really crappy ones. You think about retail, a lot of times, you would go these kind of conditioned response type of experiences, which were terrible. You got four choices. Which one are you? None of the above. Spotify obviously, and TikTok obviously have very strong, whether it's algorithms or whether it's infrastructure. But now, we're entering sort of a new era beyond things like recommendation engines where we're doing so many more interesting things with AI. What's your vision in terms of supporting some of those new applications and capabilities?
Erik Bernhardsson
>> Yeah. I mean, first of all, it's all switching to GPUs. We started playing around with GPUs in 2014 at Spotify, which was quite early, but most of this stuff wasn't running on GPUs, and I don't know if it's still running. I left a while ago, so that's a big change. And then the models have gotten very big, so that's another big change is we're working with these very large LLMs these days. There's obviously generated media, which kind of completely changes the game. So you have algorithms that can generate music these days, generate video, generate images. I mean, all these things are just complete game changer for a lot of companies. And we're starting to see that. We're starting to see a lot of companies adopting these things. Initially, we saw a lot of early stage companies adopting these things, but what we see today is a lot of demand is also coming from later stage, sometimes public companies to try to deploy these things, and that's when they often come to us, because they need help with that.
Dave Vellante
>> So interesting. So 2014, you had experience with GPUs. So I think that kind of brings you into the CUDA era and the programming of those GPUs. And again, I come back to this notion of mainstream. So our vision is, we wrote a piece a while ago, why Jamie Dimon is Sam Altman's biggest competitor. The thinking there was there's all this wonderful proprietary data sitting inside of places like JPMC. We use that as a metaphor for enterprises, and we've also written about Uber for all. Uber had to build its own infrastructure like Spotify had to build its own infrastructure. So Uber for all is, again, a metaphor for solutions or capabilities that allow enterprises to actually build this stuff much more cheaply and faster without having to hire thousands of engineers. Is that what you're bringing to the table? And what other piece parts do you see emerging? Or is it part of your TAM expansion strategy to build those other piece parts?
Erik Bernhardsson
>> Yeah, I mean, Modal focuses specifically on these machine learning problems. So our consumers are people who build the models, but I think there's a larger trend here, which you're talking about, which is I almost think of every company becoming a tech company, and also the tech company's becoming just normal companies. Spotify, for instance, today, are they a music company, are they a tech company? I don't know. It almost doesn't matter, right? They started out being very much a tech company, but the truth is thousands of other companies had similar problems like music distribution and working with large data sets, and those used to be unique things that Spotify had to solve in an idiosyncratic ways in their own ways. But today, I think those are just product you can buy from a lot of vendors.
And so I think there's an emergence of so many different infrastructure vendors and so many different tools, and I think there's also ... You look at the existing companies that previously weren't tech companies, they're buying all these vendors. So I think it's a normalization of tech, and I think you're going to see the same thing with AI, who's an AI company, who's not an AI company. I don't know. Everyone is an AI company. Everyone is going to need a lot of these things. Get like traditional companies, like you mentioned JP Morgan, how many engineers do they have? 100,000. Absurd number. They're going to need a lot. They're going to have to also buy a lot of this infrastructure. And so there's a role for many different parts of that stack, and we play into one of the lower layers. We focus a lot on running code, but there's many opportunities also higher up in the stack simplifying how to work with AI. And maybe you don't even have to think about the models and maybe you can just point and click. We're not a point of click system, we're not a low-code or no-code system. We're very much a high-code system. We focus on the software engineers, but it's all part of the same trend, I think, of the marketizing or broadening the footprint of technology and AI and the cloud and maybe other things.
Dave Vellante
>> But your basic premise is you shouldn't have to have a PHG and AI infrastructure in order to deploy AI. You're making that simpler, even though it's full stack.
Erik Bernhardsson
>> Yeah. No, I think that's right.
Dave Vellante
>> You're targeting, so what is broken about today's model generally and maybe the cloud model specifically?
Erik Bernhardsson
>> Yeah, yeah. And by the way, I will say that PhDs are actually our best customers, even if they are among the few people who understand the course of this stuff, they're also tired of bad infrastructure. So we sell to the whole spectrum of machine learning PhDs all the way to just any sort of developer.
Dave Vellante
>> But the point is you don't want them wasting their time building infrastructure.
Erik Bernhardsson
>> Exactly, exactly. They should focus on what they're good at, which is to build amazing models. Right? So going back to your question, what's broken with today's infrastructure? A lot of today's infrastructure is built for very different applications, built for backend applications. So it means you have kind of smooth workloads. They're very IO-bound in the sense that they often just coordinate work between many other systems like maybe databases, et cetera. What I'm talking about are things just to drop some names here like Kubernetes or Docker, et cetera. And that's been the dominant development paradigm and the dominant infrastructure for the last 10 years or so. But when you start to look at a lot of these new applications, you run into a lot of different problems. First of all, you have a lot of GPU capacity, or actually I should say GPU scarcity. So you have different regions all over the world with different capacity, and often when you start to scale up, you run into constraints. Maybe you run out of GPUs in one region, so you have to go to a different region, or you need to run things close to you so you have certain latency constraints because speed of light is a real thing for a lot of real-time applications. Another thing is these models are very big. So building applications that have to load up these models, load them into the GPU and do that fast and scale up reactively to user workloads, you need to really think deeply about how do you build infrastructure that can do that, especially when you look at inference, and we're very focused on inference historically. We've had a lot of use cases in inference. And when you're doing inference, fundamentally, you don't know what the user demand is. The user demand goes up and down, and so scaling up and down with these workloads is a real challenge. And so we basically decided at some point in order to support a lot of those things, we had to build basically our own stack. We decided we can't build a new stack on top of an old stack, so let's throw out the old stack, let's build our own file system, let's build our own container runtime, let's build our own scheduler and many other things. And we had to do that in order to be able to support these new types of applications. The other thing I think is very broken also is developer experience. So it wasn't as much of a problem in the old paradigm, but when you're starting to look at these AI researchers and how they're operating and how they're trying to leverage the old infrastructure, they're only into a lot of challenges. Just being able to take models and get it out in production and putting it in front of users, it's a real challenge for these developers. So we spent a lot of time super obsessed with like, how do we make it fun for engineers? How do we make them fast? How do we really improve the speed to market? Because I think that's almost as important as or more important maybe than everything else. There are so many companies out there where speed to market is really the thing that matters. So by making engineers 10x more productive, you're also gaining a massive advantage.
Dave Vellante
>> So let me ask you, but follow up on that. So if I think about the cloud today, I've got access to all these primitives and services through APIs.
Erik Bernhardsson
>> That's right. Yep.
Dave Vellante
>> I got Elastic Beanstalk, I've got Kubernetes, and you're saying this is just too complicated today. It requires too much brute force to actually scale, and you're orchestrating that in a much simpler way and streamlining that experience.
Erik Bernhardsson
>> Yeah, that's exactly right. So first of all, we're not trying to compete with the cloud vendors. We are customers of the cloud vendors.
Dave Vellante
>> Yeah, you're running on the cloud.
Erik Bernhardsson
>> Yeah, we're running on the cloud and we're super happy. We consider them partners. We use a lot of different cloud vendors, AWS, GCP, et cetera, Oracle and Azure and many others. They're very good at delivering physical data, like running physical data centers and delivering compute capacity through an API.
Dave Vellante
>> You need compute networking there.
Erik Bernhardsson
>> Exactly. The lower stuff, right?
Dave Vellante
>> Yeah.
Erik Bernhardsson
>> What I think there's a gap is slightly higher up in the stack, and that's really most visible in developer experience, like working and building these AI applications. You can build them straight on the clouds, but it's pretty hard. And even if it's hard, I mean I know how to do it. It's also annoying. It takes a lot of work. Because we think a lot of the primitives that they offer are somewhat crude, and you have to spend a lot of time configuring them and putting them together. And so I don't think that the right building blocks are there. So we think that there's a massive opportunity slightly higher up in the stack to weave together all that physical capacity and offer a different set of building blocks that is more geared towards these new applications. And by doing that, we can let engineers iterate much faster and not have to worry about a lot of the underlying physical infrastructure.
Dave Vellante
>> So you would use a Modal container versus Kubernetes, is that right?
Erik Bernhardsson
>> That's right. Yeah. We replace Kubernetes from-
Dave Vellante
>> Compare the two. To a lay person, what's the difference in terms of the developer experience?
Erik Bernhardsson
>> Yeah, yeah. So the number one thing is, like you said, the developer experience and the ability to iterate. So one of the really cool things we can do is when you're writing code in Modal, you can iterate almost as if it's like running locally. So you can have code on your local computer and you can write it, and then you can hit enter, and we can trigger that code running in the cloud within about a second. So we have that superfast feedback loop. Whereas traditionally, you would've to build a container, push that container to the cloud, trigger that container, go download logs. It is a very slow set of feedback loop, many minutes or even hours, sometimes days. So the most important thing is having that superfast feedback loop. The other thing is making everything programmable. So everything in Modal is in code. We focus a lot on Python, because that's just where machine learning engineers, how they think. Although we have other SDKs, we offer Go and TypeScript as well. And then kind of rethinking the fundamental primitives. We think more in terms of here's the function and deploy that function into the cloud, and don't have to think about necessarily scaling and containers and all these things. We sort of manage that fully automatically for you. So it's just one tool that handles all of that stuff, gets it into production, handles all the scaling up and down. In a way, using better abstractions so engineers don't have to think about all the containers and writing YAML and configuring things and running a bunch of different commands, and it all kind of just works in one single command.
Dave Vellante
>> And you do that function experience, you do that using existing serverless function from the AWS, Google, and Azure, or it's your own-
Erik Bernhardsson
>> It's our own stack.
Dave Vellante
>> Your own stack? Okay.
Erik Bernhardsson
>> Our own stack.
Dave Vellante
>> Okay. So essentially you're using the clouds for the core infrastructure.
Erik Bernhardsson
>> That's right.
Dave Vellante
>> The security that they bring around the S3 bucket or whatever it is, and then you build on top of that?
Erik Bernhardsson
>> Yeah.
Dave Vellante
>> Okay.
Erik Bernhardsson
>> Primarily the compute primitive, so EC2 in case of AWS. Running physical nodes in the cloud, that is the main thing we use. And then everything above that, we basically built our own stack.
Dave Vellante
>> Okay. And then your developers would tap whatever other services they need in storage or that.
Erik Bernhardsson
>> Exactly. Yeah, we also have built their own storage primitives, and that runs on top of S3 and R2 and Cloudflare. But again, we sort of abstract the way the complexity of that. For them, it's just a CLI or a Python SDK.
Dave Vellante
>> And you mentioned latency before. Do you run on-prem as well in a hybrid model?
Erik Bernhardsson
>> We don't do on-prem. We don't do on-prem.
Dave Vellante
>> Okay.
Erik Bernhardsson
>> We think the future is in the cloud. We think that's where the wind is going. That's what we're focusing on today. Who knows, maybe we'll do on-prem in the future. But right now, making everything work seamlessly in the cloud is our focus. We see so much revenue growth coming from that. And so as long as we ...
Dave Vellante
>> Not a bad bet.
Erik Bernhardsson
>> That keeps going up. We .
Dave Vellante
>> Definitely not a bad bet for a startup. And I'm sure your investors want you to be focused, and I want to get to that. The other question I had is around your superpowers ML, so deterministic AI. How do you see generative AI playing in your future?
Erik Bernhardsson
>> I don't think it's different from ... I mean, most of our use cases are generative AI. Me personally, I actually use the terms almost interchangeably. I used to call it machine learning a few years ago. Then everyone started calling it AI. So we think of Modal's focus as more broad than either machine learning, AI or data. We think of it as it's a broader set of ... I think it was compute-intensive workloads. And we actually have many other applications that are not even AI. We work with a number of customers who are using us for computational biotech. So I was talking to a customer the other day, and they're literally trying to cure cancer, which I find incredibly mind-blowing using Modal by running protein folding or looking at molecules, how drugs bind to proteins and trying to model that binding energy. So there's some really cool applications also outside of AI. Talked to another customer the other day using us for weather forecasting, which I also find incredibly cool. So there's many other things happening also outside of AI leveraging GPUs or compute-intensive systems, but AI is our core focus.
Dave Vellante
>> Right, right. So I'm misinterpreting your use of the term ML. I'm thinking it's deterministic machine learning, traditional machine learning versus ... No, you're saying it's the whole gamut. I mean, you're doing LLMs, you're doing AI, you're doing machine learning.
Erik Bernhardsson
>> Yeah, all things. Yeah, a lot of different stuff. Yeah.
Dave Vellante
>> Okay. Tell me about the company, how you're funded, when did you start the company. Are you pre-revenue?
Erik Bernhardsson
>> Nope.
Dave Vellante
>> You got revenue?
Erik Bernhardsson
>> We have revenue. No, we started the company, actually just a few blocks from here. I live here and our first office was just also two blocks from here.
Dave Vellante
>> Nice.
Erik Bernhardsson
>> Which is cool. Started in 2021, during the pandemic. We've sort of grown over the last four years to about 65 people now. Based here in New York, so most people are here in New York. We're about 45 engineers, so most of the team is really building the product. We're also expanding and building out a world-class go-to-market team. That's a big investment for us. In terms of funding, we raised our seed round primarily from Amplify based in SF. We raised our A round from Redpoint, also based in SF. And then most recently, we raised our B round from Lux, which is based in New York.
Dave Vellante
>> Oh, okay. And how much have you raised to date? Have you disclosed that?
Erik Bernhardsson
>> I believe the total amount is 111 million.
Dave Vellante
>> Substantial. Okay. And so I presume if you're scaling your go-to-market, you feel very strongly that you have product-market fit.
Erik Bernhardsson
>> Yeah.
Dave Vellante
>> How do you determine that as a startup?
Erik Bernhardsson
>> Yeah. I mean, first of all, revenue, right? I recently crossed about 60 million in annualized revenue.
Dave Vellante
>> 6 million?
Erik Bernhardsson
>> 60 million.
Dave Vellante
>> Six-zero?
Erik Bernhardsson
>> Yeah, yeah.
Dave Vellante
>> Okay. So substantial.
Erik Bernhardsson
>> Very strong growth.
Dave Vellante
>> Wow, okay.
Erik Bernhardsson
>> Tons of demand from customers. And historically, it's all been inbound. We basically had zero marketing or sales for a long time. The market is heating up though. We see so many customers out there demanding products like us. We're also starting to see competitors. I mean, I don't know, the market is heating up and there's more people in this category broadly. So we think now is the right time to start invest in both marketing and sales, in particular sales.
Dave Vellante
>> What are you looking for in your go-to-market professionals? Consultant kind of salespeople, or sort of volume, both?
Erik Bernhardsson
>> It's both. I mean, we both see more in the startup side, we see sort more high velocity. A lot of customers almost coming inbound, and they don't need a lot of help. They kind of know what they're doing. They just want to talk a bit about pricing and understand at scale, what does it mean to work with Modal. But then of course, there's also more like the enterprisey type customers. We're seeing a lot of digital natives and public companies, sometimes very traditional companies coming to us, and they need a little bit more handholding. They need people to figure out the procurement process, fill out the InfoSec questionnaires and other stuff. So that's more sort of traditional sales. And in that case, you typically also need some solutions engineering, solutions architects, or working a little bit more closely with them and trying to figure out how do we take their problem and how do we get it running on Modal. It's a little bit more labor-intensive.
Dave Vellante
>> Yeah. Outstanding. Well, congratulations on your success.
Erik Bernhardsson
>> Yeah, thank you.
Dave Vellante
>> And best of luck. Really appreciate you coming into our studio.
Erik Bernhardsson
>> Of course. It's been awesome.
Dave Vellante
>> Love to have you back and track your progress.
Erik Bernhardsson
>> Thank you.
Dave Vellante
>> Thank you, Erik. All right. And thank you for watching our Mixture of Expert series. NYSE Wired and theCUBE will be right back, right after this short break from the New York Stock Exchange. I'm Dave Vellante, keep it right there.
>> Hey everybody, welcome back to the New York Stock Exchange. We're here in the Buttonwood podium overlooking the options exchange. My name is Dave Vellante and you're watching the NYSE Wired and theCUBE's Mixture of Experts series. Erik Bernhardsson is here. He's the CEO and co-founder of Modal, a company that does serverless compute and really cloud innovator, startup. Super psych to have you here, Erik.
Erik Bernhardsson
>> Yeah, it's great to be here. Thanks a lot for having me.
Dave Vellante
>> You were very welcome. Why did you start the company? What's Modal all about?
Erik Bernhardsson
>> Yeah, it's a long story, but I think the shorter version is I have built a lot of AI products in my career so far. In particular, I was at Spotify for many years and built the music recommendation system and realized at some point that a lot of traditional infrastructure just isn't built for AI and for a lot of these newer types of applications. And with infrastructure, I mean this sort of software layer sort of sitting under the applications, you have the applications and this is what the machine learning engineers are building, and these could be things like training models, or deploying models, or building vibe coding platforms, or computational biotech, or other things. They only need infrastructure, right? So that infrastructure layer today is often built for very different types of applications. So we realized at some point we should just build a whole new layer that is more geared towards these types of new applications and rethink a lot of the assumptions and focus on the needs of these new applications in terms of scaling up and down, or working with large models, working with GPU capacity all over the world, et cetera. So that's why we started Modal.
Dave Vellante
>> So interesting. I mean, I never studied Spotify deeply, although as a consumer and a user of the service, I always felt like there was some really smart people behind the scenes. You were obviously one of them building what I refer to as data products. I mean, essentially, you had data sets and you would serve up services to consumers like myself that were really well-thought-out and the infrastructure underneath was something that you probably didn't get off the shelf as you had to build this, right? Is that what you're implying? And then now what you're doing is bringing that to the masses? Is that-
Erik Bernhardsson
>> No, that's exactly right. So a big part of what Spotify focused on was, how do we make it easy for people to know what to play? You have 10 million tracks, where do you start? And so my team specifically focused a lot on music recommendations and built a lot of features around that. And my experience building those products was because you didn't have the right infrastructure, you ended up spending way more time building the infrastructure than actually spending time building the applications. Then I realized, "Someone should just build this layer." And I'm sure Spotify isn't the only company, so I'm sure there's 10,000 other companies out there who also need this infrastructure layer. And that's what turned out to be quite true. We realized so far there's a lot of companies out there who need better infrastructure, especially now that we're switching to GPUs and very large models and LMs and all these things. You especially need more infrastructure.
Dave Vellante
>> So there are a lot of recommendation engines out there and a lot of really crappy ones. You think about retail, a lot of times, you would go these kind of conditioned response type of experiences, which were terrible. You got four choices. Which one are you? None of the above. Spotify obviously, and TikTok obviously have very strong, whether it's algorithms or whether it's infrastructure. But now, we're entering sort of a new era beyond things like recommendation engines where we're doing so many more interesting things with AI. What's your vision in terms of supporting some of those new applications and capabilities?
Erik Bernhardsson
>> Yeah. I mean, first of all, it's all switching to GPUs. We started playing around with GPUs in 2014 at Spotify, which was quite early, but most of this stuff wasn't running on GPUs, and I don't know if it's still running. I left a while ago, so that's a big change. And then the models have gotten very big, so that's another big change is we're working with these very large LLMs these days. There's obviously generated media, which kind of completely changes the game. So you have algorithms that can generate music these days, generate video, generate images. I mean, all these things are just complete game changer for a lot of companies. And we're starting to see that. We're starting to see a lot of companies adopting these things. Initially, we saw a lot of early stage companies adopting these things, but what we see today is a lot of demand is also coming from later stage, sometimes public companies to try to deploy these things, and that's when they often come to us, because they need help with that.
Dave Vellante
>> So interesting. So 2014, you had experience with GPUs. So I think that kind of brings you into the CUDA era and the programming of those GPUs. And again, I come back to this notion of mainstream. So our vision is, we wrote a piece a while ago, why Jamie Dimon is Sam Altman's biggest competitor. The thinking there was there's all this wonderful proprietary data sitting inside of places like JPMC. We use that as a metaphor for enterprises, and we've also written about Uber for all. Uber had to build its own infrastructure like Spotify had to build its own infrastructure. So Uber for all is, again, a metaphor for solutions or capabilities that allow enterprises to actually build this stuff much more cheaply and faster without having to hire thousands of engineers. Is that what you're bringing to the table? And what other piece parts do you see emerging? Or is it part of your TAM expansion strategy to build those other piece parts?
Erik Bernhardsson
>> Yeah, I mean, Modal focuses specifically on these machine learning problems. So our consumers are people who build the models, but I think there's a larger trend here, which you're talking about, which is I almost think of every company becoming a tech company, and also the tech company's becoming just normal companies. Spotify, for instance, today, are they a music company, are they a tech company? I don't know. It almost doesn't matter, right? They started out being very much a tech company, but the truth is thousands of other companies had similar problems like music distribution and working with large data sets, and those used to be unique things that Spotify had to solve in an idiosyncratic ways in their own ways. But today, I think those are just product you can buy from a lot of vendors.
And so I think there's an emergence of so many different infrastructure vendors and so many different tools, and I think there's also ... You look at the existing companies that previously weren't tech companies, they're buying all these vendors. So I think it's a normalization of tech, and I think you're going to see the same thing with AI, who's an AI company, who's not an AI company. I don't know. Everyone is an AI company. Everyone is going to need a lot of these things. Get like traditional companies, like you mentioned JP Morgan, how many engineers do they have? 100,000. Absurd number. They're going to need a lot. They're going to have to also buy a lot of this infrastructure. And so there's a role for many different parts of that stack, and we play into one of the lower layers. We focus a lot on running code, but there's many opportunities also higher up in the stack simplifying how to work with AI. And maybe you don't even have to think about the models and maybe you can just point and click. We're not a point of click system, we're not a low-code or no-code system. We're very much a high-code system. We focus on the software engineers, but it's all part of the same trend, I think, of the marketizing or broadening the footprint of technology and AI and the cloud and maybe other things.
Dave Vellante
>> But your basic premise is you shouldn't have to have a PHG and AI infrastructure in order to deploy AI. You're making that simpler, even though it's full stack.
Erik Bernhardsson
>> Yeah. No, I think that's right.
Dave Vellante
>> You're targeting, so what is broken about today's model generally and maybe the cloud model specifically?
Erik Bernhardsson
>> Yeah, yeah. And by the way, I will say that PhDs are actually our best customers, even if they are among the few people who understand the course of this stuff, they're also tired of bad infrastructure. So we sell to the whole spectrum of machine learning PhDs all the way to just any sort of developer.
Dave Vellante
>> But the point is you don't want them wasting their time building infrastructure.
Erik Bernhardsson
>> Exactly, exactly. They should focus on what they're good at, which is to build amazing models. Right? So going back to your question, what's broken with today's infrastructure? A lot of today's infrastructure is built for very different applications, built for backend applications. So it means you have kind of smooth workloads. They're very IO-bound in the sense that they often just coordinate work between many other systems like maybe databases, et cetera. What I'm talking about are things just to drop some names here like Kubernetes or Docker, et cetera. And that's been the dominant development paradigm and the dominant infrastructure for the last 10 years or so. But when you start to look at a lot of these new applications, you run into a lot of different problems. First of all, you have a lot of GPU capacity, or actually I should say GPU scarcity. So you have different regions all over the world with different capacity, and often when you start to scale up, you run into constraints. Maybe you run out of GPUs in one region, so you have to go to a different region, or you need to run things close to you so you have certain latency constraints because speed of light is a real thing for a lot of real-time applications. Another thing is these models are very big. So building applications that have to load up these models, load them into the GPU and do that fast and scale up reactively to user workloads, you need to really think deeply about how do you build infrastructure that can do that, especially when you look at inference, and we're very focused on inference historically. We've had a lot of use cases in inference. And when you're doing inference, fundamentally, you don't know what the user demand is. The user demand goes up and down, and so scaling up and down with these workloads is a real challenge. And so we basically decided at some point in order to support a lot of those things, we had to build basically our own stack. We decided we can't build a new stack on top of an old stack, so let's throw out the old stack, let's build our own file system, let's build our own container runtime, let's build our own scheduler and many other things. And we had to do that in order to be able to support these new types of applications. The other thing I think is very broken also is developer experience. So it wasn't as much of a problem in the old paradigm, but when you're starting to look at these AI researchers and how they're operating and how they're trying to leverage the old infrastructure, they're only into a lot of challenges. Just being able to take models and get it out in production and putting it in front of users, it's a real challenge for these developers. So we spent a lot of time super obsessed with like, how do we make it fun for engineers? How do we make them fast? How do we really improve the speed to market? Because I think that's almost as important as or more important maybe than everything else. There are so many companies out there where speed to market is really the thing that matters. So by making engineers 10x more productive, you're also gaining a massive advantage.
Dave Vellante
>> So let me ask you, but follow up on that. So if I think about the cloud today, I've got access to all these primitives and services through APIs.
Erik Bernhardsson
>> That's right. Yep.
Dave Vellante
>> I got Elastic Beanstalk, I've got Kubernetes, and you're saying this is just too complicated today. It requires too much brute force to actually scale, and you're orchestrating that in a much simpler way and streamlining that experience.
Erik Bernhardsson
>> Yeah, that's exactly right. So first of all, we're not trying to compete with the cloud vendors. We are customers of the cloud vendors.
Dave Vellante
>> Yeah, you're running on the cloud.
Erik Bernhardsson
>> Yeah, we're running on the cloud and we're super happy. We consider them partners. We use a lot of different cloud vendors, AWS, GCP, et cetera, Oracle and Azure and many others. They're very good at delivering physical data, like running physical data centers and delivering compute capacity through an API.
Dave Vellante
>> You need compute networking there.
Erik Bernhardsson
>> Exactly. The lower stuff, right?
Dave Vellante
>> Yeah.
Erik Bernhardsson
>> What I think there's a gap is slightly higher up in the stack, and that's really most visible in developer experience, like working and building these AI applications. You can build them straight on the clouds, but it's pretty hard. And even if it's hard, I mean I know how to do it. It's also annoying. It takes a lot of work. Because we think a lot of the primitives that they offer are somewhat crude, and you have to spend a lot of time configuring them and putting them together. And so I don't think that the right building blocks are there. So we think that there's a massive opportunity slightly higher up in the stack to weave together all that physical capacity and offer a different set of building blocks that is more geared towards these new applications. And by doing that, we can let engineers iterate much faster and not have to worry about a lot of the underlying physical infrastructure.
Dave Vellante
>> So you would use a Modal container versus Kubernetes, is that right?
Erik Bernhardsson
>> That's right. Yeah. We replace Kubernetes from-
Dave Vellante
>> Compare the two. To a lay person, what's the difference in terms of the developer experience?
Erik Bernhardsson
>> Yeah, yeah. So the number one thing is, like you said, the developer experience and the ability to iterate. So one of the really cool things we can do is when you're writing code in Modal, you can iterate almost as if it's like running locally. So you can have code on your local computer and you can write it, and then you can hit enter, and we can trigger that code running in the cloud within about a second. So we have that superfast feedback loop. Whereas traditionally, you would've to build a container, push that container to the cloud, trigger that container, go download logs. It is a very slow set of feedback loop, many minutes or even hours, sometimes days. So the most important thing is having that superfast feedback loop. The other thing is making everything programmable. So everything in Modal is in code. We focus a lot on Python, because that's just where machine learning engineers, how they think. Although we have other SDKs, we offer Go and TypeScript as well. And then kind of rethinking the fundamental primitives. We think more in terms of here's the function and deploy that function into the cloud, and don't have to think about necessarily scaling and containers and all these things. We sort of manage that fully automatically for you. So it's just one tool that handles all of that stuff, gets it into production, handles all the scaling up and down. In a way, using better abstractions so engineers don't have to think about all the containers and writing YAML and configuring things and running a bunch of different commands, and it all kind of just works in one single command.
Dave Vellante
>> And you do that function experience, you do that using existing serverless function from the AWS, Google, and Azure, or it's your own-
Erik Bernhardsson
>> It's our own stack.
Dave Vellante
>> Your own stack? Okay.
Erik Bernhardsson
>> Our own stack.
Dave Vellante
>> Okay. So essentially you're using the clouds for the core infrastructure.
Erik Bernhardsson
>> That's right.
Dave Vellante
>> The security that they bring around the S3 bucket or whatever it is, and then you build on top of that?
Erik Bernhardsson
>> Yeah.
Dave Vellante
>> Okay.
Erik Bernhardsson
>> Primarily the compute primitive, so EC2 in case of AWS. Running physical nodes in the cloud, that is the main thing we use. And then everything above that, we basically built our own stack.
Dave Vellante
>> Okay. And then your developers would tap whatever other services they need in storage or that.
Erik Bernhardsson
>> Exactly. Yeah, we also have built their own storage primitives, and that runs on top of S3 and R2 and Cloudflare. But again, we sort of abstract the way the complexity of that. For them, it's just a CLI or a Python SDK.
Dave Vellante
>> And you mentioned latency before. Do you run on-prem as well in a hybrid model?
Erik Bernhardsson
>> We don't do on-prem. We don't do on-prem.
Dave Vellante
>> Okay.
Erik Bernhardsson
>> We think the future is in the cloud. We think that's where the wind is going. That's what we're focusing on today. Who knows, maybe we'll do on-prem in the future. But right now, making everything work seamlessly in the cloud is our focus. We see so much revenue growth coming from that. And so as long as we ...
Dave Vellante
>> Not a bad bet.
Erik Bernhardsson
>> That keeps going up. We .
Dave Vellante
>> Definitely not a bad bet for a startup. And I'm sure your investors want you to be focused, and I want to get to that. The other question I had is around your superpowers ML, so deterministic AI. How do you see generative AI playing in your future?
Erik Bernhardsson
>> I don't think it's different from ... I mean, most of our use cases are generative AI. Me personally, I actually use the terms almost interchangeably. I used to call it machine learning a few years ago. Then everyone started calling it AI. So we think of Modal's focus as more broad than either machine learning, AI or data. We think of it as it's a broader set of ... I think it was compute-intensive workloads. And we actually have many other applications that are not even AI. We work with a number of customers who are using us for computational biotech. So I was talking to a customer the other day, and they're literally trying to cure cancer, which I find incredibly mind-blowing using Modal by running protein folding or looking at molecules, how drugs bind to proteins and trying to model that binding energy. So there's some really cool applications also outside of AI. Talked to another customer the other day using us for weather forecasting, which I also find incredibly cool. So there's many other things happening also outside of AI leveraging GPUs or compute-intensive systems, but AI is our core focus.
Dave Vellante
>> Right, right. So I'm misinterpreting your use of the term ML. I'm thinking it's deterministic machine learning, traditional machine learning versus ... No, you're saying it's the whole gamut. I mean, you're doing LLMs, you're doing AI, you're doing machine learning.
Erik Bernhardsson
>> Yeah, all things. Yeah, a lot of different stuff. Yeah.
Dave Vellante
>> Okay. Tell me about the company, how you're funded, when did you start the company. Are you pre-revenue?
Erik Bernhardsson
>> Nope.
Dave Vellante
>> You got revenue?
Erik Bernhardsson
>> We have revenue. No, we started the company, actually just a few blocks from here. I live here and our first office was just also two blocks from here.
Dave Vellante
>> Nice.
Erik Bernhardsson
>> Which is cool. Started in 2021, during the pandemic. We've sort of grown over the last four years to about 65 people now. Based here in New York, so most people are here in New York. We're about 45 engineers, so most of the team is really building the product. We're also expanding and building out a world-class go-to-market team. That's a big investment for us. In terms of funding, we raised our seed round primarily from Amplify based in SF. We raised our A round from Redpoint, also based in SF. And then most recently, we raised our B round from Lux, which is based in New York.
Dave Vellante
>> Oh, okay. And how much have you raised to date? Have you disclosed that?
Erik Bernhardsson
>> I believe the total amount is 111 million.
Dave Vellante
>> Substantial. Okay. And so I presume if you're scaling your go-to-market, you feel very strongly that you have product-market fit.
Erik Bernhardsson
>> Yeah.
Dave Vellante
>> How do you determine that as a startup?
Erik Bernhardsson
>> Yeah. I mean, first of all, revenue, right? I recently crossed about 60 million in annualized revenue.
Dave Vellante
>> 6 million?
Erik Bernhardsson
>> 60 million.
Dave Vellante
>> Six-zero?
Erik Bernhardsson
>> Yeah, yeah.
Dave Vellante
>> Okay. So substantial.
Erik Bernhardsson
>> Very strong growth.
Dave Vellante
>> Wow, okay.
Erik Bernhardsson
>> Tons of demand from customers. And historically, it's all been inbound. We basically had zero marketing or sales for a long time. The market is heating up though. We see so many customers out there demanding products like us. We're also starting to see competitors. I mean, I don't know, the market is heating up and there's more people in this category broadly. So we think now is the right time to start invest in both marketing and sales, in particular sales.
Dave Vellante
>> What are you looking for in your go-to-market professionals? Consultant kind of salespeople, or sort of volume, both?
Erik Bernhardsson
>> It's both. I mean, we both see more in the startup side, we see sort more high velocity. A lot of customers almost coming inbound, and they don't need a lot of help. They kind of know what they're doing. They just want to talk a bit about pricing and understand at scale, what does it mean to work with Modal. But then of course, there's also more like the enterprisey type customers. We're seeing a lot of digital natives and public companies, sometimes very traditional companies coming to us, and they need a little bit more handholding. They need people to figure out the procurement process, fill out the InfoSec questionnaires and other stuff. So that's more sort of traditional sales. And in that case, you typically also need some solutions engineering, solutions architects, or working a little bit more closely with them and trying to figure out how do we take their problem and how do we get it running on Modal. It's a little bit more labor-intensive.
Dave Vellante
>> Yeah. Outstanding. Well, congratulations on your success.
Erik Bernhardsson
>> Yeah, thank you.
Dave Vellante
>> And best of luck. Really appreciate you coming into our studio.
Erik Bernhardsson
>> Of course. It's been awesome.
Dave Vellante
>> Love to have you back and track your progress.
Erik Bernhardsson
>> Thank you.
Dave Vellante
>> Thank you, Erik. All right. And thank you for watching our Mixture of Expert series. NYSE Wired and theCUBE will be right back, right after this short break from the New York Stock Exchange. I'm Dave Vellante, keep it right there.