Technology
BENGALURU: AI is a particularly well-suited tech trajectory for India: Cerebras’ Feldman
BENGALURU: Andrew Feldman, co-founder and CEO of US-based Cerebras Systems, which last month released what it touts as the world’s largest and fastest artificial intelligence (AI) ‘chip’ with 4 trillion transistors, insists that while AI is not a silver bullet, it can certainly solve many of the world’s problems. In a video interview from his US office, Feldman spoke about areas where his company’s third-generation wafer-scale AI accelerator—the CS-3—scores over Nvidia’s graphics processing units (GPUs). He also shared his thoughts on how India can leverage AI and Generative AI (GenAI) tools, and whether GenAI is being oversold, among other things. Edited excerpts:
How is this large chip (CS-3) helping train AI models faster, and how do you see businesses, academic institutions and governments leveraging it?
One of the fundamental challenges in AI right now is to distribute a single model over hundreds or thousands of GPUs. You can’t fit the big matrix multipliers (matrix multiplication is a big part of the math done in deep learning models, which requires significant computing power) on a single GPU. But we can fit this on a single wafer, and so we can bring to the enterprise and the academician, the power of tens of thousands of GPUs but the programming simplicity of a single GPU, helping them do work that they wouldn’t otherwise be able to do.
We are able to tie together dozens or hundreds of these (chips) into supercomputers and make it easy to train big models. GPT-4 cited in its paper that it had 240 contributors, of which 35 are mostly doing distributed computing. Which enterprise company has 35 supercomputer jockeys, whose job it is to do distributed computing? The answer is, very few. That means it’s very difficult for them (most companies) to do big AI work. We eliminate that need (with this big chip).
Please share some examples of how companies across sectors are working with you to leverage this big chip.
Companies like GlaxoSmithKline Pharmaceuticals are using us to do genomic research and in the drug design workflow. Companies like Mayo Clinic, one of the leading medical institutions in the world, have multiple projects. Some of them are looking at using genetic data to predict which rheumatoid arthritis drug would be best for a given individual. Others are doing hospital administration—how you can predict how long a patient will stay in a hospital based on the medical history.
Customers like Total (TotalEnergies)—the giant French oil and gas company—are using us to do AI work in oil exploration. We also have government customers and those who are doing research on the Covid virus. We have government researchers who include our system in giant physics simulations, where they use what’s called a Simulation plus AI or HPC (high performance computing) plus AI (system), where the AI is doing some training work and recommending starting points for the simulator.
How’s your partnership with Abu Dhabi-based Group 42 Holding panning out for the Arabic LLM and supercomputers you’re building with them?
G42 is our strategic partner. We’ve completed two supercomputers in the US, four exaflops each, and we’ve just started the third supercomputer in Dallas, Texas. We announced that we will build nine supercomputers with them. We also saw the opportunity to train an Arabic LLM to cater to the 400 million native Arabic speakers. G42 had the data and we both had researchers who we brought together to train what is head and shoulders, the best Arabic model in the world.
We have many projects underway with them. We also trained one of the best coding models called Crystal Coder. We have also worked with M42, a JV between G42 Healthcare and Mubadala Health, and trained a medical assistant. The aspirations in the Emirates are extraordinary, and the vision and desire to be a leader in AI, exceptional.
What about India, where you have already had talks with certain companies and government officials, too?
We had lots of conversations with data centre owners, cloud providers, and with government officials in New Delhi. We have a team of about 40 engineers in Bangalore (Bengaluru) and we’re growing as fast as we can there. India has some of the great university systems—the IITs and NITs of the world. And many of the researchers working on big compute problems around the world were trained in India.
Obviously, India is one of the most exciting markets in the world, but it does not have enough supercomputers for the talent it has. So, it’s both important for sovereignty and for a collection of national issues to have better infrastructure in India to create an opportunity to keep some of its world-class talent that wants to work on the biggest supercomputers.
I think AI is a particularly well-suited technology trajectory for India. It builds on a strength in CS (computer science) and in statistics that you’ve had in your university system for generations.
Talking about your big chip, companies now appear to be focused more on fine-tuning large language models (LLMs), building smaller language models (SLMs), and doing inference (using a pre-trained AI model to make predictions or decisions on new, unseen data), rather than building large multimodal models (LMMs) and foundational models. Many such customers would do with fewer GPUs. Wouldn’t your big chip prove an overkill and too expensive for such clients?
The amount of compute you need is approximately the product of the size of the model, and times the number of tokens you train on. Now, the cost to do inference is a function of the size of the model. And so, as we’re thinking about how to deploy these models in production, there’s a preference for smaller models. However, there isn’t a preference for less accuracy. So, while the models might be 7 billion(b), 13b or 30b, the number of tokens they’re trained on is a trillion, two trillion (and more). So, the amount of compute you need hasn’t changed. In fact, in many instances, it’s gone up.
Hence, you still need huge amounts of compute, even though the models are smaller, because you’re running so many tokens through so much data. In fact, you’re trading off parameter size with data. And you’re not using less compute, you’re just allocating it differently, because that has different ramifications on the cost of inference.
I also do not subscribe to the view that as inference grows, there will be less training. As inference grows, the importance of accuracy and training increases. And so, there will be more training. If you have a good model, say, for reading pathology slides, and it’s 93% accurate, and you don’t retrain, and someone else comes up with one that’s 94% accurate, who’s going to use your model? And so, there will be tremendous pressure as these are deployed to you to be better, and better, and better. Training continues for years and years to come.
Inference will come in many flavours—there will be easy batch inference, and then there will be real-time inference in which latency matters a huge amount. If you want, the obvious example, self-driving cars that are making inference decisions in near real time. And as we move inference to harder problems, and we include it in a control system, then the inference challenge is much harder. Those are problems we’re interested in. Most of the inference problems today are pretty straightforward, and we’ve partnered with Qualcomm, because they have an excellent offering. And we wanted to be sure we could show up with a solution that did not include Nvidia.
But what about the cost comparison with GPUs?
Inference is on the rise, and our systems today are being used for what I call are real-time, very hard inference problems—predominantly for defence and security. I think there will be more of those over time. But in the next 9-12 months, the market will be dominated by much easier problems.
That said, CS-3 costs about the same as three DGX H100s (Nvidia’s AI system capable of handling demanding tasks such as generative AI, natural language processing, and deep learning recommendation models), and gives you the performance of seven or 10. And so you have a dramatic price performance advantage.
But if you want one GPU, we’re not a good choice. We begin being sort of equivalent to 40 or 50 GPUs. So, we have a higher entry point, but we’re designed for the AI practitioner who’s doing real work—you have to be interested in training models of some size, or some complexity on interesting data sets. That’s where we enter.
But Nvidia is estimated to have 80-95% global market share in AI chips, and it will be very difficult to compete in this space.
I don’t think we have to take share. I think the market is growing so fast. I mean, Nvidia added $40 billion last year to their revenue. And that market is growing unbelievably quickly. The universe of AI is expanding at an extraordinary rate. And there’ll be many winners. We did big numbers last year. We’re going to do bigger numbers this year. We’ve raised in total about $750 million till date, and the last round’s valuation was $4.1 billion.
How water and energy efficient are your big chips?
There are several interesting elements of a big chip. Each chip uses about 18 kilowatts but it replaces 40 or 50 chips that use 600 watts. Moreover, when you build one chip, it does a lot of work—you can afford more efficient cooling. Also, GPUs use air, and air is an inefficient cooler. We use water, because we can amortize a more efficient and more expensive cooling system over more compute on a wafer. And so, we generally run per-unit compute, somewhere between a third and half the power draw. Why is that? The big chip allows us both to be more efficient in our compute, to keep information on the chip, not move it and spend power in switches (electronic switches are the basic building blocks of microchips), etc. It also allows us to use a more efficient cooling mechanism.
You’ve also said that this big chip has broken Moore’s law. What exactly do you mean?
Moore’s law said the number of transistors on a single chip would double every 18 months at lower costs. But, first, that required the shrinking of the fab geometries. Second, the chips got bigger themselves. But the radical limit, which constrains everybody but Cerebras, was about 815-820 square millimetres. We obliterated the radical limits and went to 46,000 square millimetres. So, in a single chip, we were able to use more silicon to break Moore’s law. That was the insight for this workload—the cost of moving data off the chip, the cost of all these switches, and the cost that forced Nvidia to buy Mellanox (a company Nvidia acquired in March 2019 to optimize datacentre-scale workloads) which could be avoided with a big chip. While everybody else is working with 60 billion, 80 billion, 120 billion transistors, we’re at 4 trillion.
Some people believe that GenAI is being overhyped and are becoming disillusioned, given its limitations including hallucinations, lack of accuracy, copyright violations, trademarks, IP violations, etc. The other school believes that GenAI models will iron out all these kinds of problems over a period of time and achieve maturity by 2027 or 2028. What’s your view?
These arguments only exist if there are kernels of truth on both sides. AI is not a silver bullet. It allows us to attack a class of problems with computers that have historically been foreclosed to us—like images, like text. They allow us to find insight in data in a new and different way. And generally, the first step is to make existing work better—better summarization, we replace people who do character recognition, we replace analysts who looked at satellite imagery with machines, and you get a modest performance of sort of societal benefit GDP (gross domestic product) growth. But it typically takes 3-7 years, following which you begin to reorganize things around the new technology, and get the massive bump.
For instance, computers first replaced ledgers, then assistants, and then replaced typewriters. And we got a little bump in productivity. But when we moved to the cloud, and reorganized the delivery of software where you could gain access to compute anywhere, we suddenly got a huge jump in unit labour and productivity.
So, there are kernels of truth in both arguments. But to people who say this is the answer to everything, you’re clearly going to be wrong. To people who say there are obviously large opportunities to have substantial impact, you’re right.
In my view, AI is the most important technology trajectory of our generation, bar none. But it’s not going to solve every problem—it will give us answers to many problems. It solves protein folding—a problem that humans had not been able to solve until then. It has made games like chess and poker, which had been interesting to people for hundreds and hundreds of years, trivial. It will change the way wars are fought. It will change the way drugs are discovered.
But will it make me a better husband? Probably not. Will it help my friendships? Will it help my dreams and aspirations? No. Sometimes we go crazy thinking about a new technology.
Talking about crazy, what are your quick thoughts on artificial general intelligence (AGI)?
I think we will definitely have machines that can do pretty thoughtful reasoning. But I don’t think that’s AGI. That’s data-driven logical learning. I am not optimistic about AGI in the next 5-10 years, as most people constructed. I think we will get better and better at extracting insight from data, extracting logic from data, and reasoning. But I don’t think we’re close to some notion of self-awareness.
In this context, what should CXOs keep in mind when implementing AI, GenAI projects?
There are three fundamental elements when dealing with AI—the algorithm, data, and computing power. And you have to decide where you have an advantage. Many CXOs have data, and they are sitting on a gold mine if they’re invested in curated data. And AI is a mechanism to extract insight from data. This can help them think about the partnerships they need.
Consider the case of OpenAI, which had algorithms; they partnered with Microsoft for compute, and they used open-source data. GlaxoSmithKline had data, partnered with us for compute, and had internal algorithm expertise. These three parts will help your strategy, and your data will be enormously important for construction of models that solve your problems.
Technology
CAIRO: Indian-Origin Teen, Son Of IIT-JEE Topper, Wins Gold At World’s Toughest Programming Contest
CAIRO: Agastya Goel, a 17-year-old Indian-origin student from California, has won his second gold medal at the 36th International Olympiad in Informatics (IOI), which took place in Egypt. Agastya, son of Stanford professor Ashish Goel, secured the fourth overall rank at the prestigious contest, also dubbed as the world’s toughest programming competition for high school students.
Agastya scored 438.97 out of 600, while Kangyang Zhou of China topped the event with a perfect score of 600. A total of 34 students were awarded gold medals this year, including fellow Indian participant Kshitij Sodani, who ranked 21st.
This marks the 17-year-old’s second gold at the IOI. His father, Ashish Goel, made waves in 1990 by topping the IIT-JEE exam before going on to build a distinguished career at companies like Twitter and Stripe.
Who is Agastya Goel?
Agastya Goel, a junior at Henry M Gunn High School in Palo Alto, California, loves physics, computer science, and music. His hobbies include tennis, hiking, stargazing, and playing the guitar and piano.
Agastya’s interest in physics began at a young age, with his father introducing him to the subject during car rides and hikes. One memorable experience was learning about projectile motion while playing Angry Birds. This sparked a fascination with physics that has only grown stronger over the years.
Agastya’s competitive spirit was ignited through the USACO competition series, which he started preparing for in 6th grade. His hard work earned him a spot at the USACO training camp in his freshman year, followed by selection for the US IOI team.
Initially, Agastya considered majoring in computer science, but after exploring physics, he discovered a new passion. He spent 14 days intensively studying physics over winter break and was hooked.
Technology
CALIFORNIA: PM Modi Put India On World Map As Credible Innovator- IT Industry Leaders
CALIFORNIA: Indian-origin business leaders from the IT industry lauded Prime Minister Narendra Modi’s leadership for putting India on the global map as a “credible innovator and leader” for all nations to invest in and grow.
Speaking at the Indiaspora AI Summit 2024 on ‘Can Artificial intelligence be good force for future or not’ at the University of Stanford on April 26, Informatica CEO Amit Walia said, “India is perceived very differently in the global space. It’s a space of innovation. It’s a space of human capital. It’s a space of progress and innovation. And as we look forward, what Modi has been able to do is put India into the global map as a credible innovator, leader for all countries to invest in and grow.”
Speaking at the same event, Elastic CEO Ashutosh Kulkarni said that PM Modi and the government have been doing amazing work.
Mr Kulkarni said, “The Prime Minister and the government has been doing some amazing work and their focus on being open to innovation, their focus on industry, their focus on bringing the population, India is now one of the, it is the largest, most populous country in the world and bringing that population towards progress, the way they have tapped into the potential of this country has just been fantastic.”
Calling PM Modi’s era “amazing,” Mayfield Fund Managing Partner Navin Chaddha praised PM Modi’s leadership for ties between India and the US and his ability to connect the Indian diaspora which he stressed is doing extremely well in the tech industry. He noted that India is making progress in using technology and improving its infrastructure.
Navin Chaddha said, “The Modi era has been amazing, the kind of relationship they’ve been able to build with the United States. And to be able to connect the Indian diaspora, which is doing extremely well in the tech industry, hats off to them. And also the progress India is making in using technology, improving the infrastructure is just amazing. So, Mr Modi, keep doing what you’re doing.”
Rohit Jain, Speaker and Moderator – CIO, Harvard Business School, called on PM Modi to increase his guarantee, as he has always delivered. The event on Artificial Intelligence was organised to deliberate and discuss artificial intelligence.
He said, “Mr Modi needs to increase his guarantee because he’s always been delivering. So, I think his promises have been 100 per cent true. So, he now needs to promise 110 per cent and then again he’ll deliver on that. Mr Modi, thank you for what you have done, not only for India, but for the world.”
Priya Rajan, Managing Director, National Venture Coverage and Business Development, MUFG lauded PM Modi’s leadership in the entrepreneurial ecosystem. He stated that a number of VCs across the globe want to invest in India.
She said, “I see the Modi magic play is in the entrepreneurial ecosystem, which is an adventure ecosystem. So you can see the number of VCs now wanting to invest, not just where in India, investing in India, but world investors around the world wanting to do things. And you’re seeing kind of US, a ton of US global investors now setting up shops in India. I think that’s all coming back to what Modi magic has created.”
At the event, prominent Indian diaspora, along with the esteemed corporate leaders from the US shared their insights on the topic “Can Artificial Intelligence be a force for future or not.”
Rajya Sabha MP Satnam Singh Sandhu participated in the event. Notably, PM Modi has been advocating that artificial intelligence can provide a great support to help humanity’s development in the 21st century.
Taking to X, Satnam Singh Sandhu stated, “Indiaspora AI Summit, Stanford University, USA: Delighted to be a part of the summit where we had @ericgarcetti, US Ambassador to India, Indian Diaspora and business leaders from IT sectors who reiterated how PM @narendramodi Ji is leveraging AI to shape nation’s future strategy to build Viksit Bharat by 2047.”
“They highlighted PM Modi’s tech-savvy approach and his efforts to promote AI applications in various sectors like agriculture, health-care, and education for social development and inclusive growth. They also condemned the Western bias against India to rank them lower in parameters like Ease of doing business, and cite the irregularities while computing such rankings,” he added.
Earlier in December, PM Narendra Modi said that artificial intelligence is a dynamic issue and that India is working on creating a global framework for AI. He made the remarks while interacting with the participants of the Grand Finale of the Smart India Hackathon.
Speaking to the participants via video conferencing, the Prime Minister said, “AI (Artificial Intelligence) is a very dynamic issue. If one solution is brought in, the mischievous people will find another way. We have to be very alert to any new technology. It can be very useful if used within rules, but if it is used in the wrong way, it can create a lot of problems.”
“We all are aware of the generative AI-created deep fake videos that look completely real. Hence we need to be more alert before believing any photo or video. India is working on creating a global framework for AI,” said PM Modi.
Technology
SILICON VALLEY: All About Pavan Davuluri, New Head Of Microsoft Windows
SILICON VALLEY: Pavan Davuluri, an IIT Madras graduate, has been appointed the new head of Microsoft Windows and Surface. He succeeds Panos Panay, who left for Amazon last year. Mr Davuluri previously oversaw the Surface group and now leads both Windows and Surface departments.
Rajesh Jha, Microsoft’s head of experiences and devices, said, “As part of this change, we are bringing together the Windows Experiences and Windows + Devices teams as a core part of the Experiences + Devices (E+D) division. This will enable us to take a holistic approach to building silicon, systems, experiences, and devices that span Windows client and cloud for this AI era.”
“Pavan Davuluri will lead this team and continue to report to me. Shilpa Ranganathan and Jeff Johnson and their teams will report directly to Pavan. The Windows team will continue to work closely with the Microsoft AI team on AI, silicon, and experiences,” he said.
Who is Pavan Davuluri?
- Pavan Kumar Davuluri is an IIT Madras graduate. He completed his post-graduation from the University of Maryland, according to his LinkedIn
- Mr Davuluri began his career at Microsoft as a Reliability Component Manager in 2001.
- Mr Davuluri has held various leadership positions over his 23-year tenure at Microsoft, working across PC and Xbox hardware, Surface, and Windows.
- Pavan Davuluri took on the role of corporate vice president for Windows and Silicon & Systems Integration in 2021, overseeing efforts to optimise Windows for Arm-based devices.
- Mr Davuluri was recently assigned to lead Windows engineering alongside his existing responsibilities in Microsoft’s hardware endeavours. He will lead a dedicated team developing silicon systems for Microsoft’s Windows and Surface products.
-
Diplomatic News1 year agoSTOCKHOLM: Dr. Neena Malhotra appointed as the next Ambassador of India to the Kingdom of Sweden
-
Opinions4 years ago
2020 will be remembered as time of the pandemic. The fallout will be felt for years
-
Diplomatic News1 year agoMELBOURNE: Fourth India-Australia 2+2 Secretary-level Consultations
-
Diplomatic News1 year agoKINGSTON: Shri Subhash Prasad Gupta concurrently accredited as the next High Commissioner of India to St.Vincent and the Grenadines
-
Diplomatic News3 years agoROME : State Visit of Prime Minister of Italy to India
-
Diplomatic News3 years agoMUSCAT : Envoys of five nations present credentials to the President of India
-
Education3 years agoDHAKA : Sheikh Hasina lauds PM Modi for evacuating Bangladeshi students
-
Diplomatic News3 years agoDHAKA : 38th Meeting of Ministerial level Joint Rivers Commission of India and Bangladesh held at New Delhi
