💫 Resumen
Tesla launches a massive $300 million AI cluster, including 10,000 NVIDIA H100 Compute GPUs, to power its full self-driving product. OpenAI launches Chat GPT Enterprise to address privacy and security concerns among businesses. AI beats world champion drone racers, demonstrating real-world applications such as environmental monitoring and disaster reporting.
✨ Destacados📊 Transcripción
Copiar
Chatear con el video
Tesla launches $300 million AI cluster with 10,000 NVIDIA H100 Compute GPUs for training full self-driving product
00:00
The new AI cluster will be used to power several AI applications
Tesla may have the largest training datasets in the world due to real-world video training
Elon Musk showed off the newest version of full self-driving, version 12, in a live stream violating Tesla's rules for its advanced autopilot technology
Full self-driving V12 will be the first version to remove the beta label and will be entirely cameras and AI, as opposed to previous versions which mixed in other sensors.
OpenAI's dominance in AI language models may be threatened by the open-source LLaMA by Meta, while OpenAI has launched Chat GPT Enterprise to address businesses' privacy and security concerns.
03:01
Users are split between using only one prompt or five prompts during their Chat GPT session, with these two ends of the spectrum accounting for nearly 70% of all visitors.
Popular words used in Chat GPT prompts include "write", "create", "list", and "fun".
OpenAI's revenue and costs are exploding, while usage is down, but the company has launched Chat GPT Enterprise to address businesses' privacy and security concerns.
Chat GPT Enterprise features include customer data not being used for training OpenAI models, data encryption, and SOC 2 compliance, as well as an admin console, single sign-on, and unlimited usage of GPT for increased speeds and larger context sizes.
AI drone racing system beats world champion human drone racers
06:00
The AI system called Swift, designed by University of Zurich researchers, beat the best human drone racers in the world
The AI control drone was able to beat the world record by half a second, which is significant in the world of drone racing
This accomplishment has real-world applications such as environmental monitoring, disaster reporting, and rescue operations
A16Z announced a grant program to give funding to a small group of AI developers to help the open source community
Google launches AI watermark for identifying AI-generated images; updates Vertex AI platform
09:00
Google introduced an AI watermark that helps identify AI-generated images created by their AI Gen art product, Imagine.
The company also launched access to their new AI training cluster based on their custom-built TPU architecture.
Google updated its Vertex AI platform with upgrades to POM2, enhanced code generation, and new search and conversational models.
The secretive land purchases made by Silicon Valley elites for building a city from scratch have locals worried.
Apple's M1 and M2 chips are great for running large language models, and Stability AI founder believes we'll see a chat GPT level model on a mobile phone next year with a GPT4 level model the year after that.
12:00
AI pioneer Andre Carpathi explains why M2 chip is a great option for running large language models.
An unquantized 34 billion parameter version of CodeLama can run at 20 tokens per second on an M2 Ultra.
Stability AI founder predicts chat GPT level model on a mobile phone next year and GPT4 level model the year after that.
Regulators are asking for input from the public in determining how to create AI copyright policy.
00:00I say this every week, but we really did
00:02have another incredible week of AI news.
00:05Tesla has multiple AI launches. OpenAI is
00:08raking in the cash, but also struggling.
00:11AI sets records at drone racing, beating
00:13the best humans, and
00:14meta launches Code Llama,
00:16which quickly beat GPT-4 at coding tasks.
00:19We have a ton of other
00:20fantastic tech and AI stories today,
00:22so sit back, relax, and enjoy. Let's go!
00:26First, let's talk about Tesla. This week,
00:28Tesla launched a massive
00:29$300 million AI cluster,
00:31including 10,000 NVIDIA H100 Compute
00:34GPUs. By the way, this is
00:36why NVIDIA is crushing earnings
00:39lately. Their new AI cluster will be used
00:41to power several AI
00:42applications, but of course,
00:44its main use will be to continue training
00:46its full self-driving
00:47product. And they have good reason
00:49to have such a massive supercomputer.
00:51According to Tim Zaman, AI infra and AI
00:54platform engineering
00:55manager at Tesla, "Due to real-world
00:57video training, we may
00:59have the largest training
01:00datasets in the world. Hot-tier cache
01:03capability beyond 200
01:05petabytes. Orders of magnitude larger
01:08than LLMs." Also, this week, Elon Musk
01:11showed off the newest
01:12version of full self-driving,
01:14version 12, in a live stream he did while
01:17driving, violating
01:18Tesla's rules for its advanced
01:20autopilot technology. And the 45-minute
01:22demo went mainly well,
01:24except for a couple issues,
01:25including almost running a red light.
01:28Musk takes over at that point
01:29and reminds viewers that full
01:30self-driving V12 is still in beta,
01:33although V12 will be the first version to
01:35remove the beta label.
01:36But that Tesla was able to navigate many
01:39complex driving
01:40situations, including roundabouts and
01:41construction zones. Musk also mentions
01:44that V12 will be the first
01:45time that full self-driving is
01:47entirely cameras and AI, as opposed to
01:50previous versions which
01:51mixed in other sensors. Tesla has
01:53found that the best way to account for
01:55the thousands or even
01:57millions of edge cases humans
01:58experience while driving is to use a pure
02:01neural network approach, which is
02:02different from other
02:03companies like Cruise and Waymo. In true
02:05Musk Uber troll fashion, he
02:08claims in the video that he'll
02:09drive over to Mark Zuckerberg's house to
02:11initiate their much-anticipated but
02:12highly unlikely fight.
02:14Next, let's talk about OpenAI. Just a
02:16couple weeks ago, it was
02:17reported that OpenAI is burning
02:19through a tremendous amount of cash,
02:21about $700,000 a day, and costs
02:23associated with running their AI
02:24systems. But now, an article was
02:26published claiming OpenAI
02:28is on track to generate more
02:29than a billion dollars in revenue over
02:31the next 12 months, or about $80 million
02:33a month. A staggering
02:35number given they earned a total of $28
02:37million in all of last
02:38year. So it looks like burning
02:40all that cash isn't going to complete
02:41waste. It's generating a
02:43massive amount of revenue.
02:44But at the same time, another report
02:46published by Spark Toro shows visits to
02:49chat GPT are down 29%
02:51since its peak in May. And the majority
02:53of usage is for a coding
02:55assistant. I can tell you from my
02:57experience that coding assistance is my
02:59number one use case. And as we'll see in
03:01a story later in this
03:02video, that dominance may already be
03:05threatened by open source and
03:06completely free code llama by
03:08meta. And there are some other
03:10interesting findings from this report.
03:12Users are pretty split
03:13between using only one prompts during
03:15their session and using five
03:17prompts with nothing really in
03:19between. And those two ends of the
03:20spectrum accounting for
03:21nearly 70% of all visitors. And
03:24some of the most popular words used in
03:26chat GPT prompts include
03:27write, create, list and fun. This
03:30article has other awesome findings. I'll
03:32drop a link in the
03:33description below so you can check it
03:34out. So I really can't tell how open AI
03:36is doing revenue is
03:38exploding costs are exploding and
03:40usage is down. My take is that chat GPT
03:42is still settling into its
03:44baseline. Since it was such a
03:46revolutionary product, people are still
03:47figuring out how to
03:48integrate it into their lives. And to
03:50continue growing its revenue open AI has
03:52launched chat GPT enterprise.
03:55I can tell you firsthand from
03:56conversations I've had with my clients
03:58that privacy and security are
03:59the number one concern amongst
04:01businesses when considering chat GPT.
04:04Companies don't want to give
04:05sensitive data over to chat
04:07GPT to help train their models for later
04:09to be found in responses by
04:11that AI by other companies
04:12effectively leaking company secrets. Now
04:15with chat GPT enterprise,
04:17that concern has been more
04:18less quelled features from chat GPT
04:21enterprise include that
04:22customer props and customer data
04:24are not used for training open AI models
04:26data encryption at rest
04:27and in transit and their
04:30certified SOC 2 compliant. They also
04:32offer several highly
04:33requested features including an
04:35admin console, single sign on unlimited
04:38usage of GPT for increased speeds and
04:41larger context sizes.
04:43Chat GPT enterprise is a highly
04:45compelling product, a strong
04:46offering in the face of growing
04:48competition from the open source model
04:50world. With the guarantee
04:51of privacy and security,
04:53I can now recommend chat GPT as a real
04:55option amongst the open
04:56source models to companies that
04:58ask me which model they should use for
04:59their business. But this
05:00wouldn't be AI news if meta AI
05:02didn't launch something incredible and
05:05open source. At the end of last week,
05:08meta launched CodeLama,
05:09a fine tuned version of llama to
05:11explicitly trained for
05:12coding tasks. Shortly after that,
05:15multiple fine tuned versions of CodeLama
05:17were released that beat
05:18GPT for at coding problems.
05:20Yes, you heard me right, beat not just
05:24chat GPT, but GPT for
05:26also. This is an incredible
05:28accomplishment given I didn't think GPT
05:30for would have any competition in the
05:32coding realm anytime
05:33soon. GPT for has been my go to coding
05:35assistance since it was
05:36launched. But now I have a completely
05:38free and open source alternative. Not
05:40only that, but quantized versions with
05:42sizes ranging from a
05:43billion parameters all the way up to 70
05:45billion allow for pretty much any
05:47hardware to run these
05:48models. There's even a full unquantized
05:5134 billion parameter version
05:53running at over 20 tokens per
05:55second on an M2 ultra Mac. Be sure to
05:58check out the videos I made testing
06:00CodeLama versus GPT for
06:01and also the tutorial videos showing how
06:03to install CodeLama
06:04locally. Both will be in the
06:06description below. Our next story is
06:08about the constant march of AI beating
06:10humans at new things.
06:12This week, AI beat world champion drone
06:14racers. The AI system called Swift,
06:17designed by University
06:18of Zurich researchers beat the best human
06:21drone racers in the world. A feat
06:23considered impossible
06:24just a few years ago. Drone racing is a
06:26popular sport where
06:27racers navigate drones through
06:28complicated courses at speeds exceeding
06:31100 kilometers per hour,
06:32controlling them remotely
06:33through a VR like headset connected to an
06:35onboard camera. Training
06:37for this AI occurred in a
06:38simulated environment. And then the race
06:40occurred on an actual course. The AI
06:43control drone was able
06:44to beat the world record by a half of a
06:46second, which doesn't seem
06:47like much. But in the world of
06:48drone racing, everything is measured in
06:50fractions of a second. This
06:52accomplishment isn't just for
06:54fun. It actually has a lot of real world
06:56applications such as
06:58environmental monitoring,
07:00disaster reporting and rescue operations.
07:02What do you think will be
07:03the next thing that AI beats
07:04humans at? Let me know in the comments.
07:06Our next story is one that
07:08I'm very happy to be talking
07:09about. A16Z, the famed venture capital
07:12firm out of Silicon Valley
07:14seems to end up in my news
07:15videos almost every week now. This week
07:18they announced a grant
07:18program where they're giving
07:19away funding to a small group of AI
07:22developers to help open source
07:23community. Creating artificial
07:24intelligence is extremely expensive given
07:27the hardware
07:27requirements. Just look at the $300
07:30million AI cluster that Tesla just
07:32launched. The open source
07:33community gives their software away
07:35for free. So acquiring the expensive
07:36hardware to create and run
07:38open source models is nearly
07:39impossible. Now A16Z will be giving
07:42grants to some of the community's most
07:43prominent open source AI
07:45developers. Tom Jobins, also known as the
07:47bloke who I mentioned all
07:48the time, was one of the
07:50initial recipients of the grant. And I'm
07:52very happy to see this
07:53because I use his quantized models
07:55all the time. Congrats to all the grant
07:57recipients and a big thank
07:58you to A16Z for helping bolster
08:00the open source community. Next, not to
08:03be left out of the AI wave,
08:04Google made many announcements
08:05this week. First with its launch of Duet
08:07AI and Google Workspaces.
08:09This is a massive launch
08:11because Google Workspaces has 3 billion
08:13users. That number blew me
08:15away. I really didn't understand
08:16how that's possible and that's on par
08:18with Facebook. I don't
08:20know how they calculate those
08:21users but I guess it includes every Gmail
08:23user. Now Google Workspace
08:25users can access Duet AI,
08:27which Google describes as a powerful
08:29collaboration partner that
08:30can act as a coach, a source of
08:32inspiration, and productivity booster.
08:35You'll find Duet features in
08:36almost every product within the
08:38Google Workspace suite of products. And
08:40not only that, Google
08:42unveiled several new AI tools and
08:44capabilities at the Google Next
08:45conference in San Francisco.
08:47Let's take a look at some of the
08:48launches. Google's cloud service now
08:50includes 20 pre-built AI
08:52models optimized for enterprises
08:54like LAMATU and CLOD2. They also launched
08:57their new AI
08:57watermarking product Synth ID,
09:00which helps people identify AI-generated
09:02images created by
09:03their AI Gen art product,
09:05Imagine. The watermark is undetectable by
09:08the human eye but also persists even
09:10after modifications to
09:11the image are made, like filters, color
09:13changes, and brightness adjustments.
09:15Google also launched
09:16access to their new AI training cluster
09:18based on their custom-built TPU
09:20architecture, which can be
09:21used to train and fine-tune AI models.
09:24Last, Google updated its Vertex AI
09:26platform with upgrades to
09:27POM2, enhanced code generation, and new
09:30search and conversational models. Even
09:32with these launches,
09:33it still does feel like Google is playing
09:35catch-up to meta, open AI, and Microsoft.
09:37Now let's switch gears for a minute. In
09:39tech news, it's been reported that
09:41Silicon Valley Elite are
09:42building a city from scratch. According
09:44to the article in Marin Independent
09:46Journal, billionaire
09:47VC Michael Moritz and others had dreams
09:50of transforming tens
09:51of thousands of acres
09:52into a bustling metropolis that,
09:54according to the pitch,
09:55could generate thousands of jobs
09:57and be as walkable as Paris or the West
09:59Village in New York. He
10:00painted a kind of urban blank slate
10:03where everything from design to
10:05construction methods and
10:06new forms of governance can be
10:07rethought. Since the initial idea, large
10:10plots of land have been
10:11purchased and $800 million
10:13has been committed to the project from
10:16tech elites. These
10:17secretive land purchases have
10:19locals worried, though, unsure what will
10:21become of their quiet
10:22towns. Some of the investors that
10:24have been identified include Reid
10:26Hoffman, the founder of LinkedIn, Mark
10:28Andreessen of Andreessen
10:29Horowitz A16Z, Chris Dixon, Patrick and
10:32John Collison, who are
10:33the founders of Stripe,
10:35Lauren Powell Jobs, Steve Jobs' wife, and
10:37more. And this isn't the
10:38first time tech entrepreneurs
10:39have tried to affect California's
10:41significant and ongoing housing crisis.
10:43As someone who lives in
10:44California, anything to bring down the
10:46cost of living is something
10:47I'm all for. So I hope they
10:49build something incredible. Next, look
10:51out mid-journey. Another
10:53competitor is On the Horizon.
10:55ideogram this week launched in beta with
10:57a unique differentiator,
10:58being able to add text to AI
11:00generated images. Text in AI images has
11:03been a difficult problem
11:04to solve, but ideogram seems
11:06to have successfully solved it. Founded
11:08by ex-Google brain
11:09researchers, ideogram received a massive
11:12$16.5 million in funding from powerhouse
11:15investors like A16Z and Index
11:18Ventures. I don't know if just
11:20being able to add text within AI
11:22generated images is going to be enough to
11:24set them apart in such a
11:25crowded field, especially since
11:27competition is likely to add this
11:29functionality soon enough.
11:30Still, I wish them luck and the more
11:32competition, the better for consumers.
11:35Speaking of generative art, Runways Gen 2
11:37had another big release this week called
11:39Motion Slider. This feature allows you to
11:42select the number from 1 to
11:4310 to control the amount of
11:44movement in your output video. Take a
11:46look at this example. It seems like each
11:49week text to video is
11:50becoming better. Next, Apple may be well
11:53positioned to win the
11:54hardware game for AI. As it is
11:56increasingly difficult to get your hands
11:58on Nvidia GPUs, it turns
12:00out that Apple's own silicon,
12:01the M1 and the M2 are incredibly good at
12:04running AI models. In a
12:06lengthy tweet by AI pioneer Andre
12:08Carpathi, he details why the M2 chip is a
12:11great option for running
12:12large language models. And,
12:14as mentioned earlier, ex-user Georgi
12:17Gurgunov showed a video of himself
12:18running an unquantized
12:2034 billion parameter version of CodeLama
12:22at 20 tokens per second
12:24on an M2 Ultra. So, all you
12:26really need to run incredibly powerful
12:28large language models is
12:29an Apple computer. But,
12:31you may not even need a computer.
12:33According to the Stability AI
12:34founder, he believes we'll see
12:36a chat GPT level model on a mobile phone
12:39next year with a GPT4 level
12:42model the year after that.
12:43This is incredible news for the open
12:44source AI community and hints at what
12:46could be coming from
12:47the iPhone maker. Your move, Tim Apple.
12:49Now, for the AI video of the week. In
12:52what is sure to scare
12:53the pants off of Disney, ex-user Jeff
12:56Synthesize created a two and a half
12:58minute long AI-generated
13:00Pixar-like film using Mid-Journey and Gen
13:022 called Glitch. The
13:04video looks absolutely
13:05incredible and could have easily been
13:07created by Pixar. But instead it was
13:09created by one person,
13:10a very hard-working AI artist. Take a
13:12look at this 20
13:13second clip from the film.
13:37Generally, films like this take dozens if
13:41not hundreds of people to create. So,
13:43the implications for Disney are
13:44tremendous. Amid an ongoing
13:46writer strike and declining stock
13:48performance, I imagine Disney is looking
13:51very closely at AI
13:52technology to help them reduce
13:54their costs of creating amazing films. If
13:56you want to submit an
13:57entry for AI video of the week,
13:59jump in my Discord and find the AI video
14:01of the week channel. I'll
14:02link it in the description
14:03below. Our last story is about AI and
14:06copyright. As regulators
14:08race to figure out how to handle
14:09the avalanche of AI content being
14:11generated, they are now
14:13asking for input from the public
14:14in determining how to create AI copyright
14:16policy. The US Copyright
14:18Office has opened for public
14:20comment to figure out how to answer three
14:22main questions. How AI
14:23models should use copyrighted
14:25data in training, whether AI generated
14:27material can be copyrighted
14:29even without a human involved,
14:31and how copyright liability will work
14:33with AI. Just last week I
14:35reported that it was ruled
14:36AI art can't be copyrighted, but it seems
14:38that decision isn't the
14:39last word on the subject.
14:41And also last week I reported on major
14:43lawsuits filed against OpenAI for
14:45allegedly training their
14:46models on copyrighted data. It'll be
14:47interesting to see how all of
14:49these legal elements of AI play
14:50out and I'll keep you up to date all
14:52along the way. If you liked
14:54this video, please consider
14:55giving me a like and subscribe and I'll
14:57see you in the next one.