Skip to main content

22 docs tagged with "technology"

View All Tags

AI-powered drone beats human champion pilots

Artificial Intelligence (AI) has once again proven its superiority over humans, this time in the realm of drone racing. Researchers at the University of Zurich developed an AI algorithm, named Swift, that defeated three world champion drone racers. The AI managed to navigate a 3D race course at high speeds, winning 15 out of 25 races. Swift clocked the fastest lap on a course where drones reached speeds of 50mph and endured accelerations up to 5g.

Apple's Mistake

Apple's App Store approval process is severely flawed, and it's causing significant damage to their reputation among programmers. This is a stark contrast to the past when Apple was admired, almost to a fault, by its fans and developers. Now, many programmers perceive Apple as 'evil' due to the App Store's policies.

Beyond Smart

Albert Einstein is often celebrated for his intelligence, but it's his innovative ideas that truly set him apart. Intelligence is undoubtedly a prerequisite for innovation, but it's not the same thing. The distinction may seem minor, but it's significant. There are countless intelligent individuals who never make groundbreaking discoveries.

Generative AI and intellectual property — Benedict Evans

The discourse around intellectual property (IP) has been evolving for centuries, and each technological advancement brings new challenges and perspectives. Generative AI is the latest phenomenon posing intriguing questions about IP rights.

Google Gemini Eats The World – Gemini Smashes GPT-4 By 5X, The GPU-Poors

Before Covid, Google released the MEENA model, which for a short period of time, was the best large language model in the world. The blog and paper Google wrote were incredibly cute, because it specifically compared against to OpenAI. Compared to an existing state-of-the-art generative model, OpenAI GPT-2, Meena has 1.7x greater model capacity and was trained on 8.5x more data. This model required more than 14x the FLOPS of GPT-2 to train, but this was largely irrelevant because only a few months later OpenAI dropped GPT-3, which was >65x more parameters and >60x the token count, >4,000x more FLOPS. The performance difference between these two models was massive. The MEENA model sparked an internal memo written by Noam Shazeer titled "MEENA Eats The World.” In this memo, he predicted many of the things that the rest of the world woke up to after the release of ChatGPT. The key takeaways were that language models would get increasingly integrated into our lives in a variety of ways, and that they would dominate the globally deployed FLOPS. Noam was so far ahead of his time when he wrote this, but it was mostly ignored or even laughed at by key decision makers. Let’s go on a tangent about how far ahead of his time, Noam really was. He was part of the team that did the original Transformer paper, “Attention is All You Need.” He also was part of the first modern Mixture of Experts paper, Switch Transformer, Image Transformer, and various elements of LaMDA and PaLM. One of the ideas from 2018 he hasn’t yet gotten credit for more broadly is speculative decoding which we detailed here in our exclusive tell-all about GPT-4. Speculative decoding reduces the cost of low batch inference by multiple-fold. The point here is Google had all the keys to the kingdom, but they fumbled the bag. A statement that is obvious to everyone. The statement that may not be obvious is that the sleeping giant, Google has woken up, and they are iterating on a pace that will smash GPT-4 total pre-training FLOPS by 5x before the end of the year. The path is clear to 20x by the end of next year given their current infrastructure buildout. Whether Google has the stomach to put these models out publicly without neutering their creativity or their existing business model is a different discussion. Today we want to discuss Google’s training systems for Gemini, the iteration velocity for Gemini models, Google’s Viperfish (TPUv5) ramp, Google’s competitiveness going forward versus the other frontier labs, and a crowd we are dubbing the GPU-Poor. Access to compute is a bimodal distribution. There are a handful of firms with 20k+ A/H100 GPUs, and individual researchers can access 100s or 1,000s of GPUs for pet projects. The chief among these are researchers at OpenAI, Google, Anthropic, Inflection, X, and Meta, who will have the highest ratios of compute resources to researchers. A few of the firms above as well as multiple Chinese firms will 100k+ by the end of next year, although we are unsure of the ratio of researchers in China, only the GPU volumes. One of the funniest trends we see in the Bay area is with top ML researchers bragging about how many GPUs they have or will have access to soon. In fact, this has become so pervasive over the last ~4 months that it’s become a measuring contest that is directly influencing where top researchers decide to go. Meta, who will have the 2nd most number of H100 GPUs in the world, is actively using it as a recruiting tactic. Then there are a whole host of startups and open-source researchers who are struggling with far fewer GPUs. They are spending significant time and effort attempting to do things that simply don’t help, or frankly, matter. For example, many researchers are spending countless hours agonizing on fine-tuning models with GPUs that don’t have enough VRAM. This is an extremely counter-productive use of their skills and time. These startups and open-source researchers are using larger LLMs to fine-tune smaller models for leaderboard style benchmarks with broken evaluation methods that give more emphasis to style rather than accuracy or usefulness. They are generally ignorant that pretraining datasets and IFT data need to be significantly larger/higher quality for smaller open models to improve in real workloads. Yes, being efficient with GPUs is very important, but in many ways, that’s being ignored by the GPU-poors. They aren’t concerned with efficiency at scale, and their time isn’t being spent productively. What can be done commercially in their GPU-poor environment is mostly irrelevant to a world that will be flooded by more than 3.5 million H100s by the end of next year. For learning, experimenting, smaller weaker gaming GPUs are just fine. The GPU poor are still mostly using dense models because that’s what Meta graciously dropped on their lap with the LLAMA series of models. Without God’s Zuck’s good grace, most open source projects would be even worse off. If they were actually concerned with efficiency, especially on the client side, they’d be running sparse model architectures like MoE, training on these larger datasets, and implementing speculative decoding like the Frontier LLM Labs (OpenAI, Anthropic, Google Deepmind). The underdogs should be focusing on tradeoffs that improve model performance or token to token latency by upping compute and memory capacity requirements in favor of reduced memory bandwidth because that’s what the edge needs. They should be focused on efficient serving of multiple finetuned models on shared infrastructure without paying the horrendous cost penalties of small batch sizes. Instead, they continually are focused on memory capacity constraints or quantizing too far while covering their eyes about real quality decreases. To take the rant on a slight tangent, in general, model evaluation is broken. While there is a lot of effort in the closed world to improve this, the land of open benchmarks is pointless and measures almost nothing useful. For some reason there is an unhealthy obsession over the leaderboard-ification of LLMs, and meming with silly names for useless models (WizardVicunaUncensoredXPlusPlatypus). Hopefully the open efforts are redirected towards evaluations, speculative decoding, MoE, open IFT data, and clean pre-training datasets with over 10 trillion tokens, otherwise, there is no way for the open source to compete with commercial giants. While the US and China will be able to keep racing ahead, the European startups and government backed supercomputers such as Jules Verne are also completely uncompetitive. Europe will fall behind in this race due to the lack of ability to make big investments and choosing to stay GPU-poor. Even multiple Middle Eastern countries are investing more on enabling large scale infrastructure for AI. Being GPU-poor isn’t limited to only scrappy startups though. Some of the most well recognized AI firms, HuggingFace, Databricks (MosaicML), and Together are also part of this GPU-poor group. In fact, they may be the most GPU-poor groups out there with regard to both the number of world class researchers per GPU and the number of GPUs versus the ambition/potential customer demand. They have world class researchers, but all of them are limited by working on systems with orders of magnitude less capabilities. These firms have tremendous inbound from enterprises on training real models, and on the order of thousands of H100s coming in, but that won’t be enough to grab much of the market. Nvidia is eating their lunch with multiple times as many GPUs in their DGX Cloud service and various in-house supercomputers. Nvidia’s DGX Cloud offers pretrained models, frameworks for data processing, vector databases and personalization, optimized inference engines, APIs, and support from NVIDIA experts to help enterprises tune models for their custom use cases. That service has also already racked up multiple larger enterprises from verticals such as SaaS, insurance, manufacturing, pharmaceuticals, productivity software, and automotive. While not all customers are announced, even the public list of Amgen, Adobe, CCC, ServiceNow, Accenture, AstraZeneca, Getty Images, Shutterstock, Morningstar, Evozyne, Insilico Medicine, Quantiphi, InstaDeep, Oxford Nanopore, Peptone, Relation Therapeutics, ALCHEMAB Therapeutics, and Runway is quite impressive. This is a far longer list than the other players have, and Nvidia has many other undisclosed partnerships too. To be clear, revenue from these announced customers of Nvidia’s DGX cloud service is unknown, but given the size of Nvidia’s cloud spending and in-house supercomputer construction, it seems that more services can/will be purchased from Nvidia’s Cloud than HuggingFace, Together, and Databricks can hope to offer, combined. The few hundred million that HuggingFace and Together have raised collectively means they will remain GPU-poor, getting left in the dust as they will be unable to train N-1 LLMs that can serve as the base to fine tune for customers. This means they will ultimately be unable to capture high share at enterprises who can just access Nvidia’s service today anyways. HuggingFace in particular has one the biggest names in the industry, and they need to leverage that to invest a huge amount and build a lot more model, customization, and inference capabilities. Their recent round was done at too high of a valuation to garner the investment they need to compete. HuggingFace’s leaderboards show how truly blind they are because they actively hurting the open source movement by tricking it into creating a bunch of models that are useless for real usage. Databricks (MosaicML) could at least maybe catch up, due to their data and enterprise connections. The issue is they need to accelerate spend by multiple times if they want to have hopes of serving their over 7,000 customers. The $1.3B acquisition of MosaicML was a big bet on this vertical, but they also need to throw a similar amount of money at infrastructure. Unfortunately for Databricks, they can’t pay for GPUs in shares. They need to do a large offering via their upcoming private round/IPO, and use that cold hard cash to quadruple down on hardware. The economic argument falls flat on its face because they must build before customers can come, because Nvidia is throwing money at their service. To be clear, many folks are buying loads of compute not making their money back, (Cohere, Saudi Arabia, UAE), but it is a pre-requisite to compete. The picks and shovels training and inference ops firms (Databricks, HuggingFace, and Together) are behind their chief competition, who also happens to also be the source of almost all of their compute. The next largest operator of customized models is simply the fine tuning APIs from OpenAI. The key here is everyone from Meta to Microsoft to startups are simply serving as a pipeline of capital to Nvidia’s bank account. Can anyone save us from Nvidia slavery? Yes, there is one potential savior. While Google does use GPUs internally as well as a significant number sold via GCP, they a have a few Ace’s up their sleeve. These include Gemini and the next iteration which has already begun training. The most important advantage they have is their unbeatably efficient infrastructure. Before getting into Gemini and their cloud business, we will share some datapoints on their insane buildout. The chart below shows the total advanced chips added by quarter. Here we give OpenAI every benefit of the doubt. That the number of total GPUs they have will 4x over 2 years. For Google, we ignore their entire existing fleet of TPUv4 (Pufferfish), TPUv4 lite, and internally used GPUs. Furthermore, we are also not including the TPUv5e (lite), despite that likely being the workhorse for inference of smaller language models. Google’s growth in this chart is only TPUv5 (Viperfish).

Hands on with Apple Vision Pro in the wild

The long-anticipated Apple Vision Pro headset is finally here, and it's a marvel of modern design and technology. While it's not the first AR and VR headset on the market, it certainly stands out in terms of its intent, pricing, and use cases, drawing closest comparison to Microsoft's HoloLens.

Having Kids

Before I had kids, I was apprehensive about the idea of parenthood. I perceived it as an end to fun and coolness, and children seemed like little terrors to my childless self. However, having children of my own has completely transformed my perspective. The instant emotional bond and protective instincts that kicked in were overwhelming and unexpected.

How Jensen Huang’s Nvidia Is Powering the A.I. Revolution

The revelation that ChatGPT, the astonishing artificial-intelligence chatbot, had been trained on an Nvidia supercomputer spurred one of the largest single-day gains in stock-market history. When the Nasdaq opened on May 25, 2023, Nvidia’s value increased by about two hundred billion dollars. By the close of trading, Nvidia was the sixth most valuable corporation on earth, worth more than Walmart and ExxonMobil combined.

How the iMac saved Apple

The iMac was a transformative force in a stagnant computing world of the mid-1990s. When Steve Jobs returned to a struggling Apple in 1997, he, along with designer Jony Ive, created a plan to shake up the industry. The iMac was a bold move, contradicting the PC industry's norms with its self-contained unit and vibrant, translucent blue-green plastic design.

How to Be an Expert in a Changing World

If the world were static, we could have monotonically increasing confidence in our beliefs. However, this is not the case, especially for things that change, which could include practically everything. When experts are wrong, it's often because they're experts on an earlier version of the world. To avoid obsolete beliefs, one must actively protect against them. As a startup investor, I've learned that most really good startup ideas initially look like bad ideas because some change in the world just made them viable.

How You Know

Despite having read Villehardouin's chronicle of the Fourth Crusade multiple times, I realize that my recollection of its contents is minimal. This observation extends to the hundreds of books on my shelves, raising the question: What is the value of reading if so little is remembered?

Introducing Chat Notebooks: Integrating LLMs into the Notebook Paradigm

We originally invented the concept of “Notebooks” back in 1987, for Version 1.0 of Mathematica. Over the past 36 years, Notebooks have proved to be an incredibly convenient medium in which to do—and publish—work. Now, there’s a new challenge and opportunity for Notebooks: integrating LLM functionality into them. Today we’re introducing Chat Notebooks as a new kind of Notebook that supports LLM-based chat functionality. This will be built into the upcoming version of Wolfram Language (Version 13.3).

What I've Learned from Users

I recently advised Y Combinator applicants that the best advice for getting in was to explain what they've learned from users. This advice tests if they’re paying attention to users, understanding them, and recognizing the necessity of their product. Reflecting on what I've learned from YC's startups, the recurrence of similar problems across different startups stands out. Advising numerous startups reveals common issues, which is a key factor in YC's effectiveness.

What OpenAI Really Wants

OpenAI, co-founded by Sam Altman, is on a mission to build artificial general intelligence (AGI) that is safe for humanity. The company's journey has been marked by significant breakthroughs, challenges, and transformations.

Why Tim Cook Is Going All In on the Apple Vision Pro

The first time Tim Cook experienced the Apple Vision Pro, it wasn’t called that. It was years ago, before Apple Park was built. Cook recalls seeing the prototype at Mariani 1, a secretive Apple facility. This monstrous machine, with multiple screens and cameras, transported him to the moon, making him realize the potential of this technology.