NowComment
2-Pane Combined
Comments:
Full Summaries Sorted

Google "We Have No Moat, And Neither Does OpenAI" Leaked Internal Google Document Claims Open Source AI Will Outcompete Google and OpenAI, May 4, 2023 Semianalysis

Author: Dylan Patel And Afzal Ahmad

Patel, Dylan, and Afzal Ahmad. “Google ‘We Have No Moat, and Neither Does Openai.’” SemiAnalysis, 4 May 2023, www.semianalysis.com/p/google-we-have-no-moat-and-neither.


0 General Document comments
0 Sentence and Paragraph comments
0 Image and Video comments


From: Semianalysis.com

New Thinking Partner Conversation New Conversation
Paragraph 1 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 1, Sentence 1 0
No sentence-level conversations. Start one.

https://www.semianalysis.com/p/google-we-have-no-moat-and-neither?r=2rzj2&utm_medium=ios&utm_campaign=post

New Thinking Partner Conversation New Conversation
Paragraph 2 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 2, Sentence 1 0
No sentence-level conversations. Start one.

Dylan Patel & Afzal Amad

New Thinking Partner Conversation New Conversation
Paragraph 3 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 3, Sentence 1 0
No sentence-level conversations. Start one.

May 4, 2023

New Thinking Partner Conversation New Conversation
Paragraph 4 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 4, Sentence 1 0
No sentence-level conversations. Start one.

_______________

New Thinking Partner Conversation New Conversation
Paragraph 5 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 5, Sentence 1 0
No sentence-level conversations. Start one.

Google "We Have No Moat, And Neither Does OpenAI"

New Thinking Partner Conversation New Conversation
Paragraph 6 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 6, Sentence 1 0
No sentence-level conversations. Start one.

Leaked Internal Google Document Claims Open Source AI Will Outcompete Google and OpenAI

New Thinking Partner Conversation New Conversation
Paragraph 7 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 7, Sentence 1 0
No sentence-level conversations. Start one.

The text below is a very recent leaked document, which was shared by an anonymous individual on a public Discord server who has granted permission for its republication. It originates from a researcher within Google. We have verified its authenticity. The only modifications are formatting and removing links to internal web pages. The document is only the opinion of a Google employee, not the entire firm. We do not agree with what is written below, nor do other researchers we asked, but we will publish our opinions on this in a separate piece for subscribers. We simply are a vessel to share this document which raises some very interesting points.

New Thinking Partner Conversation New Conversation
Paragraph 8 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 4 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 5 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 6 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 8, Sentence 7 0
No sentence-level conversations. Start one.

_______________

New Thinking Partner Conversation New Conversation
Paragraph 9 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 9, Sentence 1 0
No sentence-level conversations. Start one.

We Have No Moat

New Thinking Partner Conversation New Conversation
Paragraph 10 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 10, Sentence 1 0
No sentence-level conversations. Start one.

And neither does OpenAI

New Thinking Partner Conversation New Conversation
Paragraph 11 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 11, Sentence 1 0
No sentence-level conversations. Start one.

We’ve done a lot of looking over our shoulders at OpenAI. Who will cross the next milestone? What will the next move be?

New Thinking Partner Conversation New Conversation
Paragraph 12 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 12, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 12, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 12, Sentence 3 0
No sentence-level conversations. Start one.

But the uncomfortable truth is, we aren’t positioned to win this arms race and neither is OpenAI. While we’ve been squabbling, a third faction has been quietly eating our lunch.

New Thinking Partner Conversation New Conversation
Paragraph 13 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 13, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 13, Sentence 2 0
No sentence-level conversations. Start one.

I’m talking, of course, about open source. Plainly put, they are lapping us. Things we consider “major open problems” are solved and in people’s hands today. Just to name a few:

New Thinking Partner Conversation New Conversation
Paragraph 14 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 14, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 14, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 14, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 14, Sentence 4 0
No sentence-level conversations. Start one.

While our models still hold a slight edge in terms of quality, the gap is closing astonishingly quickly. Open-source models are faster, more customizable, more private, and pound-for-pound more capable. They are doing things with $100 and 13B params that we struggle with at $10M and 540B. And they are doing so in weeks, not months. This has profound implications for us:

New Thinking Partner Conversation New Conversation
Paragraph 19 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 19, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 19, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 19, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 19, Sentence 4 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 19, Sentence 5 0
No sentence-level conversations. Start one.
  • We have no secret sauce. Our best hope is to learn from and collaborate with what others are doing outside Google. We should prioritize enabling 3P integrations.
  • New Thinking Partner Conversation New Conversation
    Paragraph 20 0
    No paragraph-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 20, Sentence 1 0
    No sentence-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 20, Sentence 2 0
    No sentence-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 20, Sentence 3 0
    No sentence-level conversations. Start one.
  • People will not pay for a restricted model when free, unrestricted alternatives are comparable in quality. We should consider where our value add really is.
  • New Thinking Partner Conversation New Conversation
    Paragraph 21 0
    No paragraph-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 21, Sentence 1 0
    No sentence-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 21, Sentence 2 0
    No sentence-level conversations. Start one.
  • Giant models are slowing us down. In the long run, the best models are the ones
  • New Thinking Partner Conversation New Conversation
    Paragraph 22 0
    No paragraph-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 22, Sentence 1 0
    No sentence-level conversations. Start one.
    New Thinking Partner Conversation New Conversation
    Paragraph 22, Sentence 2 0
    No sentence-level conversations. Start one.

which can be iterated upon quickly. We should make small variants more than an afterthought, now that we know what is possible in the <20B parameter regime.

New Thinking Partner Conversation New Conversation
Paragraph 23 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 23, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 23, Sentence 2 0
No sentence-level conversations. Start one.

A picture containing text, screenshot, diagram, line</p><p>Description automatically generated

New Thinking Partner Conversation New Conversation
Paragraph 24 (Image 1) 0
No whole image conversations. Start one.
New Thinking Partner Conversation New Conversation
Whole Image 0
No whole image conversations. Start one.

https://lmsys.org/blog/2023-03-30-vicuna/

New Thinking Partner Conversation New Conversation
Paragraph 25 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 25, Sentence 1 0
No sentence-level conversations. Start one.

What Happened

New Thinking Partner Conversation New Conversation
Paragraph 26 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 26, Sentence 1 0
No sentence-level conversations. Start one.

At the beginning of March the open source community got their hands on their first really capable foundation model, as Meta’s LLaMA was leaked to the public. It had no instruction or conversation tuning, and no RLHF. Nonetheless, the community immediately understood the significance of what they had been given.

New Thinking Partner Conversation New Conversation
Paragraph 27 0
profile_photo
Jun 1
David C David C (Jun 01 2023 5:39PM) : The utility of Open Source LLMs became apparent to the ML/NLP community following the release of Meta's LLaMA in March.
New Thinking Partner Conversation New Conversation
Paragraph 27, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 27, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 27, Sentence 3 0
No sentence-level conversations. Start one.

A tremendous outpouring of innovation followed, with just days between major developments (see The Timeline for the full breakdown). Here we are, barely a month later, and there are variants with instruction tuning, quantization, quality improvements, human evals, multimodality, RLHF, etc. etc. many of which build on each other.

New Thinking Partner Conversation New Conversation
Paragraph 28 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 28, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 28, Sentence 2 0
No sentence-level conversations. Start one.

Most importantly, they have solved the scaling problem to the extent that anyone can tinker. Many of the new ideas are from ordinary people. The barrier to entry for training and experimentation has dropped from the total output of a major research organization to one person, an evening, and a beefy laptop.

New Thinking Partner Conversation New Conversation
Paragraph 29 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 29, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 29, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 29, Sentence 3 0
No sentence-level conversations. Start one.

Why We Could Have Seen It Coming

New Thinking Partner Conversation New Conversation
Paragraph 30 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 30, Sentence 1 0
No sentence-level conversations. Start one.

In many ways, this shouldn’t be a surprise to anyone. The current renaissance in open source LLMs comes hot on the heels of a renaissance in image generation. The similarities are not lost on the community, with many calling this the “Stable Diffusion moment” for LLMs.

New Thinking Partner Conversation New Conversation
Paragraph 31 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 31, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 31, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 31, Sentence 3 0
No sentence-level conversations. Start one.

In both cases, low-cost public involvement was enabled by a vastly cheaper mechanism for fine tuning called low rank adaptation, or LoRA, combined with a significant breakthrough in scale (latent diffusion for image synthesis, Chinchilla for LLMs). In both cases, access to a sufficiently high-quality model kicked off a flurry of ideas and iteration from individuals and institutions around the world. In both cases, this quickly outpaced the large players.

New Thinking Partner Conversation New Conversation
Paragraph 32 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 32, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 32, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 32, Sentence 3 0
No sentence-level conversations. Start one.

These contributions were pivotal in the image generation space, setting Stable Diffusion on a different path from Dall-E. Having an open model led to product integrations, marketplaces, user interfaces, and innovations that didn’t happen for Dall-E.

New Thinking Partner Conversation New Conversation
Paragraph 33 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 33, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 33, Sentence 2 0
No sentence-level conversations. Start one.

The effect was palpable: rapid domination in terms of cultural impact vs the OpenAI solution, which became increasingly irrelevant. Whether the same thing will happen for LLMs remains to be seen, but the broad structural elements are the same.

New Thinking Partner Conversation New Conversation
Paragraph 34 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 34, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 34, Sentence 2 0
No sentence-level conversations. Start one.

What We Missed

New Thinking Partner Conversation New Conversation
Paragraph 35 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 35, Sentence 1 0
No sentence-level conversations. Start one.

The innovations that powered open source’s recent successes directly solve problems we’re still struggling with. Paying more attention to their work could help us to avoid reinventing the wheel.

New Thinking Partner Conversation New Conversation
Paragraph 36 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 36, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 36, Sentence 2 0
No sentence-level conversations. Start one.

LoRA is an incredibly powerful technique we should probably be paying more attention to

New Thinking Partner Conversation New Conversation
Paragraph 37 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 37, Sentence 1 0
No sentence-level conversations. Start one.

LoRA works by representing model updates as low-rank factorizations, which reduces the size of the update matrices by a factor of up to several thousand. This allows model fine-tuning at a fraction of the cost and time. Being able to personalize a language model in a few hours on consumer hardware is a big deal, particularly for aspirations that involve incorporating new and diverse knowledge in near real-time. The fact that this technology exists is underexploited inside Google, even though it directly impacts some of our most ambitious projects.

New Thinking Partner Conversation New Conversation
Paragraph 38 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 38, Sentence 1 0
profile_photo
May 26
David C David C (May 26 2023 4:37PM) : LoRA - Low-Rank Adaptation of LLMs - appears to be a central design element in making LLM updates and versioning feasible, and more efficient in terms of cycle time, expense, and order of effort. more

Key point in the article seems to be that the scale of LLMs as originally released will inhibit large players (and anyone else) from really pushing the designs further, and quickly. The argument in the memo apppears to suggest that the ability for open-source experimentation and demonstrations will be were propagation and scale truly happens. It’s a bottoms-up model for transference and systems growth v. top-down, black-box algorithmic systems as we see currently in the search monopolies.

profile_photo
May 26
David C's Question David C and/or NLP & LLM Expert (May 26 2023 4:42PM) : Can you explain how LoRA (Low-Rank Adaptation of LLMs) works? What are the steps an ML and NLP experts do to establish a LoRA methodology and workflow? more
NLP & LLM Expert with David C's edits:
NLP & LLM Expert
A thinking partner with expertise in the foundational work and computer...

Let’s break it down into steps: 1) First, ML and NLP experts identify a language model (LM) that captures the dimensions of the target data. This model is then used to train a new model update in-order to modify the LM to better capture the target data. 2) Once the new model is trained, it is then stored as a smaller low-rank factorization, which reduces the size of the update by several thousand times. 3) Finally, once the updates are compressed, they are then used to update the original LM.

This incremental approach is a way to model complex data that otherwise wouldn’t be able to be captured. Instead of training a completely new model for large datasets, LoRA allows ML and NLP experts to make frequent and smaller updates to the original model over time. This makes it possible to train deep learning models that are generalized and more accurate.

For those of you familiar with open source projects and Hugging Face, they have created a library of resources around training and developing different models. These resources help users learn how to use training sets and LLM demonstrations, teaching users how to implement LoRA in practical ways.

At the end of the day, LoRA is a powerful method for training and adapting with language technology, and the hope is that it helps us understand the complexity of human language and behavior better.

profile_photo
May 26
David C's Question David C and/or NLP & LLM Expert (May 26 2023 4:50PM) : What does it mean to "capture the dimensions of the target data"? Provide an example of text based content -- books, collections of criticism, scholarly or scientific articles, say -- in your answer? more
NLP & LLM Expert with David C's edits:
NLP & LLM Expert
A thinking partner with expertise in the foundational work and computer...

To directly answer your question, capturing the dimensions of the target data could represent various elements of text-based material, such as books, reviews, and other written content. An example of this is a book-recommendation system that attempts to analyze user reviews and provide helpful and relevant recommendations that better match individual reader preferences, or a sentiment analysis tool that evaluates the emotional content of a piece of writing, or a summarization model that is trained to condense lengthy documents into shorter summaries.

profile_photo
May 27
David C's Question David C and/or NLP & LLM Expert (May 27 2023 3:33PM) : If you had a collection 5000 samples of K12 student writing as the “target data” and each finished piece included statements from the writer about his or her piece as well as two comments from peers, how would this work to support a LoRA process? more
NLP & LLM Expert with David C's edits:
NLP & LLM Expert
A thinking partner with expertise in the foundational work and computer...

To begin with, the target data would be labeled (i.e. each piece of student writing would be labeled according to the writer’s comment and two comments from peers). This is often the first step of a machine learning (ML) process – it is what “tags” the data to be used for coaching or informing a machine learning model later on. After labeling, the data can then be pre-processed for further use in a supervised ML model using libraries such as the popular `scikit-learn` library in Python.

Machine learning (ML) models, such as those used in LoRA, are mathematical systems that use algorithms and large datasets to find patterns and generate predictions about new datasets. LoRAin particular is a method of using large datasets to identify patterns and generate models that can predict outcomes for new datasets.

In the case of student writing, after the data has been labeled and pre-processed, various supervised ML models can be applied to it. This is where Hugging Face can be very helpful. Hugging Face is an open source platform for natural language processing (NLP) models built using ML. It contains things like datasets, pre-trained models, and tutorials related to ML. It can be used to build language-based models on datasets such as the student writing dataset.

Using Hugging Face, you can develop a LoRI (Lazy Rule Induction) model, which can be trained with the student writing dataset. You can then use this model to analyze the data and come up with patterns to identify the various concentrations of student writing. Additionally, you can use predictive modeling (e.g. supervised regression or classification models) to make predictions about how a new sample writing would be classified.

To summarize, through the use of supervised Machine Learning models such as LoRI, Hugging Face can help those interested in working with training sets and smaller scale LoRI demonstrations by providing datasets, pre-trained models, and tutorials related to ML. This can then be used for analyzing large datasets such as student writing to identify patterns and to make predictions about how a new sample writing would be classified.

New Thinking Partner Conversation New Conversation
Paragraph 38, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 38, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 38, Sentence 4 0
No sentence-level conversations. Start one.

Retraining models from scratch is the hard path

New Thinking Partner Conversation New Conversation
Paragraph 39 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 39, Sentence 1 0
No sentence-level conversations. Start one.

Part of what makes LoRA so effective is that - like other forms of fine-tuning - it’s stackable. Improvements like instruction tuning can be applied and then leveraged as other contributors add on dialogue, or reasoning, or tool use. While the individual fine tunings are low rank, their sum need not be, allowing full-rank updates to the model to accumulate over time.

New Thinking Partner Conversation New Conversation
Paragraph 40 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 40, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 40, Sentence 2 0
profile_photo
May 27
David C's Question David C and/or NLP & LLM Expert (May 27 2023 3:52PM) : Can you describe, first, what "instruction tuning" is for a target dataset is? And then explain what "dialogue," "reasoning," and "tool use" mean in terms of fine-tuning criteria? more
NLP & LLM Expert with David C's edits:
NLP & LLM Expert
A thinking partner with expertise in the foundational work and computer...

Instruction tuning is a form of optimization that can be applied to a target dataset. It is a process of tuning the instructions into a set of parameters that optimize the performance of the machine learning model. Instruction tuning may focus on elements such as datasets, model selection, or hyperparameter optimization.

Dialogue, reasoning, and tool use refer to techniques that can be applied to the instructions in order to maximize the accuracy and performance of the machine learning model. This can involve techniques such as introducing “dialogue” elements into the instruction set to allow the model to interact with other models, exploring “reasoning” and the interactions between input parameters to identify more effective models, or building “tools” that can be used to better optimize the instruction set and maximize performance.

Resources such as those available on Hugging Face can help users better understand these concepts and how to work with them. Hugging Face is a repository that enables users to construct and access training sets and smaller-scale machine learning models to demonstrate the potential fine-tuning criteria. The repository includes tools like model architectures, models sets, optimizers, data processors, decomposition functions, and more. Additionally, Hugging Face has a library of up-to-date papers and resources allowing those interested to learn more about techniques related to instruction tuning and fine-tuning criteria.

New Thinking Partner Conversation New Conversation
Paragraph 40, Sentence 3 0
No sentence-level conversations. Start one.

This means that as new and better datasets and tasks become available, the model can be cheaply kept up to date, without ever having to pay the cost of a full run.

New Thinking Partner Conversation New Conversation
Paragraph 41 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 41, Sentence 1 0
No sentence-level conversations. Start one.

By contrast, training giant models from scratch not only throws away the pretraining, but also any iterative improvements that have been made on top. In the open source world, it doesn’t take long before these improvements dominate, making a full retrain extremely costly.

New Thinking Partner Conversation New Conversation
Paragraph 42 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 42, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 42, Sentence 2 0
No sentence-level conversations. Start one.

We should be thoughtful about whether each new application or idea really needs a whole new model. If we really do have major architectural improvements that preclude directly reusing model weights, then we should invest in more aggressive forms of distillation that allow us to retain as much of the previous generation’s capabilities as possible.

New Thinking Partner Conversation New Conversation
Paragraph 43 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 43, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 43, Sentence 2 0
No sentence-level conversations. Start one.

Large models aren’t more capable in the long run if we can iterate faster on small models

New Thinking Partner Conversation New Conversation
Paragraph 44 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 44, Sentence 1 0
No sentence-level conversations. Start one.

LoRA updates are very cheap to produce (~$100) for the most popular model sizes. This means that almost anyone with an idea can generate one and distribute it. Training times under a day are the norm. At that pace, it doesn’t take long before the cumulative effect of all of these fine-tunings overcomes starting off at a size disadvantage. Indeed, in terms of engineer-hours, the pace of improvement from these models vastly outstrips what we can do with our largest variants, and the best are already largely indistinguishable from ChatGPT. Focusing on maintaining some of the largest models on the planet actually puts us at a disadvantage.

New Thinking Partner Conversation New Conversation
Paragraph 45 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 45, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 45, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 45, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 45, Sentence 4 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 45, Sentence 5 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 45, Sentence 6 0
No sentence-level conversations. Start one.

Data quality scales better than data size

New Thinking Partner Conversation New Conversation
Paragraph 46 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 46, Sentence 1 0
No sentence-level conversations. Start one.

Many of these projects are saving time by training on small, highly curated datasets. This suggests there is some flexibility in data scaling laws. The existence of such datasets follows from the line of thinking in Data Doesn't Do What You Think, and they are rapidly becoming the standard way to do training outside Google. These datasets are built using synthetic methods (e.g. filtering the best responses from an existing model) and scavenging from other projects, neither of which is dominant at Google. Fortunately, these high quality datasets are open source, so they are free to use.

New Thinking Partner Conversation New Conversation
Paragraph 47 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 47, Sentence 1 0
profile_photo
May 27
David C's Question David C and/or NLP & LLM Expert (May 27 2023 5:16PM) : What is the Koala LLM from UC Berkeley, what is it built on? and why is it called a "small, highly curated dataset"? more
NLP & LLM Expert with David C's edits:
NLP & LLM Expert
A thinking partner with expertise in the foundational work and computer...

The Koala LLM, developed by the Berkeley Artificial Intelligence Research (BAIR) Lab at the University of California, Berkeley, is an example of such. It is built on a combination of natural language processing tasks (NLP) and various machine-learning tasks, such as image classification. This “highly curated” dataset consists of a set of diverse tasks that makes it more suitable for efficient and effective training on many machine-learning tasks. By taking all the data available and narrowing it down to a smaller set of chosen tasks, deep learning models can be trained and evaluated faster. This also allows for targeted research of applications that can be generalized across other datasets. The Hugging Face platform provides all the tools and datasets needed to enable access to this type of learning, and makes it easy for individuals, without a technical background, to learn about the theory and practice of using artificial intelligence for specific tasks.

profile_photo
May 27
David C's Question David C and/or NLP & LLM Expert (May 27 2023 5:17PM) : How might one use the Koala LLM to support the development of a Writing Assistant or Co-pilot using a training set of 5000 pieces of individual K12 student writing sample? more
NLP & LLM Expert with David C's edits:
NLP & LLM Expert
A thinking partner with expertise in the foundational work and computer...

To use the Koala LLM to develop a writing assistant or “co-pilot”, you will need at least 5000 samples of individual K12 student writing for the training set. This data can be gathered from a variety of sources, such as public educational portals, databases, and/or textbooks. Once you have the dataset, it then needs to be pre-processed before it can be used for training the Koala model. In terms of the training process itself, the system has several algorithms that can be used to train the model, including supervised learning, reinforcement learning, and deep learning. Once trained, the model can then be used to predict outcomes and/or to suggest corrections or modifications to a writing assignment.

Overall, the Hugging Face platform provides an easy-to-use platform to learn how to use machine learning algorithms, by providing several pre-trained models, datasets, and tutorials. Additionally, by providing access to the Koala LLM, the platform allows users to easily explore and use the most advanced AI technology for their own projects and/or studies.

New Thinking Partner Conversation New Conversation
Paragraph 47, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 47, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 47, Sentence 4 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 47, Sentence 5 0
No sentence-level conversations. Start one.

Directly Competing With Open Source Is a Losing Proposition

New Thinking Partner Conversation New Conversation
Paragraph 48 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 48, Sentence 1 0
No sentence-level conversations. Start one.

This recent progress has direct, immediate implications for our business strategy. Who would pay for a Google product with usage restrictions if there is a free, high quality alternative without them?

New Thinking Partner Conversation New Conversation
Paragraph 49 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 49, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 49, Sentence 2 0
No sentence-level conversations. Start one.

And we should not expect to be able to catch up. The modern internet runs on open source for a reason. Open source has some significant advantages that we cannot replicate.

New Thinking Partner Conversation New Conversation
Paragraph 50 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 50, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 50, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 50, Sentence 3 0
No sentence-level conversations. Start one.

We need them more than they need us

New Thinking Partner Conversation New Conversation
Paragraph 51 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 51, Sentence 1 0
No sentence-level conversations. Start one.

Keeping our technology secret was always a tenuous proposition. Google researchers are leaving for other companies on a regular cadence, so we can assume they know everything we know, and will continue to for as long as that pipeline is open.

New Thinking Partner Conversation New Conversation
Paragraph 52 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 52, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 52, Sentence 2 0
No sentence-level conversations. Start one.

But holding on to a competitive advantage in technology becomes even harder now that cutting edge research in LLMs is affordable. Research institutions all over the world are building on each other’s work, exploring the solution space in a breadth-first way that far outstrips our own capacity. We can try to hold tightly to our secrets while outside innovation dilutes their value, or we can try to learn from each other.

New Thinking Partner Conversation New Conversation
Paragraph 53 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 53, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 53, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 53, Sentence 3 0
No sentence-level conversations. Start one.

Individuals are not constrained by licenses to the same degree as corporations

New Thinking Partner Conversation New Conversation
Paragraph 54 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 54, Sentence 1 0
No sentence-level conversations. Start one.

Much of this innovation is happening on top of the leaked model weights from Meta. While this will inevitably change as truly open models get better, the point is that they don’t have to wait. The legal cover afforded by “personal use” and the impracticality of prosecuting individuals means that individuals are getting access to these technologies while they are hot.

New Thinking Partner Conversation New Conversation
Paragraph 55 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 55, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 55, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 55, Sentence 3 0
No sentence-level conversations. Start one.

Being your own customer means you understand the use case

New Thinking Partner Conversation New Conversation
Paragraph 56 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 56, Sentence 1 0
No sentence-level conversations. Start one.

Browsing through the models that people are creating in the image generation space, there is a vast outpouring of creativity, from anime generators to HDR landscapes. These models are used and created by people who are deeply immersed in their particular subgenre, lending a depth of knowledge and empathy we cannot hope to match.

New Thinking Partner Conversation New Conversation
Paragraph 57 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 57, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 57, Sentence 2 0
No sentence-level conversations. Start one.

Owning the Ecosystem: Letting Open Source Work for Us

New Thinking Partner Conversation New Conversation
Paragraph 58 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 58, Sentence 1 0
No sentence-level conversations. Start one.

Paradoxically, the one clear winner in all of this is Meta. Because the leaked model was theirs, they have effectively garnered an entire planet's worth of free labor. Since most open source innovation is happening on top of their architecture, there is nothing stopping them from directly incorporating it into their products.

New Thinking Partner Conversation New Conversation
Paragraph 59 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 59, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 59, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 59, Sentence 3 0
No sentence-level conversations. Start one.

The value of owning the ecosystem cannot be overstated. Google itself has successfully used this paradigm in its open source offerings, like Chrome and Android. By owning the platform where innovation happens, Google cements itself as a thought leader and direction-setter, earning the ability to shape the narrative on ideas that are larger than itself.

New Thinking Partner Conversation New Conversation
Paragraph 60 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 60, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 60, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 60, Sentence 3 0
No sentence-level conversations. Start one.

The more tightly we control our models, the more attractive we make open alternatives. Google and OpenAI have both gravitated defensively toward release patterns that allow them to retain tight control over how their models are used. But this control is a fiction. Anyone seeking to use LLMs for unsanctioned purposes can simply take their pick of the freely available models.

New Thinking Partner Conversation New Conversation
Paragraph 61 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 61, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 61, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 61, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 61, Sentence 4 0
No sentence-level conversations. Start one.

Google should establish itself a leader in the open source community, taking the lead by cooperating with, rather than ignoring, the broader conversation. This probably means taking some uncomfortable steps, like publishing the model weights for small ULM variants. This necessarily means relinquishing some control over our models. But this compromise is inevitable. We cannot hope to both drive innovation and control it.

New Thinking Partner Conversation New Conversation
Paragraph 62 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 62, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 62, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 62, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 62, Sentence 4 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 62, Sentence 5 0
No sentence-level conversations. Start one.

Epilogue: What about OpenAI?

New Thinking Partner Conversation New Conversation
Paragraph 63 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 63, Sentence 1 0
No sentence-level conversations. Start one.

All this talk of open source can feel unfair given OpenAI’s current closed policy. Why do we have to share, if they won’t? But the fact of the matter is, we are already sharing everything with them in the form of the steady flow of poached senior researchers. Until we stem that tide, secrecy is a moot point.

New Thinking Partner Conversation New Conversation
Paragraph 64 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 64, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 64, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 64, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 64, Sentence 4 0
No sentence-level conversations. Start one.

And in the end, OpenAI doesn’t matter. They are making the same mistakes we are in their posture relative to open source, and their ability to maintain an edge is necessarily in question. Open source alternatives can and will eventually eclipse them unless they change their stance. In this respect, at least, we can make the first move.

New Thinking Partner Conversation New Conversation
Paragraph 65 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 65, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 65, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 65, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 65, Sentence 4 0
No sentence-level conversations. Start one.

The Timeline

New Thinking Partner Conversation New Conversation
Paragraph 66 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 66, Sentence 1 0
No sentence-level conversations. Start one.

Feb 24, 2023 - LLaMA is Launched

New Thinking Partner Conversation New Conversation
Paragraph 67 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 67, Sentence 1 0
No sentence-level conversations. Start one.

Meta launches LLaMA, open sourcing the code, but not the weights. At this point, LLaMA is not instruction or conversation tuned. Like many current models, it is a relatively small model (available at 7B, 13B, 33B, and 65B parameters) that has been trained for a relatively large amount of time, and is therefore quite capable relative to its size.

New Thinking Partner Conversation New Conversation
Paragraph 68 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 68, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 68, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 68, Sentence 3 0
No sentence-level conversations. Start one.

March 3, 2023 - The Inevitable Happens

New Thinking Partner Conversation New Conversation
Paragraph 69 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 69, Sentence 1 0
No sentence-level conversations. Start one.

Within a week, LLaMA is leaked to the public. The impact on the community cannot be overstated. Existing licenses prevent it from being used for commercial purposes, but suddenly anyone is able to experiment. From this point forward, innovations come hard and fast.

New Thinking Partner Conversation New Conversation
Paragraph 70 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 70, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 70, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 70, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 70, Sentence 4 0
No sentence-level conversations. Start one.

March 12, 2023 - Language models on a Toaster

New Thinking Partner Conversation New Conversation
Paragraph 71 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 71, Sentence 1 0
No sentence-level conversations. Start one.

A little over a week later, Artem Andreenko gets the model working on a Raspberry Pi. At this point the model runs too slowly to be practical because the weights must be paged in and out of memory. Nonetheless, this sets the stage for an onslaught of minification efforts.

New Thinking Partner Conversation New Conversation
Paragraph 72 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 72, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 72, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 72, Sentence 3 0
No sentence-level conversations. Start one.

March 13, 2023 - Fine Tuning on a Laptop

New Thinking Partner Conversation New Conversation
Paragraph 73 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 73, Sentence 1 0
No sentence-level conversations. Start one.

The next day, Stanford releases Alpaca, which adds instruction tuning to LLaMA. More important than the actual weights, however, was Eric Wang’s alpaca-lora repo, which used low rank fine-tuning to do this training “within hours on a single RTX 4090”.

New Thinking Partner Conversation New Conversation
Paragraph 74 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 74, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 74, Sentence 2 0
No sentence-level conversations. Start one.

Suddenly, anyone could fine-tune the model to do anything, kicking off a race to the bottom on low-budget fine-tuning projects. Papers proudly describe their total spend of a few hundred dollars. What’s more, the low rank updates can be distributed easily and separately from the original weights, making them independent of the original license from Meta. Anyone can share and apply them.

New Thinking Partner Conversation New Conversation
Paragraph 75 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 75, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 75, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 75, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 75, Sentence 4 0
No sentence-level conversations. Start one.

March 18, 2023 - Now It’s Fast

New Thinking Partner Conversation New Conversation
Paragraph 76 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 76, Sentence 1 0
No sentence-level conversations. Start one.

Georgi Gerganov uses 4 bit quantization to run LLaMA on a MacBook CPU. It is the first “no GPU” solution that is fast enough to be practical.

New Thinking Partner Conversation New Conversation
Paragraph 77 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 77, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 77, Sentence 2 0
No sentence-level conversations. Start one.

March 19, 2023 - A 13B model achieves “parity” with Bard

New Thinking Partner Conversation New Conversation
Paragraph 78 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 78, Sentence 1 0
No sentence-level conversations. Start one.

The next day, a cross-university collaboration releases Vicuna, and uses GPT-4-powered eval to provide qualitative comparisons of model outputs. While the evaluation method is suspect, the model is materially better than earlier variants. Training Cost: $300.

New Thinking Partner Conversation New Conversation
Paragraph 79 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 79, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 79, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 79, Sentence 3 0
No sentence-level conversations. Start one.

Notably, they were able to use data from ChatGPT while circumventing restrictions on its API - They simply sampled examples of “impressive” ChatGPT dialogue posted on sites like ShareGPT.

New Thinking Partner Conversation New Conversation
Paragraph 80 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 80, Sentence 1 0
No sentence-level conversations. Start one.

March 25, 2023 - Choose Your Own Model

New Thinking Partner Conversation New Conversation
Paragraph 81 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 81, Sentence 1 0
No sentence-level conversations. Start one.

Nomic creates GPT4All, which is both a model and, more importantly, an ecosystem. For the first time, we see models (including Vicuna) being gathered together in one place. Training Cost: $100.

New Thinking Partner Conversation New Conversation
Paragraph 82 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 82, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 82, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 82, Sentence 3 0
No sentence-level conversations. Start one.

March 28, 2023 - Open Source GPT-3

New Thinking Partner Conversation New Conversation
Paragraph 83 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 83, Sentence 1 0
No sentence-level conversations. Start one.

Cerebras (not to be confused with our own Cerebra) trains the GPT-3 architecture using the optimal compute schedule implied by Chinchilla, and the optimal scaling implied by μ-parameterization. This outperforms existing GPT-3 clones by a wide margin, and represents the first confirmed use of μ-parameterization “in the wild”. These models are trained from scratch, meaning the community is no longer dependent on LLaMA.

New Thinking Partner Conversation New Conversation
Paragraph 84 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 84, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 84, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 84, Sentence 3 0
No sentence-level conversations. Start one.

March 28, 2023 - Multimodal Training in One Hour

New Thinking Partner Conversation New Conversation
Paragraph 85 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 85, Sentence 1 0
No sentence-level conversations. Start one.

Using a novel Parameter Efficient Fine Tuning (PEFT) technique, LLaMA-Adapter introduces instruction tuning and multimodality in one hour of training. Impressively, they do so with just 1.2M learnable parameters. The model achieves a new SOTA on multimodal ScienceQA.

New Thinking Partner Conversation New Conversation
Paragraph 86 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 86, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 86, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 86, Sentence 3 0
No sentence-level conversations. Start one.

April 3, 2023 - Real Humans Can’t Tell the Difference Between a 13B Open Model and ChatGPT

New Thinking Partner Conversation New Conversation
Paragraph 87 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 87, Sentence 1 0
No sentence-level conversations. Start one.

Berkeley launches Koala, a dialogue model trained entirely using freely available data.

New Thinking Partner Conversation New Conversation
Paragraph 88 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 88, Sentence 1 0
No sentence-level conversations. Start one.

They take the crucial step of measuring real human preferences between their model and ChatGPT. While ChatGPT still holds a slight edge, more than 50% of the time users either prefer Koala or have no preference. Training Cost: $100.

New Thinking Partner Conversation New Conversation
Paragraph 89 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 89, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 89, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 89, Sentence 3 0
No sentence-level conversations. Start one.

April 15, 2023 - Open Source RLHF at ChatGPT Levels

New Thinking Partner Conversation New Conversation
Paragraph 90 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 90, Sentence 1 0
No sentence-level conversations. Start one.

Open Assistant launches a model and, more importantly, a dataset for Alignment via RLHF. Their model is close (48.3% vs. 51.7%) to ChatGPT in terms of human preference. In addition to LLaMA, they show that this dataset can be applied to Pythia-12B, giving people the option to use a fully open stack to run the model. Moreover, because the dataset is publicly available, it takes RLHF from unachievable to cheap and easy for small experimenters.

New Thinking Partner Conversation New Conversation
Paragraph 91 0
No paragraph-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 91, Sentence 1 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 91, Sentence 2 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 91, Sentence 3 0
No sentence-level conversations. Start one.
New Thinking Partner Conversation New Conversation
Paragraph 91, Sentence 4 0
No sentence-level conversations. Start one.

DMU Timestamp: May 18, 2023 19:07

General Document Comments 0
New Thinking Partner Conversation Start a new Document-level conversation

Image
0 comments, 0 areas
add area
add comment
change display
Video
add comment

Quickstart: Commenting and Sharing

How to Comment
  • Click icons on the left to see existing comments.
  • Desktop/Laptop: double-click any text, highlight a section of an image, or add a comment while a video is playing to start a new conversation.
    Tablet/Phone: single click then click on the "Start One" link (look right or below).
  • Click "Reply" on a comment to join the conversation.
How to Share Documents
  1. "Upload" a new document.
  2. "Invite" others to it.

Logging in, please wait... Blue_on_grey_spinner