Skip to content

Commit 2d74c93

Browse files
julien-cmishig25
andauthored
No more magic comments (#1554)
* no more magic comments * Also replace h1 by actual markdown * nit: remove extra space * Fix remaining <h1>s * handling complex h1 heading * Update README.md cc @mishig25 --------- Co-authored-by: Mishig Davaadorj <[email protected]>
1 parent ef9aed0 commit 2d74c93

File tree

384 files changed

+115
-879
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

384 files changed

+115
-879
lines changed

1b-sentence-embeddings.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Train a Sentence Embedding Model with 1 Billion Training Pairs
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
**Sentence embedding** is a method that maps sentences to vectors of real numbers. Ideally, these vectors would capture the semantic of a sentence and be highly generic. Such representations could then be used for many downstream applications such as clustering, text mining, or question answering.
1412

3d-assets.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Practical 3D Asset Generation: A Step-by-Step Guide
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
## Introduction
1412

4bit-transformers-bitsandbytes.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,6 @@ authors:
1313

1414
# Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA
1515

16-
<!-- {blog_metadata} -->
17-
<!-- {authors} -->
1816

1917
LLMs are known to be large, and running or training them in consumer hardware is a huge challenge for users and accessibility.
2018
Our [LLM.int8 blogpost](https://huggingface.co/blog/hf-bitsandbytes-integration) showed how the techniques in the [LLM.int8 paper](https://arxiv.org/abs/2208.07339) were integrated in transformers using the `bitsandbytes` library.

Llama2-for-non-engineers.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,8 +6,6 @@ authors:
66
- user: abhishek
77
---
88

9-
<!-- {blog_metadata} -->
10-
<!-- {authors} -->
119

1210
# Non-engineers guide: Train a LLaMA 2 chatbot
1311

README.md

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -29,19 +29,15 @@ authors:
2929
3030
# Train your first Decision Transformer
3131
32-
<!-- {blog_metadata} -->
33-
<!-- {authors} -->
34-
3532
Your content here [...]
3633
```
3734

38-
The blog_metadata and authors HTML comments are meant to mark where in the file will be inserted the following UI elements:
35+
When published, the Hub will insert the following UI elements right after the blogpost's main header (i.e. the line that starts with a single `#`, aka. the `<h1>`):
36+
3937
- "Published on [date]"
4038
- "Update on GitHub" button
4139
- avatars of the authors that were listed in authors.
4240

43-
⚠️ Please keep the blog_metadata and authors comments exactly equal to those strings otherwise they won't be replaced.
44-
4541
5️⃣ Then, you can add your content. It's markdown system so if you wrote your text on notion just control shift v to copy/paste as markdown.
4642

4743
6️⃣ Modify `_blog.yml` to add your blogpost.

accelerate-deepspeed.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,10 +6,8 @@ authors:
66
- user: sgugger
77
---
88

9-
<h1>Accelerate Large Model Training using DeepSpeed</h1>
9+
# Accelerate Large Model Training using DeepSpeed
1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
In this post we will look at how we can leverage the **[Accelerate](https://github.com/huggingface/accelerate)** library for training large models which enables users to leverage the ZeRO features of **[DeeSpeed](https://www.deepspeed.ai)**.
1513

accelerate-large-models.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# How 🤗 Accelerate runs very large models thanks to PyTorch
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
## Load and run large models
1412

accelerate-library.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Introducing 🤗 Accelerate
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
## 🤗 Accelerate
1412

accelerate-transformers-with-inferentia2.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,6 @@ authors:
88

99
# Accelerating Hugging Face Transformers with AWS Inferentia2
1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1513

accelerated-inference.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,9 +3,8 @@ title: "How we sped up transformer inference 100x for 🤗 API customers"
33
thumbnail: /blog/assets/09_accelerated_inference/thumbnail.png
44
---
55

6-
<h1>How we sped up transformer inference 100x for 🤗 API customers</h1>
6+
# How we sped up transformer inference 100x for 🤗 API customers
77

8-
<!-- {blog_metadata} -->
98

109
🤗 Transformers has become the default library for data scientists all around the world to explore state of the art NLP models and build new NLP features. With over 5,000 pre-trained and fine-tuned models available, in over 250 languages, it is a rich playground, easily accessible whichever framework you are working in.
1110

accelerating-pytorch.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,6 @@ authors:
88
# Accelerating PyTorch distributed fine-tuning with Intel technologies
99

1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
For all their amazing performance, state of the art deep learning models often take a long time to train. In order to speed up training jobs, engineering teams rely on distributed training, a divide-and-conquer technique where clustered servers each keep a copy of the model, train it on a subset of the training set, and exchange results to converge to a final model.
1513

agents-js.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Introducing Agents.js: Give tools to your LLMs using JavaScript
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
We have recently been working on Agents.js at [huggingface.js](https://github.com/huggingface/huggingface.js/blob/main/packages/agents/README.md). It's a new library for giving tool access to LLMs from JavaScript in either the browser or the server. It ships with a few multi-modal tools out of the box and can easily be extended with your own tools and language models.
1412

ai-comic-factory.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Deploying the AI Comic Factory using the Inference API
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
We recently announced [Inference for PROs](https://huggingface.co/blog/inference-pro), our new offering that makes larger models accessible to a broader audience. This opportunity opens up new possibilities for running end-user applications using Hugging Face as a platform.
1412

ai-residency.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Announcing the 🤗 AI Research Residency Program 🎉 🎉 🎉
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311

1412
The 🤗 Research Residency Program is a 9-month opportunity to launch or advance your career in machine learning research 🚀. The goal of the residency is to help you grow into an impactful AI researcher. Residents will work alongside Researchers from our Science Team. Together, you will pick a research problem and then develop new machine learning techniques to solve it in an open & collaborative way, with the hope of ultimately publishing your work and making it visible to a wide audience.

ai-webtv.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Building an AI WebTV
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
The AI WebTV is an experimental demo to showcase the latest advancements in automatic video and music synthesis.
1412

aivsai.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77
---
88

99
# Introducing ⚔️ AI vs. AI ⚔️ a deep reinforcement learning multi-agents competition system
10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
<div align="center">
1412
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/128_aivsai/thumbnail.png" alt="Thumbnail">

ambassadors.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Student Ambassador Program’s call for applications is open!
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
As an open-source company democratizing machine learning, Hugging Face believes it is essential to **[teach](https://huggingface.co/blog/education)** open-source ML to people from all backgrounds worldwide. **We aim to teach machine learning to 5 million people by 2023**.
1412

annotated-diffusion.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,6 @@ authors:
88

99
# The Annotated Diffusion Model
1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1513

arxiv.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,6 @@ authors:
99

1010
# Hugging Face Machine Learning Demos on arXiv
1111

12-
<!-- {blog_metadata} -->
13-
<!-- {authors} -->
1412

1513
We’re very excited to announce that Hugging Face has collaborated with arXiv to make papers more accessible, discoverable, and fun! Starting today, [Hugging Face Spaces](https://huggingface.co/spaces) is integrated with arXivLabs through a Demo tab that includes links to demos created by the community or the authors themselves. By going to the Demos tab of your favorite paper, you can find links to open-source demos and try them out immediately 🔥
1614

asr-chunking.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Making automatic speech recognition work on large files with Wav2Vec2 in 🤗 Transformers
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
```
1412
Tl;dr: This post explains how to use the specificities of the Connectionist

assisted-generation.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Assisted Generation: a new direction toward low-latency text generation
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
Large language models are all the rage these days, with many companies investing significant resources to scale them up and unlock new capabilities. However, as humans with ever-decreasing attention spans, we also dislike their slow response times. Latency is critical for a good user experience, and smaller models are often used despite their lower quality (e.g. in [code completion](https://ai.googleblog.com/2022/07/ml-enhanced-code-completion-improves.html)).
1412

audio-datasets.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# A Complete Guide to Audio Datasets
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
<!--- Note to reviewer: comments and TODOs are included in this format. --->
1412

audioldm2.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# AudioLDM 2, but faster ⚡️
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
<a target="_blank" href="https://colab.research.google.com/github/sanchit-gandhi/notebooks/blob/main/AudioLDM-2.ipynb">
1412
<img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>

autoformer.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,6 @@ authors:
1010

1111
# Yes, Transformers are Effective for Time Series Forecasting (+ Autoformer)
1212

13-
<!-- {blog_metadata} -->
14-
<!-- {authors} -->
1513

1614
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1715

autonlp-prodigy.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,8 @@ authors:
55
- user: abhishek
66
---
77

8-
<h1>Active Learning with AutoNLP and Prodigy</h1>
8+
# Active Learning with AutoNLP and Prodigy
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
Active learning in the context of Machine Learning is a process in which you iteratively add labeled data, retrain a model and serve it to the end user. It is an endless process and requires human interaction for labeling/creating the data. In this article, we will discuss how to use [AutoNLP](https://huggingface.co/autonlp) and [Prodigy](https://prodi.gy/) to build an active learning pipeline.
1412

autotrain-image-classification.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Image Classification with AutoTrain
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1412

aws-marketplace.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,6 @@ authors:
99

1010
# Hugging Face Platform on the AWS Marketplace: Pay with your AWS Account
1111

12-
<!-- {blog_metadata} -->
13-
<!-- {authors} -->
1412

1513
The [Hugging Face Platform](https://aws.amazon.com/marketplace/pp/prodview-n6vsyhdjkfng2) has landed on the AWS Marketplace. Starting today, you can subscribe to the Hugging Face Platform through AWS Marketplace to pay for your Hugging Face usage directly with your AWS account. This new integrated billing method makes it easy to manage payment for usage of all our managed services by all members of your organization, including Inference Endpoints, Spaces Hardware Upgrades, and AutoTrain to easily train, test and deploy the most popular machine learning models like Llama 2, StarCoder, or BERT.
1614

aws-partnership.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,6 @@ authors:
99

1010
# Hugging Face and AWS partner to make AI more accessible
1111

12-
<!-- {blog_metadata} -->
13-
<!-- {authors} -->
1412

1513
It’s time to make AI open and accessible to all. That’s the goal of this expanded long-term strategic partnership between Hugging Face and Amazon Web Services (AWS). Together, the two leaders aim to accelerate the availability of next-generation machine learning models by making them more accessible to the machine learning community and helping developers achieve the highest performance at the lowest cost.
1614

bert-101.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,8 @@ authors:
55
- user: britneymuller
66
---
77
<html itemscope itemtype="https://schema.org/FAQPage">
8-
<h1>BERT 101 🤗 State Of The Art NLP Model Explained</h1>
8+
# BERT 101 🤗 State Of The Art NLP Model Explained
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1412

bert-cpu-scaling-part-1.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,8 +19,6 @@ authors:
1919
}
2020
</style>
2121

22-
<!-- {blog_metadata} -->
23-
<!-- {authors} -->
2422

2523
# Scaling up BERT-like model Inference on modern CPU - Part 1
2624

bert-cpu-scaling-part-2.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,6 @@ authors:
99

1010
# Scaling up BERT-like model Inference on modern CPU - Part 2
1111

12-
<!-- {blog_metadata} -->
13-
<!-- {authors} -->
1412

1513
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1614

bert-inferentia-sagemaker.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,8 @@ authors:
55
- user: philschmid
66
---
77

8-
<h1>Accelerate BERT inference with Hugging Face Transformers and AWS Inferentia</h1>
8+
# Accelerate BERT inference with Hugging Face Transformers and AWS Inferentia
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
<script async defer src="https://unpkg.com/medium-zoom-element@0/dist/medium-zoom-element.min.js"></script>
1412

bertopic.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -7,10 +7,8 @@ authors:
77
- user: davanstrien
88
---
99

10-
<h1> Introducing BERTopic Integration with the Hugging Face Hub</h1>
10+
# Introducing BERTopic Integration with the Hugging Face Hub
1111

12-
<!-- {blog_metadata} -->
13-
<!-- {authors} -->
1412

1513
[![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg 'open in colab')](https://colab.research.google.com/#fileId=https://huggingface.co/spaces/davanstrien/blog_notebooks/blob/main/BERTopic_hub_starter.ipynb)
1614

big-bird.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,8 +7,6 @@ authors:
77

88
# Understanding BigBird's Block Sparse Attention
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311
## Introduction
1412

blip-2.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,6 @@ authors:
88

99
# Zero-shot image-to-text generation with BLIP-2
1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
This guide introduces [BLIP-2](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2) from Salesforce Research
1513
that enables a suite of state-of-the-art visual-language models that are now available in [🤗 Transformers](https://huggingface.co/transformers).

bloom-inference-optimization.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,9 +5,7 @@ authors:
55
- user: Narsil
66
---
77

8-
<h1>Optimization story: Bloom inference</h1>
9-
<!-- {blog_metadata} -->
10-
<!-- {authors} -->
8+
# Optimization story: Bloom inference
119

1210
This article gives you the behind-the-scenes of how we made an efficient inference server that powers bloom.
1311
inference server that powers [https://huggingface.co/bigscience/bloom]().

bloom-inference-pytorch-scripts.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,10 +6,8 @@ authors:
66
- user: sgugger
77
---
88

9-
<h1>Incredibly Fast BLOOM Inference with DeepSpeed and Accelerate</h1>
9+
# Incredibly Fast BLOOM Inference with DeepSpeed and Accelerate
1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
This article shows how to get an incredibly fast per token throughput when generating with the 176B parameter [BLOOM model](https://huggingface.co/bigscience/bloom).
1513

bloom-megatron-deepspeed.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,8 @@ authors:
55
- user: stas
66
---
77

8-
<h1>The Technology Behind BLOOM Training</h1>
8+
# The Technology Behind BLOOM Training
99

10-
<!-- {blog_metadata} -->
11-
<!-- {authors} -->
1210

1311

1412

bloom.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -6,10 +6,8 @@ authors:
66
---
77

88

9-
<h1>🌸 Introducing The World's Largest Open Multilingual Language Model: BLOOM 🌸</h1>
9+
# 🌸 Introducing The World's Largest Open Multilingual Language Model: BLOOM 🌸
1010

11-
<!-- {blog_metadata} -->
12-
<!-- {authors} -->
1311

1412
<a href="https://huggingface.co/bigscience/bloom"><img style="middle" width="950" src="/blog/assets/86_bloom/thumbnail-2.png"></a>
1513

0 commit comments

Comments
 (0)