Xenova Profile
Xenova

@xenovacom

6,322
Followers
286
Following
196
Media
845
Statuses

Bringing the power of machine learning to the web. Currently working on Transformers.js ( @huggingface ๐Ÿค—)

Joined December 2015
Don't wanna be here? Send us removal request.
Pinned Tweet
@xenovacom
Xenova
1 year
I'm excited to announce that I have joined @huggingface ! ๐Ÿค— Continuing with Transformers.js, I aim to grow the community by bridging the gap between web development and machine learning. Thank you to everyone who has supported me so far, I can't wait to show you what's next! ๐Ÿš€
29
28
302
@xenovacom
Xenova
11 months
Introducing Whisper Web: ML-powered speech recognition directly in your browser! ๐Ÿš€ This comes with the release of Transformers.js v2.2.0, which now supports multilingual transcription and translation for over 100 different languages! ๐Ÿคฏ Check it out:
24
257
1K
@xenovacom
Xenova
22 days
Meta's Segment Anything Model (SAM) can now run in your browser w/ WebGPU (+ fp16), meaning up to 8x faster image encoding (10s โ†’ 1.25s)! ๐Ÿคฏโšก๏ธ Video is not sped up! Everything runs 100% locally thanks to ๐Ÿค— Transformers.js and onnxruntime-web! ๐Ÿ”— Demo:
17
194
1K
@xenovacom
Xenova
2 months
We just updated our in-browser Background Removal demo to use WebGPU and it's now ~50x faster! ๐Ÿคฏ ~9 seconds down to 180ms! โšก๏ธ Powered by @bria_ai_ 's RMBG-v1.4 model and ๐Ÿค— Transformers.js! ... and yes, the video is in real time! ๐Ÿคฏ
22
113
667
@xenovacom
Xenova
3 months
New @karpathy video just dropped! ๐Ÿ˜๐Ÿ”ฅ After watching, if you want to learn more about how different models (e.g., GPT4, Llama, T5, BERT) tokenize text, check out "The Tokenizer Playground": a web-app I built a few months ago with ๐Ÿค— Transformers.js! ๐Ÿ”—
@karpathy
Andrej Karpathy
3 months
New (2h13m ๐Ÿ˜…) lecture: "Let's build the GPT Tokenizer" Tokenizers are a completely separate stage of the LLM pipeline: they have their own training set, training algorithm (Byte Pair Encoding), and after training implement two functions: encode() from strings to tokens, andโ€ฆ
Tweet media one
385
2K
14K
3
106
640
@xenovacom
Xenova
6 months
Introducing Distil-Whisper Web: 49% smaller, 4.2x faster Whisper directly in your browser! ๐Ÿš€ Here is a side-by-side comparison with OpenAI's original version! ๐Ÿคฏ
16
88
573
@xenovacom
Xenova
2 months
WebGPU is the future! ๐Ÿ˜๐Ÿ”ฅ Transformers.js can now perform real-time background removal, powered by MODNet! โšก๏ธ Development for Transformers.js v3 (which adds full WebGPU support) is well underway, and we're excited to continue sharing updates and demos! Try it out yourself! ๐Ÿ‘‡
13
88
535
@xenovacom
Xenova
20 days
I know it just released, but I don't see many people talking about the Phi-3 tokenizer! ๐Ÿ‘€ Here's the full list of added special tokens... what do you notice? ๐Ÿคฏ <|assistant|> <|step|> <|function_output|> <|tag|> <|function_call|> <|system|> <|end|> <|raw|> <|continue|> <|user|>โ€ฆ
Tweet media one
18
71
515
@xenovacom
Xenova
8 months
Massive size reductions for Transformers.js models: - Whisper-tiny decoder: 50MB โ†’ 30MB (-40%) - NLLB decoder: 732MB โ†’ 476MB (-35%) - bloom: 819MB โ†’ 562MB (-31%) - gpt2: 167MB โ†’ 127MB (-24%) - bert-base: 134MB โ†’ 111MB (-17%) ... all this with ZERO drop in performance! ๐Ÿคฏ
3
86
467
@xenovacom
Xenova
3 months
YOLOv9 just released, and now it's compatible with ๐Ÿค— Transformers.js! That's right... near real-time object detection running locally in your browser: no server required! ๐Ÿคฏ Try it out yourself! ๐Ÿ‘‡ ๐Ÿ”— Demo:
7
81
457
@xenovacom
Xenova
8 months
When do you *really* need to use a vector database? ๐Ÿค” To try answer that question, I recreated my semantic image search application to run 100% in-browser with Transformers.js (no server). After loading the model and database, it only takes ~50ms to compute text embeddings andโ€ฆ
@karpathy
Andrej Karpathy
1 year
@sinclanich np.array people keep reaching for much fancier things way too fast these days
20
95
1K
33
61
432
@xenovacom
Xenova
1 year
Transformers.js v2.0 is finally here! ๐Ÿ”ฅ Run @huggingface transformers directly in your browser, with no need for a server! Some of the new features include: ๐Ÿ› ๏ธ Complete ES6 rewrite ๐Ÿ“„ Documentation + examples ๐Ÿค— Improved Hugging Face Hub integration
7
91
405
@xenovacom
Xenova
3 months
Here's a side-by-side comparison of the GPT-4, Gemma, and Llama tokenizers, tested on "The Great Gatsby" (270k characters). As @karpathy points out, the Gemma and Llama tokenizers are very similar, with the main difference being vocabulary size. One interesting thing to see isโ€ฆ
Tweet media one
@karpathy
Andrej Karpathy
3 months
Seeing as I published my Tokenizer video yesterday, I thought it could be fun to take a deepdive into the Gemma tokenizer. First, the Gemma technical report [pdf]: says: "We use a subset of the SentencePiece tokenizer (Kudo and Richardson, 2018) ofโ€ฆ
184
476
5K
13
68
390
@xenovacom
Xenova
4 months
Today we released ๐Ÿค— Transformers.js v2.14, which adds support for SAM (Segment Anything Model). This means you can now generate high-quality segmentation masks for objects in a scene, directly in your browser! ๐Ÿคฏ Demo (+ source code):
8
80
377
@xenovacom
Xenova
2 months
Grok-1 is finally out! ๐Ÿ˜ But while everyone was focused on the weights, I decided to take a look at the tokenizer. I also added it to the Tokenizer Playground! Structurally, it looks quite similar to the Llama 2 tokenizer (BPE w/ byte-fallback), with a vocabulary size of 2ยนโท =โ€ฆ
Tweet media one
8
77
358
@xenovacom
Xenova
8 months
WOW! ๐Ÿคฏ An in-browser version of ChatGPT (or HF Chat), built with ๐Ÿค— Transformers.js! Yes that's right, everything runs 100% locally in your browser, meaning no need for a server! Check it out! ๐Ÿ”—
@dhuynh95
Daniel Huynh
8 months
๐Ÿš€Just released #BlindChat : an open-source & privacy-first ChatGPT alternative! ๐Ÿ”’BlindChat combines @huggingface transformers.js (from the great @xenovacom ) with #ChatUI for a private and fully in-browser experience Try it on HF:
10
41
225
10
70
353
@xenovacom
Xenova
5 months
Distil-Whisper small is finally here! ๐Ÿ”ฅ Over 10x smaller, 5x faster, and within 3% WER of large-v2. ๐Ÿคฏ Since it's only 166M params, it can even run locally in your browser with ๐Ÿค— Transformers.js! Check it out! ๐Ÿ‘‰
9
61
356
@xenovacom
Xenova
3 months
Real-time object detection w/ ๐Ÿค— Transformers.js, running YOLOv9 directly in your browser! ๐Ÿคฏ This demo shows why on-device ML is so important: 1. Privacy - local inference means no user data is sent to the cloud 2. No server latency - empowers developers to build real-timeโ€ฆ
9
63
347
@xenovacom
Xenova
11 months
Due to popular demand, we added microphone support to Whisper Web! ๐ŸŽ™๏ธ This means you can now record and transcribe audio directly in your browser: no installation required. ๐Ÿ˜ Demo: Source code:
7
81
334
@xenovacom
Xenova
7 months
We just released ๐Ÿค— Transformers.js v2.7.0, which adds supports for ๐Ÿ—ฃ๏ธ Text to Speech w/ speecht5. This means you can now synthesize human speech directly in your browser... no server required! ๐Ÿ”ฅ Check out the demo! ๐Ÿ‘‡
15
74
322
@xenovacom
Xenova
3 months
Introducing Remove Background Web: In-browser background removal, powered by @bria_ai_ 's new RMBG-v1.4 model and ๐Ÿค— Transformers.js! That's right, everything runs 100% locally, meaning none of your images are uploaded to a server! ๐Ÿคฏ Try it out: ๐Ÿ‘‡
10
64
320
@xenovacom
Xenova
6 months
Transformers.js v2.9.0 is now out! ๐Ÿ˜ New features: ๐ŸŽฏ Zero-shot Object Detection w/ OwlViT ๐Ÿ•ต๏ธโ€โ™‚๏ธ Depth Estimation w/ DPT and GLPN ๐Ÿ“ Optical Document Understanding w/ Nougat ... and you can get started in just a few lines of code! ๐Ÿคฏ๐Ÿ‘‡
Tweet media one
6
57
314
@xenovacom
Xenova
10 months
Calling all JS developers! We just released 2 example Next.js applications which show how to use Transformers.js for client-side (in-browser) or server-side (Node.js) inference. ๐Ÿค— Building full-stack AI applications has never been this easy! ๐Ÿš€ Tutorial:โ€ฆ
5
71
312
@xenovacom
Xenova
7 months
Woah! ๐Ÿคฏ A new 20M parameter embeddings model that gives similar performance to OpenAI's text-embedding-ada-002, but is much smaller+faster! ๐Ÿ”ฅ I don't understand why >95% of developers still use closed-source embeddings models... ๐Ÿ˜… + it's compatible with ๐Ÿค— Transformers.js!
Tweet media one
@andersonbcdefg
Ben (e/sqlite)
7 months
New embeddings model, gte-tiny, is published! Distilled from gte-small, offering slightly-worse performance with half the layers. (Alternatively, same size but better performance compared to all-MiniLM-L6-v2.) ONNX models also available. Check it out! (link below)
Tweet media one
7
33
231
7
45
309
@xenovacom
Xenova
4 months
Depth Anything is now available in ๐Ÿค— Transformers.js! At just 25M parameters, the small version of the model runs great locally. Here's a demo I created which performs monocular depth estimation directly in your browser (no server needed)! ๐Ÿคฏ ๐Ÿ”— Demo:
3
54
302
@xenovacom
Xenova
4 months
TinyLlama is finally here: a 1.1B Llama model trained on 3 trillion tokens! ๐Ÿคฏ It's also compatible with ๐Ÿค— Transformers.js (see code below)! ๐Ÿ‘‡ What a way to end the year! ๐Ÿฅณ ๐Ÿ”—
Tweet media one
8
46
297
@xenovacom
Xenova
20 days
Using ๐Ÿค— Transformers.js, you can now run CLIP directly in your browser at over 20fps w/ WebGPU (@ fp16) for real-time zero-shot image classification! ๐Ÿคฏ As always, everything runs 100% locally, meaning no calls made to an API! ๐Ÿ”ฅ Try it out! ๐Ÿ‘‡ ๐Ÿ”— Demo:
6
45
282
@xenovacom
Xenova
9 months
Introducing Chat with YouTube, an AI-powered browser extension that lets you chat with YouTube videos! ๐Ÿš€ This project shows how easy it is to build conversational browser extensions using ๐Ÿค— Inference Endpoints and @Vercel 's AI SDK. + it's open source!
13
66
266
@xenovacom
Xenova
2 months
Claude 3 just released and although the weights aren't open, the tokenizer is! ๐Ÿ”ฅ If you want to calculate how many tokens you're sending to the API, check out The Tokenizer Playground, which we recently updated to include the Claude 3 tokenizer! ๐Ÿ‘‡ ๐Ÿ”—
3
54
273
@xenovacom
Xenova
1 month
Introducing MusicGen Web: AI-powered music generation directly in your browser, built with ๐Ÿค— Transformers.js! ๐ŸŽต Everything runs 100% locally, meaning no calls to an API! ๐Ÿคฏ Served as a static website... this costs $0 to host and run! ๐Ÿ”ฅ Try it out yourself! ๐Ÿ‘‡
7
58
272
@xenovacom
Xenova
6 months
We just released ๐Ÿค— Transformers.js v2.8.0, which adds a ton of new features, including: ๐Ÿ–ผ๏ธ Super-resolution and image restoration w/ Swin2SR โœ๏ธ Optical character recognition w/ TrOCR ๐Ÿ’ฌ Text-generation w/ Mistral and Falcon (<1B params) More details in ๐Ÿงต๐Ÿ‘‡
Tweet media one
5
53
269
@xenovacom
Xenova
1 year
Introducing Transformers.js โ€“ run @HuggingFace transformers directly in your browser! We currently support BERT, DistilBERT, T5, and GPT2 models, for a variety of tasks such as translation, text generation, and sentiment analysis. โ€ฆ and it's open-source!
10
53
258
@xenovacom
Xenova
10 months
We just released Transformers.js v2.4.0, which adds support for generating word-level timestamps w/ whisper! ๐Ÿคฏ That's right, you can now perfectly caption videos directly in your browser. ๐Ÿ”ฅ I can't wait to see what you build with this! Get started in just a few lines of code:
Tweet media one
10
53
253
@xenovacom
Xenova
2 months
โšก๏ธ Now with WebGPU support! โšก๏ธ Run depth estimation w/ Depth Anything in under 200ms, thanks to Transformers.js and WebGPU! Try it out yourself! ๐Ÿ‘‡
@xenovacom
Xenova
4 months
Depth Anything is now available in ๐Ÿค— Transformers.js! At just 25M parameters, the small version of the model runs great locally. Here's a demo I created which performs monocular depth estimation directly in your browser (no server needed)! ๐Ÿคฏ ๐Ÿ”— Demo:
3
54
302
7
41
252
@xenovacom
Xenova
9 months
Two annoying things about OpenAI's tokenizer playground: (1) it's capped at 50k characters, and (2) it doesn't support GPT-4 or GPT-3.5... So, I built my own version w/ Transformers.js! It can tokenize the entire "Great Gatsby" (269k chars) in 200ms! ๐Ÿš€
Tweet media one
Tweet media two
8
35
252
@xenovacom
Xenova
2 months
New features in ๐Ÿค— Transformers.js v2.16.1: ๐Ÿ”ฅ New models: APISR for Anime Super-Resolution and EfficientNet for image classification ๐Ÿ–ผ๏ธ New pipeline: Image Feature Extraction ๐Ÿ’ฌ Improved chat templating support: C4AI Command-R tool and RAG prompt generation See ๐Ÿงต for more info
5
33
211
@xenovacom
Xenova
7 months
A new open-source embeddings model with 8K context length that matches the performance of text-embedding-ada-002! ๐Ÿคฏ This is a game changer! ๐Ÿ”ฅ And now it's compatible with ๐Ÿค— Transformers.js, meaning you can generate embeddings in your browser, Node.js, or even Deno!
Tweet media one
@JinaAI_
Jina AI
7 months
Introducing jina-embeddings-v2, the world's first open-source model boasting an 8K context length. Matching the prowess of OpenAI's proprietary models, now accessible on @huggingface , signaling a significant milestone in the landscape of text embeddings.
15
102
495
2
27
200
@xenovacom
Xenova
27 days
The competition in AI music generation is heating up, with Suno and Udio leading the way. Unfortunately, neither are open source... ๐Ÿ˜ข Luckily, MusicGen is! ๐Ÿ˜ The quality is amazing and you can even run it locally in your browser with Transformers.js! ๐ŸŽต For example: prompt:โ€ฆ
6
39
200
@xenovacom
Xenova
6 months
Yes, that's right... the new Distil-Whisper models from @huggingface are fully compatible with Transformers.js! ๐Ÿค— This means you can generate high-quality transcripts directly in JavaScript: in-browser, Node, or even Deno! ๐Ÿคฏ๐Ÿš€ Get started in just 3 lines of code:
Tweet media one
@xenovacom
Xenova
6 months
Introducing Distil-Whisper Web: 49% smaller, 4.2x faster Whisper directly in your browser! ๐Ÿš€ Here is a side-by-side comparison with OpenAI's original version! ๐Ÿคฏ
16
88
573
5
25
176
@xenovacom
Xenova
2 months
Introducing the ๐Ÿค— Transformers.js WebGPU Embedding Benchmark! โšก๏ธ How much does WebGPU speed up ML models running locally in your browser? Try it out and share your results! ๐Ÿš€๐Ÿ‘‡ ๐Ÿ”—
7
28
174
@xenovacom
Xenova
10 months
AI code completion running 100% locally inside your browser, thanks to @BigCodeProject 's StarCoder models and ๐Ÿค— Transformers.js! We also got their new 1B model running at ~20 tokens per second in Node.js (CPU). ๐Ÿš€ Check out the demo!
7
35
165
@xenovacom
Xenova
25 days
Meta's Llama 3 is here, with a brand new tokenizer! ๐Ÿฆ™ I've added it to the Tokenizer Playground, so you can experiment with it in your browser: For those interested, here are the key differences over Llama 2: 1. 4x larger vocabulary (32K -> 128K). Thisโ€ฆ
Tweet media one
3
32
169
@xenovacom
Xenova
5 months
๐Ÿค— Transformers.js v2.13 - Holiday update! โ˜ƒ๏ธ In this version, we added: 1. SegFormer for semantic segmentation and image classification. 2. VITS for multilingual text-to-speech (>1000 languages). 3. CLIPSeg for zero-shot image segmentation. 4. Table Transformer for tableโ€ฆ
Tweet media one
5
30
160
@xenovacom
Xenova
6 months
This is an absolute game changer! ๐Ÿคฏ @threejs but for Gaussian Splatting! ๐Ÿ”ฅ I canโ€™t wait to see what the web-dev community builds with this! ๐Ÿค— cc @mrdoob
@dylan_ebert_
dylan
6 months
๐Ÿš€ Introducing gsplat.js - a JavaScript Gaussian Splatting library ๐Ÿš„ Speed: Optimized WebAssembly โ˜‚๏ธ Versatility: Modular, swap out the renderer ๐Ÿค— Open Source: MIT License Code: Repo:
18
273
1K
1
32
156
@xenovacom
Xenova
6 months
Transformers.js just hit 1 million total requests on @jsDelivr , with 52% of them coming in the past 30 days alone! ๐Ÿคฏ We have a ton of exciting updates coming soon, so stay tuned! I'm excited to show you what's next... ๐Ÿš€๐Ÿ”ฅ
Tweet media one
8
24
150
@xenovacom
Xenova
4 months
Image-to-LaTeX in 3 lines of JavaScript code, with ๐Ÿค— Transformers.js! This is made possible thanks to @vikparuchuri 's amazing texify2 model, which we converted to ONNX so it can run in the browser! ๐Ÿ”ฅ ๐Ÿ’ก Project idea: browser extension to convert PDFs/screenshots to LaTeX!
Tweet media one
3
26
143
@xenovacom
Xenova
5 months
To showcase the power of in-browser machine learning for real-time data visualization, I built a semantic music search app, powered by ๐Ÿค— Transformers.js and Deepscatter. Users can search over 50k songs with natural language, all running client-side (no server)! ๐Ÿคฏ
9
19
142
@xenovacom
Xenova
3 months
Nomic Embed v1.5 is out, the first open model with variable-sized Matryoshka embeddings and 8192 context! ๐Ÿคฏ It's also compatible with ๐Ÿค— Transformers.js, meaning you can perform adaptive retrieval directly in your browser! Demo showing how dimensionality affects performance ๐Ÿ‘‡
5
26
142
@xenovacom
Xenova
1 year
Yes, you *heard* that right... Transformers.js now supports automatic speech recognition w/ Whisper! Everything runs entirely inside your browser. No need to make API calls to a server! ๐Ÿคฏ #WebML ... and yes, it's open source:
6
35
140
@xenovacom
Xenova
7 months
In-browser TTS in 3 lines of code! ๐Ÿคฏ import { pipeline } from '@โ€‹xenova/transformers'; let pipe = await pipeline( 'text-to-speech', 'Xenova/speecht5_tts', { quantized: false } ); let out = await pipe('Text goes here', { speaker_embeddings: 'speaker_embeddings.bin' });
@xenovacom
Xenova
7 months
We just released ๐Ÿค— Transformers.js v2.7.0, which adds supports for ๐Ÿ—ฃ๏ธ Text to Speech w/ speecht5. This means you can now synthesize human speech directly in your browser... no server required! ๐Ÿ”ฅ Check out the demo! ๐Ÿ‘‡
15
74
322
7
27
140
@xenovacom
Xenova
10 months
Transformers.js just hit 3000 stars on GitHub! ๐Ÿคฏ The #WebML community is growing so fast, and I'm proud to be a part of it! ๐Ÿค— If you ever plan on adding in-browser machine-learning functionality to your website or web-app, check out the project:
Tweet media one
3
30
139
@xenovacom
Xenova
8 months
We just released Transformers.js v2.6.0! New features: - 14 new architectures: BLOOM, MPT, BeiT, CamemBERT, CodeLlama, GPT-J, mBART, ResNet, WavLM, and more! ๐Ÿ˜ - Over 150 newly-converted models on the Hub! ๐Ÿš€ - Huge model size reductions (up to -40%)! ๐Ÿ‘‡
3
25
133
@xenovacom
Xenova
8 months
Transformers.js v2.6.2 now supports Document Question Answering, meaning you can easily extract information from images... directly in your browser (no server needed)! ๐Ÿคฏ We also added new models like Donut, LongT5, and Blenderbot! ๐Ÿฅณ I can't wait to see what you build! ๐Ÿค—
Tweet media one
3
18
135
@xenovacom
Xenova
2 months
We just added some new features to the ๐Ÿค— Transformers.js WebGPU Embedding Benchmark: - fp16 and int8 support - Ability to change models - Lazy model loading - Options to select which tests to run On my device, I got >100x speedup with fp16 on WebGPU! โšก๏ธ
Tweet media one
@xenovacom
Xenova
2 months
Introducing the ๐Ÿค— Transformers.js WebGPU Embedding Benchmark! โšก๏ธ How much does WebGPU speed up ML models running locally in your browser? Try it out and share your results! ๐Ÿš€๐Ÿ‘‡ ๐Ÿ”—
7
28
174
4
18
128
@xenovacom
Xenova
2 months
Transformers.js v2.16 is now out! ๐Ÿš€ Here are some of the new features: ๐Ÿ’ฌ StableLM text-generation models ๐Ÿ”‰ Speaker verification and diarization models ๐Ÿ“ Improved chat templating operation coverage ๐ŸŽจ New example applications and demos Release notes:
Tweet media one
2
14
129
@xenovacom
Xenova
1 year
Generate embeddings directly in your browser (or Node.js) with the latest version of Transformers.js! ๐Ÿคฏ We can't wait to see what you make with it! The possibilities are endless! Semantic search, sentence similarity, clustering, etc... What else? ๐Ÿ‘€๐Ÿ”ฅ
Tweet media one
9
23
127
@xenovacom
Xenova
4 months
And now, SigLIP is available in ๐Ÿค— Transformers.js! To test how it fares in practice, I adapted my semantic image search demo to use SigLIP instead of CLIP, and it works great! Everything runs 100% locally in your browser (no server needed)! ๐Ÿ”ฅ ๐Ÿ”—
@NielsRogge
Niels Rogge
4 months
SigLIP by @Google is now available in ๐Ÿค— Transformers! It improves upon @OpenAI 's CLIP with a simple sigmoid loss. SOTA for linking images with text and vice versa. Demo notebook: Original meme credits: @giffmana
Tweet media one
4
89
656
5
22
129
@xenovacom
Xenova
8 months
Continuing from yesterday: - T5 decoder: 59MB โ†’ 42MB (-28%) - distilbert-base: 91MB โ†’ 68MB (-25%) - bart-base decoder: 207MB โ†’ 155MB (-25%) - roberta-base: 165MB โ†’ 126MB (-24%) #WebML for the win! ๐Ÿš€
@xenovacom
Xenova
8 months
Massive size reductions for Transformers.js models: - Whisper-tiny decoder: 50MB โ†’ 30MB (-40%) - NLLB decoder: 732MB โ†’ 476MB (-35%) - bloom: 819MB โ†’ 562MB (-31%) - gpt2: 167MB โ†’ 127MB (-24%) - bert-base: 134MB โ†’ 111MB (-17%) ... all this with ZERO drop in performance! ๐Ÿคฏ
3
86
467
4
20
123
@xenovacom
Xenova
3 months
Nomic Embed is out: a new 8K text embedding model by @nomic_ai ! ๐Ÿš€ It's also compatible with ๐Ÿค— Transformers.js, meaning you can generate embeddings directly in your browser (no server required)!
Tweet media one
@nomic_ai
Nomic AI
3 months
Introducing Nomic Embed - the first fully open long context text embedder to beat OpenAI - Open source, open weights, open data - Beats OpenAI text-embeding-3-small and Ada on short and long context benchmarks - Day 1 integrations with @langchain , @llama -index, @MongoDB
38
276
2K
2
16
116
@xenovacom
Xenova
9 months
Transformers.js v2.5.2 now supports audio classification w/ MMS and wav2vec2, meaning you can, for example, perform language identification for over 4000 languages! ๐Ÿคฏ๐Ÿ”ฅ Get started in just 3 lines of code! ๐Ÿ‘‡ Full release notes:
Tweet media one
4
18
114
@xenovacom
Xenova
1 month
Transformers.js v2.17 is out! ๐Ÿ”ฅ๐Ÿš€ New features: ๐Ÿ”ข Binary embeddings: 32x storage savings and significantly faster retrieval (with up to ~95% of the original performance)! ๐Ÿ’ฌ Improved conversational support: Pass chat messages directly to the `text-generation` pipeline.
Tweet media one
2
12
115
@xenovacom
Xenova
3 months
Did you know that HuggingChat uses Transformers.js for RAG/WebSearch? ๐Ÿคฏ It's amazing to see how far the library has come, and I'm so grateful to everyone in the community for helping make it what it is today! ๐Ÿค— HuggingChat is also open source! ๐Ÿ‘‡
2
8
114
@xenovacom
Xenova
10 months
Introducing Doodle Dash, an ML-powered web game that runs completely in your browser, thanks to Transformers.js! ๐Ÿคฏ You have 60 seconds to draw as many words as you can, while a neural network tries to guess what you're drawing in real time! Play here:
5
29
103
@xenovacom
Xenova
5 months
๐Ÿค— Transformers.js just hit 5K stars on GitHub! ๐ŸŒŸ Thank you to everyone in the community for your support and contributions... this is why open source is the best! ๐Ÿ”ฅ PS: Stay tuned for some exciting updates coming soon! ๐Ÿš€
Tweet media one
3
13
100
@xenovacom
Xenova
25 days
Snowflake just released Arctic Embed, a collection of open-source text embedding models optimized for retrieval accuracy and efficiency! โ„๏ธ ๐Ÿ“„ Apache 2.0 license ๐ŸŒ Great for in-browser use w/ ๐Ÿค— Transformers.js (22 โ†’ 335 M params) โšก๏ธ WebGPU-accelerated (>120x faster than WASM)
Tweet media one
2
14
99
@xenovacom
Xenova
3 months
Did anyone else notice the `<start_of_image>` token in Gemma's vocabulary? ๐Ÿ‘€ Are we going to see some VLM-variants soon? I also added Gemma to "The Tokenizer Playground", which you can check out if you want to learn more about how the model performs tokenization! ๐Ÿ‘‡
4
10
98
@xenovacom
Xenova
3 months
New blog post: An Introduction to Matryoshka Embedding Models ๐Ÿช† Learn how these models are able to produce embeddings of various dimensions, how they can speed up tasks like retrieval, and how you can train your own! ๐Ÿ‘‡ ๐Ÿ”—
1
17
95
@xenovacom
Xenova
3 months
๐Ÿšจ Hugging Chat Assistants are out! ๐Ÿšจ Just like OpenAI's GPTs, you can now create your own personal assistant in Hugging Chat! ๐Ÿคฏ To test it out, I built a "Prisoner Interrogation" game, where you must try to extract a secret password from a prisoner. Can you do it? ๐Ÿค”
10
19
91
@xenovacom
Xenova
3 months
Qwen1.5 is out: a collection of powerful LLMs with sizes ranging from 0.5B to 72B parameters. Even at 8-bit quantization, the smallest one (0.5B) is surprisingly good for its size! Here's a demo I made with Transformers.js (v2.15), running 100% locally in the browser w/ WASM! ๐Ÿคฏ
@JustinLin610
Junyang Lin
3 months
๐ŸŽ‰Happy to announce the release of Qwen1.5! This time, we directly opensource new models of 6 sizes, 0.5B, 1.8B, 4B, 7B, 14B, and 72B (including base, chat, AWQ, GPTQ, GGUF)! From small to huge! Blog: GitHub: HF:โ€ฆ
Tweet media one
34
117
627
4
14
92
@xenovacom
Xenova
9 months
Here's a sneak peek of my "Chat with YouTube" browser extension, made with @Vercel 's AI SDK! ๐Ÿ”ฅ It uses Llama-v2 (7B) deployed with @HuggingFace inference endpoints. ๐Ÿš€ Source code and tutorial coming soon! ๐Ÿค—
4
11
87
@xenovacom
Xenova
4 months
Introducing the ๐Ÿ› Jinja Playground: Design LLM chat templates directly in your browser with instant feedback. Built with `@โ€‹huggingface/jinja`, a minimalistic JavaScript implementation of the Jinja templating engine, specifically designed for parsing + rendering chat templates.
2
19
84
@xenovacom
Xenova
4 months
Yesterday, @MoritzLaurer released some new tiny zero-shot classifiers, so to put them to the test, I built a simple web application that sorts customer product reviews into classes chosen at runtime. Everything runs 100% locally in your browser, thanks to ๐Ÿค— Transformers.js!
@MoritzLaurer
Moritz Laurer
4 months
๐Ÿค New 0.02B, 25 MB tiny zeroshot classifiers for edge device use-cases on @huggingface ! The xtremedistil ONNX quantized version is only 13 MB and very fast on CPUs. Without quantization, it has a throughput of ~4000 full sentences (! not just tokens) per second on an A10G withโ€ฆ
1
19
90
1
19
83
@xenovacom
Xenova
5 months
Say goodbye to silent performance issues when prompting LLMs! Today we released ๐Ÿค— Transformers.js v2.12, which adds support for chat templating! ๐Ÿ’ฌ This means you can generate LLM inputs for almost any model on the @huggingface Hub, directly in your browser w/ JavaScript! ๐Ÿคฏ
Tweet media one
2
22
81
@xenovacom
Xenova
3 months
Local background removal Figma plugin, built with ๐Ÿค— Transformers.js and BRIA AI's RMBG-v1.4 model! This shows what an amazing opportunity it is for JavaScript developers to build powerful AI applications, without worrying about API/server costs! Great work @enzostvs ! ๐Ÿ”ฅ
2
15
77
@xenovacom
Xenova
1 year
Today we added code-completion to Transformers.js! ๐Ÿคฏ It's like GitHub copilot, but it runs directly in your browser (i.e., no calls to a server)! #WebML We can't wait to see what people make with it! ๐Ÿ˜ ... oh, and did we mention it's open source?
5
18
77
@xenovacom
Xenova
5 months
๐Ÿค— Transformers.js v2.11 is one of our biggest releases yet and includes 8 exciting new models! ๐Ÿงต 1. ViTMatte for image matting: separate images into foreground and background, directly in your browser! This is going to make for some cool image editing web applications! ๐Ÿ”ฅ
Tweet media one
2
17
76
@xenovacom
Xenova
1 year
HUGE seq2seq model size reductions in Transformers.js v1.1.0! whisper: 110 MB -> 61 MB (-45%) t5-small: 151 MB -> 95 MB (-37%) distilbart: 529 MB -> 336 MB (-36%) ... without any reduction in quality! ๐Ÿคฏ This is MASSIVE for #WebML !
4
9
74
@xenovacom
Xenova
5 months
We just released ๐Ÿค— Transformers.js v2.10, which adds support for: ๐ŸŽต Zero-shot audio classification w/ CLAP ๐ŸŽ™๏ธ Audio classification w/ Audio Spectrogram Transformer ๐Ÿ–ผ๏ธ Image classification w/ ConvNeXT Get started with just a few lines of code! ๐Ÿคฏ๐Ÿฅณ
Tweet media one
2
16
75
@xenovacom
Xenova
5 months
How to perform image segmentation in 3 lines of JavaScript code: import { pipeline } from '@โ€‹xenova/transformers'; const segmenter = await pipeline( 'image-segmentation', 'Xenova/face-parsing', ); const output = await segmenter('image.jpg');
2
16
74
@xenovacom
Xenova
8 months
I just tried installing the latest version of ๐Ÿค— Transformers.js with Bun 1.0: npm: 10.7 seconds Bun: 0.404 seconds (26.5x faster) ๐Ÿคฏ @bunjavascript @jarredsumner This is amazing! Great job! ๐Ÿ˜
Tweet media one
@xenovacom
Xenova
8 months
We just released Transformers.js v2.6.0! New features: - 14 new architectures: BLOOM, MPT, BeiT, CamemBERT, CodeLlama, GPT-J, mBART, ResNet, WavLM, and more! ๐Ÿ˜ - Over 150 newly-converted models on the Hub! ๐Ÿš€ - Huge model size reductions (up to -40%)! ๐Ÿ‘‡
3
25
133
2
8
68
@xenovacom
Xenova
23 days
๐Ÿค— Transformers.js + MusicGen = ๐Ÿคฏ Check it out! ๐Ÿ‘‡
@victormustar
Victor M
23 days
Mind. Blown. Web browsers can now generate music, and it's ๐Ÿ”ฅ no servers needed, it all happens right in your browser. โฌ‡๏ธ Check out the link below and get ready to jam!
10
37
217
3
14
68
@xenovacom
Xenova
1 month
@Raynerdwebsdev It sure can!
@xenovacom
Xenova
2 months
We just updated our in-browser Background Removal demo to use WebGPU and it's now ~50x faster! ๐Ÿคฏ ~9 seconds down to 180ms! โšก๏ธ Powered by @bria_ai_ 's RMBG-v1.4 model and ๐Ÿค— Transformers.js! ... and yes, the video is in real time! ๐Ÿคฏ
22
113
667
2
5
68
@xenovacom
Xenova
2 months
Amazing work @fleetwood___ ! ๐Ÿ”ฅ Distil-Whisper large running blazingly fast in your browser, powered by WebGPU! โšก๏ธ
@fleetwood___
Fleetwood
2 months
๐Ÿšจ Ratchet reaches alpha! ๐Ÿšจ With todays release of Distil Whisper Large V3 by @sanchitgandhi99 , Ratchet officially enters alpha. Check out this demo running ๐—น๐—ฎ๐—ฟ๐—ด๐—ฒ-๐˜ƒ๐Ÿฏ(!!) in the browser
7
13
85
2
3
67
@xenovacom
Xenova
1 year
I can't believe this just happened... @huggingface
Tweet media one
2
5
65
@xenovacom
Xenova
1 year
With all the hype around multi-modal models (like GPT-4), we decided to add support for CLIP to Transformers.js ๐Ÿ”ฅ CLIP can be used for image-text similarity and zero-shot image classification... and now it can run directly in your browser! Source code:
2
13
65
@xenovacom
Xenova
1 year
You asked for it, so here it is... Transformers.js now supports *timestamped* speech-to-text with whisper! ๐Ÿคฏ We can't wait to see what people will create with it! (see below for example usage and outputs) Source code:
Tweet media one
Tweet media two
1
12
64
@xenovacom
Xenova
1 year
Today we added object detection to Transformers.js! ๐Ÿคฏ This means that you can detect the location and type of objects in an image, directly in your browser (no calls to a server)! ... and did we mention it's open source?
1
15
62
@xenovacom
Xenova
2 months
Bring images to life with ๐Ÿค— Transformers.js and Depth Anything! Everything runs 100% locally in your browser (no server required)! ๐Ÿคฏ Great work on the demo @conzept__ ! ๐Ÿ”ฅ
@conzept__
Conzept
2 months
Added "image depth estimation" to the image viewer (using @xenovacom #transformersjs project) This semi 3D perspective can add new insights and flavour to many of the images. Painting: Landscape:
Tweet media one
Tweet media two
Tweet media three
1
1
12
4
11
63
@xenovacom
Xenova
8 months
My first Hugging Face collection! ๐Ÿ˜ Check out my favorite #WebML demos built with Transformers.js! ๐Ÿ”—
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
10
60
@xenovacom
Xenova
10 months
Have you ever wanted to build your own semantic image search application in @nextjs ? Well now you can, thanks to ๐Ÿค— Transformers.js and โšก๏ธ @supabase ! This comes with the release of v2.5.0, which adds support for computing embeddings with CLIP! Demo:
Tweet media one
3
16
61
@xenovacom
Xenova
1 year
With all the hype around SAM (Segment Anything Model) from @MetaAI , we decided to add image-segmentation to Transformers.js! ๐Ÿ˜ Unlike their demo, this *actually* runs in your browser, meaning no calls to a server! See how you can use it in your projects today:
Tweet media one
2
10
61
@xenovacom
Xenova
1 year
You can now transcribe audio clips longer than 30 seconds using @OpenAI 's whisper model, directly in your browser (no server)! ๐Ÿ˜ Do this by specifying a chunk length and stride length when calling the pipeline function (see example). Check it out!
Tweet media one
4
13
57
@xenovacom
Xenova
6 months
Introducing Static Templates, now live on the ๐Ÿค— Hub! Create and deploy your own website in just two clicks! Perfect for showcasing machine learning projects, demos, and more... Get started today! ๐Ÿ‘‡ ๐Ÿ”—
2
15
55
@xenovacom
Xenova
1 year
Weights on @huggingface or it didnโ€™t happenโ€ฆ ๐Ÿ˜‰
@AIatMeta
AI at Meta
1 year
Today we're releasing the Segment Anything Model (SAM) โ€” a step toward the first foundation model for image segmentation. SAM is capable of one-click segmentation of any object from any photo or video + zero-shot transfer to other segmentation tasks โžก๏ธ
144
2K
7K
0
9
55
@xenovacom
Xenova
1 year
Soon you'll be able to run @huggingface Transformers in browser extensions! ๐Ÿคฏ The models run locally inside your browser - no need for a server! We can't wait to see what people will develop when Transformers.js v2.0.0 releases! ๐Ÿš€ #WebML
Tweet media one
2
4
53
@xenovacom
Xenova
1 year
This is my petition to change @huggingface 's error page to a sad face, because it better reflects how I feel right now... ๐Ÿ˜ญ
Tweet media one
6
3
53
@xenovacom
Xenova
1 year
Transformers.js is number 1 trending on Hacker News! ๐Ÿคฏ and it has just blown up to 400 stars on GitHub?! AHHHHHHHHHHHHHH!
Tweet media one
5
7
48
@xenovacom
Xenova
8 months
Did you know you can build powerful AI applications directly in JavaScript? ๐Ÿ‘€ Well, to show how easy it is, we published an interactive video tutorial โ€” in partnership with @Scrimba โ€” where we build an Object Detection web app using Transformers.js! ๐Ÿ”—
Tweet media one
4
10
49
@xenovacom
Xenova
1 year
Transformers.js can now do zero-shot classification! ๐Ÿฅณ This means you can classify text according to classes specified at runtime (and without finetuning). What @huggingface tasks should we add next? Leave them in the comments below! ๐Ÿ‘‡
3
12
48
@xenovacom
Xenova
5 months
Build powerful AI chat applications in just a few lines of JavaScript code, with the @huggingface Inference API and Transformers.js! ๐Ÿค— Here's a demo showing how to run the latest @MistralAI 7B Instruct model for free in vanilla JS! Try it out yourself!
1
8
47