Jordan Legg PRO
AI & ML interests
Recent Activity
Organizations
Thank you!
Built a golden hour tracker in under 15 minutes with Lovable: uses your phoneβs Geolocation API, the SunCalc library, and runs fully client-side with no servers. https://goldenhour.404missing.link
What actually drives popularity of these papers, why do some papers get zero upvotes and why do some get thousands?
The answer is absolutely nothing. Yes that's right. Nothing about the actual paper itself drives popularity, the paper's popularity is driven by external factors like it's authors, external marketing and others.
So next time you see a research paper with a lot of upvotes, just remember it's not because of the efforts of the authors. Remain objective.
1οΈβ£ Faster ONNX and OpenVINO backends for SparseEncoder models
Usage is as simple as
backend="onnx" or backend="openvino" when initializing a SparseEncoder to get started, but I also included utility functions for optimization, dynamic quantization, and static quantization, plus benchmarks.2οΈβ£ New
n-tuple-scores output format from mine_hard_negativesThis new output format is immediately compatible with the MarginMSELoss and SparseMarginMSELoss for training SentenceTransformer, CrossEncoder, and SparseEncoder losses.
3οΈβ£ Gathering across devices
When doing multi-GPU training using a loss that has in-batch negatives (e.g. MultipleNegativesRankingLoss), you can now use
gather_across_devices=True to load in-batch negatives from the other devices too! Essentially a free lunch, pretty big impact potential in my evals.4οΈβ£ Trackio support
If you also upgrade
transformers, and you install trackio with pip install trackio, then your experiments will also automatically be tracked locally with trackio. Just open up localhost and have a look at your losses/evals, no logins, no metric uploading.5οΈβ£ MTEB Documentation
We've added some documentation on evaluating SentenceTransformer models properly with MTEB. It's rudimentary as the documentation on the MTEB side is already great, but it should get you started.
Plus many more smaller features & fixes (crash fixes, compatibility with datasets v4, FIPS compatibility, etc.).
See the full release notes here: https://github.com/UKPLab/sentence-transformers/releases/tag/v5.1.0
Big thanks to all of the contributors for helping with the release, many of the features from this release were proposed by others. I have a big list of future potential features that I'd love to add, but I'm
openfree/OpenAI-gpt-oss
VIDraft/gpt-oss-RAG
π― Two Models, One Space!
GPT-OSS hit #1 on HF just 2 hours after release! π
Now you can use both models conveniently in a single space.
π Model Selection Made Easy!
Just pick from the dropdown β
βββ GPT-OSS-120B (Complex tasks)
βββ GPT-OSS-20B (Quick chats)
π« How to Use (Takes 30 seconds!)
Sign in β With your HF account π
Select model β Choose what you need π
Apply β Click! β‘
Start chatting β That's it! π¬
π Perfect For:
120B β Deep analysis, professional work π§
20B β Fast responses, casual conversations β‘
No installation needed - just use it in your browser! π
β¨ Special Features
π¨ Beautiful gradient UI
π Dark mode support
π Real-time model switching
π Completely free!
π Try it now! It's really that simple!
#GPT-OSS #HuggingFace #FreeAI #EasyToUse
Made with late interaction I'd love to recreate the dataset to see a proper apache 2.0 version!
I'm using https://artificialanalysis.ai/ just because it puts everything in one place! It's not the best resource but these days I'm all about saving time.
@ThomasTheMaker if you make an issue on the repo, I'll look into it!
@ThomasTheMaker it's just the raw attention and transformer architecture in golang designed for serverless so performance will definitely be less than ggml and llama.cpp since it's not accelerated by GPU's but if you're into edge AI CPU only, this is the first, only and best way to compute attention.
Quantization can definitely be supported as it's just a math model!
We built this library at takara.ai to bring attention mechanisms and transformer layers to Go β in a form that's lightweight, clean, and dependency-free.
Weβre proud to say that every part of this project reflects what we set out to do.
- Pure Go β no external dependencies, built entirely on the Go standard library
- Core support for DotProductAttention and MultiHeadAttention
- Full transformer layers with LayerNorm, feed-forward networks, and residual connections
- Designed for edge, embedded, and real-time environments where simplicity and performance matter
Thank you to everyone who has supported this so far β the stars, forks, and feedback mean a lot.
No abstracts, just bullet points.
Start your day here: https://tldr.takara.ai
This is a pretty big update for sure. The models have improved significantly which is great for everyone involved, especially the end user. Those datasets look very promising as well!
Sounds interesting, Iβll check it out!
This is a really interesting post. Iβve been looking at the DeepSeek models for sure. This shows a pretty nice improvement, would love to see some example changes!