2024-11-24 10:44:00
github.com
SmolLM2 is a family of compact language models available in three size: 135M, 360M, and 1.7B parameters. They are capable of solving a wide range of tasks while being lightweight enough to run on-device.
You can find our most capable model 🤏 SmolLM2-1.7B-Instruct here.
New: Introducing SmolTalk, the SFT dataset of SmolLM2 🚀
Our most powerful model is SmolLM2-1.7B-Instruct
, which you can use as an assistant with transformers
, trl
, or using quantized versions with tools like llama.cpp
, MLX
, and transformers.js
. For lighter applications, you can also use the smaller models SmolLM2-360M
andSmolLM2-135M
, which are suitable for on-device usage and can be integrated similarly.
All available in this collection.
from transformers import AutoModelForCausalLM, AutoTokenizer
checkpoint = "HuggingFaceTB/SmolLM2-1.7B-Instruct"
device = "cuda" # for GPU usage or "cpu" for CPU usage
tokenizer = AutoTokenizer.from_pretrained(checkpoint)
# for multiple GPUs install accelerate and do `model = AutoModelForCausalLM.from_pretrained(checkpoint, device_map="auto")`
model = AutoModelForCausalLM.from_pretrained(checkpoint).to(device)
messages = [{"role": "user", "content": "Write a 100-word article on 'Benefits of Open-Source in AI research"}]
input_text=tokenizer.apply_chat_template(messages, tokenize=False)
inputs = tokenizer.encode(input_text, return_tensors="pt").to(device)
outputs = model.generate(inputs, max_new_tokens=50, temperature=0.2, top_p=0.9, do_sample=True)
print(tokenizer.decode(outputs[0]))
You can also use the TRL CLI to chat with the model from the terminal:
pip install trl
trl chat --model_name_or_path HuggingFaceTB/SmolLM2-1.7B-Instruct --device cpu
You can find more details on how to leverage the model for use cases such as text summarization, text rewriting and function calling in the model card: https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B-Instruct
You can use the models locally with frameworks like llama.cpp
, MLX
, and transformers.js
, which support SmolLM2.
All models are available in this collection.
A collection of lightweight AI-powered tools built with LLaMA.cpp and small language models. These tools are designed to run locally on your machine without requiring expensive GPU resources.
Further instructions on how to use the tools can be found in the smol-tools README.
You can find scripts for launching pre-training with nanotron under pre-training, we share the exact configs for training SmolLM1 and will upload SmolLM2’s configs soon.
You can find an example script to finetune SmolLM2 using TRL
and PEFT
in the finetuning
folder. We also link to our post-training scripts for SmolLM2 using the alignement handbook.
You can find more detailed evaluation of each model size in the model cards in this collection.
We use lighteval for all our evaluations, for more details refer to the evaluation README.
We released SmolTalk the SFT dataset used for building SmolLM2 instruct models. It was created with distilabel and you can check and execute the synthetic data pipelines in distilabel_pipelines README
Comparison of models finetuned on SmolTalk and Orca AgentInstruct 1M. For more details, refer to the dataset card.
Keep your files stored safely and securely with the SanDisk 2TB Extreme Portable SSD. With over 69,505 ratings and an impressive 4.6 out of 5 stars, this product has been purchased over 8K+ times in the past month. At only $129.99, this Amazon’s Choice product is a must-have for secure file storage.
Help keep private content private with the included password protection featuring 256-bit AES hardware encryption. Order now for just $129.99 on Amazon!
Support Techcratic
If you find value in Techcratic’s insights and articles, consider supporting us with Bitcoin. Your support helps me, as a solo operator, continue delivering high-quality content while managing all the technical aspects, from server maintenance to blog writing, future updates, and improvements. Support Innovation! Thank you.
Bitcoin Address:
bc1qlszw7elx2qahjwvaryh0tkgg8y68enw30gpvge
Please verify this address before sending funds.
Bitcoin QR Code
Simply scan the QR code below to support Techcratic.
Please read the Privacy and Security Disclaimer on how Techcratic handles your support.
Disclaimer: As an Amazon Associate, Techcratic may earn from qualifying purchases.