-
Notifications
You must be signed in to change notification settings - Fork 28.8k
[Community contributions] Model cards #36979
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi. I would like to work on model card for gemma 2. |
Hi. I would like to work on model card for mistral. |
Hi @stevhliu , this is my first contribution so I have a really basic question . Should I clone every repo under mistralai? I just cloned the repo mistralai/Ministral-8B-Instruct-2410, but there are many other repos under mistralai. It's ok if I need to, but I just want to be sure. |
Hey , I would like to work on the model card for llama3 . |
Hey @NahieliV, welcome! You only need to modify the mistral.md file. This is just for the model cards in the Transformers docs rather than the Hub. |
Hey @stevhliu I would like to work on the model card for qwen2_5_vl. |
@stevhliu Is it not possible to automate with an LLM? |
hi @stevhliu i would be super grateful if you can let me work on the model card for code_llama |
Hey @stevhliu, I would like to work on the |
Hey @stevhliu , i would like to contribute to |
Hey @stevhliu , I would like to contribute to vitpose model card |
Hey @stevhliu, I would like to work on the |
Hey @stevhliu , I would like to contribute to |
To the folks who have been raising PR so far , just have a doubt did you get to install EDIT : Got it up and running, had to install all the libraries to make it run successfully. Initially felt doubtful about the need to install all the libraries such as flax but yea seems like it has to be installed too. |
Hey @stevhliu, I would like to work on the phi3 model card |
As you just going to edit the docs, you need not have a complete development setup. Fork the |
Hey @stevhliu I would like to work on the model card for deberta. Hope that's alright |
hi @stevhliu I would like to add model card for |
Hey @Rishik00, DeBERTa is already taken. Do you want to work on DeBERTav2? |
Hello @stevhliu , I would like to add the model card for Gemma and Siglip2 if that's possible |
Hello @stevhliu ! |
@stevhliu Definitely! I'd love to! Is there a deadline? |
There's no deadline, so feel free to work on it whenever you have the time :) |
@stevhliu for the attention visualization image, I opened a pull request at https://huggingface.co/datasets/huggingface/documentation-images/discussions/479 . please merge it so i can include it in model card |
@stevhliu raised a PR for SigLIP2 |
@stevhliu Can I work on Audio Spectrogram Transformers? |
@stevhliu can i work on qwen2_vl |
@afafelwafi, since @saswatmeher has already raised a PR for SigLIP2, would it be ok if you picked a different model? 🤗 |
@stevhliu please merge this |
Hey, I don't think the |
@stevhliu Yeah, you're right from transformers import AlbertTokenizer, AlbertModel
import matplotlib.pyplot as plt
import torch
tokenizer = AlbertTokenizer.from_pretrained("albert/albert-base-v1")
model = AlbertModel.from_pretrained("albert/albert-base-v1", output_attentions=True) # Enable attention output
text = "Plants create energy through a process known as"
inputs = tokenizer(text, return_tensors="pt")
with torch.no_grad():
outputs = model(**inputs)
attentions = outputs.attentions # List of attention tensors per layer
# Visualize first layer, first head
plt.imshow(attentions[0][0, 0].numpy(), cmap='plasma')
plt.title("Layer 1 - Head 1 Attention")
plt.xlabel("Source Tokens")
plt.ylabel("Target Tokens")
plt.xticks(range(len(inputs.input_ids[0])), [tokenizer.decode(tok) for tok in inputs.input_ids[0]])
plt.yticks(range(len(inputs.input_ids[0])), [tokenizer.decode(tok) for tok in inputs.input_ids[0]])
plt.colorbar()
plt.show() please let me know if you think it's worth keeping or if it would be better to discard it. about the task of integrating another query for model card --- i need to update the albert.md right? |
I think it'd be better to discard it and instead try to integrate it with the existing
Yeah this is the only file you need to update! |
Hey! @stevhliu, can I work on granite? |
Hey friends! 👋
We are currently in the process of improving the Transformers model cards by making them more directly useful for everyone. The main goal is to:
Pipeline
,AutoModel
, andtransformers-cli
with available optimizations included. For large models, provide a quantization example so its easier for everyone to run the model.Compare the before and after model cards below:
With so many models in Transformers, we could really use some a hand with standardizing the existing model cards. If you're interested in making a contribution, pick a model from the list below and then you can get started!
Steps
Each model card should follow the format below. You can copy the text exactly as it is!
For examples, take a look at #36469 or the BERT, Llama, Llama 2, Gemma 3, PaliGemma, ViT, and Whisper model cards on the
main
version of the docs.Once you're done or if you have any questions, feel free to ping @stevhliu to review. Don't add
fix
to your PR to avoid closing this issue.I'll also be right there working alongside you and opening PRs to convert the model cards so we can complete this faster together! 🤗
Models
The text was updated successfully, but these errors were encountered: