Instructions to use flexsystems/flex-e2e-super-tiny-bert-model with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use flexsystems/flex-e2e-super-tiny-bert-model with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("feature-extraction", model="flexsystems/flex-e2e-super-tiny-bert-model")# Load model directly from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("flexsystems/flex-e2e-super-tiny-bert-model") model = AutoModel.from_pretrained("flexsystems/flex-e2e-super-tiny-bert-model") - Notebooks
- Google Colab
- Kaggle
| library_name: transformers | |
| tags: [] | |
| # Model Card for Super Tiny Bert | |
| This is a super tiny Bert model for testing purposes. | |
| ## Model Details | |
| This model has been generated using: | |
| ``` | |
| from transformers import BertTokenizer, BertModel, BertConfig | |
| # Define a tiny BERT configuration | |
| config = BertConfig( | |
| vocab_size=30, | |
| hidden_size=8, | |
| num_hidden_layers=2, | |
| num_attention_heads=2, | |
| intermediate_size=8, | |
| max_position_embeddings=8, | |
| ) | |
| # Initialize a tiny BERT model with the custom configuration | |
| model = BertModel(config) | |
| # Create a custom vocabulary | |
| vocab = { | |
| "[PAD]": 0, | |
| "[UNK]": 1, | |
| "[CLS]": 2, | |
| "[SEP]": 3, | |
| "[MASK]": 4, | |
| "hello": 5, | |
| "how": 6, | |
| "are": 7, | |
| "you": 8, | |
| "?": 9, | |
| "i": 10, | |
| "am": 11, | |
| "fine": 12, | |
| "thanks": 13, | |
| "and": 14, | |
| "good": 15, | |
| "morning": 16, | |
| "evening": 17, | |
| "night": 18, | |
| "yes": 19, | |
| "no": 20, | |
| "please": 21, | |
| "thank": 22, | |
| "welcome": 23, | |
| "sorry": 24, | |
| "bye": 25, | |
| "see": 26, | |
| "later": 27, | |
| "take": 28, | |
| "care": 29, | |
| } | |
| # Save the vocabulary to a file | |
| vocab_file = "vocab.txt" | |
| with open(vocab_file, "w") as f: | |
| for token, index in sorted(vocab.items(), key=lambda item: item[1]): | |
| f.write(f"{token}\n") | |
| # Initialize the tokenizer with the custom vocabulary | |
| tokenizer = BertTokenizer(vocab_file=vocab_file) | |
| # Example usage: Tokenize input text | |
| text = "Hello, how are you?" | |
| inputs = tokenizer(text, return_tensors="pt") | |
| # Forward pass through the model | |
| outputs = model(**inputs) | |
| # Extract the last hidden states | |
| last_hidden_states = outputs.last_hidden_state | |
| print("Last hidden states shape:", last_hidden_states.shape) | |
| # Save the tokenizer and model to the Hugging Face Hub | |
| model_name = "flexsystems/flex-e2e-super-tiny-bert-model" | |
| tokenizer.push_to_hub(model_name, private=False) | |
| model.push_to_hub(model_name, private=False) | |
| print(f"Tiny BERT model and tokenizer saved to the Hugging Face Hub as '{model_name}'.") | |
| ``` | |