Nikhitha Telugu Dataset Model Model ID: Nikitha-logics/Nikhitha_telugu_dataset_model
Model Type: ALBERT-based Language Model
License: Apache-2.0
Model Overview The Nikhitha Telugu Dataset Model is an ALBERT-based language model trained on a Telugu language dataset. ALBERT (A Lite BERT) is a transformer-based model designed for natural language processing tasks, optimized for efficiency and performance.
Model Details Model Size: 33.2 million parameters
Tensor Type: Float32 (F32)
Format: Safetensors
Usage To utilize this model in your projects, you can load it using the Hugging Face Transformers library:
from transformers import AlbertTokenizer, AlbertForMaskedLM
Load the tokenizer
tokenizer = AlbertTokenizer.from_pretrained("Nikitha-logics/Nikhitha_telugu_dataset_model")
Load the model
model = AlbertForMaskedLM.from_pretrained("Nikitha-logics/Nikhitha_telugu_dataset_model")
- Downloads last month
- 2