| | |
| |
|
| | import torch |
| | from transformers import AutoTokenizer, M2M100ForConditionalGeneration |
| |
|
| | |
| | DEVICE = "cuda" if torch.cuda.is_available() else "cpu" |
| | nepali_model_path = r"D:\SIH\saksi_translation\models\nllb-finetuned-nepali-en" |
| |
|
| | |
| | print("Loading Nepali tokenizer...") |
| | try: |
| | nepali_tokenizer = AutoTokenizer.from_pretrained(nepali_model_path) |
| | print("Nepali tokenizer loaded successfully.") |
| | print(nepali_tokenizer) |
| | except Exception as e: |
| | print(f"Error loading Nepali tokenizer: {e}") |
| |
|
| | |
| | print("\nLoading Nepali model...") |
| | try: |
| | nepali_model = M2M100ForConditionalGeneration.from_pretrained(nepali_model_path).to(DEVICE) |
| | print("Nepali model loaded successfully.") |
| | print(nepali_model) |
| | except Exception as e: |
| | print(f"Error loading Nepali model: {e}") |
| |
|