You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Belt_Road_Hungarian

Model Description

This model is a conversational and instruction-following large language model, fine-tuned from the foundational open-source Qwen2.5-72B-Instruct model using supervised fine-tuning (SFT).


Key Features & Use Cases

  • Exceptional Hungarian Language Proficiency: The model has been deeply optimized for Hungarian, demonstrating excellent fluency, accuracy, and a strong understanding of cultural context in conversations.
  • Multilingual Translation and Dialogue: With extensive training data that includes Hungarian, English, and Chinese content, the model excels in translation, multilingual Q&A, and cross-language communication.
  • Advanced Instruction Following: The model shows a strong ability to comprehend and execute complex instructions, including those with multiple steps and specific constraints.
  • Creative Content Generation: It is highly suitable for a wide range of creative tasks, such as writing articles, reports, scripts, and marketing copy.

System Requirements

Hardware

  • GPU VRAM: For BF16/FP16 inference (recommended), at least 4 x NVIDIA A100 (80GB) GPUs are required. The model weights alone are approximately 136GB, so device_map="auto" is necessary to distribute them across multiple cards.
  • System RAM: A minimum of 200GB is recommended.

Software

  • Python: Version 3.10 or higher.
  • Key Libraries:
    • torch: 2.1 or higher
    • transformers: 4.41.0 or higher
    • accelerate: 1.7.0 or higher
    • einops: 0.8.1 or higher
    • sentencepiece: 0.2.0 or higher

How to Use

The recommended method for loading and running the model is by using the transformers library.

# Example code snippet for inference using transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

# Specify the model path or your Hugging Face Hub repository
model_path = "your-huggingface-repo/your-model-name" # e.g., "your-user/your-qwen-model"

# Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_path)

# Load the model with device_map to distribute it across available GPUs
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.bfloat16, # or torch.float16
    device_map="auto"
)

# Example conversation prompt
messages = [
    {"role": "system", "content": "You are a helpful assistant."},
    {"role": "user", "content": "Hello, can you translate 'hello' to Hungarian and Chinese?"}
]

# Apply the chat template and generate a response
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)

generated_ids = model.generate(
    model_inputs.input_ids,
    max_new_tokens=512,
    do_sample=True,
    temperature=0.7,
    top_p=0.9
)
generated_ids = [
    output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support