Spaces:
Sleeping
Sleeping
Fouzi Takelait
commited on
Commit
·
ed3634b
1
Parent(s):
20aa2fc
Update app.py
Browse files
app.py
CHANGED
|
@@ -27,7 +27,7 @@ def translator_fn_baseline(text_in):
|
|
| 27 |
input_ids = source_tokenizer.encode(text_in, return_tensors="pt")
|
| 28 |
output_ids = model.generate(
|
| 29 |
input_ids,
|
| 30 |
-
max_length=len(text_in.split())+
|
| 31 |
bos_token_id=target_tokenizer.bos_token_id,
|
| 32 |
eos_token_id=target_tokenizer.eos_token_id,
|
| 33 |
pad_token_id=target_tokenizer.pad_token_id,
|
|
@@ -43,7 +43,7 @@ def translator_fn_roberta(text_in):
|
|
| 43 |
input_ids_pretrained_roberta = source_tokenizer_pretrained_roberta.encode(text_in, return_tensors="pt")
|
| 44 |
output_ids_pretrained_roberta = model_pretrained_roberta.generate(
|
| 45 |
input_ids_pretrained_roberta,
|
| 46 |
-
max_length=len(text_in.split())+
|
| 47 |
bos_token_id=target_tokenizer_pretrained_roberta.bos_token_id,
|
| 48 |
eos_token_id=target_tokenizer_pretrained_roberta.eos_token_id,
|
| 49 |
pad_token_id=target_tokenizer_pretrained_roberta.pad_token_id,
|
|
|
|
| 27 |
input_ids = source_tokenizer.encode(text_in, return_tensors="pt")
|
| 28 |
output_ids = model.generate(
|
| 29 |
input_ids,
|
| 30 |
+
max_length=len(text_in.split())+3,
|
| 31 |
bos_token_id=target_tokenizer.bos_token_id,
|
| 32 |
eos_token_id=target_tokenizer.eos_token_id,
|
| 33 |
pad_token_id=target_tokenizer.pad_token_id,
|
|
|
|
| 43 |
input_ids_pretrained_roberta = source_tokenizer_pretrained_roberta.encode(text_in, return_tensors="pt")
|
| 44 |
output_ids_pretrained_roberta = model_pretrained_roberta.generate(
|
| 45 |
input_ids_pretrained_roberta,
|
| 46 |
+
max_length=len(text_in.split())+3,
|
| 47 |
bos_token_id=target_tokenizer_pretrained_roberta.bos_token_id,
|
| 48 |
eos_token_id=target_tokenizer_pretrained_roberta.eos_token_id,
|
| 49 |
pad_token_id=target_tokenizer_pretrained_roberta.pad_token_id,
|