ML Skill - English to German

Hi Team ,

I am using the Enterprise Trial for using the AI Fabric and using the ML Skill - English to German Translation.

Is there any limitation on the no. of characters or number of lines the model can process.

If I just write one or 2 line or text - we get the correct response , but if we try the model with a paragraph , the output is not proper. Some random text is appearing.

Attaching the sample File -
Model = Englsih to German Translation.

English Terms and Conditions.txt (8.9 KB)

Hello @mukeshkala!

It seems that you have trouble getting an answer to your question in the first 24 hours.
Let us give you a few hints and helpful links.

First, make sure you browsed through our Forum FAQ Beginner’s Guide. It will teach you what should be included in your topic.

You can check out some of our resources directly, see below:

  1. Always search first. It is the best way to quickly find your answer. Check out the image icon for that.
    Clicking the options button will let you set more specific topic search filters, i.e. only the ones with a solution.

  2. Topic that contains most common solutions with example project files can be found here.

  3. Read our official documentation where you can find a lot of information and instructions about each of our products:

  4. Watch the videos on our official YouTube channel for more visual tutorials.

  5. Meet us and our users on our Community Slack and ask your question there.

Hopefully this will let you easily find the solution/information you need. Once you have it, we would be happy if you could share your findings here and mark it as a solution. This will help other users find it in the future.

Thank you for helping us build our UiPath Community!

Cheers from your friendly
Forum_Staff

Hi @mukeshkala,

the model used by the OOB package is called fairseq. It has a context size of 1024 so-called “tokens”, which in this case corresponds roughly to words and punctuation marks. If you use more it should throw an error, I’m a little confused as to why it gives garbage back instead… I can reproduce the same behavior on my end, though, so it’s definitely not your fault.

The way we use it does not currently allow translating multiple text sequences in batches, so to be on the safe side, I would recommend splitting the text into chunks of 900 or so words in the RPA workflow and invoking the ML skill in a loop.

I recently looked into modifying the model to allow batch translation and also wrap both en_de and de_en into one package, but will have to check with the AI Fabric team on whether we want to add that to the OOB models. It’s not very difficult to download the pretrained fairseq model and write a small wrapper around it if you have someone who knows python.

2 Likes

Thank You @alpaca ,

The Response helps, As of now I will split the text into chunks.

For Your Reference, I have attached the the Notepad file in the First Post-

Step I Performed -

  1. Read it using the Read text file .
  2. string variable was being passed to the Model.
  3. Get the Output from Model.

Thanks,
Mukesh Kala

1 Like