Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Comments

Add RoBERTa model for question answering using SQuADv2.0 dataset.#19

Open
QingtaoLi1 wants to merge 3 commits intodotnet:main from
QingtaoLi1:roberta-qa
Open

Add RoBERTa model for question answering using SQuADv2.0 dataset. #19
QingtaoLi1 wants to merge 3 commits intodotnet:main from
QingtaoLi1:roberta-qa

Conversation

@QingtaoLi1
Copy link

@QingtaoLi1 QingtaoLi1 commented Apr 8, 2022

No description provided.

GeorgeS2019 reacted with thumbs up emoji
Copy link

dnfadmin commented Apr 8, 2022
edited
Loading

CLA assistant check
All CLA requirements met.

Copy link

GeorgeS2019 commented Apr 20, 2022
edited
Loading

@QingtaoLi1 thanks for submitting a PR related to TorchText or Pytorch Deep NLP/Pytorch for HuggingFace.

I wonder if there is need now to start thinking how to organize Torchsharp so NLP tutorials and use cases are more aligned with TorchText.

Just an unrelated question, have you looked into Huggingface GPT2?

Copy link
Author

Just an unrelated question, have you looked into Huggingface GPT2?

I have not used GPT2 yet. The main difference between the GPTs and BERTs is that the former is a generative model which can output a sequence in an auto-regression format. So I believe many of code are very similar except for generation-related parts.

Copy link

GeorgeS2019 commented Apr 20, 2022
edited
Loading

@QingtaoLi1

This is an attempt to get GPT2 working using BlingFire GPT2 tokeniser and GPT2 ONNX with the preliminary generation-related parts which are the issue

If you have time and interest, hopefully, get your perspective of the challenges of getting the generation parts working.

Copy link
Author

@GeorgeS2019 Sorry but I wonder what you would like to achieve with GPT2 or BlingFire/ONNX?

Copy link

GeorgeS2019 commented Apr 20, 2022
edited
Loading

RoBERT is for question and answer.
GPT2 is for Text Generation or NLG, which is rare in .NET

I hope this answer your question.

Microsoft BlingFire...there should be a tokeniser for RoBERT...Do check out..better performance?

Copy link

@QingtaoLi1 updated the reply

Copy link
Author

QingtaoLi1 commented Apr 20, 2022
edited
Loading

I see. You want to include GPT2 in this repo, right? I currently have other work to do; may seek for some time later.

And I know some are attempting to create infrastructure for general tokenizers. I guess your thoughts of re-organizing TorchSharpExamples to align with torchtext will have similar effect on NLP systems -- this will make it easier to build different NLP models for different tasks/datasets.

Copy link

The spirit of Torchshap is to empower. NET developers to do deep AI within .NET without the need to go back to python.

Copy link
Author

I'm not turning back to python; I mean infrastructure in .NET world.

GeorgeS2019 reacted with thumbs up emoji

Copy link
Author

An unrelated question: what do you think are the main obstacles for developers to build a deep AI system?

Copy link

Copy link

Copy link
Author

Thanks very much!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Reviewers

No reviews

Assignees

No one assigned

Labels

None yet

Projects

None yet

Milestone

No milestone

Development

Successfully merging this pull request may close these issues.

AltStyle によって変換されたページ (->オリジナル) /