MOULE :Neurodivergent: · @MOULE
404 followers · 1146 posts · Server mastodon.moule.world
MOULE :Neurodivergent: · @MOULE
404 followers · 1146 posts · Server mastodon.moule.world
_dmh · @_dmh
173 followers · 552 posts · Server mastodon.social

Similarly, you may want a so-called , some structured representation of the information conveyed by a text or a linguistically motivated semantic representation of the text. These annotations are essential for my work in / , and a big struggle for the community is coming up with ways to build for different tasks, domains, genres, or languages which also have the kinds of MRs our systems use.

#meaningrepresentation #naturallanguagegeneration #nlg #corpora

Last updated 2 years ago

_dmh · @_dmh
173 followers · 544 posts · Server mastodon.social

Yesterday I began writing about , a series of threads where I talk about what the topics highlighted in my profile mean to me.

Today it's about building, which is the process of finding, building, and/or preparing a linguistic dataset. This is essential work for and a lot of / systems, as your models can only be as good as the data they are based on.

(Yesterday's thread about : mastodon.social/@_dmh/10930309)

#whathashtagsmeantome #corpus #naturallanguageprocessing #ai #ml #naturallanguagegeneration

Last updated 2 years ago

_dmh · @_dmh
137 followers · 486 posts · Server mastodon.social

In my own work, I'm working to adapt simpler models which are not as data-hungry to perform data-to-text in settings. Part of this effort will involve exploring and approaches to neural .

Follow me if you want more content related to these topics, though be advised that this is a whole-person account and I will talk about things other than work as well.

#nlg #lowresource #multitasklearning #pipeline #naturallanguagegeneration

Last updated 2 years ago

_dmh · @_dmh
137 followers · 486 posts · Server mastodon.social

This is not to say, however, that I think these models are useless. I think the interesting question is how to integrate these models into systems that express a particular meaning, a la data-to-text . Whether this involves , integrating them into the decoder for models, or some other more clever application remains to be seen. I am looking forward to seeing how /s get used for going forward.

#naturallanguagegeneration #PromptEngineering #seq2seq #llm #nlg

Last updated 2 years ago

_dmh · @_dmh
137 followers · 481 posts · Server mastodon.social

In the years since then there has been an explosive growth in interest in , but usually not grounded in expressing a particular meaning, which was historically a priority for work in . This excitement came from the relative fluency of () and models in particular, which did an impressive job of continuing an initial utterance.

#textgeneration #naturallanguagegeneration #largelanguagemodels #llm #transformer

Last updated 2 years ago

_dmh · @_dmh
137 followers · 481 posts · Server mastodon.social

Around 2015 and 2016 we saw sequence-to-sequence () models applied to data-to-text for the first time. These models were trained end-to-end and were very exciting because it raised the prospect of reducing the amount of hand-crafted one would have to do to create a system.

#seq2seq #nlg #grammarengineering #naturallanguagegeneration

Last updated 2 years ago

_dmh · @_dmh
137 followers · 480 posts · Server mastodon.social

Since the revolution in came to (/#NLProc), these tools have also become very common in . Of course, in the late 90s folks were already incorporating statistical information into their systems in an approach called 'overgenerate-and-rank', where rules were underspecified and produced grammatical and ungrammatical utterances and one relied on n-gram frequencies to rank the possible outputs correctly

#neuralnetwork #machinelearning #nlp #naturallanguageprocessing #naturallanguagegeneration #nlg

Last updated 2 years ago

_dmh · @_dmh
137 followers · 479 posts · Server mastodon.social

/#NLG has often relied on rule-based approaches to generation with a pipeline of processes. For example, a content selection module might process raw data to decide what is worth mentioning, with the result being fed into a sentence planning module to determine how that information should be expressed broadly (for example, which words to use, how to group the info into sentences, etc), and finally that sentence plan going through 'surface realisation' to become text

#naturallanguagegeneration

Last updated 2 years ago

_dmh · @_dmh
137 followers · 478 posts · Server mastodon.social

But can broadly apply to all sorts of tasks. Even () is a kind of . And other text-to-text transformations like summarisation and style-transfer also fall under the umbrella of what I call big-tent .

#naturallanguagegeneration #machinetranslation #mt #nlg

Last updated 2 years ago

_dmh · @_dmh
137 followers · 477 posts · Server mastodon.social

For my work, I usually use to refer to so-called data-to-text generation, where we are transforming some *non-linguistic* representation of information into a natural language utterance (whether that's spoken or written, long or short, formal or informal, etc)

#nlg #naturallanguagegeneration #data2text #d2t

Last updated 2 years ago

_dmh · @_dmh
137 followers · 480 posts · Server mastodon.social

Every day that I remember to do it, I'm going to explain what I mean by one of the hashtags in my profile.

I'm at the office, so today is about professional stuff: () involves developing computer systems which can express information in natural language (that is, human languages). There are a bunch of different approaches to NLG and a lot of different tasks that can be considered a part of big-tent NLG.

#whathashtagsmeantome #naturallanguagegeneration #nlg

Last updated 2 years ago

_dmh · @_dmh
137 followers · 480 posts · Server mastodon.social

Perhaps time for a new

I'm a mid-30s he/they kinda guy who works as a researcher in (specifically , ).

I like singing, running, and gaming. I'm queer, polyamorous, Catholic, and an immigrant 3 times over.

Dad to two beautiful voids, my furbabies Salem and Cricket.

#introduction #computationallinguistics #nlg #naturallanguagegeneration

Last updated 2 years ago