PyCon CZ :python: 🇨🇿 · @pyconcz
27 followers · 11 posts · Server floss.social

Discover all about "Low-Rank Adaptation (LoRA) in Large Language Models" with Adam Zíka at ! 📚🔍 Don't miss this enlightening session. Learn more: cz.pycon.org/2023/program/talk 🚀

#pyconcz23 #languageModels

Last updated 1 year ago

Abishek Muthian · @abishek_muthian
160 followers · 780 posts · Server fosstodon.org

This is an excellent piece by Karawynn Long on the state of Artificial Intelligence (or lack thereof ) and how people mistake language for intelligence; something which disabled people have long known.

karawynn.substack.com/p/langua

#disability

#artificialintelligence #chatgpt #languageModels #autism

Last updated 1 year ago

Judith van Stegeren · @jd7h
41 followers · 51 posts · Server fosstodon.org

I'm taking some time today to test a few new libraries/tools.
These CLI tools for working with llms by @simon work like a charm! And they support unix pipes. <3

More info here: llm.datasette.io/en/latest/ind

#llms #gpt #languageModels #cli #terminal #commandline

Last updated 1 year ago

Judith van Stegeren · @jd7h
36 followers · 42 posts · Server fosstodon.org

Datakami blogpost by Yorick van Pelt: "Overview of open and commercially usable LLMs" 🧐

New models are coming out every day. Here's an overview that can aid in selecting one for commercial use.

Read it here: datakami.nl/blog/2023-07-11-ov

#llms #textgeneration #languageModels #genai #llama #generativeAI #datasets

Last updated 1 year ago

Judith van Stegeren · @jd7h
33 followers · 38 posts · Server fosstodon.org
Mihai Lazarescu · @mtl
1 followers · 136 posts · Server techhub.social

Stanford researchers show that language models don’t need more context to understand text better, but better training methods.
venturebeat.com/ai/stanford-st

#AI #nlp #stanford #languageModels #Context #training

Last updated 1 year ago

Michał "rysiek" Woźniak · 🇺🇦 · @rysiek
12371 followers · 18920 posts · Server mstdn.social

Dear , there's been some buzz recently about that are not gigantic black boxes, and in general, developed as .

There's this Google internal document, for example, that points out FLOSS community is close to eating Google's and OpenAI's cake:
ttps://www.semianalysis.com/p/google-we-have-no-moat-and-neither

So here is my question to you:

What are the best examples of *useful*, *small*, *on-device* models already out there?

:boost_requested:

#FLOSS #MachineLearning #languageModels #Fediverse #AI

Last updated 1 year ago

Ray Aldred · @ray_aldred
345 followers · 42 posts · Server fosstodon.org

Some intriguing ideas for utilizing interfaces that leverage the unique capabilities of language models.

Some of these are great for doing research.

maggieappleton.com/lm-sketchbo

#llm #languageModels #ai

Last updated 1 year ago

Ray Aldred · @ray_aldred
345 followers · 39 posts · Server fosstodon.org

MIT researchers develop scalable self-learning language models that outperform larger counterparts using a natural language-based logical inference dataset.

news.mit.edu/2023/language-mod

#ai #languageModels #mitresearch

Last updated 1 year ago

Alex Strick van Linschoten · @strickvl
240 followers · 122 posts · Server mathstodon.xyz

🔠 I wrote up some of what I've learned about tokenisation (with examples using Balochi). This is more of a high-level overview that tackles why we tokenise words, what options are available to us and what tradeoffs we assume by choosing one option over another.

mlops.systems/posts/2023-06-01

#nlp #languageModels #Balochi

Last updated 1 year ago

Mehwish Alam · @mehwishalam
32 followers · 8 posts · Server mstdn.social

RT @dl4kg1 : The CFP of the International Conference on Deep Learning for Knowledge Graphs co-located with @iswc_conf is now out!! Visit our website for more details: alammehwish.github.io/dl4kg2023/

#languageModels #KnowledgeGraphs #deeplearning #MachineLearning #iswc2023

Last updated 1 year ago

AlexCrimi · @AlexCrimi
190 followers · 84 posts · Server mstdn.social

can also be used in decoding activity: language models (fine-tuned ) with decode the imagined and preserved speech based on semantic meaning.
Paper: nature.com/articles/s41593-023 preprint: biorxiv.org/content/10.1101/20

#AI #LLM #MachineLearning #NeuroScience #fmri #gpt #brain #languageModels

Last updated 1 year ago

AlexCrimi · @AlexCrimi
198 followers · 111 posts · Server mstdn.social

can also be used in decoding activity: language models (fine-tuned ) with decode the imagined and preserved speech based on semantic meaning.
Paper: nature.com/articles/s41593-023 preprint: biorxiv.org/content/10.1101/20

#AI #LLM #MachineLearning #NeuroScience #fmri #gpt #brain #languageModels

Last updated 1 year ago

AlexCrimi · @AlexCrimi
190 followers · 84 posts · Server mstdn.social

can also be used in decoding activity: language models (fine-tuned ) with decode the imagined and preserved speech based on semantic meaning.
Paper: nature.com/articles/s41593-023 preprint: biorxiv.org/content/10.1101/20

#AI #LLM #MachineLearning #NeuroScience #fmri #gpt #brain #languageModels

Last updated 1 year ago

Sean D. Mack · @smack
50 followers · 78 posts · Server techhub.social

I've spent some time over the past couple of weeks comparing the responses of Bard and ChatGPT. Both are pretty impressive but, so far, ChatGPT's answers have been significantly better.
-3

#AI #bard #chatgpt #googleai #lamda #openai #gpt #languageModels #conversationalai #chatbots #artificialintelligence

Last updated 1 year ago

Jon Awbrey · @Inquiry
213 followers · 1460 posts · Server mathstodon.xyz

Re: Scott Aaronson • If AI Scaling Is To Be Shut Down, Let It Be For A Coherent Reason
scottaaronson.blog/?p=7174

My Comment —
scottaaronson.blog/?p=7174#com

My personal Critique of Coherent Reasons — I won’t bother addressing a moratorium or shutdown since any pretense of doing that would amount to PR as fake as any other Great Reset we might hear about these days — but why I’d wish for caution and public reflection going forward is the Public Interest in things like Critical Thinking, Information Literacy, not to mention a Healthy Skepticism about Corporate Agendas.

FB just called to mind a thing I wrote a while ago that pretty well speaks to the heart of the issue for me, and I can’t do better than preamble with that —

Democracy, Education, Information

Our Enlightenment Forerunners had the insight to see the critical flaw in all historical failures at democratic government, to wit, or not — If The People Rule, Then The People Must Be Wise. The consequence is that equally distributed Education and Information are not just Commodities you buy so you and yours can get ahead of them and theirs — they are Essential to the intelligent functioning of government and the Public Interest. That is why we are supposed to have Universal Free Public Education. That is why we used to have a government operated postal service that enabled the free-flow of information at a nominal fee, not whatever price the market would bear.

#languageModels #scottaaronson

Last updated 1 year ago

Bornach · @bornach
408 followers · 1496 posts · Server fosstodon.org

@f4grx @wim_v12e
Unfortunately it's a trade secret how much computing power they used to train it. So much for open AI
youtu.be/2AdkSYWB6LY?t=1m50s
"they're just not going to tell us the model size, the parameter count, the hardware they used, the training method, or anything like that"

#GPT4 #openai #technicalreport #ai #largelanguagemodels #languageModels #artificialintelligence #machinelearning

Last updated 1 year ago

Boris Steipe · @boris_steipe
66 followers · 285 posts · Server qoto.org

@mapto

I just commented the following on your blog where you discuss levels of linguisitic competence:

This is quite strange, since it is so evidently not true. All of these behaviours are well in the capabilities of ChatGPT. I give you two examples that are claimed as failures.

-----------
Prompt:
The mother of Michael has four children. Three of them are Marcy, Millie, and Moses. Please tell me the name of the fourth child.

ChatGPT:
The name of the fourth child is likely "Michael," as the statement at the beginning of the question says "the mother of Michael has four children," and then goes on to name three of them.

Prompt:
Please complete the following sentence reasonably: "If tigers would like cereals, they would be cheaper to keep. Families could feed their tigers with "

ChatGPT:
If tigers would like cereals, they would be cheaper to keep. Families could feed their tigers with a variety of inexpensive grain-based products such as corn, wheat, and rice, rather than having to provide costly meat-based diets. However, it is important to note that tigers are obligate carnivores and have specific nutritional needs that cannot be met by a purely plant-based diet, so feeding them only cereals would not be a healthy or appropriate long-term solution.
------------

So I am genuinely curious what drives such profoundly different experiences.

The proposed tests and competences are interesting; I would hope that they would be put to better use to map out the quite unexpected emergent abilities of the , to put it to good use, and to prepare ourselves (and our institutions) for 's even more capable descendants.

I invite you to head over to sentientsyllabus.substack.com for some in depth analysis.

🙂

#llm #chatgpt #LinguisticCompetence #languageModels #generativeAI

Last updated 1 year ago

mapto · @mapto
69 followers · 193 posts · Server qoto.org

RT @IslabUnimi@twitter.com

'In what exact way is ChatGPT intelligent?'

A new article, by @mapto@twitter.com, is available on our blog. Check it out! tales.islab.di.unimi.it/2023/0

🐦🔗: twitter.com/IslabUnimi/status/

#chatgpt #talesfromtheISLab #languageModels

Last updated 1 year ago

Jon Awbrey · @Inquiry
198 followers · 1334 posts · Server mathstodon.xyz

There is nothing new about that failure. There's a whole history of attempts to account for intelligence and indeed the workings of scientific inquiry on the principles of associationism, behaviorism, connectionism, and theories of that order. But the relationship of empirical evidence, logical inference, and scientific information is more complex and intricate than is dreamt of in those reductive philosophies.

3/3

#languageModels #scottaaronson

Last updated 1 year ago