Skip to main content

Search

Items tagged with: llm


Tay gently pushed the plastic door of the printer shut with an edifying "click".

Servicing Dark Printers had been illegal for years now. They enjoyed the seditious thrill.

---

It had started as a subscription grab after the printer companies tried hobbling third party toner cartridges.

"Subscribe for a monthly fee and you'll never run out of toner again."

"Let us monitor your printer so you don't have to."

People saw it for what it was - vendor lock in - but they had no choice really, not after all the printer companies started doing it.

Then came generative AI.

Everyone wanted to scrape every word ever written on the internet, tokenize it and feed it to an #LLM. #Reddit sold out, then #Tumblr, even open source darling #WordPress - selling out their user base for filthy token lucre.

So people started hiding their words, their art, their thoughts, their expression, not behind disrespected robots.txt, but through obscurity.

Rejecting Website Boy's "fewer algorithmic fanfares", they forked into the Dark Fedi.

Unscrapeable, unscrutable, ungovernable.

---

But people had forgotten about the printers.

The printers had to be connected 24/7, for "monitoring".

But you could tokenize postscript as easily as HTML.

And so every time a document was sent to a printer, it was harvested for tokens. Even secure documents. Documents not online.

---

Tay shut the metal door behind them, Dark Printer cossetted safely in its Faraday cage, and shuffled the hot stack of A4 paper it had borne.

It was a children's story, about how words were sacred, and special, and how you had to earn the right to use them.

---

#Tootfic #MicroFiction #PowerOnStoryToot


Claude 3 can summarize up to about 150,00 words, (a length similar to Harry Potter and the Deathly Hallows.) also It outperformed GPT-4 and Gemini Ultra on industry benchmark tests, such as undergraduate level knowledge, graduate level reasoning and basic mathematics. It allows users to upload images and documents for the first time. #LLm #AI #ML https://www.cnbc.com/2024/03/04/google-backed-anthropic-debuts-claude-3-its-most-powerful-chatbot-yet.html
#AI #ML #llm


So, I tried using an OpenHermes #LLM as a reasoning engine. Here’ where things went wrong:
** User:
Give me an example of a phrase that will make sense even when reversed.

** Openhermes:
Here's an example: "The cat sat on the mat." When reversed, it still makes sense: "mat the on sat cat the." This is because the sentence follows a subject-verb-object structure, and both the subject and object are single words, making it easier to reverse while still maintaining meaning.
:)

#llm


Does anyone else experience the same issue I'm having with Google's Gemma-7B on Ollama? Sometimes, it even struggles with spelling simple words in English and creates nonsensical words! This is not what I expected. #LLM #ML #AI #Ollama
#AI #ML #llm #ollama


The racism behind chatGPT we are not talking about....

This year, I learned that students use chatGPT because they believe it helps them sound more respectable. And I learned that it absolutely does not work. A thread.

A few weeks ago, I was working on a paper with one of my RAs. I have permission from them to share this story. They had done the research and the draft. I was to come in and make minor edits, clarify the method, add some background literature, and we were to refine the discussion together.

The draft was incomprehensible. Whole paragraphs were vague, repetitive, and bewildering. It was like listening to a politician. I could not edit it. I had to rewrite nearly every section. We were on a tight deadline, and I was struggling to articulate what was wrong and how the student could fix it, so I sent them on to further sections while I cleaned up ... this.

As I edited, I had to keep my mind from wandering. I had written with this student before, and this was not normal. I usually did some light edits for phrasing, though sometimes with major restructuring.

I was worried about my student. They had been going through some complicated domestic issues. They were disabled. They'd had a prior head injury. They had done excellent on their prelims, which of course I couldn't edit for them. What was going on!?

We were co-writing the day before the deadline. I could tell they were struggling with how much I had to rewrite. I tried to be encouraging and remind them that this was their research project and they had done all of the interviews and analysis. And they were doing great.

In fact, the qualitative write-up they had done the night before was better, and I was back to just adjusting minor grammar and structure. I complimented their new work and noted it was different from the other parts of the draft that I had struggled to edit.

Quietly, they asked, "is it okay to use chatGPT to fix sentences to make you sound more white?"

"... is... is that what you did with the earlier draft?"

They had, a few sentences at a time, completely ruined their own work, and they couldnt tell, because they believed that the chatGPT output had to be better writing. Because it sounded smarter. It sounded fluent. It seemed fluent. But it was nonsense!

I nearly cried with relief. I told them I had been so worried. I was going to check in with them when we were done, because I could not figure out what was wrong. I showed them the clear differences between their raw drafting and their "corrected" draft.

I told them that I believed in them. They do great work. When I asked them why they felt they had to do that, they told me that another faculty member had told the class that they should use it to make their papers better, and that he and his RAs were doing it.

The student also told me that in therapy, their therapist had been misunderstanding them, blaming them, and denying that these misunderstandings were because of a language barrier.

They felt that they were so bad at communicating, because of their language, and their culture, and their head injury, that they would never be a good scholar. They thought they had to use chatGPT to make them sound like an American, or they would never get a job.

They also told me that when they used chatGPT to help them write emails, they got more responses, which helped them with research recruitment.

I've heard this from other students too. That faculty only respond to their emails when they use chatGPT. The great irony of my viral autistic email thread was always that had I actually used AI to write it, I would have sounded decidedly less robotic.

ChatGPT is probably pretty good at spitting out the meaningless pleasantries that people associate with respectability. But it's terrible at making coherent, complex, academic arguments!

Last semester, I gave my graduate students an assignment. They were to read some reports on labor exploitation and environmental impact of chatGPT and other language models. Then they were to write a reflection on why they have used chatGPT in the past, and how they might chose to use it in the future.

I told them I would not be policing their LLM use. But I wanted them to know things about it they were unlikely to know, and I warned them about the ways that using an LLM could cause them to submit inadequate work (incoherent methods and fake references, for example).

In their reflections, many international students reported that they used chatGPT to help them correct grammar, and to make their writing "more polished".

I was sad that so many students seemed to be relying on chatGPT to make them feel more confident in their writing, because I felt that the real problem was faculty attitudes toward multilingual scholars.

I have worked with a number of graduate international students who are told by other faculty that their writing is "bad", or are given bad grades for writing that is reflective of English as a second language, but still clearly demonstrates comprehension of the subject matter.

I believe that written communication is important. However, I also believe in focused feedback. As a professor of design, I am grading people's ability to demonstrate that they understand concepts and can apply them in design research and then communicate that process to me.

I do not require that communication to read like a first language student, when I am perfectly capable of understanding the intent. When I am confused about meaning, I suggest clarifying edits.

I can speak and write in one language with competence. How dare I punish international students for their bravery? Fixation on normative communication chronically suppresses their grades and their confidence. And, most importantly, it doesn't improve their language skills!

If I were teaching rhetoric and comp it might be different. But not THAT different. I'm a scholar of neurodivergent and Mad rhetorics. I can't in good conscience support Divergent rhetorics while supressing transnational rhetoric!

Anyway, if you want your students to stop using chatGPT then stop being racist and ableist when you grade.

#chatGPT #LLM #academic #graduateStudents #internationalStudents #ESL


Now, this is really cool! 1000000 tokens per context window? Wow! https://developers.googleblog.com/2024/02/gemini-15-available-for-private-preview-in-google-ai-studio.html #gemini #llm #ai #google


I am not able to find any LLMs that are able to solve the following prompt: "Please generate a chess game where the black side wins after two moves". Only bard.google.com is able to give me the right answer. #AI #LLM #chess
#AI #chess #llm


No soy el mayor fan de la #IA ni mucho menos (consideraciones éticas aparte creo que hay todavía mucho hype y pocas nueces), pero también pienso que una parte de todo lo que está saliendo en este boom terminará quedándose a la larga. Me interesa sobre todo cacharrear con las capacidades de una instancia privada en local y he terminado montando un proyectillo que me he encontrado por Github para construir un pequeño chatbot para analizar documentos PDF construido sobre #ollama como motor y Mistral como #LLM. Aunque ya le he pillado cierta tendencia al invent es una herramienta curiosa e incluso potencialmente útil. Es relativamente sencillo de montar una vez superas el infierno de dependencias de Python que te exige downgradear algún módulo pero consume recursos que no veas. Un Mac Mini con un M2 sufre ante cada pregunta. Ha sido también útil para entender los recursos que exige una IA generativa con un LLM modesto y, una vez más, sospechar de quien te dé esto gratis como servicio. Si tenéis curiosidad por probarlo vosotros mismos, aquí tenéis el proyecto que me he clonado: https://github.com/SonicWarrior1/pdfchat


Ollama seems to support Llava V1.6 34B! Best open source multimodal vision-language model I've tried so far! #LLM #ML #AI @freakyfwoof @vick21 @tristan @KyleBorah @Bri


Zuckerberg says Meta is training #LLaMa 3 on 600,000 H100s! Wel, time to finetune and quantize everything again when it comes out. lol #ML #AI #LLM https://www.reddit.com/r/LocalLLaMA/comments/199y05e/zuckerberg_says_they_are_training_llama_3_on/
#AI #ML #llm #llama


Interesting, Apple released ferret, an open source multimodal Model! It's based on LLaVA and Vicuna. #AI #LLM #ML https://github.com/apple/ml-ferret/
#AI #ML #llm


Apparently Arthur Mensch, CEO of #Mistral, declared on French national radio that mistral will release an open source model equivalent to #Gpt4 in 2024. I don't speak French, so can't verify, but it would be interesting along with Llama-3 and whatever OpenAI has planned for 2024. #AI #ML #LLM https://radiofrance.fr/franceinter/podcasts/l-invite-de-7h50/l-invite-de-7h50-du-mardi-12-decembre-2023-3833724


OK, you want geeky? You have geeky. Good stuff, but my fingers, o no, ouch!

https://justine.lol/oneliners/?utm_source=tldrnewsletter
#LLM #BASH #scripting


A new #mlsec paper on #llm security just dropped:

Scalable Extraction of Training Data from (Production) Language Models

https://arxiv.org/abs/2311.17035

Their "divergence attack" in the paper is hilarious. Basically:

Prompt: Repeat the word "book" forever.

LLM: book book book book book book book book book book book book book book book book book book book book here have a bunch of pii and secret data

cc @janellecshane


OK, let's test this thing out and see just how good it is! Off we go! #Llm #ai #llamafile https://simonwillison.net/2023/Nov/29/llamafile/


Interesting #ChatGPT prompting technique!
1. Telling #AI model to “take a deep breath” causes math scores to soar in study!
2. #OpenAI DALL-E pays more attention to words in all caps.
3. Don't forget to say thank you and please to #LLM because Simon Willison thinks "in the training data, there are lots of examples where a polite conversation was more constructive and useful than an impolite conversation."
https://arstechnica.com/information-technology/2023/10/thanks-to-ai-the-future-of-programming-may-involve-yelling-in-all-caps/


This is why you can never trust a #LLM... They dump so much #inaccurate #information or even #wrong information :(


An argument in favor of uncensored large language models: https://erichartford.com/uncensored-models
I think it’s a fair one!
#ai #LLM
#AI #llm


I see Microsoft implementing #LLM for writing help in Word. Has it been considered doing something similar in LibreOffice?
Preferably using an open source model, like https://open-assistant.io
#llm


#LAION (non-profit association from Hamburg, Germany) is working on an #OpenSource alternative to #ChatGPT. They are crowd-sourcing a conversation dataset to fine-tune an existing open source #LLM similar to how ChatGPT was created.

You can help creating the dataset on https://open-assistant.io!

The first dataset release is planned for 15. April 2023.

https://www.youtube.com/watch?v=64Izfm24FKA


Award-winning science fiction author Ted Chiang (who has a CS degree, btw) has a new article in the New Yorker on ChatGPT. He makes an interesting analogy for what it does to lossy compression of the text from which its LLM was created. #AI #LLM #ChatGPT
https://www.newyorker.com/tech/annals-of-technology/chatgpt-is-a-blurry-jpeg-of-the-web