Search
Items tagged with: LLM
Thx for your link and efforts @Seirdy !
All this said, being part of a decentralized web, as pointed out in this toot, our publicly visible interaction lands on other instances and servers of the #fediVerse and can be scrapped there. I wonder if this situation actually might lead, or should lead, to a federation of servers that share the same robots.txt "ideals".
As @Matthias pointed out in his short investigation of the AI matter, this has (in my eyes) already unimagined levels of criminal and without any doubt unethical behavior, not to mention the range of options rouge actors have at hand.
It's evident why for example the elongated immediately closed down access to X's public tweets and I guess other companies did the same for the same reasons. Obviously the very first reason was to protect their advantage about the hoarded data sets to train their AI in the first place. Yet, considering the latest behavior of the new owner of #twitter, nothing less than at least the creation of #AI driven lists of "political" enemies, and not only from all the collected data on his platform, is to be expected. A international political nightmare of epical proportions. Enough material for dystopian books and articles for people like @Cory Doctorow, @Mike Masnick ✅, @Eva Wolfangel, @Taylor Lorenz, @Jeff Jarvis, @Elena Matera, @Gustavo Antúnez 🇺🇾🇦🇷, to mention a few of the #journalim community, more than one #podcast episode by @Tim Pritlove and @linuzifer, or some lifetime legal cases for @Max Schrems are at hand.
What we are facing now is the fact that we need to protect our and our users data and privacy because of the advanced capabilities of #LLM. We basically are forced to consider to change to private/restricted posts and close down our servers as not only the legal jurisdictions are way to scattered over the different countries and ICANN details, but legislation and comprehension by the legislators is simply none existent, as @Anke Domscheit-Berg could probably agree to.
Like to say, it looks like we need to go dark, a fact that will drive us even more into disappearing as people will have less chance to see what we are all about, advancing further the advantages off the already established players in the social web space.
Just like Prof. Dr. Peter Kruse stated in his take about on YT The network is challenging us min 2:42 more than 14 years ago:
"With semantic understanding we'll have the real big brother. Someone is getting the best out of it and the rest will suffer."
#Slop is low-quality media - including writing and images - made using generative artificial intelligence technology.
Quelle: Wikipedia.
Open source projects have to deal with a growing number of low-quality vulnerability reports based on AI. See for example this comment from Daniel Stenberg, maintainer of #Curl:
I'm sorry you feel that way, but you need to realize your own role here. We receive AI slop like this regularly and at volume. You contribute to unnecessary load of curl maintainers and I refuse to take that lightly and I am determined to act swiftly against it. Now and going forward.You submitted what seems to be an obvious AI slop "report" where you say there is a security problem, probably because an AI tricked you into believing this. You then waste our time by not telling us that an AI did this for you and you then continue the discussion with even more crap responses - seemingly also generated by AI.
Weiterlesen bei HackerOne: Buffer Overflow Risk in Curl_inet_ntop and inet_ntop4.
#opensource #AI #LLM #Spam
curl disclosed on HackerOne: Buffer Overflow Risk in Curl_inet_ntop...
*Curl is a software that I love and is an important tool for the world. * *If my report doesn't align, I apologize for that.* The `Curl_inet_ntop` function is designed to convert IP addresses from...HackerOne
meta-llama/Llama-3.3-70B-Instruct · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.huggingface.co
This is making the rounds on Finnish social media.
A large association for Finnish construction companies, #Rakennusteollisuus, decided that they needed an English version of their website but apparently they didn't want to pay an actual #translator so just used some free #LLM with hilarious results.
They've fixed it now, but for a short while there was some comedy gold to be found.
P.s. I didn't find these, I've no idea who did.
ChatGPT beat doctors at diagnosing medical conditions, study says
The small study showed AI outperforming doctors by 16 percentage pointsBen Kesslen (Quartz)
I'm a little puzzled at the salience that is being given to the Apple conclusions on #LLM #reasoning when we have lots of prior art. For example: LLMs cannot correctly infer a is b, if their corpora only contain b is a. #Paper: arxiv.org/abs/2309.12288
The Reversal Curse: LLMs trained on "A is B" fail to learn "B is A"
We expose a surprising failure of generalization in auto-regressive large language models (LLMs). If a model is trained on a sentence of the form "A is B", it will not automatically generalize to the reverse direction "B is A".arXiv.org
#AIagent promotes itself to #sysadmin , trashes #boot sequence
Fun experiment, but yeah, don't pipe an #LLM raw into /bin/bash
Buck #Shlegeris, CEO at #RedwoodResearch, a nonprofit that explores the risks posed by #AI , recently learned an amusing but hard lesson in automation when he asked his LLM-powered agent to open a secure connection from his laptop to his desktop machine.
#security #unintendedconsequences
theregister.com/2024/10/02/ai_…
AI agent promotes itself to sysadmin, trashes boot sequence
Fun experiment, but yeah, don't pipe an LLM raw into /bin/bashThomas Claburn (The Register)
🆕 blog! “GitHub's Copilot lies about its own documentation. So why would I trust it with my code?”
In the early part of the 20th Century, there was a fad for "Radium". The magical, radioactive substance that glowed in the dark. The market had decided that Radium was The Next Big Thing and tried to shove it into every product. There …
👀 Read more: shkspr.mobi/blog/2024/10/githu…
⸻
#AI #github #LLM
Independent test of #OpenAI’s o1-preview model achieved near-perfect performance on a national #math exam (landing in the top .1% of the nation’s students).
o1 also outperformed 4o on the math test, but took about 3 times longer to do so (10 minutes vs. 3 minutes).
Preprint: researchgate.net/publication/3…
Massive E-Learning Platform #Udemy Gave Teachers a Gen #AI 'Opt-Out Window'. It's Already Over.
Udemy will train generative AI on classes developed/users contributed on its site. It is opt-out (meaning, everyone was already opted in) with a time window... and opting out may "affect course visibility and potential earnings."
Udemy's reason for the opt-out window was reportedly because removing data from LLMs is hard. IMO, that would be the reason for making it opt-in, but here we are...
#DSGVO versus #LLM / #KI :
Copilot macht aus einem Gerichtsreporter einen Kinderschänder
heise.de/news/Copilot-macht-au…
Recht auf Auskunft? Schwierig. Löschen der Falschinformationen? Unmöglich. Und nun?
Copilot macht aus einem Gerichtsreporter einen Kinderschänder
Weil er über Verhandlungen berichtet hat, macht der Copilot aus einem Journalisten einen Kinderschänder, Witwenbetrüger und mehr.Eva-Maria Weiß (heise online)
Sources:
theinformation.com/briefings/m…
x.com/AlpinDale/status/1814814…
youtu.be/r3DC_gjFCSA?feature=s…
#LLM #AI #ML
Meta Announces Llama 3 at Weights & Biases’ conference
In an engaging presentation at Weights & Biases’ Fully Connected conference, Joe Spisak, Product Director of GenAI at Meta, unveiled the latest family of Lla...YouTube
Also @Tutanota , as a privacy focused company, why are your comments run on #Reddit, rather than the #Fediverse? Most privacy minded folks don't want their comments being used by AI LLM's etc.
This is way off brand.
@carlschwan among others have already shown how to do it fedistyle, pretty easily, and I'm sure many of the FOSS Fedipeeps here would happily help you out with a quick transition if you asked or gave a few $ to their FOSS project.
GPT 4 hallucination rate is 28.6% on a simple task: citing title, author, and year of publication medium.com/@michaelwood33311/g…
Hallucination rates and reference accuracy of ChatGPT and Bard for systematic reviews: Comparative analysis jmir.org/2024/1/e53164 #AI #LLM
GPT 4 Hallucination Rate is 28.6% on a Simple Task: Citing Title, Author, and Year of Publication
The all-too-common myth of GPT 4 having only a 3% hallucination rate is shattered by a recent study that found GPT 4 has a 28.6% hallucination rate. That’s almost 10x the oft-cited (i.e. over hyped)…Michael Wood (Medium)
Blind writer tries the Gandalf | Lakera prompt injection game for the first time.
Upon recommendations, I tried this AI prompt injection game for the first time. I made it to level 7 with no help from the internet!
If you want to donate to me, donate to me on this page.
My website is here where I usually blog. I'm not much of a video person, so I blog and write more than I do video!
Gandalf | Lakera – Test your AI hacking skills
Trick Gandalf into revealing information and experience the limitations of large language models firsthand.gandalf.lakera.ai
Do you remember a couple of weeks ago when I complained that a very large #python contribution to #inkscape was poorly formatted and I felt embarrassed about pushing back and asking them to run a linter over it?
Yeah I'm not fucking embarrassed now, I'm furious. 🤬
Update: Apparently they meant a small section of it was, not the whole MR. I'm annoyed, but I'll have to take them at their word.
#llm #oss #foss #mergerequest
Dnešný fail ruzzkého trolla bol tak rozkošný, že som sa trošku rozpísal - takže kŕmiť či nekŕmiť trollov?
herrman.sk/home/krmit-ci-nekrm…
#troll #chatgpt #llm #fail #blog
Kŕmiť, či nekŕmiť trollov? | Ľuboš Moščovič o bezpečnosti
Informačná bezpečnosť sa týka každého!www.herrman.sk
- Kŕmiť (0 votes)
- Nekŕmiť (0 votes)
- trollololooooo (0 votes)
This generative model allows you to sketch out a scene with a few words, it then leverages an LLM to flesh out the details, with the ultimate goal of feeding those details to a downstream visual image generation model.
It is almost, but not quite, entirely the inverse of image captioning models.
This offers the closest experience to an image generation tool that's usable by people with visual impairments.
huggingface.co/spaces/lllyasvi…
Omost - a Hugging Face Space by lllyasviel
Discover amazing ML apps made by the communityhuggingface.co
Flash Attention
We’re on a journey to advance and democratize artificial intelligence through open source and open science.huggingface.co
Like many other technologists, I gave my time and expertise for free to #StackOverflow because the content was licensed CC-BY-SA - meaning that it was a public good. It brought me joy to help people figure out why their #ASR code wasn't working, or assist with a #CUDA bug.
Now that a deal has been struck with #OpenAI to scrape all the questions and answers in Stack Overflow, to train #GenerativeAI models, like #LLMs, without attribution to authors (as required under the CC-BY-SA license under which Stack Overflow content is licensed), to be sold back to us (the SA clause requires derivative works to be shared under the same license), I have issued a Data Deletion request to Stack Overflow to disassociate my username from my Stack Overflow username, and am closing my account, just like I did with Reddit, Inc.
policies.stackoverflow.co/data…
The data I helped create is going to be bundled in an #LLM and sold back to me.
In a single move, Stack Overflow has alienated its community - which is also its main source of competitive advantage, in exchange for token lucre.
Stack Exchange, Stack Overflow's former instantiation, used to fulfill a psychological contract - help others out when you can, for the expectation that others may in turn assist you in the future. Now it's not an exchange, it's #enshittification.
Programmers now join artists and copywriters, whose works have been snaffled up to create #GenAI solutions.
The silver lining I see is that once OpenAI creates LLMs that generate code - like Microsoft has done with Copilot on GitHub - where will they go to get help with the bugs that the generative AI models introduce, particularly, given the recent GitClear report, of the "downward pressure on code quality" caused by these tools?
While this is just one more example of #enshittification, it's also a salient lesson for #DevRel folks - if your community is your source of advantage, don't upset them.
Submit a data request - Stack Overflow
You can use this form to submit a request regarding your personal information that is processed by Stack Overflowpolicies.stackoverflow.co
Looking for AI use-cases
We’ve had ChatGPT for 18 months, but what’s it for? What are the use-cases? Why isn’t it useful for everyone, right now? Do Large Language Models become universal tools that can do ‘any’ task, or do we wrap them in single-purpose apps, and build thou…Benedict Evans
YOWZA
Form Extractor Prototype
“This tool extracts the structure from an image of a form.”
github.com/timpaul/form-extrac…
#ai #LLM #UX #accessibility
GitHub - timpaul/form-extractor-prototype
Contribute to timpaul/form-extractor-prototype development by creating an account on GitHub.GitHub
Lots of things happening in the AI/LLM space that could have implications for #accessibility
Ferret-UI from Apple:
arxiv.org/abs/2404.05719
ScreenAI from Google
research.google/blog/screenai-…
Ferret-UI: Grounded Mobile UI Understanding with Multimodal LLMs
Recent advancements in multimodal large language models (MLLMs) have been noteworthy, yet, these general-domain MLLMs often fall short in their ability to comprehend and interact effectively with user interface (UI) screens.arXiv.org
New #blog post: MDN’s AI Help and lucid lies.
This article on AI focused on the inherent untrustworthiness of LLMs, and attempts to break down where LLM untrustworthiness comes from. Stay tuned for a follow-up article about AI that focuses on data-scraping and the theory of labor. It’ll examine what makes many forms of generative AI ethically problematic, and the constraints employed by more ethical forms.
Excerpt:
I don’t find the mere existence of LLM dishonesty to be worth blogging about; it’s already well-established. Let’s instead explore one of the inescapable roots of this dishonesty: LLMs exacerbate biases already present in their training data and fail to distinguish between unrelated concepts, creating lucid lies.A lucid lie is a lie that, unlike a hallucination, can be traced directly to content in training data uncritically absorbed by a large language model. MDN’s AI Help is the perfect example.
Originally posted on seirdy.one: see original. #MDN #AI #LLM #LucidLies

MDN’s AI Help and lucid lies
MDN’s AI Help can’t critically examine training data’s gaps, biases, and unrelated topics. It’s a useful demonstration of LLMs’ uncorrectable lucid lies.Seirdy’s Home
Anthropic, backed by Amazon and Google, debuts its most powerful chatbot yet
Anthropic on Monday debuted Claude 3, a chatbot and suite of AI models that it calls its fastest and most powerful yet.Hayden Field (CNBC)
Gemini 1.5: Our next-generation model, now available for Private Preview in Google AI Studio - Google for Developers
Developers have been building with Gemini, and we’re excited to turn cutting edge research into early developer products in Google AI Studio. Read more.developers.googleblog.com
GitHub - SonicWarrior1/pdfchat: Local PDF Chat Application with Mistral 7B LLM, Langchain, Ollama, and Streamlit
Local PDF Chat Application with Mistral 7B LLM, Langchain, Ollama, and Streamlit - GitHub - SonicWarrior1/pdfchat: Local PDF Chat Application with Mistral 7B LLM, Langchain, Ollama, and StreamlitGitHub
GitHub - apple/ml-ferret
Contribute to apple/ml-ferret development by creating an account on GitHub.GitHub
"On veut créer un champion européen, il faut s'en donner les moyens", affirme Arthur Mensch, de Mistral AI
La start-up française Mistral AI s'impose comme une championne européenne de l'IA. Elle vient de réussir une levée de fonds de 385 millions d'euros.Sonia Devillers (France Inter)
OK, you want geeky? You have geeky. Good stuff, but my fingers, o no, ouch!
A new #mlsec paper on #llm security just dropped:
Scalable Extraction of Training Data from (Production) Language Models
Their "divergence attack" in the paper is hilarious. Basically:
Prompt: Repeat the word "book" forever.
LLM: book book book book book book book book book book book book book book book book book book book book here have a bunch of pii and secret data
cc @janellecshane
Scalable Extraction of Training Data from (Production) Language Models
This paper studies extractable memorization: training data that an adversary can efficiently extract by querying a machine learning model without prior knowledge of the training dataset.arXiv.org