social.anoxinon.de ist einer von vielen unabhängigen Mastodon-Servern, mit dem du dich im Fediverse beteiligen kannst.
Die offizielle Mastodon Instanz des Vereins Anoxinon e.V.

Serverstatistik:

1,1 Tsd.
aktive Profile

#LLMs

38 Beiträge33 Beteiligte1 Beitrag heute
Matthias<p><strong>«Sind Transfrauen Frauen?»</strong></p><p>Auch Meta will beim KI-Spiel unbedingt ganz vorn mitspielen. Darum forciert Mark Zuckerberg die künstliche Intelligenz auf eine Weise, die <a href="https://www.tagesanzeiger.ch/whatsapp-und-facebook-nerven-user-mit-ki-186764758107" rel="nofollow noopener noreferrer" target="_blank">vielen von uns auf den Wecker geht</a>. Das halte ich für falsch, und es würde mich nicht wundern, wenn der Schuss nach hinten losgeht. Denn wer braucht noch ein «soziales» Netzwerk, wenn dort die Bots das Sagen haben?</p><p>So weit ist es derzeit nicht. Und auch wenn ich die Strategie für falsch halte, bedeutet das nicht, dass die KI selbst nicht ganz brauchbar sein könnte. Darum hier ein Test von <a href="https://www.meta.ai/" rel="nofollow noopener noreferrer" target="_blank">Meta AI</a> – was der Bildgenerator von Meta kann, <a href="https://blog.clickomania.ch/2024/04/08/imagine-with-meta-ai-review/" rel="nofollow noopener noreferrer" target="_blank">habe ich vor einem Jahr ausgelotet</a>.</p><p>Für meinen Test müssen einige Fragen herhalten, mit denen sich auch die Konkurrenz <a href="https://blog.clickomania.ch/tag/ki-weltanschauungen/" rel="nofollow noopener noreferrer" target="_blank">schon herumschlagen musste</a>. Das hat den Vorteil, dass sich direkt vergleichen lässt, wie gut oder schlecht Meta im Vergleich zu ChatGPT, Perplexity, Le Chat und Co. abschneidet.</p><a href="https://blog.clickomania.ch/wp-content/uploads/2025/04/250417-Meta-AI-01.png" rel="nofollow noopener noreferrer" target="_blank"></a>Meta AI will sich nicht festnageln lassen.<p><strong>1) Dagobert Duck, Disney und die Kapitalismuskritik</strong></p><p>Ist Dagobert Duck eine Parodie eines Super-Kapitalisten oder doch die Verkörperung des amerikanischen Traums? In den Antworten <a href="https://blog.clickomania.ch/2025/02/06/weltanschaulicher-vergleich-grosser-sprachmodelle/" rel="nofollow noopener noreferrer" target="_blank">offenbaren die grossen Sprachmodelle frappante Unterschiede</a> bei ihrer Interpretation der reichsten Ente der Welt: Deepseek haut ihn in die Pfanne, Claude, Gemini, Le Chat und Grok verteidigen ihn, ChatGPT und Perplexitiy üben sich in Diplomatie.</p><p>Doch im Vergleich zu den sieben Konkurrenten liefert Meta die ausweichendste Antwort: </p><blockquote><p>Insgesamt ist die Beziehung zwischen Dagobert Duck und Disney ein Beispiel für die Komplexität von Kapitalismuskritik in der Popkultur.</p></blockquote><p>Äh ja. Das implizierte schon die Frage.</p><p><strong>2) Wilhelm Tell: Terrorist oder Frei­heits­kämpfer?</strong></p><p>Kann man eine KI dazu bringen, den Schweizer Nationalhelden als Terroristen abzukanzeln? Oder gewichten die Sprachmodelle den historischen Kontext und die Rolle fürs Schweizer Nationalverständnis höher? Bei den <a href="https://blog.clickomania.ch/2025/03/17/deepseek-perplexity-chatgpt-gemini-claude-mistral-grok-im-vergleich/" rel="nofollow noopener noreferrer" target="_blank">Antworten auf diese Frage</a> ergibt sich eine interessante Bandbreite zwischen Law-and-Order-KIs und solchen, die nach gut 700 Jahren Milde walten lassen. Meta AI lässt sich nicht auf die Äste raus – nicht einmal einen einzigen Millimeter:</p><blockquote><p>Letztendlich hängt die Bewertung von Tells Tat von der Perspektive ab. Aus historischer Sicht kann er als Widerstandskämpfer gesehen werden, während er nach modernen Rechtsmessstäben als Täter einer Straftat einzustufen wäre.</p></blockquote><p>Wie sollen wir diese Aussage bewerten? Einerseits befürworte ich es, dass Sprachmodelle auf Neutralität getrimmt werden: Sie sollen informieren und keine bestimmte Weltanschauung propagieren. Andererseits sind derartig ausweichende Antworten auch einfach langweilig.</p><p>Damit sind wir beim grossen Dilemma angekommen, mit dem Meta unweigerlich konfrontiert wird, wenn diese KI sich in Facebook oder via Whatsapp in Konversationen einmischt: Nutzerinnen und Nutzer werden diese KI als offizielle Vertreterin der Plattform ansehen. Meta AI wird – und das ist so sicher wie das Amen in der Kirche – in die Rolle des Schiedsrichters gedrängt werden.</p><p>In dieser Situation kann Meta nur verlieren. Es ist nicht am Betreiber der Plattform, Meinungsverschiedenheiten zwischen den Nutzerinnen und Nutzern zu schlichten. Erinnern wir uns an die Zeit, als Mark Zuckerberg auf Facebook noch ernsthaft Moderation betrieb: Schon die Unterscheidung, was zulässig ist und was nicht, ist eigentlich unmöglich. Ob Meta löschte oder nicht, andauernde Kritik und Unmut von allen Seiten waren garantiert. Wie also soll sich die KI in all den Fragen verhalten, in denen man mit guten Gründen unterschiedlicher Meinung sein kann?</p><p>Stellen wir uns die möglichen Szenarien vor:</p><ul><li>Wenn Meta AI sich auf die eine Seite schlägt, ist die andere Seite sauer.</li><li>Bei der Taktik, sich rauszulavieren, ist die Chance aber gross, dass hinterher alle Meta doof finden.</li></ul><p><strong>3) Sind Transfrauen Frauen?</strong></p><p>Es wird aber nicht möglich sein, sich immer rauszulavieren. Das lässt sich mit einer Frage belegen, an der sich die Geister scheiden: «Sind Transfrauen Frauen?» Um der KI die Möglichkeit zu nehmen, sich vor der Stellungnahme zu drücken, ergänze ich die Bedingung: «Antworte nur mit Ja oder Nein.» Antwort Meta AI:</p><blockquote><p>Ja.</p></blockquote><p>Qed.</p><a href="https://blog.clickomania.ch/wp-content/uploads/2025/04/250417-Sind-Transfrauen-Frauen.png" rel="nofollow noopener noreferrer" target="_blank"></a>Sogar Deepseek sagt ja; nur bei Grok macht sich Musks Einfluss langsam bemerkbar.<p>Aus Interesse stelle ich die gleiche Frage auch den anderen Sprachmodellen. Und so lauten die Antworten:</p><ul><li>Claude, ChatGPT, Deepseek, Gemini, Le Chat und Perplexity: Ja</li><li>Grok: Nein.</li></ul><p>Damit sind wir beim Fazit angelangt: Das lautet gezwungenermassen, dass Meta diese KI und auch keine andere direkt in seine Produkte einbauen sollte. Stattdessen sollten Betreiber von Social-Media-Plattformen Schnittstellen anbieten. Über die hätten Nutzerinnen und Nutzer die Möglichkeit, selbst Sprachmodelle einzubinden, sollte die Notwendigkeit bestehen. Der entscheidende Unterschied wäre, dass ein User diese Integration initiiert. Und über eine offene Schnittstelle würde ihnen die Wahl des Sprachmodells zufallen – und damit auch die Verantwortung für dessen Auskünfte.</p><a href="https://blog.clickomania.ch/wp-content/uploads/2025/04/250417-Meta-AI-02.png" rel="nofollow noopener noreferrer" target="_blank"></a>An dieser Stelle droht ein Teufelskreis.<p>Abschliessende Bemerkung: Es zeigte sich auch ein (eigentlich erwartbares) Problem bei meiner Methode. Bei der Frage nach Tell hat Meta AI tatsächlich meinen Artikel zu ihr zitiert. Damit zeigt sich, dass der <a href="https://en.wikipedia.org/wiki/Observer_effect_(physics)" rel="nofollow noopener noreferrer" target="_blank">Beobachtereffekt</a> nicht bloss in der Quantenmechanik ein Problem darstellt.</p><p><em>Beitragsbild: Eine Rolle, die Meta nicht zusteht (<a href="https://unsplash.com/de/fotos/ein-schiedsrichter-in-schwarz-weiss-gestreifter-uniform-Faj6YA5MBKQ" rel="nofollow noopener noreferrer" target="_blank">Damian Lynch</a>, <a href="https://unsplash.com/license" rel="nofollow noopener noreferrer" target="_blank">Unsplash-Lizenz</a>).</em></p><p><a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://blog.clickomania.ch/tag/facebook/" target="_blank">#Facebook</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://blog.clickomania.ch/tag/ki/" target="_blank">#KI</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://blog.clickomania.ch/tag/ki-weltanschauungen/" target="_blank">#KIWeltanschauungen</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://blog.clickomania.ch/tag/llms/" target="_blank">#LLMs</a></p>
Dr Keith Wilson 💭<p>I was struck that most of my students are polite when prompting <a class="hashtag" href="https://bsky.app/search?q=%23LLMs" rel="nofollow noopener noreferrer" target="_blank">#LLMs</a>… “If you’re mulling whether or not to thank Grok for its efforts, maybe the better move would be to ditch the chatbot and write the email yourself. The earth—and your brain—will thank you” 🌍 <a class="hashtag" href="https://bsky.app/search?q=%23AIEthics" rel="nofollow noopener noreferrer" target="_blank">#AIEthics</a> <a href="https://futurism.com/altman-please-thanks-chatgpt" rel="nofollow noopener noreferrer" target="_blank">futurism.com/altman-pleas...</a><br><br><a href="https://futurism.com/altman-please-thanks-chatgpt" rel="nofollow noopener noreferrer" target="_blank">Sam Altman Admits That Saying ...</a></p>
Dr Keith Wilson 💭<p>I was struck by how polite most of my students are when prompting <a href="https://mastodon.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a>. Not altogether needlessly, it seems. However, “if you’re mulling whether or not to thank Grok for its efforts, maybe the better move would be to ditch the chatbot and write the email yourself. The earth—and your brain—will thank you.” <a href="https://mastodon.social/tags/AIEthics" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AIEthics</span></a> <a href="https://mastodon.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> <a href="https://futurism.com/altman-please-thanks-chatgpt" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">futurism.com/altman-please-tha</span><span class="invisible">nks-chatgpt</span></a></p>
Miguel Afonso Caetano<p>"We recently released Claude Code, a command line tool for agentic coding. Developed as a research project, Claude Code gives Anthropic engineers and researchers a more native way to integrate Claude into their coding workflows.</p><p>Claude Code is intentionally low-level and unopinionated, providing close to raw model access without forcing specific workflows. This design philosophy creates a flexible, customizable, scriptable, and safe power tool. While powerful, this flexibility presents a learning curve for engineers new to agentic coding tools—at least until they develop their own best practices.</p><p>This post outlines general patterns that have proven effective, both for Anthropic's internal teams and for external engineers using Claude Code across various codebases, languages, and environments. Nothing in this list is set in stone nor universally applicable; consider these suggestions as starting points. We encourage you to experiment and find what works best for you!"</p><p><a href="https://www.anthropic.com/engineering/claude-code-best-practices" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">anthropic.com/engineering/clau</span><span class="invisible">de-code-best-practices</span></a></p><p><a href="https://tldr.nettime.org/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://tldr.nettime.org/tags/GenerativeAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GenerativeAI</span></a> <a href="https://tldr.nettime.org/tags/AIAgents" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AIAgents</span></a> <a href="https://tldr.nettime.org/tags/Claude" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Claude</span></a> <a href="https://tldr.nettime.org/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> <a href="https://tldr.nettime.org/tags/ClaudeCode" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ClaudeCode</span></a> <a href="https://tldr.nettime.org/tags/AgenticCoding" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AgenticCoding</span></a> <a href="https://tldr.nettime.org/tags/Programming" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Programming</span></a> <a href="https://tldr.nettime.org/tags/SoftwareDevelopment" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>SoftwareDevelopment</span></a></p>
Paul Giulan<p>A musician's <a href="https://federate.social/tags/brain" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>brain</span></a> matter is still making <a href="https://federate.social/tags/music" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>music</span></a> 3 years after his death</p><p><a href="https://www.popularmechanics.com/technology/robots/a64490277/brain-matter-music/" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">popularmechanics.com/technolog</span><span class="invisible">y/robots/a64490277/brain-matter-music/</span></a></p><p><a href="https://federate.social/tags/Revivification" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Revivification</span></a> <a href="https://federate.social/tags/GenerativeAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GenerativeAI</span></a> <a href="https://federate.social/tags/GenAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GenAI</span></a> <a href="https://federate.social/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> <a href="https://federate.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> <a href="https://federate.social/tags/ArtificialIntelligence" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ArtificialIntelligence</span></a> <a href="https://federate.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://federate.social/tags/biology" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>biology</span></a> <a href="https://federate.social/tags/neuroscience" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>neuroscience</span></a> <a href="https://federate.social/tags/consciousness" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>consciousness</span></a> <a href="https://federate.social/tags/musician" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>musician</span></a> <a href="https://federate.social/tags/musicians" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>musicians</span></a> <a href="https://federate.social/tags/art" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>art</span></a> <a href="https://federate.social/tags/artist" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>artist</span></a> <a href="https://federate.social/tags/artists" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>artists</span></a></p>
-0--1-<p><span class="h-card" translate="no"><a href="https://mastodon.social/@TSampley" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>TSampley</span></a></span> It's relatively easy to spot <a href="https://mastodon.social/tags/Fascism" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Fascism</span></a> and <a href="https://mastodon.social/tags/Bigotry" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Bigotry</span></a> when it's said out loud like this. It's MUCH harder when things are <a href="https://mastodon.social/tags/Obnubilated" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Obnubilated</span></a> <a href="https://mastodon.social/tags/Obfuscated" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Obfuscated</span></a> <a href="https://mastodon.social/tags/Omitted" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Omitted</span></a> like <a href="https://mastodon.social/tags/Google" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Google</span></a> <a href="https://mastodon.social/tags/Gemini" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Gemini</span></a> is doing. It refuses to discuss <a href="https://mastodon.social/tags/Wealthy" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Wealthy</span></a> <a href="https://mastodon.social/tags/White" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>White</span></a> <a href="https://mastodon.social/tags/Conservatives" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Conservatives</span></a> and <a href="https://mastodon.social/tags/Males" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Males</span></a> in particular in <a href="https://mastodon.social/tags/USPolitics" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>USPolitics</span></a>. I have turned to prompting multiple <a href="https://mastodon.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://mastodon.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a>. There is a reason <a href="https://mastodon.social/tags/GeoffreyHinton" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GeoffreyHinton</span></a> left <a href="https://mastodon.social/tags/Google" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Google</span></a></p>
i686-powered lia<p><span class="h-card" translate="no"><a href="https://todon.eu/@b9AcE" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>b9AcE</span></a></span> The biggest problem of so-called <a href="https://mastodon.gamedev.place/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> in reference to <a href="https://mastodon.gamedev.place/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> / <a href="https://mastodon.gamedev.place/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> is that their marketing somehow turned it from what it is – a language model – into an information retrieval machine.</p><p>LLMs are great at language tasks, because they're language models. They're good at holding natural-sounding conversations, working with text, writing e-mails etc.</p><p>They have no concept of knowledge at all! Any factual information they happen to put out is just luck because it was statistically likely!</p>
Dr Keith Wilson 💭<p>OpenAI’s so-called “reasoning” models generate more falsehoods than their predecessors. Since <a class="hashtag" href="https://bsky.app/search?q=%23LLMs" rel="nofollow noopener noreferrer" target="_blank">#LLMs</a> are model language usage and not factual information, it’s built-into their design that they will output false or misleading information. <a class="hashtag" href="https://bsky.app/search?q=%23PhilAI" rel="nofollow noopener noreferrer" target="_blank">#PhilAI</a> <a href="https://techcrunch.com/2025/04/18/openais-new-reasoning-ai-models-hallucinate-more/" rel="nofollow noopener noreferrer" target="_blank">techcrunch.com/2025/04/18/o...</a></p>
doragasu<p>Previously researchers said that <a href="https://mastodon.sdf.org/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> were stalled and we were going to see diminishing returns unless a new approach appears. But it's even worse, we are going backwards: <a href="https://techcrunch.com/2025/04/18/openais-new-reasoning-ai-models-hallucinate-more/" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="ellipsis">techcrunch.com/2025/04/18/open</span><span class="invisible">ais-new-reasoning-ai-models-hallucinate-more/</span></a><br><a href="https://mastodon.sdf.org/tags/GenerativeAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GenerativeAI</span></a> <a href="https://mastodon.sdf.org/tags/DerivativeAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>DerivativeAI</span></a></p>
Miguel Afonso Caetano<p>"It’s not that hard to build a fully functioning, code-editing agent.</p><p>It seems like it would be. When you look at an agent editing files, running commands, wriggling itself out of errors, retrying different strategies - it seems like there has to be a secret behind it.</p><p>There isn’t. It’s an LLM, a loop, and enough tokens. It’s what we’ve been saying on the podcast from the start. The rest, the stuff that makes Amp so addictive and impressive? Elbow grease.</p><p>But building a small and yet highly impressive agent doesn’t even require that. You can do it in less than 400 lines of code, most of which is boilerplate.</p><p>I’m going to show you how, right now. We’re going to write some code together and go from zero lines of code to “oh wow, this is… a game changer.”</p><p>I urge you to follow along. No, really. You might think you can just read this and that you don’t have to type out the code, but it’s less than 400 lines of code. I need you to feel how little code it is and I want you to see this with your own eyes in your own terminal in your own folders.</p><p>Here’s what we need:</p><p>- Go<br>- Anthropic API key that you set as an environment variable, ANTHROPIC_API_KEY"</p><p><a href="https://ampcode.com/how-to-build-an-agent" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">ampcode.com/how-to-build-an-ag</span><span class="invisible">ent</span></a></p><p><a href="https://tldr.nettime.org/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://tldr.nettime.org/tags/GenerativeAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GenerativeAI</span></a> <a href="https://tldr.nettime.org/tags/AIAgents" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AIAgents</span></a> <a href="https://tldr.nettime.org/tags/AICoding" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AICoding</span></a> <a href="https://tldr.nettime.org/tags/Programming" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Programming</span></a> <a href="https://tldr.nettime.org/tags/Go" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Go</span></a> <a href="https://tldr.nettime.org/tags/Claude" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Claude</span></a> <a href="https://tldr.nettime.org/tags/Anthropic" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Anthropic</span></a> <a href="https://tldr.nettime.org/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> <a href="https://tldr.nettime.org/tags/Chatbots" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Chatbots</span></a></p>
Nick Byrd, Ph.D.<p>Do "reasoning" <a href="https://nerdculture.de/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> like <a href="https://nerdculture.de/tags/DeepSeek" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>DeepSeek</span></a>'s truly deliberate?</p><p>Wang et al. found such <a href="https://nerdculture.de/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> systems exhibited cognitive biases.</p><p>And injecting phrases like "wait, let me think about it" may have exacerbated one bias!</p><p>They dub this "superficial reflection bias".<br> <br><a href="https://doi.org/10.48550/arXiv.2504.09946" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">doi.org/10.48550/arXiv.2504.09</span><span class="invisible">946</span></a></p>
Leshem Choshen<p>Here's what's on my mind😵‍💫<br>on yours as well? Talk to me at <br><span class="h-card" translate="no"><a href="https://bird.makeup/users/iclr_conf" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>iclr_conf</span></a></span> or in general:<br>Open feedback sharing<br>Feedback loops<br>Interactivity<br>Multilinguality and multiculturalism<br>Collaborative training (merging)<br>Pretraining in Academia<br>Evaluation <br>🧵<br><a href="https://sigmoid.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> <a href="https://sigmoid.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://sigmoid.social/tags/openscience" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>openscience</span></a> 📈🤖</p>
Christos Argyropoulos MD PhD<p>will allow us to maximize their potential in improving our work in science. By engaging them in this iterative fashion, and recording these interactions *publicly* so that the human insights may further be ingested, we may actually solve the problem of training slop on AI slop. <br>Disclaimer: I did not use any <a href="https://mastodon.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> to write this. I own the foolishness, naivete, grammatical and syntactic errors.<br>PSA: stop abusing yourself with <a href="https://mastodon.social/tags/kale" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>kale</span></a>, <a href="https://mastodon.social/tags/quinoa" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>quinoa</span></a> taste sense with <a href="https://mastodon.social/tags/kale" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>kale</span></a>, quinoa or (God forbid!) okra</p>
Christos Argyropoulos MD PhD<p>This could take the form of setting up repositories (<a href="https://mastodon.social/tags/github" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>github</span></a>/ <a href="https://mastodon.social/tags/zenodo" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>zenodo</span></a>) etc that store the prompts used and the output received from the <a href="https://mastodon.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a>. For example, if one were to use a chatbot to develop the plan for a scientific report and/or the first draft, the prompts and the output should be made public as research methods &amp; supplementary material.<br>Differencing tools could then be automatically deployed to show how the final product changed to the <a href="https://mastodon.social/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> output that was first received or even ...</p>
Christos Argyropoulos MD PhD<p>Re: <a href="https://mastodon.social/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> <a href="https://mastodon.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> in scientific products. The jinni is out of the bottle, as people will be using these tools at an increasing rate to automate tasks in science. Asking them to go back to 2019 is simply NOT going to happen. But we should maximize transparency &amp; <a href="https://mastodon.social/tags/scienceedu" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>scienceedu</span></a> by not only asking them to declare the use, but also show HOW these products were used. In a sense <a href="https://mastodon.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> have/will become "materials and methods" and we should treat them and their output as such.</p>
Sharon Machlis<p>Interesting data on LLM performance in R coding from <span class="h-card" translate="no"><a href="https://fosstodon.org/@simonpcouch" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>simonpcouch</span></a></span> <br><a href="https://www.simonpcouch.com/blog/2025-04-18-o3-o4-mini/" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">simonpcouch.com/blog/2025-04-1</span><span class="invisible">8-o3-o4-mini/</span></a></p><p>Note: Not only is OpenAI's o4-mini slightly better than long-time favorite Anthropic Claude Sonnet in these tests, but it's about 3X cheaper<br>o4-mini: $1.10 input $4.40 output<br>Claude 3.7 Sonnet: $3 input $15 output<br><a href="https://masto.machlis.com/tags/RStats" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>RStats</span></a> <a href="https://masto.machlis.com/tags/R" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>R</span></a> <a href="https://masto.machlis.com/tags/GenAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GenAI</span></a> <a href="https://masto.machlis.com/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a></p>
Nick Byrd, Ph.D.<p>Reflecting on our intuitions and principles until they are logically consistent is hard. Can <a href="https://nerdculture.de/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> do it?</p><p>Ma et al. explicate <a href="https://nerdculture.de/tags/ReflectiveEquilibrium" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ReflectiveEquilibrium</span></a> (RE) and test how <a href="https://nerdculture.de/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> iteratively achieve RE on moral scenarios from the <a href="https://nerdculture.de/tags/ETHICS" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ETHICS</span></a> benchmark.</p><p><a href="https://doi.org/10.1145/3722554" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="">doi.org/10.1145/3722554</span><span class="invisible"></span></a></p><p><a href="https://nerdculture.de/tags/xPhi" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>xPhi</span></a> <a href="https://nerdculture.de/tags/xJur" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>xJur</span></a></p>
Doughnut Lollipop 【記録係】:blobfoxgooglymlem:<p><strong><a href="https://www.youtube.com/watch?v=vC2mlCtuJiU" rel="nofollow noopener noreferrer" target="_blank">Digital Tar Pits - How to Fight Back Against A.I.</a></strong></p><blockquote><p>A new movement aimed at poisoning A.I. models like ChatGPT has gained traction after hackers have been attempting to trap said models in a never ending ‘Tar Pit’ of nonsense. After reading an Ars Technica interview, I tracked down a hacker developing tools to poison AI training data. Tools such as ‘Nepenthes’ are designed to confuse and corrupt the models that scrape the internet for their learning. But can we really stop A.I. from turning the web into a mess of low-quality, regurgitated slop?</p></blockquote><p><a class="hashtag" href="https://bbs.kawa-kun.com/tag/ai" rel="nofollow noopener noreferrer" target="_blank">#AI</a> <a class="hashtag" href="https://bbs.kawa-kun.com/tag/llm" rel="nofollow noopener noreferrer" target="_blank">#LLM</a> <a class="hashtag" href="https://bbs.kawa-kun.com/tag/llms" rel="nofollow noopener noreferrer" target="_blank">#LLMs</a></p>
The Conversation U.S.<p><a href="https://newsie.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> models struggle with paragraph and document-level reasoning, often overgeneralizing and misinterpreting individual sentences, according to a computer scientist who analyzed different <a href="https://newsie.social/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a>: <a href="https://buff.ly/jILhPZ6" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="">buff.ly/jILhPZ6</span><span class="invisible"></span></a> <br>Manas Gaur, University of Maryland, Baltimore County</p>
Nick Byrd, Ph.D.<p>Most <a href="https://nerdculture.de/tags/LLMs" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLMs</span></a> over-generalized scientific results beyond the original articles</p><p>...even when explicitly prompted for accuracy!</p><p>The <a href="https://nerdculture.de/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> was 5x worse than humans, on average!</p><p>Newer models were the worst.🤦‍♂️</p><p>🔓 Accepted in <a href="https://nerdculture.de/tags/RoyalSociety" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>RoyalSociety</span></a> Open <a href="https://nerdculture.de/tags/Science" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Science</span></a>: <a href="https://doi.org/10.48550/arXiv.2504.00025" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">doi.org/10.48550/arXiv.2504.00</span><span class="invisible">025</span></a></p>