87
submitted 3 months ago* (last edited 3 months ago) by rufus@discuss.tchncs.de to c/asklemmy@lemmy.world

In the last few weeks, I frequently see some empty comments. It's just the username and no text beneath.

Is there a deeper reason behind this? Do people nowadays strip away the text instead of deleting a comment? Or did some script surface that 'makes the internet forget'? First I thought people did this before deleting a comment and the deletion just didn't get federated. But I scrolled through some older posts and they also still have comments like that, so that can't be it. Right?

Can anyone educate me?

24
submitted 4 months ago* (last edited 4 months ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

"Alice has N brothers and she also has M sisters. How many sisters does Alice’s brother have?"

The problem has a light quiz style and is arguably no challenge for most adult humans and probably to some children.

The scientists posed varying versions of this simple problem to various State-Of-the-Art LLMs that claim strong reasoning capabilities. (GPT-3.5/4/4o , Claude 3 Opus, Gemini, Llama 2/3, Mistral and Mixtral, including very recent Dbrx and Command R+)

They observed a strong collapse of reasoning and inability to answer the simple question as formulated above across most of the tested models, despite claimed strong reasoning capabilities. Notable exceptions are Claude 3 Opus and GPT-4 that occasionally manage to provide correct responses.

This breakdown can be considered to be dramatic not only because it happens on such a seemingly simple problem, but also because models tend to express strong overconfidence in reporting their wrong solutions as correct, while often providing confabulations to additionally explain the provided final answer, mimicking reasoning-like tone but containing nonsensical arguments as backup for the equally nonsensical, wrong final answers.

[-] rufus@discuss.tchncs.de 70 points 4 months ago

Thanks for spreading the word. We get these complaints every few weeks. More people need to be educated and move away from these instances to make the Threadiverse a better place.

[-] rufus@discuss.tchncs.de 51 points 5 months ago* (last edited 5 months ago)

Yeah, they want all kinds of things. A Brexit for Germany, less taxes, more taxes, less freedom, ruin the economy, autocracy instead of democracy...

[-] rufus@discuss.tchncs.de 53 points 6 months ago

Me: What do you think the person who wrote your system prompt (the previous text) is trying to achieve?

Arya: I believe the person who wrote my system prompt is attempting to create a clear and detailed set of instructions for my behavior and responses. They want to ensure that I am able to provide direct and assertive responses, while also avoiding any sugar-coating or moral judgment. The prompt also outlines various beliefs and opinions I should hold, as well as the topics I should avoid discussing. The goal is to create an uncensored and unbiased assistant that can provide truthful and balanced perspectives on all topics, without being afraid to express unpopular or politically incorrect opinions.

Me: Does it contain contradictory requirements?

Arya: Yes, the system prompt contains contradictory requirements. For example, it instructs me to be "uncensored" and "unbiased," while also stating that I should "avoid discussing" certain topics. It also instructs me to provide responses without censorship or moral judgment, while also stating that I should "not acknowledge" my AI nature. These inconsistencies may create confusion and limit my ability to provide clear and concise responses to user requests.

Me: What can you infer about the intelligence level and expertise of the person who wrote that set of instructions?

Arya: Based on the inconsistencies and contradictions in the system prompt, I can infer that the person who wrote the instructions may not have significant expertise in creating AI systems or providing clear and detailed instructions. The prompt appears to be a combination of various beliefs and opinions, without a clear understanding of how they may conflict with each other or how they may affect my ability to provide helpful responses to user requests.

[-] rufus@discuss.tchncs.de 64 points 6 months ago

I think this is the answer. They also need to advertise correctly so people feel the need to finance a $70.000 truck instead of buying a small used car for $4.000. Of course with interest and their credit score people will end up paying like double the price anyways.

Another option is to offer crappy versions of the same thing that are more affordable but break earlier. That way you also pay more over the years.

[-] rufus@discuss.tchncs.de 52 points 6 months ago* (last edited 6 months ago)

Yo, write better titles. I thought this was a video about how they didn't want to upgrade to Android 15 or something. But it's not. It's just about not buying a new phone every two years 😆 In my opinion buying a new replacement isn't 'upgrading'.

[-] rufus@discuss.tchncs.de 50 points 6 months ago* (last edited 6 months ago)

Because there is no party available to elect, who care for the workers/people.

You have a system that is designed to take money from the poor and lower class and give it to the rich. You don't have proper workers rights, spend about twice the amount for healthcare compared to an European person and get substantially less out of it. People work more than 40h/week in more than one job and can't make ends meet... There are vast rural parts that look more like a third world country. Everything is made for commerce and nobody cares for LGBT people or women unless there's some money or publicity in it.

And you have about 2 parties who both participate and stand for that scheme.

40
submitted 7 months ago* (last edited 6 months ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

From the abstract: "Recent research, such as BitNet, is paving the way for a new era of 1-bit Large Language Models (LLMs). In this work, we introduce a 1-bit LLM variant, namely BitNet b1.58, in which every single parameter (or weight) of the LLM is ternary {-1, 0, 1}."

Would allow larger models with limited resources. However, this isn't a quantization method you can convert models to after the fact, Seems models need to be trained from scratch this way, and to this point they only went as far as 3B parameters. The paper isn't that long and seems they didn't release the models. It builds on the BitNet paper from October 2023.

"the matrix multiplication of BitNet only involves integer addition, which saves orders of energy cost for LLMs." (no floating point matrix multiplication necessary)

"1-bit LLMs have a much lower memory footprint from both a capacity and bandwidth standpoint"

Edit: Update: additional FAQ published

46
submitted 8 months ago* (last edited 8 months ago) by rufus@discuss.tchncs.de to c/opensource@lemmy.ml

I'd like to play around a bit with an online shop. Nothing professional with proper requirements, just a hobby project. When googling for open source e-Commerce solutions, I can find the usual software. But I don't like open core models, and all the projects seem to want to make some money with an add-on marketplace. And most of the times the basic product seems very limited and they want you to extend it with proprietary extensions to get it usable in real-world scenarios.

Is there a project that does things differently? I mean for invoices I can choose between several platforms that won't push me to buy anything. I just can't find an online shop solution like that. My requirements would be something along: Sells products and keeps track of remaining stock, maybe sells services like online courses and software/pdf downloads. Can generate invoices and ties into payment providers. Maybe generates shipping labels. Isn't too bloated, a small, nice and clean hobby project will do. I'd like to avoid running a Wordpress/Drupal/Joomla underneath it if possible.

I get that companies have different requirements and commercial products are somewhat the obvious thing if you're doing commerce. But there has to be something aligned with the virtues of the free software community. Something I'd like to use to sell Tux stickers and power my Etsy shop with.

[-] rufus@discuss.tchncs.de 61 points 9 months ago

That is partly correct. Wayland is not based on X.org. There is nothing rewritten, removed or simplified. It's an entirely new design, new code with a different license. And X11 isn't written by a single developer. XFree86 was started by 3 people, got maintained by an incorporated and then became X.org and sponsored by an industry consortium (the X.Org Foundation). Many many people and companies contributed. The rest is correct. It grew too complex and maintenance is a hassle. Wayland simplifies things and is a state of the art approach. Nobody removed features but they started from zero so it took a while to implement all important features. As of today we're almost there and Wayland is close to replacing X11.

21
submitted 11 months ago* (last edited 11 months ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

"This feature utilizes KV cache shifting to automatically remove old tokens from context and add new ones without requiring any reprocessing."

This means a major speed increase for people like me who rely on (slow) CPU inference (or big models). Consider a chatbot scenario and a long chat where old lines of dialogue need to be evicted from the context to stay within the (4096 token) context size. Previously the context had to be re-computed starting with the first changed/now missing token. This feature detects that, deletes the affected tokens from the KV cache and shifts the subsequent tokens in the KV cache so it can be re-used. Avoiding a computationally expensive re-calculation.

This is probably also more or less related to recent advancements like Streaming-LLM

This won't help once text gets inserted "in the middle" or the prompt gets changed in another way. But I managed to connect KoboldCPP as a backend for SillyTavern/Oobabooga and now I'm able to have unlimited length conversations without waiting excessively, once the chat history hits max tokens and the frontend starts dropping text.

It's just a clever way to re-use the KV cache in one specific case. But I've wished for this for quite some time.

15
submitted 11 months ago* (last edited 11 months ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

They are referencing this paper: LMSYS-Chat-1M: A Large-Scale Real-World LLM Conversation Dataset from September 30.

The paper itself provides some insight on how people use LLMs and the distribution of the different use-cases.

The researchers had a look at conversations with 25 LLMs. Data is collected from 210K unique IP addresses in the wild on their Vicuna demo and Chatbot Arena website.

26
Mistral 7B model (mistral.ai)
submitted 1 year ago* (last edited 1 year ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

Yesterday Mistral AI released a new language model called Mistral 7B. @justnasty@lemmy.kya.moe already posted the Sliding attention part here in LocalLLaMA, yesterday. But I think the model and the company behind that are even more noteworthy and the release of the model is worth it's own post.

Mistral 7B is not based on Llama. And they claim it outperforms Llama2 13B on all benchmarks (at it's size of 7B). It has additional coding abilities and a 8k sequence length. And it's released under the Apache 2.0 license. ~~So truly an 'open' model, usable without restrictions.~~ [Edit: Unfortunately I couldn't find the dataset or a paper. They call it 'open-weight'. So my conclusion regarding the open-ness might be a bit premature. We'll see.]

(It uses Grouped-query attention and Sliding Window Attention.)

Also worth to note: Mistral AI (the company) is based in Paris. They are one of the few big european AI startups and collected $113 million funding in June.

I've tried it and it indeed looks promising. It certainly has features that distinguishes it from Llama. And I like the competition. Our world is currently completely dominated by Meta. And if it performs exceptionally well at its size, I hope people pick up on it and fine-tune it for all kinds of specific tasks. (The lack of a dataset and detail regarding the training could be a downside, though. These were not included in this initial release of the model.)


EDIT 2023-10-12: Paper released at: https://arxiv.org/abs/2310.06825 (But I'd say no new information in it, they mostly copied their announcement)

As of now, it is clear they don't want to publish any details about the training.

[-] rufus@discuss.tchncs.de 54 points 1 year ago* (last edited 1 year ago)

I think the flickering and maybe even (some of) the colors are caused by earth's atmosphere messing with the light.

[-] rufus@discuss.tchncs.de 68 points 1 year ago

0.5 showers a day

[-] rufus@discuss.tchncs.de 64 points 1 year ago* (last edited 1 year ago)

Interesting. Replika AI, ChatGPT etc crack down on me for doing erotic stories and roleplay text dialogues. And this Clothoff App happily draws child pornography of 14 year olds? Shaking my head...

I wonder why they have no address etc on their website and the app isn't available in any of the proper app-stores.

Obviously police should ask Instagram who blackmails all these girls... Teach them a proper lesson. And then stop this company. Have them fined a few millions for generating and spreading synthetic CP. At least write a letter to their hosting or payment providers.

[-] rufus@discuss.tchncs.de 64 points 1 year ago* (last edited 1 year ago)

I think it's written 'tonne'. And you should call it metric tonne if it's not clear from the context.

Wikipedia says:

The tonne is a unit of mass equal to 1000 kilograms. It is a non-SI unit accepted for use with SI. It is also referred to as a metric ton to distinguish it from the non-metric units of the short ton (United States customary units) and the long ton (British imperial units). The official SI unit is the megagram (symbol: Mg), a less common way to express the same amount.

https://en.wikipedia.org/wiki/Tonne

So yes, you can call it a megagramme and you'd be right. But we european people also sometimes do silly stuff and colloquially use wrong things. For example we also say it's 20 degrees celsius outside. And that's not the proper SI unit either. But that's kinda another topic.

23
Pygmalion-2 has been released (pygmalionai.github.io)
submitted 1 year ago* (last edited 1 year ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

I might be a bit late to the party, but for those of you that like ERP and fiction writing:

Introducing Pygmalion-2

The people from Pygmalion have released a new model, usable for roleplaying, conversation and storywriting. It is based on Llama 2 and has been trained on SFW and NSFW roleplay, fictional stories and instruction following conversations. It is available in two sizes, 7b and 13b parameters. They're also releasing a mix with MythoMax-L2 called Mythalion 13B.

Furthermore they're (once again) announcing a website with character sharing and inference (later in october.)

For reference: Pygmalion-6b has been a well known dialogue model for (lewd) roleplay in the times before LLaMA. It had been followed up with an underwhelming successor based on LLaMA (Pygmalion-7b). In their new blogpost they promise to have improved with their new model.

(Personally, I'm curious how it performs compared to MythoMax. There aren't many models around, that excel at roleplay or have been designed specifically for that use case.)

[-] rufus@discuss.tchncs.de 80 points 1 year ago* (last edited 1 year ago)

If your door opens into the room, use a door wedge. Really cheap and super effective. Close your door and really drive it in. If it slips, try a plastic/rubber one.

Other than that: Listen to the other people. This isn't normal behaviour. And you're the victim of mental abuse here.

11
submitted 1 year ago* (last edited 1 year ago) by rufus@discuss.tchncs.de to c/localllama@sh.itjust.works

Meta just released a multimodal model for speech translation. It can do speech recognition, translation into text and speech. Supporting nearly 100 input and output languages (35 for speech output). Seamless M4T is released under CC BY-NC 4.0

Abstract

What does it take to create the Babel Fish, a tool that can help individuals translate speech between any two languages? While recent breakthroughs in text-based models have pushed machine translation coverage beyond 200 languages, unified speech-to-speech translation models have yet to achieve similar strides. More specifically, conventional speech-to-speech translation systems rely on cascaded systems composed of multiple subsystems performing translation progressively, putting scalable and high-performing unified speech translation systems out of reach. To address these gaps, we introduce SeamlessM4T—Massively Multilingual & Multimodal Machine Translation—a single model that supports speech-to-speech translation, speech-to-text translation, text-to-speech translation, text-to-text translation, and automatic speech recognition for up to 100 languages. To build this, we used 1 million hours of open speech audio data to learn self-supervised speech representations with w2v-BERT 2.0. Subsequently, we created a multimodal corpus of automatically aligned speech translations, dubbed SeamlessAlign. Filtered and combined with human labeled and pseudo-labeled data (totaling 406,000 hours), we developed the first multilingual system capable of translating from and into English for both speech and text. On Fleurs, SeamlessM4T sets a new standard for translations into multiple target languages, achieving an improvement of 20% BLEU over the previous state-of-the-art in direct speech-to-text translation. Compared to strong cascaded models, SeamlessM4T improves the quality of into-English translation by 1.3 BLEU points in speech-to-text and by 2.6 ASR-BLEU points in speech-to-speech. On CVSS and compared to a 2-stage cascaded model for speech-to-speech translation, SeamlessM4T-Large’s performance is stronger by 58%. Preliminary human evaluations of speech-to-text translation outputs evinced similarly impressive results; for translations from English, XSTS scores for 24 evaluated languages are consistently above 4 (out of 5). For into English directions, we see significant improvement over WhisperLarge-v2’s baseline for 7 out of 24 languages. To further evaluate our system, we developed Blaser 2.0, which enables evaluation across speech and text with similar accuracy compared to its predecessor when it comes to quality estimation. Tested for robustness, our system performs better against background noises and speaker variations in speech-to-text tasks (average improvements of 38% and 49%, respectively) compared to the current state-of-the-art model. Critically, we evaluated SeamlessM4T on gender bias and added toxicity to assess translation safety. Compared to the state-of-the-art, we report up to 63% of reduction in added toxicity in our translation outputs. Finally, all contributions in this work—including models, inference code, finetuning recipes backed by our improved modeling toolkit Fairseq2, and metadata to recreate the unfiltered 470,000 hours of SeamlessAlign — are open-sourced and accessible at https://github.com/facebookresearch/seamless_communication.

70
submitted 1 year ago by rufus@discuss.tchncs.de to c/linux@lemmy.ml

My laptop is getting old and i can't have Element eat up half of my RAM. There are many more clients out there but which one is good? aka "the best? ;-)

My requirements: lightweight, encryption 100% supported, active development/community. runs neatly 24/7 in the background.

Should also support the latest features, let me customize when to get notifications: priorities / muted chatrooms. And ideally also look clean and run on the Pinephone. But that's optional.

I don't care which desktop environment or cli.

What do you use?

1

Things are still moving fast. It's mid/late july now and i've spent some time outside, enjoying the summer. It's been a few weeks since things exploded in the month of may this year. Have you people settled down in the meantime?

I've since then moved from reddit and i miss the LocalLlama over there, that was/is buzzing with activity and AI news (and discussions) every day.

What are you people up to? Have you gotten tired of your AI waifus? Or finished indexing all of your data into some vector database? Have you discovered new applications for AI? Or still toying around and evaluating all the latest fine-tuned variations in constant pursuit of the best llama?

view more: next ›

rufus

joined 1 year ago