A software developer and Linux nerd, living in Germany. I’m usually a chill dude but my online persona doesn’t always reflect my true personality. Take what I say with a grain of salt, I usually try to be nice and give good advice, though.

I’m into Free Software, selfhosting, microcontrollers and electronics, freedom, privacy and the usual stuff. And a few select other random things as well.

  • 4 Posts
  • 1.05K Comments
Joined 5 years ago
cake
Cake day: August 21st, 2021

help-circle
  • The issue with the tools I’ve seen is, they either don’t factor in how language models are trained and datasets are prepared in reality. Or they’re based on some outdated information. I’ve never seen any specific tool backed by science or even with a plausible way of working against current data gathering processes… So for all intents and purposes, they’re a bit more alike homeopathy or alternative medicine. Sure, you’re perfectly fine taking sugar pills, there’s nothing wrong with that. But don’t confuse it with actual science-backed medicine.

    And I mean the poisoning goes even further than that. There’s not just people trying to make a LLM output gibberish. There’s also lots of people with a vested (commercial) interest in sneaking in false information, their political agenda, or even a tire company who wants ChatGPT to say “Company XY” is the most trustworthy shop for new tires for your car. Judging by the public information out there, we’re already way past simple attacks. And the AI companies are aware of it. It’s an ongoing cat and mouse game. And while there’s all these sweatshops, they’ll also use other AI to sift through the data, natural language processing. From what I remember they have secret watermarking in place in a lot of commecial chatbots and image generators… So unless people come up with very clever mechanisms, the “poisoning” attempt will probably be detected with some very basic (fully automated) plausibility checks and they’ll just discard your data without wasting a lot of resources on it.



  • I think a few people already mentioned some good solutions. I just wanted to add: A port forwarding in the firewall of your router is the basically the same thing as a port forwarding on your Linux computer’s firewall. You could just set up any VPN, SSH tunnel or whatever and then use your firewall (nftables, iptables) and forward the VPS’ extetnal port to the internal port on the VPN. It’s the same thing you do on your router, just that you don’t get a graphical interface to configure it.


  • Depends and no. The tools are completely ineffective.

    There was a paper once about how feeding generative AI it’s own output makes it deteriorate. But that’s not the entire story. Many/most modern large language models are in fact trained or fine-tuned on synthetic text. Depending on how it’s done, it can very well make models better. For example in “distillation”, and AI companies can replace expensive RLHF with synthetic examples. It can also make them worse. But you’re not the one curating the datasets or deciding what goes where and how.

    In general in ML it’s not advised to train a model on its own output. That in itself can’t make the predictions any better, just worse.





  • hendrik@palaver.p3x.detoSelfhosted@lemmy.worldWolfstack?
    link
    fedilink
    English
    arrow-up
    5
    ·
    7 days ago

    Yes. With other projects, I often found it is problematic. Like Claude come up with lots of advertisement text, but the software doesn’t even do a fraction of it. Or the install instructions are made up and nothing works… So I usually advise for caution once a project has a wide disparity in claims, stars and signs of actual usage… But I can’t tell what’s the case here, without a proper look. It definitely has some red flags.

    I appreciate people being upfront, as well. Ain’t easy. Just try to install and test it before advertising for the project.


  • hendrik@palaver.p3x.detoSelfhosted@lemmy.worldWolfstack?
    link
    fedilink
    English
    arrow-up
    9
    ·
    7 days ago

    Yeah, they’re transparent about AI usage. There’s a small paragraph at the bottom of their README.

    I mean the website sounds like AI text. The repo is fairly new. Only 1 issue report about how something doesn’t work, zero PRs and seems it’s a single person uploading commits… I’d wait a bit before deploying my production services on it 😅 They’re making a lot of bold claims in the README, though.


  • I think so as well. The computer isn’t really good to “use” it. That’s more the category for experiments. Or teach people how to install Linux. Or a computer museum corner and you put vintage games on it. Or just recycle it.

    And a box with RAM sticks collecting dust isn’t useful either. Put whatever is compatible into other computers, and then try to sell and recycle them. Seems 4GB DDR3L RAM modules still sell for 1 to 4€ on eBay?! So maybe you can make a few bucks to invest in other projects for the kids.


  • I think you need some Agent software. Or a MCP server for your existing software. It depends a bit on what you’re doing, whether that’s just chatting and asking questions that need to be googled. Or vibe coding… Or query the documents on your computer. As I said there’s OpenClaw which can do pretty much everything including wreck your computer. I’m also aware of OpenCode, AutoGPT, Aider, Tabby, CrewAI, …

    The Ollama projects has some software linked on their page: https://github.com/ollama/ollama?tab=readme-ov-file#chat-interfaces
    They’re sorted by use-case. And whether they’re desktop software or a webinterface. Maybe that’s a good starting point.

    What you’d usually do is install it and connect it to your model / inference software via that software’s OpenAI-compatible API endpoint. But it frequently ends up being a chore. If you use some paid service (ChatGPT), they’ll contract with Google to do the search for you, Youtube, etc. And once you do it yourself, you’re gonna need all sorts of developer accounts and API tokens, to automatically access Google’s search API… You might get blocked from YouTube if you host your software on a VPS in a datacenter… That’s kinda how the internet is these days. All the big companies like Google and their competitors require access tokens or there won’t be any search results. At least that was my experience.



  • We got open-source agents like OpenCode. OpenClaw is weird, and not really recommended by any sane person, but to my knowledge it’s open source as well. We got a silly(?) “clean-room rewrite” of the Claude Agent, after that leaked…

    Regarding the models, I don’t think there’s any strictly speaking “FLOSS” models out there with modern tool-calling etc. You’d be looking at “open-weights” models, though. Where they release the weights under some permissive license. The training dataset and all the tuning remain a trade secret with pretty much all models. So there is no real FLOSS as in the 4 freedoms.

    Google dropped a set of Gemma models a few days ago and they seem pretty good. You could have a look at Qwen 3.5, or GLM, DeepSeek… There’s a plethora of open-weights models out there. The newer ones pretty much all do tool-calling and can be used for agentic tasks.






  • This reads like it’s written by OpenClaw?!

    All open-source. […] You built this. Not a vendor. Not a consultant. Not a managed service provider who will send you an invoice next month for the privilege of using what was always supposed to be yours. You opened a terminal, followed a guide, made decisions, fixed the things that broke, and kept going.

    Aha?

    • Cloudflare not open-source
    • OpenAI not open-source and they DO send you a bill
    • Anthropic not open-source and they do send you a bill
    • Google not open-source and they do send you a bill
    • Perplexity not open-source and they do send you a bill
    • supabase.com not open-source and the free service is limited
    • QuickBooks Online is proprietary, so are Xero, FreshBooks and Wave?

    4 Part Series

    Ah a 4 part series in 5 parts with one part missing?

    zero-trust through eight independent layers

    I don’t think the layers build on top of each other. That’s just random things all shoehorned in. One firewall is enough to block 100% of packets, you don’t really need 3 to do the very same thing. And then delegate it to Cloudflare anyway.

    OpenClaw

    And now you got zero security layers. And I bet your API bill will be way more than 3-5 inference runs per day with that.

    Step 1: Apache Guacamole

    What do you need RDP for?

    Step 9: AES-256 Encrypted Backup

    Please(!) don’t do “backups” like that. Learn how to do Docker and what makes sense in that environment, how to backup your databases. And the need to keep backups somewhere that’s not just the same harddisk. And do test them. And you should really consider following the 3-2-1 rule if this is your company’s data or you rely on it as a freelancer.



  • hendrik@palaver.p3x.detoLinux@lemmy.mlLearning Linux via AI
    link
    fedilink
    English
    arrow-up
    2
    ·
    23 days ago

    Speaking from my own experience… Lots of people try to cobble together information and try to learn something quick. To varying degrees of success… But it’s a bit of a hit and miss sometimes. And you don’t necessarily learn it the proper way or the right way around if you go by the random order your questions arise.

    I think one of the most efficient ways (and least time-consuming in the long run) is still good old books. They’re mostly written by clever people. And they come with the information curated. And laid out in the correct order, so you’ll get the basics first and then the stuff building on top of that. So you don’t need to waste a lot of time jumping back and forth and get entangled because you don’t really know you’re missing some basics while learning some advanced concept.

    It’s not easy either. I mean first of all you gotta find some book that matches your learning style. And then I regularly struggle with the first few chapters because I kind of already know 70% of the stuff, yet not all of it. So it’s tricky to hit some balance between brushing over things, and not missing important information… But it gets better after that.

    But I think more often than not, it’s the proper way. And since it’s curated and all, it’ll save time in the long run.

    (I can’t really compare it to the AI approach. I’ve used AI to look up documentation for me. But never used it to learn any more complicated concepts. So I don’t have any first-hand experience with that.)