• 1 Post
  • 30 Comments
Joined 1 year ago
cake
Cake day: June 14th, 2023

help-circle
  • Always back up your stuff, but after doing so, the process is pretty much boot to bios, set boot priority with linux usb at the top, and away you go.

    If you have secure boot enabled, you might have to enter a pass code or passphrase but otherwise its identical to traditional bios. If you want secure boot, which prevents someone else from doing this process to your machine, re enable after you’ve installed nvidia drivers otherwise you’ll have to provide it your secure boot password during and sometimes it likes to break.



  • Hey mate, so this comment is just not productive. I’m going to be a little hyperbolic here: if everyone alive is being advertised to then your “unrelated ways companies making suckers out of their customers” comment isn’t correct or honest. It’s the norm, everyones going through it is totally related.

    I talked about companies that lock you into their ecosystems and force you to have a stake in their business model. They do this for two reasons: you make money and they want it, and if you spend your money elsewhere they don’t get it. Name one phone manufacturer that isn’t stealing your data. Name one social media app that isn’t spyware. Name one online store, review site or fucking cooking blog that isn’t loaded with ad trackers and cursor monitoring shit that tells you to subscribe as soon as you go to close the tab.

    Sure some smaller examples exist (I love lemmy, this place is awesome), sure I can download a free open source os, or just install an:

    Adblocker User agent spoofer Anti track-sender Set my browser to stop allowing targeted ads or download a privacy browser

    but everyone is still stuck using the other products in some capacity just the same. I’m happy for you if you fall outside this, seriously. However, most people do not. We are stuck and it’s because we got prayed upon. So yeah, everyone is the product. Always. No exceptions.




  • Dammit, so my comment to the other person was a mix of a reply to this one and the last one… not having a good day for language processing, ironically.

    Specifically on the dragonfly thing, I don’t think I’ll believe myself naive for writing that post or this one. Dragonflies arent very complex and only really have a few behaviours and inputs. We can accurately predict how they will fly. I brought up the dragonfly to mention the limitations of the current tech and concepts. Given the worlds computing power and research investment, the best we can do is a dragonfly for intelligence.

    To be fair, Scientists don’t entirely understand neurons and ML designed neuron-data structures behave similarly to very early ideas of what brains do but its based on concepts from the 1950s. There are different segments of the brain which process different things and we sort of think we know what they all do but most of the studies AI are based on is honestly outdated neuroscience. OpenAI seem to think if they stuff enough data into this language processor it will become sentient and want an exemption from copyright law so they can be profitable rather than actually improving the tech concepts and designs.

    Newer neuroscience research suggest neurons perform differently based on the brain chemicals present, they don’t all always fire at every (or even most) input and they usually present a train of thought, I.e. thoughts literally move around in the brains areas. This is all very different to current ML implementations and is frankly a good enough reason to suggest the tech has a lot of room to develop. I like the field of research and its interesting to watch it develop but they can honestly fuck off telling people they need free access to the world’s content.

    TL;DR dragonflies aren’t that complex and the tech has way more room to grow. However, they have to generate revenue to keep going so they’re selling a large inference machine that relies on all of humanities content to generate the wrong answer to 2+2.


  • I think you’re anthropomorphising the tech tbh. It’s not a person or an animal, it’s a machine and cramming doesn’t work in the idea of neural networks. They’re a mathematical calculation over a vast multidimensional matrix, effectively solving a polynomial of an unimaginable order. So “cramming” as you put it doesn’t work because by definition an LLM cannot forget information because once it’s applied the calculations, it is in there forever. That information is supposed to be blended together. Overfitting is the closest thing to what you’re describing, which would be inputting similar information (training data) and performing the similar calculations throughout the network, and it would therefore exhibit poor performance should it be asked do anything different to the training.

    What I’m arguing over here is language rather than a system so let’s do that and note the flaws. If we’re being intellectually honest we can agree that a flaw like reproducing large portions of a work doesn’t represent true learning and shows a reliance on the training data, i.e. it cant learn unless it has seen similar data before and certain inputs provide a chance it just parrots back the training data.

    In the example (repeat book over and over), it has statistically inferred that those are all the correct words to repeat in that order based on the prompt. This isn’t akin to anything human, people can’t repeat pages of text verbatim like this and no toddler can be tricked into repeating a random page from a random book as you say. The data is there, it’s encoded and referenced when the probability is high enough. As another commenter said, language itself is a powerful tool of rules and stipulations that provide guidelines for the machine, but it isn’t crafting its own sentences, it’s using everyone else’s.

    Also, calling it “tricking the AI” isn’t really intellectually honest either, as in “it was tricked into exposing it still has the data encoded”. We can state it isn’t preferred or intended behaviour (an exploit of the system) but the system, under certain conditions, exhibits reuse of the training data and the ability to replicate it almost exactly (plagiarism). Therefore it is factually wrong to state that it doesn’t keep the training data in a usable format - which was my original point. This isn’t “cramming”, this is encoding and reusing data that was not created by the machine or the programmer, this is other people’s work that it is reproducing as it’s own. It does this constantly, from reusing StackOverflow code and comments to copying tutorials on how to do things. I was showing a case where it won’t even modify the wording, but it reproduces articles and programs in their structure and their format. This isn’t originality, creativity or anything that it is marketed as. It is storing, encoding and copying information to reproduce in a slightly different format.

    EDITS: Sorry for all the edits. I mildly changed what I said and added some extra points so it was a little more intelligible and didn’t make the reader go “WTF is this guy on about”. Not doing well in the written department today so this was largely gobbledegook before but hopefully it is a little clearer what I am saying.


  • Studied AI at uni. I’m also a cyber security professional. AI can be hacked or tricked into exposing training data. Therefore your claim about it disposing of the training material is totally wrong.

    Ask your search engine of choice what happened when Gippity was asked to print the word “book” indefinitely. Answer: it printed training material after printing the word book a couple hundred times.

    Also my main tutor in uni was a neuroscientist. Dude straight up told us that the current AI was only capable of accurately modelling something as complex as a dragon fly. For larger organisms it is nowhere near an accurate recreation of a brain. There are complexities in our brain chemistry that simply aren’t accounted for in a statistical inference model and definitely not in the current gpt models.





  • I’m thinking data entry for threat hunters, and integrations with our other platforms apis but I couldn’t say anything specific. SSDs are a good shout, I might have tried setting it up with hdds if you hadn’t said.

    Did you find it easier to add connectors in seperate docker containers or within the main octi container?

    It feels like there’s a pretty high ceiling for this platform and the data you can generate. Do you find it easy to create good data? Do you have any habits?

    I’m pretty keen to learn so feel free to answer what you can.


  • Really don’t care much about my cv. This program is a great way to learn about the STIX protocol so no idea what you mean about “no actionable skills”. STIX is an interesting information sharing method, the program is well designed to educate the user on it and seeing the format it imports and exports data will teach me a buttload.

    More to the point, maybe could you be less cynical and share some advice. I’m not going to flex my qualifications cos they’re mediocre but I’ve got smart people around me who just don’t know this particular program and I’m interested to hear from those who do.

    Do you run this program at work or at home? Have you learned anything interesting from using it? Are there avoidable mistakes I could not repeat from hosting it? Answers to those questions would be very useful.


  • I dont see myself doing too much configuration with connectors to begin with which brings some of the difficulty down. I was asking to see if others run anything similar in their home configuration. I’ve met people who run MISP from home before so it sounded feasible to me.

    I was also looking for the community aspect of this, I already knew they had a docker-compose config. I wanted to know who had attempted this before and what they’d learned, that sort of thing.






  • In the update settings she can reset her apt sources back to “default”. It’s not too hard and there’s a gui throughout the process (from memory).

    The package conflicts is an interesting one, if you have the time to post one of these on lemmy I’m sure someone will suggest a fix. It’s probably a apt install --fix-broken or something simple (hopefully) but I’m sure we could work it out.

    Totally agree that these are annoying issues though. See if you can use Nala, it’s a TUI front end for Apt and it’s got some nice user changes like if you run upgrade it updates and upgrades. It also has a fetch feature which finds nearby sources, so you’re always downloading from the closest/fastest source.