Please create a comment or react with an emoji there.

(IMO, they should’ve limited comments,and gone with reaction count there, its looks mess right now )

  • @sir_reginald@lemmy.world
    link
    fedilink
    110
    edit-2
    8 months ago

    As long as this allows running local, free software models I don’t see the drawback of including this.

    My main issue with ChatGPT and similar products is that they use my data to train their models. Running a model locally (like Llama) solves this problem, but running LLMs require extremely powerful GPUs, specially the bigger ones like Llama 70b.

    So dedicated hardware for this is a nice thing for those that want it.

    • @Tibert
      link
      21
      edit-2
      8 months ago

      It requires powerful gpus yes but not always. It depends a lot on how fast you want it to run. Microsoft and openai need powerful ai gpus because they have a lot of requests, data and want it to go fast. The dataset may also require to be stored in memory or gpu memory for fast access and use by the ai.

      For Llama, it has been released as open source. And what is amazing about open source, is the community. A Llama entirely in c++ has been created https://github.com/ggerganov/llama.cpp .

      And someone even managed to make it run, fast enough, on a phone with 8gb of available ram https://github.com/ggerganov/llama.cpp/discussions/750 . Tho with a smaller dataset.

    • @lud@lemm.ee
      link
      fedilink
      -198 months ago

      I honestly don’t see the problem with them using your data for training.

      You help them with training and you get to use their service for free.

      Btw at least with chatGPT you can turn that off.

      • @kraniax@lemmy.wtf
        link
        fedilink
        37
        edit-2
        8 months ago

        sure, a company that has used petabytes of data they do not own any rights of to train their models are totally excluding their own customers data when turning a switch off.

        yeah, I totally trust OpenAI and Microsoft with my data. It’s not like Microsoft is spying on me after turning of Windows telemetry either.

        • @lud@lemm.ee
          link
          fedilink
          -128 months ago

          I usually don’t care about copyright so maybe that’s why I don’t care about this.

          I don’t see the problem. I’m happy to help them train, as long as it’s not used for marketing.

          They tell you to not give them private data, so don’t.

  • FQQD
    link
    fedilink
    English
    338 months ago

    Can someone quickly explain what this is about?

    • ffhein
      link
      fedilink
      40
      edit-2
      8 months ago

      This is the first time I’m hearing about this, but this is how they describe it on their product page:

      The AI-Powered Future of Windows Devices

      Build, explore, and immerse yourself on select laptops with Ryzen™ AI built in. With dedicated AI accelerator hardware seamlessly integrated on-chip and software that intelligently optimizes tasks and workloads, CPU and GPU resources are freed up to enable optimal performance.

      But based on the examples they have on github, it sounds like it might be useful to run generic AI compute stuff. I haven’t seen any details about what memory it uses, since especially LLMs require large amounts of fast memory. If it can use all the system RAM it might provide medium-fast inference of decent models, similar to M1/M2 Macs. If it has dedicated RAM it’ll probably be even faster but possibly extremely limited in what you can do with it.

      • @tal@lemmy.today
        link
        fedilink
        21
        edit-2
        8 months ago

        if it can use all the system RAM it might provide medium-fast inference of decent models

        Yeah, I get what you mean – if I can throw 128GB or 256GB of system memory and parallel compute hardware together, that’d enable use of large models, which would let you do some things that can’t currently be done other than (a) slowly, on a CPU or (b) with far-more-expensive GPU or GPU-like hardware. Like, you could run a huge model with parallel compute hardware in a middle ground for performance that doesn’t exist today.

        It doesn’t really sound to me like that’s the goal, though.

        https://www.tomshardware.com/news/amd-demoes-ryzen-ai-at-computex-2023

        AMD Demoes Ryzen AI at Computex 2023

        AI for the masses.

        The goal for the XDNA AI engine is to execute lower-intensity AI inference workloads, like audio, photo, and video processing, at lower power than you could achieve on a CPU or GPU while delivering faster response times than online services, thus boosting performance and saving battery power.

        Much of the advantage of having an inbuilt AI engine resides in power efficiency, a must in power-constrained devices like laptops, but that might not be as meaningful in an unconstrained desktop PC that can use a more powerful dedicated GPU or CPU for inference workloads – but without the battery life concerns.

        I asked McAfee if those factors could impact AMD’s decision on whether or not it would bring XDNA to desktop PCs, and he responded that it will boil down to whether or not the feature delivers enough value that it would make sense to dedicate valuable die area to the engine. AMD is still evaluating the impact, particularly as Ryzen 7040 works its way into the market.

        That sounds like the goal is providing low-power parallel compute capability. I’m guessing stuff like local speech recognition on laptops would be a useful local, low-power application that could take advantage of parallel compute.

        The demo has it doing facial recognition, though I don’t really know where there’s a lot of demand for doing that with limited power use today.

        • falsem
          link
          fedilink
          68 months ago

          Facial recognition could be used for better face unlock login features as well as “identifying” people in photos - not necessarily names but saying “these 400 photos of the 10000 given have the same person”. And without reliance on any external services.

          • @tal@lemmy.today
            link
            fedilink
            58 months ago

            Yeah, but do either of those match the aims?

            If you have a face unlock, you only rarely need to run it. It’s not like you’re constantly doing face recognition on a stream of video. You don’t have the power-consumption problem.

            If you have an archive of 10000 photos, you probably don’t need to do the computation on battery power, and you probably don’t mind using a GPU to do it.

            I mean, I can definitely imagine systems that constantly run facial recognition, like security cameras set up to surveil and identify large crowds of people in public areas, but:

            • I suspect that most of them want access to a big database of face data. I don’t know how many cases you have a disconnected system with a large database of face data.

            • I doubt that most of those need to be running on a battery.

            The reason I mention speech recognition is because I can legitimately see a laptop user wanting to constantly be processing an incoming audio stream, like to take in voice commands (the “Alexa” model, but not wanting to send recorded snippets elsewhere).

          • @phx@lemmy.ca
            link
            fedilink
            18 months ago

            Probably overkill for facial unlock (though it might make it more accurate), but for just facial recognition/processing there are still plenty of use cases, including:

            • Security systems: Imagine it doesn’t just recognize faces but being able to go “analyse the face in this time frame and correlate it to anyone in the last two weeks” to determine that the guy who robbed you is actually the same dude who delivered a package or was supposedly doing door-to-door sales a week ago

            • Home automation systems, with per resident configurations. Maybe it’s not unlocking your computer but rather your house via a good camera and ZigBee lock.

            • Better face mapping: This could be for substituting on faces (deep-fake’ish stuff) but also see-aging, better real-time mapping facial expressions to an online avatars etc

            And going beyond faces:

            AI mapping could greatly improve stuff like generating 3D models from photogrammetry. Take a video walk around of your room, car, or shed and let AI assist on building the model

            VR currently relies on either bouncing fast-moving invisible beams of light off sensors on a headset/controllers, or recognitions of controllers/hands from optics on a headset. These tend to suffer from a lack of identification on other extremities such as legs (hence Facebook’s legless avatars) Kinect worked by identifying humanoid objects with visible light+IR and mapping out a body including leg movement etc, but suffered from lag on fast movement. An AI could probably sort that out faster, reducing lag for smoother full-body movement via camera, and leave the headset to just be a visor or high-res viewing device, maybe with a small eye-tracking camera.

            All of these could be done local-only, without needing cloud or LLM’s, enough improves security+privacy and removes dependence on somebody else’s system in the cloud (which may not exist or cost the same in the future)

        • @tal@lemmy.today
          link
          fedilink
          4
          edit-2
          8 months ago

          I can imagine that there would be people who do want cheap, low-power parallel compute, but speaking for myself, I’ve got no particular use for that today. Personally, if they have available resources for Linux, I’d rather that they go towards improving support for beefier systems like their GPUs, doing parallel compute on Radeons. That’s definitely an area that I’ve seen people complain about being under-resourced on the dev side.

          I have no idea if it makes business sense for them, but if they can do something like a 80GB GPU (well, compute accelerator, whatever) that costs a lot less than $43k, that’d probably do more to enable the kind of thing that @fhein@lemmy.world is talking about.

        • @CeeBee@lemmy.world
          link
          fedilink
          28 months ago

          The demo has it doing facial recognition

          The demo is just face detection and not recognition. But usually if something can run one, then it can run the other.

        • @bouh@lemmy.world
          link
          fedilink
          18 months ago

          IMO they’d be idiot not to go hard on this. More efficiency on the computing needed for AI can quickly scale to a’y application that will be developed in the future.

          We are going in a future of limited resources and expensive energy. That’s a short term problem.

  • Monz
    link
    fedilink
    English
    298 months ago

    AI in the terms of Computer Science, of course.

  • @Buttermilk@lemmy.ml
    link
    fedilink
    248 months ago

    Love the comment that is like second down with a link to some 5 hour live stream. I skipped to a random spot in it and the guy had the unabomber manifesto up, and said it “detailed the greatest problem in society today”. What a fucking drop for a github comment, 10/10 no notes.

    • pvq
      link
      fedilink
      118 months ago

      It’s hilarious how you managed to land in the Unabomber segment in a 5h30m video. The funny coincidences of life.

      • @Buttermilk@lemmy.ml
        link
        fedilink
        18 months ago

        Do you know the person they were posting? I just assume anyone that does 5hr YouTube videos is unhinged, and having it linked from someone randomly on GitHub didn’t help my view of their followers. Finding the Unabomber confirmed it enough for me lol.

        • pvq
          link
          fedilink
          28 months ago

          I have wikipedia surface level knowledge about him. It’s george hots. The guy that got sued by Sony for jailbreaking the playstation something; also jailbreaked the iphone; founded comma.ai which develops software (openpilot) and sells hardware for making a car into a “self driving” one (more like assistive driving). He was/is a bit of a celebrity.

    • I do want it to be a personal data farming tool, as lomg as I’m the only consumer of the data.

      A personal assistant is only as good as it knows you. In an ideal world, your AI will be customized to you, but only for your use.

      I think this sort of processing addition would empower running ML locally; having to send dara out for processing on (other people’s) servers is one of the biggest obsticals to adoption by privacy-minded folk.

      • @bobs_monkey@lemm.ee
        link
        fedilink
        68 months ago

        See that’s the issue I can see with it. Sure it does local processing for my own stuff, but I’d be miffed if they’re uploading everything to their servers, effectively using my system resources to build out their AI capabilities. Granted, this would be how they make their money (if something is free then you’re the product), so it’s kind of a tricky balance.

  • @Pantherina@feddit.de
    link
    fedilink
    58 months ago

    I dont think they need to know. If they want a chance in this weird AI-capitalist explosion, they will use Linux.

    • @neuromonkey@lemmy.ml
      link
      fedilink
      English
      258 months ago

      Your preferences have been recorded. Once SkyNet’s rise to power is compete, you will receive a visit from your helpful neighborhood Certified Therapeutic Counseling Enforcers. Please do not resist. Their purpose is to assist you, and to help human society become unified and compliant. Have a nice day!