• 0 Posts
  • 77 Comments
Joined 2 years ago
cake
Cake day: June 19th, 2023

help-circle
  • Yep! Give granite a try. I think that would be perfect for this use case both in terms of able to answer your queries and doing them quickly, without a GPU by just using modern CPU. I was getting above 30 tokens per second on my 10th gen i5, which kind of blew my mind.

    Thinking models like r1 will be better at things like troubleshooting a faulty furnace, or user problems, so there’s benefits in pushing those envelopes. However, if all you need is to give basic instructions, have it infer your intent, and finally perform the desired tasks, then smaller mixture of experts models should be passable even without a GPU.



  • Depending on what you want to do with it, and what your expectations are; the smaller distilled versions could work on CPU, but most likely will need extra help on top, just like other similar sized models.

    This being a reasoning model, you might get a more well thought out results out of it, but at the end of the day, smaller parameter space (easiest to think as ‘less vocabulary’), smaller capabilities.

    If you just want something to very quickly chat back and forth with on a CPU, try IBM’s granite3.1-moe:3b, which is very fast even on a modern CPU, but doesn’t really excel in complex problems without additional support (ie: RAG or tool use).






  • Completely agree with you on the news vs science aspect. At the same time, it is worth considering that not all science researches are evergreen… I know this all too well; as a UX researcher in the late 2000s / early 2010s studying mobile UX/UI, most of the stuff our lab has done was basically irrelevant the year after they were published. Yet, the lab preserved and continues to conduct studies and add incremental knowledge to the field. At the pace generative AI/LLMs are progressing, studies against commercially available models in 2023 is largely irrelevant in the space we are in, and while updated studies are still important, I feel older articles doesn’t shine an appropriate light on the subject in this context.

    A lot of words to say that despite the linked article being a scientific research, since the article is dropped here without context nor any leading discussion, it leans more towards the news spectrum, and gives off the impression that OP just want to leverage the headline to strike emotion and reinforce peoples’ believes on outdated information.




  • chiisana@lemmy.chiisana.nettoTechnology@lemmy.ml*Permanently Deleted*
    link
    fedilink
    arrow-up
    1
    arrow-down
    4
    ·
    edit-2
    7 months ago

    Linguistic question: is it misogyny if it originates from women? Reason for asking is because I genuinely don’t know if it is like racism against own race kind of situation, and the article appears to have been written by two women.

    Edit: lol Lemmy showing their true colors. Would rather dodge and avoid the hard questions, downvote and continue to circle jerk themselves about anti-AI. Love it. Keep it up Lemmy!


  • What’s that joke? Think of how stupid the average person is, and realize half of them are stupider than that?

    Same idea here.

    You’d find about half of people whose creativity level being lower than the “average” (technically, mean). If Gen AI is learnt from the totality of our collective knowledge, it should help those on the lower half of the curve much more than those above the curve. However, since Gen AI itself is not able to create new concepts, the collective end up creating more of the same stuff that Gen AI is regurgitating from its training material.

    I don’t think this is necessarily a bad thing. This doesn’t apply only to creativity but all spectrum of general knowledge, and should help with raising equity and equality for the humanity at large.




  • The amount of people who would pay is going to be near zero in the grand scheme of things.

    Next time you’re anywhere where you could discretely look at people’s phones, see how many of them run apps with ads. Most apps will offer very cheap IAP to remove ads, but people choose to not pay it. Vast majority of the users have already decided that their time wasted on ads are worth less than whatever tiny monetary cost it would be to remove them. Same thing here: Vast majority of the users have already decided they’re not going to pay to get rid of the ads. This in turn means due to how few people who would be willing to pay, it is not going to be nearly sufficient to keep the infrastructure required up and running, as well as keep the creators compensated for creating the content.


  • Japan has nicovideo.jp as well. Russia has Yandex Efir (gone through a couple rebrands, Efir was the name in 2020 when we were discussing deals; it was operating under another name prior, and I think it is superseded by dzen). Off to the side I think vK also has a small video delivery presence like how Facebook has videos in their feeds. China has several platforms: Tencent Video (owned by Tencent), Youku as you’ve called out (owned by Alibaba), XiGua (ByteDance), Haokan (Baidu), and then slew of smaller ones like KuaiShou, BiliBili and that video thing WeChat tries to push. None of these are public service operated by the State, by the way. List really goes on… and I’d know, because I’ve worked in the space for almost 12 years now.

    China’s great firewall aside, all these platforms are tiny in comparison, and in the grand scheme of things, and barely have any reach. In general, these regional are all taking a backseat just like Nebula and alike — if creators’ content are hyperlocal/super niche, they might be okay with smaller regional platforms; but if they’re trying to extend their reach and monetization (to ensure they have money to continue producing content), the creators’ presence on these platforms are really just auxiliary to their primary presence on YouTube.

    Getting viewers to these smaller platforms is going to pose a significant chicken or the egg problem — creators aren’t incentivized to be there because lack of viewer, viewers aren’t incentivized to go there because lack of content. Worse yet, I’ve also seen situations where creators are paid for some period of exclusivity and then when the deal lapses they just go straight back to YouTube.

    Real competitors do not exist, and likely will not exist for the foreseeable future. YouTube is the million pound behemoth when everyone else barely registers on the radar.


  • That’s a drop in the pond in the grand scheme of things. You just out source that out to rights management companies and absolve yourself from that obligation behind safe harbour. This is basically what they’re doing in this department. They’ve built Content ID for digital finger printing, and then invented an entire market for rights management companies on both sides of the equation.

    On the other hand, 500 hours of video footage got uploaded to YouTube every minute per YouTube in 2022 (pdf warning). 30 minutes of video game content (compresses better), just the 720p variant using avc1 codec is about 443MB of space. Never mind all the other transcodes or higher bitrates. So say 800MB per hour of 720p content; 500 hours of content per minute means 400GB of disk space requirement, per minute; 500TB of disk space per day.

    That’s just video uploaded to YouTube. I don’t even know how much is being watched regularly, but even if we assume at least one view per video, that’s 500TB of bandwidth in and then 500TB of bandwidth out per day.

    Good luck scaling that on public budget.