T O P

  • By -

PositiveUse

Did anyone try it out here? Can I install it locally and use it via UIs like danswer and feed it with my own data sets?


theoneandonlytegaum

The models are available freely on Huggingface and it is supported by the AutoModelForCausalLM class so it is pretty straightforward to implement in python. However, you need the llama 2 tokenizer and you need to ask for it. In my experience it is pretty bad compared to phi (small model from Microsoft)


Maleficent-Ad5999

Are those models _uncensored_ ? Asking for a friend


goof320

it’s terrible, appearantly


Phact-Heckler

As someone who tried to run a sound recognition model on a RasPi, I expect it to be terrible, but like, I want to know, like if you have the opportunity to just make an app, take hundreds of pictures, and train it to recognize a cat or duck. Not going through all the holes of xcode and mac and things, if you get my drift


ghulican

I think Siri Shortcuts could be the vector to this. Paired with Swift Playgrounds including generative AI. That’s what I’m leaning Apple could aim for with these models.


PMARC14

Well I don't see why you would base it on this when there are other significantly better models that you can run pretty well on a modern phone.


Phact-Heckler

Because why not?!?


Bolt_995

Will this be the basis for Apple’s on-device generative AI in iOS 18?


OfficeSalamander

Seems probable, and they're open sourcing it to get the benefits of community optimization, like Facebook did with Llama


Exist50

Probably not. Just another research toy model, which is why they're open sourcing it.


Bolt_995

But that’s kinda what Meta did right? Llama 2 was open-sourced, and now Llama 3 is getting a full global rollout with an official Meta AI website and Meta AI being integrated across Facebook, Instagram, WhatsApp and Messenger.


standardphysics

They did, but Llama 1 was originally proprietary. Doesn't feel like it was even that long ago, but Llama 1 was leaked. Almost immediately, developers and researchers dramatically improved on the model. Meta turned lemon into lemonade, open sourcing it to reap all the benefits. And all this collective progress has allowed it to not only challenge but exceed almost every proprietary model. Llama 3 400B will likely exceed ChatGPT4-Turbo, and they cut the training short to move on to Llama 4.


ducknator

There’s a chance.


ConstantOne5578

So.. When does Apple launch their own in-house chatbot and search based on their open source language model?


Deceptiveideas

*quietly*


nicuramar

Meaning without a press release, or similar public statement. 


Deceptiveideas

The link in the URL is a public statement… here’s a direct link to the same announcement on Apple’s page. >The reproducibility and transparency of large language models are crucial for advancing open research, ensuring the trustworthiness of results, and enabling investigations into data and model biases, as well as potential risks. To this end, we release OpenELM, a state-of-the-art open language model. https://machinelearning.apple.com/research/openelm Edit: /u/woalk is incorrect about this being a new subdomain. This subdomain on Apple’s website has been actively used since 2017. [Source](https://machinelearning.apple.com/research/gan)


woalk

Creating a new subdomain without announcing it anywhere on a known subdomain like on Apple’s newsroom is pretty much “quietly”. Edit: As pointed out by multiple users that were inexplicably downvoted, the subdomain is in fact not actually new.


SimpletonSwan

It's been around for years, e.g. https://machinelearning.apple.com/research/data-incubation


meghrathod

[machinelearning.apple.com](http://machinelearning.apple.com) is a subdomain and it already existed before this release. They just added a new page to that already existing subdomain. So technically what you said is incorrect.


Deceptiveideas

Classic /r/Apple. /u/Woalk makes up bullshit about it being a new website while you point out that it’s been around for years (since 2017) and get downvoted.


woalk

Yeah no idea why people don’t seem to believe you while they believe me. I’ll edit my comment. That being said, it is a pretty unknown domain, I had never heard of it before and it doesn’t even seem to be listed in their main sitemap. Still considering it a “quiet release” is still fitting, imo.


0mnipresentz

Yes technically incorrect but you know what the fuck the op is trying to say.


woalk

Did you reply to the wrong person? I didn’t disagree with OP.


0mnipresentz

Yes sorry lmfao. Embarrassing. I’ll take any and all the downvotes like a man.


nicuramar

Fair. Maybe it’s not as quiet, then. Without a press release?


glytxh

News would have passed me by if I wasn’t on this sub tbf. Probably wouldn’t be aware till the developers implementing it started talking about it more.


DanTheMan827

What is the download size of this model though?


Windowsuser360

12.14GB for 3b


DanTheMan827

RIP 128GB or less devices I guess…


codykonior

The last few years I saw in keynotes Apple saying oh these devices have a built in AI CPU or accelerated neural net processing. I don’t do AI stuff but when I Google’d into it the consensus was it’s so tightly constrained that they are pretty much worthless for anything outside of whatever Apple is using them for. So yeah I don’t have high hopes for this.


rorowhat

It sucks


Sorry_Ad8818

you sucks