Noa Roggendorff's picture

Noa Roggendorff

nroggendorff

AI & ML interests

None

Recent Activity

Organizations

Gradio-Blocks-Party's profile picture Blog-explorers's profile picture ZeroGPU Explorers's profile picture MLX Community's profile picture Dev Mode Explorers's profile picture Glide's profile picture None yet's profile picture

nroggendorff's activity

replied to their post 3 days ago
view reply

I am not sure if that makes sense, I am under the impression that, if the space is not running(not started), no models can be actively loaded in the space.

Can you share your relevant workflow(docker-compose, app code, etc) so i can see more clearly whats happening?

I might be able to aid in a solution, its possible that there is an issue in the workflow itself.

EDIT: I looked at the spaces, Do you mean this space as an example? 'https://huggingface.co/spaces/nroggendorff/train-llama'
Because this space shows a missing "CUDA_HOME" env var, most your other spaces throwing errors about missing CUDA drivers or are paused. These are configuration errors.

Could you tell me the space and error message?
I might be able to help you fix it.

That’s the one.

replied to their post 3 days ago
replied to their post 4 days ago
view reply

it's pretty specific to my workflow, but spaces now don't get cuda until after they start, so you can't load models or anything until an app is running

reacted to their post with β€οΈπŸ€—πŸš€ 7 days ago
view post
Post
2287
I'm not really doing much on HuggingFace right now due to their new Docker space policies, so if you want to keep up with most of what I'm up to, follow my [instagram](https://sly.sh/ig)
Β·
posted an update 7 days ago
view post
Post
2287
I'm not really doing much on HuggingFace right now due to their new Docker space policies, so if you want to keep up with most of what I'm up to, follow my [instagram](https://sly.sh/ig)
Β·
replied to their post 16 days ago
reacted to their post with πŸ€—πŸš€ 17 days ago
view post
Post
3620
200
Β·
posted an update 18 days ago
view post
Post
3620
200
Β·
reacted to clem's post with πŸ€— 19 days ago
view post
Post
4610
We just crossed 1,500,000 public models on Hugging Face (and 500k spaces, 330k datasets, 50k papers). One new repository is created every 15 seconds. Congratulations all!
Β·
reacted to their post with πŸ˜”πŸ”₯ 20 days ago
view post
Post
2770
to the nvidia employee that won't respond to my emails: hear me now.

you have made a semi-powerful to irrelevant enemy. you have been warned
  • 4 replies
Β·
posted an update 20 days ago
view post
Post
2770
to the nvidia employee that won't respond to my emails: hear me now.

you have made a semi-powerful to irrelevant enemy. you have been warned
  • 4 replies
Β·
posted an update 29 days ago
replied to their post about 1 month ago
view reply

it's a similar architecture to image generation, so.. kinda? diffusion llms aren't very popular though, so there isn't a ton of research on them. transformers is a much more reliable model type for now.

edit: it's not really a super serious experiment, they are more for testing if a logical response is possible this way.
this is also kinda why q and a bots are really bad, people just found that that format doesn't scale very well at all

edit 2: (i said one of, because another huge reason is quality data scarcity and lack of flexibility. with incremental models like gpts, you can have any number of roles and stuff, whereas input-output models just have that)

posted an update about 1 month ago
view post
Post
2818
We're using RLHF on diffusion models, right? Just making sure..
Β·
replied to clem's post about 1 month ago
reacted to m-ric's post with πŸš€ about 2 months ago
view post
Post
2543
For those who haven't come across it yet, here's a handy trick to discuss an entire GitHub repo with an LLM:

=> Just replace "github" with "gitingest" in the url, and you get the whole repo as a single string that you can then paste in your LLMs