Hacker News new | past | comments | ask | show | jobs | submit | gradascent's comments login

They're all names of popular Kendrick Lamar songs. Don't ask me why though


Perhaps AI generated?


Or maybe the final moment was a sigh of acceptance and gratitude for the live he lived. Nobody knows but him.


On a similar note, I really hope that the AI companies that don't make it, but have invested a lot in curating and annotating high quality datasets, would release them to the public. Autonomous car and robotics companies in particular since that kind of data doesn't exist on the internet as abundantly as, say, natural language text.


If you want to gain familiarity with the kind of terminology you mentioned here, but don't have a background in graduate-level mathematics (or even undergrad really), I highly recommend Andrew Ng's "Deep Learning Specialization" course on Coursera. It was made a few years ago but all of the fundamental concepts are still relevant today.


Fei Fei Li and Andrej Karpathy's Stanford CS231N course is also a great intro to the basic of the math from an engineering forward perspective. I'm pretty sure all the materials are online. You build up from the basic components to an image focused CNN.


I didn't find it confusing at all. I think it's totally ok to re-use phrasing made famous by someone else - this is how language evolves after all.


Thank you, I appreciate it.


Interesting! I would like to learn more about how AI is being applied to robotics. Do you have any suggestions for how to keep up with developments/ideas in this field?


These two links could be a good start:

ALOHA-2: https://aloha-2.github.io/

RT-X: https://robotics-transformer-x.github.io/


In October I wrote a blogpost on this subject: https://hlfshell.ai/posts/llms-and-robotics-papers-2023/

..and plan to do an updated version soon for much of what's been released since. I've also done work related to LLM and robotics integration, also on that site.

Happy to chat about it.


Working my way through your blog post and it is so refreshing. Unfortunately my algorithm currently is showing me takes which are extreme on either end (like in your blog post).

> Technology’s largest leaps occur when new tools are provided to those that want to make things.

I love this sentence. And the general attitude of curiosity of your post.


Thanks! Appreciate the kind words. I should have in the next month or so (interviewing and finishing my Master's, so there's been delays) a follow up that follows more advancements in the router style VLA, sensoiromotor VLM, and advances in embedding enriched vision models in general.

If you want a great overview of what a modern robotics stack would look like with all this, https://ok-robot.github.io/ was really good and will likely make it into the article. It's a VLA combined with existing RL methods to demonstrate multi-tasking robots, and serves as a great glimpes into what a lot of researchers are working on. You won't see these techniques in robots in industrial or commercial settings - we're still too new at this to be reliable or capable enough to deploy these on real tasks.


Very cool. I'm curious - did you find the results from your mixture of experts model to be (qualitatively) better than with the standard approach?


Thanks! So this is something I tried and qualitatively I didn't see a huge difference. I'd like to swap out my hand rolled modules with standard pytorch modules for self attention etc. and train it on the wikipedia English split. That's on my to-do list for sure.


I run some tests. Single model of the same size is better than MoE. Single expert out of N is better than model of the same size (i.e. same as expert). 2 experts are better than one. That was on small LLM, not sure if it scales.


fyi I think you have bias and variance the wrong way around. Over-fitting indicates high variance


Thank you for catching that. Corrected.


Then perhaps a method emerges out of this to make training faster (but not inference) - do early training on highly quantized (even ternary) weights, and then swap out the weights for fp16 or something and fine-tune? Might save $$$ in training large models.


Because it shifts the burden (or at least appearance) of responsibility from those experiencing homelessness to the government orgs tasked with housing them.


Uh... how does "unhoused" do that? Or, I don't see how unhoused is synonymous with "the government has not provided these people with a house". The opposite of unhoused would be housed. Is everyone that is housed in that position because the government provided a house for them?


Wrongly so I’d argue. It’s your own responsibility to secure a place for yourself (to live, and in society generally). Failure to do this is personal, not collective.


It does not. That is not how language works.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: