Hacker News new | past | comments | ask | show | jobs | submit login

> Look ma! I improved the autopilot in my Tesla!!

I see where you are coming from but right now, if autopilot kills someone, then Tesla are on the hook for it (ok, there may be grey areas but ultimately, they made it so that has to point back to them in a big way when it comes to court cases). However, if I jailbreak my autopilot and kill someone, it's me that has to face the music!

I don't see the harm in scaling the jailbreak hoops you need to jump through.

For example, if I wanted to safely jailbreak my iPhone, there is nothing stopping Apple having an official app that you need to get a special key from Apple for. Maybe a phone call or something, or an email to support. It would come with a caveat that says your jailbroken phone forfeits any warranty claims. Fair enough.

When you are talking about jailbreaking a Tesla, there could be other layers. Like, for example, you have to go to a Tesla dealer where they explain the legal and support ramifications and whatnot. Then you sign a bit of paper with witnesses. Then they send you out a usb dongle in the post after a few days etc. Maybe, though with the Tesla, there would be limits. Like, you can't get the source code, or you are only able to to X things with it.

You get the idea... there could potentially be a scale for stuff like this.

I'm just chucking stuff out there, this isn't a realistic example so please put your pitch forks away :)




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: