Hacker News new | past | comments | ask | show | jobs | submit | burningion's comments login

Author here.

This is a genuine concern of mine! I don't want to build something that generates slop.

Rather, I think whenever we change the costs / process of things, new possibilities open up.

As an example, last night I re-watched Starship Troopers for the six-hundredth time. I'm a huge fan of Paul Verhoeven.

What if I could watch a custom edit of Starship Troopers on demand, and this edit surprised me with something new? I don't know exactly how this would look, but maybe it's interesting?

It's tough to predict the future and how things will change.

But I'd rather be participating in its creation, trying to make it better.


>What if I could watch a custom edit of Starship Troopers on demand, and this edit surprised me with something new? I don't know exactly how this would look, but maybe it's interesting?

this is not interesting whatsoever actually


lol alright, appreciate the skepticism.

What if instead of an algorithm designed to hold your attention captive to sell you shit, a feed of videos created to help you focus on what you aspire to learn / be / do?

idk probably bullshit too, but why not?


We don't advance as a society unless people ask new questions. Having folk willing to spend some time answering those questions (in public, no less!) helps others. It's really, really damn hard to predict how advancements in one area can help another.

All that said, thanks for your interesting new question, and thanks for spending time on it :D


You have no idea what you're doing do you

Bfr


>What if I could watch a custom edit of Starship Troopers on demand, and this edit surprised me with something new? I don't know exactly how this would look, but maybe it's interesting?

Is this what you want to do? https://www.youtube.com/watch?v=6sUR6ylVH7E


There is a high time cost to actual video editing and managing the details well, quite different than generating one template/perception of it, which is more where social media slop is at.


Author here.

Yes, this is a big feature I've been working on, should be ready for a beta by the end of the month.

I allude to it in the post, but good search (for editing) is a challenge, and necessitates a mix of embeddings/vector search and text models.


Derushing in general is the most time consuming, so not only language pattern recognition but also image recognition: "From the rushes, extract all the sequences with bicycle crashes to give me a pile of clips to use in my edit" !


Yes, agreed.

I film a bunch of skateboarding, and it can take tens of tries to land a trick. Similarly, there's usually an unique sound that signals a trick was finally landed.

Good multi-modal search and discovery is a huge part of cracking the editing problem.


Looks like https://kino.ai addresses that derushing stage, but as a specialized tool rather than as a function inside a video editor - which makes a lot of sense to me.


Tens? It sometimes takes my crew hundreds of tries (all on DV tapes).

How far have you been able to come with search for trick variations? It would be interesting to see a system that can reliably recognize what’s switch, nollie vs fakie etc. Then have it generate a list of all tricks for each skater and perhaps outstanding fails. Just some thoughts.


Detect the cheer everyone makes when the trick lands. Lots of proxy indicators to key off of.


> I allude to it

And that’s why I read the comments to see if anyone mentioned it.

To be able to literally take the source files used to put the video together and edit each piece individually would be great.

I wanted to create a car driving down a road covered in arches if greenery. I got lots of great options but I wanted a particular combination of options. If I could do something like that with video, that would be terrific


Author here. I imagine that being one of the components you can "plug in" to what I'm building.

Imagine taking in a prompt, which describes the video you'd like generated. At render time you pass along variables which get injected to describe the specifics for your audience.

We can then adjust the video edit according to that audience, including mixing generated and non-generated content.


Just finished re-reading Emily Wilson's translation of the Odyssey a few weeks ago.

If you haven't read the Odyssey before, I think her translation is accessible enough to just jump right in.

My favorite part of the story is how central luck is to everything.

The characters constantly accept the role of luck in what they do, and the potential of landing on the wrong side of it.

Few modern stories give luck and randomness such prominence, and downplay our own ability to entirely control outcomes.


That's right, but from the Greeks' point of view it was not luck: it was the diving fate, determined by the Gods. The modern equivalent is religious people who believe that everything is determined by the Jewish god.


Wouldn't the modern equivalent be scientific materialists who believe it is determined by luck? The ancient greeks believed that fate and consequence were determined by divine act, but not necessarily decided by them. You could after all find yourself a pawn in a power struggle between gods, or by being favored by one be used as a weapon against them by a rival or enemy. This is a really different conception of divine interference than what modern abrahamic religious people believe.

The abrahamic religions all more or less believe god is benevolent and acting in our best interest within the constraint of allowing us also to act freely. This seems fairly different both from what ancient greeks believed about fate and modern secular beliefs about luck and coincidence.


The secular belief of luck is that nobody is determining anything: no gods, no divine providence, etc. So, completely unrelated to the Greek beliefs.

As for modern religious people, everyone has a different version of what their god can do or not. Theologians may spend their whole lives trying to support one version of another.


Not luck, but the Μοίραι, the Fates, the most powerful deities in Greek mythology, whose will not even Zeus could change.

If Wilson is translating them as "luck", she's done you and all her readers a disservice, but I get a feeling that's just one of many. Popular as her translations might be, and despite the glowing review int he article it seems she is taking great liberties with the content of the epics and presenting it through a strongly distorting modern lens [1].

Source: I'm Greek, grew up in Greece, went to school in Greece and you betcha we studied the Iliad and the Odyssey (in absolutely abominable modern Greek translations) in high school. I read them back and forth multiple times and drew stick figure cartoons on the margins, usually of Corinthian-helmeted warriors disemboweling each other. a.k.a. The Iliad.

___________________________

[1] Apparently:

https://casa-kvsa.org.za/legacy/AC63-Whitaker-18DEC2019.pdf


> If you haven't read the Odyssey before, I think her translation is accessible enough to just jump right in.

I keep checking, but it is continuously checked out at my local library :D One day I'll get lucky. (Yes, I know about holds; I don't like them; I don't like the sense of obligation, and I enjoy the hunt; yes, it's a quirk.)


Sounds like a TTRPG Live Play.

"such and such would narratively be the best thing to do here"

"natural one"

"well fuck"


You were downvoted, but as a user, I’ve experienced the same inconsistency with deploys.

There’s also a hard container size limit I’ve run into multiple times. If you add a dependency and go over the size limit, your app won’t deploy unless you switch to a GPU instance, which is substantially more expensive.


Don't switch to GPU instances just to get around the container size limit!

What are you trying to do and what's the size of your container? My instinctive response here is that you're not holding it the way we expect you to: that some of what you're sticking in your container, we'd expect you to keep in a volume.


One thing you can do to minimize container size is have multi-stage builds in your Dockerfile so that you are not affected by cache bloat.


I see the limitation used to be 3, but was raised over a year to 8 for most modern GPUs:

https://videocardz.com/newz/nvdia-geforce-gpus-now-support-u...

Anyone know what sort of a difference "unlimited" NVENC streams means in practice for a 4090?


Found this reference to show the potential capabilities of each card:

https://www.elpamsoft.com/?p=Plex-Hardware-Transcoding

Seems a 4090 is capable of 22 concurrent h264 / av1 / hevc streams, quite a bit less than the default unlocked 8.


Is decoding limited then?

That link above talks about only 1x nvdec task. What does this mean in practice? Clearly it doesn't mean just decoding a single h264 stream??


So I think if I'm reading it right, it means there's 1x NVDEC chips. Each chip is capable of handling multiple encodings / decodings, as each of the transcodings list means concurrent streams of NVDEC -> NVENC.


The iPhone 15 Pro now has support for L5 GPS, and there's a (partially) operational constellation. This supposedly improves performance, but there's no real information beyond a mention on the iPhone stats page:

https://www.apple.com/iphone-15-pro/specs/

The "Precision dual-frequency GPS" is it, and the full constellation is scheduled to be operational in 2027.


I wrote about the difference between the growth of AI and the growth of Open Source last year:

https://www.makeartwithpython.com/blog/what-is-happening-wit...

tldr; too much capital necessary for wide participation, secrecy and paranoia is a key part of building a moat, and questionable training data. the high salaries just add to the pressure.


The problem is this tool (Autopilot) advertises itself as something other than what it is. (Driver assist.) The report also states that this tool in particular behaves differently from its peers in that it discourages the driver from participating in corrective driving while it's engaged, leading to less attentive use of the tool.


See my other reply.

A name change will not fix inattentive drivers. You’re falling into exactly the same trap I’ve described: excusing everything but the driver.


Yes! https://reactflow.dev/ is what I’ve been using, has been great.


Cool! Thanks for sharing. It explains why this type of ui is being used so often. Building this from scratch isn’t the hardest thing in the world, but I can imagine it is somewhat of a challenge to get just right


Consider applying for YC's W25 batch! Applications are open till Nov 12.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: