

I use Shotcut for more or less any video operations that require re-encoding. It’s great for basic editing but also simple transcoding jobs too
I use Shotcut for more or less any video operations that require re-encoding. It’s great for basic editing but also simple transcoding jobs too
It’s a slightly different use case, I default to Losslesscut and switch to Shotcut when I need a vfilter or if I’m just generally willing to concede to making a lossy cut.
Shotcut is way more flexible but I can make a quick clip in Losslesscut with probably 1/3 the number of user effort/inputs. Let alone trying to remember every ffmpeg parameter under the sun just to get consistent usable output
That’s what I’ve always assumed it does since back when quicktime player barely even ran on my PC yet for timeline operations it was significantly more responsive than WMP/MPC.
For Losslesscut I just get around this by encoding my input from source using keyint=n:scenecut=0 in ffmpeg where n is a manually set keyframe interval.
So e.g. if my expected cut occurs on a frame that occurs at t+10 seconds of footage, n can be the same as the fps and then there’ll always be a keyframe exactly at timestamp 00:00:01, 00:00:02 and so on. I can then open it in losslesscut and easily snap to the frame I want and make the cut losslessly.
Yeah the first encode generally means a lossy transcode by the time I get to my final video but being realistic that’d be a part of my workflow either way and this way it’s less
Sure, but this is largely because currently each client doesn’t need to aggregate the whole fediverse. In a decentralised network, you can’t split the sum total of processing required to run the fediverse equally amongst peers. Each peer would need to do more or less the same aggregation job, so the total processing required would be exponentially more than with the current setup. You could still argue it’s a negligible processing cost per client, but it’s certainly way less efficient overall even if we assume perfect i/o etc in the p2p system and even if the client only needs to federate the user selected content
Also just practically deploying a client app that can federate/aggregate constantly in the background (kinda required for full participation) and scale with the growth of fedi without becoming a resource hog I imagine would be pretty tough, like maybe possible yeah but I feel like it makes sense why it isn’t like that also
The instance is the aggregator, if it’s P2P then the aggregation is done by the client. In a torrent swarm you contribute bandwidth, not processing power
There’s caveats to that these days. Official streaming, in practice, sure. But with a debrid/similar service and sufficient bandwidth, you can pirate stream files with equivalent quality to uncompressed Blurays