If it’s just one large file I just play it on a podcast app (in my case Podcast Addict) and if it’s multiple I stich them together with ffmpeg before sending it to my phone.
If it’s just one large file I just play it on a podcast app (in my case Podcast Addict) and if it’s multiple I stich them together with ffmpeg before sending it to my phone.
Every day, I have an app that connects the audio of my pc to my phone through wifi and using bluetooth would just make the whole setup unusable for calls + wifi has better range and audio quality than bluetooth. The whole setup is PC->Ethernet->Router->wifi->Phone for ~30 ms of latency and the app is soundwire.
They are, but training models is hard and inference (actually using them) is (relatively) cheap. If you make a a GPT-3 size model you don’t always need the full H100 with 80+ gb to run it when things like quantization show that you can get 99% of its performance at >1/4 the size.
Thus NVIDIA selling this at 3k as an ‘AI’ card, even though it wont be as fast. If they need top speed for inference though, yea, H100 is still the way they would go.
If it crashes hard I look forward to all the cheap server hardware that will be in the secondhand market in a few years. One I’m particularly excited about is the 4000 sff, single slot, 75w, 20GB, and ~3070 performance.
Both are bad, Apple needs to stop the parts locking (like not being able to use all the features on a replaced screen without their approval), and Android companies have to be better about giving support for their phones.
Keep both accountable, let’s leave the console wars bs in 2008.