Looks very Norwegian, and there seems to be a crossing on the road used to prevent sheep from leaving the area (while cars still can drive over it) which is also something we have in Norway.
Looks very Norwegian, and there seems to be a crossing on the road used to prevent sheep from leaving the area (while cars still can drive over it) which is also something we have in Norway.
Compute becomes cheaper and larger undertakings happen. LLMs are huge, but there is new tech moving things along. The key part in LLMs, the transformer is getting new competition that may surpass it, both for LLMs and other machine learning uses.
Otherwise, cheaper GPUs for us gamers would be great.
It’s a link to an image on github not sure why it doesn’t work for you. Try just looking at the repo then:
(Windows only warning, unless someone wants to add Linux support)
I didn’t really search around for GUIs way back, but ended up making a basic GUI because I wanted to learn programming.
With just having options as checkboxes for YouTube-dl. It has served me well all these years. It was literally the thing I made while learning programming so the code is pretty janky when I look back at it though…
Sideloading apps The home screen layout (I’m sure this can be changed up though), gotta love launchers Live backgrounds that also work with launchers More styling options such as app icons for home screens. While less relevant with gestures now, their navigation setup The punchouts and larger things in the screen. I hate them, and hate that on android too. Apples lock in, esp things like file transfers. Google has some too of course, but Apple is worse.
Quick google shows that Kanban is a method. Mainlu around picking up things as the come, but also limiting how much can happen at once.
The project I’m has a team that uses Kanban for the “Maintenance” tasks/development, take what is at the top of the board and do it. Adapt if higher priority things comes around, such as prod bugs. Our developments teams are trying to implement Scrum, where interruptions are to be avoided if possible during sprints. You plan a sprint, try to do that work, and can present it, and iterate when users inevitably changes criteria.
In the meme, kanban does somewhat make sense, since getting armrests is never going to get a high priority as part of building a rocket. Scrum isn’t exactly right, but I can see where it’s coming from. They are all agile methods though.
I kinda get where he is coming for though. AI is being crammed into everything, and especially in things where they are not currently suited to be.
After learning about Machine learning, you kind realize that unlike “regular programs” that ML gives you “roughly what you want” answers. Approximations really. This is all fine and good for generating images for example, because minor details being off of what you wanted probably isn’t too bad. A chat bot itself isn’t wrong here, because there are many ways to say the same thing. The important thing is that there is a definite step after that where you evaluate the result. In simpler ML you can even figure out the specifics of the process, but for the most part we evaluate what the LLM said or if the image is accurate to our expectations. But we can’t control or constrain the output to exactly our needs, because our restrictions largely are just input in a almost finished approximation engine.
The problem is, that companies take these approximation engines, put them in their product and consider their output fact. Like Ai chatbots doing customer support, and make up facts like the user that was told about rules that didn’t exist for an airline, or the search engines that parrot jokes or harmful advice. Sure you and I might realize that these things come from a machine that doesn’t actually think about it’s answers, but others don’t. And throwing a “*this might be wrong because its AI” on it is not an acceptable waiver of accountability.
Despite this, I use chatgpt and gemini a lot to help me program, they get a lot of things wrong but also do great. It’s a great tool, exactly because I step in after the approximation step, review and decide. I’m aware of the limits. But putting these things in front of “users” without a review step means you are advertising that you are either unaware of this flaw, or just see the cost-benefit analysis and see that if noting else it’ll generate interest during the hype.
There is a huge potential, but throwing AI into a situation where facts are needed when it’s only making rough guesses, is the wrong way about it.
Completely true, but also compression can make anything bad. I’ve seen 480p better 1080p simply because the 480p was using more bitrate, where the 1080p is encoded without enough relatively speaking.
It’s worth adding I greatly prefer MS Auth style authentication, since I don’t have to find the right entry to read the Auth code and then write it on the other computer. Instead MS pops a notification and you either type or select the right number, verify with fingerprint and done. Much more convenient.
It often tells you what you login into and where you are attempt to log in from, so it’s a few extra layers of security for those that have that awareness to check those details.
Played a lot of rainbow six siege, where you have to shoot those 360° security cameras when you are attacking. So, now I’m trained to spot those on instinct.
I disabled my adblock for Twitter to see a update about game server maintenance. It showed me random posts, nothing from this year. Literally unusable site when you can’t even see the latest tweets. Had to have other people tell me maintenance was extended…
I made do with my IDE, even after getting a developer job. Outside shenanigans involving a committed password, and the occasional empty commit to trigger a build job on GitHub without requiring a new review to be approved, I still don’t use the commandline a lot.
But it’s true, if you managed to commit and push, you are OK. Even the IDE will make fixing most merges simple.
Already been explained a few times, but GPU encoders are hardware with fixed options, with some leeway in presets and such. They are specialized to handle a set of profiles.
They use methods which work well in the specialized hardware. They do not have the memory that a software encoder can use for example to comb through a large amount of frames, but they can specialize the encoding flow and hardware to the calculations. Hardware encoded can not do everything software encoders do, nor can they be as thorough because of constraints.
Even the decoders are like that, for example my player will crash trying to hardware decode AV1 encoded with super resolution frames, frames that have a lower resolution that are supposed to be upscale by the decoder. (a feature in AV1, that hardware decoder profiles do not support, afaik.)
Here’s a little article which highlights jxl well. https://chipsandcheese.com/2021/02/28/modern-data-compression-in-2021-part-2-the-battle-to-dethrone-jpeg-with-jpeg-xl-avif-and-webp/
I do not think it’s mentioned there, but I think webp and also it’s indirect successor avif afaik, both lack progressive loading which is not optimal for website loading. It’s has incremental loading which I think is akin the the old dial up time of loading top to bottom row for row. They proclaim progressive decoding is costly on memory and cpu, but progressive gives the best user experience imo.
Lastly a fringe issue, re-encooding multiple times. The good old reason why jpgs turn into trash over time because people encode instead of save images. Or because sites re-encode when uploading. Jxl wins here. It also is very easy to see why jpg turns into what it does rather quickly.
https://www.reddit.com/r/AV1/comments/ju18pz/generation_loss_comparing_jpeg_webp_jxl_and_avif/
All of them are OK, except mkv is less a file type and more a container. What should be specified is the code for video, which for most things I’d say AV1, but high res movies might not be the most suitable. Throw in opus for the audio track, and you can use mkv, but might as well use webm anyways since it’s more clear what’s behind it. (though can still be other things)
I’d also add that jxl should be the standard for lossy images. Better than jpg. And you want something other than png for massive images because that quickly gets costly in terms of size due to png being lossless.
Norway has something similar, you own the inside usually and the HOA own the outside, including the houses themselves. Live in one, largely a good thing but some things come slow since they need to be voted for of course. Generally worth it, since you get good deals on things like internet. It’s cheaper but it’s also something you usually have to use and the only option. Eg only that provider of internet.
I’m my case, they are also responsible for my balanced ventilation, my exterior doors and my water heater. So when the time comes, they handle it. Shared costs cover snow plowing, the shared community building, upkeep of garage, outdoors and the buildings, and things like water bills and taxes paid. In particular, HOAs purchases do not need to pay a 2.5% of the purchase price fee when you purchase a home. This itself saves you quite a bit, and makes up for some of the extra you pay in monthly costs. (but pretty much all of those are at least going somewhere that benefit you anyways)
The downsides are, there are special rules so some people that have membership may have a right to take over the winning bid in a sale. I myself used this to purchase my place, having gotten 10 years of seniority in “HOA company”. You spend the seniority with your purchase, but also are not allowed to own more than one part. Also, no long term renting so there aren’t any companies buying and renting out and things like that. You have to live in the HOA.