I could easily see companies, especially enterprise-level companies, expect code that was generated with AI to have some level of ownership attributed to that AI. Whether a simple "Co-Authored-by Copilot" byline on the commit is the right way to do that is another question though.
How common is this that they would even care about it anyways? I've run Firefox exclusively for the last 2 decades and have never once run into a site that told me I needed to switch to Chromium for compatibility.
In the past Vivaldi used their own user agent string and they ran into a bunch of issues. And they are a chrome derivative! They had to default to the chrome user agent. Here are the examples they cite in their announcement of the decision:
"On Google.com if you present a Vivaldi user agent and arrive via a redirect, the search text box will be misaligned
On Google Docs if you present a Vivaldi user agent you will receive a warning
On Facebook’s WhatsApp web interface if you present a Vivaldi user agent, you cannot enter the site and are advised to switch to one of our competitors
On Microsoft Teams (chat and collaboration website), presenting a Vivaldi user agent will stop you from being able to use the website
On Netflix, presenting a Vivaldi user agent results in a suggestion to install Silverlight to play videos… yes… really… Silverlight!"
When these mega-companies block new competitors it really ought to be seen as collusion. Google, Facebook, and Microsoft certainly have the resources to test and approve the occasional new browser.
They don’t even have the resources to test the most common browsers on every scenario of every page of every application, let alone fix every issue such testing would find.
Common enough that Mozilla has full-time engineers working on triaging compatibility issues, so they can either be fixed in Firefox or reported to webmasters. Here are the reports they get: https://webcompat.com/issues
Your experience may be different, but every time I hit the Cloudflare "checking if your connection is secure" turnstyle, it goes into an infinite loop on Firefox. It's the only reason I still have Chrome on any personal device. It may be tracker and privacy settings rather than just Firefox on its own, but I'm not going to run combinatorial experiments to figure out exactly what Cloudflare is looking for, especially since it's probably a moving target.
A lot of systems seem to silently fail on Firefox - my broadband supplier's website failed at the last step of the onboarding process. I managed to get charged for installation (connection by the network operator, UK) twice, have onboarding emails sent, but not have the appointment in my account.
Used Edge, went through completely.
Can't guarantee it was Firefox/browser issues. But this is not that uncommon an occurrence.
I've never had one tell me; they just don't work, or they get stuck in a loop until they consume all my RAM and the gecko engine crashes. That is, assuming they even show me the page, instead of telling me to go away because they they think I'm a bot.
A few years ago I was maintaining the website for a major brand whose products you probably use. To my horror the website did not support Firefox. I gave them a very minimal estimate on what it would cost to support Firefox along with the estimated percentage of Firefox users in their target market. They were not interested.
I actually ran into such issues, in particular with commercial websits. Some browsers I use do not work for my online transactions for instance - annoyingly the local bank I use for logging into my account as well. It is basically the bank hijacking my money and forcing me into using a specific browser (or, at the least, very few; they improved compatibility a bit in the last years, but there were more issues in the past here). It is just a reality of the situation that some websites don't work well on certain browsers.
I switched bank in 2021 and it was hard. No bank advertises "we do compliant chip tan" and no bank advertises "we do not buy an app framework that scans for customs roms".
Switching banks is hard, because all of them suck, are underdocumented and a moving target.
I suppose today it is more difficult. At the time I only had to get a guarantee from the bank that they support Firefox in addition to IE. Getting that guarantee did take them a week, and only happened because I found a sympathetic bank manager who was willing to pressure them on my behalf.
And honestly, I've only ever once encountered a website that required Widevine. And that site was a media site. So if you don't watch DRMed movies in your browser then you don't need Widevine in my experience.
I've found widewine a blessing because news sites that autoplay trash seem to be the only group that uses it (other than paid media platforms like Netflix and Spotify).
The blessing is I can just reject it and it blocks all their videos from playing/downloading.
How so? I use Firefox for all leisure activities, including extensive YouTube usage, and have never noticed any issue. I’m running uBlock Origin and Sponsor Block. I’m logged into a dedicated Google account I made solely for browsing YouTube (so I can keep the viewing history without linking it too obviously to my main Google account).
Page load takes twice as long as Chrome, videos buffer more slowly, and memory usage grows much faster. After a dozen tabs it starts visibly lagging when you press play/pause, the exact same session works flawlessly in Chrome.
It doesn't seem to affect everyone equally. Pretending to be Chrome sometimes helps - not too long ago someone found a piece of code that introduced seconds of busy delay for any non-Chrome user agent.
I second that. Loading the pages takes longer before the video starts playing. I also regularly run into bot detection where no video will play at all until I log in.
Not for every YouTube Premium subscriber in populated areas of California with decent internet connections watching popular videos, even with a userscript automatically selecting highest quality. Always instant, zero wait, very few exceptions.
It’s non-Google sites where Firefox may not be as well supported as Chrome, IME.
Is there somewhere you'd recommend that I can read more about the pros/cons of TOTP? These authenticator apps are the most common 2FA second factor that I encounter, so I'd like to have a good source for info to stay safe.
I'm sorry, but this is selling good engineers very short. If you didn't nest your utils folder 8 folders deep, it seems pretty obvious that one should check the utils folder before writing another utility function. This stuff should also be caught in code reviews. Maybe the new guy didn't know that util function existed, but surely you did when you reviewed their MR? Obviously mistakes like that can happen, but I've found that to be the exception rather than the rule, even in some of the gnarlier codebases I've worked in.
Assuming they even have code reviews - in your experience, in a situation where the person writing the code didn't check if it already exists, the reviewer will check that and then tell them to delete their already finished implementation and use that existing thing?
I wouldn't say you should explicitly check, necessarily. More like, you go to implement the widget and when you open the appropriate file to get started, it's already there.
Is that not something that was already possible with basically every AI provider by prompting it to develop learning steps and not to provide you with a direct answer? I've used this quite a bit when learning new topics and pretty much every provider does this without a specialized model.
It's really nice to have something like this baked in. I can see this being handy if it's connected to external learning resources / sites to have a more focused area of search for it's answers. Having hard defined walls in the system prompt to prevent just asking for the answer seems pretty handy to me, particularly in a school setting.
Yeah, for sure. I wasn't asking from the framing of saying it's a bad idea, my thoughts were more driven by this seeming like something every other major player can just copy with very little effort because it's already kind of baked into the product.
Hardware-wise the peak is obviously the M-series. Ditching x86 while simultaneously nearly flawlessly emulating x86 apps via Rosetta - making the transition to ARM64 completely painless - was a landmark achievement.
I think it is CPU / SOC Wise. There is no reason why you cant have old MacBook with M Series. ( Apart from Memory ). All things about previous MacBook Pro still stand. And they could still have all the features while being thinner and lighter.
Current gen is the most optimized but I don’t get that “oh that clever” feeling from interacting with it. It’s all been simplified, which makes it solid and reliable, but that’s about it.
As a non Apple user, yeah, M series are neat in the sense that the premium you pay goes into barring the competition from accessing the current nodes at TSMC, making Apple look good on benchmarks for 12-18 months or so. Apple used to have something else to offer, a sense of novelty, excitement, taste, and couldn't care less about performance. Apple of today is just Samsung/Gates' Microsoft "look at how big mine is!", with more bucks and even more user-hostile practices.
Easy to disagree on this one. MacBooks are easily the best-manufactured computers money can buy. The entry level MacBook is just unbeatable for value, which is very unusual for Apple.
Depends on what you value. To me, MB Pro's keyboard is terrible, and MacOS is abysmal. On an ideological level, I defend right to repair, right to upgrade and oppose vendor lock-in. What does that leave me with? An admittedly decent CPU, a good display and speakers? That's pretty weak to entertain Apple's consumer-hostile charade with my own money.
In fact, now is the best time in the last 20 years for either: fully integrated SoC’s inside laptops (with all the pros and cons of better battery life, lower heat, smaller size - but irrepairability) and almost entirely modular laptops.
I understand that most people want socketed CPU’s in machines, but speaking genuinely storage used to be upgraded more than ram, and ram more than a CPU; CPU’s limit how much ram we can have so having soldered RAM isn’t that big of a deal in reality to most people.
I feel like a heathen saying it, because emotionally I don’t want it to be true, but it’s definitely the truth.
> In fact, now is the best time in the last 20 years for either
> having soldered RAM isn’t that big of a deal in reality to most people.
Precisely. This whole sub-thread is a response to sacralising Apple for their M chips, pretending there is no compelling alternative, and subsequently giving Apple a free pass for consumer-hostile/commercially-dubious practices.
At any point in time, Apple has the lead, there's no argument there, but if you can afford to wait 12/18 months, you get about the same performance in a repairable/extensible package. That makes Apple's performance less stellar, especially when the same people laud those devices' life expectancy (my daily-driver ThinkPad is specced from early 2017, in 2025 it wouldn't care having bought it "old" in 2018).
I get it, this is an affluent forum, people like the latest and greatest, and a lot of Apple's marketing strategy is about validation and status, that's not terribly rational and healthy, though.
You can - and you’ll readily find (as is expected and forgivable from a low run, low R&D budget machine) the build quality is abysmal next to the Mac. The case is flexy, the battery lasts a fraction of the time, the trackpad is nowhere near as good, the processors are anemic or badly thermally managed by comparison.
And if you have something like the MNT Reform laptop[0] (which is even more to the extreme end of repairability) that those things you mention are even worse.
So, there's the rub, and it should be clear: repairability is coming with a trade-off.. Sleekness, performance, battery life.
We as consumers would rather have ultra portable, high performance laptops that feel rigid in our hands over bulky devices that could be useful for 1.5x as long (or survive more wear potentially).
Framework is banking on people who are aware of this tradeoff and want to buy a device going the opposite direction; which is GREAT!.
For example all smartphones are going larger and there's no choice for someone like me to get a small phone these days: I am forced. You're not! Someone is allowing you to make the tradeoff, and instead of understanding that these are actual trade-offs, you'd rather complain that they exist at all.
Come on, there’s no way you wrote that down unironically and didn’t struggle breathing through the strong chemical copium smells.
> goes into barring the competition from accessing the current nodes at TSMC
I know it’s en vogue to hate on Apple and make them out to be this big evil corporation, but you’re naming it sound as if they’ve been jerking off while sitting on TSMC’s capacity just to fuck with the competition and purely to make it impossible to compete, when in reality they’ve continued to make exponential improvements on their silicon platform.
> making Apple look good on benchmarks for 12-18 months or so
What are you on about?
They’ve essentially been in a league of their own since the M1, especially if you take into consideration the power envelope and how performance is available with just passive cooling.
There isn’t really anything like it.
Even the salty argument of Apple hogging TMSC nodes just crumbles apart if you give more than a second of thought.
For starters, yes, sure Apple is great at managing their logistics and supply chain, which is why, when Cook was in charge of that, it impressed Jobs so much and it proved to be so essential to Apple’s success, that Jobs decided to hand pick Cook as his successor.
I don’t see how that is a useful argument against Apple, moral or otherwise.
Nothing is stopping competitors from optimizing their process to the point where they can call TSMC and offer to buy their capacity for the next year or two. To say nothing of the efforts made outside of TSMC like Samsung GAAFET 3nm and MBCFET 2nm process and whatever Intel is dicking around with on their 2nm process.
More importantly though, it’s silly to make it seem as if that’s the only reason for the fruits of Apple’s labor.
Take AMD’s HX 370 for example, released last year, courtesy of TSMC’s N4P process.
It still struggled to provide a PPA similar to the M1 Pro, which wasn’t only 3 years older at the time, it was a product of TSMC’s older N5 process.
Clearly having access to newer TSMC nodes isn’t a guaranteed win.
> and couldn't care less about performance
You’ve got it mixed up. Apple has never cared about raw specs, but they always have and always will care about performance.
If you’re inclined to read their every move through the big bad filter then you might say they never cared about raw performance because they’ve always been able to get more out of less and this way they could charge high spec prices without the high spec cost (and without, historically, advertising specs), and it clearly worked out for them.
Their stuff is being sold as if it’s given away for free, in doing so they’ve proven that the average user couldn’t give two fucks about bigger numbers as long as it works well, and their competitors have to pack their phones and other devices with higher specs and cooling solutions like vapor chambers (something Apple has managed to avoid so far) to keep up.
In a way they’ve always had to care more about performance than their competitors because they’ve mostly worked with hardware that’s “lesser” on paper to maximize their margins.
> to offer, a sense of novelty, excitement, taste
I don’t know about you but single-handedly making x86_64 look like an ancient joke with something that would’ve been considered a silly mobile processor 10 years ago is quite novel and exiting. If nothing else it lit a fire under Intel, even if they’ve seemed to have decided to let themselves be turned into a well done steak.
This was essentially what Intel had in mind with their Atom series for netbooks back in the day and Intel never managed to crack the code.
I remember being amazed when I received my developer transition kit, running macOS on an A12Z like it was nothing.
Even now, if I want to be more comfortable and do some coding or video editing work on the couch I can use my off-the-shelve base model M3 MacBook Air to do most of what I can on my M1 Max, that’s quite the leap in performance in such a short time.
There’s no accounting for taste or course and what I like might not be to your liking, and there is plenty about Apple that deserve legitimate criticism, so I don’t understand the need to make something out of nothing in this instance.
I'm not taking away from Apple's push towards ARM, that was ballsy, and well executed (also, they had little choice but to ditch Intel, and with AMD not being an option it's pretty obvious in retrospect). That said, I'm tired of the rhetoric and attitude that somehow Apple's chips are made of angel dust or something, especially on this "tech"/"science" forum.
> You’ve got it mixed up. Apple has never cared about raw specs, but they always have and always will care about performance.
Apple a decade and a half ago was selling you "unique" products or clever features. Today's Apple announcements is Tim showing you benchmarks.
> I don’t know about you but single-handedly making x86_64 look like an ancient joke
No, they haven't. They did put intel to shame, but so did AMD, and that came as a surprise to nobody.
What do you mean by "in a position"? Do i think it would be successful? of course not, that's mad.
Do I think israel is inclined to try, or otherwise, risk failure on the back of US blood and treasure? More or less, yes -- i think that's quite likely.
The US invasion and occupation of vietnam, afganistan, iraq, etc. were all mad. The US foreign policy elite are not very competent because america doesnt receive any real blowback from its failures -- so there's no conditioning mechanism to force it into instutitonal competence.
Do I think such an elite would do one more stupid thing? yes, its actually far more improbable that they'd learn caution
They've bankrupted america, caused half the world to turn against them -- all the while presiding over the rise and enrichment of a peer competitor (china). You could not describe a more incompent, warmongering, self-destructive set of foreign policy institutions.
It's what happens when you are isolated on your own continent and rarely have to pay for your decisions.
Operations are defined by goals. If you want to invade or launch a special forces op into your enemy territory, you need a small and attainable goal. Not "eliminate all nuclear threats" but more like "clear this area of nuclear materiel" in any areas you consider suspect. Otherwise you end up deploying troops that never come home.
Israel's state government is absolutely filled to the brim with war hawks - but they're not stupid. The situation they want to contain is too large to fix with IDF ground forces, they necessarily have to involve US force structures to seriously challenge Iran. And even then, it feels likely that we'd be looking at an Afghan War situation where guerrilla combat absolutely shreds the modern forces the further they push in.
Look, I don't want to get pissy because your track-record in this comment chain is mostly on-point. Boots are about to deploy on Iranian soil, and it's going to be a deliberate bloodbath for the first few days. Israel is going to piss and moan until America sends over more assets and materiel, at which point we'll be firmly in WWIII territory. It's downright bad, and you're not at all hyperbolic to lay things out like this.
...but I will repeat myself - this is an attack of opportunity for Israel, not a desperate scramble to destroy nuclear assets. Israel's long-term goal is to become the unquestioned geopolitical power of the Levant, even outside America's auspices. They can do that by leveraging the dumb-as-a-brick administration to provoke Iran into a response, at which point they will fight until attrition forces them both to retreat. Now Israeli forces are the de-facto security guarantor in the region, and we already know they draw their borders however they like.
Mind you, this isn't the last you'll hear about "Iran's nuclear program" - it hasn't outlived it's usefulness, quite yet. Israel will continue targeting them not until nuclear assets are destroyed, but until America perceives itself to be backed into a corner with no choice but to search Iran door-to-door for a hidden bomb. (Stretch Goal - +100 Brownie Points: get America to launch a tactical nuclear weapon on Iran and increase the escalation ladder beyond what any peer power can compete with.)
Was just looking at their release strategy. This is being handled by people that have experienced the hell that is dependency management in the JS ecosystem. Kudos to them.
We use it for an application that aggregates data for consumption by several different teams that all consume different subsets of the data. When you have a pretty simple use-case it's really not that bad to get a decently functioning API off the ground, and because it's self-documenting we can spend our time on more mission critical work.
reply