I feel like LLMs[1] are going to cause a kind of "divorce" between those who love making software and those who love selling software. It was difficult for these two groups to communicate and coordinate before, and now it is _excruciating_. What little mutual tolerance and slack there was, is practically gone.
Open source was always[2] a fragile arrangement based on the kind of trust that involves looking at things through one's fingers (turning a blind eye may be more idiomatic in English), and we are at the point where you just have to either shut your eyes, or otherwise stop pretending that the situation can be salvaged at all.
Just a thought I had: some people think that LLM-shaming is declasse, and maybe it is, but I think that perhaps we _should_ LLM-shame, until the AI-companies train their LLMs to actually give attribution, if nothing else (I mean if it can memorize entire blocks of code, why can't it memorize where it saw that code? Would this not, potentially, _improve_ the attribution-situation, to levels better than even the pre-LLM era? Oh right, because plagiarism might actually be the product).
[1]: Not blaming the tech itself, but rather the people who choose to use it recklessly, and an industry that is based almost entirely on getting mega-corporations to buy startups that, against the odds, have acquired a decent number of happy-ish customers, that can now be relentlessly locked-in and up-sold to.
Just to add some perspective to this comparison: the US massacred four _million_ people in South East Asia, during the Vietnam war. That is 2/3rds of a holocaust. The Iraq War (second one), cost between half a million and a million lives (estimates vary, and it only includes violent deaths directly caused by American troops -- the war itself caused an increase in crime and murder and out-migration).
I could go on, but Tienanmen does not compare to most of the things the US has done outside of its own borders from 1946 to the present. And no, we (I am American) cannot justify a body count in the millions, just because our victims are communist/authoritarian/theocratic. Note also that we only number 5% of the world's population, and that if we compared body-counts as percentage of populations, instead of as absolute numbers, I doubt we even have enough people to settle that debt.
Even worse, if the world internalizes that it is fine to murder millions of foreigners, just because they are oddballs that their citizens cannot empathize with, the _we_ are going to have a big problem -- we appear much more odd to the world than the world does to us.
I am surprised that our shenanigans have been tolerated for nearly a century.
And the US massacred four _million_ people in South East Asia, during the Vietnam war. That is 2/3rds of a holocaust. The Iraq War (second one), cost between half a million and a million lives (estimates vary, and it only includes violent deaths directly caused by American troops -- the war itself caused an increase in crime and murder and out-migration).
I could go on, but Tienanmen does not compare to most of the things the US has done outside of its own borders from 1946 to the present. And no, we (I am American) cannot justify a body count in the millions, just because our victims are communist/authoritarian/theocratic. Note also that we only number 5% of the world's population, and that if we compared body-counts as percentage of populations, instead of as absolute numbers, I doubt we even have enough people to settle that debt.
Even worse, if the world internalizes that it is fine to murder millions of foreigners, just because they are oddballs that their citizens cannot empathize with, the _we_ are going to have a big problem -- we appear much more odd to the world than the world does to us.
I am surprised that our shenanigans have been tolerated for nearly a century.
Around the time of the pandemic, a company wanted to make some Javascript code do a kind of transformation over large number of web-pages (a billion or so, fetched as WARC files from the web archive). Their engineers suggested setting up SmartOS VMs and deploying Manta (which would have allowed the use of the Javascript code in a totally unmodified way -- map-reduce from the command-line, that scales with the number storage/processing nodes) which should have taken a few weeks at most.
After a bit of googling and meeting, the higher ups decided to use AWS Lambdas and Google Cloud Functions, because that's what everyone else was doing, and they figured that this was a sensible business move because the job-market must be full of people who know how to modify/maintain Lambda/GCF code.
Needless to say, Lambda/GCF were not built for this kind of workload, and they could not scale. In fact, the workload was so out-of-distribution, that the GCP folks moved the instances (if you can call them that) to a completely different data-center, because the workload was causing performance problems, for _other_ customers in the original data-center.
Once it became clear that this approach cannot scale to a billion or so web-pages, it was decided to -- no, not to deploy Manta or an equivalent -- but to build a custom "pipeline" from scratch, that would do this. This system was in development for 6 months or so, and never really worked correctly/reliably.
This is the kind of thing that happens when non-engineers can override or veto engineering decisions -- and the only reason they can do that, is because the non-engineers sign the paychecks (it does not matter how big the paycheck is, because market will find a way to extract all of it).
One of the fallacies of the tech-industry (I do not mean to paint with too broad a brush, there are obviously companies out there that know what they are doing) is that there are trade-offs to be made between business-decisions and engineering-decisions. I think this is more a kind of psychological distortion or a false-choice (forcing an engineering decision on the basis of what the job market will be like some day in the future -- during a pandemic no less -- is practically delusional). Also, if such trade-offs are true trade-offs, then maybe the company is not really an engineering company (which is fine, but that is kind of like a shoe-store having a few podiatrists on staff -- it is wasteful, but they can now walk around in white lab-coats, and pretend to be a healthcare institution instead of a shoe-store).
Personally, I believe that the tech industry sustains itself via technical debt, much like the real economy sustains itself on real debt. In some sense, everyone is trying to gaslight everyone else into incurring as much technical debt as possible, so that a way to service the debt can be sold. Most of the technical debt is not necessary, and if people were empowered to just not incur it, I suspect it would orient tech companies towards making things that actually push the state of the art forward.
There was a moment ca. 2020 when everyone was losing their minds over Lambda and other cloud services like SQS and S3 because they're "so cheap!!11". Innumeracy is a hell of a drug.
A lot of criticism of k8s is always centered about some imagined perfect PaaS, or related to being in very narrow goldilocks zone where the costs of "serverless" are easier to bear...
> Personally, I believe that the tech industry sustains itself via technical debt, much like the real economy sustains itself on real debt. In some sense, everyone is trying to gaslight everyone else into incurring as much technical debt as possible, so that a way to service the debt can be sold.
This feels like a reminder that everything "Cloud" is still basically the same as IBM's ancient business model. We've always just been renting time on someone else's computers, and those someone else people are always trying to rent more time. The landlords shift, but the game stays the same.
It seems that many high-quality things (or otherwise aspirational things) take on Esperanto names (disclosure, I am an Esperantist). While Monero is no doubt a cool crypto-currency, it is even cooler that it has inspired some crypto-curious people to learn Esperanto[1] instead!
While I am here, I might as well give you a brief Esperanto lesson. Mono = money, ero = piece/quantum. So, "pano" = bread, "panero" = bread-crumb. Thus, "monero" = coin.
Many previous international currencies (all of them created with Swiss involvement), were also given Esperanto names: Spesmilo (thousand speso's (speso is analogous to "penny")), Stelo (star).
There is even a luxury watch-brand (from Switzerland) called "Movado", which is Esperanto for "Movement" (made back when watches were made with mechanical movements).
And I also learned, from the linked thread (disclosure, I am a participant), that there is a soft-drink called "Mirinda". This is an adjective that means "awe-worthy".
Back in 2014, I did an analysis of (single threaded) CPU-efficiency and RAM-efficiency of various data-structures (skiplists, slablists, avl-trees, rb-trees, b-trees):
I used whatever I could find on the internet at the time, so the comparison compares both algorithm and implementation (they were all written in C, but even slight changes to the C code can change performance -- uuavl performs much better than all other avl variants, for example). I suspect that a differently-programmed skip-list would not have performed quite so poorly.
The general conclusion from all this, is that any data-structure that can organize itself _around_ page-sizes and cache-sizes, will perform very well compared to structures that cannot.
I've taken to telling people, that if they see me write a long piece, that lacks em-dashes, then they should assume that I am under duress, and send help.
I hate to tell you this, but by saying things like this, you are not much better than the authoritarians/totalitarians you seem to despise.
I do understand where the revulsion comes from, however. My own family, on my mother's side, during WW2, was reduced in size. Four out of every ten were liquidated by the occupying Nazis, and then two out of every ten (or one out of every three, if you adjust the denominator) were liquidated by the triumphant communists (to say nothing of battlefield deaths in-between, and the blanket dispossession that those left alive experienced).
That said, I do think that ideas and ideologies should be evaluated on their own merits, and should not be reactively shunned, because of an atrocity that happened generations ago. The reality, is that good ideas get hijacked and used by opportunists to benefit themselves at the expense of their neighbors. By being so vehemently against an _idea_, you create new opportunities for these kinds of pathological opportunists -- you create an entirely new category of scapegoats, whose pleas and screams and tears you can ignore, while they get dragged away, and separated from their homes and their families and their friends.
I'm not saying that you should cuddle up next to a communist, but I am saying that you should evaluate all people and communities in a nuanced and thoughtful way, befitting a member of _this_ community of absolute oddballs.
I think it is even simpler than you suggest. Cloning an established product is more efficient in terms of both effort (even if costs are subsidized turnaround time is still a measurable physical quantity) and politics (nobody ever got fired for cloning -- if the clone becomes popular, you win, if the clone does not become popular, the West loses).
It is the difference between "safety" and "liveness" (the two kinds of correctness guarantees in computer programming). Communist societies are, at their extreme, "safety" societies: they try to guarantee that nothing bad ever happens. Capitalist/market societies are, at their extreme, "liveness" societies: they try to guarantee that something good/interesting _eventually_ happens (even if bad things have to keep happening).
A "safety" mindset is sympathetic to cloning, because it does not have to deal with much uncertainty. A "liveness" mindset is not sympathetic to cloning, because it has already been done, and profit/monopoly opportunity is minimal.
I feel like LLMs[1] are going to cause a kind of "divorce" between those who love making software and those who love selling software. It was difficult for these two groups to communicate and coordinate before, and now it is _excruciating_. What little mutual tolerance and slack there was, is practically gone.
Open source was always[2] a fragile arrangement based on the kind of trust that involves looking at things through one's fingers (turning a blind eye may be more idiomatic in English), and we are at the point where you just have to either shut your eyes, or otherwise stop pretending that the situation can be salvaged at all.
Just a thought I had: some people think that LLM-shaming is declasse, and maybe it is, but I think that perhaps we _should_ LLM-shame, until the AI-companies train their LLMs to actually give attribution, if nothing else (I mean if it can memorize entire blocks of code, why can't it memorize where it saw that code? Would this not, potentially, _improve_ the attribution-situation, to levels better than even the pre-LLM era? Oh right, because plagiarism might actually be the product).
[1]: Not blaming the tech itself, but rather the people who choose to use it recklessly, and an industry that is based almost entirely on getting mega-corporations to buy startups that, against the odds, have acquired a decent number of happy-ish customers, that can now be relentlessly locked-in and up-sold to.
[2]: I mentioned a specific example of good old fashioned, pre-LLM, human plagiarism here: https://news.ycombinator.com/item?id=46540608
reply