There's an almost inhuman amount of mind over matter psychology when it comes to endurance running. Unless you can duplicate reality multiple times and swap out the shoes without anyone knowing to do properly scientific testing, we can't know for sure what did it. (The shoes probably helped.)
You don't need to be competing on the world stage to enjoy some of the benefits of Alpha flys or those pumas. 500 for the new Adidas does seem a little silly though.
While the foam may last longer than older EVA foam shoes, the outsoles of the shoes have gotten ridiculously thin these days.
The continental rubber outsole on these Adidas Adios Pro EVO 3 shoes are so thin (less than two sheets of paper, I think), that they don't even appear in side/profile views of the shoes. The outsole doesn't even extend the length of the entire shoe, it stops around the middle of the shoe. So heel strikers aren't welcome and will have loads of fun in wet weather. see https://www.adidas.com/us/adizero-adios-pro-evo-3/KH7678.htm...
In general, these high stack, forward-leaning shoes are meant for going straight ahead - imagine ladies' high heel shoes with an inch and a half of foam on the bottom - any sharp turns will force the runner to slow down or they'll twist their ankles. Looking at the London Marathon course, https://www.londonmarathonevents.co.uk/london-marathon/cours..., there's about twenty ninety-degree or sharper turns.
I usually don't mind, but tend to split reviews into two types. Either I understand the context and can quickly do an in depth review, or I have to take some time to actually learn about the code by reviewing the surrounding systems, experimenting with it, etc. But in both cases I would at least run the code and verify correctness.
I think it becomes a chore when there are too many trivial mistakes, and you feel like your time would have been better spent writing it yourself. As models and agent frameworks improve I see this happening less and less.
I drag a tiny fetch wrapper around with error/json handling, timeouts and basic interceptor support. It doesn't cover everything axios does but it's nice enough and I haven't had to touch it in a couple years.
I often do similar... though most of the time the past couple years, I'm generating the client from OpenAPI integration on the backend that uses fetch as its' base.
You might be surprised, there are plenty of low effort attacks out there that just install a crypto miner and phone home periodically without doing much to cover it up.
That's interesting, I actively use both and usually find it to be a toss up which one performs better at a given task. I generally find Claude to be better with complex tool calls and Codex to be better at reviewing code, but otherwise don't see a significant difference.
If you want to find an advocate for Codex that can give a pretty good answer as to why they think it's better, go ask Eric Provencher. He develops https://repoprompt.com/. He spends a lot of time thinking in this space and prefers Codex over Claude, though I haven't checked recently to see if he still has that opinion. He's pretty reachable on Discord if you poke around a bit.
Quite irrelevant what factions think. This or that model may be superior for these and those use cases today, and things will flip next week.
Also. RLHF mean that models spit out according to certain human preference, so it depends what set of humans and in what mood they've been when providing the feedback.
On the contrary, I very much care about what the other factions think because I want to know if things have already flipped and the easiest way to do so is just ask someone who's been using the tool. Of course the correct thing to do is to set up some simple evals, but there is a subjective aspect to these tools that I think hearing boots on the ground anecdata helps with.
Haven't done it in a while, but I've done some tasks with both Codex and Claude to compare. In all cases I asked both to put their analysis and plans for implementation into a .md file. Then I asked the other agent to analyze said file for comparison.
In general, Claude was impressed by what Codex produced and noted the parts where it (i.e. Claude) had missed something vs. Codex "thinking of it".
From a "daily driver" perspective I still use Claude all the time as it has plan mode, which means I can guarantee that it won't break out and just do stuff without me wanting it to. With Codex I have to always specify "Don't implement/change, just tell me" and even then it sometimes "breaks out" and just does stuff. Not usually when I start out and just ask it to plan. But after we've started implementation and I review, a simple question of "Why did you do X?" will turn into a huge refactoring instead of just answering my question.
To be fair, that's what most devs do too (at least at first), when you ask them "Why did you do X" questions. They just assume that you are trying to formulate a "Do Y instead of X" as a question, when really you just don't understand their reasoning but there really might be a good reason for doing X. But I guess LLMs aren't sure of themselves, so any questioning of their reasoning obliterates their ego and just turns them into submissive code monkeys (or rather: exposes them as such) vs. being software engineers that do things for actual reasons (whether you agree with them or not).
For that I'm not so sure. I tried both early 2025 and was disappointed in their ability to deal with a TCA based app (iOS) and Jetpack compose stuff on Android, but I assume Opus 4.6 and GPT 5.4 are much better.
Addictive patterns in games and other online activity is a bit less innocent than you are portraying it: knowingly causing harm is too low a standard. A lot of the profitability of online games, prediction markets, etc. comes from the whales. The whales are probably addicted. If your business is a whale hunt you are possibly causing harm at least to the extent that addiction is dangerous.
They'd find another method. Why are we allowing this in the first place?
I don't have an answer to fix this whole mess, but it starts with our attitude towards addiction. We've built a system that rewards addiction in all sorts of places. Granted, every addiction is different, and I'm of the opinion that it's not (drug = bad), it's how you use it and react to it. We can control the latter, but we choose to ignore it because we're too busy with anything else. This is a tale as old as time...
For me, it feels like you could cut this problem down substantially by eliminating section 230 protection on any algorithmically elevated content. Everywhere. Full stop.
If you write or have an algorithm created that pushes content to users, in ANY fashion, that is endorsement. You want that content to be seen, for whatever odd reason, and if it's harmful to your users, you should be held responsible for it. It's one thing if some random asshole messages me on Telegram trying to scam me; there's little Telegram can do (though a fucking "do not permit messages from people not in my contacts" setting would be nice) but there is nothing at all that "makes" Facebook shovel AI bullshit at people, apart from it juices engagement, either by genuine engagement or ironic/ragebaiting.
And AI bullshit is just annoying, I've seen "Facebook help" groups that are clearly just trawling to get people's account info, I've seen scam pages and products, all kinds of shit, and either it pisses people off so Facebook passes it around, or they give Facebook money and Facebook shoves it into the feeds of everyone they can.
It's fucking disgusting and there's no reason to permit it.
I don't see a good way to make a definite legal distinction between the icky stuff versus normal an unobjectionable things which are, technically, also forms of elevation-by-algorithm:
rank_by_age(items) // Good
rank_by_age_and_poster_reputation(items) // Probably
rank_by_on_topic_ness(items, forum_subject)
rank_by_likes(items)
rank_by_engagement_likelihood(items) // Bad?
rank_by_positive_sentiment_toward_clients(items) // Bad
rank_by_age(items) // Good
rank_by_age_and_poster_reputation(items) // Probably
rank_by_on_topic_ness(items, forum_subject)
rank_by_likes(items)
<-- here -->
rank_by_engagement_likelihood(items) // Bad?
rank_by_positive_sentiment_toward_clients(items) // Bad
Age is deterministic. When was the thing posted?
Poster reputation is deterministic. How many times has this poster received positive feedback based on their content?
On-topic-ness is deterministic, if a bit fuzzy. That said I think the likes will reflect this, if you post a thread about cooking potatoes in the gopro subreddit, your post will be downvoted and probably removed via other means in which case it's presence in the feed is already null.
Likes are again, deterministic. How many people upvoted it?
In contrast:
Engagement likelihood is clearly a subjective, theoretical measure. An algorithm is going to parse a database for other posts like this, see how much attention it got, and say "is this likely to drive engagement." That's what I'm talking about.
And positive sentiment towards clients I can't quite read? I'm guessing you're referring to like, community sponsors but I'm not 100% certain. But that almost certainly is a subjective one too, and even if not, it's giving people with money the ability to put their thumb on the scale.
I don't think "deterministic" is the right term to capture this concept. An if-statement which bans posts containing a political phrase would be 100% deterministic, or one which prioritizes anything from a username on a list.
> On-topic-ness is deterministic, if a bit fuzzy.
If you permit that exception (even for good reasons) then it reveals how the original "algorithmic elevation" is too vague and unenforceable.
All someone needs is a ToS footnote like "this forum is provided for truthful international news and engaging with $COMPANY in a positive way." Poof, loophole. Anything the moderator (or moderator-algorithm) decides is "untrue" or "negative" becomes off-topic and can be pushed down.
Eliminating section 230 protections would heavily disfavor any kind of intellectually stimulating content, because it's hard for a platform to scalably verify that nobody's making defamatory claims. But pointless clickbait, heavily filtered Instagram models, etc. don't really have liability concerns on a video-by-video level. To me it seems like this makes the problem worse.
It’s not eliminating section 230 entirely, it’s eliminating it for algorithmically promoted content. If you have a site that has user content and you present that content in a neutral fashion, section 230 applies. If you pick and choose what content to present to users (manually or by algorithm), you’re no longer a neutral platform, and shouldn’t be getting the benefit of 230.
I understand that. My point is that this would mean algorithmic feeds can only contain vapid, pointless content with no liability concerns. To me, it doesn't improve the world to require that Instagram and Youtube exclusively serve slop, even if that might cause some number of people to abandon them for non-algorithmic platforms with better content.
Literally every social media site I'm aware of has had, in varying strengths and at varying times, many still currently, a movement among users asking for a fucking chronological ordered feed. Just, what the fuck my friends are saying, in the reverse order that they said it, displayed in a list.
Not only is this seemingly the most desired feed among end users, it was also the default one. MySpace didn't have a choice in the matter, they had to show a chronological timeline, because they didn't have a machine-learning algorithm nor a way to make one. They could tweak it based on engagement metrics but on the whole, it was just here's what all your friends have posted, in reverse order, scroll away. And then eventually you'd hit the end where it's like "you're up to date" and then you go on with your fucking day.
But of course platforms hate that. They want you there, all day, scrolling through an infinite deluge of bullshit, amongst which they can park ads. And we know they hate this, because not only have platforms refused to bring back chronological feeds, they actively removed them if they existed at one time. Not only is this doable, it's the most efficient way that requires the least compute from their servers, but platforms reliably chose the inverse... because it makes them more money.
Also specifically on this:
> My point is that this would mean algorithmic feeds can only contain vapid, pointless content
The vast majority of these sites is vapid, pointless content RIGHT NOW, even if it attempts to convince you it isn't.
Literally every social media site I'm aware of has a chronological ordered feed of people you've chosen to follow. Facebook does, Instagram does, Youtube does. It's just not the homepage, and most people don't care enough about what feed they get to go navigate to it every time they open the app. Would it be nice to make them let you put it on the homepage? Sure, I'd support that.
The current state of affairs is that Youtube and Instagram have brought back fascism and the measles, so if the complaint here is "it's impossible to moderate algorithmic content at scale and so the platforms would become incredibly risk averse," I think I'd take that alternative. I also don't think effectively forcing a breakup of the current online media monopolies is a bad thing either - if you can't actually mitigate the damage of your platform because you're too big, then maybe you shouldn't be that big.
> If you write or have an algorithm created that pushes content to users, in ANY fashion, that is endorsement
Yes. People make free speech arguments about this, but the list and order of stuff returned by algorithmic non-directed (+) lists is clearly a form of endorsement. Even more so is advertising, which undergoes a bidding process. Pages which show ads should be liable if those ads are fraudulent, especially if they're so obviously fraudulent that casual readers suspect them immediately.
(+) Returning a list of stuff in a user-specified query, on the other hand, is not endorsement. Chronological or alphabetical order or distance-based or even random is fine.
Note that section 230 is, of course, US specific and other countries manage without it.
For context, facebook is so dystopian when I login once every few years that I’m not sure I’ll ever use it again. And, I hate wading through the YouTube cesspool to find some educational content I like. But, I don’t think it makes sense to ban a/b testing or optimization in general. Some company could use it, for example, to figure out how to teach math to kids in a way that’s as engaging as possible. This would be “more addictive” technically.
That's a good point, I'm not 100% sure it's worth throwing away the potentially beneficial uses. There might not be a solution that's both feasible to implement and avoids banning useful things. In the end I usually come back to it being the parent's responsibility to monitor usage, limit screen time, etc., but it hasn't been working so well in practice.
Not enough to diffuse liability. 15 years ago when recommender algorithms were the new hotness, I saw every single group of students introduced to the idea immediately grasp the implication that the endgame would involve pandering to base instincts. If someone didn't understand this, it's because
> It is difficult to get a man to understand something, when his salary depends on his not understanding it. - Upton Sinclair
reply