Off-and-on trying out an account over at @tal@oleo.cafe due to scraping bots bogging down lemmy.today to the point of near-unusability.

  • 20 Posts
  • 1.01K Comments
Joined 2 years ago
cake
Cake day: October 4th, 2023

help-circle

  • I don’t know what the situation is for commercial games — I don’t know if there’s a marketplace like that — but I do remember someone setting up some repository for free/Creative Commons assets a while back.

    goes looking

    https://opengameart.org/

    It’s not highly-structured in the sense that someone can upload, say, a model in Format X and someone else can upload a patch against that model or something like that with improvements and changes, though. Like, it’s not quite a “GitHub of assets”.

    I haven’t looked at it over time, but I also don’t think that we’ve had an explosion in inter-compatible assets there. Like, it’s not like a community forms around a particular collection of chibi-style sprite artwork at a particular resolution, and then lots of libre games use those assets, the way RPGMaker or something has collections of compatible commercial assets.

    I’m sure that there must be some sort of commercial asset marketplace out there, probably a number, though I don’t know if any span all game asset types or if they permit easily republishing modifications. I know that I’ve occasionally stumbled across a website or two that have individuals sell 3D models.


  • My first question is “why is that the case?”

    Like, is FAIR being (rationally) chosen because people simply cannot afford the private plans on offer, and private plans don’t provide a minimal-enough level of coverage? If so, maybe the problem is actually that we need more availability of housing, that people are financially-stretched too far.

    Or are people irrationally getting too little fire insurance, and FAIR just provides an opportunity to do that? Then you’d think that we should improve the information available about fire insurance plans.

    Or is FAIR providing a better cost-for-value, in which case one would want to look a breakdown of why private plans would cost more — like, is the market not competitive?

    My own gut guess is that the most-likely largest culprit is the first, because I am comfortable saying that California has a very real undersupply of housing, which makes housing highly-unaffordable in California, and causes people to be under greater financial pressure. Like, we’d like to have more housing, which would reduce housing prices, which would permit people to spend less on housing, which would permit people to be less-financially-stretched, which would let people, among other things, spend more on insurance for that housing. I don’t know that that’s the dominant factor, but I’m pretty sure that it is a factor.

    searches for an affordability metric

    https://www.affordabilityindex.org/rankings/states/

    On this metric, California ranks #43 out of 50 states plus DC on housing affordability, measuring what housing costs relative to income. That’s not the bottom of the bin, so it’s likely that one can’t chalk it up only to that, but it isn’t great, and I’d bet that it is a substantial factor.

    takes another look at the metric

    I’d also guess that it’s pretty good odds that the ratio being computed (income to price) is very probably using pre-tax income, and California is exceptionally high in absolute cost of housing among the states, second only to Hawaii. Because we use a progressive income tax system, having higher income means that each additional dollar in income goes less towards making housing affordable as income rises — instead, some of it goes towards effectively subsidizing standard of living in other states that have lower median income. So you’d expect the affordability issues in California to be more-severe than just that ratio suggests; California’s higher income would have less real-world effect than lower housing prices in other states relative to the ratio that the metric uses.

    EDIT: This affordability metric ranks California at #47 on housing affordability out of 50 states plus DC.

    https://www.realtor.com/research/state-report-cards-2025/

    It’s also based on median (I assume pre-tax) income relative to house price.




  • I think that you have two factors here. GDC isn’t specific to PC gaming, and additionally, a lot of titles will see both PC and console releases.

    For a game that is intended to see only a PC release, my guess is that that that might affect system requirements of the game.

    For games that see console releases, things like “will fewer people have consoles” — because current-gen consoles are very unlikely to change spec, just price, is how this manifests itself. “Is the Playstation 6 going to be postponed” is a big deal if you were going to release a game for that hardware.





  • As it currently exists on other platforms, Gaming Copilot lets you ask guide-like questions about the game you’re currently playing. Microsoft’s official site offers an example question like “Can you remind me what materials I need to craft a sword in Minecraft?”

    I haven’t used consoles for a few generations, but historically, switching between a game and a Web browser on a console wasn’t all that great, and text entry wasn’t all that great. I dunno if things have improved, but it was definitely a pain in the neck to refer to a website in-game historically.

    On Linux, Wayland, I swap between fullscreen desktops when playing games, and often have a Web browser with information relevant to the game on another desktop. If it helps enable some approximation of a workflow like that for console players, that doesn’t sound unreasonable.

    There are other objections I’d have, like not really wanting someone logging what my voice sounds like or giving Microsoft even more data on me to profile with via my searches. But it sounds to me like the basic functionality has a point.



  • What makes this worse is that git servers are the most pathologically vulnerable to the onslaught of doom from modern internet scrapers because remember, they click on every link on every page.

    The especially disappointing thing is that, for the specific case that Xe was running into, a better-written scraper could just recognize that this is a public git repository and just git clone the thing and get all the useful code without the overhead. Like, it’s not even “this scraper is scraping data that I don’t want it to have”, but “this scraper is too dumb to just scrape the thing efficiently and is blowing both the scraper’s resources and the server’s resources downloading innumerable redundant copies of the data”.

    It’s probably just as well, since the protection is relevant for other websites, and he probably wouldn’t have done it if he hadn’t been getting his git repo hammered, but…

    EDIT: Plus, I bet that the scraper was requesting a ton of files at once from the server, since he said that it was unusable. Like, you have a zillion servers to parallelize requests over. You could write a scraper that requested one file at once per server, which is common courtesy, and you’re still going to be bandwidth constrained if you’re schlorping up the whole Internet. Xe probably wouldn’t have even noticed.


  • https://en.wikipedia.org/wiki/National_Helium_Reserve

    The National Helium Reserve, also known as the Federal Helium Reserve, was a strategic reserve of the United States, which once held over 1 billion cubic meters (about 170,000,000 kg)[a] of helium gas.

    The Bureau of Land Management (BLM) transferred the reserve to the General Services Administration (GSA) as surplus property, but a 2022 auction[10] failed to finalize a sale.[11] On June 22, 2023, the GSA announced a new auction of the facilities and remaining helium.[12] The auction of the last helium assets was due to take place in November, 2023.[13] Though the last of the Cliffside reserve was to be sold by November 2023, more natural gas was discovered at the site than was previously known, and the Bureau of Land Management extended the auction to January 25, 2024 to allow for increased bids.[14] In 2024 the remaining reserve was sold to the highest bidder, Messer Group.[15]

    Arguably not the best timing on that.


  • Sure. What that guy is using is actually not the most-interesting diagram style, IMHO, for automatic layout of network maps, if you want large-scale stuff, which is where the automatic layout gets more interesting. I have some scripts floating around somewhere that will generate very large network maps — run a bunch of traceroutes, geolocate IPs, dump the results into an sqlite database, and then generate an automatically laid-out Internet network map. I don’t want to go to the trouble of anonymizing the addresses and locations right now, but if you have a graphviz graph and want to try playing with it, I used:

    goes looking

    Ugh, it’s Python 2, a decade-and-a-half old, and never got ported to Python 3. Lemme gin up an example for the non-hierarchical graphviz stuff:

    graph.dot:

    graph foo {
        a--b
        a--d
        b--c
        d--e
        c--e
        e--f
        b--d
    }
    

    Processed with:

    $ sfdp -Goverlap=prism -Gsep=+5 -Gesep=+4 -Gremincross -Gpack -Gsplines=true -Tpdf -o graph.pdf graph.dot
    

    Generates something like this:

    That’ll take a ton of graphviz edges and nicely lay them out while trying to avoid crossing edges and stuff, in a non-hierarchical map. Get more complicated maps that it can’t use direct lines on, it’ll use splines to curve lines around nodes. You can create massive network maps like this. Note that I was last looking at graphviz’s automated layout stuff about 15 years ago, so it’s possible that they have better layout algorithms now, but this can deal with enormous numbers of nodes and will do reasonable things with them.

    I just grabbed his example because it was the first graphviz network map example that came up on a Web search.


  • tal@lemmy.todaytoTechnology@lemmy.worldDigg Shut Down Again
    link
    fedilink
    English
    arrow-up
    9
    ·
    edit-2
    2 days ago

    We faced an unprecedented bot problem

    When the Digg beta launched, we immediately noticed posts from SEO spammers noting that Digg still carried meaningful Google link authority. Within hours, we got a taste of what we’d only heard rumors about. The internet is now populated, in meaningful part, by sophisticated AI agents and automated accounts. We knew bots were part of the landscape, but we didn’t appreciate the scale, sophistication, or speed at which they’d find us. We banned tens of thousands of accounts. We deployed internal tooling and industry-standard external vendors. None of it was enough. When you can’t trust that the votes, the comments, and the engagement you’re seeing are real, you’ve lost the foundation a community platform is built on.

    This isn’t just a Digg problem. It’s an internet problem. But it hit us harder because trust is the product.

    It’s a social media problem. It’s going to be hard to provide pseudonymity, low-cost accounts relatively freely, and counter bots spamming the system to manipulate it. The model worked well in an era before there were very human-like bots that were easy to produce.

    It might be possible to build webs of trust with pseudonyms. You can make a new pseudonym, but the influence and visibility gets tied to, for example, what users or curators that you trust trust, so the pseudonym has less weight until it acquires reputation. I do not think that a single global trust “score” will work, because you can always have bot webs of trust.

    Unfortunately, the tools to unmask pseudonyms are also getting better, and throwing away pseudonyms occasionally or using more of them is one of the reasonable counters to unmasking, and that doesn’t play well with relying more on reputation.







  • You have all your devices attached to a console server with a serial port console set up on the serial port, and if they support accessing the BIOS via a serial console, that enabled so that you can access that remotely, right? Either a dedicated hardware console server, or some server on your network with a multiport serial card or a USB to multiport serial adapter or something like that, right? So that if networking fails on one of those other devices, you can fire up minicom or similar on the serial console server and get into the device and fix whatever’s broken?

    Oh, you don’t. Well, that’s probably okay. I mean, you probably won’t lose networking on those devices.