Hacker Newsnew | past | comments | ask | show | jobs | submit | ColinWright's commentslogin

Trying to pay a bill. On the website ... it took 24 minutes to navigate to the right place. Then they needed 2FA, so they emailed it to her.

Now she's supposed to open her email while keeping the web page open. It took 5 minutes to do that, find the email, copy down the code, close the email ...

Web site has timed out.

Just one of many examples.


Looking at the front page now and I cannot see it. Do you have a link?

(pause)

Ah, now on the second page ...

Thanks for letting me know.


This is a fascinating experiment! I've just been reading the first few paragraphs of the paper ... easily readable, intended to be accessible by anyone.

In Gauss's time mathematicians would solve problems, publish the solutions in an encrypted form, and then challenge their contemporaries to solve the problems.

Here the authors of a paper on the arXiv say:

"To assess the ability of current AI systems to correctly answer research-level mathematics questions, we share a set of ten math questions which have arisen naturally in the research process of the authors. The questions had not been shared publicly until now; the answers are known to the authors of the questions but will remain encrypted for a short time."

Tao says:

"... the challenge is to see whether 10 research-level problems (that arose in the course of the authors research) are amenable to modern AI tools within a fixed time period (until Feb 13).

"The problems appear to be out of reach of current "one-shot" AI prompts, but were solved by human domain experts, and would presumably a fair fraction would also be solvable by other domain experts equipped with AI tools. They are technical enough that a non-domain-expert would struggle to verify any AI-generated output on these problems, so it seems quite challenging to me to have such a non-expert solve any of these problems, but one could always be surprised."


A slide rule is casually used by the General in "Fantastic Voyage"


Funny you ask ... I lieterally just now pulled out one of my slide rules to keep track of the required run rate for a cricket T20 game.

Fastest and best feedback for whether the batting team is ahead of the rate.


Does it implement Duckworth-Lewis-Stern ;)


No, but a friend of mine made a nomogram to do that.


The title I've chosen here is carefully selected to highlight one of the main points. It comes (lightly edited for length) from this paragraph:

Far more insidious, however, was something else we discovered:

More than two-thirds of these articles failed verification.

That means the article contained a plausible-sounding sentence, cited to a real, relevant-sounding source. But when you read the source it’s cited to, the information on Wikipedia does not exist in that specific source. When a claim fails verification, it’s impossible to tell whether the information is true or not. For most of the articles Pangram flagged as written by GenAI, nearly every cited sentence in the article failed verification.


FWIW, this is a fairly common problem on Wikipedia in political articles, predating AI. I encourage you to give it a try and verify some citations. A lot of them turn out to be more or less bogus.

I'm not saying that AI isn't making it worse, but bad-faith editing is commonplace when it comes to hot-button topics.


Any articles where newspapers are the main source are basically just propaganda. An encyclopaedia should not be in the business of laundering yellow journalism into what is supposed to be a tertiary resource. If they banned this practice, that would immediately deal with this issue.


A blanket dimsissal is a simple way to avoid dealing with complexity, here both in understanding the problem and forming solutions. Obviously not all newspapers are propaganda and at the same time not all can be trusted; not everything in the same newspaper or any other news source is of the same accuracy; nothing is completely trustworthy or completely untrustworthy.

I think accepting that gets us to the starting line. Then we need to apply a lot of critical thought to sometimes difficult judgments.

IMHO quality newspapers do an excellent job - generally better than any other category of source on current affairs, but far from perfect. I remember a recent article for which they intervied over 100 people, got ahold of secret documents, read thousands of pages, consulted experts .... That's not a blog post or Twitter take, or even a HN comment :), but we still need to examine it critically to find the value and the flaws.


> Obviously not all newspapers are propaganda

citation needed


There is literally no source without bias. You just need to consider whether you think a sources biases are reasonable or not


See you should work for a newspaper. You have the gumption.


That's not what I'm saying. I mean citations that aren't citations: a "source" that doesn't discuss the topic at all or makes a different claim.


That is probably 95% of wikipedia articles. Their goal is to create a record of what journalists consider to be true.


Submitted title was "For most flagged articles, nearly every cited sentence failed verification".

I agree, that's interesting, and you've aptly expressed it in your comment here.


People here are claiming that this is true of humans as well. Apart from the fact that bad content can be generated much faster with LLMs, what's your feeling about that criticism? It's there any measure of how many submissions before LLMs make unsubstantiated claims?

Thank you for publishing this work. Very useful reminder to verify sources ourselves!


I have indeed seen that with humans as well, including in conference papers and medical journals. The reference citations in papers is seen by many authors as another section they need to fill to get their articles accepted, not as a natural byproduct of writing an article.


"Never interrupt your enemy when he is making a mistake."

-- Napoleon.


Click on the link to read about this:

    403 ERROR
    The request could not be satisfied.
    Request blocked. We can't connect to the server for this app
    or website at this time.  There might be too much traffic or
    a configuration error.  Try again later, or contact the app
    or website owner.

    If you provide content to customers through CloudFront, you
    can find steps to troubleshoot and help prevent this error by
    reviewing the CloudFront documentation.

    Generated by cloudfront (CloudFront)
    Request ID: IFiQvbhPlrP5MaRdM5km5yAdFAEmvC_IUx2LA899aXly11zm3wAoKg==
Is it the HN "Hug of Death" ?


I doubt HN could knock over an Amazon site. They're probably geoblocking regions where that service never operated.


> I doubt HN could knock over an Amazon site

Not every amazon site is cloud-scale. Niche product like this might be running bare metal under someone's desk


Same right now from Spain. I guess it could be a geoblock, although I expected something more from Amazon


No, apparently eu-west-1 went castors up earlier. I wouldn't be surprised if there was something related to this error.

The site came back around eu-west-1 which, while correlation isn't causation, it does look meaningfully in causation's direction and wiggle an eyebrow suggestively.


One of the talks I give has this in it. The talk includes Continued Fractions and how they can be used to create approximations. That the way to find 355/113 as an excellent approximation to pi, and other similarly excellent approximations.

I also talk about the Continued Fraction algorithm for factorising integers, which is still one of the fastest methods for numbers in a certain range.

Continued Fractions also give what is, to me, one of the nicest proofs that sqrt(2) is irrational.


Thanks! Do you have a version of that talk published anywhere? I tried searching your YouTube channel [1] for a few things like "golden ratio" "ratio", "irrational"... but didn't find anything.

[1] https://www.youtube.com/@colinwright/


That's not my channel. Alas, my name is fairly common, and there are some proper whackos who publish widely.

I only have one video in my channel ... never really got going, but I keep promising myself I'll start doing more. My channel is here:

https://www.youtube.com/@ColinTheMathmo

I don't have that talk on video, but I can probably sketch the content for you if you're interested, and then give pointers to pages with the details.

How to contact me is in my profile ... I'm happy to write a new thing.


Aw thanks, but I think the Mathologer and Numberphile videos are sufficient for me if you haven't already uploaded yours. I don't want to bother you doing extra work for little return!


I honestly should sketch out the talk anyway. I haven't seen anyone else bring together the proof that sqrt(2) is irrational, and the Continued Fraction method of factoring.

Yeah, maybe I'll hack out a sketch tomorrow, show it to a few people, and get them to tell me what's missing so I can flesh it out.

I owe Chalkdust an article anyway.


Yes. We used it for the structures underlying the digital fade algorithm for marine radar images.

It's probably no longer "Commercial In Confidence" ... I should probably write it up sometime.


Could you elaborate please.


Hmm.

This was a long time ago, so we didn't have GPUs or fancy rendering h/ware. We addressed every pixel individually.

So a radar image was painted to the screen, and then the next update was painted on top of that. But that just gives the live radar image ... we wanted moving objects to leave "snail trails".

So what you do for each update is:

* Decrement the existing pixel;

* Update the pixel with the max of the incoming value and the decremented value.

This then leaves stationary targets in place, and anything that's moving leaves a trail behind it so when you look at the screen it's instantly obvious where everything is, and how fast they're moving.

Ideally you'd want to decrement every pixel by one every tenth of a second or so, but that wasn't possible with the h/ware speed we had. So instead we decremented every Nth pixel by D and cycled through the pixels.

But that created stripes, so we needed to access the pixels in a pseudo-random fashion without leaving stripes. The area we were painting was 1024x1024, so what we did was start at the zeroth pixel and step by a prime number size, wrapping around. But what prime number?

We chose a prime close to (2^20)/phi. (Actually we didn't, but that was the starting point for a more complex calculation)

Since phi has no good rational approximation, this didn't leave stripes. It created an evenly spread speckle pattern. The rate of fade was controlled by changing D, and it was very effective.

Worked a treat on our limited hardware (ARM7 on a RiscPC) and easy enough to program directly in ARM assembler.


Thanks for the story.

What's decrementing a pixel ?

I(x,y,t+1) = I(x,y,t) - c ?


Exactly that ... for a given pixel, reducing the existing level/brightness by some value, the default is usually 1, or a fixed percentage.


Ah! Now I understand.

I was stepping out with my wife for a day out and had read your reply very cursorily. That reading had left me quite puzzled -- "I would have done exponentially weighted moving average (EWMA) over time for trails. Why is \phi important here in any form. Is \phi the weight of the EWMA ?".

Now I get it, decrementing the pixels were quite peripheral to the main story.

The main story is that of finding a scan sequence that (a) cycles through a set of points without repetition and (b) without obvious patterns discernible to the eye.

In this, the use \phi is indeed neat. I don't think it would have occurred to me. I would have gone with some shift register sequence with cycle length 1024 * 1024 or a space filling curve on such a grid.

This becomes even more interesting if you include the desiderata that the minimum distance between any two temporally adjacent pixels must not be small (to avoid temporal hot spots).

Finding MiniMax, min over temporal adjacency, max over all 1024* 1024! sequences, might be intractable.

Another interesting formulation could be, that for any fixed kxk sized disc that could be drawn on the grid, the temporal interval between any two "revisit" events need to be independent of the disk's position on the grid.

I think this is the road to small discrepancy sequences of quasi Monte Carlo.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: