Hacker Newsnew | past | comments | ask | show | jobs | submit | fourier456's commentslogin

I think you have a point. The paper has load bearing reliance on other preprints. I think soon we see a workflow where AI (ChatGPT) can identify precise transitions in the argument that do not require full formalization to falsify. Link - https://chatgpt.com/share/693cc655-ca94-800c-870a-a5c78fb10d...


I did the same thing (roughly) with Gemini 3.0 Pro.

Prompt: "Analyze the mathematics in the paper. Find mistakes, note instabilities (if present) and generally criticize as needed."

https://gemini.google.com/share/6860fdaea334

The mathematics here are far beyond me, but it's interesting that Gemini more or less concurs with chatgpt with respect to "load bearing reliance on other preprints".

Gemini's summary (there's much more in the link above that builds up to this):

The mathematics is largely robust but relies on heavy "black box" theorems (Iritani's blowup formula, Givental's equivariant mirror symmetry) to bypass difficult geometric identifications in the non-archimedean setting. The primary instability is the lack of a non-archimedean Riemann-Hilbert correspondence, which limits the "enhanced" atoms theory. The core results on cubic fourfolds hold because the specific spectral decomposition (eigenvalues $0, 9, 9\zeta, 9\zeta^2$) is robust and distinct22, but the framework's extension to finer invariants (integral structures) is currently obstructed.


Okay, wait though like I want to know the full transcript because that actually is a better / softer benchmark if you measure in terms of the necessary human input.


Tbf you can always ask an LLM to explain it /s


yea dating apps can be useful for other groups, lgbt, finding people with cultural connections, etc.


This just isn’t really a good book. One sign; it’s an arXiv book.


Rhetorical question, what do you think?


This comment seems misguided. I am a grad student working with LLMs in a CS department. I promise you CS has (and will have) much to say about theoretical underpinnings of transformers.


I can't see what you might be referring to. Even topics around quantization are fundamentally about information theory and not really computer science per se. It could just as easily be an explanation that your CS department is better-funded and more interested in tackling those problems -- statistics departments tend to be a bit behind the curve in this regard, but that doesn't magically move an entire branch of science somewhere else in the taxonomy.

Maybe if you can explain what you mean instead of merely providing some vague assertion, we could have a real conversation about it. In my experience, though, the hype train around software engineering has incited CS departments to retcon a lot of things as CS-related so that their grant writers have an easier time securing funding.


I can’t wait for like 2-3 years from now when we can explore this with vision-GPTs trained with different languages.


How does your strategy work? Perhaps distinguish it from naive approaches.


Example:

Is the pytorch convention that one or zero should be used for mask values you do not want to attend to?

ChatGPT will give a correct answer and sample implementation, if the implementation is broken or uses a non resistant api then I just tell it what's wrong and ask it if it knows better. Getting to the correct answer has proven vastly more efficient than wading through google results.

Alternatively if I'm asking a conceptual question, chatgpt will give a few directions, I can ask to dive deeper on one, and then I ask it for a citation or google search terms to confirm the result. If I fail to confirm, I'll tell it that and see if it corrects.

If I have some code to write - and I'm feeling lazy. I can just tell chatgpt my requirements and ask it if it has any questions. I keep clarifying the questions until it says it can write the code ( which it usually does with 95% correctness for non trivial asks) - if you tell it where it made mistakes it will usually correct, and if not porting the code to a working state just means changing a few function calls which chat gpt hallucinated into existence.


This annoys me. If you understand the last two words, it’s very likely you understand the first. Just like OP above.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: