Hacker Newsnew | past | comments | ask | show | jobs | submit | fnord123's commentslogin

I don't mind the AI features per se, but is there a configuration setting to sent the traffic through a local AI Gateway to prevent the AI from receiving private information? At the very least to track what is sent over the wire.


Found it: browser.ml.chat.provider

On thing that troubles me is that code reviews are also an educational moment for seniors teaching juniors as well as an opportunity for people who know a system to point out otherwise undocumented constraints of the system. If people slack on reviews with the agent it means these other externalities suffer.

Are being handling this at all? Is it no longer needed because it gets rolled into AGENTS.md?


I find working with Ai a lot like working with a junior employee... with the junior employee they learn and get better (skill level and at dealing with me) but with Ai the mentoring lessons reset once you type /clear

Skills are a positive development for task preferences, agents.md for high level context, but a lot of the time its just easier to do things the way your Ai wants.


> educational moment for seniors teaching juniors

You see, this is no longer necessary - companies are firing all the non-seniors, are not hiring any juniors, and delegating everything to AI. This is the future apparently!


> the AA-Omniscience Hallucination Rate Benchmark which puts 3.0 Pro among the higher hallucinating models. 3.1 seems to be a noticeable improvement though.

As sibling comment says, AA-Omniscience Hallucination Rate Benchmark puts Gemini 3.0 as the best performing aside from Gemini 3.1 preview.

https://artificialanalysis.ai/evaluations/omniscience


You are misreading the benchmark.

https://artificialanalysis.ai/#aa-omniscience-hallucination-...

If you look at the results 3.0 hallucinates an awful lot, when it's wrong.

It's just not wrong that often.

(And it looks like 3.1 does better on both fronts)


Be careful. You get good at what your practice.


> Others disallow even empty bottles at security screening

I haven't encountered this. Could you name some?


> Stop citing single studies as definitive. They are not. Check if the ones you are reading or citing have been replicated.

And from the comments:

> From my experience in social science, including some experience in managment studies specifically, researchers regularly belief things – and will even give policy advice based on those beliefs – that have not even been seriously tested, or have straight up been refuted.

Sometimes people use fewer than one non replicatable studies. They invent studies and use that! An example is the "Harvard Goal Study" that is often trotted out at self-review time at companies. The supposed study suggests that people who write down their goals are more likely to achieve them than people who do not. However, Harvard itself cannot find such a study existing:

https://ask.library.harvard.edu/faq/82314


Definitely ignore single studies, no matter how prestigious the journal or numerous the citations.

Straight-up replications are rare, but if a finding is real, other PIs will partially replicate and build upon it, typically as a smaller step in a related study. (E.g., a new finding about memory comes out, my field is emotion, I might do a new study looking at how emotion and your memory finding interact.)

If the effect is replicable, it will end up used in other studies (subject to randomness and the file drawer effect, anyway). But if an effect is rarely mentioned in the literature afterwards...run far, FAR away, and don't base your research off it.

A good advisor will be able to warn you off lost causes like this.


Check out the “Jick Study,” mentioned in Dopesick.

https://en.wikipedia.org/wiki/Addiction_Rare_in_Patients_Tre...


> America is sick. The data is clear.

Can we inform dictionaries and encyclopaedia that data is now a mass noun and it is considered archaic to use data as a plural of datum?


OED records the first usage as a mass noun in 1702, I think the ship's already sailed there


Which scheme implementation? Guile?


All of them.


To elaborate, the scheme spec requires tco.


Which scheme is embeddable and lightweight?

And what does lightweight mean? Does it mean low memory footprint or does it mean few-lines-of-code-to-introduce or does it mean zero-dependencies?


https://en.wikipedia.org/wiki/SIOD is lighter weight than almost anything. Way smaller than Lua.

Also look at Hedgehog Lisp. The bytecode compiler (runs on a PC) is separate from the interpreter, i.e. there is no REPL. But it means that the interpreter is only about 20KB of code. It's quite practical. It's not Scheme but rather is a functional Lisp (immutable data including AVL trees as the main lookup structure) and it is tail recursive. https://github.com/sbp/hedgehog


I meant which scheme implementation is "lightweight" and also meant to ask what "lightweight means".


For a functionnal language, TCO is really a must have. How would you do the equivalent of loops without it ?


For a purely functional language. Scheme is not that.

https://www.gnu.org/software/guile/manual/html_node/while-do...


And you are wrong because you can define a loop as iterating recursively over a list with just car and cdr.


Can but since other constructs exist that doesn't by itself make TCO "must have"


I wasn't asking which scheme has TCO.

I was asking which scheme if the 20-50 of them was "lightweight" and embeddable.


This is why TFA used Segway as an example.


Indeed. The complement of No Tux No Bux is If Tux Then Bux.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: