Scikit-learn defaults its logistic regression to ridge regression because the l2 penalty is on by default. So plenty of people are using it without knowing.
scikit-learn.org/stable/modul...
Scikit-learn defaults its logistic regression to ridge regression because the l2 penalty is on by default. So plenty of people are using it without knowing.
scikit-learn.org/stable/modul...
It's a consequence of some EU and UK regulation where people need to be able to opt out of ad targeting. I think the ads option is supposed to be the same as before and the payment is new to allow the opt out. Facebook already have rights to everything you post anyhow.
www.bbc.com/news/article...
Yeah, I feel the same about most of the universes beyond stuff. Tarkir: Dragonstorm and Edge of Eternities were good, though the Avatar the Last Airbender set is really nicely designed even if it's not in universe. The Bloomburrow starters worked fairly well, but I could never win with the otter one
They do have starter kits for some sets which have two 60 card precon decks in them. I picked up the Bloomburrow ones last year to teach my kids, but the power level was a bit odd. They did one for the final fantasy set this summer - magic.wizards.com/en/news/anno...
๐จAnnouncement!๐จ
#Java peers, registration is now open for the 2026 #JavaOne conference (Mar 17-19).
Don't wait, sign up now to reserve your seat โก๏ธ social.ora.cl/60117vBFh
๐จ Time Sensitive ๐จ
#Java peers...the 2026 #JavaOne call for papers closes soon. Get your proposals in by the Nov 11 deadline โก๏ธ
social.ora.cl/6017A7UwL
Is your US shipping still on hold? I ordered a bunch of Faction Paradox books which just arrived (thanks! looking forward to reading them) but I forgot to get the EDAs collection before the tariffs arrived.
๐ขReminder๐ข
#Java peers, #JavaOne is returning next year (Mar 17-19, 2026). PLEASE SIGN UP FOR UPDATES if you're interested in attending, speaking, or sponsoring:
โก๏ธ go.oracle.com/LP=149517
There are countless papers I've reviewed where they say they used a 2GHz Intel processor when reporting performance numbers and I have to ask "which one of the 25 years worth of 2GHz Intel processors did you use?" System environment reporting standards are just terrible.
Many conferences also use openreview.net, so the reviews (for accepted papers after publication) tend to be open as well.
I've only ever published in CS so I can't compare it to elsewhere, but it's typically 20-25% at the top conferences. Not sure about the journals. So much stuff is on ArXiv now too, so lots of things become well known before they finish going through peer review.
The best journal in Machine Learning (jmlr.org) and most of the best conferences are entirely open access and free to publish in. It's been this way for at least 15 years and is working pretty well.
Everyone remembers the first time (and the nth time) they got burned by floating point numbers and this is a great description of exactly how it happens when sampling from a multivariate Gaussian. I'm still sad about a tiny change in Java's Math.exp changing my trained linear model's predictions.
I've derived a bunch of amusement from asking different LLMs about various obscure Doctor Who spinoffs and watching each one straight up lie to me. The 90s novels must not have been in the books they scraped from the internet.
I watched that one last night. Things would have been much better for everyone over the next 6 years if they'd decided to drop her in the gamma quadrant.
Yeah, I have a rough idea what JVM startup looks like, and a similarly rough idea of how processes get launched & what libc does to bring up a C program, but a web browser is far beyond my expertise. Even node.js or CPython are systems I don't understand startup for.
Feels like the difficulty decrease from using a higher level language to implement fizz buzz is not commensurate with the difficulty increase in describing how the language runtime works. It's much easier to describe how a (*nix) C program executes than how Java, Python or JavaScript executes.
A MacBook Pro with 5 stickers on the lid, closed on a table. The stickers say "I'm not hording I'm archiving", "Tribuo", "ONNX Runtime", "Machine Learning", and the final one is the Streetlight Manifesto band logo.
Thanks for the stickers, they are great. We were considering making a similar one with a pile of old Sun servers on, we're the office hardware horders.
#Java developers!
@JavaOne is BACK (18-20 Mar). Early bird registration is available through Feb 9th. Please (re)join us at THE conference that started it all...
Blog โก๏ธ social.ora.cl/6015Qne5z
Event โก๏ธ javaone.com
#JavaOne #OpenJDK #JDK
Good job my new laptop still has a bunch of space for stickers because I need that one.
Is it a transformer decoder? If it's encoder or encoder/decider then an auxillary loss head like BERT has for NSP will train that into the representation. If it's a decoder then that's more of a pain as the place where you do the prediction matters and sticking it on the last token is bad.
Are you including things like Mission to the Unknown to get there, or people who are more Doctor adjacent?
Learn how we used machine learning to develop a static profiler for @graalvm.org Native Image, and achieved a 7.5% improvement in runtime performance:
medium.com/graalvm/mach...
#GraalVM #Java #MachineLearning
Early Wikipedia? Before the tooling and culture built up it was a lot more vulnerable to trolling and misleading edits. I remember my teachers continually saying to check its sources rather than use Wikipedia text directly.
Given how they work getting to zero hallucinations requires incredibly sharp probability distributions over the next token, leading to no variability. So I suspect the hallucination problem isn't going away with generative systems.