2
points
lopuhin
Joined 153 karma
I work at https://www.zyte.com/ on https://www.zyte.com/automatic-extraction/
Github: https://github.com/lopuhin/ Kaggle: https://www.kaggle.com/lopuhin/
- lopuhinContext window size of 400k is not new, gpt-5, 5.1, 5-mini, etc. have the same. But they do claim they improved long context performance which if true would be great.
- you can rent them for less then $2/h in a lot of places (maybe not in the drawer)
- I find OpenAI's new flex processing more attractive, as it has the same 50% discount, but allows to use the same API as regular chat mode, so you can still do stuff where Batch API won't work (e.g. evaluating agents), and in practice I found it to work well enough when paired with client-side request caching: https://platform.openai.com/docs/guides/flex-processing?api-...
- it's pretty difficult to package native python dependencies for wasmtime or other wasi runtimes, e.g. lxml
- Crazy amount of breakage...
Here is a PR which reverts this: https://github.com/pypa/setuptools/pull/4911
Interesting that maintainers of setuptools still only postpone the depreciation date for a year, so we can probably expect more issues like this in the future.
- Congrats on the launch! How much does it cost? And what is the sandboxing technology?
- I find it strange that the author is really happy with the quality of string comparison here https://pgaleone.eu/ai/coding/2025/01/26/using-ai-for-coding... and while it would kind of work, it's a very weird piece of code from ML standpoint, e.g. it's training a TF-IDF vectorizer on just two strings being compared, which at best won't change anything (unless the same word is repeated within one product), and is a super weird thing to do as for better quality you'd probably want to train that on some corpus, or not bother at all. And also it compare two strings as bags of words, which again is not the end of the world but maybe not what the author wants here, and if they want this then it's not the easiest way of doing it. So it's taking some things which can be useful when comparing texts (tf-idf and cosine similarity) but then applying them in a weird way which does not let them show their strengths.
- It's a 600B+ mixture of experts and yes it's described in the paper, GitHub, etc.
- Why is this doubtful, did you spot any suspicious things in their paper? They make the weights and a lot of training details open as well, which leaves much less room for making stuff up, e.g. you could check training compute requirements from active weight size (which they can't fake as they released the weights) and fp8 training used.
- With distilled models being released, it's very likely they'd be soon served by other providers at a good price and perf, unlike the full R1 which is very big and much harder to serve efficiently.
- I don't think so, what they show on CS video is exactly the Dust2 map, not just something similar/inspired by it.
- I think GraalPython does have a GIL, see https://github.com/oracle/graalpython/blob/master/docs/contr... - and if by "there is no such thing on those platforms" you mean JVM/CLR not having a GIL, C also does not have a GIL but CPython does.
- Curious which model was used? Sorry if I missed that. Looks like an important detail to mention when doing an evaluation.
- Also I don't think you can use NIM packages in production without a subscription, and I wasn't able to find the cost without signing up. Also NIM package for Mistral Nemo is not yet available anyways.
- The README says they plan to add llama.cpp support which should cover a lot of targets, also they have tinygrad already integrated I think.
- Not quite the same, OpenAI was initially quite open, while Ilia is currently very explicitly against opening or open-sourcing research, e.g. see https://www.theverge.com/2023/3/15/23640180/openai-gpt-4-lau...
- yes looks like a bug in the example to me, feel free to report to https://github.com/fchollet/ARC-AGI/issues :)
- Really nice to see no-GIL Python become closer and closer to reality.
- I think they mean "chemical composition", e.g. how much hydrogen / helium / other elements there are in the star, deduced from their spectra