Hacker Newsnew | past | comments | ask | show | jobs | submit | aziis98's commentslogin

I would actually merge html and js in a single language and bring the layout part of css too (something like having grid and flexbox be elements themselves instead of display styles, more typst kind of showed this is possible in a nice way) and keep css only for the styling part.

Or maybe just make it all a single lispy language


I know this is a bit out of scope for these image editing models but I always try this experiment [1] of drawing a "random" triangle and then doing some geometric construction and they mess up in very funny ways. These models can't "see" very well. I think [2] is still very relevant.

[1]: https://chatgpt.com/share/6941c96c-c160-8005-bea6-c809e58591...

[2]: https://vlmsareblind.github.io/


> Pointing capability: Gemini 3 has the ability to point at specific locations in images by outputting pixel-precise coordinates. Sequences of 2D points can be strung together to perform complex tasks, such as estimating human poses or reflecting trajectories over time

Does somebody know how to correctly prompt the model for these tasks or even better provide some docs? The pictures with the pretty markers are appreciated but that section is a bit vague and without references


For my CMS I’d love to get an AI to nicely frame a picture in certain aspect ratios. Like of I provide an image, give me coordinates for a widescreen, square, portrait, and 4x3 using a photographers eye.

Any model that can do that? I tried looking in huggingface but didn’t quite see anything.


Simon Wilson has some good blogs on this: https://simonwillison.net/2024/Aug/26/gemini-bounding-box-vi...


This is a small TailwindCSS alternative based on a css template literal. I was inspired by styled-components and EmotionCSS, which however do not work well with ViteJS and specifically Preact.

I wanted something more optimized than other CSS-in-JS alternatives that generate CSS at runtime, so I created a ViteJS plugin. It extracts all style snippets, replaces them with classes like css-a1b2c3, and injects all the corresponding styles into a CSS file in place of an @extracted-css directive.

This provides a better experience than Tailwind, as you can use all CSS language features without learning new conventions while maintaining a per-component styling approach. This turns out to be more inspectable in the browser's dev-tools, as snippets are extracted as-is and are not fragmented across thousands of small classes.

There is also a preact options hook that adds a custom classList attribute, which maps to clsx for easy class composition (similarly to VueJS, Svelte, etc.).

P.S. I know other frameworks exist, but I have really been enjoying using Preact for frontend development recently.


It's not clear to me what problem this is trying to solve

Edit: Ok so this is just a dag cell evaluator on top of jsonata [1] (that looks like to be a small data processing language to evaluate expressions)

[1]: https://github.com/jsonata-js/jsonata


It appears to by cyclic.


I didn't know about https://e2b.dev/ but I was looking for something exactly like that. Does anyone know about any self hostable alternatives?


You should look at [Modal](https://modal.com/), not affiliated.


Lovable runs on Modal Sandboxes.


It appears that e2b runs Firecracker microVMs (https://e2b.dev/blog/how-manus-uses-e2b-to-provide-agents-wi...)

It shouldn't be too hard to get a Firecracker orchestrator running locally - the articles here were very helpful when I was doing this myself: https://jvns.ca/blog/2021/01/23/firecracker--start-a-vm-in-l...


Can we run an agent in Github CI? I create a Github issue with my requirements, the CI runner runs Aider scripting mode (it connects to say, OpenAI api) which looks at the comments, maintains a branch, runs tests, takes browser screenshots and then gives its replies.

https://aider.chat/docs/scripting.html


you can host e2b: https://github.com/e2b-dev/infra/blob/main/self-host.md

we host it for https://www.definite.app/. You'd need pretty heavy usage to beat e2b's pricing.


Played with https://www.freestyle.sh this weekend, really solid


You can check out https://daytonaio-ai.framer.website/, it’s also could be self-hosted


I think this article can be interesting:

https://www.seangoedecke.com/inference-batching-and-deepseek...

Here is an example of what happens

> The only way to do fast inference here is to pipeline those layers by having one GPU handle the first ten layers, another handle the next ten, and so on. Otherwise you just won’t be able to fit all the weights in a single GPU’s memory, so you’ll spend a ton of time swapping weights in and out of memory and it’ll end up being really slow. During inference, each token (typically in a “micro batch” of a few tens of tokens each) passes sequentially through that pipeline of GPUs


I'm just can't wait for a globally scaled rag system. I think that will be a turning point for search engines.

For now there is only https://exa.ai/ that is currently doing something similar it seems.


I successfully converted a typst report to md/mdx last week using this technique. For complex layout primitives I just told the llm to write a comment with a warning todo of the missing part it wasn't able to convert


Yeah I really like Astro too, I often just start with the minimal template [1] that is just a couple of files. I used to do many experiments with custom SSG some time ago but since Astro came out I can't change back anymore.

If you don't want js on the frontend you can just use it as a nice html templating engine. It also renders markdown automatically for you as it recognizes different formats by the file extension.

[1]: https://github.com/withastro/astro/tree/main/examples/minima...


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: