Hacker Newsnew | past | comments | ask | show | jobs | submit | LeonidBugaev's commentslogin

It does not implement the Auth :)

(mcp auth is terrible btw)


I couldn’t find any great examples of MCP auth, so made this demonstrate an oauth flow recently - https://github.com/OBannon37/chatgpt-deep-research-connector...


For my app I'm bypassing MCP auth and doing the regular oauth2 flow to connect users to external apps.

Then I pass the stored oauth token directly to my (private) MCP servers alongside a bearer token.


You should check https://probeai.dev/ too. Thats one of those building blocks which makes AI trully understand the code.


To put it simple:

A2A is for communication between the agents. MCP is how agent communicate with its tools.

Important aspect of A2A, is that it has a notion of tasks, task rediness, and etc. E.g. you can give it a task and expect completely in few days, and get notified via webhook or polling it.

For the end users for sure A2A will cause a big confusing, and can replace a lot of current MCP usage.


If an agent could wrap itself in an MCP server, would that make A2A redundant?


I have the same problem come out in my mind.

What if I wrap the agent as a tool in MCP?

Since the agents I got from the 'A2A' protocol is passed as tools to another Agent...

https://github.com/google/A2A/blob/72a70c2f98ffdb9bd543a57c8...


you mean wrap mcp server in itself?


Hello HN!

I'm building Probe https://probeai.dev/ for a while now, and this this docs-mcp project is showcase of its capable. Giving you a local semantic search over any codebase or docs without indexing.

Feel free to ask any questions!


Nope, it is simply fresh issues with "help wanted" and "good first issue" labels.


I do maintain big OSS projects and and try to contribute as well.

However contribution experience can very bad, if you follow the path of picking the most famous objects. Good luck contributing to Node, Rust, Shadcn and etc - they do not need your contribution, their PR queue is overloaded and they can't handle it. Plus you need to get to their internal circles first, though quite complex process.

The world is much bigger. There are so many help required from the smaller but still active projects.

Just recently I raised 3 small PRs, and they reviewed the same day!

As a my respect to all the OSS community, I have build https://helpwanted.dev/ website, which in the nutshell shows latest "help wanted" and "good first issue" issues, from all over github in the last 24 hours.

You would be amazed how many cool projects out of there looking for the help!


One of the cases when AI not needed. There is very good working algorithm to extract content from the pages, one of implementations: https://github.com/buriy/python-readability


Some years ago I compared those boilerplate removal tools and I remember that jusText was giving me the best results out of the box (tried readability and few other libraries too). I wonder what is the state of the art today?


This is worth having a look at: https://mixmark-io.github.io/turndown/

With some configuration you can get most of the way there.


oh AI is optional here. I do use readability to clean the html before converting to .md.


Last time I tried readability it worked well with articles but struggled with other kinds of pages. Took away far more content than I wanted it to.


How do you achieve the same things without AI here using that tool?


"How do you do it without AI" is a question I (sadly) expect to see more often.


Feel free to answer then, how do you do the same functions this does with gpt(3/4) without AI?

Edit -

This is an excellent use of it, a free text human input capable of doing things like extracting summaries. It does not seem to be used at all for the basic task of extracting content, but for post filtering.


I think “copy from a PDF” could be improved with AI. It’s been 30 years and I still get new lines in the middle of sentences when I try to copy from one.


That's a great use case, you might be able to do this if you've got a copy and paste on the command line with

https://github.com/simonw/llm

In between. An alias like pdfwtf translating to "paste | llm command | copy"


i've long assumed that is a "feature" of PDF akin to DRM. Making copying text from a PDF makes sense from a publisher's standpoint.


Meh, it’s just the “how does it work?” question. How content extractors work is interesting and not obvious nor trivial.

And even when you see how readability parser works, AI handles most of the edge cases that content extractors fail on, so they are genuinely superseded by LLMs.


I was honestly expecting it to be mostly black magic, but it looks like the meat of the project is a bunch of (surely hard won) regexes. Nifty.


> I was … expecting it to be mostly black magic, but … the meat of the project is a bunch of … regexes

Wait, regexes are the epitome of black magic. What do you consider as black magic?


Macros? Any situation where code edits other code?

Sure, I could not write a regex engine, but the language itself can be fine if you keep it to straightfoward stuff. Unlike the famous e-mail parsing regex.


how is it compared to mozilla/readability?


it uses readibility but does some additional stuff like relink images to local paths etc., which I needed


I have had challenges with readability. The output is good for blogs but when we try it for other type of content, it misses on important details even when the page is quite text-heavy just like blog.


yeah that’s correct. i put a checkbox to disable readability filter if needed…


Plus in production, with high load, Redis cluster is way more common, which kind of solve single-threaded concern.


I've always found redis cluster to just bring problems with it.


What I really appreciate about Rails, is the strong vision, and not becoming another bloated framework for building "enterprise grade" applications. This is a 20-year-old framework, which does not afraid to radically change with time, and still being seen as Punk compared to rest

It was so heartwarming to remember my Rails story, its been 20 years ago since I started using it! https://twitter.com/buger/status/1723040883325460818


I love to see more activity in this area!

I'm maintainer of GoReplay https://github.com/buger/goreplay and work in this area for the last 10 years.

It is quite hard problem to solve, because you have to deal with state difference between test and production environments. Love your approach to mocking dependencies, and leveraging OpenTelementry. It potentially can solve some of state issues. But still require modifying user code. I wonder if it can be done purely using OpenTelementry (e.g. you depend on typical OTel setup), and then read the data directly from OTel DB.

Cheers!


Thanks Leonid! Your vote of confidence means a lot.

OTel for go requires user code changes. Languages that allow monkey-patching (java, js, python, etc.).

> I wonder if it can be done purely using OpenTelementry (e.g. you depend on typical OTel setup), and then read the data directly from OTel DB.

OTel doesn't work out of the box. OTel usually doesn't collect request or response for any network or db call. 90% of my time is spent on extending the individual agents' code; so that they can collect additional required information, and perform "replay".


Go replay has been one of the inspiration Leonid, so glad you checked out CodeParrot :)

Typical Otel implementation don’t capture some request data esp parameters and replay part is missing among few other issues, so we need to extend it.


thank you Leonid for the GoReplay. A great ecosystem of products will be built on top of it.


I hope so! But I also hope that I will be also able to monetise some of this movent. GoReplay dual licensed under AGPL and Commercial license. I also sell special appliance licenses.

If anyone in this thread wants to build a product based on GoReplay technology (capture network traffic directly, via AWS Traffic Mirroring or k8s), sent me message :)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: