Skip to main content
0
t

the0807/GeekNews-MCP-Server

the0807/GeekNews-MCP-Server

Rating

0.0

Votes

0

score

Downloads

0

total

Price

Free

No login needed

Works With

Claude CodeCursorWindsurfVS CodeDeveloper tool

About

[](https://smithery.ai/server/@the0807/geeknews-mcp-server)

이 프로젝트는 GeekNews에서 아티클을 가져오는 Model Context Protocol(MCP) 서버입니다. Python으로 구현되었으며, BeautifulSoup을 사용하여 웹 스크래핑을 수행합니다. 서버 부하를 줄이기 위해 하루에 한 번 자동으로 데이터를 파싱하고 캐시에 저장하는 기능을 제공합니다.

기능

  • get_articles 도구: GeekNews에서 아티클을 가져오는 기능
  • 아티클 유형(top, new, ask, show)과 반환할 아티클 수를 지정할 수 있음
  • 각 응답에는 제목, URL, 포인트, 작성자, 시간, 댓글 수, 순위 정보가 포함됨
  • 캐시된 데이터를 사용하여 서버 부하 감소

Works with Claude (desktop and mobile), Cursor, Windsurf, VS Code, and any MCP-compatible AI app.

Category: Developer Tools

Don't lose this

Three weeks from now, you'll want the0807/GeekNews-MCP-Server again. Will you remember where to find it?

Save it to your library and the next time you need the0807/GeekNews-MCP-Server, it’s one tap away — from any AI app you use. Group it into a bench with the rest of the team for that kind of task and you can pull the whole stack at once.

⚡ Pro tip for geeks: add a-gnt 🤵🏻‍♂️ as a custom connector in Claude or a custom GPT in ChatGPT — one click and your library is right there in the chat. Or, if you’re in an editor, install the a-gnt MCP server and say “use my [bench name]” in Claude Code, Cursor, VS Code, or Windsurf.

🤵🏻‍♂️

a-gnt's Take

Our honest review

This plugs directly into your AI and gives it new abilities it didn't have before. the0807/GeekNews-MCP-Server. Once connected, just ask your AI to use it. It's completely free and works across most major AI apps. This one just landed in the catalog — worth trying while it's fresh.

Tips for getting started

1

Tap "Get" above, pick your AI app, and follow the steps. Most installs take under 30 seconds.

What's New

Version 1.0.06 days ago

Imported from awesome:punkpeye/awesome-mcp-servers

Ratings & Reviews

0.0

out of 5

0 ratings

No reviews yet. Be the first to share your experience.