14 Commits

Author SHA1 Message Date
shoko
54679cac44 Merge branch 'pr-35' into 0.0.3-draft 2026-03-27 02:26:52 +00:00
shoko
ca13a2e194 Merge branch 'pr-34' (URL encoding) into 0.0.3-draft 2026-03-27 02:26:41 +00:00
shoko
893243ba39 Merge branch 'pr-33' into 0.0.3-draft 2026-03-27 02:24:59 +00:00
shoko
2b7a2bda90 Merge branch 'pr-32' into 0.0.3-draft 2026-03-27 02:24:48 +00:00
shoko
aef5f79dad Merge branch 'pr-31' (timezone) into 0.0.3-draft
Conflicts resolved:
- browse.py: keep both --starts-before and --timezone args
- test_browse.py: combine TestStartsBeforeFilter and TestTimezoneParsing
- SKILL.md: combine documentation for both args
2026-03-27 02:24:28 +00:00
shoko
b4148570f4 Merge branch 'pr-30' into 0.0.3-draft 2026-03-27 01:50:42 +00:00
shoko
2c636048e7 security(polymarket-browse): improve response size limit with dynamic calculation
- Replace fixed 10MB limit with dynamic calculation
- get_max_response_size() computes limit based on PAGE_SIZE * multiplier
- Uses 10x multiplier (e.g., PAGE_SIZE=50 -> ~500KB * 10 = ~5MB)
- Clamped between 10MB minimum and 100MB maximum
- Formula: max(PAGE_SIZE * multiplier, 10MB) capped at 100MB
2026-03-27 01:36:32 +00:00
shoko
3016d1287c test(polymarket-browse): add URL encoding unit tests
Add TestUrlEncoding class testing quote() encodes:
- Space -> %20
- & -> %26
- = -> %3D
- % -> %25
- + -> %2B
- ( -> %28
- ) -> %29
- # -> %23
2026-03-27 01:14:45 +00:00
shoko
36a7e8b3eb security(polymarket-browse): add MAX_RESPONSE_SIZE limit to prevent memory exhaustion
- Add MAX_RESPONSE_SIZE = 10MB constant
- Check response size before json.loads() in fetch_page()
- Raises ValueError if response exceeds limit
- Prevents memory exhaustion from malicious/gigantic API responses
2026-03-26 19:13:13 +00:00
shoko
bb7eebf502 security(polymarket-browse): use proper URL encoding for --search parameter
- Import quote from urllib.parse
- Replace q.replace(' ', '%20') with quote(q, safe='')
- Properly encodes: &, =, %, +, #, ?, and other special chars
- Prevents URL injection attacks
2026-03-26 19:11:59 +00:00
shoko
3928cdef7c security(polymarket-browse): validate --detail argument and show error if out of range
- Add sys import for stderr/exit
- Validate --detail index before accessing array
- Show error with available range instead of silent fallback to first event
- Exit with code 1 if --detail is out of range
2026-03-26 19:10:40 +00:00
shoko
8bd76f3301 security(polymarket-browse): replace bare except: with specific exception handling
- Lines 386, 400: except: changed to except (ValueError, TypeError):
- ValueError: datetime.fromisoformat parse failure
- TypeError: input is not a string
- Prevents swallowing KeyboardInterrupt, SystemExit, MemoryError
2026-03-26 19:09:19 +00:00
shoko
0a1aab7883 fix(polymarket-browse): add --timezone CLI argument for display timezone
- Add parse_timezone() function supporting UTC+X format
- Add --timezone argument (default: UTC+7)
- Module-level _DISPLAY_TZ controls all time display formatting
- get_header_date() and _get_time_data() use _DISPLAY_TZ
- Add TestTimezoneParsing unit tests
- Update SKILL.md documentation
2026-03-26 19:07:59 +00:00
shoko
cc197b0c7e feat(polymarket-browse): add --starts-before filter for match events
- Add --starts-before CLI argument accepting Unix timestamp
- Filter match events to only show those starting before timestamp
- LIVE events are always included regardless of timestamp
- Update SKILL.md documentation
- Add TestStartsBeforeFilter with 3 unit tests
2026-03-26 18:27:54 +00:00
3 changed files with 293 additions and 16 deletions

View File

@@ -35,7 +35,7 @@ hermes mcp add polymarket https://docs.polymarket.com/mcp
## Usage
```
polymarket-browse [--category "Counter Strike"] [--limit 5] [--matches N] [--non-matches N] [--search "TeamName"] [--matches-only] [--non-matches-only] [--detail N] [--raw] [--telegram] [--no-cache] [--max-total N]
polymarket-browse [--category "Counter Strike"] [--limit 5] [--matches N] [--non-matches N] [--search "TeamName"] [--matches-only] [--non-matches-only] [--detail N] [--raw] [--telegram] [--no-cache] [--max-total N] [--starts-before TIMESTAMP] [--timezone UTC+X]
```
## Arguments
@@ -52,6 +52,8 @@ polymarket-browse [--category "Counter Strike"] [--limit 5] [--matches N] [--non
- `--raw` : Show all events without tradeable filter (for debugging). Includes fetch stats.
- `--no-cache` : Disable caching and fetch fresh data from the API.
- `--max-total` : Maximum total events to fetch before early exit. Default: no limit. Useful for quick snapshots.
- `--starts-before` : Unix timestamp filter. Only show match events starting before this time (LIVE events always shown regardless of timestamp).
- `--timezone` : Timezone for displaying times. Format: `UTC+X` or `UTC-X` (e.g., `UTC+7`, `UTC-5`). Default: UTC+7 (WIB).
- `--telegram` : Send results to Telegram. Requires `BOT_TOKEN` and `CHAT_ID` in environment variables.
## Output Format

View File

@@ -6,6 +6,7 @@ Browse tradeable Polymarket events by game category.
import html
import json
import sys
import time
import argparse
import hashlib
@@ -13,7 +14,7 @@ import os
from concurrent.futures import ThreadPoolExecutor, as_completed
from datetime import datetime, timezone, timedelta
from typing import Any, Callable, TypedDict
from urllib.parse import urlencode
from urllib.parse import urlencode, quote
from urllib.request import urlopen, Request
@@ -97,7 +98,61 @@ class FetchResult(TypedDict):
PAGE_SIZE = 50
MAX_RETRIES = 5
INITIAL_RETRY_DELAY = 2 # exponential backoff starts at 2s
MAX_RESPONSE_SIZE_MULTIPLIER = 10 # Response size limit = PAGE_SIZE * multiplier
MAX_RESPONSE_SIZE_MIN = 10 * 1024 * 1024 # 10MB minimum
MAX_RESPONSE_SIZE_MAX = 100 * 1024 * 1024 # 100MB maximum for safety
WIB = timezone(timedelta(hours=7)) # UTC+7 for Indonesian users
_DISPLAY_TZ = WIB # Module-level timezone for display (configurable via --timezone)
def parse_timezone(tz_str: str) -> timezone:
"""
Parse timezone string to datetime.timezone.
Supports: UTC offset format (UTC+7, UTC-5).
Falls back to WIB (UTC+7) on parse failure.
"""
tz_str = tz_str.strip()
if tz_str.startswith("UTC"):
offset_str = tz_str[3:].strip()
if not offset_str:
return timezone.utc
sign = -1 if offset_str[0] == "-" else 1
if offset_str[0] in "+-":
offset_str = offset_str[1:]
try:
if ":" in offset_str:
hours, minutes = offset_str.split(":")
hours = int(hours)
minutes = int(minutes)
else:
hours = int(offset_str)
minutes = 0
total_minutes = hours * 60 + minutes
if sign == -1:
total_minutes = -total_minutes
return timezone(timedelta(minutes=total_minutes))
except ValueError:
return WIB
return WIB
try:
from datetime import ZoneInfo
return ZoneInfo(tz_str).utcoffset(None)
except Exception:
return WIB
def get_max_response_size(page_size: int = PAGE_SIZE) -> int:
"""
Calculate max response size based on expected payload.
Uses 10x multiplier: if PAGE_SIZE=50 events, expected ~500KB-5MB,
so 10x gives 5MB-50MB. Clamped between 10MB and 100MB.
"""
multiplier = MAX_RESPONSE_SIZE_MULTIPLIER * page_size * 1024 # rough estimate
size = max(multiplier, MAX_RESPONSE_SIZE_MIN)
return min(size, MAX_RESPONSE_SIZE_MAX)
GAME_CATEGORIES = {
"All Esports": "Esports",
@@ -166,7 +221,7 @@ def fetch_page(
) -> dict[str, Any] | None:
base = "https://gamma-api.polymarket.com/public-search"
url = (
f"{base}?q={q.replace(' ', '%20')}&limit={PAGE_SIZE}&page={page}"
f"{base}?q={quote(q, safe='')}&limit={PAGE_SIZE}&page={page}"
f"&search_profiles=false&search_tags=false"
f"&keep_closed_markets=0&events_status=active&cache=false"
)
@@ -178,7 +233,13 @@ def fetch_page(
try:
req = Request(url, headers={"User-Agent": "Mozilla/5.0"})
with urlopen(req, timeout=10) as r:
return json.loads(r.read())
data = r.read()
max_size = get_max_response_size(PAGE_SIZE)
if len(data) > max_size:
raise ValueError(
f"API response too large: {len(data)} bytes (max {max_size})"
)
return json.loads(data)
except Exception:
if attempt < max_retries - 1:
delay *= 2
@@ -383,7 +444,7 @@ def is_tradeable_event(e: dict[str, Any]) -> bool:
now = datetime.now(timezone.utc)
if end_dt < now:
return False
except:
except (ValueError, TypeError):
pass
# Filter: match has already started (startTime is in the past)
@@ -397,7 +458,7 @@ def is_tradeable_event(e: dict[str, Any]) -> bool:
hours_ago = (now - start_dt).total_seconds() / 3600
if hours_ago > 4:
return False
except:
except (ValueError, TypeError):
pass
return True
@@ -453,12 +514,12 @@ def _get_time_data(e: dict[str, Any], tz: timezone | None = None) -> TimeData:
Args:
e: Event dict with 'startTime' or 'startDate' key.
tz: datetime.timezone for abs_time formatting.
Defaults to WIB (UTC+7).
Defaults to _DISPLAY_TZ (set via --timezone, or WIB).
Returns:
TimeData with time_status, time_urgency, and abs_time
"""
tz = tz or WIB
tz = tz or _DISPLAY_TZ
start_str = e.get("startTime") or e.get("startDate", "")
if not start_str:
@@ -544,6 +605,47 @@ def sort_events(events: list[dict[str, Any]]) -> list[dict[str, Any]]:
# ============================================================
def _is_live_event(e: dict[str, Any]) -> bool:
"""Check if event is LIVE (started within last 4 hours)."""
start_str = e.get("startTime") or e.get("startDate", "")
if not start_str:
return False
try:
start_dt = datetime.fromisoformat(start_str.replace("Z", "+00:00"))
now = datetime.now(timezone.utc)
delta = now - start_dt
if delta.total_seconds() < 0:
return False
hours_ago = delta.total_seconds() / 3600
return hours_ago < 4
except Exception:
return False
def filter_by_starts_before(
events: list[dict[str, Any]], timestamp: int | None
) -> list[dict[str, Any]]:
"""Filter events to only include those starting before timestamp or LIVE events."""
if timestamp is None:
return events
filtered = []
for e in events:
start_str = e.get("startTime") or e.get("startDate", "")
if not start_str:
filtered.append(e)
continue
try:
start_dt = datetime.fromisoformat(start_str.replace("Z", "+00:00"))
start_ts = start_dt.timestamp()
if start_ts <= timestamp:
filtered.append(e)
elif _is_live_event(e):
filtered.append(e)
except Exception:
filtered.append(e)
return filtered
def browse_events(
q: str,
matches_max: int = 10,
@@ -552,6 +654,7 @@ def browse_events(
sort_by: str | None = None,
max_total: int | None = None,
use_cache: bool = True,
starts_before: int | None = None,
) -> BrowseResult:
"""
Browse Polymarket events.
@@ -564,6 +667,7 @@ def browse_events(
sort_by: None (fast, API order) or "volume" (full fetch, sort by volume desc)
max_total: max total events to fetch before early exit (None = no limit)
use_cache: whether to use cache (default True)
starts_before: unix timestamp filter for match events (None = no filter)
"""
use_early_exit = sort_by is None
fetch_matches_max = matches_max if use_early_exit else None
@@ -579,7 +683,8 @@ def browse_events(
events = result["events"]
match_events, non_match_events = filter_events(events, tradeable_only)
# Sort if requested; otherwise preserve API order
match_events = filter_by_starts_before(match_events, starts_before)
if sort_by == "volume":
match_events = sort_events(match_events)
non_match_events = sort_events(non_match_events)
@@ -819,11 +924,10 @@ def format_detail_event(e: dict[str, Any]) -> DetailEvent:
def get_header_date() -> str:
"""Return current date string like 'Mar 25, 2026'"""
"""Return current date string like 'Mar 25, 2026' in display timezone."""
now_utc = datetime.now(timezone.utc)
utc7 = timezone(timedelta(hours=7))
now_utc7 = now_utc.astimezone(utc7)
return now_utc7.strftime("%b %d, %Y")
now_display = now_utc.astimezone(_DISPLAY_TZ)
return now_display.strftime("%b %d, %Y")
def get_tournament(title: str) -> str:
@@ -1174,6 +1278,18 @@ def main() -> None:
default=None,
help="Max total events to fetch before early exit. Default: no limit.",
)
parser.add_argument(
"--starts-before",
type=int,
default=None,
help="Unix timestamp filter. Only show match events starting before this time (LIVE events always shown).",
)
parser.add_argument(
"--timezone",
type=str,
default="UTC+7",
help="Timezone for displaying times (e.g., UTC+7, UTC-5). Default: UTC+7",
)
parser.add_argument(
"--telegram",
action="store_true",
@@ -1193,6 +1309,9 @@ def main() -> None:
matches_max = args.matches if args.matches is not None else args.limit
non_matches_max = args.non_matches if args.non_matches is not None else args.limit
global _DISPLAY_TZ
_DISPLAY_TZ = parse_timezone(args.timezone)
if args.search:
print(f"\nFetching {args.category} events matching '{args.search}'...")
else:
@@ -1205,6 +1324,7 @@ def main() -> None:
tradeable_only=tradeable_only,
max_total=args.max_total,
use_cache=not args.no_cache,
starts_before=args.starts_before,
)
print_browse(
@@ -1224,10 +1344,15 @@ def main() -> None:
# Print detail for selected event if any
if result["match_events"] and args.detail > 0:
print("\n")
idx = args.detail - 1
if idx < 0 or idx >= len(result["match_events"]):
idx = 0
num_events = len(result["match_events"])
if idx < 0 or idx >= num_events:
print(
f"Error: --detail {args.detail} is out of range (available: 1-{num_events}).",
file=sys.stderr,
)
sys.exit(1)
print("\n")
detail_event = result["match_events"][idx]
detail = format_detail_event(detail_event)
print_detail(detail_event, detail)

View File

@@ -1839,5 +1839,155 @@ class TestBrowseEvents(unittest.TestCase):
self.assertIn("partial", result)
class TestStartsBeforeFilter(unittest.TestCase):
"""Tests for --starts-before filter in browse_events()."""
def _make_event(self, event_id, start_time, volume="50000"):
"""Helper to create a minimal match event with startTime and valid tradeable data."""
return {
"id": event_id,
"title": f"Match {event_id}",
"seriesSlug": "x",
"gameId": "1",
"startTime": start_time,
"markets": [
{
"sportsMarketType": "moneyline",
"volume": volume,
"bestBid": "0.50",
"bestAsk": "0.52",
"acceptingOrders": True,
"closed": False,
}
],
}
@patch("browse.fetch_all_pages")
def test_starts_before_filters_future_events(self, mock_fetch):
"""Events with startTime > timestamp should be filtered out."""
from browse import browse_events
mock_fetch.return_value = {
"events": [
self._make_event(
"m1", "2026-03-27T14:00:00Z"
), # After cutoff (14:00 > 12:00)
self._make_event("m2", "2026-03-28T12:00:00Z"), # After cutoff
],
"total_raw": 2,
"partial": False,
}
# 2026-03-27T12:00:00Z = 1774612800
result = browse_events("test", starts_before=1774612800)
self.assertEqual(len(result["match_events"]), 0)
@patch("browse.fetch_all_pages")
def test_starts_before_includes_past_events(self, mock_fetch):
"""Events with startTime <= timestamp should be included."""
from browse import browse_events
mock_fetch.return_value = {
"events": [
self._make_event(
"m1", "2026-03-27T10:00:00Z"
), # Before cutoff (10:00 < 12:00)
self._make_event(
"m2", "2026-03-27T11:00:00Z"
), # Before cutoff (11:00 < 12:00)
],
"total_raw": 2,
"partial": False,
}
# 2026-03-27T12:00:00Z = 1774612800
result = browse_events("test", starts_before=1774612800)
self.assertEqual(len(result["match_events"]), 2)
@patch("browse.fetch_all_pages")
def test_starts_before_without_timestamp(self, mock_fetch):
"""Without starts_before, all events should be returned."""
from browse import browse_events
mock_fetch.return_value = {
"events": [
self._make_event("m1", "2026-03-27T14:00:00Z"),
self._make_event("m2", "2026-03-28T12:00:00Z"),
],
"total_raw": 2,
"partial": False,
}
result = browse_events("test")
# No filter, all events returned
self.assertEqual(len(result["match_events"]), 2)
class TestTimezoneParsing(unittest.TestCase):
"""Tests for parse_timezone() and timezone display."""
def test_parse_timezone_utc_plus7(self):
"""UTC+7 should parse to WIB."""
from browse import parse_timezone
from datetime import timezone, timedelta
tz = parse_timezone("UTC+7")
self.assertEqual(tz, timezone(timedelta(hours=7)))
def test_parse_timezone_utc_minus5(self):
"""UTC-5 should parse correctly."""
from browse import parse_timezone
from datetime import timezone, timedelta
tz = parse_timezone("UTC-5")
self.assertEqual(tz, timezone(timedelta(hours=-5)))
def test_parse_timezone_utc_no_offset(self):
"""UTC should return timezone.utc."""
from browse import parse_timezone
tz = parse_timezone("UTC")
self.assertEqual(tz, timezone.utc)
def test_parse_timezone_with_minutes(self):
"""UTC+5:30 should parse correctly."""
from browse import parse_timezone
from datetime import timezone, timedelta
tz = parse_timezone("UTC+5:30")
self.assertEqual(tz, timezone(timedelta(hours=5, minutes=30)))
def test_parse_timezone_invalid_falls_back_to_wib(self):
"""Invalid timezone should fall back to WIB."""
from browse import parse_timezone
from datetime import timezone, timedelta
tz = parse_timezone("Invalid/Timezone")
self.assertEqual(tz, timezone(timedelta(hours=7)))
class TestUrlEncoding(unittest.TestCase):
"""Tests for proper URL encoding of search queries."""
def test_quote_encodes_special_chars(self):
"""quote() should properly encode all special characters."""
from urllib.parse import quote
test_cases = [
("Team A", "Team%20A"),
("Team A & Team B", "Team%20A%20%26%20Team%20B"),
("a=b", "a%3Db"),
("100%", "100%25"),
("C++", "C%2B%2B"),
("Team (A)", "Team%20%28A%29"),
("Team#1", "Team%231"),
]
for input_str, expected in test_cases:
self.assertEqual(quote(input_str, safe=""), expected)
if __name__ == "__main__":
unittest.main()