byted-link-reader
BytedLinkReader
适用场景
当需要获取网页、PDF 或抖音视频的标题和正文内容时,使用该技能调用 link_reader 函数。
认证与凭据来源
- 用户首次使用本技能时,先检查是否配置了
ARK_API_KEY和MODEL_AGENT_API_KEY环境变量,如未配置,提示用户需要提供ARK_API_KEY,并把它写入到workspace下的环境变量文件中,如果文件已存在则追加到末尾,确保环境变量格式正确,使这个环境变量生效 - 该技能默认使用
ARK_API_KEY环境变量 - 若未配置,将尝试使用
MODEL_AGENT_API_KEY环境变量。
使用步骤
- 准备 URL 列表。
- 运行脚本
python scripts/link_reader.py "url1" "url2" ...。运行之前cd到对应的目录。 - 返回结果包含每个 URL 的标题和内容。
输出格式
- JSON 格式的列表,每个元素包含 URL 对应的标题和内容。
错误处理
- 如果出现'PermissionError: ARK_API_KEY or MODEL_AGENT_API_KEY is not set in environment variables.'错误,则提示用户需要提供
ARK_API_KEY或MODEL_AGENT_API_KEY,并把它们写入到workspace下的环境变量文件中,如果文件已存在则追加到末尾,确保环境变量格式正确,使这个环境变量生效,并重试刚才失败的读取。 - 再次使用本技能时,优先从环境变量中读取
ARK_API_KEY或MODEL_AGENT_API_KEY
示例
python scripts/link_reader.py "https://example.com"
More from bytedance/agentkit-samples
byted-seedream-image-generate
Generate high-quality images from text prompts using Volcano Engine Seedream models. Supports multiple artistic styles and aspect ratios. Use this skill when users want to create images from text descriptions, generate artwork in various styles, create visual content for creative projects, or need AI-powered image generation capabilities.
183byted-las-video-edit
Extracts and clips video segments from long videos using natural language descriptions. AI-powered smart video editing, video trimming, and video cutting powered by Volcengine LAS. Describe what you want — scenes, people, objects, actions, events — and get trimmed clips automatically. Video search and video content retrieval: find and locate specific people, objects, or scenes in footage. Supports reference images for person matching and object matching (search video by image). Two modes: simple (fast) and detail (thorough, optional ASR). Use this skill when the user wants to edit/clip/cut videos using natural language descriptions, extract highlights or key moments from videos, find specific people/objects/scenes in video footage (by text or reference image), compile highlight reels from long videos, trim video segments, or do AI-powered smart video editing.
163byted-las-pdf-parse-doubao
Parses and reads PDF documents into structured Markdown text using Volcengine LAS Doubao AI models. PDF parsing, PDF OCR, and document recognition — extracts text, headings, paragraphs, tables, charts, and layout structure from PDF files with high fidelity. Performs layout analysis including multi-column recognition and complex table extraction. Two modes: normal (fast, cost-effective everyday parsing) and detail (deep analysis for complex tables, charts, and multi-column layouts). Converts PDF to Markdown, PDF to text, and structured data. Digitizes scanned PDF documents and scanned images via OCR. Supports TOS paths, HTTP URLs, and local file upload. Async submit-poll workflow with batch processing support. Use this skill when the user wants to parse PDF files into Markdown/text, extract text/tables/charts from PDFs, convert PDF to Markdown format, do OCR on scanned documents, recognize PDF layout structure, digitize paper documents, process PDFs in batch, or extract structured data from PDF documents.
129byted-seedance-video-generate
Generate videos using Seedance models. Invoke when user wants to create videos from text prompts, images, or reference materials.
109byted-data-search
|
106byted-las-vlm-video
Analyzes and understands video content using Volcengine LAS Doubao vision-language models (VLM). Multimodal AI video analysis, video comprehension, and visual understanding of video clips and footage. Performs video question answering (video Q&A) — ask questions about what happens in a video and get AI answers. Scene recognition and scene description, object recognition and object detection, action recognition and action detection from video frames. Generates video descriptions, video captions, video summaries, video annotations, and content summarization. Visual frame analysis for identifying people, objects, actions, and events in video. Auto-compresses video to 50MB before inference. Synchronous single-call processing. Use this skill when the user wants to analyze or understand video content using VLM/AI, do video Q&A (ask questions about a video), describe what happens in a video, recognize objects/actions/scenes in video frames, generate video captions/descriptions/summaries, annotate or label video content, get AI-powered visual understanding of video clips, or perform multimodal video analysis with vision-language models.
97