# vision-agent **Repository Path**: fengying11/vision-agent ## Basic Information - **Project Name**: vision-agent - **Description**: No description available - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: add-new-tools - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-02-10 - **Last Updated**: 2025-02-10 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README
VisionAgent [![](https://dcbadge.vercel.app/api/server/wPdN8RCYew?compact=true&style=flat)](https://discord.gg/wPdN8RCYew) ![ci_status](https://github.com/landing-ai/vision-agent/actions/workflows/ci_cd.yml/badge.svg) [![PyPI version](https://badge.fury.io/py/vision-agent.svg)](https://badge.fury.io/py/vision-agent) ![version](https://img.shields.io/pypi/pyversions/vision-agent)
## VisionAgent VisionAgent is a library that helps you utilize agent frameworks to generate code to solve your vision task. Check out our discord for updates and roadmaps! The fastest way to test out VisionAgent is to use our web application which you can find [here](https://va.landing.ai/). ## Installation ```bash pip install vision-agent ``` ```bash export ANTHROPIC_API_KEY="your-api-key" export OPENAI_API_KEY="your-api-key" ``` > **_NOTE:_** We found using both Anthropic Claude-3.5 and OpenAI o1 to be provide the best performance for VisionAgent. If you want to use a different LLM provider or only one, see 'Using Other LLM Providers' below. ## Documentation [VisionAgent Library Docs](https://landing-ai.github.io/vision-agent/) ## Examples ### Counting cans in an image You can run VisionAgent in a local Jupyter Notebook [Counting cans in an image](https://github.com/landing-ai/vision-agent/blob/main/examples/notebooks/counting_cans.ipynb) ### Generating code You can use VisionAgent to generate code to count the number of people in an image: ```python from vision_agent.agent import VisionAgentCoderV2 from vision_agent.models import AgentMessage agent = VisionAgentCoderV2(verbose=True) code_context = agent.generate_code( [ AgentMessage( role="user", content="Count the number of people in this image", media=["people.png"] ) ] ) with open("generated_code.py", "w") as f: f.write(code_context.code + "\n" + code_context.test) ``` ### Using the tools directly VisionAgent produces code that utilizes our tools. You can also use the tools directly. For example if you wanted to detect people in an image and visualize the results: ```python import vision_agent.tools as T import matplotlib.pyplot as plt image = T.load_image("people.png") dets = T.countgd_object_detection("person", image) # visualize the countgd bounding boxes on the image viz = T.overlay_bounding_boxes(image, dets) # save the visualization to a file T.save_image(viz, "people_detected.png") # display the visualization plt.imshow(viz) plt.show() ``` You can also use the tools for running on video files: ```python import vision_agent.tools as T frames_and_ts = T.extract_frames_and_timestamps("people.mp4") # extract the frames from the frames_and_ts list frames = [f["frame"] for f in frames_and_ts] # run the countgd tracking on the frames tracks = T.countgd_sam2_video_tracking("person", frames) # visualize the countgd tracking results on the frames and save the video viz = T.overlay_segmentation_masks(frames, tracks) T.save_video(viz, "people_detected.mp4") ``` ## Using Other LLM Providers You can use other LLM providers by changing `config.py` in the `vision_agent/configs` directory. For example to change to Anthropic simply just run: ```bash cp vision_agent/configs/anthropic_config.py vision_agent/configs/config.py ``` > **_NOTE:_** VisionAgent moves fast and we are constantly updating and changing the library. If you have any questions or need help, please reach out to us on our discord channel.